cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

vignesh22
by New Contributor
  • 1707 Views
  • 2 replies
  • 0 kudos

Pipelines are expected to have at least one table Error While running DLT pipeline

Error :Pipelines are expected to have at least one table defined butno tables were found in your pipeline I wrote simple code as phase 1 debug%sql CREATE OR REFRESH STREAMING TABLE test_table AS SELECT "hello" as greeting; Can u plz help what's wrong...

  • 1707 Views
  • 2 replies
  • 0 kudos
Latest Reply
RiyazAliM
Honored Contributor
  • 0 kudos

Hey @vignesh22 - Adding to what @Takuya-Omi san has mentioned - the instantiation of streaming table in your definition is incorrect. You're trying to create a stream table using a batch source which will result in the DLT Analysis Exception as descr...

  • 0 kudos
1 More Replies
saikrishna1020
by New Contributor
  • 940 Views
  • 1 replies
  • 1 kudos

Community Edition Data recovery

I was using Databricks Community Edition for some practice work, and I had created a few notebooks as part of my learning. However, when I recently tried to log in, I received a message saying, "We were not able to find a Community Edition." Now, non...

  • 940 Views
  • 1 replies
  • 1 kudos
Latest Reply
Brahmareddy
Esteemed Contributor
  • 1 kudos

Hi saikrishna1020,How are you doing today? I totally understand how upsetting it can be to lose work you’ve put effort into. With Databricks Community Edition, unfortunately, inactivity for an extended period (usually 14–30 days) can cause the worksp...

  • 1 kudos
dplatform_user
by New Contributor II
  • 987 Views
  • 1 replies
  • 0 kudos

INVALID_PARAMETER_VALUE.LOCATION_OVERLAP when trying to copy from s3 location

Hi,Currently we are getting an issue when we try to copy a file from s3 location using dbutils.fs.cp, please see example below:source = s3://test-bucket/external/zones/{database_name}/{table_name}/test.csvdestination = s3://test-bucket/external/desti...

  • 987 Views
  • 1 replies
  • 0 kudos
Latest Reply
Brahmareddy
Esteemed Contributor
  • 0 kudos

Hi dplatform_user,How are you doing today?, As per my understanding, this error is actually a common one when working with external storage paths that overlap with Unity Catalog-managed locations. The error message is basically saying that your sourc...

  • 0 kudos
Mado
by Valued Contributor II
  • 20196 Views
  • 4 replies
  • 0 kudos

Resolved! How to enforce delta table column to have unique values?

Hi,I have defined a delta table with a primary key:%sql   CREATE TABLE IF NOT EXISTS test_table_pk ( table_name STRING NOT NULL, label STRING NOT NULL, table_location STRING NOT NULL,   CONSTRAINT test_table_pk_col PRIMARY KEY(table_name) ...

image
  • 20196 Views
  • 4 replies
  • 0 kudos
Latest Reply
SibbirSihan
New Contributor II
  • 0 kudos

CREATE TABLE table_name (id_col1 BIGINT GENERATED ALWAYS AS IDENTITY,id_col2 BIGINT GENERATED ALWAYS AS IDENTITY (START WITH -1 INCREMENT BY 1),id_col3 BIGINT GENERATED BY DEFAULT AS IDENTITY,id_col4 BIGINT GENERATED BY DEFAULT AS IDENTITY (START WIT...

  • 0 kudos
3 More Replies
Punit_Prajapati
by Databricks Partner
  • 3180 Views
  • 3 replies
  • 6 kudos

Resolved! SERVERLESS SQL WAREHOUSE

Hello All,I have two questions regarding the serverless SQL warehouse which are following:1.) If I create a small Serverless SQL Warehouse in Databricks that shows 12 DBUs/hour, will I be charged 12 DBUs even if I don’t run any queries in that hour? ...

  • 3180 Views
  • 3 replies
  • 6 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 6 kudos

Shua42 hits the nail on the head. If I can be so bold as to summarize: You are only charged when the Warehouse is running regardless of how much or how little you use it.  We do have an auto stop feature you can configure. Essentially, you set a time...

  • 6 kudos
2 More Replies
640913
by New Contributor III
  • 14921 Views
  • 3 replies
  • 1 kudos

%pip install requirements.txt - path not found

Hi everyone,I was just testing things out to come up with a reasonable way of working with version management in DB and was inspired by the commands specified here. For my team and I, it makes no sense to put the requirements file in the dbfs locatio...

  • 14921 Views
  • 3 replies
  • 1 kudos
Latest Reply
Rajat-TVSM
New Contributor III
  • 1 kudos

Hello, were you able to find a solution to this?

  • 1 kudos
2 More Replies
notwarte
by New Contributor III
  • 2052 Views
  • 4 replies
  • 0 kudos

Unity Catalog storage amounts

Hi,I am using Azure and I do have predictive optimization enable on the catalog. I have wrote a script to calculate the data amounts of all of the tables -  looping over all of the tables and running "describe detail".All of the tables amount to ~ 1....

wiselka_1-1744630645723.png
  • 2052 Views
  • 4 replies
  • 0 kudos
Latest Reply
Isi
Honored Contributor III
  • 0 kudos

Hey @notwarte,Using the __databricks_internal catalog to trace the underlying storage location is a solid approach for investigating their footprint.Regarding your question about storage duplication: yes, materialized views in Databricks do store a p...

  • 0 kudos
3 More Replies
Malthe
by Valued Contributor II
  • 987 Views
  • 1 replies
  • 0 kudos

Parametrize DLT pipeline

If I'm using Databricks Asset Bundles, how would I parametrize a DLT pipeline based on a static configuration file.In pseudo-code, I would have a .py-file:import dlt # Something that pulls a pipeline resource (or artifact) and parses from JSON table...

  • 987 Views
  • 1 replies
  • 0 kudos
Latest Reply
Emmitt18Lefebvr
New Contributor II
  • 0 kudos

Hello!To parametrize a Databricks DLT pipeline with a static configuration file using Asset Bundles, include your JSON/YAML config file in the bundle. In your DLT pipeline code, read this file using Python's file I/O (referencing its deployed path). ...

  • 0 kudos
dc-rnc
by Contributor
  • 4460 Views
  • 3 replies
  • 2 kudos

Resolved! DAB | Set tag based on job parameter

Hi Community.Since I wasn't able to find a way to set a job tag dynamically at runtime based on a parameter that is passed to the job, I was wondering if it is possible or if there is an equivalent way to do it.Thank you. Regards.

  • 4460 Views
  • 3 replies
  • 2 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 2 kudos

Based on the provided context, it appears that there isn't a direct way within Databricks to dynamically set job tags at runtime based on a parameter passed to the job. However, there are alternative approaches you can consider to work around this li...

  • 2 kudos
2 More Replies
cs_de
by New Contributor II
  • 2095 Views
  • 4 replies
  • 3 kudos

How do I deploy or run one job if I have multiple jobs in a Databricks Asset Bundle?

How do I deploy or run a single job if I have 2 or more jobs defined in my asset bundle?$databricks bundle deploy job1 #does not work I do not see a flag to identify what job to run.

  • 2095 Views
  • 4 replies
  • 3 kudos
Latest Reply
mark_ott
Databricks Employee
  • 3 kudos

I haven't done it with multiple jobs, but I think under resources you name multiple jobs, then when you deploy you just call that job key.  

  • 3 kudos
3 More Replies
Chris_sh
by Databricks Partner
  • 4645 Views
  • 2 replies
  • 1 kudos

[STREAMING_TABLE_OPERATION_NOT_ALLOWED.REQUIRES_SHARED_COMPUTE]

Currently trying to refresh a Delta Live Table using a Full Refresh but an error keeps coming up saying that we have to use a shared cluster or a SQL warehouse. I've tried both a shared cluster and a SQL warehouse and the same error keeps coming up. ...

  • 4645 Views
  • 2 replies
  • 1 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 1 kudos

You are not using "No Isolation Shared" mode, right?  Also, can you share the chunk of code that is causing the failure? Thanks, Louis.

  • 1 kudos
1 More Replies
guest0
by New Contributor III
  • 2353 Views
  • 6 replies
  • 3 kudos

Spark UI Simulator Not Accessible

Hello,The Spark UI Simulator is not accessible since the last few days. I was able to refer to it last week, at https://www.databricks.training/spark-ui-simulator/index.html. I already have access to partner academy (if that is any relevant).  <Error...

Data Engineering
simulator
spark-ui
  • 2353 Views
  • 6 replies
  • 3 kudos
Latest Reply
guest0
New Contributor III
  • 3 kudos

Just a short update: the request I raised was closed saying there is no active support contract with the org (from the email I used) to look into this. Perhaps someone else could try raising a request using the link above.

  • 3 kudos
5 More Replies
Vasu_Kumar_T
by Databricks Partner
  • 1464 Views
  • 3 replies
  • 1 kudos

Data Migration using Bladebridge

Hi,We are planning to migrate from Teradata to Databricks using Bladebridge. Going through various portals, I am not able to conclude the component that facilitates Data movement between Teradata and databricks.Please clarify end to end tool and acti...

  • 1464 Views
  • 3 replies
  • 1 kudos
Latest Reply
RiyazAliM
Honored Contributor
  • 1 kudos

I'm not aware if blade bridge has data movement tool handy with them.I don't see anything advertised by them though.Let me know if you find anything on this.

  • 1 kudos
2 More Replies
yashojha1995
by New Contributor
  • 1027 Views
  • 1 replies
  • 0 kudos

Error while running update statement using delta lake linked service through ADF

Hi All, I am getting the below error while running an update query in a lookup activity using the delta lake linked service:ErrorCode=AzureDatabricksCommandError,Hit an error when running the command in Azure Databricks. Error details: <span class='a...

  • 1027 Views
  • 1 replies
  • 0 kudos
Latest Reply
RiyazAliM
Honored Contributor
  • 0 kudos

Hi @yashojha1995 EOL while scanning string literal hints that there might be a syntax error in the update query.could you share your update query here, and any other info such as how are you creating a Linked service to your delta lake? Does it mean ...

  • 0 kudos
Dharinip
by Contributor
  • 7780 Views
  • 5 replies
  • 3 kudos

Resolved! How to decide on creating views vs Tables in Gold layer?

We have the following use case:We receive raw form of data from an application and that is ingested in the Iron Layer. The raw data is in the JSON FormatThe Bronze layer will the first level of transformation. The flattening of the JSON file happens ...

  • 7780 Views
  • 5 replies
  • 3 kudos
Latest Reply
artus2050189155
New Contributor II
  • 3 kudos

The whole medallion architecture is unnecesarily complex.   Bronze, Silver, Gold.  Some places I have seen people do -  RAW , Trusted RAW , Silver, Trusted Silver, Gold

  • 3 kudos
4 More Replies
Labels