cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

umahesb3
by New Contributor
  • 1171 Views
  • 0 replies
  • 0 kudos

Facing issues databricks asset bundle, All jobs are getting Deployed into specified targets Instead

Facing issues databricks asset bundle, All jobs are getting Deployed into specified targets Instead of defined target following was files i am using resourser yaml and databricks yml file , i am using Databricks CLI v0.240.0 , i am using databricks b...

  • 1171 Views
  • 0 replies
  • 0 kudos
MariuszK
by Valued Contributor II
  • 570 Views
  • 2 replies
  • 0 kudos

Changes to deletion behavior of Materialized View and Streaming Tables defined by Delta Live Table

Hi,Sometime ago, I got a message that there will be a change (starting from 01/31/2025) in "deletion behavior of Materialized View and Streaming Tables defined by Delta Live Table", but when I remove dlt pipeline, it also removes related tables, will...

  • 570 Views
  • 2 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hi @MariuszK, The users will need to explicitly call DROP MATERIALIZED VIEW to delete MVs and DROP TABLE to delete STs, when deleting DLT pipelines. https://home.databricks.com/account-alert-deletion-behavior-change-for-materialized-view-and-streamin...

  • 0 kudos
1 More Replies
muir
by New Contributor II
  • 600 Views
  • 3 replies
  • 2 kudos

Resolved! Instance Pool Usage

We have instance pools setup with a maximum capacity and are looking at ways to monitor the usage to help with our capacity planning.I have been using the system tables to track how many nodes are being used within a pool at a point in time but it ap...

  • 600 Views
  • 3 replies
  • 2 kudos
Latest Reply
TuckerGage
New Contributor II
  • 2 kudos

I also using it and it's working properly.

  • 2 kudos
2 More Replies
Ruby8376
by Valued Contributor
  • 971 Views
  • 1 replies
  • 2 kudos

Tableau analytics integration with databricks delta lake

Hii there!!Currently, we are exploring options for reporting on Salesforce. We extract data from salesforce via databricks and store it in delta lake.Is there a connector by which data can be pulled from databricks into Tableau/CRM analytics??I know ...

  • 971 Views
  • 1 replies
  • 2 kudos
Latest Reply
emillion25
New Contributor III
  • 2 kudos

Hello @ruby Were you able to resolve this? I know it's been a while, but I believe we now have multiple ways to connect Tableau and Databricks.1. Use the Native Databricks Connector for TableauTableau has a built-in Databricks connector, making it ea...

  • 2 kudos
tonykun_sg
by New Contributor II
  • 697 Views
  • 5 replies
  • 0 kudos

Delta sharing for external table to external users who has no access to external storage?

We used delta sharing (authentication type: token) to generate the config.share file and share with external users not from our organisation, the users faced the "FileNotFoundError" error while using python "delta_sharing.load_as_pandas" method to re...

  • 697 Views
  • 5 replies
  • 0 kudos
Latest Reply
Isi
Contributor III
  • 0 kudos

Hello @tonykun_sg,It looks like ADLS Gen2 might be restricting access to the data through an ACL, which is why Databricks allows access but the underlying files remain protected. Could you check with your team to temporarily enable access for testing...

  • 0 kudos
4 More Replies
ggsmith
by Contributor
  • 1113 Views
  • 8 replies
  • 3 kudos

Resolved! Workflow SQL Task Query Showing Empty

I am trying to create a SQL task in Workflows. I have my query which executes successfully in the SQL editor, and it is saved in a repo.However, when I try to execute the task, the below error shows.Query text can not be empty: BAD_REQUEST: Query tex...

ggsmith_0-1738014329449.png ggsmith_1-1738014420683.png ggsmith_2-1738014505322.png
  • 1113 Views
  • 8 replies
  • 3 kudos
Latest Reply
ggsmith
Contributor
  • 3 kudos

It ended up being that the query wasn't actually saved. Once I manually clicked save, the query preview showed and the task ran successfully. I'm really surprised that was the reason. I had moved the query around to different folders and closed and r...

  • 3 kudos
7 More Replies
nguyenthuymo
by New Contributor III
  • 389 Views
  • 2 replies
  • 0 kudos

my query works with All-purpose cluster but return NULL with SQL Warehouse

Hi,(1) On SQL warehouse, I created a table in unity catalog from data source file vw_businessmetrics_1000.json in ADLS blob.USE CATALOG `upreport`;USE SCHEMA `test_genie`;-- Create the external table from the JSON fileCREATE EXTERNAL TABLE IF NOT EXI...

databricks_question_2.png databricks_question.png
  • 389 Views
  • 2 replies
  • 0 kudos
Latest Reply
nguyenthuymo
New Contributor III
  • 0 kudos

Hi @Ayushi_Suthar Thank you very much. I tried with the Classic and Pro and it did not work.My solution is: drop the table and recreate as a delta table then loading data from json to the delta table. Now it works. Probably, the SQL warehouse only su...

  • 0 kudos
1 More Replies
ankitmit
by New Contributor III
  • 995 Views
  • 5 replies
  • 0 kudos

How to specify path while creating tables using DLT

Hi All,I am trying to create table using DLT and would like to specify the path where all the files should reside.I am trying something like this:dlt.create_streaming_table( name="test", schema="""product_id STRING NOT NULL PRIMARY KEY, ...

Data Engineering
Databricks
dlt
Unity Catalog
  • 995 Views
  • 5 replies
  • 0 kudos
Latest Reply
joma
New Contributor II
  • 0 kudos

tengo un inconveniente igual. no me gusta guardar con un nombre aleatorio dentro de __unitystorage java.lang.IllegalArgumentException: Cannot specify an explicit path for a table when using Unity Catalog. Remove the explicit path:

  • 0 kudos
4 More Replies
Sunflower7500
by New Contributor II
  • 1456 Views
  • 4 replies
  • 2 kudos

Databricks PySpark error: OutOfMemoryError: GC overhead limit exceeded

I have a Databricks pyspark query that has been running fine for the last two weeks but am now getting the following error despite no changes to the query: OutOfMemoryError: GC overhead limit exceeded.I have done some research on possible solutions a...

Sunflower7500_0-1738624317697.png
  • 1456 Views
  • 4 replies
  • 2 kudos
Latest Reply
loic
Contributor
  • 2 kudos

When you say: "I have a Databricks pyspark query that has been running fine for the last two weeks but am now getting the following error despite no changes to the query: OutOfMemoryError: GC overhead limit exceeded."Can you tell us how do you execut...

  • 2 kudos
3 More Replies
g96g
by New Contributor III
  • 357 Views
  • 2 replies
  • 0 kudos

Streaming with Medalion Architchture and star schema Help

What are the best practices for implementing non-stop streaming in a Medallion Architecture with a Star Schema?Use Case:We have operational data and need to enable near real-time reporting in Power BI, with a maximum latency of 3 minutes. No Delta li...

  • 357 Views
  • 2 replies
  • 0 kudos
Latest Reply
MadhuB
Valued Contributor
  • 0 kudos

@g96g I've setup a near real-time (30-minute latency) streaming solution that ingests data from SQL Server into Delta Lake.Changes in the source SQL Server tables are captured using Change Data Capture (CDC) and written to CSV files in a data lake.A ...

  • 0 kudos
1 More Replies
ac567
by New Contributor III
  • 901 Views
  • 3 replies
  • 0 kudos

Resolved! com.databricks.backend.common.rpc.DriverStoppedException

com.databricks.backend.common.rpc.DriverStoppedException: Driver down cause: driver state change (exit code: 143)facing this cluster issue while i deploy and run my workflow through asset bundle. i have tried everything to update in spark configurati...

  • 901 Views
  • 3 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Awesome, good to hear!

  • 0 kudos
2 More Replies
Kayla
by Valued Contributor II
  • 1978 Views
  • 14 replies
  • 6 kudos

New error: middleware.base:exception while intercepting server message

We started getting a very weird error at random from Databricks. This is from cells that routinely work, and after it happens once it will happen on every cell. It appears to be including full text of a .py file we're importing, that I've had to remo...

  • 1978 Views
  • 14 replies
  • 6 kudos
Latest Reply
Kayla
Valued Contributor II
  • 6 kudos

@TKr Hey everybody - sorry that you experienced these issues. We identified the issue and reverted the feature causing it. Things should be back to normal already.I'm glad to hear that. Are you a Databricks employee?Referring to your question, we did...

  • 6 kudos
13 More Replies
cdn_yyz_yul
by New Contributor II
  • 550 Views
  • 3 replies
  • 4 kudos

Resolved! Shoud data in Raw /Bronze be in Catalog?

Hello,What are the benefits of not "registering" Raw data into Unity Catalog when the data in Raw will be in its original format, such as .csv, .json, .parquet, etc?An example scenario could be:Data arrives at Landing as .zip; The zip will be verifie...

  • 550 Views
  • 3 replies
  • 4 kudos
Latest Reply
cdn_yyz_yul
New Contributor II
  • 4 kudos

Thanks @Rjdudley I meant to say, the scenario is:Data arrives at Landing as .zip;   The zip will be verified for correctness, and then unzipped, the extracted files will be saved to Raw as-is, in a pre-defined folder structure. Unity Catalog will not...

  • 4 kudos
2 More Replies
Sega2
by New Contributor III
  • 1680 Views
  • 1 replies
  • 0 kudos

Adding a message to azure service bus

I am trying to send a message to a service bus in azure. But I get following error:ServiceBusError: Handler failed: DefaultAzureCredential failed to retrieve a token from the included credentials.This is the line that fails: credential = DefaultAzure...

  • 1680 Views
  • 1 replies
  • 0 kudos
Latest Reply
Panda
Valued Contributor
  • 0 kudos

@Sega2 - Can you explicitly useClientSecretCredential and try

  • 0 kudos
noorbasha534
by Contributor III
  • 958 Views
  • 1 replies
  • 0 kudos

Spot instances usage in Azure Databricks

Hi all,as per the below article -https://community.databricks.com/t5/technical-blog/optimize-costs-for-your-data-and-ai-workloads-with-azure-and-aws/ba-p/662411. it is possible to choose the number of spot instances using 'availability' parameter. Bu...

  • 958 Views
  • 1 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hello @noorbasha534, Thanks for your question! 1. 'availability' Parameter: The 'availability' parameter in Azure Databricks controls whether the compute uses on-demand or spot instances. The values for this parameter are: ON_DEMAND_AZURE: This value...

  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels