cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Naga05
by New Contributor III
  • 429 Views
  • 4 replies
  • 2 kudos

Databricks app with parameters from databricks asset bundle

HelloooI tried out setting up a Databricks App using asset bundle, where i was able to successfully parameterize the sql warehouse id which was specified on specific targets. However i was unable to get values of other variables from the targets, the...

  • 429 Views
  • 4 replies
  • 2 kudos
Latest Reply
Naga05
New Contributor III
  • 2 kudos

Found that this is an implementation in progress on the Databricks CLI. https://github.com/databricks/cli/issues/3679

  • 2 kudos
3 More Replies
smoortema
by Contributor
  • 356 Views
  • 2 replies
  • 3 kudos

Resolved! handling both Pyspark and Python exceptions

In a Python notebook, I am using error handling according to the official documentation.  try:[some data transformation steps]except PySparkException as ex:[logging steps to log the error condition and error message in a table]However, this catches o...

  • 356 Views
  • 2 replies
  • 3 kudos
Latest Reply
mark_ott
Databricks Employee
  • 3 kudos

To handle both PySpark exceptions and general Python exceptions without double-logging or overwriting error details, the recommended approach is to use multiple except clauses that distinguish the exception type clearly. In Python, exception handlers...

  • 3 kudos
1 More Replies
tom_1
by New Contributor III
  • 1314 Views
  • 5 replies
  • 1 kudos

Resolved! BUG in Job Task of Type DBT

Hi, just wanted to let the Databricks Team know, that there is a bug in the task ui.Currently it is not possible to save a task of "Type: dbt" if the "SQL Warehouse" is set to "None (Manual)".Some weeks ago this was possible, also the "Profiles Direc...

tom_1_0-1741870684542.png tom_1_1-1741870779606.png
  • 1314 Views
  • 5 replies
  • 1 kudos
Latest Reply
Aishu95
New Contributor II
  • 1 kudos

I am facing this bug still. I don't want to select any SQL warehouse, what do I do? and from where can I pass the profiles directory

  • 1 kudos
4 More Replies
Navi991100
by New Contributor II
  • 241 Views
  • 3 replies
  • 1 kudos

Resolved! I recently made new account on databricks under Free edition

It by default made SQL warehouse compute, but I want all-purpose compute, as I want test and learn capabilities of PySpark and Databricks.I can't connect with the serverless compute in the notebook; it gives a mean  error as follows: "An error occurr...

Navi991100_0-1759078594989.png
  • 241 Views
  • 3 replies
  • 1 kudos
Latest Reply
belforte
New Contributor II
  • 1 kudos

In the free Databricks edition, to use PySpark you need to create and start a cluster, since the SQL Warehouse is only for SQL queries; go to Compute > Create Cluster, set up a free cluster, click Start, and then attach your notebook to it this will ...

  • 1 kudos
2 More Replies
yit
by Contributor III
  • 289 Views
  • 1 replies
  • 1 kudos

How to implement MERGE operations in Lakeflow Declarative Pipelines

Hey everyone,We’ve been using Autoloader extensively for a while, and now we’re looking to transition to full Lakeflow Declarative Pipelines. From what I’ve researched, the reader part seems straightforward and clear.For the writer, I understand that...

  • 289 Views
  • 1 replies
  • 1 kudos
Latest Reply
saurabh18cs
Honored Contributor II
  • 1 kudos

Hi @yit Lakeflow supports upsert/merge semantics natively for Delta tables unlile ForEachBatchInstead of writing custom forEachBatch code, you declare the merge keys and update logic in your pipeline configuration.Lakeflow will automatically generate...

  • 1 kudos
vishal_balaji
by New Contributor II
  • 494 Views
  • 2 replies
  • 1 kudos

Unable to access metrics from Driver node on localhost:4040

Greetings,I am trying to setup monitoring in Grafana for all my databricks clustersI have added 2 things as part of thisUnder Compute > Configuration > Advanced > Spark > Spark Config, I have addedspark.ui.prometheus.enabled trueUnder init_scripts, I...

  • 494 Views
  • 2 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @vishal_balaji ,You're following guides that were prepared for OSS Apache Spark. For sure localhost won't work in this case because in Databricks all compute is cloud-based. Please follow below guide how to configure it properly on databricks:Data...

  • 1 kudos
1 More Replies
saurabh18cs
by Honored Contributor II
  • 465 Views
  • 4 replies
  • 3 kudos

Autoloader - File Notification Mode

Hello All,We have started to consume source messages/files via autoloader directory listing mode at the moment and want to convert this to file notification mode instead so consumption can be faster with no more entire directories/folder scanning. I ...

  • 465 Views
  • 4 replies
  • 3 kudos
Latest Reply
saurabh18cs
Honored Contributor II
  • 3 kudos

Hi @K_Anudeep @szymon_dybczak how do i understand a situation when 100 jobs are running in parallel with minimal latency needed. does autoloader directly connect to the cloud queue service ? or databricks stores and manages detected files somewhere? ...

  • 3 kudos
3 More Replies
jmeer
by New Contributor II
  • 1595 Views
  • 5 replies
  • 2 kudos

Cannot click Compute tab

Hi, I want to change the cluster I am using. However, when I click on the "Compute" tab on the platform, I get automatically redirected to the "SQL Warehouses" page. I am not able to click and enter the "Compute" page. How can I solve this? Thank you

  • 1595 Views
  • 5 replies
  • 2 kudos
Latest Reply
efchea
New Contributor II
  • 2 kudos

I had the same problem, but I figured it out. You might be working in a serverless space of Databricks, so you don't have the option to create a cluster. To fix the compute issue, create a new notebook. After that, look at the right of your screen af...

  • 2 kudos
4 More Replies
RobsonNLPT
by Contributor III
  • 360 Views
  • 2 replies
  • 0 kudos

Foreign Catalog Wrong Mapping - Azure SQL Database Binary Column

Hi all.I've used foreign catalog attached to azure sql databases and never had problems except in 2 situations:1) Foreign Catalogs don't support sql schemas/objects like [xxxx.yyyy].tablename. The workaround is creating views on sql database2) This i...

  • 360 Views
  • 2 replies
  • 0 kudos
Latest Reply
Isi
Honored Contributor III
  • 0 kudos

Hello @RobsonNLPT One thing that might help to narrow this down: could you check whether the problem occurs for the entire column, or if some of the batches you receive actually contain the full (non-truncated) value?If some batches are complete but ...

  • 0 kudos
1 More Replies
DataDev
by New Contributor
  • 470 Views
  • 5 replies
  • 3 kudos

Schedule databricks job based on custom calendar

I want to schedule the databricks jobs based on the custom calender, like skip the job run on random days or holidays.#databricks @DataBricks @DATA 

  • 470 Views
  • 5 replies
  • 3 kudos
Latest Reply
Advika
Databricks Employee
  • 3 kudos

Hello @DataDev! Did the suggestions shared above help address your question? If so, please consider marking one or more responses as the accepted solution. If you found another approach that worked for you, sharing it with the community would be real...

  • 3 kudos
4 More Replies
shan-databricks
by New Contributor III
  • 275 Views
  • 3 replies
  • 3 kudos

How to load all the previous day's data only into the newly added column of the existing delta table

How to load all the previous day's data only into the newly added column of the existing delta table? Is there any option available to do that without writing any logic?

  • 275 Views
  • 3 replies
  • 3 kudos
Latest Reply
Advika
Databricks Employee
  • 3 kudos

Hello @shan-databricks! Did the suggestions shared above help resolve your concern? If so, please consider marking one of the responses as the accepted solution. If you found a different approach that worked for you, it would be great if you could sh...

  • 3 kudos
2 More Replies
philsch
by New Contributor III
  • 1073 Views
  • 8 replies
  • 3 kudos

Resolved! How to create a managed iceberg table via REST catalog

We're iceberg's java lib to write managed iceberg tables in databricks. We actually can create these tables using databricks as iceberg REST catalog. But this only works when we provide a partitioning spec. This is then picked up as cluster_columns f...

  • 1073 Views
  • 8 replies
  • 3 kudos
Latest Reply
liko
Databricks Employee
  • 3 kudos

Why are you using the iceberg-core Java library instead of an existing open source Iceberg client (like Apache Spark)? Any of these can create a table with partitions when using Unity Catalog.

  • 3 kudos
7 More Replies
chirag_nagar
by New Contributor
  • 1883 Views
  • 1 replies
  • 1 kudos

Resolved! uidance Required for Informatica to Databricks Workflow Migration Using AI

Hi Team,I am currently exploring approaches to convert Informatica PowerCenter workflows into Databricks-compatible code using AI capabilities. As part of this effort, I would like to highlight that Informatica generates individual XML files for each...

  • 1883 Views
  • 1 replies
  • 1 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 1 kudos

Greetings @chirag_nagar , as you can imagine or know, migrations are extremely complex and time consuming.  There are a few approaches to migrations but I want to focus on one - Bladebridge.  This is a free tool provided by Databricks that is AI powe...

  • 1 kudos
nikhilshetty4
by New Contributor III
  • 978 Views
  • 8 replies
  • 6 kudos

Issue with Autoloader cleanSource=MOVE Not Working as Expected

Hi everyone, I've been trying to explore on cleanSource option in Autoloader to move files from the source to an archive location after they're processed and loaded into a table. I used the following simple code to test this functionality. While the ...

  • 978 Views
  • 8 replies
  • 6 kudos
Latest Reply
Ozear
New Contributor II
  • 6 kudos

any update on this ?

  • 6 kudos
7 More Replies
mgcasas-aws
by New Contributor
  • 1886 Views
  • 1 replies
  • 1 kudos

Resolved! Azure Databricks Serverless private connection to S3 bucket

I'm looking for technical references to connect an Azure Databricks serverless workspace to an S3 bucket over a private site-to-site VPN connection. Found the following to connect AWS (consumer) to Azure (provider), but I'm looking for the other way....

  • 1886 Views
  • 1 replies
  • 1 kudos
Latest Reply
Sai_Ponugoti
Databricks Employee
  • 1 kudos

Hello @mgcasas-aws Thank you for your question! We’re currently working on a solution for private cross-cloud Delta Sharing (Azure → AWS). In the meantime, here’s a possible approach: Update your Azure Storage Account network settings from private e...

  • 1 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels