cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Kabil
by New Contributor
  • 254 Views
  • 0 replies
  • 0 kudos

useing dlt metadata as runtime parameter

i have started using DLT pipeline, and i have common code which is used by multiple DLT pipeline. now i need to read metadata information like name of the pipeline and start time of the pipeline during run time, but since im using common code and pip...

  • 254 Views
  • 0 replies
  • 0 kudos
797646
by New Contributor II
  • 747 Views
  • 5 replies
  • 2 kudos

Resolved! Calculated measures not working in Dashboards for queries with big result

Queries with big result are executed on cluster. If we specify calculated measure as something like cal1 ascount(*) / count(distinct field1) it will wrap it in backticks as `count(*) / count(distinct field1) ` as `cal1`functions are not identified in...

  • 747 Views
  • 5 replies
  • 2 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 2 kudos

Hello Team, Could you all try with all caps? COUNT(DISTINCT xxx)

  • 2 kudos
4 More Replies
abelian-grape
by New Contributor III
  • 548 Views
  • 5 replies
  • 0 kudos

Trigger a Databricks Job When there is an insert to a Snowflake Table?

I need to automatically trigger a Databricks job whenever a new row is inserted to a Snowflake table. Additionally, I need the job to receive the exact details of the newly inserted row as parameters.What are the best approaches to achieve this? I’m ...

  • 548 Views
  • 5 replies
  • 0 kudos
Latest Reply
ashraf1395
Honored Contributor
  • 0 kudos

I think lamba function/ event bridge would be a good way - You can query your snowflake table there and create logic for any new row insert mabe CDC etc and then you send a job trigger using databricks API / databricks SDK where you can pass your new...

  • 0 kudos
4 More Replies
Trifa
by New Contributor II
  • 753 Views
  • 3 replies
  • 1 kudos

Resolved! Override DLT Fille Refresh using a Job parameter

HelloI have a Job with a DLT pipeline as a first task. From time to time, I want to execute this Job with a Full Refresh of the DLT pipeline. How could I override my default "full_refresh = false" ?This was possible before using the Legacy parameters...

Trifa_0-1701170537015.png
  • 753 Views
  • 3 replies
  • 1 kudos
Latest Reply
adriennn
Valued Contributor
  • 1 kudos

@Trifa luckily, it's simple to implement. You can be the guys are going to release Pipeline Parameters® a week after you have deployed your solution though 

  • 1 kudos
2 More Replies
Kumarn031425
by New Contributor
  • 287 Views
  • 1 replies
  • 0 kudos

Automating Migration of Delta Live Tables Pipelines Across Environments Using Azure DevOps CI/CD

I am seeking guidance on automating the migration of Delta Live Tables (DLT) pipelines across various environments—specifically from development to testing, and ultimately to production—utilizing Azure DevOps for Continuous Integration and Continuous...

  • 287 Views
  • 1 replies
  • 0 kudos
Latest Reply
ashraf1395
Honored Contributor
  • 0 kudos

Hi there @Kumarn031425 , I guess, This video tutorial will answer most of your questions : https://youtu.be/SZM49lGovTg?si=X7Cwp0Wfqlo1OnuSHere , deployment of workspace resources using databricks azure devops and databeicks asset bundles tutorial is...

  • 0 kudos
WYO
by New Contributor II
  • 144 Views
  • 1 replies
  • 0 kudos
  • 144 Views
  • 1 replies
  • 0 kudos
Latest Reply
ashraf1395
Honored Contributor
  • 0 kudos

Hi there @WYO , I dont think we have a way to add multiple notebooks for a single dlt pipeline from dlt pipeline configuration settings.But there can be another way - you can create a single notebooks which has multiple code blocks which use the run ...

  • 0 kudos
Meghana89
by New Contributor II
  • 314 Views
  • 2 replies
  • 0 kudos

Read Write Stream Data from Event Hub to Databricks Delta Lake

I am trying to read streaming data from EventHub which is in JSON Format, able to read data in a data frame but the body type was coming as binary I have converted it to string and decoded but while implementing the write stream I am facing an ongoin...

  • 314 Views
  • 2 replies
  • 0 kudos
Latest Reply
Meghana89
New Contributor II
  • 0 kudos

@SantoshJoshi  Thanks for reply please find the code snippet belowfrom pyspark.sql import functions as Ffrom pyspark.sql.types import StringTypeimport base64# Define the Event Hubs connection stringconnectionString = endpoint (replace with endpoint f...

  • 0 kudos
1 More Replies
908314
by New Contributor II
  • 427 Views
  • 3 replies
  • 2 kudos

Cluster logs stopped getting written to S3

We have two Databricks Workspaces and since a couple of days ago, cluster logs are not getting persisted to S3, in both workspaces. Driver logs are available in Databricks UI only when the job is active. Haven't seen any errors in the job logs relate...

  • 427 Views
  • 3 replies
  • 2 kudos
Latest Reply
adriantaut
New Contributor II
  • 2 kudos

Hello, Facing same issue in both our Workspaces, our Cluster logs suddenly stopped being delivered to S3 on 12th of March. There were no changes on Cluster settings nor IAM side, also all IAM Permissions should be in place according to Databricks Off...

  • 2 kudos
2 More Replies
DylanStout
by Contributor
  • 418 Views
  • 2 replies
  • 0 kudos

Resolved! DBFS folder access

When trying to open a folder in dbfs, mnt in my case, my whole team gets the following error message - Uncaught Error: No QueryClient set, use QueryClientProvider to set one. Reloading the page results in this error not showing up anymore, but the fo...

DylanStout_1-1741874948742.png
  • 418 Views
  • 2 replies
  • 0 kudos
Latest Reply
DylanStout
Contributor
  • 0 kudos

Compute had to be assigned first before being able to open the folder, this was done automatically before.The error is however not clear at all that this has to be done and that this is causing the error.

  • 0 kudos
1 More Replies
sparklez
by New Contributor II
  • 356 Views
  • 2 replies
  • 1 kudos

DAB fails to deploy from first try "TLS protocol version not supported"

 HelloI'm using the Databricks extension on VSCode, and when i attempt to deploy i often get this error"Error: Post "https://adb-xxxxxx.xx.azuredatabricks.net/api/2.0/workspace-files/import-file/Workspace%2FUsers%2FXXXX@XXX.com%2F.bundle%2Fxxxx%2Floc...

sparklez_0-1741344021912.png
  • 356 Views
  • 2 replies
  • 1 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 1 kudos

Hi @sparklez, Is your network blocking TLS 1.2? Databricks supports TLS 1.2  and TLSv1.3 so wondering if any of these is blocked.

  • 1 kudos
1 More Replies
azam-io
by New Contributor II
  • 385 Views
  • 2 replies
  • 0 kudos

Passing Different Job Parameters or Environment Variables Based on Targets in DABs

Hi, I am currently using DABs in my project and have encountered an issue. Specifically, I need to pass different job parameters or environment variables depending on the target and for specific jobs.Could you please provide guidance on how to approa...

  • 385 Views
  • 2 replies
  • 0 kudos
Latest Reply
Nivethan_Venkat
Contributor
  • 0 kudos

Hi @azam-io,Please find the best practices here for deploying jobs using Asset Bundles configuration: https://docs.databricks.com/aws/en/dev-tools/bundles/settingsIf you would like to pass different parameters for your same job in different environme...

  • 0 kudos
1 More Replies
tom_1
by New Contributor III
  • 404 Views
  • 3 replies
  • 1 kudos

Resolved! BUG in Job Task of Type DBT

Hi, just wanted to let the Databricks Team know, that there is a bug in the task ui.Currently it is not possible to save a task of "Type: dbt" if the "SQL Warehouse" is set to "None (Manual)".Some weeks ago this was possible, also the "Profiles Direc...

tom_1_0-1741870684542.png tom_1_1-1741870779606.png
  • 404 Views
  • 3 replies
  • 1 kudos
Latest Reply
Gerard_Sola
New Contributor II
  • 1 kudos

This bug has appeared again in Azure Databricks. Could you please review it?Thanks!

  • 1 kudos
2 More Replies
mlivshutz
by New Contributor II
  • 278 Views
  • 2 replies
  • 0 kudos

How to configure DAB bundles to run serverless

I am following the guidelines in https://docs.databricks.com/aws/en/dev-tools/bundles/jobs-tutorial to setup the job for serverless. It says to "omit the job_clusters configuration from the bundle configuration file." It sounds like the idea is to si...

  • 278 Views
  • 2 replies
  • 0 kudos
Latest Reply
mlivshutz
New Contributor II
  • 0 kudos

Hi, @ashraf1395 , Thank you for looking at my question. My cli is 0.243, which is current as of today (3/17/25).The task definition within resources/dbx_backfill_emotion_job.yml:tasks: - task_key: dbx_backfill_base_fields_x_1 # job_...

  • 0 kudos
1 More Replies
noorbasha534
by Contributor III
  • 343 Views
  • 1 replies
  • 0 kudos

Databricks Jobs API - Throttling

Dear all,I am planning to execute a script that fetches databricks jobs status every 10 minutes. I have around 500 jobs in my workspace. The APIs I use are listed below - list runs, get all job runs.I was wondering if this could cause throttling as t...

  • 343 Views
  • 1 replies
  • 0 kudos
Latest Reply
koji_kawamura
Databricks Employee
  • 0 kudos

Hi @noorbasha534  Different limitations are implemented at API endpoints. The "/jobs/runs/list" has a limitation of 30 requests/second. The number of concurrent task executions is limited up to 2000. These limits work separately, so the job list API ...

  • 0 kudos
brickster_2018
by Databricks Employee
  • 8631 Views
  • 6 replies
  • 2 kudos
  • 8631 Views
  • 6 replies
  • 2 kudos
Latest Reply
VasuBajaj
New Contributor II
  • 2 kudos

A .CRC file (Cyclic Redundancy Check) is an internal checksum file used by Spark (and Hadoop) to ensure data integrity when reading and writing files.Data Integrity Check – .CRC files store checksums of actual data files. When reading a file, Spark/H...

  • 2 kudos
5 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels