cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Sadam97
by New Contributor III
  • 782 Views
  • 2 replies
  • 1 kudos

databricks job cancel does not wait for termination of streaming tasks

We have created databricks jobs and each has multiple tasks. Each task is 24/7 running streaming with checkpoint enabled. We want it to be stateful when cancel and run the job but it seems like, when we cancel the job run it kill the parent process a...

  • 782 Views
  • 2 replies
  • 1 kudos
Latest Reply
Vidhi_Khaitan
Databricks Employee
  • 1 kudos

If the “reporting” layer is essentially micro-batching over bounded backlogs, run it with availableNow (or a scheduled batch job) so each run is naturally bounded and exits cleanly on its own, no manual cancel. This greatly reduces chances of partial...

  • 1 kudos
1 More Replies
Srajole
by New Contributor
  • 857 Views
  • 1 replies
  • 1 kudos

Write data issue

My Databricks job is completing successful but my data is not written into the target table, source path is correct, each n every thing is correct, but I am not sure y data is not written into the delta table.

  • 857 Views
  • 1 replies
  • 1 kudos
Latest Reply
Vidhi_Khaitan
Databricks Employee
  • 1 kudos

hi @Srajole ,There are a bunch of possibilities as to why the data is not being written into the table -You’re writing to a path different from the table’s storage location, or using a write mode that doesn’t replace data as expected.spark.sql("DESCR...

  • 1 kudos
dbr_data_engg
by New Contributor III
  • 2121 Views
  • 2 replies
  • 0 kudos

Using Databrick Bladebridge or Lakebridge for SQL Migration

Getting Transpile Error while executing command for Databrick Bladebridge or Lakebridge,databricks labs lakebridge transpile --source-dialect mssql --input-source "<Path>/sample.sql" --output-folder "<Path>\output"Error :TranspileError(code=FAILURE, ...

  • 2121 Views
  • 2 replies
  • 0 kudos
Latest Reply
Abhimanyu
Databricks Partner
  • 0 kudos

did you find a solution? 

  • 0 kudos
1 More Replies
juanjomendez96
by Contributor
  • 1328 Views
  • 2 replies
  • 3 kudos

Resolved! Best practices for compute usage

Hello there!I am writing this open message to know how you guys are using the computes in your work cases.Currently, in my company, we have multiple compute instances that can be differentiated into two main types:Clusters with a large instance for b...

  • 1328 Views
  • 2 replies
  • 3 kudos
Latest Reply
radothede
Valued Contributor II
  • 3 kudos

Hello @juanjomendez96 ,to my best knowledge and experience autoscaled shared cluster (using smaller instances) works good for most 2nd-case scenario (clusters for ad-hoc/development team usage).This approach allows You to reuse the resources across t...

  • 3 kudos
1 More Replies
VicS
by Databricks Partner
  • 1575 Views
  • 1 replies
  • 1 kudos

Resolved! How to install SAP JDBC on job cluster via asset bundles

I'm trying to use the SAP JDBC driver to read data in my Spark application which I deploy via asset bundles with job computes.I was able to install the SAP JDBC Driver on a general purpose cluster by adding the jar (com.sap.cloud.db.jdbc:ngdbc:2.25.9...

  • 1575 Views
  • 1 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @VicS ,To add a Maven package to a job task definition , in libraries, specify a maven mapping for each Maven package to be installed. For each mapping, specify the following: resources: jobs: my_job: # ... tasks: - task_...

  • 1 kudos
abueno
by Contributor
  • 1089 Views
  • 1 replies
  • 1 kudos

Resolved! Python If Statement with multiple "and" conditions, if not default column value

Python 3.10.12 I am trying to get these filter results: example:If "column1" = '2024' and column2 in ('DE','SC') then 'value1" else 'value2'If "column1" = "2023" and column2 in ('DE,'SC')then "value3 else "value4"if the row/record does not fit the cr...

  • 1089 Views
  • 1 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @abueno .I'm assuming you're asking how to do this in pyspark. You can use when and otherwise conditional functions to achieve your expected result:from pyspark.sql import SparkSession from pyspark.sql.types import StructType, StructField, StringT...

  • 1 kudos
pogo
by New Contributor III
  • 577 Views
  • 1 replies
  • 1 kudos

Resolved! Cognito as IdP provider for Delta Share

I am trying to setup a delta sharing Recipient using OIDC Federation with the Issuer URL being cognito idp endpoint.Are there any examples, other than EntraID, for the values of Subject Claim/Subject/Audiences in the OIDC Policy for Cognito or Google...

  • 577 Views
  • 1 replies
  • 1 kudos
Latest Reply
pogo
New Contributor III
  • 1 kudos

We managed to figure how to make machine to machine authentication to work.when you setup cognito pool for m2m scenario you add App Clientand then set App Client as both `sub` and Audience in databricks recepient OIDC Policy:2. Set `aud` claim to the...

  • 1 kudos
LakehouseOMG14
by New Contributor II
  • 4271 Views
  • 7 replies
  • 3 kudos

Resolved! Salesforce with Databricks connectivity

Can we connect Salesforce with Databricks.I want to do both push and pull activity using Databricks and Salesforce.Do There any challenge while using ODBC?Please help me with detailed approach.Thanks a ton.

  • 4271 Views
  • 7 replies
  • 3 kudos
Latest Reply
ManojkMohan
Honored Contributor II
  • 3 kudos

Have resolved it Step 1: Create a Service PrincipalLog in to your Databricks Workspace and navigate to the Admin Settings page by clicking your email in the bottom-left corner and selecting "Admin Settings".Go to the Identity and access tab and click...

  • 3 kudos
6 More Replies
ChristianRRL
by Honored Contributor
  • 1456 Views
  • 2 replies
  • 1 kudos

Autoloader Error Loading and Displaying

Hi there,I'd appreciate some assistance with troubleshooting what is supposed to be a (somewhat) simple use of autoloader. Below are some screenshots highlighting my issue:When I attempt to create the dataframe via spark.readStream.format("cloudFiles...

ChristianRRL_0-1750702687568.png ChristianRRL_1-1750702720386.png
  • 1456 Views
  • 2 replies
  • 1 kudos
Latest Reply
lingareddy_Alva
Esteemed Contributor
  • 1 kudos

Hi @ChristianRRL  This is a common issue with Spark Structured Streaming and the display() function.The error occurs because you're trying to display a streaming DataFrame, which requires special handling. Here are several solutions:1. Use writeStrea...

  • 1 kudos
1 More Replies
Karl
by New Contributor II
  • 8733 Views
  • 2 replies
  • 0 kudos

Resolved! DB2 JDBC Connection from Databricks cluster

Has anyone successfully connected to a DB2 database on ZOS from a Databricks cluster using a JDBC connection?I also need to specify an SSL certificate path and not sure if I need to use an init script on the cluster to do so.Any examples would be ver...

  • 8733 Views
  • 2 replies
  • 0 kudos
Latest Reply
Ayushi_Suthar
Databricks Employee
  • 0 kudos

Hi @Karl , Greetings!  I've outlined the steps below to connect from Databricks to IBM DB2 using JDBC:Step 1: Obtain the DB2 JDBC Driver Visit the IBM website to download the appropriate JDBC driver for DB2 on z/OS.Reference Document: IBM DB2 JDBC Dr...

  • 0 kudos
1 More Replies
pogo
by New Contributor III
  • 1058 Views
  • 2 replies
  • 2 kudos

Resolved! Delta sharing to pandas error

we are on a trial databricks premium workspace (fully managed by databricks)we are trying to test Delta Sharing feature, where we are sharing UC table with a recipient using python client (outside of databricks). We are using `delta-sharing` python l...

  • 1058 Views
  • 2 replies
  • 2 kudos
Latest Reply
pogo
New Contributor III
  • 2 kudos

Yes - you are right, works on trial - i was able to setup S3 as an external location and configured the UC schema to use this S3 ext location, and then was able to query data from an external python client.

  • 2 kudos
1 More Replies
thiagoawstest
by Contributor
  • 4963 Views
  • 3 replies
  • 0 kudos

create databricks scope by reading AWS secrets manager

Hi, I have datbricks on AWS, I created some secrets in AWS Secrets Manger, I would need to create the scopes based on AWS secrets manager.When I use Azure's Key Vault, when creating the scope, it uses the option -scope-backend-type AZURE_KEYVAULT, bu...

  • 4963 Views
  • 3 replies
  • 0 kudos
Latest Reply
Yeshwanth
Databricks Employee
  • 0 kudos

Hi @thiagoawstest  Step 1: Create Secret ScopeYou can create a secret scope using the Databricks REST API as shown below: python import requests import json # Define the endpoint and headers url = "https://<databricks-instance>/api/2.0/secrets/scope...

  • 0 kudos
2 More Replies
SusmithaBadam
by New Contributor II
  • 1065 Views
  • 1 replies
  • 0 kudos

Liquid clustering not improved performance

Hi There,I have a table of 160 GB with partition applied on country and yearmonth columns, I maintain a previous history of 6 years and replace the partitions (latest 2 months) to add the new data.I use overwrite mode to replace the effected partitio...

  • 1065 Views
  • 1 replies
  • 0 kudos
Latest Reply
Renu_
Valued Contributor II
  • 0 kudos

Hi @SusmithaBadam, based on your use case, partitioned tables are performing better because they work kind of like labeled folders. When you group by, it can quickly go to the exact folder instead of scanning everything, so it’s much faster.Liquid cl...

  • 0 kudos
Suki
by New Contributor III
  • 1243 Views
  • 2 replies
  • 0 kudos

Issue with Resetting Checkpoint Metadata in DLT with Unity Catalog

Hi Community,Hope someonne can help with this DLT question.I am currently working in a Databricks environment using Delta Live Tables (DLT) with Unity Catalog enabled, and I'm encountering a blocker related to schema evolution and checkpoint metadata...

  • 1243 Views
  • 2 replies
  • 0 kudos
Latest Reply
T0M
Contributor
  • 0 kudos

I feel you.Probably not the way to go, but did to try to destroy and re-deploy your pipeline?

  • 0 kudos
1 More Replies
Mohan_Baabu1
by New Contributor III
  • 4383 Views
  • 4 replies
  • 3 kudos

Resolved! Best Practices for Designing Bronze Layer with SQL Server Source in Medallion Architecture

Hi Databricks Experts,I'm working on a Medallion Architecture implementation in Databricks, where the source data is coming from SQL Server. I would like some advice on how to handle the bronze layer correctly and cost-effectively.Should I create a b...

  • 4383 Views
  • 4 replies
  • 3 kudos
Latest Reply
pgo
New Contributor III
  • 3 kudos

Create the bronze table using Auto Loader and store it in Delta format. Although it might seem like you'll only read from bronze once to populate the silver layer, in real-world production scenarios, you'll often need to re-read from bronze—for repro...

  • 3 kudos
3 More Replies
Labels