cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

yinan
by New Contributor
  • 139 Views
  • 5 replies
  • 2 kudos
  • 139 Views
  • 5 replies
  • 2 kudos
Latest Reply
Khaja_Zaffer
Contributor
  • 2 kudos

Hello @yinan Good day!!Databricks, being a cloud-based platform, does not have direct built-in support for reading data from a truly air-gapped (completely offline, no network connectivity) Cloudera Distribution for Hadoop (CDH) environment.  In such...

  • 2 kudos
4 More Replies
Kurgod
by New Contributor II
  • 159 Views
  • 2 replies
  • 0 kudos

Using Databricks to transform cloudera lakehouse on-prem without bringing the data to cloud

I am looking for a solution to connect databricks to cloudera lakehouse hosted on-prem and transform the data using databricks without bringing the data to databricks delta tables or cloud storage. once the transformation is done the data need to be ...

  • 159 Views
  • 2 replies
  • 0 kudos
Latest Reply
BR_DatabricksAI
Contributor III
  • 0 kudos

Hello, What is your data volume? You can connect using  jdbc/odbc but this process will be slower if the data volume is too high.Another way of connecting is if your cloudera storage is in HDFS then you can also connect through HDFS API as well.  

  • 0 kudos
1 More Replies
azam-io
by New Contributor II
  • 618 Views
  • 4 replies
  • 2 kudos

How can I structure pipeline-specific job params separately in Databricks Asset Bundle.

Hi all, I am working with databricks asset bundle and want to separate environment-specific job params (for example, for "env" and "dev") for each pipeline within my bundle. I need each pipeline to have its own job params values for different environ...

  • 618 Views
  • 4 replies
  • 2 kudos
Latest Reply
Michał
New Contributor
  • 2 kudos

Hi azam-io, were you able to solve your problem? Are you trying to have different parameters depending on the environment, or a different parameter value? I think the targets would allow to specify different parameters per environment / target. As fo...

  • 2 kudos
3 More Replies
seefoods
by Contributor II
  • 2403 Views
  • 2 replies
  • 1 kudos

Resolved! assets bundle

Hello Guys,I am working on assets bundle. So i want to make it generic for all team like ( analytics, data engineering), Someone could you share a best practice for this purpose ? Cordially, 

  • 2403 Views
  • 2 replies
  • 1 kudos
Latest Reply
Michał
New Contributor
  • 1 kudos

Hi seefoods, Were you able to achieve that generic asset bundle setup? I've been working on something, potentially, similar, and I'd be happy to discuss it, hoping to share experiences. While what I have works for a few teams, it is focused on declar...

  • 1 kudos
1 More Replies
SharathE
by New Contributor III
  • 1908 Views
  • 3 replies
  • 1 kudos

Incremental refresh of materialized view in serverless DLT

Hello, Every time that I run a delta live table materialized view in serverless , I get a log of "COMPLETE RECOMPUTE" . How can I achieve incremental refresh in serverless in DLT pipelines?

  • 1908 Views
  • 3 replies
  • 1 kudos
Latest Reply
drewipson
New Contributor III
  • 1 kudos

Make sure you are using the aggregates and SQL restrictions outlined in this article. https://docs.databricks.com/en/optimizations/incremental-refresh.htmlIf a SQL function is non-deterministic (current_timestamp() is a common one) you will have a CO...

  • 1 kudos
2 More Replies
korijn
by New Contributor II
  • 664 Views
  • 4 replies
  • 0 kudos

Git integration inconsistencies between git folders and job git

It's a little confusing and limiting that the git integration support is inconsistent between the two options available.Sparse checkout is only supported when using a workspace Git folder, and checking out by commit hash is only supported when using ...

  • 664 Views
  • 4 replies
  • 0 kudos
Latest Reply
_J
New Contributor II
  • 0 kudos

Same here, could be a good improvement for the jobs layer guys!

  • 0 kudos
3 More Replies
IONA
by New Contributor III
  • 385 Views
  • 6 replies
  • 7 kudos

Resolved! Getting data from the Spark query profiler

When you navigate to Compute > Select Cluster > Spark UI > JDBC/ODBC There you can see grids of Session stats and SQL stats. Is there any way to get this data in a query so that I can do some analysis? Thanks

  • 385 Views
  • 6 replies
  • 7 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 7 kudos

 Hi @IONA ,As @BigRoux  correctly suggested there no native way to get stats from JDBC/ODBC Spark UI.1. You can try to use query history system table, but it has limited number of metrics %sql SELECT * FROM system.query.history 2. You can use /api/2....

  • 7 kudos
5 More Replies
LeoGaller
by New Contributor II
  • 7539 Views
  • 4 replies
  • 4 kudos

What are the options for "spark_conf.spark.databricks.cluster.profile"?

Hey guys, I'm trying to find what are the options we can pass to spark_conf.spark.databricks.cluster.profileI know looking around that some of the available configs are singleNode and serverless, but there are others?Where is the documentation of it?...

  • 7539 Views
  • 4 replies
  • 4 kudos
Latest Reply
s3
New Contributor II
  • 4 kudos

Recently I got stuck with the same issue. However, in the new view of the form/template to create a policy, you have and option to delete the setting "spark_conf.spark.databricks.cluster.profile" by clicking on the "bin" icon. Once you did that, you ...

  • 4 kudos
3 More Replies
Yulei
by New Contributor III
  • 29298 Views
  • 7 replies
  • 1 kudos

Resolved! Could not reach driver of cluster

 Hi, Rencently, I am seeing issue Could not reach driver of cluster <some_id> with my structure streaming job when migrating to unity catalog and found this when checking the traceback:Traceback (most recent call last):File "/databricks/python_shell/...

  • 29298 Views
  • 7 replies
  • 1 kudos
Latest Reply
omsingh
New Contributor II
  • 1 kudos

It seems like a temporary connectivity or cluster initialization glitch. So if anyone else runs into this, try re-running the job before diving into deeper troubleshooting - it might just work!Hope this helps someone save time.

  • 1 kudos
6 More Replies
ChristianRRL
by Valued Contributor III
  • 101 Views
  • 1 replies
  • 0 kudos

Can schemaHints dynamically handle nested json structures? (Part 2)

Hi there, I'd like to follow up on a prior post:https://community.databricks.com/t5/data-engineering/can-schemahints-dynamically-handle-nested-json-structures/m-p/130209/highlight/true#M48731Basically I'm wondering what's the best way to set *both* d...

  • 101 Views
  • 1 replies
  • 0 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 0 kudos

I am not aware on schemahints supporting wildcards for now.  It would be awesome to have though, I agree.So I think you are stuck with what is already proposed in your previous post, or exploding the json or other transformations.

  • 0 kudos
minhhung0507
by Valued Contributor
  • 80 Views
  • 1 replies
  • 1 kudos

Could not reach driver of cluster

I am running a pipeline job in Databricks and it failed with the following message:Run failed with error message Could not reach driver of cluster 5824-145411-p65jt7uo. This message is not very descriptive, and I am not able to identify the root ca...

minhhung0507_0-1756870994085.png
  • 80 Views
  • 1 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @minhhung0507 ,Typically this error could appear when there's a high load on the driver node. Another reason could be related to high garbage collection on driver node as well as high memory and cpu which leads to throttling, and prevents the driv...

  • 1 kudos
elgeo
by Valued Contributor II
  • 6012 Views
  • 7 replies
  • 8 kudos

Clean up _delta_log files

Hello experts. We are trying to clarify how to clean up the large amount of files that are being accumulated in the _delta_log folder (json, crc and checkpoint files). We went through the related posts in the forum and followed the below:SET spark.da...

  • 6012 Views
  • 7 replies
  • 8 kudos
Latest Reply
michaeljac1986
New Contributor
  • 8 kudos

What you’re seeing is expected behavior — the _delta_log folder always keeps a history of JSON commit files, checkpoint files, and CRCs. Even if you lower delta.logRetentionDuration and run VACUUM, cleanup won’t happen immediately. A couple of points...

  • 8 kudos
6 More Replies
erigaud
by Honored Contributor
  • 9609 Views
  • 7 replies
  • 6 kudos

Resolved! SFTP Autoloader

Hello, Don't know if it is possible, but I am wondering if it is possible to ingest files from a SFTP server using autoloader ? Or do I have to first copy the files to my dbfs and then use autoloader on that location ? Thank you !

  • 9609 Views
  • 7 replies
  • 6 kudos
Latest Reply
Anonymous
Not applicable
  • 6 kudos

Hi @erigaud  We haven't heard from you since the last response from​, @BriceBuso  and I was checking back to see if her suggestions helped you. Or else, If you have any solution, please share it with the community, as it can be helpful to others.  Al...

  • 6 kudos
6 More Replies
chiruinfo5262
by New Contributor II
  • 519 Views
  • 4 replies
  • 0 kudos

Trying to convert oracle sql to databricks sql but not getting the desired output

ORACLE SQL: COUNT( CASE WHEN TRUNC(WORKORDER.REPORTDATE) BETWEEN SELECTED_PERIOD_START_DATE AND SELECTED_PERIOD_END_DATE THEN 1 END ) SELECTED_PERIOD_BM,COUNT( CASE WHEN TRUNC(WORKORDER.REPORTDATE) BETWEEN COMPARISON_PERIOD_START_DATE AND COMPARISON_...

  • 519 Views
  • 4 replies
  • 0 kudos
Latest Reply
Granty
New Contributor
  • 0 kudos

This is a helpful comparison! I've definitely run into similar date formatting issues when migrating queries. The Oracle TRUNC function and Databricks' DATE_FORMAT/CAST combo can be tricky to reconcile. Speaking of needing a break after debugging SQL...

  • 0 kudos
3 More Replies
james_
by New Contributor II
  • 224 Views
  • 5 replies
  • 0 kudos

Low worker utilisation in Spatial SQL

I am finding low worker node utilization when using Spatial SQL features. My cluster is DBR 17.1 with 2x workers and photon enabled.When I view the cluster metrics, they consistently show one worker around 30-50% utilized, the driver around 15-20%, a...

  • 224 Views
  • 5 replies
  • 0 kudos
Latest Reply
james_
New Contributor II
  • 0 kudos

Thank you again, @-werners- . I have a lot still to learn about partitioning and managing spatial data. Perhaps I mainly need more patience!

  • 0 kudos
4 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels