cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

abelian-grape
by New Contributor III
  • 1472 Views
  • 1 replies
  • 0 kudos

Near real time processing with CDC from snowflake to databricks

Hi I would like to configure near real time streaming on Databricks to process data as soon as a new data finish processing on snowflake e.g. with DLT pipelins and Auto Loader. Which option would be better for this setup? Option A)Export the Snowpark...

  • 1472 Views
  • 1 replies
  • 0 kudos
Latest Reply
saurabh18cs
Honored Contributor III
  • 0 kudos

it is like latency vs complexity and cost. you have to choose for yourself for me option A sounds reasonable

  • 0 kudos
Sans
by New Contributor III
  • 6242 Views
  • 9 replies
  • 3 kudos

Unable to create new compute in community databricks

Hi Team,I am unable to create computer in databricks community due to below error. Please advice.Bootstrap Timeout:Node daemon ping timeout in 780000 ms for instance i-0ab6798b2c762fb25 @ 10.172.246.217. Please check network connectivity between the ...

  • 6242 Views
  • 9 replies
  • 3 kudos
Latest Reply
drag7ter
Contributor
  • 3 kudos

The same get this error regularly in eu-west-1 workspace. So many issues. Did databricks try to check this issue, as it could be a bug? No any response so far? 

  • 3 kudos
8 More Replies
jyothib
by New Contributor II
  • 3164 Views
  • 2 replies
  • 3 kudos

Resolved! System tables latency

How much time is the latency of system tables#unitycatalog

  • 3164 Views
  • 2 replies
  • 3 kudos
Latest Reply
raphaelblg
Databricks Employee
  • 3 kudos

@jyothib at the current moment, system tables are still under Public Preview stage (more details at: https://docs.databricks.com/en/admin/system-tables/index.html)We don’t offer data freshness SLOs for system tables at this point and there are no pla...

  • 3 kudos
1 More Replies
Kanna
by New Contributor II
  • 2518 Views
  • 1 replies
  • 4 kudos

Resolved! Autoloader clarification

Hi team,Good day! I would like to know how we can perform an incremental load using Autoloader.I am uploading one file to DBFS and writing it into a table. When I upload a similar file to the same directory, it does not perform an incremental load; i...

  • 2518 Views
  • 1 replies
  • 4 kudos
Latest Reply
boitumelodikoko
Databricks Partner
  • 4 kudos

Hi @Kanna,Good day! Based on the issue you’re encountering, I believe the problem stems from missing deduplication or upsert logic in your current implementation. Here's an approach that combines the power of Databricks Autoloader and Delta Lake to h...

  • 4 kudos
harlemmuniz
by New Contributor II
  • 5320 Views
  • 8 replies
  • 1 kudos

Issue with Job Versioning with “Run Job” tasks and Deployments between envinronments

Hello,I am writing to bring to your attention an issue that we have encountered while working with Databricks and seek your assistance in resolving it.When running a Job of Workflow with the task "Run Job" and clicking on "View YAML/JSON," we have ob...

  • 5320 Views
  • 8 replies
  • 1 kudos
Latest Reply
saurabh18cs
Honored Contributor III
  • 1 kudos

Hi , Sorry if I don't understand your usecase, are your trying to start/stop databricks job via terraform? for this reason do you want to harcode job-id??

  • 1 kudos
7 More Replies
Kumar4567
by New Contributor II
  • 8460 Views
  • 4 replies
  • 0 kudos

disable downloading files for specific group of users ?

I see we can disable/enable download button for entire workspace using download button for notebook results.is there a way to disable/enable this just for specific group of users ?

  • 8460 Views
  • 4 replies
  • 0 kudos
Latest Reply
_anonymous
New Contributor II
  • 0 kudos

To future adventurers, the feature described by responder to OP does not exist.

  • 0 kudos
3 More Replies
MauricioS
by Databricks Partner
  • 2382 Views
  • 3 replies
  • 2 kudos

Delta Live Tables - Dynamic Target Schema

Hi all,I have a requirement where I need to migrate a few jobs from standard databricks notebooks that are orchestrated by Azure Data Factory to DLT Pipelines, pretty straight forward so far. The tricky part is that the data tables in the catalog are...

image.png
  • 2382 Views
  • 3 replies
  • 2 kudos
Latest Reply
fmadeiro
Contributor II
  • 2 kudos

@MauricioS Great question!Databricks Delta Live Tables (DLT) pipelines are very flexible, but by default, the target schema specified in the pipeline configuration (such as target or schema) is fixed. That said, you can implement strategies to enable...

  • 2 kudos
2 More Replies
Jfoxyyc
by Valued Contributor
  • 7936 Views
  • 6 replies
  • 2 kudos

Is there a way to catch the cancel button or the interrupt button in a Databricks notebook?

I'm running oracledb package and it uses sessions. When you cancel a running query it doesn't close the session even if you have a try catch block because a cancel or interrupt issues a kill command on the process. Is there a method to catch the canc...

  • 7936 Views
  • 6 replies
  • 2 kudos
Latest Reply
gustavo_woiler
New Contributor II
  • 2 kudos

I was having the same issue and I think I was finally able to solve it!When you simply except and capture the KeyboardInterrupt signal and do not raise it, the notebook gets into an endless cycle of "interrupting..." and never does anything.However, ...

  • 2 kudos
5 More Replies
pranitha
by New Contributor II
  • 1316 Views
  • 3 replies
  • 0 kudos

instance_id in compute.node_timelines

I am trying to fetch active worker nodes from system tables using the code like below:select count(distinct instance_id)from system.compute.node_timelines where cluster_id = "xx"groupy by instance_id,start_time,end_timesIt gives an output like 20 but...

  • 1316 Views
  • 3 replies
  • 0 kudos
Latest Reply
pranitha
New Contributor II
  • 0 kudos

Hi @Alberto_Umana , Thanks for replying.Even if we add the driver node it should be around 16-17 right, not like 20. I checked for al the clusters, for every cluster there is a difference of 5-7 nodes between max_worker count and count(distinct insta...

  • 0 kudos
2 More Replies
TejeshS
by Contributor II
  • 1720 Views
  • 3 replies
  • 0 kudos

Event based Alert based on certain events from System Audit tables

We need to implement an event-based trigger system that can detect any manual intervention performed by users. Upon detection of such an event, the system should automatically send a warning email. The events can be generated through DLT or other pro...

  • 1720 Views
  • 3 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

Unfortunately the system events are only tracked via the system table, only option to have more recent data will be to re execute the query each time is needed.

  • 0 kudos
2 More Replies
thedatacrew
by Databricks Partner
  • 3072 Views
  • 6 replies
  • 0 kudos

Delta Live Tables - skipChangeCommits in SQL

Hi,Could anyone tell me if the skipChangeCommits option is supported in SQL mode? I can use it successfully using Python, but it doesn't look like it is supported by SQL.It seems to be a glaring omission from the SQL support, or support for this will...

thedatacrew_0-1736866714336.png
  • 3072 Views
  • 6 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Unfortunately there is no ETA on this yet. If I know about it, will let you know!

  • 0 kudos
5 More Replies
kyrrewk
by New Contributor II
  • 1273 Views
  • 3 replies
  • 0 kudos

Monitor progress when using databricks-connect

When using databricks-connect how can you monitor the progress? Ideally, we want something similar to what you get in the Databricks notebook, i.e., information about the jobs/stages. We are using Python.

  • 1273 Views
  • 3 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

I would suggest you to submit this as a feature request through https://docs.databricks.com/en/resources/ideas.html#ideas 

  • 0 kudos
2 More Replies
matthiasn
by Databricks Partner
  • 3880 Views
  • 6 replies
  • 0 kudos

Resolved! Use temporary table credentials to access data in Databricks

Hi everybody,I tested the temporary table credentials API. I works great, as long as I use the credentials outside of Databricks (e.g. in a local duckdb instance).But as soon as I try to use the short living credentials (Azure SAS for me) in Databric...

  • 3880 Views
  • 6 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

Hello Matthias, many thanks for sharing this valuable information, it is great to hear your issue got resolved.

  • 0 kudos
5 More Replies
Hubert-Dudek
by Databricks MVP
  • 5341 Views
  • 2 replies
  • 3 kudos

Bridging the SQL-Python Gap

Python often edges out SQL with its metaprogramming capabilities. However, dbt bridges this gap with Jinja templates. Introducing simple "for" loops, especially for parameter iteration, in databricks SQL could significantly enhance the user experienc...

ezgif-5-0f8c73663b.gif
  • 5341 Views
  • 2 replies
  • 3 kudos
Latest Reply
Greg_c
New Contributor II
  • 3 kudos

Was this solved @Rajeev45  Do you have any docs?

  • 3 kudos
1 More Replies
sahasimran98
by New Contributor II
  • 1995 Views
  • 3 replies
  • 0 kudos

Data Volume Read/Processed for a Databricks Workflow Job

Hello All, I have a DBx instance hosted on Azure and I am using the Diagnostic Settings to collect Databricks Jobs related logs in log analytics workspace. So far, from the DatabricksJobs table in Azure Loganalytics, I am able to fetch basic job rela...

  • 1995 Views
  • 3 replies
  • 0 kudos
Latest Reply
saurabh18cs
Honored Contributor III
  • 0 kudos

Hi @sahasimran98 I think you're right this is more valid for synapse where such configuration exist but you can still give a try for databricks and let us know here the results. otherwise try to find some spark-monitoring package in github for databr...

  • 0 kudos
2 More Replies
Labels