cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

MyProfile
by New Contributor
  • 110 Views
  • 1 replies
  • 0 kudos

Disable Public Network Access on Databricks Managed Storage Account - Deny Assignment

Issue Description:I am attempting to disable public network access on the Azure Databricks managed storage account. However, I am encountering the following error:Failed to save resource settings — access is denied due to a deny assignment created by...

  • 110 Views
  • 1 replies
  • 0 kudos
Latest Reply
Sumit_7
Honored Contributor III
  • 0 kudos

@MyProfile This would be helpful, check once - https://learn.microsoft.com/en-us/answers/questions/1707749/managed-storage-accounts-compliance

  • 0 kudos
tsam
by New Contributor
  • 174 Views
  • 3 replies
  • 0 kudos

Driver memory utilization grows continuously during job

I have a batch job that runs thousands of Deep Clone commands, it uses a ForEach task to run multiple Deep Clones in parallel. It was taking a very long time and I realized that the Driver was the main culprit since it was using up all of its memory ...

tsam_2-1776095245905.png
  • 174 Views
  • 3 replies
  • 0 kudos
Latest Reply
nayan_wylde
Esteemed Contributor II
  • 0 kudos

You’re seeing (a monotonic / stair‑step climb in driver RAM over thousands of DEEP CLONE statements) is a very common pattern when the driver is not “holding data”, but holding metadata, query artifacts, and per‑command state that accumulates faster ...

  • 0 kudos
2 More Replies
abhijit007
by Databricks Partner
  • 142 Views
  • 2 replies
  • 2 kudos

Redshift to Databricks Migration with Lakebridge

We are currently performing an assessment for a client’s Redshift to Databricks migration, and we would like to better understand the enhanced capabilities of Lakebridge for this use case.We would appreciate clarification on the following points:Scop...

  • 142 Views
  • 2 replies
  • 2 kudos
Latest Reply
pradeep_singh
Contributor III
  • 2 kudos

There is a nice course on Partner Academy as well . It uses SQL Server as a target system for migration but you can follow the same steps for Redshift as well . https://partner-academy.databricks.com/learn/courses/4326/lakebridge-for-sql-source-syste...

  • 2 kudos
1 More Replies
ittzzmalind
by New Contributor II
  • 620 Views
  • 1 replies
  • 0 kudos

Resolved! Accessing Azure Databricks Workspace via Private Endpoint and On-Premises Proxy

Public access to the Azure Databricks workspace is currently disabled. Access is required through a Private Link (private endpoint – api_ui).A private endpoint has already been configured successfully:Virtual Network: Vnet-PE-ENDPOINTSubnet: Snet-PE-...

  • 620 Views
  • 1 replies
  • 0 kudos
Latest Reply
anuj_lathi
Databricks Employee
  • 0 kudos

This is a classic hub-spoke + on-premises hybrid networking scenario. Here's how to architect it end-to-end. Architecture Overview The traffic flow will be: VM (VNet-App) --> ExpressRoute/VPN Gateway --> On-Prem Proxy Server --> ExpressRoute/VPN Gate...

  • 0 kudos
ittzzmalind
by New Contributor II
  • 258 Views
  • 1 replies
  • 1 kudos

Resolved! Delta Sharing with Materialized View - recepient data not refreshing when using Open Protocol

Scenario: Delta Sharing with Materialized ViewProvider Side Setup :->A Delta Share was created.->A materialized view was added to the share.->Recipients Created-> 1). Open Delta Sharing recipient       Accessed using Python (import delta_sharing)->2)...

  • 258 Views
  • 1 replies
  • 1 kudos
Latest Reply
Ashwin_DSA
Databricks Employee
  • 1 kudos

Hi @ittzzmalind, This is expected behaviour and is mainly due to how Delta Sharing handles materialized views for open (non-Databricks) recipients versus Databricks-to-Databricks recipients. For Databricks-to-Databricks recipients, the shared materia...

  • 1 kudos
ittzzmalind
by New Contributor II
  • 271 Views
  • 1 replies
  • 1 kudos

Resolved! Databricks Workspace - Unknow IP access

Azure monitor log showing unknow ip authentication requests to Databricks workspace . -- When searched ip below url, result showing its from AZURE CLOUD : <Region> (the region is same as workspace)https://azureipranges.azurewebsites.net/SearchFor -- ...

  • 271 Views
  • 1 replies
  • 1 kudos
Latest Reply
Ashwin_DSA
Databricks Employee
  • 1 kudos

Hi @ittzzmalind, Because the IP is in the same Azure region but not listed in the Azure Databricks control plane ranges, it’s very likely not a Databricks owned control plane IP. It’s typically either a user or service coming from another Azure resou...

  • 1 kudos
sai_sakhamuri
by Databricks Partner
  • 794 Views
  • 1 replies
  • 1 kudos

Resolved! Databricks optimization for query perfomance and pipeline run

I am currently working on optimizing several Spark pipelines and wanted to gather community insights on advanced performance tuning. Typically, my workflow for traditional SQL optimization involves a deep dive into the execution plan to identify bott...

  • 794 Views
  • 1 replies
  • 1 kudos
Latest Reply
lingareddy_Alva
Esteemed Contributor
  • 1 kudos

Hi @sai_sakhamuri You're clearly past the basics. Let me give you a practitioner-level breakdown of each layer you mentioned, plus a few things that often get overlooked.Spark Catalyst Optimizer — Working With the Rules EngineCatalyst operates in fou...

  • 1 kudos
databrciks
by New Contributor III
  • 510 Views
  • 3 replies
  • 1 kudos

Resolved! Parametrize the DLT pipeline for dynamic loading of many tables

I need to load many tables into Bronze layer connecting to sql server DB. How can i pass the tables names dynamically in DLT. Means one code pass many tables and load into bronze layer

  • 510 Views
  • 3 replies
  • 1 kudos
Latest Reply
databrciks
New Contributor III
  • 1 kudos

Hi Ashwin Thanks for the quick response. Yes I want to pass all the tables through config parameter/param file and load that into bronze layerI will try this approach. Thanks 

  • 1 kudos
2 More Replies
ittzzmalind
by New Contributor II
  • 218 Views
  • 2 replies
  • 0 kudos

DLT Pipeline Error -key not found: all_info_dlt_cx_utils_cod resulting in a NoSuchElementException.

Databricks ETL pipeline, specifically an error with the @DP.expectorfail decorator causing the pipeline update to fail. The error message indicated a 'key not found: all_info_dlt_cx_utils_cod ' resulting in a NoSuchElementException.Note: if we commen...

  • 218 Views
  • 2 replies
  • 0 kudos
Latest Reply
ittzzmalind
New Contributor II
  • 0 kudos

@MoJaMa Thanks for the reply, The issue was in the code, corrected code worked

  • 0 kudos
1 More Replies
demo-user
by New Contributor III
  • 378 Views
  • 2 replies
  • 0 kudos

S3A Connector Trying to Use AWS STS on Non-AWS S3 Endpoint

Hi everyone,I’m trying to write Delta tables to my S3-compatible (non-AWS) endpoint, and it was writing perfectly fine last week with the same setup. Now, without any changes on my end, it’s failing and giving me anUnknownException: (com.amazonaws.se...

  • 378 Views
  • 2 replies
  • 0 kudos
Latest Reply
aleksandra_ch
Databricks Employee
  • 0 kudos

Hi @demo-user , Can you share more information about your setup: Cluster type and DBR versionS3-compatible storage implementation (MinIO / something else?) AFAIK this is not supposed to work as Delta client in DBR relies on AWS STS to perform S3 comm...

  • 0 kudos
1 More Replies
rwhitepwt
by New Contributor II
  • 516 Views
  • 4 replies
  • 2 kudos

Resolved! Netsuite Data Connector Not Available

I see that the Azure Databricks Data Connector for Netsuite is in Public Preview. Unfortunately I am unable to see it in my instance. I have gone into Preview and selected it as enabled, have downloaded the JAR file from Netsuite and have setup the i...

  • 516 Views
  • 4 replies
  • 2 kudos
Latest Reply
Ashwin_DSA
Databricks Employee
  • 2 kudos

Hi @rwhitepwt, From what I can see, having the NetSuite connector in Public Preview doesn’t automatically guarantee that the tile appears in every workspace. In addition to enabling the preview and creating the UC connection + uploading the SuiteAnal...

  • 2 kudos
3 More Replies
Malthe
by Valued Contributor II
  • 307 Views
  • 4 replies
  • 0 kudos

429 XHR requests against jobs endpoint (RESOURCE_EXHAUSTED)

The Databricks UI sends thousands of repeated queries on the form:/ajax-api/2.0/jobs/get?include_acls=true&job_id=<redacted>Getting a 429 Too Many Requests response. It seems to rotate on a list of job ids.It just keeps trying ...

  • 307 Views
  • 4 replies
  • 0 kudos
Latest Reply
Malthe
Valued Contributor II
  • 0 kudos

Just notify your engineers, they'll know how to handle this and won't need more information. It's a no-brainer to fix this.

  • 0 kudos
3 More Replies
ashraf1395
by Honored Contributor
  • 2411 Views
  • 4 replies
  • 0 kudos

Getting error while using Live.target_table in dlt pipeline

I have created a target table in the same dlt pipeline. But when I read that table in different block of notebook with Live.table_path. It is not able to read it Here is my code block 1 Creating a streaming table # Define metadata tables catalog = sp...

  • 2411 Views
  • 4 replies
  • 0 kudos
Latest Reply
IM_01
Contributor III
  • 0 kudos

Hi @ashraf1395 Were you able to add expectations to the append flow table. If yes could you please share the approach

  • 0 kudos
3 More Replies
toast_2001
by New Contributor II
  • 354 Views
  • 2 replies
  • 1 kudos

Resolved! Non-existent schema on redeployment of DAB with external volumes.

Hi all,DAB issue.My setup:Running CLI v0.294 on Python 3.12.11.Deployement is mode direct and using standard serverless compute.External locations in ADLS ST Container (container per ext loc).I'm attempting to deploy a bundle according to the followi...

config_file.png example_error.png
  • 354 Views
  • 2 replies
  • 1 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 1 kudos

Hi @toast_2001, I did some digging and have a few helpful tips/tricks to assist your troubleshooting. So let me walk through what's likely happening and what to actually do about it. The error tells you that on the second deployment, DAB is trying to...

  • 1 kudos
1 More Replies
Labels