cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

IM_01
by Contributor III
  • 559 Views
  • 8 replies
  • 0 kudos
  • 559 Views
  • 8 replies
  • 0 kudos
Latest Reply
IM_01
Contributor III
  • 0 kudos

Hi @Ashwin_DSA Thanks for the response â€Œâ€ŒI was thinking if the results are precomputed using cube and persisted using mv that would be retrieve results faster than metric view. Could u please let me know if my understanding is correct

  • 0 kudos
7 More Replies
Diehl
by New Contributor III
  • 192 Views
  • 1 replies
  • 0 kudos

Resolved! Auto Loader with ignoreMissingFiles and useManagedFileEvents fails on Classic Compute

Hi everyone,I am seeing an unexpected behavior with Auto Loader when using Managed File Events on Classic Compute.The error message itself seems inconsistent with the behavior I am seeing:[FAILED_READ_FILE.DBR_FILE_NOT_EXIST] Error while reading file...

  • 192 Views
  • 1 replies
  • 0 kudos
Latest Reply
Diehl
New Contributor III
  • 0 kudos

Just sharing a solution in case anyone runs into the same issue.The error was caused by the cluster configuration including spark.master: "local[*]". After removing this setting, the error stopped occurring and the Auto Loader finished correctly.This...

  • 0 kudos
lrm_data
by New Contributor
  • 182 Views
  • 2 replies
  • 0 kudos

Lakeflow Connect - SQL Server - Issues restarting after failure

Has anyone else run into a situation where a breaking schema change on a SQL Server source table leaves their Lakeflow Connect pipeline in a state it can't recover from — even after destroying and recreating the pipeline?Here's what happened to us:- ...

  • 182 Views
  • 2 replies
  • 0 kudos
Latest Reply
abhi_dabhi
Databricks Partner
  • 0 kudos

Hi @lrm_data yes, this one catches a lot of people. A few things to check on the SQL Server side that commonly block recovery even after destroy + recreate:Stale lakeflow_* capture instance. SQL Server allows only 2 capture instances per table. If bo...

  • 0 kudos
1 More Replies
vg33
by New Contributor
  • 116 Views
  • 1 replies
  • 0 kudos

Network Configuration

I have a Databricks workspace on AWS (serverless compute). I created a network policy with "Allow access to all destinations" enabled and attached it to my workspace. When I run a Python notebook and try to make an HTTP request or curl to any externa...

  • 116 Views
  • 1 replies
  • 0 kudos
Latest Reply
Lu_Wang_ENB_DBX
Databricks Employee
  • 0 kudos

Most likely the egress policy change hasn’t actually taken effect on the serverless compute that’s running your notebook. Check these things in order: Verify the network policy itself (Account Console → Security → Networking → Context-based ingress ...

  • 0 kudos
TX-Aggie-00
by Databricks Partner
  • 99 Views
  • 0 replies
  • 0 kudos

Sharepoint Connector Site Limitation

Hey All!We are trying out the Beta connector for SharePoint and found that the connector will not work at the root-level site.  Is there a reason for this limitation.  It is unfortunately a hard blocker for us to use the native connector.  MUST_START...

  • 99 Views
  • 0 replies
  • 0 kudos
mdee
by Databricks Partner
  • 107 Views
  • 2 replies
  • 1 kudos

LDP Materialized View Incremental Refreshes - Changeset Size Thresholds

Is there any documentation available around the changeset size thresholds for materialized view incremental refreshes?  Are these configurable at all?  Are they constant or do the thresholds change depending on the number of rows/size of the material...

  • 107 Views
  • 2 replies
  • 1 kudos
Latest Reply
emma_s
Databricks Employee
  • 1 kudos

Hi, On top of Pradeep's reply, which I'd recommend trying, I'd also suggest you raise a support ticket for this. They will potentially be able to tweak the settings in the backend (not guaranteed), but it may help. Thanks,Emma

  • 1 kudos
1 More Replies
lrm_data
by New Contributor
  • 82 Views
  • 1 replies
  • 0 kudos

**Lakeflow Connect SQL Server — Snapshots Firing Outside Configured Full Refresh Window?**

Has anyone else seen full refresh snapshots trigger outside of their configured refresh window in Lakeflow Connect?Here's our situation:- We have a full refresh window configured to restrict snapshot operations to off-hours- On at least one occasion,...

  • 82 Views
  • 1 replies
  • 0 kudos
Latest Reply
Sumit_7
Honored Contributor II
  • 0 kudos

@lrm_data This is very unlike case for the refresh to be triggered outside the configured window. Though I would still suggest to check the Configured Window and Auto Full Refresh policy once to be sure.If still persists, then you may raise a support...

  • 0 kudos
MyProfile
by New Contributor
  • 102 Views
  • 1 replies
  • 0 kudos

Disable Public Network Access on Databricks Managed Storage Account - Deny Assignment

Issue Description:I am attempting to disable public network access on the Azure Databricks managed storage account. However, I am encountering the following error:Failed to save resource settings — access is denied due to a deny assignment created by...

  • 102 Views
  • 1 replies
  • 0 kudos
Latest Reply
Sumit_7
Honored Contributor II
  • 0 kudos

@MyProfile This would be helpful, check once - https://learn.microsoft.com/en-us/answers/questions/1707749/managed-storage-accounts-compliance

  • 0 kudos
Raghu_Bindingan
by New Contributor III
  • 5551 Views
  • 5 replies
  • 2 kudos

Truncate delta live table and try to repopulate it in the pipeline

Has anyone attempted to truncate a delta live gold level table that gets populated via a pipeline and then tried to repopulate it by starting the pipeline. I have this situation wherein i need to reprocess all data in my gold table, so i stopped the ...

  • 5551 Views
  • 5 replies
  • 2 kudos
Latest Reply
sanjivsingh
New Contributor
  • 2 kudos

My Blog on thishttps://medium.com/@singh.sanjiv/truncate-and-load-streaming-live-table-8f840eb424d1

  • 2 kudos
4 More Replies
leopold_cudzik
by New Contributor II
  • 140 Views
  • 1 replies
  • 0 kudos

Resolved! Lakehouse sync tables over rolling history

Hi,we're exploring replacing one of the use cases we are running in our clour provider with a Databricks pipelines. We currently have explored possibility to subscribe to an eventhub using SDP pipelines, feedding our iot data into a Delta table where...

  • 140 Views
  • 1 replies
  • 0 kudos
Latest Reply
Ashwin_DSA
Databricks Employee
  • 0 kudos

Hi @leopold_cudzik, The pattern you are suggesting is feasible, but it’s much easier to manage if you separate history ingestion from the 7-day serving view instead of cleaning the streaming sink table in place. A common architecture on Databricks wo...

  • 0 kudos
kevinleindecker
by New Contributor II
  • 411 Views
  • 6 replies
  • 1 kudos

SQL Warehouse error: "Cannot read properties of undefined (reading 'data')" when querying system tab

Queries that previously worked started failing in SQL Warehouse (Dashboards) without any changes on our side.The query succeeds, but fails to render results with error:"Cannot read properties of undefined (reading 'data')"This happens with:- system.b...

  • 411 Views
  • 6 replies
  • 1 kudos
Latest Reply
Esgario
New Contributor II
  • 1 kudos

Same problem here. I have previously reported this issue, and it had been resolved at the time. However, the problem has now reoccurred.When ingesting large tables (over 100k rows), the system is unable to properly render the data, preventing the tab...

  • 1 kudos
5 More Replies
DineshOjha
by New Contributor III
  • 105 Views
  • 1 replies
  • 0 kudos

Environment-Specific Schemas in SQL Files

Hi Databricks Community,We are looking for the recommended best practice for managing environment-specific configurations (catalogs and schemas) within our SQL files when using Databricks Asset Bundles.Our Setup:We use Databricks Asset Bundles to pac...

  • 105 Views
  • 1 replies
  • 0 kudos
Latest Reply
lingareddy_Alva
Esteemed Contributor
  • 0 kudos

Hi @DineshOjha The best approach is parameterized SQL with widget-based defaults in your Python wrapper, wired to DABs target variables. Why this works on both fronts: Engineers run the notebook interactively and widget defaults kick in (dev values)....

  • 0 kudos
PNC
by Databricks Partner
  • 171 Views
  • 4 replies
  • 0 kudos

Materialized view creation fails

Hi,I have ran into a problem when creating materialized view.Here's my simple query I'm trying to run:%sql create or replace materialized view catalog.schema.mView_test as select * from catalog.schema.table limit 10;I'm getting following error:Encoun...

  • 171 Views
  • 4 replies
  • 0 kudos
Latest Reply
balajij8
Contributor
  • 0 kudos

There are multiple requirements for materialized views. You can check belowYou must use a Unity Catalog enabled pro or serverless SQL warehouse.To incrementally refresh a materialized view from Delta tables, the source tables must have row tracking e...

  • 0 kudos
3 More Replies
Labels