cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

cpayne_vax
by New Contributor III
  • 29289 Views
  • 16 replies
  • 9 kudos

Resolved! Delta Live Tables: dynamic schema

Does anyone know if there's a way to specify an alternate Unity schema in a DLT workflow using the @Dlt.table syntax? In my case, I’m looping through folders in Azure datalake storage to ingest data. I’d like those folders to get created in different...

  • 29289 Views
  • 16 replies
  • 9 kudos
Latest Reply
surajitDE
Contributor
  • 9 kudos

if you add these settings in the pipeline JSON, the issue should get fixed:"pipelines.setMigrationHints" = "true""pipelines.enableDPMForExistingPipeline" = "true"I tried it on my side, and now it no longer throws the materialization error.

  • 9 kudos
15 More Replies
JuanSeValencia
by New Contributor III
  • 21258 Views
  • 21 replies
  • 16 kudos

Resolved! Was DBFS disabled from community edition?

HelloI'm trying to use the Upload data to DBFS... from Databricks community edition but it's disabled. I'm also trying to activate it using the path Settings>Advanced>Other but the option is not longer in the list. Is this a temporary or permanent mo...

  • 21258 Views
  • 21 replies
  • 16 kudos
Latest Reply
lisagjh
New Contributor II
  • 16 kudos

Noticed there a several questions/'workarounds' on this topic. I may have overlooked but again, has this feature been removed or is it currently integrated naturally? Appreciate the assistance, in advance.

  • 16 kudos
20 More Replies
seefoods
by Valued Contributor
  • 909 Views
  • 2 replies
  • 1 kudos

Resolved! dashboard cost databricks - AWS instance

Hello guys, I hope your day going well ! I have some question about your dashboard cost publish on github databricks quick lab: this dashboard describe both the cost of databricks and AWS instance EC2 or just databricks cost ? Thanx

  • 909 Views
  • 2 replies
  • 1 kudos
Latest Reply
BS_THE_ANALYST
Databricks Partner
  • 1 kudos

@seefoods, also worth noting that you might not be able to see the system tables without a metastore admin granting the appropriate access. That was something I learned in Azure : https://docs.databricks.com/aws/en/admin/ All the best,BS

  • 1 kudos
1 More Replies
IONA
by New Contributor III
  • 2101 Views
  • 1 replies
  • 3 kudos

Move/Merge Catalogs

HiWe have a couple of unity catalogs. In each are schema's, some used, some old, same named temp_data etc. etc Within those schema are tables with the same erratic approach. The result of a couple of years of bad housekeeping. (who not experienced th...

  • 2101 Views
  • 1 replies
  • 3 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 3 kudos

For moving tables and entire schemas consider looking at the "SYNC" command.   The SYNC Command - Your Primary Migration Tool Databricks provides a SYNC command specifically designed for this exact use case - migrating tables from one location to ano...

  • 3 kudos
liu
by Databricks Partner
  • 1009 Views
  • 3 replies
  • 2 kudos

Permission issue for pandas to read local files

I can use pandas to read local files in a notebook, such as those located in tmp.However, when I run two consecutive notebooks within the same job and read files with pandas in both, I encounter a permission error in the second notebook stating that ...

  • 1009 Views
  • 3 replies
  • 2 kudos
Latest Reply
BS_THE_ANALYST
Databricks Partner
  • 2 kudos

@liu is there anything preventing you from moving this file into a Volume on the Unity Catalog & then reading it from there?Based on your original message, the "permission" message is making me feel like it's having an issue with opening the file; pe...

  • 2 kudos
2 More Replies
PratikRudra
by New Contributor
  • 1117 Views
  • 1 replies
  • 0 kudos

unable to create table on external location

Currently trying to connect a table on external location and it fails with error -[UNAUTHORIZED_ACCESS] Unauthorized access: PERMISSION_DENIED: request not authorized SQLSTATE: 42501which seems like a pretty straight forward error but unable to find ...

  • 1117 Views
  • 1 replies
  • 0 kudos
Latest Reply
Khaja_Zaffer
Esteemed Contributor
  • 0 kudos

Hello @PratikRudra Thank you for sharing the error: I think probably there is a component that is missing. Writing table metadata (for example, to the _delta_log directory) requires the CREATE EXTERNAL TABLE capability on the external location; this ...

  • 0 kudos
tana_sakakimiya
by Contributor
  • 1914 Views
  • 7 replies
  • 7 kudos

Resolved! MATERIALZIED VIEW TRIGGER ON UPDATE with external table as upstream table

goal: implement event driven architecture without trigger on file arrivalI would like to know whether materialzied view can update itself when its source table which is external table updated.given that the source external table referencing data in d...

  • 1914 Views
  • 7 replies
  • 7 kudos
Latest Reply
tana_sakakimiya
Contributor
  • 7 kudos

it seems that my idea is a bad idea because it seems that materialzied view doesn't support incremental udpate for external locationIncremental refresh for materialized views - Azure Databricks | Microsoft Learn

  • 7 kudos
6 More Replies
IONA
by New Contributor III
  • 670 Views
  • 1 replies
  • 0 kudos

Changing paths to tables

HiMy organization has many notebooks that reference tables in schemas with the three part pathcatalog.schema.tablenameWith a lack of foresight we hardcoded all of these paths in the code and now the inevitable is happening and there is a need to rest...

  • 670 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @IONA ,Definitely, I would say that’s even a common practice. Create a feature branch and make the necessary changes there. But once a day, merge into that feature branch all the changes that have appeared on your main branch. That way, you will a...

  • 0 kudos
Malthe
by Valued Contributor II
  • 2807 Views
  • 10 replies
  • 4 kudos

Resolved! Intermittent failures with auto loader on Azure

We've been using auto loader to ingest data from a storage account on Azure (format "cloudFiles").Today, we're starting to see failures during the setup of event notification:25/09/11 19:06:28 ERROR MicroBatchExecution: Non-interrupted exception thro...

  • 2807 Views
  • 10 replies
  • 4 kudos
Latest Reply
Khaja_Zaffer
Esteemed Contributor
  • 4 kudos

Hello @Malthe @Saska @MehdiJafariWhen it was showing such error, did you try any alternative methods like using databricks service credentials. For example  stream = spark.readStream.format("cloudFiles")  .option("cloudFiles.format", "json")  .option...

  • 4 kudos
9 More Replies
ManojkMohan
by Honored Contributor II
  • 1722 Views
  • 1 replies
  • 3 kudos

Resolved! Simulating Real Time Streaming in Databricks free edition

Use Case:Kafka real time steaming network telemetry logsIn a real use case approx. 40 TB of data can be real time streamed in a dayArchitecture:issue encountered:  when i try to simulate kakfa like streaming in databricks itself , as this is a free e...

ManojkMohan_0-1757780238326.png
  • 1722 Views
  • 1 replies
  • 3 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 3 kudos

Hi @ManojkMohan ,Unfortunately there is no workaround here. Free Edition supports only serverless compute. And serverless has following streaming limitation - one of which you just encountered - there is no support for default or time-based trigger i...

  • 3 kudos
LucasAntoniolli
by New Contributor III
  • 1577 Views
  • 4 replies
  • 4 kudos

Resolved! Problems with cluster shutdown in DLT

[Issue] DLT finishes processing, but cluster remains active due to log write errorHi everyone, I'm running into a problem with my DLT pipeline and was hoping someone here could help or has experienced something similar.Problem DescriptionThe pipeline...

  • 1577 Views
  • 4 replies
  • 4 kudos
Latest Reply
nayan_wylde
Esteemed Contributor II
  • 4 kudos

Can you please try one more option. If you’re on Preview, move to Current (or vice versa). Sometimes the regression only exists in one channel.

  • 4 kudos
3 More Replies
IONA
by New Contributor III
  • 810 Views
  • 1 replies
  • 0 kudos

Decoupling power bi reports from schemas

HiI'm sure many of you have power bi reports that use the native Databricks connector to pull data from a schema to fuel a wonderful dashboard. If the source schemes was moved to a different catalog or a table renamed then the pbi connection would be...

  • 810 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @IONA ,You can parametrize your data source using power query parameters. Just define parameter for catalog, schema etc. I have such setup in one of the project. We parametrized our PBI semantic models and when we are deploying semantic model to a...

  • 0 kudos
RevanthV
by Contributor
  • 1185 Views
  • 3 replies
  • 6 kudos

Resolved! Issue with Auto Liquid clustering

 I have written data to a table using clusterByAuto set to trueBut the clustering keys are not selected automatically when i do a desc detail on the table.Screenshot belowWhy are clustering columns not being selected automatically?Repro steps:Create ...

  • 1185 Views
  • 3 replies
  • 6 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 6 kudos

Hi @RevanthV ,As @K_Anudeep  correctly suggested it could be the case that your table is to small to benefit from liquid clustering.Another possibility it that you're using runtime lower than 15.4 LTS.

  • 6 kudos
2 More Replies
Soumenkumar
by New Contributor
  • 696 Views
  • 1 replies
  • 2 kudos
  • 696 Views
  • 1 replies
  • 2 kudos
Latest Reply
K_Anudeep
Databricks Employee
  • 2 kudos

Hello @Soumenkumar ,I believe OGG has a Databricks target that stages to cloud storage (ADLS on S3) and then runs MERGE into Delta tables. This is designed for UC and documented in the official Oracle Docs.In Unity Catalog, create a storage credentia...

  • 2 kudos
Labels