cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

tana_sakakimiya
by Contributor
  • 938 Views
  • 1 replies
  • 2 kudos

Resolved! What is "External tables backed by Delta Lake"?

Goal: event-driven without implementing job triggereed on file arrivalI see hope to incrementally update materialized views which have external tables as their sources.This is quite a game changer if it works for various data formats.(since MV starte...

  • 938 Views
  • 1 replies
  • 2 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 2 kudos

Hi @tana_sakakimiya ,Yes, only external tables that are in delta format are supported. Databricks supports other table formats, but to be able to use this particular feature, your table needs to be in Delta format. But if you have parquet files it's ...

  • 2 kudos
andr3s
by New Contributor II
  • 44459 Views
  • 8 replies
  • 2 kudos

SSL_connect: certificate verify failed with Power BI

Hi, I'm getting this error with Power BI:Any ideas?Thanks in advance,Andres

Screenshot 2023-05-19 154328
  • 44459 Views
  • 8 replies
  • 2 kudos
Latest Reply
GaneshKrishnan
New Contributor II
  • 2 kudos

In the proxy setup, PowerBI is not aware of process to fetch intermediate certificate like a browser. hence it fails. Recent PowerBI comes with additional option such as"Automatic Proxy Discovery (Optional): Enabled"Implementation (optional) : 2.0(be...

  • 2 kudos
7 More Replies
cpayne_vax
by New Contributor III
  • 29296 Views
  • 16 replies
  • 9 kudos

Resolved! Delta Live Tables: dynamic schema

Does anyone know if there's a way to specify an alternate Unity schema in a DLT workflow using the @Dlt.table syntax? In my case, I’m looping through folders in Azure datalake storage to ingest data. I’d like those folders to get created in different...

  • 29296 Views
  • 16 replies
  • 9 kudos
Latest Reply
surajitDE
Contributor
  • 9 kudos

if you add these settings in the pipeline JSON, the issue should get fixed:"pipelines.setMigrationHints" = "true""pipelines.enableDPMForExistingPipeline" = "true"I tried it on my side, and now it no longer throws the materialization error.

  • 9 kudos
15 More Replies
JuanSeValencia
by New Contributor III
  • 21275 Views
  • 21 replies
  • 16 kudos

Resolved! Was DBFS disabled from community edition?

HelloI'm trying to use the Upload data to DBFS... from Databricks community edition but it's disabled. I'm also trying to activate it using the path Settings>Advanced>Other but the option is not longer in the list. Is this a temporary or permanent mo...

  • 21275 Views
  • 21 replies
  • 16 kudos
Latest Reply
lisagjh
New Contributor II
  • 16 kudos

Noticed there a several questions/'workarounds' on this topic. I may have overlooked but again, has this feature been removed or is it currently integrated naturally? Appreciate the assistance, in advance.

  • 16 kudos
20 More Replies
seefoods
by Valued Contributor
  • 912 Views
  • 2 replies
  • 1 kudos

Resolved! dashboard cost databricks - AWS instance

Hello guys, I hope your day going well ! I have some question about your dashboard cost publish on github databricks quick lab: this dashboard describe both the cost of databricks and AWS instance EC2 or just databricks cost ? Thanx

  • 912 Views
  • 2 replies
  • 1 kudos
Latest Reply
BS_THE_ANALYST
Databricks Partner
  • 1 kudos

@seefoods, also worth noting that you might not be able to see the system tables without a metastore admin granting the appropriate access. That was something I learned in Azure : https://docs.databricks.com/aws/en/admin/ All the best,BS

  • 1 kudos
1 More Replies
IONA
by New Contributor III
  • 2109 Views
  • 1 replies
  • 3 kudos

Move/Merge Catalogs

HiWe have a couple of unity catalogs. In each are schema's, some used, some old, same named temp_data etc. etc Within those schema are tables with the same erratic approach. The result of a couple of years of bad housekeeping. (who not experienced th...

  • 2109 Views
  • 1 replies
  • 3 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 3 kudos

For moving tables and entire schemas consider looking at the "SYNC" command.   The SYNC Command - Your Primary Migration Tool Databricks provides a SYNC command specifically designed for this exact use case - migrating tables from one location to ano...

  • 3 kudos
liu
by Databricks Partner
  • 1011 Views
  • 3 replies
  • 2 kudos

Permission issue for pandas to read local files

I can use pandas to read local files in a notebook, such as those located in tmp.However, when I run two consecutive notebooks within the same job and read files with pandas in both, I encounter a permission error in the second notebook stating that ...

  • 1011 Views
  • 3 replies
  • 2 kudos
Latest Reply
BS_THE_ANALYST
Databricks Partner
  • 2 kudos

@liu is there anything preventing you from moving this file into a Volume on the Unity Catalog & then reading it from there?Based on your original message, the "permission" message is making me feel like it's having an issue with opening the file; pe...

  • 2 kudos
2 More Replies
PratikRudra
by New Contributor
  • 1117 Views
  • 1 replies
  • 0 kudos

unable to create table on external location

Currently trying to connect a table on external location and it fails with error -[UNAUTHORIZED_ACCESS] Unauthorized access: PERMISSION_DENIED: request not authorized SQLSTATE: 42501which seems like a pretty straight forward error but unable to find ...

  • 1117 Views
  • 1 replies
  • 0 kudos
Latest Reply
Khaja_Zaffer
Esteemed Contributor
  • 0 kudos

Hello @PratikRudra Thank you for sharing the error: I think probably there is a component that is missing. Writing table metadata (for example, to the _delta_log directory) requires the CREATE EXTERNAL TABLE capability on the external location; this ...

  • 0 kudos
tana_sakakimiya
by Contributor
  • 1920 Views
  • 7 replies
  • 7 kudos

Resolved! MATERIALZIED VIEW TRIGGER ON UPDATE with external table as upstream table

goal: implement event driven architecture without trigger on file arrivalI would like to know whether materialzied view can update itself when its source table which is external table updated.given that the source external table referencing data in d...

  • 1920 Views
  • 7 replies
  • 7 kudos
Latest Reply
tana_sakakimiya
Contributor
  • 7 kudos

it seems that my idea is a bad idea because it seems that materialzied view doesn't support incremental udpate for external locationIncremental refresh for materialized views - Azure Databricks | Microsoft Learn

  • 7 kudos
6 More Replies
IONA
by New Contributor III
  • 672 Views
  • 1 replies
  • 0 kudos

Changing paths to tables

HiMy organization has many notebooks that reference tables in schemas with the three part pathcatalog.schema.tablenameWith a lack of foresight we hardcoded all of these paths in the code and now the inevitable is happening and there is a need to rest...

  • 672 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @IONA ,Definitely, I would say that’s even a common practice. Create a feature branch and make the necessary changes there. But once a day, merge into that feature branch all the changes that have appeared on your main branch. That way, you will a...

  • 0 kudos
Malthe
by Valued Contributor II
  • 2811 Views
  • 10 replies
  • 4 kudos

Resolved! Intermittent failures with auto loader on Azure

We've been using auto loader to ingest data from a storage account on Azure (format "cloudFiles").Today, we're starting to see failures during the setup of event notification:25/09/11 19:06:28 ERROR MicroBatchExecution: Non-interrupted exception thro...

  • 2811 Views
  • 10 replies
  • 4 kudos
Latest Reply
Khaja_Zaffer
Esteemed Contributor
  • 4 kudos

Hello @Malthe @Saska @MehdiJafariWhen it was showing such error, did you try any alternative methods like using databricks service credentials. For example  stream = spark.readStream.format("cloudFiles")  .option("cloudFiles.format", "json")  .option...

  • 4 kudos
9 More Replies
ManojkMohan
by Honored Contributor II
  • 1722 Views
  • 1 replies
  • 3 kudos

Resolved! Simulating Real Time Streaming in Databricks free edition

Use Case:Kafka real time steaming network telemetry logsIn a real use case approx. 40 TB of data can be real time streamed in a dayArchitecture:issue encountered:  when i try to simulate kakfa like streaming in databricks itself , as this is a free e...

ManojkMohan_0-1757780238326.png
  • 1722 Views
  • 1 replies
  • 3 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 3 kudos

Hi @ManojkMohan ,Unfortunately there is no workaround here. Free Edition supports only serverless compute. And serverless has following streaming limitation - one of which you just encountered - there is no support for default or time-based trigger i...

  • 3 kudos
LucasAntoniolli
by New Contributor III
  • 1581 Views
  • 4 replies
  • 4 kudos

Resolved! Problems with cluster shutdown in DLT

[Issue] DLT finishes processing, but cluster remains active due to log write errorHi everyone, I'm running into a problem with my DLT pipeline and was hoping someone here could help or has experienced something similar.Problem DescriptionThe pipeline...

  • 1581 Views
  • 4 replies
  • 4 kudos
Latest Reply
nayan_wylde
Esteemed Contributor II
  • 4 kudos

Can you please try one more option. If you’re on Preview, move to Current (or vice versa). Sometimes the regression only exists in one channel.

  • 4 kudos
3 More Replies
IONA
by New Contributor III
  • 812 Views
  • 1 replies
  • 0 kudos

Decoupling power bi reports from schemas

HiI'm sure many of you have power bi reports that use the native Databricks connector to pull data from a schema to fuel a wonderful dashboard. If the source schemes was moved to a different catalog or a table renamed then the pbi connection would be...

  • 812 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @IONA ,You can parametrize your data source using power query parameters. Just define parameter for catalog, schema etc. I have such setup in one of the project. We parametrized our PBI semantic models and when we are deploying semantic model to a...

  • 0 kudos
Labels