cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

kranthi2
by New Contributor III
  • 1924 Views
  • 2 replies
  • 2 kudos

Resolved! alter DLT Materialized View alter column set MASK

I am trying to mask a column on a DLT materialized view - this is created using DLT syntax. I am not able set the column masking after creation. Appreciate any workaround.alter DLT Materialized View alter column set MASK

  • 1924 Views
  • 2 replies
  • 2 kudos
Latest Reply
kranthi2
New Contributor III
  • 2 kudos

Thank you. I will submit the idea.

  • 2 kudos
1 More Replies
prasadvaze
by Valued Contributor II
  • 29826 Views
  • 15 replies
  • 12 kudos

Resolved! How to query delta lake using SQL desktop tools like SSMS or DBVisualizer

Is there a way to use sql desktop tools? because delta OSS or databricks does not provide desktop client (similar to azure data studio) to browse and query delta lake objects.I currently use databricks SQL , a webUI in the databricks workspace but se...

  • 29826 Views
  • 15 replies
  • 12 kudos
Latest Reply
prasadvaze
Valued Contributor II
  • 12 kudos

DSR is Delta Standalone Reader. see more here - https://docs.delta.io/latest/delta-standalone.htmlIts a crate (and also now a py library) that allows you to connect to delta tables without using spark (e.g. directly from python and not using pyspa...

  • 12 kudos
14 More Replies
oleh_v
by New Contributor
  • 1108 Views
  • 2 replies
  • 0 kudos

Upload of .bin file >400mb

I try to upload to local workspace folder with .bin extension.It is required to have it locally.I tried load from DBFS, but loading files over 265mb is not allowed with cluster. I tried to upload manually but failed with same error "OSError: [Errno5]...

  • 1108 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kartheek_Katta
New Contributor II
  • 0 kudos

Hello Slash,Thank you for your response. I'm encountering the same issue as described. I tried running the provided code in my Databricks workspace, but I received an error. My question is how the script is expected to access local files, especially ...

  • 0 kudos
1 More Replies
Meghana_Vasavad
by New Contributor III
  • 2271 Views
  • 4 replies
  • 0 kudos

Resolved! Discrepancies in Data Engineering GitHub Repositories and Errors in Databricks Notebooks

Hi team,I am writing to express my concerns regarding the recent Databricks webinar on "Data Intelligence with Databricks". During the session, there was mention of two GitHub repositories associated with the notebooks being used as mentioned below.1...

  • 2271 Views
  • 4 replies
  • 0 kudos
Latest Reply
Sujitha
Databricks Employee
  • 0 kudos

Hi @Meghana_Vasavad could you please file a ticket with Databricks SupportThey will help you with this request. 

  • 0 kudos
3 More Replies
raghu2
by New Contributor III
  • 4571 Views
  • 5 replies
  • 0 kudos

Resolved! Error deploying a DAB

I followed steps listed in this article.After creating and validation of bundle with default template, during deployment using this command:databricks bundle deploy -t dev --profile zzI get this message:Building mySecPrj...Error: build failed mySecPr...

  • 4571 Views
  • 5 replies
  • 0 kudos
Latest Reply
ADB0513
New Contributor III
  • 0 kudos

@daniel_sahal I am receiving the same error and tried your solution and am still getting the invalid command bdist_wheel.  Any other suggestions?  Thanks

  • 0 kudos
4 More Replies
brickster_2018
by Databricks Employee
  • 20448 Views
  • 3 replies
  • 1 kudos
  • 20448 Views
  • 3 replies
  • 1 kudos
Latest Reply
Hugh_Ku
New Contributor II
  • 1 kudos

I've also run into the same issue, customised docker image does not give DATABRICKS_RUNTIME_VERSION as env. I believe there are still many issues in how customised docker image is used in databricks cluster.Can anyone from databricks help answer it?

  • 1 kudos
2 More Replies
varshini_reddy
by New Contributor III
  • 2519 Views
  • 6 replies
  • 0 kudos

Databricks UC enabled but Lineage not found for one table

Databricks UC enabled but Lineage not found for one table whereas i can see the lineage for the other two, any idea on why is it?. Im performing few transformations to bronze data , taking good_data_transformed as a dataframe, creating a temp view fo...

  • 2519 Views
  • 6 replies
  • 0 kudos
Latest Reply
filipniziol
Esteemed Contributor
  • 0 kudos

It is because of using temp view. To debug further you would need to write all the source tables, transformations, target tables, actual lineage and expected lineage, but as a rule of thumb if the lineage is lost when using temp view.Lineage is captu...

  • 0 kudos
5 More Replies
karolinalbinsso
by New Contributor II
  • 4198 Views
  • 2 replies
  • 3 kudos

Resolved! How to access the job-Scheduling Date from within the notebook?

I have created a job that contains a notebook that reads a file from Azure Storage. The file-name contains the date of when the file was transferred to the storage. A new file arrives every Monday, and the read-job is scheduled to run every Monday. I...

  • 4198 Views
  • 2 replies
  • 3 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 3 kudos

Hi, I guess the files are in the same directory structure so that you can use cloud files autoloader. It will incrementally read only new files https://docs.microsoft.com/en-us/azure/databricks/spark/latest/structured-streaming/auto-loaderSo it will ...

  • 3 kudos
1 More Replies
csmcpherson
by New Contributor III
  • 1427 Views
  • 1 replies
  • 1 kudos

Resolved! Workflow file watch - capture filename trigger

With respect to the file watch trigger in workflows, how can we capture what files and or path was identified as raising the trigger?  I'd like to use this information to set parameters based upon the file name and the file path  Thank you!  https://...

  • 1427 Views
  • 1 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @csmcpherson ,This is currently not supported, but databricks team is working on that idea according to below thread:Solved: File information is not passed to trigger job on f... - Databricks Community - 39266As a workaround, if you use autoloader...

  • 1 kudos
Govardhana
by New Contributor
  • 5318 Views
  • 1 replies
  • 1 kudos

Interview question for ADF

Hello,I am trying to attending interviews for Data Engineer i have 3 years Experince I am looking realtime interview question ,If any one have could you please share.Thank you,Govardhana

  • 5318 Views
  • 1 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @Govardhana ,There are plenty of those question on the internet. Below, is the one that's actually quite good:- Top 40+ Azure Data Factory Interview Questions 2024 (k21academy.com)

  • 1 kudos
Djelany
by New Contributor II
  • 4889 Views
  • 3 replies
  • 1 kudos

Resolved! DLT Event Logs

Hi,Does anyone know what details:planning_information:technique_information[0]:cost under planning_information event type means in my DLT workflow system event logs? For context, I'm trying to track the cost per run of my DLT workflow and I do not ha...

  • 4889 Views
  • 3 replies
  • 1 kudos
Latest Reply
adriennn
Valued Contributor
  • 1 kudos

you can enable system.billing schema and see the costs of the runs from the usage table.

  • 1 kudos
2 More Replies
jay971
by New Contributor II
  • 2732 Views
  • 3 replies
  • 0 kudos

Error: Cannot use legacy parameters because the job has job parameters configured.

I created a job which has two Job Parameters. How can I use Databricks CLI to pass different values to those parameters. 

  • 2732 Views
  • 3 replies
  • 0 kudos
Latest Reply
jay971
New Contributor II
  • 0 kudos

The job ran but did not pick up the values from the CLI.

  • 0 kudos
2 More Replies
Saf4Databricks
by New Contributor III
  • 3288 Views
  • 2 replies
  • 0 kudos

Reading single file from Databricks DBFS

I have a Test.csv file in FileStore of DBFS in Databricks Community edition. When I try to read the file using With Open, I get the following error:FileNotFoundError: [Errno 2] No such file or directory: '/dbfs/FileStore/tables/Test.csv' import os wi...

  • 3288 Views
  • 2 replies
  • 0 kudos
Latest Reply
Saf4Databricks
New Contributor III
  • 0 kudos

@EricRM It should work. Please see the accepted response from this same forum here. So, we still need to find a cause of the error. Following is the detailed error message. Maybe, this will help readers understand the issue better and help it resolve...

  • 0 kudos
1 More Replies
databicky
by Contributor II
  • 22366 Views
  • 13 replies
  • 4 kudos
  • 22366 Views
  • 13 replies
  • 4 kudos
Latest Reply
FerArribas
Contributor
  • 4 kudos

Hi @Hubert Dudek​,​Pandas API doesn't support abfss protocol.You have three options:​If you need to use pandas, you can write the excel to the local file system (dbfs) and then move it to ABFSS (for example with dbutils)Write as csv directly in abfss...

  • 4 kudos
12 More Replies
sakuraDev
by New Contributor II
  • 1202 Views
  • 1 replies
  • 2 kudos

Resolved! how does autoloader handle source outage

Hey guys,I've been looking for some docs on how autoloader manages the source outage, I am currently running the following code: dfBronze = (spark.readStream .format("cloudFiles") .option("cloudFiles.format", "json") .schema(json_schema_b...

sakuraDev_0-1725478024362.png
  • 1202 Views
  • 1 replies
  • 2 kudos
Latest Reply
filipniziol
Esteemed Contributor
  • 2 kudos

Hi @sakuraDev ,1. Using the availableNow trigger to process all available data immediately and then stop the query. As you noticed your data was processed once and now you need to trigger the process once again to process new files.2. Changing the tr...

  • 2 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels