cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

aseufert
by New Contributor III
  • 2048 Views
  • 8 replies
  • 4 kudos

Dynamic Value References Not Working

I can't get the dynamic value references to work in my jobs. I can use the deprecate references (e.g. job_id) but not the new references (e.g. job.id). As a test, I set a text widget called MyJobID following the example that will receive the dynamic ...

  • 2048 Views
  • 8 replies
  • 4 kudos
Latest Reply
themattmorris
New Contributor III
  • 4 kudos

For what it's worth, it looks like job-level parameters were added with this update as well. I was wondering why I was unable to use those, but those are also working for me now.

  • 4 kudos
7 More Replies
thibault
by Contributor
  • 1198 Views
  • 1 replies
  • 0 kudos

Resolved! Import notebook content into a python file

Hi, I have a workflow based on python scripts. How can I import the content of a notebook where a class and functions are defined?I know how to import python files into notebooks, but the other way around doesn't seem as straight forward.

  • 1198 Views
  • 1 replies
  • 0 kudos
Latest Reply
thibault
Contributor
  • 0 kudos

Found a solution executing a notebook, using the databricks api to download the notebook content as bytes :1. set environment variables DATABRICKS_HOST and DATABRICKS_TOKEN2. w = WorkspaceClient()with w.workspace.download(notebook_path) as n: note...

  • 0 kudos
pgruetter
by Contributor
  • 1295 Views
  • 3 replies
  • 2 kudos

Help me understand streaming logic with Delta Tables

Hello allI have a delta table in bronze layer, let's call it BRZ. It contains 25B rows and many duplicates. It has a version 0 and a version 1, nothing else yet. I then create a silver table SLV by running one deduplication batch job. This creates ve...

  • 1295 Views
  • 3 replies
  • 2 kudos
Latest Reply
pgruetter
Contributor
  • 2 kudos

Thanks for the confirmation. Not sure I see everything as your text gets truncated, but it basically confirms that it should work.Anyway: It looks like the incremental load is working. The problem here is, that we receive late arriving facts that tou...

  • 2 kudos
2 More Replies
Nathant93
by New Contributor II
  • 1434 Views
  • 1 replies
  • 0 kudos

Resolved! Date formatting

Does anyone know how to change the format of a date like this Dec 17 2016 8:22PMinto yyyy-MM-dd hh:mm:ss?Thanks

  • 1434 Views
  • 1 replies
  • 0 kudos
Latest Reply
Krishnamatta
New Contributor III
  • 0 kudos

 Convert to timestamp first and then format to stringselect  date_format(to_timestamp('Dec 17 2016 8:22PM', 'MMM dd yyyy h:ma'), "yyyy-MM-dd HH:mm:ss")Here is the documentation for this:https://docs.databricks.com/en/sql/language-manual/sql-ref-datet...

  • 0 kudos
Randy
by New Contributor III
  • 616 Views
  • 1 replies
  • 0 kudos

Resolved! Unable to Write Table to Synapse 'x' has a data type that cannot participate in a columnstore index.

We have a process that creates a table in Synapse then attempts to write the Data generated in Databricks to it. We are able to create the table no problem but when we go to copy the data we keep getting an error that the column has a data type that ...

  • 616 Views
  • 1 replies
  • 0 kudos
Latest Reply
Randy
New Contributor III
  • 0 kudos

Resolved

  • 0 kudos
learnerbricks
by New Contributor II
  • 3405 Views
  • 4 replies
  • 0 kudos

Unable to save file in DBFS

I have took the azure datasets that are available for practice. I got the 10 days data from that dataset and now I want to save this data into DBFS in csv format. I have facing an error :" No such file or directory: 'No such file or directory: '/dbfs...

  • 3405 Views
  • 4 replies
  • 0 kudos
Latest Reply
pardosa
New Contributor II
  • 0 kudos

Hi,after some exercise you need to aware folder create in dbutils.fs.mkdirs("/dbfs/tmp/myfolder") it's created in /dbfs/dbfs/tmp/myfolderif you want to access path to_csv("/dbfs/tmp/myfolder/mytest.csv") you should created with this script dbutils.fs...

  • 0 kudos
3 More Replies
knutasm
by New Contributor III
  • 2418 Views
  • 6 replies
  • 6 kudos

Run Delta Live Tables as service principal

How to run a delta live tables pipeline in production? It uses the owner's (creator's) permissions for writing to tables, and I can't change the owner of a UC-enabled pipeline after creation. I don't want regular users to have write access to prod ta...

  • 2418 Views
  • 6 replies
  • 6 kudos
Latest Reply
js54123875
New Contributor III
  • 6 kudos

same!

  • 6 kudos
5 More Replies
ilarsen
by Contributor
  • 916 Views
  • 2 replies
  • 1 kudos

Auto Loader and source file structure optimisation

Hi.  I have a question, and I've not been able to find an answer.  I'm sure there is one...I just haven't found it through searching and browsing the docs. How much does it matter (if it is indeed that simple) if source files read by auto loader are ...

  • 916 Views
  • 2 replies
  • 1 kudos
Latest Reply
Kaniz
Community Manager
  • 1 kudos

Hi @ilarsen, According to the Azure Databricks documentation, Auto Loader incrementally and efficiently processes new data files as they arrive in cloud storage. Auto Loader can load data files from Azure data lake Storage Gen2 (ADLS Gen2) using hier...

  • 1 kudos
1 More Replies
Rubini_MJ
by New Contributor
  • 3264 Views
  • 1 replies
  • 0 kudos

Resolved! Other memory of the driver is high even in a newly spun cluster

Hi Team Experts,    I am experiencing a high memory consumption in the other part in the memory utilization part in the metrics tab. Right now am not running any jobs but still out of 8gb driver memory 6gb is almost full by other and only 1.5 gb is t...

  • 3264 Views
  • 1 replies
  • 0 kudos
Latest Reply
User16539034020
Contributor II
  • 0 kudos

Hello,  Thanks for contacting Databricks Support.  Seems you are concern with high memory consumption in the "other" category in the driver node of a Spark cluster. As there are no logs/detail information provided, I only can address several potentia...

  • 0 kudos
houstonamoeba
by New Contributor III
  • 2070 Views
  • 7 replies
  • 1 kudos

Resolved! examples on python sdk for install libraries

Hi Everyone,I'm planning to use databricks python cli "install_libraries"can some one pls post examples on function install_libraries https://github.com/databricks/databricks-cli/blob/main/databricks_cli/libraries/api.py

  • 2070 Views
  • 7 replies
  • 1 kudos
Latest Reply
Loop-Insist
New Contributor II
  • 1 kudos

Here you go using Python SDKfrom databricks.sdk import WorkspaceClientfrom databricks.sdk.service import computew = WorkspaceClient(host="yourhost", token="yourtoken")# Create an array of Library objects to be installedlibraries_to_install = [compute...

  • 1 kudos
6 More Replies
JVesely
by New Contributor III
  • 631 Views
  • 1 replies
  • 0 kudos

Resolved! DLT CDC SCD-1 pipeline not showing stats when reading from parquet file

Hi,I followed the tutorial here: https://docs.databricks.com/en/delta-live-tables/cdc.html#how-is-cdc-implemented-with-delta-live-tablesThe only change I did is that data is not appended to a table but is read from a parquet file. In practice this me...

  • 631 Views
  • 1 replies
  • 0 kudos
Latest Reply
JVesely
New Contributor III
  • 0 kudos

My bad - waiting a bit and doing a proper screen refresh does show the numbers. 

  • 0 kudos
Anonymous
by Not applicable
  • 3446 Views
  • 8 replies
  • 2 kudos
  • 3446 Views
  • 8 replies
  • 2 kudos
Latest Reply
djhs
New Contributor III
  • 2 kudos

I also tried to leverage this endpoint (inferred from devtools): https://<workspace_id>.cloud.databricks.com/sql/api/dashboards/import with the exported dashboard (the dbdash file) in the request payload. It returns a 200 but nothing happens. Maybe s...

  • 2 kudos
7 More Replies
rt-slowth
by Contributor
  • 1909 Views
  • 5 replies
  • 1 kudos

Resolved! CRAS in @dlt

The Delta Table created as a result of the Dataframe returned by @dlt.create_table is confirmed to be overwritten when checked with the DECREASE HISTORY command.I want this to be handled as a CRAS, or CREATE AS SELECT, but how can I do this in python...

  • 1909 Views
  • 5 replies
  • 1 kudos
Latest Reply
siddhathPanchal
New Contributor III
  • 1 kudos

Hi @rt-slowth You can review this open source code base of Delta to know more about the DeltaTableBuilder's implementation in Python.  https://github.com/delta-io/delta/blob/master/python/delta/tables.py

  • 1 kudos
4 More Replies
msj50
by New Contributor III
  • 7450 Views
  • 11 replies
  • 1 kudos

Spark Running Really slow - help required

My company urgently needs help, we are having severe performance problems with spark and are having to switch to a different solution if we don't get to the bottom of it. We are on 1.3.1, using spark SQL, ORC Files with partitions and caching in me...

  • 7450 Views
  • 11 replies
  • 1 kudos
Latest Reply
Kaniz
Community Manager
  • 1 kudos

Hi @msj50 , Thank you for posting your question in our community! We are happy to assist you. To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers your ...

  • 1 kudos
10 More Replies
Labels
Top Kudoed Authors