cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Gilg
by Contributor II
  • 3267 Views
  • 2 replies
  • 0 kudos

APPLY_CHANGES with json data

Hi Team,I am building a DLT pipeline and planning to use APPLY_CHANGES from Bronze to Silver. In the bronze table, a column has a json value. This value contains questions and answers as key, value pair and can change depending on list of questions h...

  • 3267 Views
  • 2 replies
  • 0 kudos
SaraCorralLou
by New Contributor III
  • 38897 Views
  • 5 replies
  • 2 kudos

Resolved! Error: The spark driver has stopped unexpectedly and is restarting. Your notebook will be automatically reattached.

What is the problem?I am getting this error every time I run a python notebook on my Repo in Databricks.BackgroundThe notebook where I am getting the error is a notebook that creates a dataframe and the last step is to write the dataframe to a Delta ...

  • 38897 Views
  • 5 replies
  • 2 kudos
Latest Reply
Anonymous
Not applicable
  • 2 kudos

Hi @Sara Corral​ Thank you for posting your question in our community! We are happy to assist you.To help us provide you with the most accurate information, could you please take a moment to review the responses and select the one that best answers y...

  • 2 kudos
4 More Replies
Kratik
by New Contributor III
  • 2522 Views
  • 0 replies
  • 0 kudos

Spark submit job running python file

I have a spark submit job which is running one python file called main.py.The other file is alert.py which is being imported in main.py.Also main.py is using multiple config files.Alert.py is passed in --py-files and other config files are passed as ...

Data Engineering
pyfiles
spark
submit
  • 2522 Views
  • 0 replies
  • 0 kudos
Anonymous
by Not applicable
  • 5772 Views
  • 3 replies
  • 2 kudos
  • 5772 Views
  • 3 replies
  • 2 kudos
Latest Reply
sajith_appukutt
Databricks Employee
  • 2 kudos

User sessions automatically timeout after six hours of idle time. This is not configurable like @Kunal Gaurav​  mentioned. Please raise a feature request if you have a requirement to configure this.Now, in Azure you could configure AAD refresh token ...

  • 2 kudos
2 More Replies
alexkit
by New Contributor II
  • 3856 Views
  • 4 replies
  • 3 kudos

ASP1.2 Error create database in Spark Programming with Databricks training

I'm on Demo and Lab in Dataframes section. I've imported the dbc into my company cluster and has run "%run ./Includes/Classroom-Setup" successfully. When i run the 1st sql command %sql CREATE TABLE IF NOT EXISTS events USING parquet OPTIONS (path "/m...

  • 3856 Views
  • 4 replies
  • 3 kudos
Latest Reply
KDOCKX
New Contributor II
  • 3 kudos

I had the same issue and solved it like this:In the includes folder, there is a reset notebook, run the first command, this unmounts all mounted databases.Go back to the ASP 1.2 notebook and run the %run ./Includes/Classroom-Setup codeblock.Then run ...

  • 3 kudos
3 More Replies
MonishKumar
by New Contributor
  • 3153 Views
  • 1 replies
  • 0 kudos

SFTP - JSchException: Algorithm negotiation fail

When I tried to read the SFTP (CSV file) in Databricks I'm getting the below error"JSchException: Algorithm negotiation fail"Code:var df = spark.read.options(Map("header"->"true","host"->"20.118.190.30","username"->"user","password"->"pass","fileForm...

Data Engineering
SFTP Spark SCALA Databricks CSV JSCH
  • 3153 Views
  • 1 replies
  • 0 kudos
Latest Reply
User16752239289
Databricks Employee
  • 0 kudos

@MonishKumar Could you provide the entire exception ?From the one line error message, I suspect this is due to the SSL cipher suites required by the SFTP server is not available on cluster. You can run below to get the cipher suites that sftp require...

  • 0 kudos
Manjula_Ganesap
by Contributor
  • 2153 Views
  • 1 replies
  • 0 kudos

Delta Live Table Graph different with no change in Notebook code

I have a DLT code to create 40+ bronze tables. The tables are created on top of the latest parquet files for each of those tables. While executing the pipeline, sometimes I notice that the graph is different than the regular one i see. I do not under...

Manjula_Ganesap_0-1694001824330.png Manjula_Ganesap_1-1694001865082.png
  • 2153 Views
  • 1 replies
  • 0 kudos
Latest Reply
Manjula_Ganesap
Contributor
  • 0 kudos

@Retired_mod  - Thank you for your response. There is no change in the table dependencies. The code to create the individual raw tables look like this: The input to this is always the same 40 tables with only the underlying parquet file changing. I c...

  • 0 kudos
FabriceDeseyn
by Contributor
  • 1637 Views
  • 0 replies
  • 0 kudos

merge breaking persistance of dataframe

Hi allIn the minimal example below you can see that executing a merge statement trigger recomputation of a persisted dataframe. How does this happen?   from delta.tables import DeltaTable table_name = "hive_metastore.default.test_table" # initializ...

FabriceDeseyn_1-1694011507567.png
  • 1637 Views
  • 0 replies
  • 0 kudos
RP2007
by New Contributor
  • 3040 Views
  • 2 replies
  • 1 kudos

I would like to know why I am getting this error when I tried to earn badges for lakehouse fundamen

 I would like to know why I am getting this error when I tried to earn badges for lakehouse fundamentals. I can't access the quiz page. Can you please help on this?Getting below error:-403FORBIDDENYou don't have permission to access this page2023-08-...

  • 3040 Views
  • 2 replies
  • 1 kudos
Latest Reply
APadmanabhan
Databricks Employee
  • 1 kudos

Hello Both, This link would be of help.

  • 1 kudos
1 More Replies
guostong
by New Contributor III
  • 8508 Views
  • 2 replies
  • 0 kudos

Resolved! how to set jobs permission with rest api

create job with cli, but can not set the permission with cli,have to use rest api to set permission:https://docs.databricks.com/api/workspace/permissions/setbelow is my command in windows to set permission:curl -X PUT https://my-workspace-url.azureda...

  • 8508 Views
  • 2 replies
  • 0 kudos
Latest Reply
guostong
New Contributor III
  • 0 kudos

thank you, the new permission list should be a whole list, not the new permission

  • 0 kudos
1 More Replies
parimalpatil28
by New Contributor III
  • 1453 Views
  • 0 replies
  • 0 kudos

Looking for Upload file to dbfs using "/api/2.0/dbfs/put"

Hello,I am trying to upload the file from local linux machine to dbfs using request.post(<URI>,<Headers>, params={"path": "dbfs:/tmp", "contents": local_path}) and getting the error b'{"error_code":"INVALID_PARAMETER_VALUE","message":"You must provid...

  • 1453 Views
  • 0 replies
  • 0 kudos
804082
by New Contributor III
  • 2953 Views
  • 2 replies
  • 1 kudos

Backup/Export Databricks SQL Column Comments

We've had users make comments on tables/columns throughout Databricks SQL using the Data Explorer UI. I'm looking for a way to backup these comments, but when I run DESCRIBE TABLE, the comment column is always null despite being non-null in Data Expl...

  • 2953 Views
  • 2 replies
  • 1 kudos
Latest Reply
shan_chandra
Databricks Employee
  • 1 kudos

@804082  - Markdown does not render when returned by DESCRIBE statements. we can view them on the Data Explorer UI.Reference:   https://docs.databricks.com/en/data/markdown-data-comments.html#document-data-with-markdown-comments

  • 1 kudos
1 More Replies
venkat94
by New Contributor
  • 1445 Views
  • 1 replies
  • 0 kudos

Databricks Job RUns API

/api/2.1/jobs/runs/list Currently returns all jobs which got executed within specified time lines which we provide as input. Is There any way where we can get only specific jobs as per their status(only success)?

Data Engineering
API
azure
Databricks
jobruns
  • 1445 Views
  • 1 replies
  • 0 kudos
Latest Reply
BilalAslamDbrx
Databricks Employee
  • 0 kudos

@venkat94 thanks for the feedback. We are working on updating the Jobs Runs API so you can filter runs by status e.g. only success. Stay tuned in the next couple of months.

  • 0 kudos
dbdude
by New Contributor II
  • 2064 Views
  • 1 replies
  • 0 kudos

Re-running DLT Pipeline Does Not Add Data After Delete

I am using DLT and unity catalog and using managed tables. The first table in this pipeline is a live streaming table. I first did this in the SQL editor:DELETE FROM my_table;This appears to have deleted all the records, which I wanted since now when...

  • 2064 Views
  • 1 replies
  • 0 kudos
Latest Reply
BilalAslamDbrx
Databricks Employee
  • 0 kudos

@Mo is correct! 

  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels