cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

mriccardi
by New Contributor II
  • 1523 Views
  • 4 replies
  • 1 kudos

Spark streaming: Checkpoint not recognising new data

Hello everyone!We are currently facing an issue with a stream that is not updating new data since the 20 of July.We've validated and bronze table has data that silver doesn't have.Also seeing the logs the silver stream is running but writing 0 files....

  • 1523 Views
  • 4 replies
  • 1 kudos
Latest Reply
mriccardi
New Contributor II
  • 1 kudos

Also the trigger is configured to run once, but when we start the job it never ends, it keeps in an endless loop.

  • 1 kudos
3 More Replies
thains
by New Contributor III
  • 1022 Views
  • 1 replies
  • 0 kudos

Resolved! Error: cannot create mws storage configurations: default auth: cannot configure default credentials.

I’ve run into an error that I can't figure out how to debug. We're trying to use terraform through a service account. I don’t know if it’s a permissions issue on Databricks, in our account, or in AWS, but it seems that something is being blocked some...

  • 1022 Views
  • 1 replies
  • 0 kudos
Latest Reply
thains
New Contributor III
  • 0 kudos

Ok. I found the issue here. We had a *second* place where we were setting up the databricks provider, which I had not updated with the proper client credentials.

  • 0 kudos
g96g
by New Contributor III
  • 5770 Views
  • 4 replies
  • 2 kudos

Resolved! If exists in databrickds sql

what is the equivalent of "IF EXISTS" in databricks? I would like to first check something later after that use the insert into statement.

  • 5770 Views
  • 4 replies
  • 2 kudos
Latest Reply
WWoman
New Contributor II
  • 2 kudos

Is there a way to check if a table exists, without trying to drop it? something like :select table_name from system_catalogs where database_name = 'mydb' and schema_name = 'myschema' and object_name = 'mytab';

  • 2 kudos
3 More Replies
DataGirl
by New Contributor
  • 3710 Views
  • 5 replies
  • 2 kudos

Multi value parameter on Power BI Paginated / SSRS connected to databricks using ODBC

Hi All, I'm wondering if anyone has had any luck setting up multi valued parameters on SSRS using ODBC connection to Databricks? I'm getting "Cannot add multi value query parameter" error everytime I change my parameter to multi value. In the query s...

  • 3710 Views
  • 5 replies
  • 2 kudos
Latest Reply
TechMG
New Contributor II
  • 2 kudos

Hello,I am facing similar kind of issue.  I am working on Power BI paginated report and databricks is my source for the report. I was trying to pass the parameter by passing the query in expression builder as mentioned above. However, I have end up w...

  • 2 kudos
4 More Replies
deng_dev
by New Contributor III
  • 580 Views
  • 1 replies
  • 0 kudos

Run Query from another notebook in streaming job

Hi!We want to run query located in another notebook every streaming microbatch.We were trying dbutils.run.notebook but we always get errorContext not valid. If you are calling this outside the main thread, you must set the Notebook context via dbutil...

  • 580 Views
  • 1 replies
  • 0 kudos
Latest Reply
norbitek
New Contributor II
  • 0 kudos

Query Parameters means that you have to pass all parameters as a part of URL after question mark not in the body"/api/1.2/commands/status?clusterId=$cid&contextId=$ec_id&commandId=$command_id"

  • 0 kudos
williamwjs
by New Contributor II
  • 2665 Views
  • 5 replies
  • 1 kudos

Issue with Could not initialize class $linec4a1686037264c21b0e58b369fab8f2d59.$read$

Our job is written in Scala on DataBricks. It used to have the same problem, but was managed to work with putting all case classes in a separate cell. However, lately it started to fail again due to the same error:Could not initialize class $linec4a1...

  • 2665 Views
  • 5 replies
  • 1 kudos
Latest Reply
williamwjs
New Contributor II
  • 1 kudos

Hi @Kaniz , may I ask if there's any updates to this issue? Thank you!

  • 1 kudos
4 More Replies
Databricks143
by New Contributor III
  • 448 Views
  • 1 replies
  • 0 kudos

Not able to generate Excel file in blob through databricks

Hi team,.I am using cluster 9.1 8n databricks not able to generate Excel file in blob  and  below are conf Cluster:9.1.8Spark version -3.1.1Scala version 3.1.1Library:Com.crealyticsSpark.excel_2.12Version-3.1.1_0.18.2Dependency:Org.apachr.poi-poi-5.2...

  • 448 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @Databricks143, It appears that you’re encountering an issue while trying to generate an Excel file in Azure Databricks.    Let’s troubleshoot this step by step:   Library Dependencies: Ensure that the necessary libraries are correctly installed i...

  • 0 kudos
dwiltse12
by New Contributor II
  • 1721 Views
  • 2 replies
  • 1 kudos

Tableau Delta Sharing

Does anyone have any recent examples of using Tableau and Delta Sharing? The video below mentions using web connector but this connector has been depreciated in Tableau 2023.1. https://www.youtube.com/watch?v=Yg-5LXH9K1I&t=913shttps://help.tableau.co...

  • 1721 Views
  • 2 replies
  • 1 kudos
Latest Reply
JohnMT
New Contributor II
  • 1 kudos

Hi,I am still trying to figure out how to use delta sharing with tableau. I've looking for information for a month without any success. As mentionned before, web data connector is deprecatedAny help would be appreciated.thanks, Johnattan  

  • 1 kudos
1 More Replies
AG2
by New Contributor III
  • 1051 Views
  • 1 replies
  • 0 kudos

Orchestration

Is it possible to use redwood orchestration over Databricks ?

  • 1051 Views
  • 1 replies
  • 0 kudos
Latest Reply
Miguel_Suarez
New Contributor III
  • 0 kudos

Hi @AG2 , We don't currently support Redwood Orchestration over Databricks. Best, Miguel

  • 0 kudos
Hubert-Dudek
by Esteemed Contributor III
  • 3000 Views
  • 2 replies
  • 0 kudos

Resolved! dlt append_flow = multiple streams into a single Delta table

With the append_flow method in Delta Live Tables, you can effortlessly combine data from multiple streams into a single Delta table.

dlt_target.png
  • 3000 Views
  • 2 replies
  • 0 kudos
Latest Reply
jose_gonzalez
Moderator
  • 0 kudos

Thank you for sharing this information @Hubert-Dudek 

  • 0 kudos
1 More Replies
Hubert-Dudek
by Esteemed Contributor III
  • 1299 Views
  • 1 replies
  • 3 kudos

row-level concurrency

Databricks Runtime 14.2 now has row-level concurrency generally available and enabled by default for Delta tables with deletion vectors. This feature dramatically reduces conflicts between concurrent write operations.

142.png
  • 1299 Views
  • 1 replies
  • 3 kudos
Latest Reply
jose_gonzalez
Moderator
  • 3 kudos

Thank you for sharing this @Hubert-Dudek !!!

  • 3 kudos
grazie
by Contributor
  • 526 Views
  • 0 replies
  • 0 kudos

Run a job as different service principals

We currently have several workflows that are basically copies with the only difference being that they run with different service principals and so have different permissions and configuration based on who is running. The way this is managed today is...

  • 526 Views
  • 0 replies
  • 0 kudos
reshmir18
by New Contributor II
  • 662 Views
  • 2 replies
  • 0 kudos

Unable to setcheckpointdir in unitycatalog enabled workspace

I have a Unity catalog enabled workspace where I am trying to setCheckpointDir during runtime. The method looks to authenticate using fs.azure.account.key instead of storage credentials. I am using databricks access connector which has "Storage Blob ...

Data Engineering
autoloader
Databricks
storagecredentials
streaming
unitycatalog
  • 662 Views
  • 2 replies
  • 0 kudos
Latest Reply
reshmir18
New Contributor II
  • 0 kudos

@Kaniz I have provided all the necessary permissions and were able to browse through the folders of the container added as an external location.I don't understand why the method setcheckpointdir looks for account key when the access is already provid...

  • 0 kudos
1 More Replies
Anup
by New Contributor III
  • 1125 Views
  • 2 replies
  • 1 kudos

Resolved! Copy Into : Pattern for sub-folders

While trying to ingest data from the S3 bucket, we are running into a situation where the data in s3 buckets is in sub-folders of multiple depths.Is there a good way of specifying patterns for the above case?We tried using the following for a depth o...

  • 1125 Views
  • 2 replies
  • 1 kudos
Latest Reply
Kaniz
Community Manager
  • 1 kudos

Hi @Anup, When dealing with data in S3 buckets that are organized into sub-folders of varying depths, specifying patterns can be challenging.   However, there are some approaches you can consider:   Wildcard Patterns: You’ve already used a wildcard p...

  • 1 kudos
1 More Replies
MinMin
by New Contributor II
  • 1363 Views
  • 4 replies
  • 0 kudos

Extra underscore behind ".xlsm" and ".xlsx" after exporting excel files from Databricks

Hi all, I tried to export several excel files from Databricks. But there will always be one extra underscore behind ".xlsm" and ".xlsx", if I export them and try to open the files on local system. I have to manually remove the underscore from the fil...

  • 1363 Views
  • 4 replies
  • 0 kudos
Latest Reply
DH_Fable
New Contributor II
  • 0 kudos

Hi, did you find a solution this? I have the same/similar problem where when I save a dataframe from a Databricks notebook using to_excel() it saves the file with extension ".xlsx_" rather then "xlsx", meaning to open I have to manually download and ...

  • 0 kudos
3 More Replies
Labels
Top Kudoed Authors