cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Anske
by New Contributor III
  • 1614 Views
  • 4 replies
  • 0 kudos

how to stop dataframe with federated table source to be reevaluated when referenced (cache?)

Hi,Would anyone happen to know whether it's possible to cache a dataframe in memory that the result of a query on a federated table?I have a notebook that queries a federated table, does some transformations on the dataframe and then writes this data...

  • 1614 Views
  • 4 replies
  • 0 kudos
Latest Reply
Anske
New Contributor III
  • 0 kudos

@daniel_sahal , this is the code snippet:lsn_incr_batch = spark.sql(f"""select start_lsn,tran_begin_time,tran_end_time,tran_id,tran_begin_lsn,cast('{current_run_ts}' as timestamp) as appendedfrom externaldb.cdc.lsn_time_mappingwhere tran_end_time > '...

  • 0 kudos
3 More Replies
amar1995
by New Contributor II
  • 3150 Views
  • 4 replies
  • 0 kudos

Performance Issue with XML Processing in Spark Databricks

I am reaching out to bring attention to a performance issue we are encountering while processing XML files using Spark-XML, particularly with the configuration spark.read().format("com.databricks.spark.xml").Currently, we are experiencing significant...

  • 3150 Views
  • 4 replies
  • 0 kudos
Latest Reply
shan_chandra
Databricks Employee
  • 0 kudos

@amar1995 - Can you try this streaming approach and see if it works for your use case (using autoloader) - https://kb.databricks.com/streaming/stream-xml-auto-loader

  • 0 kudos
3 More Replies
johnp
by New Contributor III
  • 2237 Views
  • 1 replies
  • 0 kudos

Call databricks notebook from azure flask app

I have an Azure web app running flask web server.  From flask server, I want to run some queries on the data  stored in ADLS Gen2 storage.   I already created Databricks notebooks running these queries.  The flask server will pass some parameters in ...

  • 2237 Views
  • 1 replies
  • 0 kudos
Latest Reply
feiyun0112
Honored Contributor
  • 0 kudos

you can use databricks SDKhttps://docs.databricks.com/en/dev-tools/sdk-python.html#create-a-job 

  • 0 kudos
Kanti1989
by New Contributor II
  • 1810 Views
  • 4 replies
  • 0 kudos

Pyspark execution error

I am getting a error message when executing a simple pyspark code. Can anyone help me with this.  

Kanti1989_0-1713522601530.png
  • 1810 Views
  • 4 replies
  • 0 kudos
Latest Reply
AmanSehgal
Honored Contributor III
  • 0 kudos

Could you please share the entire error message?Are you running the code locally or on databricks?

  • 0 kudos
3 More Replies
miaomia123
by New Contributor
  • 925 Views
  • 1 replies
  • 0 kudos

LLM using DataBrick

Is there any coding example for how to use LLM?

  • 925 Views
  • 1 replies
  • 0 kudos
Latest Reply
jose_gonzalez
Databricks Employee
  • 0 kudos

I would like to share the following links https://www.databricks.com/product/machine-learning/large-language-models https://docs.databricks.com/en/large-language-models/index.html

  • 0 kudos
BrianJ
by New Contributor II
  • 2741 Views
  • 5 replies
  • 4 kudos

{{job.trigger.type}} not working and throws error on Edit Parameter from Job page

Following the instruction on the Job Parameter Dynamic values, I am able to use {{job.id}}{{job.name}}{{job.run_id}}{{job.repair_count}}{{job.start_time.[argument]}}However, when I set trigger_type as trigger_type: {{job.trigger.type}} and hit SAVE, ...

BrianJ_1-1713544000542.png BrianJ_0-1713544144110.png
  • 2741 Views
  • 5 replies
  • 4 kudos
Latest Reply
BrianJ
New Contributor II
  • 4 kudos

Thanks everyone, I decided to use the Sparkcontext instead. dbutils.notebook.entry_point.getDbutils().notebook().getContext().toJson()

  • 4 kudos
4 More Replies
Phani1
by Valued Contributor II
  • 2227 Views
  • 0 replies
  • 0 kudos

Boomi integrating with Databricks

Hi Team,Is there any impact when integrating Databricks with Boomi as opposed to Azure Event Hub? Could you offer some insights on the integration of Boomi with Databricks?https://boomi.com/blog/introducing-boomi-event-streams/Regards,Janga

  • 2227 Views
  • 0 replies
  • 0 kudos
Espenol1
by New Contributor II
  • 8059 Views
  • 4 replies
  • 2 kudos

Resolved! Using managed identities to access SQL server - how?

Hello! My company wants us to only use managed identities for authentication. We have set up Databricks using Terraform, got Unity Catalog and everything, but we're a very small team and I'm struggling to control permissions outside of Unity Catalog....

  • 8059 Views
  • 4 replies
  • 2 kudos
Latest Reply
Espenol1
New Contributor II
  • 2 kudos

Thanks a lot. Then I guess we will try to use dbmanagedidentity for most of our needs, and create service principals +secret scopes when there are more specific needs, such as for limiting access to sensitive data. A bit of a hassle to scale, probabl...

  • 2 kudos
3 More Replies
Snoonan
by Contributor
  • 11506 Views
  • 6 replies
  • 0 kudos

Resolved! Unity catalog issues

Hi all,I have recently enabled Unity catalog in my DBX workspace. I have created a new catalog with an external location on Azure data storage.I can create new schemas(databases) in the new catalog but I can't create a table. I get the below error wh...

  • 11506 Views
  • 6 replies
  • 0 kudos
Latest Reply
daniel_sahal
Esteemed Contributor
  • 0 kudos

@Snoonan First of all, check the networking tab on the storage account to see if it's behind firewall. If it is, make sure that Databricks/Storage networking is properly configured (https://learn.microsoft.com/en-us/azure/databricks/security/network/...

  • 0 kudos
5 More Replies
Carlton
by Contributor
  • 1180 Views
  • 1 replies
  • 0 kudos

Help Refactor T-SQL Code to Databricks SQL

Hello CommunityCan someone help refactor the following T-SQL Code to Databricks SQLCONVERT(DECIMAL(26, 8), ISNULL(xxx.xxxxxxx * ISNULL(RH.xxxxx, 1 / NULLIF(ST.xxxxxx, 0)), ST.xxxxx)) AS AmountWhen I attempt to execute the above code I get the followi...

  • 1180 Views
  • 1 replies
  • 0 kudos
Latest Reply
AmanSehgal
Honored Contributor III
  • 0 kudos

You can use CAST instead.Eg:SELECT cast('2024' as int);

  • 0 kudos
EWhitley
by New Contributor III
  • 2283 Views
  • 0 replies
  • 0 kudos

Custom ENUM input as parameter for SQL UDF?

Hello  - We're migrating from T-SQL to Spark SQL. We're migrating a significant number of queries."datediff(unit, start,end)" is different between these two implementations (in a good way).  For the purpose of migration, we'd like to stay as consiste...

  • 2283 Views
  • 0 replies
  • 0 kudos
andre_rizzatti
by New Contributor II
  • 2511 Views
  • 3 replies
  • 0 kudos

Ingest __databricks_internal catalog - PERMISSION DENIED

Good morning, I have a DLT process with CDC incremental load and I need to ingest the history as CDC transactions are only recent. To do this I need to ingest data in the __databricks_internal catalog. In my case, as I am full admin, I can do it, how...

image.png
  • 2511 Views
  • 3 replies
  • 0 kudos
Latest Reply
andre_rizzatti
New Contributor II
  • 0 kudos

The tables do not have specific configuration, and the user who is receiving the error is in a group that has full permission in the INTERNAL catalog

  • 0 kudos
2 More Replies
amde99
by New Contributor
  • 1277 Views
  • 2 replies
  • 0 kudos

How can I throw an exception when a .json.gz file has multiple roots?

I have a situation where source files in .json.gz sometimes arrive with invalid syntax containing multiple roots separated by empty braces []. How can I detect this and thrown an exception? Currently the code runs and picks up only record set 1, and ...

  • 1277 Views
  • 2 replies
  • 0 kudos
Latest Reply
Lakshay
Databricks Employee
  • 0 kudos

Schema validation should help here.

  • 0 kudos
1 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels