cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

alsetr
by New Contributor III
  • 695 Views
  • 1 replies
  • 0 kudos

Disable Databricks-generated error messages

Since Databricks Runtime 12.2 Databricks started to wrap spark exceptions in their own exceptions.https://learn.microsoft.com/en-us/azure/databricks/error-messages/While for some users it might be handy, for our team it is not convinient, as we canno...

  • 695 Views
  • 1 replies
  • 0 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 0 kudos

Databricks does not use vanilla spark.They added optimizations like the AQE, unity catalog etc.So looking for the error in the spark source code will not always work (in a lot of cases it will)

  • 0 kudos
deano2025
by New Contributor II
  • 1216 Views
  • 2 replies
  • 0 kudos

Resolved! How to create an external location that accesses a public s3 bucket

Hi,I'm trying to create an external location that accesses a public s3 bucket (for open data). However, I'm not having any success. I'm confused to what to specify as the storage credential (IAM role) since its a public bucket that is out of my contr...

  • 1216 Views
  • 2 replies
  • 0 kudos
Latest Reply
deano2025
New Contributor II
  • 0 kudos

Thanks @Isi Now that you've explained external locations, I think it does indeed make sense that they are probably unnecessary in this case. Thanks for clarifying!

  • 0 kudos
1 More Replies
carlos_tasayco
by Contributor
  • 978 Views
  • 1 replies
  • 0 kudos

Materializing tables in custom schemas is not supported.

Hello,I have been seeing this:https://www.databricks.com/blog/publish-multiple-catalogs-and-schemas-single-dlt-pipelineNow dlt pipelines support multiple schemas, however is not working my case:Did I do something wrong?Thanks in advance 

carlos_tasayco_0-1742491316716.png carlos_tasayco_1-1742491343464.png
  • 978 Views
  • 1 replies
  • 0 kudos
Latest Reply
MauricioS
New Contributor III
  • 0 kudos

Hi Carlos,Hope you are doing well, did you get any update on this issue, I'm currently running into the same problem.

  • 0 kudos
hims_2021
by New Contributor
  • 611 Views
  • 1 replies
  • 0 kudos

Unable to export object using /api/2.0/workspace/export API

Hi ,I was using /api/2.0/workspace/export API in power automate workflow to export to excel from data brick to sharepoint. This functionality was working fine till yesterday. Today onwards it is throwing below error while calling the APIAction 'HTTP_...

  • 611 Views
  • 1 replies
  • 0 kudos
Latest Reply
lingareddy_Alva
Honored Contributor III
  • 0 kudos

@hims_2021 This error indicates an encoding issue when trying to export an Excel file from Databricks to SharePoint via Power Automate. The specific error message about being "Unable to translate bytes [9A] at index 11" suggests that Power Automate i...

  • 0 kudos
804082
by New Contributor III
  • 5663 Views
  • 8 replies
  • 2 kudos

Resolved! DLT Direct Publishing Mode

Hello,I'm working on a DLT pipeline and have a block of SQL that runs...USE CATALOG catalog_a; USE SCHEMA schema_a; CREATE OR REFRESH MATERIALIZED VIEW table_a AS SELECT ... FROM catalog_b.schema_b.table_b;Executing this block returns the following.....

  • 5663 Views
  • 8 replies
  • 2 kudos
Latest Reply
Dorsey
New Contributor II
  • 2 kudos

I'm in EastUS and i don't have that option on my previews page. Also it only works with serverless?

  • 2 kudos
7 More Replies
moski
by New Contributor II
  • 14496 Views
  • 9 replies
  • 8 kudos

Databricks short cut to split a cell

Is there a shortcut to split a cell into two in Dtabricks notebook as in Jupiter notebook? in jupyter notebook it is Shift/Ctr/-

  • 14496 Views
  • 9 replies
  • 8 kudos
Latest Reply
Harshjot
Contributor III
  • 8 kudos

 Hi @mundy Jim​ / All, Attached are two snapshots so first snapshot with one cell if pressed Ctrl+Alt+Minus split into two.  

  • 8 kudos
8 More Replies
LearnDB1234
by New Contributor III
  • 1700 Views
  • 3 replies
  • 1 kudos

Resolved! How to Update Identity Column for a Databricks Table

Hi All,I have a databricks table with the below DDL:CREATE TABLE default.Test ( ID BIGINT GENERATED ALWAYS AS IDENTITY (START WITH 1 INCREMENT BY 1), StopFromDateTime TIMESTAMP, StopToDateTime TIMESTAMP, User STRING) USING delta TBLPROPERTIE...

  • 1700 Views
  • 3 replies
  • 1 kudos
Latest Reply
pdiamond
Contributor
  • 1 kudos

If you recreate the table using BIGINT GENERATED BY DEFAULT  instead of BIGINT GENERATED ALWAYS you can manipulate the column values."When using the clause GENERATED BY DEFAULT AS IDENTITY, insert operations can specify values for the identity column...

  • 1 kudos
2 More Replies
ramyav7796
by New Contributor II
  • 1520 Views
  • 1 replies
  • 0 kudos

add custom logs and save in a folder logs

Hi,I am trying to add custom logging functionality for my code. Please refer to the code I am using, I am trying to save my log files by creating a logs folder in my users workspace. My intent is to store dynamic custom log files each time I run my n...

  • 1520 Views
  • 1 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

Here are some suggestions for your consideration.   The issue with your custom logging setup seems to stem from attempting to save the log files in a path under "/Workspace/Users/ramya.v@point32health.org/CD/", which is not directly writable by your ...

  • 0 kudos
r0nald
by New Contributor II
  • 10199 Views
  • 4 replies
  • 1 kudos

UDF not working inside transform() & lambda (SQL)

Below is toy example of what I'm trying to achieve, but don't understand why it fails. Can anyone explain why, and suggest a fix or not overly bloated workaround?%sqlcreate or replace function status_map(status int)returns stringreturn map(10, "STATU...

  • 10199 Views
  • 4 replies
  • 1 kudos
Latest Reply
DattaWalake
Databricks Employee
  • 1 kudos

Scope of lambda implementation within transform function only ,which fails with udf because lambda variable bindings (e in your case) are not available for use within the UDF's scope. we can use below workaround for above example which generate same ...

  • 1 kudos
3 More Replies
User16826994223
by Databricks Employee
  • 2147 Views
  • 2 replies
  • 1 kudos

Does Databricks have a data processing agreement?

Does Databricks have a data processing agreement?

  • 2147 Views
  • 2 replies
  • 1 kudos
Latest Reply
liam_noah
New Contributor II
  • 1 kudos

Yes, Databricks typically provides a Data Processing Agreement (DPA) to comply with data protection regulations like GDPR. It's important for businesses to thoroughly review these agreements to ensure alignment with their data privacy policies. You c...

  • 1 kudos
1 More Replies
hadoan
by New Contributor II
  • 1934 Views
  • 3 replies
  • 1 kudos

How to define DLT table with cyclic reference

 @Dlt.table def table_A(): return ( dlt.read_stream(...) ) @dlt.table def table_join_A_and_C(): df_A = dlt.read_stream(table_A) df_C = dlt.read_stream(table_C) return ( ....df_A.join(df_C) ) @dlt.table def table_C(): return ( ...

  • 1934 Views
  • 3 replies
  • 1 kudos
Latest Reply
dilipdiwakar
New Contributor II
  • 1 kudos

Could you please describe best approach here. Thanks

  • 1 kudos
2 More Replies
Dejian
by New Contributor II
  • 1469 Views
  • 3 replies
  • 0 kudos

DLT Append Flow Parameterization

Hi All,I'm currently using DLT append flow to merge multiple streaming flows into one output.While trying to make the append flow into a dynamic function for scalability, the dlt append flow seem to have some errors.stat_table = f"{catalog}.{bronze_s...

  • 1469 Views
  • 3 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

The error you're encountering occurs because Delta Live Tables (DLT) append flows currently do not support streaming aggregations or other transformations on streaming DataFrames unless a watermark is applied properly to handle late data. Based on yo...

  • 0 kudos
2 More Replies
AntonDBUser
by New Contributor III
  • 4806 Views
  • 1 replies
  • 0 kudos

Oracle Lakehouse Federaton with CA Certificate

Hi!We have been pulling data from Oracle to Databricks by installing Oracle Driver and certificates directly in the cluster. We are now looking into using Lakehouse Federation for Oracle instead, but it seems like the connection doesn't pick up the c...

  • 4806 Views
  • 1 replies
  • 0 kudos
Latest Reply
Isi
Honored Contributor III
  • 0 kudos

Hi @AntonDBUser ,Thanks for sharing your experience, we’re looking into using Lakehouse Federation with Oracle too.I haven’t tested this myself, but one idea that came to mind is whether switching from a serverless cluster to a standard (Pro) cluster...

  • 0 kudos
mridultuteja
by New Contributor II
  • 3052 Views
  • 6 replies
  • 1 kudos

external table not being written to data lake

I was following a tutorial to learn databricks from https://youtu.be/7pee6_Sq3VYGreat video btwI am stuck here at 2:52:24I am trying to create an external table directly to data lake but i am facing some weird issue saying no such location exists.I h...

mridultuteja_0-1746405222446.png mridultuteja_1-1746405246157.png
  • 3052 Views
  • 6 replies
  • 1 kudos
Latest Reply
Isi
Honored Contributor III
  • 1 kudos

Hey @mridultuteja To register an external location, you have to first create a Storage Credential, and then create the External Location.This process allows Databricks to securely access data stored in Azure Data Lake Storage Gen2 (ADLS Gen2), while ...

  • 1 kudos
5 More Replies
SeekingSolution
by New Contributor II
  • 584 Views
  • 1 replies
  • 0 kudos

Unity Catalog Enablement

Hello,After scouring documentation yesterday, I was finally able to get unity catalog enabled and assigned to my workspace. Or so I thought. When I run the CURRENT METASTORE() command I get the below error:However, when I look at my catalog I can see...

SeekingSolution_0-1746620101890.png SeekingSolution_1-1746620144801.png SeekingSolution_2-1746620282198.png
  • 584 Views
  • 1 replies
  • 0 kudos
Latest Reply
Nivethan
New Contributor III
  • 0 kudos

Hi,Please check if the cluster you are using to run the query as well upgraded to Unity Catalog. Also, follow the best practices outlined here for enablement: https://docs.databricks.com/aws/en/data-governance/unity-catalog/enable-workspacesBest Rega...

  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels