cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

deng_dev
by New Contributor III
  • 361 Views
  • 1 replies
  • 0 kudos

Getting "Job aborted" exception while saving data to the database

Hi!We have job, that runs every hour. It extracts data from the API and saves to the databricks table.Sometimes job fails with error "org.apache.spark.SparkException". Here is the full error:An error occurred while calling o7353.saveAsTable. : org.ap...

  • 361 Views
  • 1 replies
  • 0 kudos
Latest Reply
jose_gonzalez
Moderator
  • 0 kudos

Do you have any NULL values in your data? Please verify that you data is valid

  • 0 kudos
532664
by New Contributor III
  • 1841 Views
  • 11 replies
  • 3 kudos

Resolved! Replay(backfill) DLT CDC using kafka

Hello,We are receiving DB CDC binlogs through Kafka and synchronizing tables in OLAP system using the apply_changes function in Delta Live Table (DLT). A month ago, a column was added to our table, but due to a type mismatch, it's being stored incorr...

  • 1841 Views
  • 11 replies
  • 3 kudos
Latest Reply
jcozar
Contributor
  • 3 kudos

Thank you @532664 for your detailed response! That's seems to me a very good solution, and it also helps me with my doubts

  • 3 kudos
10 More Replies
Prashant777
by New Contributor II
  • 3172 Views
  • 4 replies
  • 0 kudos

Error in SQL statement: UnsupportedOperationException: Cannot perform Merge as multiple source rows matched and attempted to modify the same

My code:- CREATE OR REPLACE TEMPORARY VIEW preprocessed_source ASSELECT  Key_ID,  Distributor_ID,  Customer_ID,  Customer_Name,  ChannelFROM integr_masterdata.Customer_Master;-- Step 2: Perform the merge operation using the preprocessed source tableM...

  • 3172 Views
  • 4 replies
  • 0 kudos
Latest Reply
Tread
New Contributor II
  • 0 kudos

Hey as previously stated you could drop the duplicates of the columns that contain the said duplicates(code you can find online pretty easily), I have had this problem myself and it came when creating a temporary view from a dataframe, the dataframe ...

  • 0 kudos
3 More Replies
sunkam
by New Contributor
  • 2018 Views
  • 4 replies
  • 0 kudos

Unable to read from azure blob using SAS token\\

I have tried many times all the answers from the internet and stackover flowI have already created the config section before this steps, it passed but this below step is not executing.

image
  • 2018 Views
  • 4 replies
  • 0 kudos
Latest Reply
aockenden
New Contributor III
  • 0 kudos

We were getting this problem when using directory-scoped SAS tokens. While I know there are a number of potential issues that can cause this problem, one potential explanation is that it turns out there is an undocumented spark setting needed on the ...

  • 0 kudos
3 More Replies
Hemendra_Singh
by New Contributor II
  • 752 Views
  • 1 replies
  • 1 kudos

Resolved! Unity catalog - external table and managed table

do the external tables which we create or manage through unity catalog supports acid properties and time traveling, and if we go for the performance issue which is more faster to query and why ?

  • 752 Views
  • 1 replies
  • 1 kudos
Latest Reply
Kaniz
Community Manager
  • 1 kudos

Hi @Hilium, External tables in the Unity Catalog reference an external storage path. They are used when you require direct access to the data outside of Azure Databricks clusters or Databricks SQL warehouses. However, the ACID properties and time-tra...

  • 1 kudos
Bilal1
by New Contributor III
  • 16097 Views
  • 6 replies
  • 2 kudos

Resolved! Simply writing a dataframe to a CSV file (non-partitioned)

When writing a dataframe in Pyspark to a CSV file, a folder is created and a partitioned CSV file is created. I have then rename this file in order to distribute it my end user.Is there any way I can simply write my data to a CSV file, with the name ...

  • 16097 Views
  • 6 replies
  • 2 kudos
Latest Reply
Bilal1
New Contributor III
  • 2 kudos

Thanks for confirming that that's the only way

  • 2 kudos
5 More Replies
jorgemarmol
by New Contributor II
  • 1123 Views
  • 4 replies
  • 0 kudos

Delta Live Tables: Too much time to do the "setting up"

Hello community!Recently I have been working in delta live table for a big project. My team and me have been studying so much and finally we have built a good pipeline with CDC that load 608 entities (and, therefore, 608 delta live tables and 608 mat...

jorgemarmol_0-1688633577282.png
  • 1123 Views
  • 4 replies
  • 0 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 0 kudos

Interesting...DLT probably spends x seconds/table for the setup.If you have time, you could do some tests to see if the table setup scales linearly (1 table, 5 sec for setup, 10 tables 50 sec etc).If you do, please share the outcome.

  • 0 kudos
3 More Replies
gardener
by New Contributor III
  • 527 Views
  • 2 replies
  • 0 kudos

Resolved! Url column issue in UC information_schema.schemata view definition

Hi, I recently observed that, after creating a new catalog (without a managed location) in Unity Catalog, a column named 'url' is included in the definition of the information_schema.schemata view.However, there is no url column in the underlying tab...

  • 527 Views
  • 2 replies
  • 0 kudos
Latest Reply
Kaniz
Community Manager
  • 0 kudos

Hi @gardener, Based on the Databricks documentation, the information_schema.schemata view should contain the following columns: catalog_name: Catalog containing the schema.schema_name: Name of the schema.schema_owner: User or group (principal) that c...

  • 0 kudos
1 More Replies
N_M
by New Contributor III
  • 1604 Views
  • 5 replies
  • 1 kudos

Resolved! ignoreCorruptFiles behavior with CSV and COPY INTO

HiI'm using the COPY INTO command to insert new data (in form of CSVs) into an already existing table.The SQL query takes care of the conversion of the fields to the target table schema (well, there isn't other way to do that), and schema update is n...

Data Engineering
COPY INTO
ignoreCorruptFiles
  • 1604 Views
  • 5 replies
  • 1 kudos
Latest Reply
N_M
New Contributor III
  • 1 kudos

I actually found an option that could solve the newline issue I mentioned in my previous post:setting spark.sql.csv.parser.columnPruning.enabled to false withspark.conf.set("spark.sql.csv.parser.columnPruning.enabled", False)will consider malformed r...

  • 1 kudos
4 More Replies
datakilljoy
by New Contributor II
  • 801 Views
  • 1 replies
  • 0 kudos

Best practice for Azure Key vault secrets in spark config

HelloI created a compute in which I refer the secret inside the spark config like this: spark.hadoop.fs.azure.account.key.xxxxxxxxxx.dfs.core.windows.net {{secrets/kv-xxxxxxx-xxxx/secret-name}}  This, however, gives me the following warning.   I've l...

datakilljoy_0-1704724007789.png
Data Engineering
cluster
compute
spark
  • 801 Views
  • 1 replies
  • 0 kudos
Latest Reply
datakilljoy
New Contributor II
  • 0 kudos

Extra info: I have used the format following the instructions on this page for spark configurationhttps://learn.microsoft.com/en-us/azure/databricks/connect/storage/azure-storage#:~:text=Use%20the%20following%20format%20to%20set%20the%20cluster%20Spa...

  • 0 kudos
merca
by Valued Contributor II
  • 916 Views
  • 1 replies
  • 0 kudos

Resolved! Problems with DLT, Unity catalog and external connection

I have following code:org = spark.read.table("catalog.dbo.organisation") @dlt.create_table() def organization(): return orgThe catalog is an external azure sql database (using external connector)When i validate this in Delta live table workflow I...

  • 916 Views
  • 1 replies
  • 0 kudos
Latest Reply
Sumit671
New Contributor III
  • 0 kudos

use preview channel while create pipeline instead of current

  • 0 kudos
leelee3000
by New Contributor III
  • 629 Views
  • 3 replies
  • 2 kudos

Development Feedback Loop

I've noticed that the current development cycle for DLT jobs is quite time-consuming. The process of coding, saving, running in a workflow, and debugging seems arduous, and the feedback loop is slow. Is there a way to run DLT jobs without relying on ...

  • 629 Views
  • 3 replies
  • 2 kudos
Latest Reply
Kaniz
Community Manager
  • 2 kudos

Hi @leelee3000, Developing and iterating on Delta Live Tables (DLT) jobs can be time-consuming when relying solely on traditional workflows.   Databricks Jobs: Databricks jobs allow you to orchestrate multiple tasks within a Databricks job, creating ...

  • 2 kudos
2 More Replies
prapot
by New Contributor II
  • 5335 Views
  • 2 replies
  • 2 kudos

Resolved! How to write a Spark DataFrame to CSV file with our .CRC in Azure Databricks?

val spark:SparkSession = SparkSession.builder() .master("local[3]") .appName("SparkByExamples.com") .getOrCreate()//Spark Read CSV Fileval df = spark.read.option("header",true).csv("address.csv")//Write DataFrame to address directorydf.write...

  • 5335 Views
  • 2 replies
  • 2 kudos
Latest Reply
Nw2this
New Contributor II
  • 2 kudos

Will your csv have the name prefix 'part-' or can you name it whatever you like?

  • 2 kudos
1 More Replies
hukel
by Contributor
  • 1169 Views
  • 6 replies
  • 0 kudos

Unsupported datatype 'TimestampNTZType' with liquid clustering

I'm experimenting with liquid clustering and have some questions about compatible types  (somewhat similar to Liquid clustering with boolean columns ).Table created as CREATE TABLE IF NOT EXISTS <TABLE> ( _time DOUBLE , timestamp TIMESTAMP_NT...

  • 1169 Views
  • 6 replies
  • 0 kudos
Latest Reply
Wojciech_BUK
Contributor III
  • 0 kudos

Hi,just educated guess:There is limitation in liquid clustering docs: You can only specify columns with statistics collected for clustering keysPerhaps it is related to data types for which you can collect statistics?But i could not find related docs...

  • 0 kudos
5 More Replies
NathanE
by New Contributor II
  • 1728 Views
  • 2 replies
  • 1 kudos

Java 21 support with Databricks JDBC driver

Hello,I was wondering if there was any timeline for Java 21 support with the Databricks JDBC driver (current version is 2.34).One of the required change is to update the dependency to arrow to version 13.0 (current version is 9.0.0).The current worka...

Data Engineering
driver
java21
JDBC
  • 1728 Views
  • 2 replies
  • 1 kudos
Latest Reply
Fabich
New Contributor II
  • 1 kudos

Hello @Kaniz Any update on this topic of Java 21 ? Any timeline ?Our clients really want to upgrade to Java 21 and we don't want to disable Arrow for performance reasons

  • 1 kudos
1 More Replies
Labels
Top Kudoed Authors