cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

minhhung0507
by Valued Contributor
  • 4379 Views
  • 8 replies
  • 6 kudos

CANNOT_UPDATE_TABLE_SCHEMA

I'm encountering a puzzling schema merge issue with my Delta Live Table. My setup involves several master tables on Databricks, and due to a schema change in the source database, one of my Delta Live Tables has a column (e.g., "reference_score") that...

  • 4379 Views
  • 8 replies
  • 6 kudos
Latest Reply
Brahmareddy
Esteemed Contributor
  • 6 kudos

Dear Hung,Thank you so much for the kind words—I’m really glad the suggestions were helpful! You're absolutely doing the right thing by trying those options first before going for a full table drop. Testing with a new table and checking schema hints ...

  • 6 kudos
7 More Replies
databicky
by Contributor II
  • 1367 Views
  • 2 replies
  • 0 kudos

How to acknowledge the incidents from the databricks

I want to connect to the service now from the databricks and i want to acknowledge the incidents like assigned to the person and i need to change the status of the incident and i want to update the worknotes as well, how can i achieve this by the hel...

  • 1367 Views
  • 2 replies
  • 0 kudos
Latest Reply
Isi
Honored Contributor III
  • 0 kudos

Hi @databicky Just a quick clarification: this is the Databricks community forum, not official Databricks Support. Think of it like a place where users can share ideas, help each other, or sometimes just browse, responses are not guaranteed here.If y...

  • 0 kudos
1 More Replies
whalebone711
by New Contributor
  • 1015 Views
  • 1 replies
  • 0 kudos

Issue with Delta Lake Table Optimization

Hi Databricks Community, I’m currently working on optimizing a Delta Lake table, but I’m encountering some performance issues during the vacuum process. I’ve attached a screenshot of the error message I’m receiving, along with the relevant code snipp...

  • 1015 Views
  • 1 replies
  • 0 kudos
Latest Reply
Isi
Honored Contributor III
  • 0 kudos

Hey @whalebone711 Could you re-attach the screenshot? Looks like it doesn't appear in the post.Best,Isi

  • 0 kudos
Jorge3
by New Contributor III
  • 6762 Views
  • 2 replies
  • 2 kudos

Databricks Asset Bundle artifacts with module out of the bundle root (sync path)

Hello everyone!I’m currently working on a project with shared functionalities across different Databricks bundles. I have separate folders for each bundle, along with a common libs/ folder that holds some Python modules intended to be shared across b...

  • 6762 Views
  • 2 replies
  • 2 kudos
Latest Reply
VZLA
Databricks Employee
  • 2 kudos

Hi @Jorge3, where you able to get this issue resolved? I believe your artifact build path points outside the synced directory structure, and after syncing ../libs, libs should be available within the bundle root, so the artifact path should be update...

  • 2 kudos
1 More Replies
michael_mehrten
by Databricks Partner
  • 50916 Views
  • 26 replies
  • 14 kudos

Resolved! How to use Databricks Repos with a service principal for CI/CD in Azure DevOps?

Databricks Repos best-practices recommend using the Repos REST API to update a repo via your git provider. The REST API requires authentication, which can be done one of two ways:A user / personal access tokenA service principal access tokenUsing a u...

  • 50916 Views
  • 26 replies
  • 14 kudos
Latest Reply
pbz
New Contributor II
  • 14 kudos

For anyone coming here in the future, this should explain it: https://docs.databricks.com/aws/en/repos/ci-cd-techniques-with-repos#authorize-a-service-principal-to-access-git-foldersBasically:1. Go to your service in settings -> identity and access -...

  • 14 kudos
25 More Replies
pedrojunqueira
by New Contributor II
  • 18515 Views
  • 5 replies
  • 3 kudos

Resolved! Generating Personal Access Token to service principle databricks cli

Hi I am having issues generating personal access token to my service principle.I followed the steps from here my `~/.databrickscfg` has the following```[my-profile-name]host = <account-console-url>account_id = <account-id>azure_tenant_id = <azure-ser...

  • 18515 Views
  • 5 replies
  • 3 kudos
Latest Reply
PabloCSD
Valued Contributor II
  • 3 kudos

I want something similar, to use a service principal token instead of a PAT, have you ever done this?https://community.databricks.com/t5/administration-architecture/use-a-service-principal-token-instead-of-personal-access-token/m-p/91629

  • 3 kudos
4 More Replies
ushnish_18
by New Contributor
  • 852 Views
  • 1 replies
  • 0 kudos

Facing error which submitting lab assessment of Delivery Specialization: UC Upgrade

Hi,I attended the lab assessment of Delivery Specialization: UC Upgrade and while submitting my answers the grade is not getting updated successfully due to which the status is being showed as failed, although all the checkpoints got validated succes...

  • 852 Views
  • 1 replies
  • 0 kudos
Latest Reply
Arpita_S
Databricks Employee
  • 0 kudos

Hi @ushnish_18, Can you share the user ID or email you used to access the lab so the team can take a look? Alternatively, you can send all the details, including screenshots, here for the team to investigate in detail and guide you appropriately. Tha...

  • 0 kudos
HarryRichard08
by New Contributor II
  • 1186 Views
  • 1 replies
  • 0 kudos

Acess to s3 in aws

My Problem:My Databricks workspace (Serverless Compute) is in AWS Account A, but my  S3 bucket is in AWS Account B.Works in Shared Compute because i  am  manually setting access_key and secret_key.Does NOT work in Serverless Compute 

  • 1186 Views
  • 1 replies
  • 0 kudos
Latest Reply
SP_6721
Honored Contributor II
  • 0 kudos

Hi @HarryRichard08 ,I came across a similar thread of yours. Were you able to find a resolution for this?

  • 0 kudos
minhhung0507
by Valued Contributor
  • 4402 Views
  • 10 replies
  • 5 kudos

Performance issue with Spark SQL when working with data from Unity Catalog

  We're encountering a performance issue with Spark SQL when working with data from Unity Catalog. Specifically, when I use Spark to read data from a Unity Catalog partition created by DLT and then create a view, the executor retrieval is very slow. ...

  • 4402 Views
  • 10 replies
  • 5 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 5 kudos

you can read the physical parquet files with spark.read.parquet().The trick is to know which files are the current ones.

  • 5 kudos
9 More Replies
NikosLoutas
by Databricks Partner
  • 1952 Views
  • 3 replies
  • 2 kudos

Spark DataFrame Checkpoint

Good morning,I am having a difficulty when trying to checkpoint a PySpark DataFrame.The DataFrame is not involved in a DLT pipeline so I am using the df.checkpoint(eager=True) command, to truncate the logical plan of df and materialize it as files wi...

  • 1952 Views
  • 3 replies
  • 2 kudos
Latest Reply
saurabh18cs
Honored Contributor III
  • 2 kudos

your volume approach is also good idea

  • 2 kudos
2 More Replies
libpekin
by New Contributor II
  • 3767 Views
  • 1 replies
  • 0 kudos

Generating Multiple Excels for SQL Query

Hello,I am getting "OSError: Errno 95: Operation not supported for the code below. I have 'openpyxl 3.1.5' installed on the cluster and have imported all required modules. I am sure this is something small, but I can't put my finger on it why this is...

  • 3767 Views
  • 1 replies
  • 0 kudos
Latest Reply
Renu_
Valued Contributor II
  • 0 kudos

Hi @libpekin, is filepath pointing to a DBFS path? Because writing directly to DBFS paths using to_excel() is not supported due to DBFS limitations with certain file operations, especially random writes.As a workaround, first save the Excel file to a...

  • 0 kudos
nielsehlers
by New Contributor
  • 973 Views
  • 1 replies
  • 1 kudos

Visualizations ignore Timezones

Databricks inline visualizations (bar charts / line charts) etc. ignore Timezones and always display the UTC Time on the X-Axis.

  • 973 Views
  • 1 replies
  • 1 kudos
Latest Reply
Advika
Community Manager
  • 1 kudos

Hello @nielsehlers! As a workaround, you can use the from_utc_timestamp() function to convert UTC timestamps to your desired time zone before visualising:SELECT from_utc_timestamp(column_name, 'Asia/Kolkata’) AS alias_name FROM table_name;

  • 1 kudos
user1234567899
by New Contributor II
  • 2059 Views
  • 2 replies
  • 0 kudos

Lineage not visible for table created in DLT

Hello,I've been struggling for two days with missing lineage information for the silver layer table, and I'm unsure what I'm doing incorrectly.I have a DLT pipeline with DPM public preview enabled. Data is ingested from an S3 bucket into the bronze t...

  • 2059 Views
  • 2 replies
  • 0 kudos
Latest Reply
obitech01
New Contributor II
  • 0 kudos

I'm having the same exact issue as the poster above.I'm using a newly created DLT pipeline as of April 1st 2025 (Unity catalog enabled, Serverless).  I get lineage for all tables and views involved in my pipeline except for the streaming table.The st...

  • 0 kudos
1 More Replies
aranjan99
by Contributor
  • 1560 Views
  • 4 replies
  • 0 kudos

Disabled billing system schema accidentally and now I cannot enable it again

We accidentally disabled billing system table schema and now getting this error on re-enabling it:Error: billing system schema can only be enabled by Databricks.How can we resolve the same. We have not purchased any of the support contracts and hence...

  • 1560 Views
  • 4 replies
  • 0 kudos
Latest Reply
aranjan99
Contributor
  • 0 kudos

00640919

  • 0 kudos
3 More Replies
prashantjjain33
by New Contributor II
  • 2161 Views
  • 3 replies
  • 0 kudos

databricks_error_message:REQUEST_LIMIT_EXCEEDED:

A databricks job failed unexpectedly with below error. There ware only 5 jobs running at that time and no major operations. what could be the root cause and how can we avoid this in futureCluster '0331-xxxxxx-zs8i8pcn' was terminated. Reason: INIT_SC...

  • 2161 Views
  • 3 replies
  • 0 kudos
Latest Reply
saurabh18cs
Honored Contributor III
  • 0 kudos

are you generating any databricks tokens in this process? if yes there is a limit of 600.

  • 0 kudos
2 More Replies
Labels