cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

pavlos_skev
by New Contributor III
  • 1340 Views
  • 1 replies
  • 0 kudos

Potential Unity Catalog Bug: Invalid configuration value detected for fs.azure.account.keyInvalid

Hello,We are migrating to Unity Catalog (UC), and for very few of our tables, we get the below error when trying to write or even display them. We are using UC enabled clusters, usually with runtime versions 12.2 LTS. The below error, when it happens...

  • 1340 Views
  • 1 replies
  • 0 kudos
Latest Reply
User16539034020
Databricks Employee
  • 0 kudos

Hello,  Thanks for contacting Databricks Support.  The error message indicates a problem with the configuration key fs.azure.account.key. This configuration key is used to provide the access key for the Azure Data Lake Storage account. Not sure if th...

  • 0 kudos
GriffLehman
by New Contributor II
  • 1338 Views
  • 1 replies
  • 0 kudos

PROBLEM- Missing data in "Last Run" column in Databricks Workflows UI

Hello,I am having a pretty major problem with the Databricks Workflows UI- when I look at the list of jobs, the "Last Run" column does not have any data in it. This is kind of a big problem because now I don't have a good way of getting visibility in...

  • 1338 Views
  • 1 replies
  • 0 kudos
Latest Reply
GriffLehman
New Contributor II
  • 0 kudos

  

  • 0 kudos
MariuszJ
by New Contributor II
  • 1809 Views
  • 1 replies
  • 2 kudos

cdc type 2 - not closing records

Hi,i would not mind small advice, i do have dlt cdc typ2, the definition dlt.create_streaming_table('`my_table_dlt_cdc`') dlt.apply_changes( target = 'my_table_dlt_cdc', source = 'source', keys = ['id'], sequence_by = col('snapshot_date'), ...

  • 1809 Views
  • 1 replies
  • 2 kudos
Latest Reply
MariuszJ
New Contributor II
  • 2 kudos

For anyone who sees this post in the future. I was missing one argumentapply_as_deletes  

  • 2 kudos
ehpogue
by New Contributor III
  • 18214 Views
  • 9 replies
  • 3 kudos

how do i re-enable tab complete / autocomplete?

yesterday all of my notebooks seemingly changed to have python formatting (which seems to be in this week's release), but the unintended consequence is that shift + tab (which used to show docstrings in python) now just un-indents code, and tab inser...

  • 18214 Views
  • 9 replies
  • 3 kudos
Latest Reply
Data_33
New Contributor II
  • 3 kudos

i also facing the same in databricks now.

  • 3 kudos
8 More Replies
Rpabs
by New Contributor
  • 1042 Views
  • 0 replies
  • 0 kudos

Unable to load historical data via DLT workflow

HiI'm trying to load historical data with the DLT workflow and facing a Out of Memory or Executor heartbeat lost error. The historical data loads fine with normal processing, but fails within DLT workflow. Tried repartitioning and scaling cluster, bu...

  • 1042 Views
  • 0 replies
  • 0 kudos
DeltaTrain
by New Contributor II
  • 2063 Views
  • 1 replies
  • 0 kudos

Access Control in hive_metastore Based on Cluster Type

Hello Databricks Community, I asked the same question on the Get Started Discussion page but feels like here is the right place for this question. I'm reaching out with a query regarding access control in the hive_metastore. I've encountered behavior...

DeltaTrain_0-1691618617261.png DeltaTrain_1-1691618617263.png
  • 2063 Views
  • 1 replies
  • 0 kudos
Latest Reply
User16752239289
Databricks Employee
  • 0 kudos

That is expected. The single user mode is the legacy standard + UC ACL enabled. https://docs.databricks.com/en/archive/compute/cluster-ui-preview.html#how-does-backward-compatibility-work-with-these-changes For your case, you need the hive table acl ...

  • 0 kudos
DataEng1
by New Contributor
  • 3237 Views
  • 0 replies
  • 0 kudos

data type that cannot participate in a columnstore index Error

Hi AllI am trying to insert DF into Synapse table. I need to insert string type columns in DF into Nvarchar fields in Synapse table. I am getting the error ' data type that cannot participate in a columnstore index Error'   Can someone guide on the i...

  • 3237 Views
  • 0 replies
  • 0 kudos
YSDPrasad
by New Contributor III
  • 7136 Views
  • 3 replies
  • 3 kudos

Resolved! NoClassDefFoundError: scala/Product$class

import com.microsoft.azure.sqldb.spark.config.Configimport com.microsoft.azure.sqldb.spark.connect._import com.microsoft.azure.sqldb.spark.query._val query = "Truncate table tablename"val config = Config(Map( "url"     -> dbutils.secrets.get(scope = ...

  • 7136 Views
  • 3 replies
  • 3 kudos
Latest Reply
Anonymous
Not applicable
  • 3 kudos

@Someswara Durga Prasad Yaralgadda​ :The NoClassDefFoundError error occurs when a class that was available during the compile time is not available at the runtime. This could be due to a few reasons, including a missing dependency or an incompatible ...

  • 3 kudos
2 More Replies
adriennn
by Valued Contributor
  • 2455 Views
  • 1 replies
  • 0 kudos

Delay when updating Bronze and Silver tables in the same notebook (DBR 13.1)

I created a notebook that uses Autoloader to load data from storage and append it to a bronze table in the first cell, this works fine and Autoloader picks up new data when it arrives (the notebook is ran using a Job).In the same notebook, a few cell...

  • 2455 Views
  • 1 replies
  • 0 kudos
Latest Reply
adriennn
Valued Contributor
  • 0 kudos

Thanks @Retired_mod, in a case where it's not possible or not practical to implement a pipeline  with DLTs, what would be that "retry mechanism" based on ? I.e., is there an API other that the table history that can be leveraged to retry until "it wo...

  • 0 kudos
Nino
by Contributor
  • 2161 Views
  • 1 replies
  • 0 kudos

cluster nodes unavailable scenarios

Concerning job cluster configuration, I'm trying to figure out what happens if AWS node type availability is smaller than the minimum number of workers specified in the configuration json (either availabilty<num_workers or, for autoscaling, availabil...

  • 2161 Views
  • 1 replies
  • 0 kudos
Latest Reply
Nino
Contributor
  • 0 kudos

thanks, @Retired_mod , useful info!My specific scenario is running a notebook task with Job Clusters, and I've noticed that I get the best overall notebook run time by going without Autoscaling, setting the cluster configuration with a fixed `num_wor...

  • 0 kudos
Fiona
by New Contributor II
  • 4979 Views
  • 3 replies
  • 3 kudos

Resolved! Reading a protobuf file in a Databricks notebook

I have proto files (offline data storage) that I'd like to read from a Databricks notebook. I found this documentation (https://docs.databricks.com/structured-streaming/protocol-buffers.html), but it only covers how to read the protobuf data once the...

  • 4979 Views
  • 3 replies
  • 3 kudos
Latest Reply
StephanK
New Contributor II
  • 3 kudos

If you have proto files in offline data storage, you should be able to read them with:input_df = spark.read.format("binaryFile").load(data_path) 

  • 3 kudos
2 More Replies
DE-cat
by New Contributor III
  • 1931 Views
  • 1 replies
  • 0 kudos

err:setfacl: Option -m: Invalid argument LibraryDownloadManager error

When starting a DB job using 13.3 LTS (includes Apache Spark 3.4.1, Scala 2.12) cluster, I am seeing a lots of these errors in log4j output. Any ideas? Thx23/09/11 13:24:14 ERROR CommandLineHelper$: Command [REDACTED] failed with exit code 2 out: err...

Data Engineering
LibraryDownloadManager
  • 1931 Views
  • 1 replies
  • 0 kudos
DBUser2
by New Contributor III
  • 1966 Views
  • 1 replies
  • 0 kudos

Databricks sql using odbc issue

Hi,I'm connecting to a Databricks instance on Azure from a Windows Application using Simba ODBC driver, and when running SQL statements on delta tables, like INSERT, UPDATE, DELETE commands using Execute, the result doesn't indicate the no. of rows a...

  • 1966 Views
  • 1 replies
  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels