cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Dhruv-22
by Contributor
  • 4 Views
  • 0 replies
  • 0 kudos

Can't mergeSchema handle int and bigint?

I have a table which has a column of data type 'bigint'. While overwriting it with new data, given that I do full loads, I used 'mergeSchema' to handle schema changes. The new data's datatype was int. I thought mergeSchema can easily handle that, but...

  • 4 Views
  • 0 replies
  • 0 kudos
saicharandeepb
by New Contributor III
  • 18 Views
  • 1 replies
  • 1 kudos

How to Retrieve DBU Count per Compute Type for Accurate Cost Calculation?

Hello Everyone,We are currently working on a cost analysis initiative to gain deeper insights into our Databricks usage. As part of this effort, we are trying to calculate the hourly cost of each Databricks compute instance by utilizing the Azure Ret...

  • 18 Views
  • 1 replies
  • 1 kudos
Latest Reply
BS_THE_ANALYST
Esteemed Contributor II
  • 1 kudos

@saicharandeepb  have you looked at the system billing tables in Databricks yet? https://learn.microsoft.com/en-us/azure/databricks/admin/system-tables/billing There seems to be a field that can display the unit usage in DBU. Same in this table aswel...

  • 1 kudos
Dhruv-22
by Contributor
  • 1659 Views
  • 2 replies
  • 0 kudos

Resolved! Understanding least common type in databricks

I was reading the data type rules and found about least common type.I have a doubt. What is the least common type of STRING and INT? The referred link gives the following example saying the least common type is BIGINT.-- The least common type between...

  • 1659 Views
  • 2 replies
  • 0 kudos
Latest Reply
Dhruv-22
Contributor
  • 0 kudos

The question is solved here - link

  • 0 kudos
1 More Replies
Dhruv-22
by Contributor
  • 51 Views
  • 4 replies
  • 2 kudos

Resolved! Least Common Type is different in Serverless and All Purpose Cluster.

The following statement gives different outputs in different computes.In Databricks, 15.4 LTS%sqlSELECT typeof(coalesce(5, '6'));-- OutputstringIn Serverless, environment version 4%sqlSELECT typeof(coalesce(5, '6'));-- OutputbigintThere are other cas...

  • 51 Views
  • 4 replies
  • 2 kudos
Latest Reply
MuthuLakshmi
Databricks Employee
  • 2 kudos

@Dhruv-22 Regarding your 1st question, I'm not sureYou can refer to https://docs.databricks.com/aws/en/sql/language-manual/parameters/ansi_mode#system-default to understand what happens when ansi mode is disabled

  • 2 kudos
3 More Replies
anusha98
by Visitor
  • 40 Views
  • 2 replies
  • 2 kudos

Regarding : How to use Row_number() in dlt pipelines

We have two streaming tables : customer_info and customer_info_history and we  joined them using full join to create temp table in pyspark and now we want to eliminate the de-duped records from this temp table. Tried using row_number() but facing bel...

  • 40 Views
  • 2 replies
  • 2 kudos
Latest Reply
K_Anudeep
Databricks Employee
  • 2 kudos

Hello @anusha98 , You’re hitting a real limitation of Structured Streaming: non-time window functions (like row_number() over (...)) aren’t allowed on streaming DFs. You need to use agg().max() to get the “latest value per key” @dlt.table(name="temp_...

  • 2 kudos
1 More Replies
AmarKap
by Visitor
  • 55 Views
  • 1 replies
  • 0 kudos

Lakeflow Pipelines Trying to Read accented file with spark.readStream but failure

Trying to read a accented file(French characters) but the spark.readStream function is not working and special characters turn into something strange(ex. �)             spark.readStream            .format("cloudfiles")            .option("cloudFiles....

  • 55 Views
  • 1 replies
  • 0 kudos
Latest Reply
K_Anudeep
Databricks Employee
  • 0 kudos

Hello @AmarKap , When Spark decodes CP1252 bytes as UTF-8/ISO-8859-1, you’ll see the replacement char like � Can you read the file as : df = (spark.readStream.format("cloudFiles").option("cloudFiles.format", "text").option("encoding", "windows-1252")...

  • 0 kudos
Gustavo_Az
by Contributor
  • 1602 Views
  • 1 replies
  • 0 kudos

Doubt with range_join hints optimization, using INSERT INTO REPLACE WHERE

HelloIm optmizing a big notebook and have encountered many times the tip from databricks that says "Unused range join hints". Reading the documentation for reference, I have been able to supress that warning in almost all cells, but some of then rema...

range_joins.JPG
  • 1602 Views
  • 1 replies
  • 0 kudos
Latest Reply
Prajapathy_NKR
New Contributor
  • 0 kudos

Hi @Gustavo_AzTry to use explain to understand what's happening. https://spark.apache.org/docs/latest/sql-ref-syntax-qry-explain.html

  • 0 kudos
Jonathan_
by New Contributor II
  • 263 Views
  • 6 replies
  • 6 kudos

Slow PySpark operations after long DAG that contains many joins and transformations

We are using PySpark and notice that when we are doing many transformations/aggregations/joins of the data then at some point the execution time of simple task (count, display, union of 2 tables, ...) become very slow even if we have a small data (ex...

  • 263 Views
  • 6 replies
  • 6 kudos
Latest Reply
Jonathan_
New Contributor II
  • 6 kudos

Hi,We forgot to say that we were using a single node cluster (E class with 16 cores). Often in our projects we need to used library that works mainly with data in memory. We also need to remember that here we are not referring to a large data.When we...

  • 6 kudos
5 More Replies
Marthinus
by New Contributor III
  • 63 Views
  • 3 replies
  • 1 kudos

[Databricks Asset Bundles] Bug: driver_node_type_id not updated

Working with databricks asset bundles (using the new python-based definition), if you have a job_cluster defined using driver_node_type_id, and then update it to no longer have it defined, but only node_type_id, the driver node_type never gets update...

  • 63 Views
  • 3 replies
  • 1 kudos
Latest Reply
dkushari
Databricks Employee
  • 1 kudos

Thanks for the details. The way it works is once you set it when you first define the job cluster it does not change when you later remove driver_node_type_id from the spec (i.e., omit it), Databricks does not automatically revert it to match the nod...

  • 1 kudos
2 More Replies
EndreM
by New Contributor III
  • 2130 Views
  • 1 replies
  • 0 kudos

Replay stream to migrate to liquid cluster

The documentation is sparse about how to migrate a partition table to a liquid cluster as the Alter table suggested in the documentation doesnt work when its a partitioned table.The comments on this forum suggest replaying the stream. And this is wha...

  • 2130 Views
  • 1 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

Greetings @EndreM , I did some digging internally and I have come up with some helpful tips/tricks to help guide you through this issue: Based on your situation, you're encountering several common challenges when migrating a partitioned table to liqu...

  • 0 kudos
soumiknow
by Contributor II
  • 2000 Views
  • 1 replies
  • 0 kudos

Unable to create databricks group and add permission via terraform

I have the following terraform code to create a databricks group and add permission to a workflow: resource "databricks_group" "dbx_group" { display_name = "ENV_MONITORING_TEAM" } resource "databricks_permissions" "workflow_permission" { job_id ...

Data Engineering
databricks groups
Terraform
  • 2000 Views
  • 1 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

Greetings @soumiknow , I did some digging internally and found something that may help: Based on the information gathered, I can now draft a comprehensive response to this Databricks Community question about the Terraform authentication issue. ## Dra...

  • 0 kudos
smoortema
by New Contributor III
  • 173 Views
  • 2 replies
  • 2 kudos

Resolved! How to make FOR cycle and dynamic SQL and variables work together

I am working on a testing notebook where the table that is tested can be given as a widget. I wanted to write it in SQL. The notebook does the following steps in a cycle that should run 10 times:1. Store the starting version of a delta table in a var...

  • 173 Views
  • 2 replies
  • 2 kudos
Latest Reply
smoortema
New Contributor III
  • 2 kudos

Thank you! I realised that the example I gave was bad. However, what I was missing is that I did not know how to set a variable in SQL scripting. Including the SET command within the sql string does not work, you have to use the EXECUTE IMMEDIATE ......

  • 2 kudos
1 More Replies
AbhishekNakka
by New Contributor II
  • 56 Views
  • 1 replies
  • 0 kudos

Databricks professional data engineer

Hi, i wanted to know i anyone has given databricks professional data engineering exam recently after oct 2025. I wanted to know if the syllabus has been updated or not ?

  • 56 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @AbhishekNakka ,Yes, the syllabus has been updated. The current exam objectives you can find at below link:Databricks Certified Data Engineer Professional September 2025 - Exam Guide.docxDatabricks Certified Data Engineer Professional | Databricks

  • 0 kudos
DatabricksEngi1
by New Contributor III
  • 133 Views
  • 4 replies
  • 0 kudos

Resolved! Problem in VS Code Extention

Until a few days ago, I was working with Databricks Connect using the VS Code extension, and everything worked perfectly.In my .databrickscfg file, I had authentication configured like this:  [name]host:token: When I ran my code, everything worked fi...

  • 133 Views
  • 4 replies
  • 0 kudos
Latest Reply
dkushari
Databricks Employee
  • 0 kudos

Hi @DatabricksEngi1 - Please ensure you have a Python Venv set up for each Python version that you use with Databricks Connect. Also, I have given step-by-step ways to debug the issue, clear the cache, etc [Read the files and instructions carefully b...

  • 0 kudos
3 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels