cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

tarunnagpal
by New Contributor III
  • 1033 Views
  • 7 replies
  • 3 kudos

Lakebridge questions

We have a few questions before we propose Lakebridge as the migration tooling for one of our customers, where the requirement is to migrate from Redshift to Databricks. We need help with your quick response so we can proceed with the next steps:Our u...

  • 1033 Views
  • 7 replies
  • 3 kudos
Latest Reply
sky_bricks
Visitor
  • 3 kudos

Hi community,We’re currently planning a migration from an on-premise SQL Server data warehouse (with associated SSIS packages) to Databricks Unity Catalog. As part of this effort, we’re evaluating the use of Lakebridge for assessment, conversion, and...

  • 3 kudos
6 More Replies
Jonathan_
by New Contributor III
  • 355 Views
  • 7 replies
  • 6 kudos

Slow PySpark operations after long DAG that contains many joins and transformations

We are using PySpark and notice that when we are doing many transformations/aggregations/joins of the data then at some point the execution time of simple task (count, display, union of 2 tables, ...) become very slow even if we have a small data (ex...

  • 355 Views
  • 7 replies
  • 6 kudos
Latest Reply
Jonathan_
New Contributor III
  • 6 kudos

It's a cluster with 128 GO of memory, when looking in Spark UI there is 54 GO for storage memory. Honestly I don't think it's memory issue like I said it's a small data and if we do checkpoint at same point then continu we don't have the problem afte...

  • 6 kudos
6 More Replies
rajg
by New Contributor
  • 128 Views
  • 1 replies
  • 1 kudos

Cannot export embedded dashboard widget as CSV or other formats except PNG

I’ve integrated a Databricks dashboard into my web application for all my users, following the guidelines in this article:Embedding Databricks Dashboards.This integration worked perfectly initially. However, I’m now encountering an issue with exporti...

rajg_0-1760793195625.png rajg_1-1760793235941.png rajg_2-1760793253909.png
  • 128 Views
  • 1 replies
  • 1 kudos
Latest Reply
stbjelcevic
Databricks Employee
  • 1 kudos

Hi @rajg , Based on the link you shared, it looks to me like you have an external embedding situation? If so, this is a feature that is not currently available, but it is a commonly requested feature. External dashboard embedding is currently in Publ...

  • 1 kudos
Leladams
by New Contributor III
  • 13898 Views
  • 10 replies
  • 2 kudos

What is the best way to read in a ms access .accdb database into Databricks from a mounted drive?

I am currently trying to read in .accdb files from a mounted drive. Based on my research it looks like I would have to use a package like JayDeBeApi with ucanaccess drivers or pyodbc with ms access drivers.Will this work?Thanks for any help.

  • 13898 Views
  • 10 replies
  • 2 kudos
Latest Reply
Anonymous
Not applicable
  • 2 kudos

Hi @Leland Adams​ Hope you are doing well. Thank you for posting your question and giving us additional information. Do you think you were able to solve the query?We'd love to hear from you.

  • 2 kudos
9 More Replies
DBU100725
by New Contributor II
  • 287 Views
  • 2 replies
  • 0 kudos

URGENT: Delta writes to S3 fail after workspace migrated to Premium

Delta writes to S3 fail after workspace migrated to Premium (401 “Credential was not sent or unsupported type”)SummaryAfter our Databricks workspace migrated from Standard to Premium, all Delta writes to S3 started failing with:com.databricks.s3commi...

  • 287 Views
  • 2 replies
  • 0 kudos
Latest Reply
dkushari
Databricks Employee
  • 0 kudos

Hi @DBU100725 - Are you using a No isolation shared cluster? Can you check if this was turned ON in your account?  

  • 0 kudos
1 More Replies
Shefali
by New Contributor
  • 421 Views
  • 1 replies
  • 1 kudos

Lakebridge conversion tool: Incorrect Databricks SQL script generated

Hi Team,I was able to successfully install and use the Lakebridge code conversion tool to convert my SQL Server script into a Databricks SQL script. However, the generated script contains several syntax errors. Could you please let me know if I might...

  • 421 Views
  • 1 replies
  • 1 kudos
Latest Reply
AbhaySingh
New Contributor
  • 1 kudos

Hi there!Known lakebase issues are listed here:https://github.com/databrickslabs/lakebridge/issuesDoes any of this apply to your use case?1. Variable scope errors in WHERE clauses or subqueries  2. DELETE/UPDATE FROM statements incorrectly converted ...

  • 1 kudos
Oumeima
by New Contributor
  • 1819 Views
  • 1 replies
  • 1 kudos

I can't use my own .whl package in Databricks app with databricks asset bundles

I am building a databricks app using databricks asset bundles. I need to use a helpers packages that i built as an artifact and using in other resources outside the app. The only way to use it is to have the built package inside the app source code f...

  • 1819 Views
  • 1 replies
  • 1 kudos
Latest Reply
stbjelcevic
Databricks Employee
  • 1 kudos

Hi @Oumeima , One potential way around this is to upload the wheel file into a Unity Catalog volume or workspace file. For the volume route, reference it directly in your app’s requirements.txt using an absolute /Volumes/<catalog>/<schema>/<volume>/....

  • 1 kudos
Davila
by New Contributor II
  • 1524 Views
  • 1 replies
  • 1 kudos

Issue with Root Folder Configuration in Databricks Asset Bundles for DLT Pipelines

I'm currently working with Databricks Asset Bundles to deploy my DLT pipelines, but I’ve encountered an issue I can't resolve.The problem is that I’m unable to configure the root folder within the Asset Bundle in a way that lets me define a custom pa...

Davila_1-1750959884979.png Davila_2-1750960256363.png
  • 1524 Views
  • 1 replies
  • 1 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 1 kudos

Hey @Davila , I did some digging and have come with some things you can think about as you work through your issue. Here’s a clear way to think about what you’re seeing and how to proceed. What’s going on That “Root folder” field in the DLT UI is in...

  • 1 kudos
lauraxyz
by Contributor
  • 1985 Views
  • 6 replies
  • 0 kudos

Notebook in path workspace/repos/.internal/**_commits/** was unable to be accessed

I have a workflow job (source is git) to access a notebook and execute it.  From the job, it failed with error:Py4JJavaError: An error occurred while calling o466.run. : com.databricks.WorkflowException: com.databricks.NotebookExecutionException: FAI...

  • 1985 Views
  • 6 replies
  • 0 kudos
Latest Reply
lauraxyz
Contributor
  • 0 kudos

Just some clarification:  the caller notebook can be found with no issues, no matter the task's source is GIT or WORKSPACE.  However, the callee notebook, which is called by the caller notebook with dbutils.notebook.run(), cannot be found if the call...

  • 0 kudos
5 More Replies
JordanYaker
by Contributor
  • 2325 Views
  • 2 replies
  • 0 kudos

Integration options for Databricks Jobs and DataDog?

I know that there is already the Databricks (technically Spark) integration for DataDog. Unfortunately, that integration only covers the cluster execution itself and that means only Cluster Metrics and Spark Jobs and Tasks. I'm looking for somethin...

  • 2325 Views
  • 2 replies
  • 0 kudos
Latest Reply
greg-0935
New Contributor
  • 0 kudos

Personally, I'm using their Data Jobs Monitoring product https://docs.datadoghq.com/data_jobs/databricks/ that works great and gives the right insights both for my high level job execution stats and Spark deeper metrics

  • 0 kudos
1 More Replies
Dhruv-22
by Contributor
  • 74 Views
  • 2 replies
  • 1 kudos

Resolved! Can't mergeSchema handle int and bigint?

I have a table which has a column of data type 'bigint'. While overwriting it with new data, given that I do full loads, I used 'mergeSchema' to handle schema changes. The new data's datatype was int. I thought mergeSchema can easily handle that, but...

  • 74 Views
  • 2 replies
  • 1 kudos
Latest Reply
Chiran-Gajula
New Contributor
  • 1 kudos

Hi Dhruv,Delta won't automatically upcast unless you explicitly handle it. Cast the column Lob_Pk to LongType (which maps to BIGINT in SQL/Delta). Try below snippetfrom pyspark.sql.functions import colfrom pyspark.sql.types import LongTypecrm_retail_...

  • 1 kudos
1 More Replies
saicharandeepb
by New Contributor III
  • 79 Views
  • 3 replies
  • 1 kudos

How to Retrieve DBU Count per Compute Type for Accurate Cost Calculation?

Hello Everyone,We are currently working on a cost analysis initiative to gain deeper insights into our Databricks usage. As part of this effort, we are trying to calculate the hourly cost of each Databricks compute instance by utilizing the Azure Ret...

  • 79 Views
  • 3 replies
  • 1 kudos
Latest Reply
nayan_wylde
Honored Contributor III
  • 1 kudos

1. Is there a documented way to retrieve the DBU count per VM or compute type?Yes, but it's not directly exposed via a single API or table. The DBU consumption rate depends on:Compute type (Jobs Compute, All-Purpose Compute, SQL Compute, etc.)VM inst...

  • 1 kudos
2 More Replies
Marthinus
by New Contributor III
  • 101 Views
  • 4 replies
  • 1 kudos

[Databricks Asset Bundles] Bug: driver_node_type_id not updated

Working with databricks asset bundles (using the new python-based definition), if you have a job_cluster defined using driver_node_type_id, and then update it to no longer have it defined, but only node_type_id, the driver node_type never gets update...

  • 101 Views
  • 4 replies
  • 1 kudos
Latest Reply
Chiran-Gajula
New Contributor
  • 1 kudos

There is no built-in way in Databricks Asset bundles or terraform to automatically inherit the value of driver_node_type_id for node_type_id."You must set both explicitly in your configuration"You can always see your updated detail resource from the ...

  • 1 kudos
3 More Replies
Dhruv-22
by Contributor
  • 1687 Views
  • 2 replies
  • 0 kudos

Resolved! Understanding least common type in databricks

I was reading the data type rules and found about least common type.I have a doubt. What is the least common type of STRING and INT? The referred link gives the following example saying the least common type is BIGINT.-- The least common type between...

  • 1687 Views
  • 2 replies
  • 0 kudos
Latest Reply
Dhruv-22
Contributor
  • 0 kudos

The question is solved here - link

  • 0 kudos
1 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels