cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

DmitriyLamzin
by New Contributor II
  • 4850 Views
  • 2 replies
  • 1 kudos

applyInPandas function hangs in runtime 13.3 LTS ML and above

Hello, recently I've tried to upgrade my runtime env to the 13.3 LTS ML and found that it breaks my workload during applyInPandas.My job started to hang during the applyInPandas execution. Thread dump shows that it hangs on direct memory allocation: ...

Data Engineering
pandas udf
  • 4850 Views
  • 2 replies
  • 1 kudos
Latest Reply
Marcin_Milewski
New Contributor II
  • 1 kudos

Hi @Debayan the link just redirects to the same thread? Is there any update on this issue?We share some similar issue on job hanging using mapInPandas.   

  • 1 kudos
1 More Replies
Sanjeev
by New Contributor II
  • 1564 Views
  • 3 replies
  • 1 kudos

Unverified Commits via Databricks Repos: Seeking Solution for GitHub Verification

The team is integrating Databricks Repos with Personal Access Tokens (PAT) to commit code directly to GitHub. Our organization requires signed commits for verification purposes.Issue: When committing via Databricks Repos, the commits appear as unveri...

Data Engineering
data engineering
  • 1564 Views
  • 3 replies
  • 1 kudos
Latest Reply
Sanjeev
New Contributor II
  • 1 kudos

Can you please share the link to this doc DB-I-3082. I couldn't find it.

  • 1 kudos
2 More Replies
Danny_Lee
by Valued Contributor
  • 1072 Views
  • 1 replies
  • 0 kudos

UI improvement - open multiple workspace notebooks

Hi all,I have an idea for a feature to open multiple workbooks.  Currently, right-clicking a notebook in the Workspace will allow you to "Open in new tab".  If I multi-select notebooks, I only have option to Move or Move to trash.  Why not allow a us...

  • 1072 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

Many thanks for your feedback and great idea. We have created idea DBE-I-1544, this will be analyzed by our team and if approved it can be implemented in the near future.

  • 0 kudos
MOlliver
by New Contributor
  • 6799 Views
  • 1 replies
  • 0 kudos

DBT or Delta Live Tables

Quick question, when would people use DBT over Delta Live Tables? Or better yet can you use DBT to create Delta Live Tables?

  • 6799 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

Delta Live Tables (DLT): DLT is an ETL (Extract, Transform, Load) framework designed to simplify the creation and management of data pipelines. It uses a declarative approach to build reliable data pipelines and automatically manages infrastructure a...

  • 0 kudos
Vishwanath_Rao
by New Contributor II
  • 2278 Views
  • 2 replies
  • 0 kudos

Photon plan invariant violated Error

We've run into a niche error where we get the below message only on our non prod environment, with the same data, with the same code as our prod environment.org.apache.spark.sql.execution.adaptive.InvalidAQEPlanException: Photon plan invariant violat...

  • 2278 Views
  • 2 replies
  • 0 kudos
Latest Reply
JAC2703
New Contributor II
  • 0 kudos

Hey, did you raise a ticket and get a resolution to this? 

  • 0 kudos
1 More Replies
felix_immanuel
by New Contributor III
  • 5305 Views
  • 4 replies
  • 2 kudos

Resolved! Error while Deploying Asset Bundle using Azure Devops

Hi,I'm trying to deploy the Asset Bundle using Azure DevOps, it is giving me this error Step: databricks bundle validate -t dev========================== Starting Command Output ===========================2024-09-02T05:41:19.9113254Z Error: failed du...

  • 5305 Views
  • 4 replies
  • 2 kudos
Latest Reply
sampo
New Contributor II
  • 2 kudos

I had similar error message but then using correct environment variables in the pipeline solved the problem. Especially setting DATABRICKS_HOST point to the account. More detailed description is here Databricks Asset Bundle OAuth Authentication in Az...

  • 2 kudos
3 More Replies
Kibour
by Contributor
  • 9832 Views
  • 3 replies
  • 2 kudos

Resolved! Import from repo

Hi all,I am trying the new "git folder" feature, with a repo that works fine from the "Repos". In the new folder location, my imports from my own repo don't work anymore. Anyone faced something similar?Thanks in advance for sharing your experience

  • 9832 Views
  • 3 replies
  • 2 kudos
Latest Reply
Yuki
Contributor
  • 2 kudos

Hello,When I had the same issue even using 14.3+, I tried this codes.```pythonimport sysimport pprintpprint.pprint(sys.path)```And I noticed that the path was throwed by old(Legacy) Repos folder. I managed the same name of the folder & Repos.Then I r...

  • 2 kudos
2 More Replies
vyasakhilesh
by New Contributor
  • 791 Views
  • 1 replies
  • 0 kudos

Error Creating Table from delta location dbfs

[UC_FILE_SCHEME_FOR_TABLE_CREATION_NOT_SUPPORTED] Creating table in Unity Catalog with file scheme dbfs is not supported. Instead, please create a federated data source connection using the CREATE CONNECTION command for the same table provider, then ...

  • 791 Views
  • 1 replies
  • 0 kudos
Latest Reply
agallard2
New Contributor III
  • 0 kudos

Hi @vyasakhilesh,The error you're seeing, [UC_FILE_SCHEME_FOR_TABLE_CREATION_NOT_SUPPORTED], occurs because Unity Catalog in Databricks does not support creating tables directly from DBFS (Databricks File System) locations.In this case, you're trying...

  • 0 kudos
SethParker
by New Contributor III
  • 1619 Views
  • 6 replies
  • 0 kudos

Resolved! SQL View Formatting in Catalog - Can you turn it off?

It appears as though Databricks now formats SQL View definitions when showing them in the Catalog.  Our solution is based on views, and we have comment tags in those views.  We format these views so that it is easy for us to find and update parts of ...

SethParker_0-1730757882698.png SethParker_1-1730757945105.png
  • 1619 Views
  • 6 replies
  • 0 kudos
Latest Reply
SethParker
New Contributor III
  • 0 kudos

Thank you!  I will submit that request.In case anyone else stumbles upon this post, here is a function you can add that will return the view definition from information_schema, unformatted with an ALTER statement at the top:DROP FUNCTION IF EXISTS <c...

  • 0 kudos
5 More Replies
John_Rotenstein
by New Contributor II
  • 6679 Views
  • 7 replies
  • 2 kudos

ODBC on Windows -- Where to specify Catalog name?

We are attempting to connect a Windows ODBC application to Unity Catalog.The Configure the Databricks ODBC and JDBC drivers documentation has a section titled "ODBC configuration and connection parameters" that mentions a configuration parameter call...

  • 6679 Views
  • 7 replies
  • 2 kudos
Latest Reply
PiotrU
Contributor II
  • 2 kudos

It's quite interesting - I am using Mac, Simba spark ODBC 2.8.2 - and, If I will not add "Catalog" parameter - in UI, I will only see default one (If I will have access to it) - that doesn't mean I cannot query other one, It's just not listed in the ...

  • 2 kudos
6 More Replies
ws4100e
by New Contributor III
  • 7461 Views
  • 9 replies
  • 0 kudos

DLT piplines with UC

I try to run a (very simple) DLT pipeline in with a resulting materialized table is published in UC schema with a managed storage location defined (within an existing EXTERNAL LOCATION). Accoding to the documentation: Publishing to schemas that speci...

  • 7461 Views
  • 9 replies
  • 0 kudos
Latest Reply
ImranA
Contributor
  • 0 kudos

Difference between the Preview and Current Channel?

  • 0 kudos
8 More Replies
maranBH
by New Contributor III
  • 29708 Views
  • 5 replies
  • 11 kudos

Resolved! How to import a function to another notebook using Repos without %run?

Hi all,I was reading the Repos documentation: https://docs.databricks.com/repos.html#migrate-from-run-commandsIt is explained that, one advantage of Repos is no longer necessary to use %run magic command to make funcions available in one notebook to ...

  • 29708 Views
  • 5 replies
  • 11 kudos
Latest Reply
JakubSkibicki
Contributor
  • 11 kudos

Due to new functionalies in Runtime 16.0 regarding autoload i came across this autoload. Performaed a practical test. It works. However had some problems at first.As in solution the key was that definitions are places in a file.py not a notebook.

  • 11 kudos
4 More Replies
lakshgisprog
by New Contributor II
  • 2409 Views
  • 3 replies
  • 0 kudos

Create a simple Geospatial Table with Geography type column

Hello AllI am looking for guidance on creating an simple US states table with shape as Geography column type. I do not want to use Apache Sedona (due to cluster limitations). I am going to create an Node JS application which is going to query this ge...

  • 2409 Views
  • 3 replies
  • 0 kudos
Latest Reply
lakshgisprog
New Contributor II
  • 0 kudos

Thank you for prompt response. Yes, I have gone through the blog. I followed the same process ; for examples, buildings table have geometry which stores as binary type. My question, how to store geometry in 'Geography' type.  The GEOGRAPHY data type ...

  • 0 kudos
2 More Replies
zed
by New Contributor III
  • 2670 Views
  • 5 replies
  • 0 kudos

Can't pass dynamic parameters to non-notebook Python job (spark_python_task)

I need to access the date of a given job running as a non-notebook Python job (spark_python_task). I want to pass a value from the cli when running it and being available to access the value in the scriptI tried the approaches in the attached image w...

  • 2670 Views
  • 5 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

Can you confirm if this solution applies to you https://community.databricks.com/t5/data-engineering/retrieve-job-level-parameters-in-spark-python-task-not-notebooks/td-p/75324 ?

  • 0 kudos
4 More Replies
KFries
by New Contributor II
  • 2939 Views
  • 2 replies
  • 3 kudos

SQL Notebook Tab Spacing

My SQL notebooks in databricks suffer from having at least several different counts of spaces between tab marks.  It makes it very difficult to maintain pretty code spacing.  What sets the tab spacing in SQL language notebooks, and how is it set/adju...

  • 2939 Views
  • 2 replies
  • 3 kudos
Latest Reply
louisv-bambaw
New Contributor II
  • 3 kudos

I’m experiencing the same issue with SQL cell indentation in Databricks notebooks. While editing, I’ve noticed that the indentation level can vary from one cell to another - sometimes it’s two spaces, other times it’s four. This inconsistency makes i...

  • 3 kudos
1 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels