cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Ritesh-Dhumne
by Databricks Partner
  • 556 Views
  • 3 replies
  • 1 kudos

Resolved! Databricks Scenarios

I’m a data engineer with some experience in Databricks. I’m looking for real-life scenarios that are commonly encountered by data engineers. Could you also provide details on how to implement these scenarios?

  • 556 Views
  • 3 replies
  • 1 kudos
Latest Reply
Raman_Unifeye
Honored Contributor III
  • 1 kudos

Generic topic. Here are few latest article to help you on thishttps://community.databricks.com/t5/get-started-guides/getting-started-with-databricks-build-a-simple-lakehouse/tac-p/139492#M29https://community.databricks.com/t5/announcements/big-book-o...

  • 1 kudos
2 More Replies
ShaneCorn
by Contributor
  • 399 Views
  • 2 replies
  • 1 kudos

What are the best ways to implement transcription in podcast apps?

I am starting this discussion for everyone who can answer my query.

  • 399 Views
  • 2 replies
  • 1 kudos
Latest Reply
nayan_wylde
Esteemed Contributor II
  • 1 kudos

1. Use Speech-to-Text Models via MLflowIntegrate open-source models like OpenAI Whisper, Hugging Face Wav2Vec2, or AssemblyAI API.Log the model in MLflow for versioning and reproducibility.Deploy as a Databricks Model Serving endpoint for real-time t...

  • 1 kudos
1 More Replies
Dorothy80Galvin
by New Contributor II
  • 3365 Views
  • 3 replies
  • 1 kudos

How can I Resolve QB Desktop Update Error 15225?

I'm encountering QB Desktop update error 15225. What could be causing this issue, and how can I resolve it? It's disrupting my workflow, and I need a quick fix.

  • 3365 Views
  • 3 replies
  • 1 kudos
Latest Reply
jamessmith11
New Contributor II
  • 1 kudos

If you're seeing Update Error 15225, don’t worry — it’s usually fixable. First, check that your internet connection is stable and make sure your computer’s date and time are correct. Then, open Internet Options and verify that SSL settings are turned...

  • 1 kudos
2 More Replies
bianca_unifeye
by Databricks MVP
  • 449 Views
  • 1 replies
  • 2 kudos

Databricks One Lake

Microsoft Ignite always brings exciting updates but the real question is: what do these announcements actually mean for the business, not just for technology teams?That’s exactly what this article is about. I’m breaking down the new Databricks–OnevLa...

  • 449 Views
  • 1 replies
  • 2 kudos
Latest Reply
Raman_Unifeye
Honored Contributor III
  • 2 kudos

Great article @bianca_unifeye Teh move is certainly going to build and unifye the Governance bridge between Azure Databricks and OneLake.

  • 2 kudos
surajitDE
by Contributor
  • 455 Views
  • 1 replies
  • 1 kudos

Resolved! Databricks Dashboard Issue: No Mouse-Based Navigation When Dashboard Tabs Exceed the Top Ribbon

When dashboards have many pages, the top tab bar overflows and can’t be navigated using the mouse. Only left keyboard arrow, right keyboard arrow works, which is slow and inconvenient not user friendly.Expected: ability to scroll tabs with mouse e.g....

  • 455 Views
  • 1 replies
  • 1 kudos
Latest Reply
Advika
Community Manager
  • 1 kudos

Hello @surajitDE! You can use the horizontal scroll bar to navigate through the dashboard pages.If you’re on a trackpad, you can simply scroll horizontally. If you’re using a mouse, you can: Hold Shift and scroll with the mouse wheel (easiest), orDra...

  • 1 kudos
esistfred
by New Contributor III
  • 4816 Views
  • 4 replies
  • 6 kudos

Resolved! How to use variable-overrides.json for environment-specific configuration in Asset Bundles?

Hi all,Could someone clarify the intended usage of the variable-overrides.json file in Databricks Asset Bundles?Let me give some context. Let's say my repository layout looks like this:databricks/ ├── notebooks/ │ └── notebook.ipynb ├── resources/ ...

  • 4816 Views
  • 4 replies
  • 6 kudos
Latest Reply
esistfred
New Contributor III
  • 6 kudos

It does. Thanks for the reponse. I also continued playing around with it and found a way using the variable-overrides.json file. I'll leave it here just in case anyone is interested:Repository layout:databricks/ ├── notebooks/ │ └── notebook.ipynb ...

  • 6 kudos
3 More Replies
chris0991
by New Contributor III
  • 3176 Views
  • 4 replies
  • 1 kudos

Best practices for optimizing Spark jobs

What are some best practices for optimizing Spark jobs in Databricks, especially when dealing large datasets? Any tips or resources would be greatly appreciated! I’m trying to analyze data on restaurant menu prices so that insights would be especiall...

  • 3176 Views
  • 4 replies
  • 1 kudos
Latest Reply
Coffee77
Honored Contributor II
  • 1 kudos

In addition to above cool comments, try to use clusters with VMs enabled for disk caching as well. This caches data at parquet files level in VM local storage, acting as a great complement to spark caching.

  • 1 kudos
3 More Replies
bianca_unifeye
by Databricks MVP
  • 440 Views
  • 0 replies
  • 1 kudos

Agent Bricks Webinar

Our Databricks x Unifeye Meetup community just hit 150 members!  A huge milestone, especially considering we’ve consistently had 50+ people joining every webinar. The momentum is real, and the audience keeps growing! This week, we’re taking it one s...

1763050202042.jpg
  • 440 Views
  • 0 replies
  • 1 kudos
CookDataSol
by New Contributor II
  • 874 Views
  • 2 replies
  • 1 kudos

Resolved! SQL cell v spark.sql in notebooks

I am fairly new to Databricks, and indeed Python, so apologies if this has been answered elsewhere but I've been unable to find it.I have been mainly working in notebooks as opposed to the SQL editor, but coding in SQL where possible using SQL cells ...

  • 874 Views
  • 2 replies
  • 1 kudos
Latest Reply
CookDataSol
New Contributor II
  • 1 kudos

Thanks Louis, really good explanation and helpful examples!

  • 1 kudos
1 More Replies
Judith
by New Contributor III
  • 7759 Views
  • 5 replies
  • 3 kudos

Connect to Onelake using Service Principal, Unity Catalog and Databricks Access Connector

We are trying to connect Databricks to OneLake, to read data from a Fabric workspace into Databricks, using a notebook. We also use Unity Catalog. We are able to read data from the workspace with a Service Principal like this:from pyspark.sql.types i...

Judith_0-1739892045239.png Judith_1-1739891020619.png
  • 7759 Views
  • 5 replies
  • 3 kudos
Latest Reply
Coffee77
Honored Contributor II
  • 3 kudos

As commented you need to assign "Storage Blob Data Contributor or Storage Account Contributor to the service principal you're using in the "connection" provided to the "external location". Another more advanced and even better option would be to use ...

  • 3 kudos
4 More Replies
Radix95
by New Contributor II
  • 4600 Views
  • 1 replies
  • 0 kudos

Updating a Delta Table in Delta Live Tables (DLT) from Two Event Hubs

I am working with Databricks Delta Live Tables (DLT) and need to ingest data from two different Event Hubs. My goal is to:Ingest initial data from the first Event Hub (Predictor) and store it in a Delta Table (data_predictions).Later, update this tab...

  • 4600 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

To achieve robust, persistent CDC (Change Data Capture)–style updates in Databricks DLT with your scenario—while keeping data_predictions as a Delta Table (not a Materialized View)—you need to carefully avoid streaming joins and side effects across s...

  • 0 kudos
ChristianRRL
by Honored Contributor
  • 3912 Views
  • 1 replies
  • 1 kudos

Resolved! Databricks UMF Best Practice

Hi there, I would like to get some feedback on what are the ideal/suggested ways to get UMF data from our Azure cloud into Databricks. For context, UMF can mean either:User Managed FileUser Maintained FileBasically, a UMF could be something like a si...

Get Started Discussions
Data ingestion
UMF
User Maintained File
User Managed File
  • 3912 Views
  • 1 replies
  • 1 kudos
Latest Reply
mark_ott
Databricks Employee
  • 1 kudos

Several effective patterns exist for ingesting User Managed Files (UMF) such as CSVs from Azure into Databricks, each with different trade-offs depending on governance, user interface preferences, and integration with Microsoft 365 services. Common A...

  • 1 kudos
Ibrahim1
by New Contributor
  • 4393 Views
  • 1 replies
  • 0 kudos

DLT detecting changes but not applying them

We have three source tables used for a streaming dimension table in silver. Around 50K records are changed in one of the source tables, and the DLT pipeline shows that it has updated those 50K records, but they remain unchanged. The only way to pick ...

  • 4393 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

The most likely reason your DLT pipeline shows 50K updates but the records remain unchanged is related to how Delta Live Tables (DLT) handle streaming tables, update logic, and schema constraints. When the target table uses an auto-increment ID (espe...

  • 0 kudos
nachii_rajput
by Databricks Partner
  • 3704 Views
  • 1 replies
  • 0 kudos

Issue with Disabled "Repair DAG", "Repair All DAGs" Buttons in Airflow UI, functionality is working.

We are encountering an issue in the Airflow UI where the 'Repair DAG' and 'Repair All DAGs' options are disabled when a specific task fails. While the repair functionality itself is working properly (i.e., the DAGs can still be repaired through execu...

  • 3704 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

The issue with the 'Repair DAG' and 'Repair All DAGs' options being disabled in the Airflow UI when using the Databricks Workflow Operator is a known UI-specific problem that does not affect backend execution or the actual repair functionality. While...

  • 0 kudos
Sudheer2
by New Contributor III
  • 3647 Views
  • 1 replies
  • 1 kudos

How to Fetch Azure OpenAI api_version and engine Dynamically After Resource Creation via Python?

Hello,I am using Python to automate the creation of Azure OpenAI resources via the Azure Management API. I am successfully able to create the resource, but I need to dynamically fetch the following details after the resource is created:API Version (a...

  • 3647 Views
  • 1 replies
  • 1 kudos
Latest Reply
jamesl
Databricks Employee
  • 1 kudos

Hi Sudheer,  It's been a while since you posted, but are you still facing this issue? Here are a few things you could check if needed:  API version In Azure OpenAI, api-version is a query parameter on the data-plane (inference) requests, not a proper...

  • 1 kudos
Labels