cancel
Showing results for 
Search instead for 
Did you mean: 
Warehousing & Analytics
Engage in discussions on data warehousing, analytics, and BI solutions within the Databricks Community. Share insights, tips, and best practices for leveraging data for informed decision-making.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

MadelynM
by Databricks Employee
  • 3271 Views
  • 0 replies
  • 0 kudos

[Recap] Data + AI Summit 2024 - Warehousing & Analytics | Improve performance and increase insights

Here's your Data + AI Summit 2024 - Warehousing & Analytics recap as you use intelligent data warehousing to improve performance and increase your organization’s productivity with analytics, dashboards and insights.  Keynote: Data Warehouse presente...

Screenshot 2024-07-03 at 10.15.26 AM.png
Warehousing & Analytics
AI BI Dashboards
AI BI Genie
Databricks SQL
  • 3271 Views
  • 0 replies
  • 0 kudos
koral
by New Contributor
  • 3394 Views
  • 1 replies
  • 0 kudos

Technical Issue with Databricks Marketplace Lead Notifications

Hi,We publish listings on the Databricks Marketplace.Several times a week, we receive a lead notification stating: "db sp has installed 'Etsy Dataset'."It’s always the same dataset and the same name, so it’s likely a technical error.Does anyone know ...

koral_0-1744190031248.png
  • 3394 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Yes, this is a known issue reported by other Databricks Marketplace publishers: receiving frequent lead notifications stating "db sp has installed 'Etsy Dataset'" for the same dataset and name, which suggests an underlying technical error in the lead...

  • 0 kudos
bobmclaren
by New Contributor II
  • 4516 Views
  • 1 replies
  • 0 kudos

Comparing Methods for Scheduling Streaming updates via dbt

We are trying to schedule updates to streaming tables and materialized views in Azure Databricks that we have defined in dbt.Two options we are considering are `SCHEDULE CRON` and just scheduling `dbt run` commands via CI/CD. The `SCHEDULE CRON` opti...

  • 4516 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Your understanding of the trade-offs between using SCHEDULE CRON on streaming tables/materialized views and the dbt run orchestration in Azure Databricks is largely correct, and you’ve identified the two main pain points: orchestration of dependencie...

  • 0 kudos
Ajbi
by New Contributor II
  • 3617 Views
  • 1 replies
  • 0 kudos

Power BI refresh from Databricks workflow

We recently enabled the preview feature to refresh power bi datasets from data bricks workflow, but when we run the job data changes aren't being applied despite a completed status in refresh history. Could anyone clarify whether this is intended pre...

Warehousing & Analytics
Databricks Power BI Task
power bi refresh
  • 3617 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

This is a known issue with the Power BI/Databricks integration preview for refreshing Power BI datasets from Databricks workflows. Even when a job completes and the refresh history shows success, actual data changes—such as newly inserted rows—are of...

  • 0 kudos
ZoraidaHS
by New Contributor
  • 3885 Views
  • 1 replies
  • 0 kudos

Disjunction on static filters for a Widget

I would like to be able to express something like: WHERE columnA = "valueA" OR columnB = "valueB" but on the Widget. I only see the possibility that chaining filters that are processed as AND operator. Am I missing something?

  • 3885 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Most widget filter interfaces only support multiple filters using an AND operator, meaning each condition must be true for a row to be included. If you want to use an OR operator, such as filtering where columnA = "valueA" OR columnB = "valueB", most...

  • 0 kudos
JZhang
by New Contributor
  • 3898 Views
  • 1 replies
  • 0 kudos

Issues in visualizing tables when exporting dashboard to HTML

Hello communityI am facing an issue when I export a dashboard to an HTML: the problem is that the HTML has white background but the exported tables have black colors, like the example I am showing below Does someone know how to handle this visualizat...

JZhang_0-1740495725310.png
  • 3898 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

To resolve the issue of dashboard tables exporting to HTML with mismatched colors—specifically black table backgrounds against a white HTML page—review and adjust your dashboard’s CSS or formatting settings before export. The problem typically occurs...

  • 0 kudos
Aminsnh
by New Contributor
  • 4633 Views
  • 1 replies
  • 0 kudos

Adding customized shortcut keys

Hi all, I need to add a shortcut key for R's pip operator (%>%) to my Databricks notebook. I want the operator to be written in my code snippet when I hold down the shortcut keys (shift + ctrl + m). Is there a straightforward way to add such shortcut...

  • 4633 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Databricks notebooks do not natively support defining custom keyboard shortcuts to insert specific code snippets like R’s pipe operator () within the notebook editor UI. The list of standard shortcuts available is limited to navigation, cell manageme...

  • 0 kudos
patilsuhasv
by New Contributor
  • 4133 Views
  • 2 replies
  • 0 kudos

Dela Table and history

Hi All,How can I maintain 7 years of transactional data in delta table? Can I have log retention of 7 days, but data retention of 7 years?Appreciate your response.Thanks and regards Suhas

  • 4133 Views
  • 2 replies
  • 0 kudos
Latest Reply
Isi
Honored Contributor III
  • 0 kudos

Hi @patilsuhasv,Maintaining 7 years of transactional data in a Delta Lake table is technically possible, but it requires careful configuration — and it’s important to understand the recent changes Databricks announced (effective December 2025) regard...

  • 0 kudos
1 More Replies
AndreasWagner
by New Contributor II
  • 6700 Views
  • 2 replies
  • 1 kudos

PowerBI connected to SAP Databricks

hi everyone, does somebody have experience with connecting PowerBI to SAP Databricks in the BDC? I have quite a few SAP customers interested in that ... many thanks, Andreas

  • 6700 Views
  • 2 replies
  • 1 kudos
Latest Reply
DanielFroehler
New Contributor II
  • 1 kudos

Hi @WiliamRosa ,do you know which feature is missing in SAP Databricks that it doesnt work? As Andreas said- everybody is asking that question.KR Daniel

  • 1 kudos
1 More Replies
CEH
by New Contributor II
  • 1061 Views
  • 5 replies
  • 5 kudos

Union of tiny dataframes exhausts resource, memory error

As part of a function I create df1 and df2 and aim to stack them and output the results.  But the results do not display within the function, nor if I output the results and display after.results = df1.unionByName(df2, allowMissingColumns=False)displ...

  • 1061 Views
  • 5 replies
  • 5 kudos
Latest Reply
Advika
Community Manager
  • 5 kudos

Hello @CEH! Did any of the suggestions above help resolve the issue?If so, please mark the most helpful reply as the accepted solution. Or, if you found another fix, please share it with the community so others can benefit as well.

  • 5 kudos
4 More Replies
Akshay_Petkar
by Valued Contributor
  • 9891 Views
  • 9 replies
  • 7 kudos

Resolved! Need a Sample MERGE INTO Query for SCD Type 2 Implementation

Can anyone provide a sample MERGE INTO SQL query for implementing SCD Type 2 in Databricks using Delta Tables?

  • 9891 Views
  • 9 replies
  • 7 kudos
Latest Reply
jeffreyaven
Databricks Employee
  • 7 kudos

Here is a simple example using an upstream Delta table with ChangeDataFeed enabled, using table_changes() to get the records with their corresponding operation, this is a 2 step process you need to close out modified or deleted recordsadd new rows (i...

  • 7 kudos
8 More Replies
Kaz
by New Contributor II
  • 11061 Views
  • 4 replies
  • 2 kudos

Show full logs on job log

Is it possible to show the full logs of a databricks job? Currently, the logs are skipped with:*** WARNING: max output size exceeded, skipping output. ***However, I don't believe our log files are more than 20 MB. I know you can press the logs button...

  • 11061 Views
  • 4 replies
  • 2 kudos
Latest Reply
jkb7
New Contributor III
  • 2 kudos

Any news on this topic?Have the limits on the notebook-cell-log output been resolved?

  • 2 kudos
3 More Replies
tarunnagar
by Contributor
  • 691 Views
  • 3 replies
  • 2 kudos

Tips for Streamlining Spark Job Development and Debugging in Databricks

Hi everyone,I’m looking to improve the efficiency of developing and debugging Spark jobs within Databricks and wanted to get insights from the community. Spark is incredibly powerful, but as projects grow in complexity, it can become challenging to m...

  • 691 Views
  • 3 replies
  • 2 kudos
Latest Reply
Suheb
Contributor
  • 2 kudos

Developing and debugging Spark jobs in Databricks can be challenging due to the distributed nature of Spark and the volume of data processed. To streamline your workflow:Leverage Notebooks for Iterative Development:Use Databricks notebooks to write a...

  • 2 kudos
2 More Replies
mausch
by New Contributor
  • 4044 Views
  • 1 replies
  • 0 kudos

CalledProcessError when running dbt

I've been trying to run a dbt project (sourced in Azure DevOps) in Databricks Workflows, but I get this error message:  CalledProcessError: Command 'b'\nmkdir -p "/tmp/tmp-dbt-run-1124228490001263"\nunexpected_errors="$(cp -a -u "/Workspace/Repos/.in...

  • 4044 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

The error you encountered when running your dbt project in Databricks Workflows comes from Databricks trying to copy the entire repository, including the virtual environment (venv) folder and its cached bytecode files (__pycache__), into a temporary ...

  • 0 kudos
rajanator
by New Contributor
  • 979 Views
  • 1 replies
  • 1 kudos

Resolved! Intermittent 400 Error with Power BI Desktop - ODBC Connection to SQL Warehouse

Hi all,I'm experiencing an intermittent connection issue between Power BI Desktop and our Azure Databricks SQL Warehouse and looking for help troubleshootingError Message:ODBC: ERROR [HY000] [Microsoft][ThriftExtension] (14) Unexpected response from ...

  • 979 Views
  • 1 replies
  • 1 kudos
Latest Reply
mark_ott
Databricks Employee
  • 1 kudos

The intermittent ODBC error you’re seeing in Power BI when connecting to Azure Databricks is a recognized issue related to SSL validation interruptions or proxy interference in the Simba ThriftExtension layer. The behavior—random occurrences, tempora...

  • 1 kudos
Bakkie
by New Contributor III
  • 4202 Views
  • 3 replies
  • 4 kudos

Resolved! Databricks Apps based on Streamlit could not find a valid JAVA_HOME installation

We are launching our first Databricks Apps based on Streamlit.The App works when simply running the notebook in our workspace, but fails after deployment due to "could not find a valid JAVA_HOME installation" when running in the system environment.We...

  • 4202 Views
  • 3 replies
  • 4 kudos
Latest Reply
NandiniN
Databricks Employee
  • 4 kudos

Databricks Apps (which use a lightweight, container-based runtime) do not automatically include JVM, best is to use the databricks package to not have dependency issues.

  • 4 kudos
2 More Replies