cancel
Showing results for 
Search instead for 
Did you mean: 
Warehousing & Analytics
Engage in discussions on data warehousing, analytics, and BI solutions within the Databricks Community. Share insights, tips, and best practices for leveraging data for informed decision-making.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

MadelynM
by Databricks Employee
  • 1665 Views
  • 0 replies
  • 0 kudos

[Recap] Data + AI Summit 2024 - Warehousing & Analytics | Improve performance and increase insights

Here's your Data + AI Summit 2024 - Warehousing & Analytics recap as you use intelligent data warehousing to improve performance and increase your organization’s productivity with analytics, dashboards and insights.  Keynote: Data Warehouse presente...

Screenshot 2024-07-03 at 10.15.26 AM.png
Warehousing & Analytics
AI BI Dashboards
AI BI Genie
Databricks SQL
  • 1665 Views
  • 0 replies
  • 0 kudos
boitumelodikoko
by Contributor
  • 1129 Views
  • 9 replies
  • 6 kudos

Resolved! Assistance Needed: Issues with Databricks SQL Queries and Performance

Hi everyone,I hope you're all doing well.I'm experiencing some challenges with Databricks SQL, and I wanted to reach out to see if others have encountered similar issues or have suggestions for troubleshooting. Below is a summary of the problems I'm ...

  • 1129 Views
  • 9 replies
  • 6 kudos
Latest Reply
boitumelodikoko
Contributor
  • 6 kudos

Hi @Walter_C,Thank you for your input and support regarding the challenges I’ve been experiencing with Databricks SQL.I followed up with support, and they confirmed that these are known issues currently under review. Here’s a summary of the response:...

  • 6 kudos
8 More Replies
EWhitley
by New Contributor III
  • 5008 Views
  • 3 replies
  • 3 kudos

Resolved! Retrieve task name within workflow task (notebook, python)?

Using workflows, is there a way to obtain the task name from within a task?EX: I have a workflow with a notebook task. From within that notebook task I would like to retrieve the task name so I can use it for a variety of purposes.Currently, we're re...

  • 5008 Views
  • 3 replies
  • 3 kudos
Latest Reply
ttamas
New Contributor III
  • 3 kudos

Hi @EWhitley,Would {{task.name}} help in getting the current task name?https://docs.databricks.com/en/workflows/jobs/parameter-value-references.htmlPass context about job runs into job t 

  • 3 kudos
2 More Replies
boitumelodikoko
by Contributor
  • 2065 Views
  • 2 replies
  • 0 kudos

Resolved! Internal Error During Spark SQL Phase Optimization – Possible Bug in Spark/Databricks Runtime

We are experiencing the following issues.Description:I encountered an issue while executing a Spark SQL query in Databricks, and it seems to be related to the query optimization phase. The error message suggests an internal bug within Spark or the Sp...

  • 2065 Views
  • 2 replies
  • 0 kudos
Latest Reply
boitumelodikoko
Contributor
  • 0 kudos

Update:Response from the Databricks Team.SymptomsInternal Error During Spark SQL Phase Optimization.CauseDataBricks PG Engineering team confirmed that this is indeed a bug in CASE WHEN optimization & they are working on the fix for this issue.Resolut...

  • 0 kudos
1 More Replies
iscpablogarcia
by New Contributor II
  • 333 Views
  • 1 replies
  • 2 kudos

How can i set the workflow status to Skipped?

I have a Python script workflow with 2 tasks: Task A and Task B.When task A has data, this is shared to Task B via createOrReplaceGlobalTempView with no issues.The goal is: When A has no data, skip the Task B and also set the workflow status to "Skip...

iscpablogarcia_0-1737152423551.png
  • 333 Views
  • 1 replies
  • 2 kudos
Latest Reply
Walter_C
Databricks Employee
  • 2 kudos

To achieve the goal of setting the workflow status to "Skipped" when Task A has no data, you can use the "Run if" conditional task type in Databricks Jobs. This allows you to specify conditionals for later tasks based on the outcome of other tasks.ht...

  • 2 kudos
daviddekoning
by New Contributor II
  • 269 Views
  • 1 replies
  • 1 kudos

Resolved! Container Service on Windows base container

I have some legacy software that only runs on Windows, but that can be driven via Python. Is it possible to set up compute resources that run Databricks Container Service on a windows base image, so that I can then add this legacy software and work w...

  • 269 Views
  • 1 replies
  • 1 kudos
Latest Reply
Walter_C
Databricks Employee
  • 1 kudos

Unfortunately this is not possible, as part of the requirements you need to use an Ubuntu image: https://docs.databricks.com/en/compute/custom-containers.html#option-2-build-your-own-docker-base 

  • 1 kudos
Dnirmania
by Contributor
  • 290 Views
  • 1 replies
  • 3 kudos

Foreign Catalog - AWS Redshift

Hello EveryoneRecently, my team integrated an AWS Redshift database as a foreign catalog in Azure Databricks. We can successfully run SELECT queries and create regular views on top of the foreign catalog table. However, when attempting to create a ma...

  • 290 Views
  • 1 replies
  • 3 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 3 kudos

Hi @Dnirmania, Materialized views in SQL often use serverless Delta Live Tables pipelines, which might be causing the connection timeout due to IP whitelisting restrictions. Serverless compute might not be able to connect to the federated source if t...

  • 3 kudos
VCA50380
by Contributor
  • 619 Views
  • 5 replies
  • 2 kudos

Resolved! Best practice to materialize data

Hello all,I'm quite new to Databricks world, and currently in the process of analyzing a migration from Oracle on-premise (with a lot of SQL, PL/SQL, custom things, etc.) to Databricks.Let's try to illustrate my situation in Oracle (summary):. Let's ...

  • 619 Views
  • 5 replies
  • 2 kudos
Latest Reply
Walter_C
Databricks Employee
  • 2 kudos

Hello, sure if any additional information is needed just let us know and we will be happy to assist.

  • 2 kudos
4 More Replies
Akshay_Petkar
by Contributor III
  • 3227 Views
  • 5 replies
  • 1 kudos

Resolved! Migrate Azure Synapse Analytics data to Databricks

I have to migrate the data from Azure Synapse Analytics to Databricks. Could anyone share the different approaches to migrate data, and from those, which is the best approach to use?

  • 3227 Views
  • 5 replies
  • 1 kudos
Latest Reply
raushan_kumar
New Contributor II
  • 1 kudos

Can we mount the Syanpse storage account ADLS Gen2 directly to Databricks and then we can only  migrate pipeline/code to databricks

  • 1 kudos
4 More Replies
VKe
by New Contributor III
  • 3584 Views
  • 4 replies
  • 0 kudos

install Oracle Instant Client

Hello all,I want to install Oracle Instant Client to be able to use python-oracledb in Thick-Mode because one of our databases is old and cannot be reached in Thin-Mode. I have tried the solution from this post, but it doesn't help me. It seems that ...

  • 3584 Views
  • 4 replies
  • 0 kudos
Latest Reply
NelsonE
New Contributor III
  • 0 kudos

What about shared compute, is it possible? Neither of those scripts is working for me really...

  • 0 kudos
3 More Replies
KosmaS
by New Contributor III
  • 335 Views
  • 1 replies
  • 2 kudos

SQL Warehouse - Turning off caching

Hi All,I'd like to do some benchmarking and I need to turn off caching on my SQL Warehouse Server. However, whatever I try I receive quite high level of caching after running my queries (>60%). I tried to turn off my server, but it automatically wake...

  • 335 Views
  • 1 replies
  • 2 kudos
Latest Reply
Walter_C
Databricks Employee
  • 2 kudos

To disable query result caching, you can run SET use_cached_result = false in the SQL editor.

  • 2 kudos
jay99
by New Contributor II
  • 823 Views
  • 4 replies
  • 1 kudos

Create multiple SQL warehouse with custom access control

We set up the SQL warehouse IAM role in the settings option. This is applied to all warehouses. How do I create sql warehouses with multiple IAM roles to maintain access control.

  • 823 Views
  • 4 replies
  • 1 kudos
Latest Reply
Walter_C
Databricks Employee
  • 1 kudos

Unfortunately there is no way to restrict the access the compute has, the restrictions are being performed via the users permissions. Only option here will be to submit a feature request through https://docs.databricks.com/en/resources/ideas.html#ide...

  • 1 kudos
3 More Replies
Akshay_Petkar
by Contributor III
  • 2583 Views
  • 7 replies
  • 5 kudos

Need a Sample MERGE INTO Query for SCD Type 2 Implementation

Can anyone provide a sample MERGE INTO SQL query for implementing SCD Type 2 in Databricks using Delta Tables?

  • 2583 Views
  • 7 replies
  • 5 kudos
Latest Reply
bhanu_gautam
Contributor III
  • 5 kudos

@JissMathew  and @David_Torrejon , Thanks for sharing the example

  • 5 kudos
6 More Replies
chari
by Contributor
  • 8245 Views
  • 3 replies
  • 2 kudos

What is databricks SQL, spark SQL and how are they different from MS SQL ?

Hello Databricks Community,I have a hard time understanding how is Databricks SQL different from microsoft SQL ? Also, why does databricks provide spark SQL ? If you direct me to a well-written webpage or document its of immense help!Thanks,

  • 8245 Views
  • 3 replies
  • 2 kudos
Latest Reply
Rahul_Saini
New Contributor II
  • 2 kudos

Databricks SQL and Spark SQL are built for distributed big data analytic. Databricks SQL is great for business intelligence tools and uses Delta Lake for efficient data storage. Spark SQL works with Spark's programming features for data processing. U...

  • 2 kudos
2 More Replies
nomnomnom543
by New Contributor II
  • 1127 Views
  • 3 replies
  • 2 kudos

Databricks SQL Wildcard Operator Not Parsed Correctly

Hello there,Wasn't sure if this was just an error on my part, but I'm using a Databricks Pro SQL warehouse and unity catalogue to pull some data from my tables. I'm having this issue where whenever I try and use a wildcard operator with my LIKE claus...

  • 1127 Views
  • 3 replies
  • 2 kudos
Latest Reply
Rahul_Saini
New Contributor II
  • 2 kudos

Hi @nomnomnom543 ,Try this SELECT * FROM table_name WHERE LEFT(column_name, LENGTH('string')) = 'string';

  • 2 kudos
2 More Replies
amelia1
by New Contributor II
  • 1715 Views
  • 1 replies
  • 0 kudos

Local pyspark read data using jdbc driver returns column names only

Hello,I have an Azure sql warehouse serverless instance that I can connect to using databricks-sql-connector. But, when I try to use pyspark and jdbc driver url, I can't read or write.See my code belowdef get_jdbc_url(): # Define your Databricks p...

  • 1715 Views
  • 1 replies
  • 0 kudos
Latest Reply
NandiniN
Databricks Employee
  • 0 kudos

The error does not look specific to the warehouse that you are connecting to. The error message "Unrecognized conversion specifier [msg] starting at position 54 in conversion pattern" indicates that there is an issue with the logging configuration in...

  • 0 kudos