cancel
Showing results for 
Search instead for 
Did you mean: 
Warehousing & Analytics
Engage in discussions on data warehousing, analytics, and BI solutions within the Databricks Community. Share insights, tips, and best practices for leveraging data for informed decision-making.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

MadelynM
by Databricks Employee
  • 3392 Views
  • 0 replies
  • 0 kudos

[Recap] Data + AI Summit 2024 - Warehousing & Analytics | Improve performance and increase insights

Here's your Data + AI Summit 2024 - Warehousing & Analytics recap as you use intelligent data warehousing to improve performance and increase your organization’s productivity with analytics, dashboards and insights.  Keynote: Data Warehouse presente...

Screenshot 2024-07-03 at 10.15.26 AM.png
Warehousing & Analytics
AI BI Dashboards
AI BI Genie
Databricks SQL
  • 3392 Views
  • 0 replies
  • 0 kudos
Akshay_Petkar
by Valued Contributor
  • 812 Views
  • 0 replies
  • 0 kudos

Issue with Column Name Conflict While Importing .gz File into Spark DataFrame

I'm encountering an issue while importing a .gz file containing JSON data into a Spark DataFrame in Databricks. The error indicates a column name conflict. Could you please advise on how to resolve this issue and handle duplicate column names during ...

  • 812 Views
  • 0 replies
  • 0 kudos
PM0
by New Contributor
  • 1477 Views
  • 0 replies
  • 0 kudos

Rate limit Usage

We are having Databricks tables and we have an API which query that data, the query will be dynamic and API would allow user to query anything. However users can query lot of data and consume lot of DBU but generic rate limiting wont help, as any sin...

  • 1477 Views
  • 0 replies
  • 0 kudos
pauloquantile
by New Contributor III
  • 17140 Views
  • 6 replies
  • 1 kudos

PowerBI "Token expired while fetching results: TEAuthTokenExpired."

Hi everyone,We are at the moment stumbeling upon a big challenge with loading data into PowerBI. I need some advice!To give a bit of conext: we introduced Databricks instead of Azure Synapse for a client of ours. We are currently busy with moving all...

pauloquantile_0-1693824616520.png
  • 17140 Views
  • 6 replies
  • 1 kudos
Latest Reply
viralpatel
New Contributor II
  • 1 kudos

@Retired_mod Is there any further update from Databricks which can be helpful here OR what @pauloquantile mentioned is the only workaround solution?

  • 1 kudos
5 More Replies
Spoon_Man
by New Contributor II
  • 1710 Views
  • 2 replies
  • 2 kudos

Converting an Entity Attribute Value Data Model to Struct datatype

I'm new to databricks and I have a source data model that stores the data as Name-Value pairs (i.e. normalised) in two columns in the table.EntityIDNameValue1Field1SomeValue11Field2SomeValue21Field3SomeValue32Field1SomeValue12Field3SomeValue3The defi...

  • 1710 Views
  • 2 replies
  • 2 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 2 kudos

Your first approach didn't work, because named_struct needs it's arguments on odd postition to be foldable.So you can think of it in following way, at compile time compiler needs to "see" this value. That's why even if you prepared proper expression ...

  • 2 kudos
1 More Replies
Executivecars
by New Contributor
  • 1528 Views
  • 2 replies
  • 0 kudos

Looking for Chauffeur service Melbourne? Visit Executive cars for Chauffeur service Melbourne. ✓24/7 Operation ✓100+ Vehicles ✓ Chauffeur Cars Melbour...

Looking for Chauffeur service Melbourne? Visit Executive cars for Chauffeur service Melbourne. ✓24/7 Operation ✓100+ Vehicles ✓ Chauffeur Cars Melbourne.

  • 1528 Views
  • 2 replies
  • 0 kudos
Latest Reply
FiveLittleFish
New Contributor II
  • 0 kudos

If you're in Melbourne and need reliable chauffeur services, Executive Cars seems like a great option with its extensive fleet and round-the-clock availability. Safety and reliability are crucial in the transportation industry. Speaking of which, hav...

  • 0 kudos
1 More Replies
KrisMcDougal
by New Contributor
  • 1905 Views
  • 2 replies
  • 2 kudos

Is there any way to see who ran a notebook?

I'm trying to find information to see who runs a notebook. I'm able to see who created the notebook, and I can find out when the notebook is ran, but there doesn't seem to be any information on who ran it, only who created the notebook. 

  • 1905 Views
  • 2 replies
  • 2 kudos
Latest Reply
Rishabh-Pandey
Databricks MVP
  • 2 kudos

SELECT * FROM system.logs WHERE event_type = 'notebook_run' ORDER BY timestamp DESC;@KrisMcDougal with this code you can check from which user id the first command of the notebook got started and you will get to know who started the notebook.

  • 2 kudos
1 More Replies
OluPopoola
by New Contributor II
  • 1404 Views
  • 1 replies
  • 0 kudos

Cant Use Delta Live Tables to read MSK using IAM authenthication

Hi AllI am trying to use Delta Live Tables to connect to MSK.We have set up serverless MSK clusters that use IAM for its authetication. I cannot connect to it from a dlt notebook. The same code near enough works on normal clusters that have java libr...

  • 1404 Views
  • 1 replies
  • 0 kudos
Latest Reply
OluPopoola
New Contributor II
  • 0 kudos

Just rephrasing the question:I am trying to use the DLT to connect to serverless MSK clusters authenticated by IAM. The code works on ordinary clusters but doesn't work when run on DLT clusters. I think the issue is the authentication because we can ...

  • 0 kudos
Akshay_Petkar
by Valued Contributor
  • 1341 Views
  • 1 replies
  • 1 kudos

Impact of Overwriting Databricks SQL Tables on Versioning

When changes are made to a Databricks SQL table, a new version is created. If changes to the table are made using Spark or Python in a notebook and the table is overwritten, will a new version be created, or will it remain as version number 0?

  • 1341 Views
  • 1 replies
  • 1 kudos
Latest Reply
Rishabh-Pandey
Databricks MVP
  • 1 kudos

When changes are made to a Databricks SQL table (Delta table) using Spark or Python in a notebook, and the table is overwritten, a new version will indeed be created. It will not remain as version number 0. Each overwrite operation increments the ver...

  • 1 kudos
MadelynM
by Databricks Employee
  • 3392 Views
  • 0 replies
  • 0 kudos

[Recap] Data + AI Summit 2024 - Warehousing & Analytics | Improve performance and increase insights

Here's your Data + AI Summit 2024 - Warehousing & Analytics recap as you use intelligent data warehousing to improve performance and increase your organization’s productivity with analytics, dashboards and insights.  Keynote: Data Warehouse presente...

Screenshot 2024-07-03 at 10.15.26 AM.png
Warehousing & Analytics
AI BI Dashboards
AI BI Genie
Databricks SQL
  • 3392 Views
  • 0 replies
  • 0 kudos
Akshay_Petkar
by Valued Contributor
  • 1993 Views
  • 1 replies
  • 0 kudos

Databrics offers auto tuning?

is databricks offers auto tunning?

  • 1993 Views
  • 1 replies
  • 0 kudos
Latest Reply
raphaelblg
Databricks Employee
  • 0 kudos

Yes, First of all, open source spark already has a set of auto-tuning features denominated Adaptive Query Execution (AQE). Here are more details: https://spark.apache.org/docs/latest/sql-performance-tuning.html#adaptive-query-execution. For even bett...

  • 0 kudos
Shaimaa
by New Contributor II
  • 1616 Views
  • 1 replies
  • 0 kudos

running a query against multiple parquet files from a folder

I am runninng a query against multiple parquet files:SELECT SUM(CASE WHEN match_result.year_incorporated IS NOT NULL AND match_result.year_incorporated != '' THEN 1 ELSE 0 END) FROM parquet.`s3://folder_path/*`for some files, the field `year_incorpo...

  • 1616 Views
  • 1 replies
  • 0 kudos
Latest Reply
daniel_sahal
Databricks MVP
  • 0 kudos

@Shaimaa The column type mismatch between the files could be an issue here.For example: if in one file column 'xyz' is a type of INTEGER and in another one the same column is a type of STRING, Spark will give you a schema conversion error.Below is a ...

  • 0 kudos
Shaimaa
by New Contributor II
  • 3143 Views
  • 3 replies
  • 0 kudos

null object while running a query against parquet

I am running this query against parquet:SELECT SUM(CASE WHEN match_result.ecommerce.has_online_payments THEN 1 ELSE 0 END) FROM parquet.`s3://folder_path/*`when all the values of the object `match_result.ecommerce` are null, I get the following erro...

  • 3143 Views
  • 3 replies
  • 0 kudos
Latest Reply
Shaimaa
New Contributor II
  • 0 kudos

None of these solutions with coalesce work because it's "match_result.ecommerce" that is null not "match_result.ecommerce.has_online_payments". So it's still trying to extract a value from a null. Help me modify the query accordingly please. 

  • 0 kudos
2 More Replies
dbph
by New Contributor II
  • 2577 Views
  • 1 replies
  • 1 kudos

How to use Serverless as DBT-CLI compute?

Hi,we'd like to use serverless as the compute for DBT-CLI (of course we already used Serverless SQL before) in a DBT workflow.I configured a normal DBT-task and tried to run a dbt-run command, which i previously tested sucessfully on my local machine...

  • 2577 Views
  • 1 replies
  • 1 kudos
Latest Reply
dbph
New Contributor II
  • 1 kudos

Hi, thanks for your help!Unfortunately our problem is not the Databricks-CLI we're using on local machines. The problem is the DBT-CLI which we are trying to run on serverless compute inside a Databricks-Workflow.I already tried adding the code you p...

  • 1 kudos
Akshay_Petkar
by Valued Contributor
  • 2036 Views
  • 2 replies
  • 0 kudos

Databricks External Table (ADLS) Access with Power BI?

Can we to connect Power BI directly to an Azure Databricks external table located in ADLS using cluster credentials?

  • 2036 Views
  • 2 replies
  • 0 kudos
Latest Reply
Wojciech_BUK
Valued Contributor III
  • 0 kudos

It depends- if you want to connect Power BI to table (wither external or managed) you can do it  BUT it will be through cluster (each time you run something on Power BI desktop - it will wake up cluster or SQL endpoint and you will access data in thi...

  • 0 kudos
1 More Replies
jakubk
by Contributor
  • 5205 Views
  • 2 replies
  • 3 kudos

Resolved! unity catalog information schema columns metadata out of sync with table - cant refresh

I'm using unity catalogI've changed the schema of my table by overwriting it with a newer file df.write \ .format('delta') \ .partitionBy(partitionColumn) \ .mode("overwrite") \ .option("overwriteSchema", "true") \ ...

  • 5205 Views
  • 2 replies
  • 3 kudos
Latest Reply
daniel_sahal
Databricks MVP
  • 3 kudos

@jakubk Try runningREPAIR TABLE <table_name> SYNC METADATA 

  • 3 kudos
1 More Replies