cancel
Showing results for 
Search instead for 
Did you mean: 
Get Started Discussions
Start your journey with Databricks by joining discussions on getting started guides, tutorials, and introductory topics. Connect with beginners and experts alike to kickstart your Databricks experience.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Phani1
by Valued Contributor II
  • 5926 Views
  • 5 replies
  • 1 kudos

Azure Synapse vs Databricks

 Hi team,Could you kindly provide your perspective on the cost and performance comparison between Azure Synapse and Databricks SQL Warehouse/serverless, as well as their respective use cases? Thank you.

  • 5926 Views
  • 5 replies
  • 1 kudos
Latest Reply
Witold
Honored Contributor
  • 1 kudos

@Suncat There hasn't been any major changes for than a year: https://learn.microsoft.com/en-us/azure/synapse-analytics/spark/apache-spark-version-support E.g. I don't believe we will see support for Spark 3.5 at all. At least, apparently it's support...

  • 1 kudos
4 More Replies
n1399
by New Contributor II
  • 705 Views
  • 1 replies
  • 0 kudos

On Demand Pool Configuration & Policy definition

I'm using Job cluster and created compute policies for library management and now I'm trying to use pools in databricks. I'm getting error like this : Cluster validation error: Validation failed for azure_attributes.spot_bid_max_price from pool, the ...

  • 705 Views
  • 1 replies
  • 0 kudos
Latest Reply
lingareddy_Alva
Honored Contributor III
  • 0 kudos

@n1399 The error "Validation failed for azure_attributes.spot_bid_max_price from pool, the value must be present" suggests that the spot bid max price is required, but it’s either missing or not correctly inherited from the compute policy when using ...

  • 0 kudos
Rachana2
by New Contributor II
  • 724 Views
  • 3 replies
  • 0 kudos

Databricks lineage

Hello,I am trying to get the table lineage i.e upstreams and downstreams of all tables in unity catalog into my local database using API calls. I need my db to be up to date, if the lineage is updated in one of the in databricks, i have to update sam...

  • 724 Views
  • 3 replies
  • 0 kudos
Latest Reply
SantoshJoshi
New Contributor III
  • 0 kudos

Hi @Rachana2,As @Alberto_Umana has mentioned I'd check table_lineage / column_lineage tables, as maintaining a lineage through a bespoke pipeline/tooling may not be a right approach.Can you please explain your use case which explains why you don't wa...

  • 0 kudos
2 More Replies
joseroca99
by New Contributor II
  • 2325 Views
  • 6 replies
  • 0 kudos

Resolved! File found with %fs ls but not with spark.read

Code: wikipediaDF = (spark.read  .option("HEADER", True)  .option("inferSchema", True)  .csv("/databricks-datasets/wikipedia-datasets/data-001/pageviews/raw/pageviews_by_second.tsv"))display(bostonDF) Error: Failed to store the result. Try rerunning ...

  • 2325 Views
  • 6 replies
  • 0 kudos
Latest Reply
xx123
New Contributor III
  • 0 kudos

I have the exact same issue. Seems like limiting the the display() method works as a temporary solution, but I wonder if there's any long term one. The idea would be to have the possibility of displaying larger datasets within a notebook. How to achi...

  • 0 kudos
5 More Replies
j_h_robinson
by New Contributor II
  • 795 Views
  • 1 replies
  • 1 kudos

Resolved! Spreadsheet-Like UI for Databricks

We are currently entering data into Excel and then uploading it into Databricks.  Is there a built-in spreadsheet-like UI within Databricks that can update data directly in Databricks? 

  • 795 Views
  • 1 replies
  • 1 kudos
Latest Reply
Advika_
Databricks Employee
  • 1 kudos

Hello, @j_h_robinson! Databricks doesn’t have a built-in spreadsheet-like UI for direct data entry or editing. Are you manually uploading the Excel files or using an ODBC driver setup? If you’re doing it manually, you might find this helpful: Connect...

  • 1 kudos
h_h_ak
by Contributor
  • 4407 Views
  • 5 replies
  • 2 kudos

Resolved! Understanding Autoscaling in Databricks: Under What Conditions Does Spark Add a New Worker Node?

I’m currently working with Databricks autoscaling configurations and trying to better understand how Spark decides when to spin up additional worker nodes. My cluster has a minimum of one worker and can scale up to five. I know that tasks are assigne...

  • 4407 Views
  • 5 replies
  • 2 kudos
Latest Reply
filipniziol
Esteemed Contributor
  • 2 kudos

Hi @h_h_ak ,Short Answer:Autoscaling primarily depends on the number of pending tasks.Workspaces on the Premium plan use optimized autoscaling, while those on the Standard plan use standard autoscaling.Long Answer:Databricks autoscaling responds main...

  • 2 kudos
4 More Replies
Yuki
by Contributor
  • 2240 Views
  • 4 replies
  • 5 kudos

Resolved! Why is the ipynb format recommended?

In this document, https://docs.databricks.com/aws/en/notebooks/notebook-format,Jupyter (.ipynb) format is recommended.> Select File from the workspace menu, select Notebook format, and choose the format you want. You can choose either Jupyter (.ipynb...

  • 2240 Views
  • 4 replies
  • 5 kudos
Latest Reply
Nivethan_Venkat
Contributor III
  • 5 kudos

Hi @Yuki,One other risk that we foresee / encountered recently is how the notebooks will look in your pull requests of external repos (Azure Devops or GitHub). It will be very hard for a pull request reviewer to understand on the code / notebook read...

  • 5 kudos
3 More Replies
Adil
by New Contributor
  • 3032 Views
  • 1 replies
  • 0 kudos

Find value in any column in a table

Hi,I'm not sure if this is a possible scenario, but is there, by any chance a way to query all the columns of a table for searching a value? Explanation: I want to search for a specific value in all the columns of a databricks table. I don't know whi...

  • 3032 Views
  • 1 replies
  • 0 kudos
Latest Reply
sashiDatabricks
New Contributor II
  • 0 kudos

I also have this same requirement now and cant find the solution for this yet. Any help would be good. thanks

  • 0 kudos
Phani1
by Valued Contributor II
  • 1381 Views
  • 2 replies
  • 0 kudos

Solace to Azure Data Lake Storage

 Hi Team,What is the most effective method for performing data ingestion from Solace to Azure Data Lake Storage (ADLS) utilizing an Azure Databricks notebook? Any recommendations would be greatly appreciated.Regards,Phani 

  • 1381 Views
  • 2 replies
  • 0 kudos
Latest Reply
SravanThotakura
New Contributor II
  • 0 kudos

Here is the sample script to invoke the connectorval struct_stream = spark.readStream.format("solace").option("host", "").option("vpn", "").option("username", "").option("password", "").option("queue", "").option("connectRetries", 3).option("reconnec...

  • 0 kudos
1 More Replies
EricMa
by New Contributor III
  • 8263 Views
  • 20 replies
  • 4 kudos

Mounting Data IOException

Hello,I am currently taking a course from Coursera for data science using SQL. For one of our assignments we need to mount some data by running a script that has been provided to us by the class. When I run the script I receive the following error. I...

IOException.jpg IOException_Classroom-Setup.jpg
  • 8263 Views
  • 20 replies
  • 4 kudos
Latest Reply
raghdafaris
New Contributor II
  • 4 kudos

Hello all, we came up with a solution: to download the data directly instead of mounting it. The community version is limited, and we don't have access to S3 unless we create our own aws account, load the data there, and then mount our account on dat...

  • 4 kudos
19 More Replies
Shravanshibu
by New Contributor III
  • 3990 Views
  • 6 replies
  • 3 kudos

Unable to install a wheel file which is in my volume to a serverless cluster

I am trying to install a wheel file which is in my volume to a serverless cluster, getting the below error@ken@Retired_mod Note: you may need to restart the kernel using %restart_python or dbutils.library.restartPython() to use updated packages. WARN...

  • 3990 Views
  • 6 replies
  • 3 kudos
Latest Reply
tom-ph
New Contributor II
  • 3 kudos

Same issue here. Any solution?

  • 3 kudos
5 More Replies
Sangamswadik
by New Contributor III
  • 1234 Views
  • 4 replies
  • 1 kudos

Resolved! Agents and Inference table errors

Hi, I'm trying to deploy a rag model from GCP databricks. I've added an external gpt4o endpoint and enabled inference table in settings. But when Im trying to deploy agents I'm still getting the inference table not enabled error. (I've registered the...

Sangamswadik_0-1740977753362.png Sangamswadik_1-1740978040054.png
  • 1234 Views
  • 4 replies
  • 1 kudos
Latest Reply
MariuszK
Valued Contributor III
  • 1 kudos

The Model Serving is supported in your region so it can be another problem or limitation.

  • 1 kudos
3 More Replies
Nik_Vanderhoof
by Contributor
  • 1014 Views
  • 2 replies
  • 0 kudos

Resolved! DatabricksWorkflowTaskGroup

Hello,I recently learned about the DatabricksWorkflowTaskGroup operator for Airflow that allows one to run multiple Notebook tasks on a shared job compute cluster from Airflow.Is a similar feature possible to run multiple non-Notebook tasks from Airf...

Get Started Discussions
Airflow
workflow
  • 1014 Views
  • 2 replies
  • 0 kudos
Latest Reply
Nik_Vanderhoof
Contributor
  • 0 kudos

Thank you!

  • 0 kudos
1 More Replies
anil_reddaboina
by New Contributor II
  • 757 Views
  • 2 replies
  • 0 kudos

Databricks tasks are not skipping if running tasks using Airflow DatabricksworkflowTaskgroup

Currently we are facing a challenge with below use case:The Airflow DAG has 4 tasks (Task1, Task2, Task3 and Task4) and The dependency is like thisTask 1>> Task2 >> Task3 >> Task4 (All tasks are spark-jar task typesIn Airflow DAG for Task2, there is ...

  • 757 Views
  • 2 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hi @anil_reddaboina, Databricks allows you to add control flow logic to tasks based on the success, failure, or completion of their dependencies. This can be achieved using the "Run if" dependencies fiel: https://docs.databricks.com/aws/en/jobs/run-i...

  • 0 kudos
1 More Replies
shkelzeen
by New Contributor II
  • 2459 Views
  • 3 replies
  • 1 kudos

Databricks JDBC driver multi query in one request.

Can I run multi query in one command using databricks JDBC driver and would databricks execute one query faster then running multi queries in one script?  

  • 2459 Views
  • 3 replies
  • 1 kudos
Latest Reply
NandiniN
Databricks Employee
  • 1 kudos

Yes, you can run multiple queries in one command using the Databricks JDBC driver.The results will be displayed in separate tables. When you run the multiple queries, they are all still individual queries. Running multiple queries in a script will no...

  • 1 kudos
2 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels