cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

rednirusmart
by New Contributor
  • 772 Views
  • 0 replies
  • 0 kudos

Rednirus Mart is a Third-Party Pharma Manufacturer and Supplier. If you are looking For Pharma Contract manufacturers For Ayurvedic Medicine Manufactu...

Rednirus Mart is a Third-Party Pharma Manufacturer and Supplier. If you are looking For Pharma Contract manufacturers For Ayurvedic Medicine Manufacturer Company in your region. Rednirus Mart is one of the leading one and their products are manufactu...

Third Party Pharma Manufacturers
  • 772 Views
  • 0 replies
  • 0 kudos
pjp94
by Contributor
  • 2968 Views
  • 1 replies
  • 3 kudos

Use '%sql' inside a python cmd cell?

Hi so I want to essentially execute a sql query if a condition is met. So one of my cells in my python notebook is a sql query (%sql followed by the query). Is there any way to put that in an 'IF' statement ie if an environment variable = some value,...

  • 2968 Views
  • 1 replies
  • 3 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 3 kudos

in python cell just use:query = "SELECT 1"spark.sql(query)

  • 3 kudos
RasmusOlesen
by New Contributor III
  • 9524 Views
  • 4 replies
  • 2 kudos

Upgrading from Spark 2.4 to 3.2: Recursive view errors when using

We get errors like this,Recursive view `x` detected (cycle: `x` -> `x`).. in our long-term working code, that has worked just fine in Spark 2.4.5 (Runtime 6.4), when we run it on a Spark 3.2 cluster (Runtime 10.0).It happens whenever we have,<x is a ...

  • 9524 Views
  • 4 replies
  • 2 kudos
Latest Reply
arkrish
New Contributor II
  • 2 kudos

This is a breaking change introduced in Spark 3.1 From Migration Guide: SQL, Datasets and DataFrame - Spark 3.1.1 Documentation (apache.org)In Spark 3.1, the temporary view will have same behaviors with the permanent view, i.e. capture and store runt...

  • 2 kudos
3 More Replies
Ryan_Chynoweth
by Esteemed Contributor
  • 1728 Views
  • 0 replies
  • 0 kudos

Azure_DAAM

Attached to this post we have added an ADLS Gen2 access recommendation to have the ideal security and governance over your data. The best practice involves leveraging Cluster ACLs, cluster configuration, and secret ACLs to handle user access over you...

  • 1728 Views
  • 0 replies
  • 0 kudos
sarvesh
by Contributor III
  • 1424 Views
  • 1 replies
  • 3 kudos

Audit Vertica tables in Spark!

I am trying to use Audit from Vertica in spark and getting correct table size from it, but the minimum size Audit function can find is bytes, but we are getting data in bits even smaller than bytes. val size = f"select audit('table_name');"

  • 1424 Views
  • 1 replies
  • 3 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 3 kudos

Rather everything will be in bytes. Spak sql have built in methods to get table size but also in bytes:spark.sql("ANALYZE TABLE df COMPUTE STATISTICS NOSCAN")spark.sql("DESCRIBE EXTENDED df ").filter(col("col_name") === "Statistics").show(false)

  • 3 kudos
Hayley
by Databricks Employee
  • 3936 Views
  • 1 replies
  • 2 kudos

What is the best way to do EDA in Databricks?

Are there example notebooks to quickstart the exploratory data analysis?

  • 3936 Views
  • 1 replies
  • 2 kudos
Latest Reply
Hayley
Databricks Employee
  • 2 kudos

A quick way to start exploratory data analysis is to use the EDA notebook that is created when you use Databricks AutoML. Then you can use the notebook generated as is, or as a starting point for modeling. You’ll need a cluster with Databricks Runtim...

  • 2 kudos
Hubert-Dudek
by Esteemed Contributor III
  • 5068 Views
  • 2 replies
  • 9 kudos

Login to ideas.databricks.com from Azure or community accounts.

When I go to ideas.databricks.com it display me screen asking about workspace (so I put there for example westeurope.azuredatabricks.net):then it redirect to login and then to... my azure workspace instead of ideas.When I want to use community (I put...

image.png image.png
  • 5068 Views
  • 2 replies
  • 9 kudos
Latest Reply
Anonymous
Not applicable
  • 9 kudos

@Hubert Dudek​ - Would you be happy to send a message to sales@databricks.com, please? We think they're the ones best suited to help you.

  • 9 kudos
1 More Replies
Logan_Data_Inc_
by New Contributor II
  • 3372 Views
  • 3 replies
  • 4 kudos

Resolved! Will the database be deleted when the cluster terminates ?

Hello,I'm working in Databricks Community Edition. So I will terminate my cluster after my work because anyways it will be terminated after 2 hours. I'm creating a database to store all my transformed data. Will the database be deleted when I termina...

  • 3372 Views
  • 3 replies
  • 4 kudos
Latest Reply
Anonymous
Not applicable
  • 4 kudos

@Hubert Dudek​ - Thanks for answering so quickly!!@Sriram Devineedi​ - If Hubert's answer solved the issue for you, would you be happy to mark his answer as best? That helps others know where to look.

  • 4 kudos
2 More Replies
Joseph_B
by Databricks Employee
  • 2400 Views
  • 1 replies
  • 0 kudos

How should I tune hyperparameters when fitting models for every item?

My dataset has an "item" column which groups the rows into many groups. (Think of these groups as items in a store.) I want to fit 1 ML model per group. Should I tune hyperparameters for each group separately? Or should I tune them for the entire...

  • 2400 Views
  • 1 replies
  • 0 kudos
Latest Reply
Joseph_B
Databricks Employee
  • 0 kudos

For the first question ("which option is better?"), you need to answer that via your understanding of the problem domain.Do you expect similar behavior across the groups (items)?If so, that's a +1 in favor of sharing hyperparameters. And vice versa....

  • 0 kudos
AryaMa
by New Contributor III
  • 3501 Views
  • 3 replies
  • 4 kudos

Sending email from databricks to google drive attachement

https://stackoverflow.com/questions/67088891/send-email-from-databricks-notebook-with-attachmenti have to send the attachment to the organisation google drive folder directly instead of email any suggestionssample email with attachement codemsg.atta...

  • 3501 Views
  • 3 replies
  • 4 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 4 kudos

Maybe just use azure logic apps or power automate (trigger as http requests with json and then all actions there)

  • 4 kudos
2 More Replies
Sri_H
by New Contributor III
  • 4441 Views
  • 4 replies
  • 4 kudos

Did not receive my certificate after passing Associate Dev

I have completed my Databricks associate developer assessment on 12/05/2021 and received a pass result. On 12/08/2021 I have received an email stating my digital badge for this assessment is available. However, I do not see this badge or my completio...

  • 4441 Views
  • 4 replies
  • 4 kudos
Latest Reply
Sri_H
New Contributor III
  • 4 kudos

Hi @Kaniz Fatma​ , I have still not received the certificate and Badge. Can you please help follow-up once regarding this?Thank You.

  • 4 kudos
3 More Replies
cconnell
by Contributor II
  • 1112 Views
  • 0 replies
  • 2 kudos

www.linkedin.com

Here is an article I wrote that puts Databricks in a historical context (why was it developed?) and provides introductory steps to help a newbie get started. Feel free to copy/link as you want.https://www.linkedin.com/pulse/databricks-introduction-ch...

  • 1112 Views
  • 0 replies
  • 2 kudos
cconnell
by Contributor II
  • 1198 Views
  • 1 replies
  • 1 kudos

www.linkedin.com

Importing JSON to Databricks (PySpark) is simple in the simple case. But of course there are wrinkles for real-world data. Here are some tips/tricks to help...https://www.linkedin.com/pulse/json-databricks-pyspark-chuck-connell/

  • 1198 Views
  • 1 replies
  • 1 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 1 kudos

Nice article

  • 1 kudos
William_Scardua
by Valued Contributor
  • 7598 Views
  • 4 replies
  • 3 kudos

Resolved! Update Databricks Runtime

Hy guys,I need to upgrade my databricks runtime (current 8.0 ) What the precautions should I take ?Thank you very much

  • 7598 Views
  • 4 replies
  • 3 kudos
Latest Reply
Steward475
New Contributor II
  • 3 kudos

If you want to know the version of Databricks runtime in Azure after creation: Go to Azure Data bricks portal => Clusters => Interactive Clusters => here you can find the run time version. For more details, refer "Azure Databricks Runtime versions".R...

  • 3 kudos
3 More Replies
test_data
by New Contributor III
  • 4782 Views
  • 1 replies
  • 1 kudos
  • 4782 Views
  • 1 replies
  • 1 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 1 kudos

Please use Repos and in admin settings please enable "Files in Repo" than you will be able to import class in notebook:from repo_folder.subfolders.file import your_class

  • 1 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels