cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Hoviedo
by New Contributor III
  • 1060 Views
  • 0 replies
  • 0 kudos

load data from sql server : python works well but spark does not

Hi, i have a problem with my on-premise sql conection from databricks.i have my python code wich use pymssql and it is working well but my spark code doest not and i am using the same credential for both, my spark code is """# Read data from SQL Serv...

  • 1060 Views
  • 0 replies
  • 0 kudos
pjp94
by Contributor
  • 8959 Views
  • 0 replies
  • 0 kudos

Run threadpool on multiple nodes

I've ran a dual multiprocessing and multithreading solution in python before using the multiprocessing and concurrent futures python modules. However, since the multiprocessing module only runs on the driver node, I have to instead use sc.parallelize...

Data Engineering
parallelization
threading
  • 8959 Views
  • 0 replies
  • 0 kudos
Torlynet
by New Contributor III
  • 7163 Views
  • 2 replies
  • 3 kudos

Resolved! Azure Databricks workspace AAD authentication issue

I am trying to log in to my workspace, but it takes a very long time to evaluate. Sometimes, it simply fails to do so, and I am prompted with the below message. {"error_code":"TEMPORARILY_UNAVAILABLE","message":"Authentication is temporarily unavaila...

  • 7163 Views
  • 2 replies
  • 3 kudos
Latest Reply
Torlynet
New Contributor III
  • 3 kudos

Hi Menotron.Eventually, I realized it myself. Thank you for your comment

  • 3 kudos
1 More Replies
shreyassharmabh
by New Contributor II
  • 4579 Views
  • 2 replies
  • 1 kudos

How to check programmatically job cluster is unity catalog enabled or not in databricks

Is there any way to check job cluster is unity catalog enabled or not in databricks using python.I tried with jobs api https://{host_name}/api/2.0/jobs/get?job_id={job_id}, but I didn't that cluster is unity catalog enabled or not.Could anyone sugges...

  • 4579 Views
  • 2 replies
  • 1 kudos
Latest Reply
KarenZak
New Contributor II
  • 1 kudos

To check if a job cluster is Unity catalog enabled in Databricks programmatically using Python, you can use the Databricks REST API. Here's an example of how you can do it:Import the required modules:import requestsSet up the necessary variables:host...

  • 1 kudos
1 More Replies
chorongs
by New Contributor III
  • 4691 Views
  • 2 replies
  • 1 kudos

Resolved! Sequential vs concurrency optimization questions from query!

Preparing for databricks eligibility!Is the content below correct?"If the queries are running sequentially then scale up (increase the size of the cluster from 2x small to 4x large)If the queries are running concurrently or with many users then scale...

  • 4691 Views
  • 2 replies
  • 1 kudos
Latest Reply
Anonymous
Not applicable
  • 1 kudos

Scaling in Databricks involves two aspects: vertical scaling (scale up) and horizontal scaling (scale out). Vertical Scaling (Scale Up): If your queries are running sequentially, meaning one query at a time, and you want to improve performance for a...

  • 1 kudos
1 More Replies
kellybe
by New Contributor II
  • 6210 Views
  • 6 replies
  • 0 kudos

Databricks SQL format_string in LOCATION

Hi,I'm trying to assign a location to a new database in Databricks SQL. Normally I'd do this in Python since we specify storage account names from secret scopes, however I'm attempting to do all of this from a SQL warehouse. When doing this I seem to...

  • 6210 Views
  • 6 replies
  • 0 kudos
Latest Reply
pcbzmani
New Contributor II
  • 0 kudos

Hello @kellybe ,CREATE DATABASE IF NOT EXISTS new_database LOCATION format_string('abfss://container-name@%s.dfs.core.windows.net/', select SECRET('secret-scope', 'storage-account-name')); Add Select before secert 

  • 0 kudos
5 More Replies
kurt
by New Contributor
  • 1090 Views
  • 0 replies
  • 0 kudos

DLT & Publishing to Feature Store

Hi,Is there an example of incorporating Databricks Feature Store into DLT pipelines?  Is this possible natively via a Python notebook part of the pipeline (FYI - docs say needs ML Runtime?).  If not completely DLT-able, what is the best current way t...

  • 1090 Views
  • 0 replies
  • 0 kudos
Mbinyala
by New Contributor II
  • 30666 Views
  • 2 replies
  • 1 kudos

Connecting confluent to databricks.

Hi!!Can someone tell me how to connect the confluent cloud to Databricks? I am new to this so please elaborate on your answer.

  • 30666 Views
  • 2 replies
  • 1 kudos
Latest Reply
VaibB
Contributor
  • 1 kudos

You might want to watch this as well https://www.confluent.io/resources/online-talk/innovate-faster-and-easier-with-confluent-and-databricks-on-azure/?utm_medium=sem&utm_source=google&utm_campaign=ch.sem_br.nonbrand_tp.prs_tgt.dsa_mt.dsa_rgn.india_ln...

  • 1 kudos
1 More Replies
Gim
by Contributor
  • 3159 Views
  • 1 replies
  • 3 kudos

Columns with DEFAULT missing error during INSERT

I am really confused about the DEFAULT capability of Databricks SQL. I looked at the documentation for the minimum required DBR to get the capability yet we still need to enable it as a table property? I updated my cluster's DBR from 12.2 to 13.1.Any...

Gim_0-1688465259125.png
  • 3159 Views
  • 1 replies
  • 3 kudos
Latest Reply
BriceBuso
Contributor II
  • 3 kudos

Hello @Gim, Got the same problem. Tried with the instruction "GENERATED ALWAYS AS (CAST(CURRENT_DATE() AS DATE))" but code is returning "Error in SQL statement: DeltaAnalysisException: current_date() cannot be used in a generated column" If you find ...

  • 3 kudos
erigaud
by Honored Contributor
  • 4667 Views
  • 2 replies
  • 1 kudos

Incrementally load SQL Server table

I am accessing an on premise SQL Server table. The table is relatively small (10 000 rows), and I access it usingspark.read.jdbc(url=jdbcUrl, table = query)Every day there are new records in the on prem table that I would like to append in my bronze ...

  • 4667 Views
  • 2 replies
  • 1 kudos
Latest Reply
erigaud
Honored Contributor
  • 1 kudos

As I said, there is no unique identifier in the table that would allow me to do any sort of Join between my source table and my bronze table. 

  • 1 kudos
1 More Replies
JLL
by New Contributor II
  • 1301 Views
  • 1 replies
  • 2 kudos

Shorten query run time

Challenges in query long run time; what are the recommended steps to improve performance 

  • 1301 Views
  • 1 replies
  • 2 kudos
Latest Reply
erigaud
Honored Contributor
  • 2 kudos

The question needs more precision : is it the cluster startup that takes a while ? If yes, try serverless warehousesAre there many queries running in parallel and that is where you see a slow down ? Each cluster can only run 10 queries in parallel, s...

  • 2 kudos
christo_M
by New Contributor
  • 2854 Views
  • 4 replies
  • 0 kudos

Cost Optimization

How can I optimize the cost on our Databricks platform ? Despite some optimization actions I've taken so far it's still difficult to lower the cost. I tried different technics like Vacuum , or shutting down a cluster running after 30 mins but still d...

  • 2854 Views
  • 4 replies
  • 0 kudos
Latest Reply
erigaud
Honored Contributor
  • 0 kudos

Make sure you're using a cluster that is the right size for your workload. You can greatly reduce the costs by using smaller clusters.

  • 0 kudos
3 More Replies
EcuaCrisCar
by New Contributor III
  • 1367 Views
  • 0 replies
  • 0 kudos

Sending a personalized message to email.

Greetings community, I am new to using databricks and for some time I have tried some scripts in notebook. I would like your help on a task: Carry out a personalized mailing where, First, a query of the number of records in the test table is performe...

Data Engineering
SENDEMAIL SQL
  • 1367 Views
  • 0 replies
  • 0 kudos

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels