cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

SuMiT1
by New Contributor III
  • 13 Views
  • 1 replies
  • 0 kudos

Databricks to snowflake data load

Hi Team, I’m trying to load data from Databricks into Snowflake using the Snowflake Spark connector. I’m using a generic username and password, but I’m unable to log in using these credentials directly. In the Snowflake UI, I can only log in through ...

  • 13 Views
  • 1 replies
  • 0 kudos
Latest Reply
nayan_wylde
Honored Contributor III
  • 0 kudos

@SuMiT1  The recommended method to connect to snowflake from databricks is OAuth with Client Credentials Flow.This method uses a registered Azure AD application to obtain an OAuth token without user interaction.Steps:Register an app in Azure AD and c...

  • 0 kudos
ayush667787878
by New Contributor
  • 3069 Views
  • 1 replies
  • 0 kudos

not able to install library in normal site while in community version it working please help

I am not able to install library in normal version while in community editioin i am able to add libray using compute how to install in normal databricks same as community edition. â€ƒ 

SCR-20250211-qxiz.png ayush667787878_0-1739282136368.png
  • 3069 Views
  • 1 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

Greetings @ayush667787878 ,  The difference comes down to the type of compute you’re using. In your standard Databricks workspace, you’re running on Serverless — which doesn’t support installing additional libraries on the cluster. In Community Editi...

  • 0 kudos
pooja_bhumandla
by New Contributor II
  • 22 Views
  • 1 replies
  • 0 kudos

When to Use and when Not to Use Liquid Clustering?

 Hi everyone,I’m looking for some practical guidance and experiences around when to choose Liquid Clustering versus sticking with traditional partitioning + Z-ordering.From what I’ve gathered so far:For small tables (<10TB), Liquid Clustering gives s...

  • 22 Views
  • 1 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

Greeting @pooja_bhumandla ,    Thanks for laying out your current understanding — here’s practical guidance, trade-offs, and field rules-of-thumb for choosing between Liquid Clustering and traditional partitioning + Z-ordering, along with gotchas to ...

  • 0 kudos
StephanieAlba
by Databricks Employee
  • 3083 Views
  • 2 replies
  • 0 kudos

Is it possible to turn off the redaction of secrets? Is there a better way to solve this?

As part of our Azure Data Factory pipeline, we utilize Databricks to run some scripts that identify which files we need to load from a certain source. This list of files is then passed back into Azure Data Factory utilizing the Exit status from the n...

  • 3083 Views
  • 2 replies
  • 0 kudos
Latest Reply
joanafloresc
New Contributor II
  • 0 kudos

Hello, as of today, is it still not possible to unredact secret names?

  • 0 kudos
1 More Replies
osamR
by New Contributor
  • 26 Views
  • 1 replies
  • 2 kudos

Cannot Login to Databricks Customer Academy (Data Engineering courses)

When I am login to the course it shows me this error.You are not authorized to access https://customer-academy.databricks.com. Please select a platform you can access from the following list.

  • 26 Views
  • 1 replies
  • 2 kudos
Latest Reply
Advika
Databricks Employee
  • 2 kudos

Hello @osamR! If you’re encountering this error while trying to log in to Customer Academy, it may indicate that your account is registered under the Partner Academy. You can try accessing your courses through the Partner Academy.However, if your org...

  • 2 kudos
Adam_Borlase
by New Contributor III
  • 25 Views
  • 0 replies
  • 0 kudos

Error trying to edit Job Cluster via Databricks CLI

Good Day all,After having issues with Cloud resources allocated to Lakeflow jobs and Gateways I am trying to apply a policy to the cluster that is allocated to the Job. I am very new to a lot of the databricks platform and the administration so all h...

  • 25 Views
  • 0 replies
  • 0 kudos
GANAPATI_HEGDE
by New Contributor III
  • 53 Views
  • 8 replies
  • 0 kudos

Unable to run sql alert task using databricks job using service principal

I am trying to run a SQL alert task in data bricks job, Service principal is the run_as as set to the job, and task fails with below error message.  I also checked if SPN can be granted permission on SQL alert and SQL query, turns out only user id or...

GANAPATI_HEGDE_0-1761623911949.jpeg
  • 53 Views
  • 8 replies
  • 0 kudos
Latest Reply
GANAPATI_HEGDE
New Contributor III
  • 0 kudos

unfortunately, only these options are available in my UI, is it the new UI?

  • 0 kudos
7 More Replies
crami
by New Contributor
  • 34 Views
  • 1 replies
  • 0 kudos

Declative Pipeline: Can pipeline or job be deployed run_as using asset bundle

Hi, I have very interesting scenario. I am trying to use Declarative pipelines for first time. The platform team has made workspace artefacts as devops based deployment [infra as code], meaning, I cannot create compute. I have to create compute with ...

  • 34 Views
  • 1 replies
  • 0 kudos
Latest Reply
donna567taylor
  • 0 kudos

@crami wrote:Hi, I have very interesting scenario. I am trying to use Declarative pipelines for first time. The platform team has made workspace artefacts as devops based deployment [infra as code], meaning, I cannot create compute. I have to create ...

  • 0 kudos
Akshay_Petkar
by Valued Contributor
  • 12 Views
  • 0 replies
  • 0 kudos

%run notebook fails in Job mode with Py4JJavaError (None.get), but works in interactive notebook

 Hi everyone,I’m facing an issue when executing a Databricks job where my notebook uses %run to include other notebooks. I have a final notebook added as a task in a job, and inside that notebook I use %run to call another notebook that contains all ...

  • 12 Views
  • 0 replies
  • 0 kudos
Dimitry
by Contributor III
  • 22 Views
  • 2 replies
  • 0 kudos

databricks notebook parameter works in interactive mode but not in the job

Hi guys I've added a parameter "files_mask " to a notebook, with a default value.The job running this notebook broke with error: com.databricks.dbutils_v1.InputWidgetNotDefined: No input widget named files_mask is definedCode: mask = dbutils.widgets....

  • 22 Views
  • 2 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @Dimitry ,Do you use python or scala in your notebook?

  • 0 kudos
1 More Replies
Alessandro
by New Contributor II
  • 1656 Views
  • 2 replies
  • 0 kudos

Resolved! Update jobs parameter, when running, from API

Hi, When a Job is running, I would like to change the parameters with an API call.I know that I can set parameters value from API when I start a job from API, or that I can update the default value if the job isn't running, but I didn't find an API c...

  • 1656 Views
  • 2 replies
  • 0 kudos
Latest Reply
XueChunmei
New Contributor II
  • 0 kudos

Hi, Alessandro, I am trying to set job parameters value from API when I start a job from API call within python notebook, however, it has never succeeded, the job can be triggered, but always with job parameters' default values instead values from AP...

  • 0 kudos
1 More Replies
DM0341
by Visitor
  • 41 Views
  • 1 replies
  • 0 kudos

SQL Stored Procedures - Notebook to always run the CREATE query

I have a stored procedure that is saved as a query file. I can run it and the proc is created. However I want to take this one step further. I want my notebook to run the query file called sp_Remit.sql so if there is any changes to the proc between t...

  • 41 Views
  • 1 replies
  • 0 kudos
Latest Reply
mynameiskevin
  • 0 kudos

Something like this?import os query_name = "test_query.sql" query_path = os.path.abspath(query_name) # Read query contents with open(query_path, "r") as f: query_str = f.read() # Run it spark.sql(query_str)You can read the script from the sql ...

  • 0 kudos
fundat
by New Contributor II
  • 42 Views
  • 1 replies
  • 1 kudos

Course - Introduction to Apache Spark

Hi,In the course Introduction to Apache Spark; according to Apache Spark Runtime Architecture; Page 6 of 15. It says that :The cluster manager allocates resources and assigns tasks......Workers perform tasks assigned by the driverCan you help me plea...

fundat_3-1761596488970.png
  • 42 Views
  • 1 replies
  • 1 kudos
Latest Reply
BS_THE_ANALYST
Esteemed Contributor II
  • 1 kudos

Hi @fundat Perhaps the picture is useful here:Give this blog a read, I think this will answer some of your questions: https://medium.com/@knoldus/understanding-the-working-of-spark-driver-and-executor-4fec0e669399 .All the best,BS

  • 1 kudos
dhruvs2
by Visitor
  • 38 Views
  • 1 replies
  • 2 kudos

How to trigger a Databricks job only after multiple other jobs have completed

We have a use case where Job C should start only after both Job A and Job B have successfully completed.In Airflow, we achieve this using an ExternalTaskSensor to set dependencies across different DAGs.Is there a way to configure something similar in...

  • 38 Views
  • 1 replies
  • 2 kudos
Latest Reply
BS_THE_ANALYST
Esteemed Contributor II
  • 2 kudos

Hey @dhruvs2 You could use Lakeflow Jobs for this. You can add a job as a task:Then you can just follow the docs from here: https://docs.databricks.com/aws/en/jobs/ there's loads of great sections / tutorials.To answer your specific question:When con...

  • 2 kudos
Sakthi0311
by Visitor
  • 53 Views
  • 2 replies
  • 0 kudos

How to enable Liquid Clustering on an existing Delta Live Table (DLT) and syntax for enabling it

 Hi all,I’m working with Delta Live Tables (DLT) and want to enable Liquid Clustering on an existing DLT table that was already created without it.Could someone please clarify:How can I enable Liquid Clustering on an existing DLT table (without recre...

  • 53 Views
  • 2 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @Sakthi0311 ,For SQL language you can enable LC for materialized views and streaming tables. So the syntax looks following:If you want to use automatic clustering then use CLUSTER BY AUTO. 

  • 0 kudos
1 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels