cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

shweta_m
by New Contributor III
  • 3134 Views
  • 3 replies
  • 4 kudos

Resolved! Best Practices for Managing ACLs on Jobs and Job Clusters in Databricks

 Hi all,I’m setting up access control for Databricks jobs and have two questions:Ephemeral Job Clusters: Since job clusters are created at runtime, is it best practice to set ACLs on the job itself? The /api/2.0/permissions/clusters/{cluster_id} endp...

  • 3134 Views
  • 3 replies
  • 4 kudos
Latest Reply
shweta_m
New Contributor III
  • 4 kudos

Thanks! @juan_maedo @saurabh18cs 

  • 4 kudos
2 More Replies
susmitsircar
by New Contributor III
  • 611 Views
  • 9 replies
  • 0 kudos

Proposal: Switch to Zstd Compression for Parquet to Reduce S3 Costs

We are thinking to change the Spark configuration for Parquet files to use zstd compression.Configuration: spark.sql.parquet.compression.codec = zstdThis will only affect new data written by our Spark jobs. All existing data will remain compressed wi...

  • 611 Views
  • 9 replies
  • 0 kudos
Latest Reply
susmitsircar
New Contributor III
  • 0 kudos

Yes my believe is it should support 7.3 LTS as well, we will prove it with thorough testingThanks for the discussion. Cheers

  • 0 kudos
8 More Replies
parthesh24
by New Contributor II
  • 585 Views
  • 3 replies
  • 3 kudos

Resolved! from pyspark.ml.stat import KolmogorovSmirnovTest is not working on Serverless compute.

 Hi everyone,I am trying to run a Kolmogorov–Smirnov (KS) test on a Spark DataFrame column in Databricks using the built-in pyspark.ml.stat.KolmogorovSmirnovTest. The goal is to apply the KS test directly on Spark DataFrame data without converting it...

  • 585 Views
  • 3 replies
  • 3 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 3 kudos

 Hi @parthesh24 ,It looks more like KolmogorovSmirnovTest module under the hood is trying to access SparkContext which is not supported in serverless.  You can check it yourself by trying to use sparkContext in serverless

  • 3 kudos
2 More Replies
pinikrisher
by New Contributor II
  • 219 Views
  • 1 replies
  • 1 kudos

Dashboard tagging

How can i tag dashbaord? i do not see any place to add tagging to it?

  • 219 Views
  • 1 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @pinikrisher ,Unforutnately, you can't. Tagging is currently supported on catalogs, schemas, tables, table columns, volumes, views, registered models, and model version

  • 1 kudos
s3anil
by New Contributor II
  • 731 Views
  • 6 replies
  • 2 kudos

databricks dashboard deployment error

Hi, i am trying to deploy a dashboard using a bundle and github action. but i am getting an error on CI even though the dashboard is deployed. im using the latest version of CLI from https://raw.githubusercontent.com/databricks/setup-cli/main/install...

  • 731 Views
  • 6 replies
  • 2 kudos
Latest Reply
s3anil
New Contributor II
  • 2 kudos

@szymon_dybczak ,@nayan_wylde , I checked the permissions and the SP has 'can manage' access on the folder.

  • 2 kudos
5 More Replies
heli123
by New Contributor III
  • 293 Views
  • 2 replies
  • 2 kudos

Resolved! Lakehouse monitoring dashboard shows no data

Hello, I am replicating the demo for Lakehouse monitoring found here: https://notebooks.databricks.com/demos/lakehouse-monitoring/index.htmlFor some reason, my dashboards show empty, i.e., they say 'no data' - like nothing fits the criteria from the ...

Data Engineering
lakehouse monitoring
ml monitoring
  • 293 Views
  • 2 replies
  • 2 kudos
Latest Reply
Khaja_Zaffer
Contributor III
  • 2 kudos

Hello @heli123 Can you share the image again? looks like didnt upload well. 

  • 2 kudos
1 More Replies
ashfire
by New Contributor II
  • 400 Views
  • 3 replies
  • 3 kudos

Databricks model serving endpoint returns 403 Unauthorized access to workspace when using service

I deployed a simple Iris model in Databricks Model Serving and exposed it as an endpoint. I’m trying to query the endpoint using a service principal. I can successfully fetch the access token with the following databricks_token() function:def databri...

  • 400 Views
  • 3 replies
  • 3 kudos
Latest Reply
ashfire
New Contributor II
  • 3 kudos

Hi @szymon_dybczak, Thanks for your comment.One of the admins in this workspace tried using the token generated via client id and secret, and were able to successfully get a response from the serving endpoint using this same above mentioned code.Coul...

  • 3 kudos
2 More Replies
aranjan99
by Contributor
  • 472 Views
  • 2 replies
  • 0 kudos

How to switch serverless dlt pipeline to cost optimized mode from performance optimized

We have a few serverless dlt pipelines that we want to optimize for cost as we are ok with an increased latency. Where can I change the pipeline to run on cost optimized mode. I dont see this option in UI or API

  • 472 Views
  • 2 replies
  • 0 kudos
Latest Reply
wawefog260
New Contributor II
  • 0 kudos

Hello!To enable cost-optimized mode for your serverless DLT pipeline, switch it to Triggered mode and edit the schedule trigger—there you’ll find the option to disable “Performance optimized.” This setting isn’t visible in the main UI or API unless t...

  • 0 kudos
1 More Replies
elgeo
by Valued Contributor II
  • 39714 Views
  • 13 replies
  • 6 kudos

SQL Stored Procedure in Databricks

Hello. Is there an equivalent of SQL stored procedure in Databricks? Please note that I need a procedure that allows DML statements and not only Select statement as a function provides.Thank you in advance

  • 39714 Views
  • 13 replies
  • 6 kudos
Latest Reply
SanthoshU
New Contributor II
  • 6 kudos

how to connect the stored procedures to power bi report builder, seems like it is not working 

  • 6 kudos
12 More Replies
MauGomes
by New Contributor
  • 198 Views
  • 1 replies
  • 2 kudos

Resolved! Access to Databricks partner academy

Hi Team,My company is a Databricks Partner. But I can't get registered for Databricks Partner Academy.I have followed the following steps for Partner Academy RegistrationOpen https://partner-academy.databricks.com/learn in your web browser.Click Logi...

  • 198 Views
  • 1 replies
  • 2 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 2 kudos

Hi @MauGomes ,Don't worry, you already did the best thing you could. Check below thread with exact same issue . The user submitted a ticket and it was resolved by service desk. So, just wait patiently for reply Solved: authorized to access https://pa...

  • 2 kudos
zensardigital
by New Contributor II
  • 330 Views
  • 3 replies
  • 0 kudos

Convert a Managed Table to Streaming Table

HiI have applied transformations on a set of streaming tables and saved it as a managed table....How can i change the Managed table to a Streaming table with minimal changesRegardsZD

  • 330 Views
  • 3 replies
  • 0 kudos
Latest Reply
zensardigital
New Contributor II
  • 0 kudos

I am just writing the dataframe to delta table.....Are you suggesting me to first define a STREAMING TABLE (using the DLT definition) and then save the dataframe into that table? 

  • 0 kudos
2 More Replies
Naga05
by New Contributor III
  • 435 Views
  • 4 replies
  • 2 kudos

Databricks app with parameters from databricks asset bundle

HelloooI tried out setting up a Databricks App using asset bundle, where i was able to successfully parameterize the sql warehouse id which was specified on specific targets. However i was unable to get values of other variables from the targets, the...

  • 435 Views
  • 4 replies
  • 2 kudos
Latest Reply
Naga05
New Contributor III
  • 2 kudos

Found that this is an implementation in progress on the Databricks CLI. https://github.com/databricks/cli/issues/3679

  • 2 kudos
3 More Replies
smoortema
by Contributor
  • 372 Views
  • 2 replies
  • 3 kudos

Resolved! handling both Pyspark and Python exceptions

In a Python notebook, I am using error handling according to the official documentation.  try:[some data transformation steps]except PySparkException as ex:[logging steps to log the error condition and error message in a table]However, this catches o...

  • 372 Views
  • 2 replies
  • 3 kudos
Latest Reply
mark_ott
Databricks Employee
  • 3 kudos

To handle both PySpark exceptions and general Python exceptions without double-logging or overwriting error details, the recommended approach is to use multiple except clauses that distinguish the exception type clearly. In Python, exception handlers...

  • 3 kudos
1 More Replies
tom_1
by New Contributor III
  • 1318 Views
  • 5 replies
  • 1 kudos

Resolved! BUG in Job Task of Type DBT

Hi, just wanted to let the Databricks Team know, that there is a bug in the task ui.Currently it is not possible to save a task of "Type: dbt" if the "SQL Warehouse" is set to "None (Manual)".Some weeks ago this was possible, also the "Profiles Direc...

tom_1_0-1741870684542.png tom_1_1-1741870779606.png
  • 1318 Views
  • 5 replies
  • 1 kudos
Latest Reply
Aishu95
New Contributor II
  • 1 kudos

I am facing this bug still. I don't want to select any SQL warehouse, what do I do? and from where can I pass the profiles directory

  • 1 kudos
4 More Replies
Navi991100
by New Contributor II
  • 248 Views
  • 3 replies
  • 1 kudos

Resolved! I recently made new account on databricks under Free edition

It by default made SQL warehouse compute, but I want all-purpose compute, as I want test and learn capabilities of PySpark and Databricks.I can't connect with the serverless compute in the notebook; it gives a mean  error as follows: "An error occurr...

Navi991100_0-1759078594989.png
  • 248 Views
  • 3 replies
  • 1 kudos
Latest Reply
belforte
New Contributor II
  • 1 kudos

In the free Databricks edition, to use PySpark you need to create and start a cluster, since the SQL Warehouse is only for SQL queries; go to Compute > Create Cluster, set up a free cluster, click Start, and then attach your notebook to it this will ...

  • 1 kudos
2 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels