cancel
Showing results for 
Search instead for 
Did you mean: 
Administration & Architecture
Explore discussions on Databricks administration, deployment strategies, and architectural best practices. Connect with administrators and architects to optimize your Databricks environment for performance, scalability, and security.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

martkev
by New Contributor III
  • 286 Views
  • 6 replies
  • 0 kudos

Skepticism about U2M OAuth: Does Snowflake Federation Actually Switch User Identity per Query?

Hi everyone,I'm currently setting up Snowflake federation with Databricks using Microsoft Entra ID (U2M OAuth). However, I'm skeptical that the connection truly switches the user identity dynamically for each Databricks user (https://docs.databricks....

  • 286 Views
  • 6 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Snowflake federation with Databricks using Microsoft Entra ID (U2M OAuth) is intended to support per-user identity propagation—that is, each Databricks user is supposed to have queries executed under their own Snowflake identity at query time, rather...

  • 0 kudos
5 More Replies
Escarigasco
by New Contributor III
  • 43 Views
  • 2 replies
  • 2 kudos

Resolved! Azure Databricks Meters vs Databricks SKUs from system.billing table

When it comes to DBU, I am being charged by Azure for the following meters:- Premium Jobs Compute DBU <-- DBUs that my job computes are spending- Premium Serverless SQL DBU <-- DBUs that the SQL Warehouse compute is spending- Premium All-Purpose Phot...

  • 43 Views
  • 2 replies
  • 2 kudos
Latest Reply
Escarigasco
New Contributor III
  • 2 kudos

Thank you Bianca, great answer!

  • 2 kudos
1 More Replies
Nisha_Tech
by New Contributor II
  • 663 Views
  • 5 replies
  • 0 kudos

Databricks Asset Bundle Deployment Fails in GitHub Actions with Federated Identity Credentials

I am using a service principal with workspace admin access to deploy Databricks asset bundles. The deployment works successfully via Jenkins using the same credentials and commands. However, when attempting the deployment through GitHub Actions, I en...

  • 663 Views
  • 5 replies
  • 0 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 0 kudos

Environment variables override .databrickscfg, that's why it is probably failing to OIDC. Make sure that you have correct specification in your databricks.yml so it will be source of true. Smth like: - name: Deploy bundle env: DATABRICKS_HOST: ...

  • 0 kudos
4 More Replies
AlekseiDiaz
by New Contributor II
  • 168 Views
  • 2 replies
  • 0 kudos

Internet Access from Serverless Databricks - free trial

Hi community. I started to use databricks quick set up free trial and I have been trying to access internet from a python notebook but I haven't been able to do so. Even my UI is different. Is it becasue I am using free trial?

  • 168 Views
  • 2 replies
  • 0 kudos
Latest Reply
AlekseiDiaz
New Contributor II
  • 0 kudos

I changed the set up and I linked it to aws workspace. It doesn't raise any error now.But I was using requests

  • 0 kudos
1 More Replies
158576
by New Contributor
  • 90 Views
  • 1 replies
  • 0 kudos

mount cifs volume on all purpose compute results in permission denied

I have all networking already set, nslookup resolves NAS server IP and connectivity is enabled from worker nodes to nas server. I am able to mount the same nas drive outside of databricks, I mean standalone linux vm in the same VPC where worker nodes...

  • 90 Views
  • 1 replies
  • 0 kudos
Latest Reply
siva-anantha
Contributor
  • 0 kudos

Hello, Could you provide with more information about why you want to attach NAS drive to Databricks cluster, please? I am no expert in Storage. As far as I understand, NAS will suffer with IO and Replication Bottlenecks, when attached to Distributed ...

  • 0 kudos
Dharma25
by New Contributor III
  • 176 Views
  • 2 replies
  • 1 kudos

Task Hanging issue on DBR 15.4

Hello,I am running strucutred streaming pipeline with 5 models loaded using pyfunc.spark_udf. Lately we have been noticing very strange issue of tasks getting hanged and batch is taking very long time finishing its execution.CPU utilization is around...

Screenshot 2025-11-27 at 10.24.41 PM.png
  • 176 Views
  • 2 replies
  • 1 kudos
Latest Reply
bianca_unifeye
New Contributor III
  • 1 kudos

On DBR 15.4 the DeadlockDetector: TASK_HANGING message usually just means Spark has noticed some very long-running tasks and is checking for deadlocks. With multiple pyfunc.spark_udf models in a streaming query the tasks often appear “stuck” because ...

  • 1 kudos
1 More Replies
noorbasha534
by Valued Contributor II
  • 226 Views
  • 3 replies
  • 0 kudos

Azure VM quota for databricks jobs - demand prediction

Hey folks,a quick check -wanted to gather thoughts on how you manage demand for azure VM quota so you don't run into quota limits issues.In our case, we have several data domains (finance, master data, supply chain...) executing their projects in Dat...

  • 226 Views
  • 3 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

Yes, Azure Databricks compute policies let you define “quota-like” limits, but only within Databricks, not Azure subscription quotas themselves. You still rely on Azure’s own quota system for vCPU/VM core limits at the subscription level.​ What you c...

  • 0 kudos
2 More Replies
old_school
by New Contributor II
  • 145 Views
  • 3 replies
  • 0 kudos

Cap on OIDC (max 20) Enable workload identity federation for GitHub Actions

Hi Databricks community,I have followed below page and created Github OIDCs but there seems to be a cap on how many OIDC's a Service Principal can create (20 max). Is there any work around for this or some other solution apart from using Client ID an...

  • 145 Views
  • 3 replies
  • 0 kudos
Latest Reply
stbjelcevic
Databricks Employee
  • 0 kudos

I can't speak for specifically why, but allowing wildcards creates security risks and most identity providers and standards guidance require exact, pre-registered URLs.

  • 0 kudos
2 More Replies
Raman_Unifeye
by Contributor III
  • 264 Views
  • 5 replies
  • 2 kudos

Prevent Access to AI Functions Execution

As a workspace Admin, I want to prevent unexpected API costs from unrestricted usage of AI Functions (AI_QUERY() etc.), how can we control that only a particular group-users can execute AI Functions ?I understand the function execution cost can be vi...

  • 264 Views
  • 5 replies
  • 2 kudos
Latest Reply
Raman_Unifeye
Contributor III
  • 2 kudos

ok, so it has to be done at individual end-point and function level 

  • 2 kudos
4 More Replies
tinodj
by New Contributor II
  • 257 Views
  • 4 replies
  • 0 kudos

Real-time output missing when using “Upload and Run File” from VS Code

I am running Python files on a Databricks cluster using the VS Code Databricks extension, specifically the “Upload and Run File” command.I cannot get real-time output in the Debug Console. I have checked the official docs:https://learn.microsoft.com/...

  • 257 Views
  • 4 replies
  • 0 kudos
Latest Reply
tinodj
New Contributor II
  • 0 kudos

Yes, prints and loggings are viewable in driver logs as they happen. If the same file is run in databricks Web UI they are viewable on output window as they happen as well. But, when run through VS code, unfortunately they are not visible in the debu...

  • 0 kudos
3 More Replies
sparkplug
by New Contributor III
  • 1388 Views
  • 12 replies
  • 5 kudos

Resolved! I need a switch to turn off Data Apps in databricks workspaces

HiHow do I disable Data Apps on my workspace. This is really annoying that Databricks pushes new features without any option to disable them. At least you should have some tools to control access before rolling it out. It seems you only care about fe...

  • 1388 Views
  • 12 replies
  • 5 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 5 kudos

@Raman_Unifeye , I don't have visibility into the roadmap. However, if you are a customer you can always log a feature request. Cheers, Louis.

  • 5 kudos
11 More Replies
Marco37
by Contributor II
  • 2971 Views
  • 13 replies
  • 6 kudos

Resolved! Install python packages from Azure DevOps feed with service principal authentication

At the moment I install python packages from our Azure DevOps feed with a PAT token as authentication mechanism. This works well, but I want to use a service principal instead of the PAT token.I have created an Azure service principal and assigned it...

Marco37_0-1753975679472.png Marco37_1-1753975813527.png Marco37_2-1753975934347.png
  • 2971 Views
  • 13 replies
  • 6 kudos
Latest Reply
FilipD
New Contributor II
  • 6 kudos

I'm kinda late to the party but what is the suggested way of retriving the access token rn? Using some #bash or python code stored in global init script or cluster scoped init scripts? I don't want to stored this code in the notebook.Idea is to block...

  • 6 kudos
12 More Replies
JeremySu
by New Contributor III
  • 963 Views
  • 4 replies
  • 4 kudos

Resolved! A question about Databricks Fine-grained Access Control (FGAC) cost on dedicated compute

Hi All,recently, while testing Fine-grained Access Control (FGAC) on dedicated compute, I came across something that seems a bit unusual, and I’d like to ask if anyone else has seen similar behavior.I created a view with only one record, and had anot...

JeremySu_0-1761878010180.png
  • 963 Views
  • 4 replies
  • 4 kudos
Latest Reply
JeremySu
New Contributor III
  • 4 kudos

Hi @mark_ott Thank you very much for providing such a detailed and insightful explanation.This clearly resolves our confusion as to why an FGAC query that ran for only a few seconds ultimately incurred the DBU consumption shown on the bill, due to th...

  • 4 kudos
3 More Replies
kfadratek
by New Contributor
  • 308 Views
  • 1 replies
  • 0 kudos

Resolved! Issue Using Private CA Certificates for Databricks Serverless Private Git → On-Prem GitLab Connectio

Hi everyone,I’m trying to properly configure Databricks Serverless Private Git to connect to our on-premises GitLab, but I'm running into issues with private CA certificates.Following the latest Databricks recommendations, our connection to GitLab go...

  • 308 Views
  • 1 replies
  • 0 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 0 kudos

Hello @kfadratek ,  thanks for the detailed context — Let's take a look at what could be causing the SSL verification to fail with a custome CA in Serverless Private Git and discuss some approaches that might resolve it. What’s likely going wrong   B...

  • 0 kudos
trailblazer
by New Contributor III
  • 411 Views
  • 2 replies
  • 3 kudos

Resolved! Azure Databricks Cluster Pricing

Hi, I am trying to workout a rough total pricing of Azure Databricks Cluster using the following assumption. I want to spin a cluster on D13 v2 vms with 9 executors, so in total 1+9 = 10 nodes. I want to use the cluster for 10 hours a day, 30 hours a...

trailblazer_0-1762440739961.png
  • 411 Views
  • 2 replies
  • 3 kudos
Latest Reply
nayan_wylde
Esteemed Contributor
  • 3 kudos

Here is the simple calculation I use based on dollars and assuming the infra is in EUS.Cost ComponentsAzure VM Cost (D13 v2)On-demand price: $0.741/hour per VMMonthly VM cost:10 VMs×300 hours×$0.741=$2,223Yearly VM cost:10×3600×$0.741=$26,6762 2. Dat...

  • 3 kudos
1 More Replies