cancel
Showing results for 
Search instead for 
Did you mean: 
Administration & Architecture
Explore discussions on Databricks administration, deployment strategies, and architectural best practices. Connect with administrators and architects to optimize your Databricks environment for performance, scalability, and security.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

JanJaros
by New Contributor
  • 1057 Views
  • 0 replies
  • 0 kudos

Databricks OAUTH(OIDC) with ORY Network

Hi,we are trying to setup OIDC AUTH for Databricks with our Ory Network account. So far we have been using it without any issues with all of our apps and now we wanted to set it up also for Databricks. Unfortunately after many attempts with different...

  • 1057 Views
  • 0 replies
  • 0 kudos
AnkurMittal008
by New Contributor III
  • 5001 Views
  • 2 replies
  • 0 kudos

Databricks App : Limitations

I have some questions regarding Databricks App.1) Can we use Framework other than mentioned in documentation( Streamlit,Flask,Dash,Gradio,Shiny).2) Can we allocate compute more than 2 vCPU and 6GB memory to any App.3) Any other programming language o...

  • 5001 Views
  • 2 replies
  • 0 kudos
Latest Reply
Ismael-K
Databricks Employee
  • 0 kudos

1.) You can use most Python-based application frameworks, including some beyond those mentioned above.(Reference here) 2.) Currently, app capacity is limited to 2 vCPUs and 6 GB of RAM. However, future updates may introduce options for scaling out an...

  • 0 kudos
1 More Replies
andreapeterson
by Contributor
  • 1393 Views
  • 4 replies
  • 2 kudos

Resolved! OAuth API for service user

Is there a way to programmatically create an OAuth secret for for a workspace service principal via API/SDK? As of now, the only way I can see doing this is through UI

  • 1393 Views
  • 4 replies
  • 2 kudos
Latest Reply
vr
Contributor III
  • 2 kudos

@andreapeterson isn't it the API you are looking for?https://docs.databricks.com/api/azure/account/serviceprincipalsecrets/createIt is an account-level API, but, counterintuitively, when we create service principals in the workspace, they propagate i...

  • 2 kudos
3 More Replies
aswinkks
by New Contributor III
  • 581 Views
  • 1 replies
  • 0 kudos

Load assignment during Distributed training

Hi,I wanted to confirm, in a distributed training, if there is any way that I can control what kind/amount of load/data can be send to specific worker nodes, manually ..Or is it completely automatically handled by spark's scheduler, and we don't have...

  • 581 Views
  • 1 replies
  • 0 kudos
Latest Reply
Renu_
Valued Contributor II
  • 0 kudos

From what I know, Spark automatically handles how data and workload are distributed across worker nodes during distributed training, you can't manually control exactly what or how much data goes to a specific node. You can still influence the distrib...

  • 0 kudos
BalajiM
by New Contributor
  • 716 Views
  • 1 replies
  • 0 kudos

Running Driver Intensive workloads on all purpose compute

Recently observed when we run a driver intensive code on a all purpose compute. The parallel runs of the same pattern/kind jobs are getting failedExample:Job triggerd on all purpose compute with compute stats of 4 core and 8 gigs ram for driverLets s...

  • 716 Views
  • 1 replies
  • 0 kudos
Latest Reply
Aviral-Bhardwaj
Esteemed Contributor III
  • 0 kudos

this will help you # Cluster config adjustmentsspark.conf.set("spark.driver.memory", "16g") # Double current allocationspark.conf.set("spark.driver.maxResultSize", "8g") # Prevent large collects

  • 0 kudos
Inna_M
by New Contributor III
  • 904 Views
  • 3 replies
  • 3 kudos

Resolved! how to modify workspace creator

Hi. Our 3 workspaces were created by a consultant who is no longer with us.  The workspace shows his name still . How can we change it? Also what kind of account should be a creator of a workspace: Service principal, AD account, AD group? Pease see b...

Inna_M_0-1748370746975.png
  • 904 Views
  • 3 replies
  • 3 kudos
Latest Reply
MoJaMa
Databricks Employee
  • 3 kudos

You can file an Azure ticket and request them to contact Databricks Support, who will contact Databricks Engineering, to change that 'workspace owner'. If you have a Databricks Account Team, they can also file an internal ticket to Databricks Enginee...

  • 3 kudos
2 More Replies
vr
by Contributor III
  • 908 Views
  • 4 replies
  • 0 kudos

Terraforming Git credentials for service principals

I am terraforming service principals in my Databricks workspace and it works great until I need to assign Git credentials to my SP. In the UI we have these options to configure credentials on service principal page:However the Terraform resource I fo...

vr_0-1747957962260.png
  • 908 Views
  • 4 replies
  • 0 kudos
Latest Reply
Rjdudley
Honored Contributor
  • 0 kudos

You're a little bit ahead of me in this process, so I haven't tried the solution yet, but it looks like you create a git credential resource for the service principal.  This requires a token, which I think must be generated in the console.  My refere...

  • 0 kudos
3 More Replies
n-var
by New Contributor
  • 827 Views
  • 1 replies
  • 0 kudos

Unity Catalog: 403 Error When Connecting S3 via IAM Role and Storage Credential

Hi,We're currently setting up Databricks Unity Catalog on AWS. We created an S3 bucket and assigned an IAM role (databricks-storage-role) to give Databricks access.Note: Databricks doesn't use the IAM role directly. Instead, it requires a Storage Cre...

  • 827 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

Have you follow any specific guide for the creation of the same? Are you setting up a Unity Catalog Metastore or the default storage for the workspace?For the Metastore creation have you follow steps in https://docs.databricks.com/aws/en/data-governa...

  • 0 kudos
PabloCSD
by Valued Contributor II
  • 2385 Views
  • 7 replies
  • 1 kudos

How to install (mssql) drivers to jobcompute?

Hello, I'm having this issue with job-computes:The snippet of the code is as follows: 84 if self.conf["persist_to_sql"]: 85 # persist to sql 86 df_parsed.write.format( 87 "com.microsoft.sqlserver.jdbc.spark" 88...

PabloCSD_0-1745527653462.png
  • 2385 Views
  • 7 replies
  • 1 kudos
Latest Reply
NandiniN
Databricks Employee
  • 1 kudos

For a job compute, you would have to go init script route. Can you please highlight, the cause of the failure of library installation via init script? 

  • 1 kudos
6 More Replies
Vittorio
by New Contributor II
  • 1985 Views
  • 1 replies
  • 0 kudos

Static IP for existing workspace

Is there a way to have static IP addresses for Azure Databricks without creating new workspace?We have worked a lot in 2 workspaces (dev and main), but now we need static IP addresses for both to work with some APIs. Do we really have to recreate the...

  • 1985 Views
  • 1 replies
  • 0 kudos
Latest Reply
Rjdudley
Honored Contributor
  • 0 kudos

I don't think so, at least not on Azure.  What you need to do depends on how you set up your workspaces.  In Azure, if you just use a default install, a NAT gateway is created and configured for you, so you probably already have a static IP.If you us...

  • 0 kudos
AlbertWang
by Valued Contributor
  • 9160 Views
  • 11 replies
  • 0 kudos

What is the best practice for connecting Power BI to Azure Databricks?

I refer this document to connect Power BI Desktop and Power BI Service to Azure Databricks.Connect Power BI to Azure Databricks - Azure Databricks | Microsoft LearnHowever, I have a couple of quesitions and concerns. Can anyone kindly help?It seems l...

  • 9160 Views
  • 11 replies
  • 0 kudos
Latest Reply
sandeepmankikar
Contributor
  • 0 kudos

To securely connect Power BI to Azure Databricks, avoid using PATs and instead configure a Databricks Service Principal with SQL Warehouse access. Power BI Service does not support Client Credential authentication, so Service Principal authentication...

  • 0 kudos
10 More Replies
oricaruso
by New Contributor II
  • 757 Views
  • 1 replies
  • 0 kudos

Gcs databricks community

Hello,I would like to know if it is possible to connect my Databricks community account with a Google cloud storage account via a notebook.I tried to connect it via the json key of my gcs service account but the notebook always gives this error when ...

  • 757 Views
  • 1 replies
  • 0 kudos
Latest Reply
SP_6721
Contributor III
  • 0 kudos

Hi @oricaruso To connect to GCS, you typically need to set the service account JSON key in the cluster’s Spark config, not just in the notebook. However, since the Community Edition has several limitations, like the absence of secret scopes, restrict...

  • 0 kudos
cmathieu
by New Contributor III
  • 2180 Views
  • 3 replies
  • 0 kudos

Multiple feature branches per user using Databricks Asset Bundles

I'm currently helping a team migrating to DABs from dbx and they would like to be able to work on multiple features at the same time.What I was able to do is pass the current branch as a variable in the root_path and various names, so when the bundle...

  • 2180 Views
  • 3 replies
  • 0 kudos
Latest Reply
cmilligan262
New Contributor II
  • 0 kudos

@cmathieu can you provide an example inserting the branch name? I'm trying to do the same thing

  • 0 kudos
2 More Replies
mzs
by Contributor
  • 1029 Views
  • 1 replies
  • 0 kudos

Resolved! Removing compute policy permissions using Terraform

By default, the "users" and "admins" groups have CAN_USE permission on the Personal Compute policy.I'm using Terraform and would like to prevent regular users from using this policy to create additional compute clusters.I haven't found a way to do th...

  • 1029 Views
  • 1 replies
  • 0 kudos
Latest Reply
mzs
Contributor
  • 0 kudos

I learned the Personal Compute policy can be turned off at the account level:https://learn.microsoft.com/en-us/azure/databricks/admin/clusters/personal-compute#manage-policy 

  • 0 kudos
harishgehlot
by New Contributor III
  • 1185 Views
  • 4 replies
  • 0 kudos

Resolved! Shall we opt for multiple worker nodes in dab workflow template if our codebase is based on pandas.

Hi team, I am working in a databricks asset bundle architecture. Added my codebase repo in a workspace. My question to do we need to opt for multiple worker nodes like num_worker_nodes > 1 or autoscale with range of worker nodes if my codebase has mo...

  • 1185 Views
  • 4 replies
  • 0 kudos
Latest Reply
harishgehlot
New Contributor III
  • 0 kudos

Thanks @Shua42 . You really helped me a lot.

  • 0 kudos
3 More Replies