cancel
Showing results for 
Search instead for 
Did you mean: 
Administration & Architecture
Explore discussions on Databricks administration, deployment strategies, and architectural best practices. Connect with administrators and architects to optimize your Databricks environment for performance, scalability, and security.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

ac0
by Contributor
  • 5033 Views
  • 3 replies
  • 1 kudos

Delta Live Table pipeline steps explanation

Does anyone have documentation on what is actually occurring in each of these steps?Creating update Waiting for resourcesInitializingSetting up tablesRendering graphFor example, what is the difference between initializing and setting up tables? I am ...

  • 5033 Views
  • 3 replies
  • 1 kudos
Latest Reply
Mounika_Tarigop
Databricks Employee
  • 1 kudos

Yes, loading data (full refresh/refresh) into all streaming tables and refreshing materialized views are part of the "Setting up table" step in a Delta Live Tables (DLT) pipeline when running in trigger mode.In triggered mode, materialized views are ...

  • 1 kudos
2 More Replies
Dex
by New Contributor
  • 1347 Views
  • 1 replies
  • 0 kudos

How to use Manged Identitify within Databricks Azure to access Blob Container?

Hi,My organization has asked that all blob storage accounts be accessed via managed identity. Several data brick notebooks are affected, so I'm currently trying to see how to set up a managed identity.We've added the Databricks resource provider to t...

  • 1347 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

Have you followed the instructions available in docs https://learn.microsoft.com/en-us/azure/databricks/data-governance/unity-catalog/azure-managed-identities 

  • 0 kudos
JSilverberg
by New Contributor
  • 3946 Views
  • 1 replies
  • 0 kudos

Resolved! Azure SCIM Provisioning Failures due to Resource Exhaustion

I had to make a significant change to the group membership in an Entra SCIM Provisioned group to Databricks, and the connector removed ALL users from the group (and sync'd) and then when I fixed, by adding a replacement dynamic group (with about 30% ...

  • 3946 Views
  • 1 replies
  • 0 kudos
Latest Reply
parthSundarka
Databricks Employee
  • 0 kudos

Hi @JSilverberg , This occurs when the number of requests goes beyond the rate limits mentioned here. https://learn.microsoft.com/en-us/azure/databricks/resources/limits#:~:text=Identity,No To prevent these issues in the future, it is best to do thes...

  • 0 kudos
RohitKulkarni
by Contributor II
  • 6741 Views
  • 1 replies
  • 0 kudos

Data Bricks Architect ceritification

Hello Team,I am planning to pursue the Databricks Architect certification. Could you please let me know which certification I should opt for?If you have any study material or relevant links, kindly share them.Your support would be highly appreciated....

  • 6741 Views
  • 1 replies
  • 0 kudos
Latest Reply
Rjdudley
Honored Contributor
  • 0 kudos

The Architect credential is an accreditation, not a certification.  Accreditations are less rigorous and less expensive than certifications.  You didn't say which platform you are on, so here are links to the learning plans (which have the exams) for...

  • 0 kudos
F_Goudarzi
by New Contributor III
  • 1995 Views
  • 1 replies
  • 1 kudos

Pros and cons of putting all various Databricks workspaces (dev, qa , prod) under one metastore

Hi there, If we have separate workspaces for each and every environment, then how we should go about structuring the metastore? What are the pros and cons of putting all workspaces under one metastore instead of having separate metastore for each?Tha...

  • 1995 Views
  • 1 replies
  • 1 kudos
Latest Reply
Walter_C
Databricks Employee
  • 1 kudos

Hello Fatima, many thanks for your question. Please first note that if all the workspaces belong to the same account id and are on the same cloud region, they will all need to be associated with the same metastore as you can only have 1 metastore per...

  • 1 kudos
Dnirmania
by Contributor
  • 8497 Views
  • 7 replies
  • 9 kudos

Resolved! Databricks best practices for azure storage account

Hello EveryoneCurrently, We are in process of building azure databricks and have some doubt regarding best practices to follow for azure storage account which we will be using to store data. Can anyone help me finding best practices to follow for sto...

  • 8497 Views
  • 7 replies
  • 9 kudos
Latest Reply
bhanu_gautam
Valued Contributor III
  • 9 kudos

Thanks for sharing @Rjdudley @szymon_dybczak @filipniziol 

  • 9 kudos
6 More Replies
enr0c
by New Contributor
  • 3350 Views
  • 1 replies
  • 0 kudos

Budget Policy - Service Principals don't seem to be allowed to use budget policies

ObjectiveTransfer existing DLT pipeline to new owner (service principal). Budget policies enabled.Steps to reproduceCreated a service principalAssigned it group membership of a group that is allowed to use a budget policyEnsured it has access to the ...

a203a1bb-80d2-4c51-bc2f-dbd379b6e5e8.png Screenshot 2024-11-20 090829.png Screenshot 2024-11-20 090938.png Screenshot 2024-11-20 091107.png
Administration & Architecture
budget-policy
service-principal
  • 3350 Views
  • 1 replies
  • 0 kudos
Latest Reply
Abishek
Databricks Employee
  • 0 kudos

“Thanks for your question! I’m looking into this and will get back to you.”

  • 0 kudos
chethankumar
by New Contributor III
  • 1022 Views
  • 3 replies
  • 1 kudos

Resolved! How to add existing recipient to existing delta share

I created a recipient in the Databricks console and also set up a Delta Share. Now, I’d like to link this existing recipient to the Delta Share. Is there a way to accomplish this using Terraform?

  • 1022 Views
  • 3 replies
  • 1 kudos
Latest Reply
chethankumar
New Contributor III
  • 1 kudos

Hi @Naïm Thanks for your response. It seems your answer is helping me, but I'm facing another issue. The owner of my recipient is a group, not an individual user. I'm running this Terraform script using a service principal that is a member of that gr...

  • 1 kudos
2 More Replies
noorbasha534
by Valued Contributor II
  • 2239 Views
  • 3 replies
  • 0 kudos

Resolved! Control plane set-up

Dear all,In this video from Databricks, Azure Databricks Security Best Practices - https://www.youtube.com/watch?v=R1X8ydIR_Bc&t=623sduring this duration in the video 13.25 - 14.35the presenter talks about benefits of private endpoints. He makes the ...

  • 2239 Views
  • 3 replies
  • 0 kudos
Latest Reply
parthSundarka
Databricks Employee
  • 0 kudos

Hi @noorbasha534, Does this control plane then contains management services for several customers? - Yes, Control Plane has management services that are used across customers in the region.  Due to which the presenter says traffic can be isolated fro...

  • 0 kudos
2 More Replies
Aria
by New Contributor III
  • 3904 Views
  • 5 replies
  • 1 kudos

policy_id in databricks asset bundle workflow

We are using databricks asset bundle for code deployment and biggest issue I am facing is that policy_id is different in each environment.I tried with environment variable sin azure devops and also with declaring the variables in databricks.yaml and ...

  • 3904 Views
  • 5 replies
  • 1 kudos
Latest Reply
maikl
New Contributor III
  • 1 kudos

Solved by the lookup function https://docs.databricks.com/en/dev-tools/bundles/variables.html#retrieve-an-objects-id-value

  • 1 kudos
4 More Replies
staskh
by New Contributor III
  • 932 Views
  • 1 replies
  • 0 kudos

Is it possible to disable file download in Volumes interface?

Workspace security administration panel offers to disable downloads in notebook folders and workspaces. However, it seems that even if all those downloads are disabled, the "Volumes" panel of Unity Catalog still offers a file download button. Is it p...

  • 932 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @staskh ,Unfortunately, I don't think it is possible to disable it via UI currently. But volumes are governed by UC permission, so maybe you can try to set permission to read/write to approved group of users and take permission of users who should...

  • 0 kudos
Useit
by New Contributor
  • 3725 Views
  • 1 replies
  • 0 kudos

Issues with Delta Sharing API when using Service Principal Token

Hello,I am currently working with the Delta Sharing API and have encountered an issue when using a Service Principal token for authentication. The API call returns the following error:[CANNOT_INFER_EMPTY_SCHEMA] Can not infer schema from empty datase...

  • 3725 Views
  • 1 replies
  • 0 kudos
Latest Reply
Mounika_Tarigop
Databricks Employee
  • 0 kudos

Please find the response below:  1) The Delta Sharing API supports both personal access tokens and service principal tokens for authentication. 2) Service principals need to be granted specific roles and permissions to access data. This includes assi...

  • 0 kudos
hellobuddy
by New Contributor
  • 3428 Views
  • 1 replies
  • 0 kudos
Administration & Architecture
Azure Databricks
cluster
  • 3428 Views
  • 1 replies
  • 0 kudos
Latest Reply
Mounika_Tarigop
Databricks Employee
  • 0 kudos

The below documentation shows how to install libraries in a cluster.   https://docs.databricks.com/en/libraries/cluster-libraries.html#install-a-library-on-a-cluster

  • 0 kudos
manojpatil04
by New Contributor III
  • 951 Views
  • 3 replies
  • 2 kudos

GCP Databricks GKE cluster with 4 nodes

I am working on setting up GCP Databricks and successfully created first GCP-Databricks workspace, but what I observed is it is incurring additional charges even i am using 14days free trail. It is GKE cluster with 4 nodes which are spin up as part o...

  • 951 Views
  • 3 replies
  • 2 kudos
Latest Reply
manojpatil04
New Contributor III
  • 2 kudos

Thank you @Louis_Frolio,Just want to dig more into this as is there any way to reduce this nodes using CLI or creating customer managed network.

  • 2 kudos
2 More Replies
karen_c
by New Contributor II
  • 3592 Views
  • 2 replies
  • 2 kudos

Resolved! Databricks All-purpose compute Pricing

Hello, I am now struggling how to calculate the cost of my job cluster.My configuration is as below:If I have to run the above cluster 18 hours per day, in Standard Tier and East Asia Region, how much will be the pricing of the cluster?Any help provi...

karen_c_0-1733467288235.png karen_c_1-1733467313502.png
  • 3592 Views
  • 2 replies
  • 2 kudos
Latest Reply
Takuya-Omi
Valued Contributor III
  • 2 kudos

@karen_c Let me make a small correction.It seems that you have checked the option for Spot Instances, which should make the cost slightly lower. Please refer to the far-right column of the attached pricing table for more details.Additionally, you hav...

  • 2 kudos
1 More Replies