cancel
Showing results for 
Search instead for 
Did you mean: 
Administration & Architecture
Explore discussions on Databricks administration, deployment strategies, and architectural best practices. Connect with administrators and architects to optimize your Databricks environment for performance, scalability, and security.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Carsten03
by New Contributor III
  • 26897 Views
  • 11 replies
  • 6 kudos

Resolved! Run workflow using git integration with service principal

Hi,I want to run a dbt workflow task and would like to use the git integration for that. Using my personal user I am able to do so but I am running my workflows using a service principal.I added git credentials and the repository using terraform. I a...

  • 26897 Views
  • 11 replies
  • 6 kudos
Latest Reply
Coffee77
Honored Contributor II
  • 6 kudos

On the other hand, another approach you could use. Configure your tasks with relative paths to notebooks and deploy all of them with DAB. Your job will reference directly the deployed notebook, no need to access GIT from jobs/notebooks. That is deleg...

  • 6 kudos
10 More Replies
hv_sg3
by New Contributor
  • 269 Views
  • 1 replies
  • 1 kudos

Enable Compute Policy Management and Compute Policy Admin Role

Hi,I have an account with an Enterprise plan and wanted to change some features of the compute policy for a cluster i wanted to create in a workspace I am an Admin of. But I cannot because the fields are read-only.Co-Pilot directed me to look for an ...

  • 269 Views
  • 1 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @hv_sg3 ,That's weird. As an workspace admin you should be able to do that. Could you attach some screens?

  • 1 kudos
Marco37
by Contributor II
  • 5403 Views
  • 13 replies
  • 6 kudos

Resolved! Install python packages from Azure DevOps feed with service principal authentication

At the moment I install python packages from our Azure DevOps feed with a PAT token as authentication mechanism. This works well, but I want to use a service principal instead of the PAT token.I have created an Azure service principal and assigned it...

Marco37_0-1753975679472.png Marco37_1-1753975813527.png Marco37_2-1753975934347.png
  • 5403 Views
  • 13 replies
  • 6 kudos
Latest Reply
FilipD
New Contributor II
  • 6 kudos

I'm kinda late to the party but what is the suggested way of retriving the access token rn? Using some #bash or python code stored in global init script or cluster scoped init scripts? I don't want to stored this code in the notebook.Idea is to block...

  • 6 kudos
12 More Replies
APJESK
by Contributor
  • 423 Views
  • 1 replies
  • 0 kudos

Resolved! Can anyone share Databricks security model documentation or best-practice references

Can anyone share Databricks security model documentation or best-practice references

  • 423 Views
  • 1 replies
  • 0 kudos
Latest Reply
Coffee77
Honored Contributor II
  • 0 kudos

Here is the official documentation of Databricks: https://docs.databricks.com/aws/en/security/  Do you need to dive deeper into any specific area?

  • 0 kudos
chandru44
by New Contributor II
  • 405 Views
  • 1 replies
  • 1 kudos

Moving Databricks Metastore Storage Account Between Azure Subscriptions

I have two Azure subscriptions: one for Prod and another for Non-Prod. During the initial setup of the Non-Production Databricks Workspace, I configured the metastore storage account in the Non-Prod subscription. However, I now want to move this meta...

chandru44_0-1763266537882.png
  • 405 Views
  • 1 replies
  • 1 kudos
Latest Reply
Coffee77
Honored Contributor II
  • 1 kudos

Assuming the metastore is the same for your DEV and PROD environments and what you want is just to use the same storage account + container to place managed tables, volumes, etc. in theory you just need to copy all content from your source storage ac...

  • 1 kudos
pdiamond
by Contributor
  • 423 Views
  • 1 replies
  • 2 kudos

Resolved! Lakebase query history / details

Is there somehwere in Databricks that I can see details about queries run againt one of my Lakebase databases (similar to query history system tables)?What I'm ultimately trying to figure out is where the time is being spent between when I issue the ...

  • 423 Views
  • 1 replies
  • 2 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 2 kudos

Hi @pdiamond ,Currently in beta there's a feature that let's you monitor active queries:https://docs.databricks.com/aws/en/oltp/projects/active-queriesAlso in beta there's Lakebase SQL editor that will allow you to analyze queries:https://docs.databr...

  • 2 kudos
RDE305
by New Contributor II
  • 461 Views
  • 1 replies
  • 1 kudos

A single DLT for Ingest - feedback on this architecture

What are your thoughts on this Databricks pipeline design?Different facilities will send me backups of a proprietary transactional database containing tens of thousands of tables. Each facility may have differences in how these tables are populated o...

  • 461 Views
  • 1 replies
  • 1 kudos
Latest Reply
nayan_wylde
Esteemed Contributor II
  • 1 kudos

Your design shows strong alignment with the Medallion Architecture principles and addresses schema variability well, but there are some scalability and governance considerations worth discussing. Also Pre-Bronze, Building a schema registry early is e...

  • 1 kudos
FabianGutierrez
by Contributor
  • 1798 Views
  • 3 replies
  • 0 kudos

Looking for experiences with DABS CLI Deployment, Terraform and Security

Hi Community,I hope my topic finds you well. Within our Databricks landscape we decided to use DABS (Databricks Asset Bundles) however we found out (the hard way) that it uses Terraform for Deployment purposes. This is a concern now for Security and ...

  • 1798 Views
  • 3 replies
  • 0 kudos
Latest Reply
Coffee77
Honored Contributor II
  • 0 kudos

Try to use always service principals to deploy your asset bundles. If desired take a look here: https://www.youtube.com/watch?v=5WreXn0zbt8 Concerning terraform state, it is indeed generated, take a look at this picture extracted from one of my deplo...

  • 0 kudos
2 More Replies
eshwari
by New Contributor III
  • 1243 Views
  • 1 replies
  • 1 kudos

Resolved! Restricting Catalog and External Location Visibility Across Databricks Workspaces

Restricting Catalog and External Location Visibility Across Databricks Workspaces I am facing exact similar issue, But I don't want to create separate metastore. and I have added environment name as a prefix to all external locations. All the locatio...

  • 1243 Views
  • 1 replies
  • 1 kudos
Latest Reply
mark_ott
Databricks Employee
  • 1 kudos

You can hide or scope external locations and catalogs so they are only visible within their respective Databricks workspaces—even when using a shared metastore—by using "workspace binding" (also called isolation mode or workspace-catalog/workspace-ex...

  • 1 kudos
LEE_SUKJUN
by New Contributor II
  • 297 Views
  • 1 replies
  • 0 kudos

Inquire the location of creating a metastore resource

 I created Databricks on AWS today.Of course, we're planning to switch to paid.By the way, the Metastore is in the US Region after I was born.I'm a Korea APJ, and is the Metastore all only run in the US?Does the Metastore have no impact if I query or...

LEE_SUKJUN_0-1762932473218.png
  • 297 Views
  • 1 replies
  • 0 kudos
Latest Reply
Coffee77
Honored Contributor II
  • 0 kudos

Hi @LEE_SUKJUN , I think the general principle should be to keep all components (metastore, workspace, and cloud storage) in the same region in order to avoid cross-region latency, data egress costs, and compliance issues.Concerning number of metasto...

  • 0 kudos
DazMunro
by New Contributor
  • 642 Views
  • 2 replies
  • 0 kudos

Using or integrating SIlver or Gold Zone data in an Operational API

I am looking to understand what sort of approach we can take to use Silver or Gold zone data in an Operational style API, or even if we should. We have data that makes it's way to the Silver and Gold zones in our Medallion Architecture and it kind of...

  • 642 Views
  • 2 replies
  • 0 kudos
Latest Reply
Rjdudley
Honored Contributor
  • 0 kudos

Databricks is an analytics system and isn't optimized to perform as an OLTP system.  Additionally, Databricks compute can scale to zero if you set it to do so.  This means if you want to use gold/silver data in a real-time way you need to keep a clus...

  • 0 kudos
1 More Replies
pranav5
by New Contributor II
  • 609 Views
  • 1 replies
  • 1 kudos

Resolved! Trying to Backup Dashboards and Queries from our Workspace.

We are using a databricks workspace and our IT team is decommissioning it as our time with it is being done. I have many queries and dashboards developed. I want to copy these, unfortunately when i download using zip or .dbc these queries or dashboar...

  • 609 Views
  • 1 replies
  • 1 kudos
Latest Reply
bianca_unifeye
Databricks MVP
  • 1 kudos

NotebooksThese are the easiest assets to back up.You can export them individually or in bulk as:.dbc – Databricks archive format (can re-import directly into a new workspace).source or .py – raw code export (ideal for version control)To download in b...

  • 1 kudos
jzu
by New Contributor II
  • 1599 Views
  • 5 replies
  • 1 kudos

Problem with Metastore

Hello community.We are facing an issue when deploying and configuring metastore using terraform. We are using Azure Devops pipeline for deployment. The identity running the pipeline is a managed identity and it's set as account admin in Account porta...

  • 1599 Views
  • 5 replies
  • 1 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 1 kudos

@jzu , is this a new error or is it the same as before. I need more details please. Louis.

  • 1 kudos
4 More Replies
ambigus9
by Contributor
  • 4744 Views
  • 2 replies
  • 0 kudos

R-studio on Dedicated Cluster Invalid Access Token

Hello!! Currently I have an R-studio installed on a Dedicated Cluster over Azure Databricks, here are the specs:I must to make enfasis over the Access mode: Manual and Dedicated to a Group.Here, we install R-studio using a notebook with the following...

ambigus9_0-1743020318837.png error token access.png
  • 4744 Views
  • 2 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

You’re seeing two key issues with your RStudio Server on Azure Databricks: RStudio stops working after 1–2 days. You get permission errors using sparklyr and can’t update the Connections pane. Let’s address each: 1. RStudio Server Stops Working A...

  • 0 kudos
1 More Replies
Carelytix
by New Contributor III
  • 377 Views
  • 1 replies
  • 1 kudos

Resolved! Signing up BAA requiring Compliance Security Profile activation

Hello folks —  I need help with enabling compliance security profile for my account. I need this to execute/sign a HIPPA BAA on my account. For this, I need an enhanced security & compliance add-on for this. I first reached out to help@databricks.com...

  • 377 Views
  • 1 replies
  • 1 kudos
Latest Reply
Carelytix
New Contributor III
  • 1 kudos

I was able to turn this feature on by upgrading the plan to "Enterprise". Thanks!

  • 1 kudos