cancel
Showing results for 
Search instead for 
Did you mean: 
Administration & Architecture
Explore discussions on Databricks administration, deployment strategies, and architectural best practices. Connect with administrators and architects to optimize your Databricks environment for performance, scalability, and security.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

LuukDSL
by New Contributor
  • 143 Views
  • 3 replies
  • 0 kudos

Running jobs as service principal, while pulling code from Azure DevOps

In our Dataplatform, our jobs are defined in a dataplatform_jobs.yml within a Databricks Asset Bundle, and then pushed to Databricks via an Azure Devops Pipeline (Azure Devops is where our codebase resides). Currently, this results in workflows looki...

LuukDSL_0-1751983798686.png
  • 143 Views
  • 3 replies
  • 0 kudos
Latest Reply
ilir_nuredini
Valued Contributor
  • 0 kudos

Hello @LuukDSL ,Could you share a snippet of your CI/CD YAML file so we can give more specific advice?I’ve connected Azure DevOps to Databricks using ARM credentials, and it was set the pipeline’s "Run as" user to a service principal, no extra steps ...

  • 0 kudos
2 More Replies
Sisi
by New Contributor
  • 173 Views
  • 1 replies
  • 1 kudos

VS Code - ipynb vs py execution - spark issue

Databricks Connect works inside VS Code notebook but the same code fails in a standalone script withValueError: default auth: cannot configure default credentialsI’m developing locally with **Databricks Connect 16.1.6** and VS Code.Inside a Jupyter n...

  • 173 Views
  • 1 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @Sisi ,I think what's happening here is when you debug with option "Debug current file with Databricks Connect" then VS Code is using Databricks extension, which automatically handles authentication and sets up proper configuration.The regular Pyt...

  • 1 kudos
Sharanya13
by Contributor
  • 185 Views
  • 1 replies
  • 1 kudos

Lakebase use cases

1. What are the use cases for Lakebase?  When should I use the Lakebase Postgres over delta tables?2. What are the differences between open-source Postgres and Lakebase?3. Should I utilize Lakebase for all OLTP requirements?

  • 185 Views
  • 1 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @Sharanya13 ,1. Use Lakebase whenever you have application workload (OLTP) and you require low latency. For analytical workloads use Lakehouse. Here you have couple of example use cases from documentation:Serving data and/or features from the lake...

  • 1 kudos
carlosjuribe
by New Contributor III
  • 304 Views
  • 6 replies
  • 2 kudos

Out of memory error when installing environment dependencies of UC Python UDF

Hi,I've created a small UC Python UDF to test whether it works with custom dependencies (new PP feature), and every time I'm getting OOM errors with this message: [UDF_ENVIRONMENT_USER_ERROR.OUT_OF_MEMORY] Failed to install UDF dependencies for <cata...

  • 304 Views
  • 6 replies
  • 2 kudos
Latest Reply
Khaja_Zaffer
New Contributor II
  • 2 kudos

I tried with cluster, spent some couple of hours to load some libraries but unable to do. may be someone else can help you on this. 

  • 2 kudos
5 More Replies
bdanielatl
by New Contributor II
  • 676 Views
  • 2 replies
  • 3 kudos

Resolved! Markdown Cells Do Not Render Consistently

When I am creating a notebook in the UI editor on DataBricks, markdown cells do not always render after I run them. They still appear in 'editing mode'. See the screenshot below, it should have rendered a H1.Again, this behavior is not consistent. So...

bdanielatl_0-1736861864741.png
  • 676 Views
  • 2 replies
  • 3 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 3 kudos

Hi @bdanielatl, Thank you for reporting the issue with markdown cells not rendering consistently. This appears to be a known issue that has been encountered by other users as well. I will report it internally.

  • 3 kudos
1 More Replies
mbanxp
by New Contributor II
  • 611 Views
  • 5 replies
  • 4 kudos

Metastore deletion issues

Good afternoon, I have an issue with my metastore in North Europe.All my workspaces got detached:If I go to Databricks console, I can see the metastore in North Europe I created.However, when I select the metastore in North Europe, I get the followin...

mbanxp_0-1751622293898.png mbanxp_1-1751621952878.png mbanxp_2-1751621985852.png mbanxp_3-1751622036461.png
  • 611 Views
  • 5 replies
  • 4 kudos
Latest Reply
mbanxp
New Contributor II
  • 4 kudos

I solved the issue by deleting all the asignments before deleting the metastore.1. Access to Databricks CLI and authenticate 2. List metastores>> databricks account metastores list 3. List wotrkspaces and check assignments>> databricks account worksp...

  • 4 kudos
4 More Replies
EjB
by New Contributor
  • 351 Views
  • 1 replies
  • 1 kudos

Drop schema or catalog using cascade function

Hello In Databricks (non-Unity Catalog), I have two schemas (schema_a and schema_b) that both use the same root location in DBFS or external storage like ADLS.Example:abfss://container@storage_account.dfs.core.windows.net/data/project/schema_aabfss:/...

  • 351 Views
  • 1 replies
  • 1 kudos
Latest Reply
ilir_nuredini
Valued Contributor
  • 1 kudos

Hello @EjB For the given example, here is the response:Will DROP SCHEMA schema_a CASCADE remove or affect tables in schema_b?No, unless:1. The tables in schema_a are managed tables, AND2. Tables in schema_b store their data physically inside /schema_...

  • 1 kudos
Kutbuddin
by New Contributor II
  • 1475 Views
  • 3 replies
  • 0 kudos

[INTERNAL_ERROR] Query could not be scheduled: HTTP Response code: 503. Please try again later

We have a databricks job configured to run a dbt project. The dbt cli compute cluster being used is serverless with a serverless sql warehouse. We encountered this error during a run. SQLSTATE: XX000Any idea why this occurred?

  • 1475 Views
  • 3 replies
  • 0 kudos
Latest Reply
Amine8089
New Contributor II
  • 0 kudos

Hi,We are experiencing same recurring HTTP errors throughout the day when executing queries on Databricks. The specific error message we receive is: "[INTERNAL_ERROR] Query could not be scheduled: HTTP Response code: 503. Please try again later. SQLS...

  • 0 kudos
2 More Replies
antonionuzzo
by New Contributor III
  • 582 Views
  • 2 replies
  • 3 kudos

Resolved! System tables performance optimization

Hi AllAre there any Databricks lab projects or GitHub repositories that leverage system tables to provide dashboards or code for monitoring, and more importantly, for optimizing workflows and clusters based on usage?

  • 582 Views
  • 2 replies
  • 3 kudos
Latest Reply
Sharanya13
Contributor
  • 3 kudos

+1 to @szymon_dybczak. I would also add the dashboards for DB SQL Warehouse monitoring.

  • 3 kudos
1 More Replies
ziad
by New Contributor III
  • 3789 Views
  • 7 replies
  • 1 kudos

Resolved! Creating Azure Databricks Workspace Without NAT Gateway

Hello,Recently, when I create a new Databricks workspace on Azure, it automatically create a NAT Gateway which incurs additional cost !When creating the workspace, I don't choose secure cluster connectivity, so I'm expecting not to have a NAT Gateway...

nat 1.jpg nat 2.jpg
  • 3789 Views
  • 7 replies
  • 1 kudos
Latest Reply
Khaja_Zaffer
New Contributor II
  • 1 kudos

Hello @OmarSEKkOUTI The solution is simple: try to create a vnet-injected workspace. like here: Azure Vnet injection  This is to avoid creating a NAT gateway.  Also, Databricks says:If you enable secure cluster connectivity on your workspace that use...

  • 1 kudos
6 More Replies
AndyK
by New Contributor II
  • 445 Views
  • 2 replies
  • 0 kudos

Databricks Apps with FastAPI

I have a FastAPI Databricks App deployed to a hosted workspace. How can I call the API from external tools like Postman?P.S. I was able to call the API within the same browser.

  • 445 Views
  • 2 replies
  • 0 kudos
Latest Reply
SP_6721
Contributor III
  • 0 kudos

Hi @AndyK ,External tools don’t share browser’s login context, so you’ll need to add a bearer token in the Authorization header manually.Start by generating an access token. You can either use the Databricks CLI (databricks auth token) to get an OAut...

  • 0 kudos
1 More Replies
MBV3
by Contributor
  • 526 Views
  • 6 replies
  • 0 kudos

Unable to see sample data in Hive Metastore after moving to GCE

Hi,We have recently moved from GKE to GCE, it is taking forever to load the sample data in the manged delta tables.Even running simple select sql statements are taking forever. Totally clueless here, any help will be appreciatedThanks

  • 526 Views
  • 6 replies
  • 0 kudos
Latest Reply
MBV3
Contributor
  • 0 kudos

Hi All,Strangely after struggle for 2 days we figured out that we can't run the cluster in scalable mode, so after selecting single node mode we are able to execute queries and job. It seems there is a bug in the Databrick's GKE to GCE migration. Won...

  • 0 kudos
5 More Replies
Chinu
by New Contributor III
  • 291 Views
  • 1 replies
  • 0 kudos

Best Approach to Retrieve Policy IDs Across Multiple Workspaces

Hi, I’m aware of the API endpoint api/2.0/policies/clusters/list to fetch a list of policy IDs and names. However, we have 50 different workspaces, and I need to retrieve the specific policy ID and name for each one.Could you advise on the most effic...

  • 291 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @Chinu ,Databricks does not provide a global API to query all workspaces in a single call. I guess your only option for now is to use scripting approach.

  • 0 kudos
noorbasha534
by Valued Contributor
  • 381 Views
  • 1 replies
  • 0 kudos

Oauth Token federation

Dear allHas anyone tried oauth token federation for authentication with Databricks REST APIs?appreciate if there is a re-usable code snippet to achieve the same.

  • 381 Views
  • 1 replies
  • 0 kudos
Latest Reply
noorbasha534
Valued Contributor
  • 0 kudos

Documentation says - '''An account federation policy enables all users and service principals in your Azure Databricks account to access Databricks APIs using tokens from your identity provider. '''' But, in the examples that are provided, I do not s...

  • 0 kudos