cancel
Showing results for 
Search instead for 
Did you mean: 
Administration & Architecture
Explore discussions on Databricks administration, deployment strategies, and architectural best practices. Connect with administrators and architects to optimize your Databricks environment for performance, scalability, and security.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

SunilPoluri
by New Contributor
  • 1125 Views
  • 2 replies
  • 0 kudos

Create Databricks managed service principal programatically ?

For the current Databricks service principal API or the Databricks SDK, an ID is required. However, when dealing with Databricks-managed service principals, you typically only have the name. For registering with cloud providers, like Microsoft Entra ...

  • 1125 Views
  • 2 replies
  • 0 kudos
Latest Reply
giladba
New Contributor III
  • 0 kudos

Have you found a solution on how to programmatically create a Databricks managed service principal?

  • 0 kudos
1 More Replies
gyorgyjelinek
by New Contributor II
  • 321 Views
  • 2 replies
  • 2 kudos

Resolved! Default schema in SQL Editor is not 'default' when unity catalog is set as default catalog

In workspace settings: Workspace admin - advanced - other - Default catalog for the workspace is set to different than hive_metastore, it is set to a `Unity Catalog` catalog - the expected behaviour is copied here from the related more info panel:"Se...

  • 321 Views
  • 2 replies
  • 2 kudos
Latest Reply
gyorgyjelinek
New Contributor II
  • 2 kudos

Hi @Alberto_Umana ,Thank you for the explanation. I mark your comment as the accepted solution as it contains the current implementation logic and the work around. Good to know that the more info panel is a bit misleading as of now because the SQL Ed...

  • 2 kudos
1 More Replies
charl-p-botha
by New Contributor III
  • 674 Views
  • 10 replies
  • 3 kudos

Error "Integrating Apache Spark with Databricks Unity Catalog Assets via Open APIs" on Azure

Great blog post: https://community.databricks.com/t5/technical-blog/integrating-apache-spark-with-databricks-unity-catalog-assets/ba-p/97533I have attempted to reproduce this with Azure Databricks, and ADLS gen2 as the storage backend.Although I'm ab...

  • 674 Views
  • 10 replies
  • 3 kudos
Latest Reply
charl-p-botha
New Contributor III
  • 3 kudos

Thanks @dkushari I looked at the github issue you posted, but it has to do specifically with DELTA_UNSUPPORTED_SCHEMA_DURING_READ when streaming *from* a delta table.The specific error I'm seeing is a key error for the Azure storage account hosting t...

  • 3 kudos
9 More Replies
gyorgyjelinek
by New Contributor II
  • 220 Views
  • 0 replies
  • 0 kudos

How to calculate accurate usage cost for a longer contractual period?

Hi Experts!I work on providing and accurate total cost (in DBU and USD as well) calculation for my team for the whole ongoing contractual period. I'v checked the following four options:Account console: Manage account - Usage - Consumption (Legacy): t...

  • 220 Views
  • 0 replies
  • 0 kudos
kevindenham
by New Contributor
  • 239 Views
  • 1 replies
  • 0 kudos

Python User Input During Run-time

I'm new to Advana and was putting together a Python script that requires user interactions during run-time. However, the program pauses at an 'input()' command without a console cell to accept responses. Am I missing something about this instance of ...

  • 239 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

This is a common issue because Jupyter Notebooks are not designed to handle interactive input in the same way as a standard Python script run in a terminal. In Jupyter Notebooks, the input() function does not work as expected because the notebook int...

  • 0 kudos
MarcusC
by New Contributor III
  • 399 Views
  • 2 replies
  • 2 kudos

Databricks Apps

I've created a Databricks App that is essentially a REST Api in Flask. It returns a json object. Data is in sql-warehouse. It works when calling directly in browser but I want to access it with rest api and Bearer token. But when I login with a servi...

  • 399 Views
  • 2 replies
  • 2 kudos
Latest Reply
Walter_C
Databricks Employee
  • 2 kudos

Does the Service Principal has permission access to all the resources involved within you app?

  • 2 kudos
1 More Replies
AnkurMittal008
by New Contributor III
  • 166 Views
  • 1 replies
  • 3 kudos

Disable Catalog for predictive optimization

Let's we disabled predictive optimization for a specific Catalog name "CatalogXYZ" and after that enabled the predictive optimization at Account level. After that can schema owner for the Schema "CatalogXYZ.TestSchema" can enable the predictive optim...

  • 166 Views
  • 1 replies
  • 3 kudos
Latest Reply
Walter_C
Databricks Employee
  • 3 kudos

If predictive optimization is disabled for the catalog "CatalogXYZ" and then enabled at the account level, the schema owner for "CatalogXYZ.TestSchema" cannot enable predictive optimization for this schema. This is because the predictive optimization...

  • 3 kudos
Phani1
by Valued Contributor II
  • 203 Views
  • 1 replies
  • 0 kudos

Downstream usage control on Serverless

Hi All,We've noticed a significant increase in our Databricks  Serverless usage due to downstream system activity. We would like to reduce overall consumption by serverless. Please suggest us the possible ways and  best practices we can implement to ...

  • 203 Views
  • 1 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hello @Phani1, You might want to review this document: https://docs.databricks.com/en/compute/serverless/best-practices.html Let me know if you have any further question.

  • 0 kudos
amberleong
by New Contributor
  • 183 Views
  • 1 replies
  • 0 kudos

How to only allow one git branch, one folder?

Users are able to switch branches in the git UI. How to restrict or only allow one branch?Also, for the sparse checkout, how to view only one folder (without files from root)

  • 183 Views
  • 1 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hello @amberleong, To restrict users from switching branches in the Git UI and only allow one branch, you can implement branch protection rules in your Git repository, directly from your source code git tool.

  • 0 kudos
dtb_usr
by New Contributor II
  • 523 Views
  • 2 replies
  • 0 kudos

Okta SSO Unified login in GCP

Hi,There are versions of this question posted already but they seem to refer to legacy features. Our organisation uses google workspace IP provisioned via Okta as the first landing point and all apps are secured behind this. We have purchased Databri...

  • 523 Views
  • 2 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hello @dtb_usr, It is possible to use OKTA IdP to log into Databricks in GCP, please refer to: https://docs.gcp.databricks.com/en/admin/users-groups/scim/okta.html

  • 0 kudos
1 More Replies
as-emp
by New Contributor
  • 661 Views
  • 1 replies
  • 0 kudos

Datadog, OpenTelemetry, and Databricks container service

We have successfully gotten Datadog agent(s) installed and running on databricks clusters via init script - this part seems to be working fine. We are working on instrumenting our jobs using the OpenTelemetry endpoint feature of the Datadog agent, wh...

  • 661 Views
  • 1 replies
  • 0 kudos
Latest Reply
arjun_kr
Databricks Employee
  • 0 kudos

The agent installations via the init script would install the agents in the Spark containers (All user workloads + spark processes run in the container). The users don't have direct access to the host machine and can't install any agents. You may nee...

  • 0 kudos
Ravichcv85
by New Contributor
  • 252 Views
  • 1 replies
  • 0 kudos

Community edition login

Hi Am not able to login to community edition, its saying not a member , can someone please help?

  • 252 Views
  • 1 replies
  • 0 kudos
Latest Reply
Rjdudley
Contributor II
  • 0 kudos

I'll ask the dumb question first--did you sign up for it?  Although both Databricks Community and Databricks Community Cloud Edition have similar names and are run by Databricks, they do not share a login.  You need to register separately for each.

  • 0 kudos
AnkurMittal008
by New Contributor III
  • 445 Views
  • 2 replies
  • 1 kudos

Resolved! Enable Predictive optimization

 In case to use predictive optimization we should first enable this at account level ? If this is the case then by doing this each of the catalogue/schema/table in Account will start using predictive optimization by default? should we first disable t...

  • 445 Views
  • 2 replies
  • 1 kudos
Latest Reply
AnkurMittal008
New Contributor III
  • 1 kudos

Thanks a lot @SparkJun , In documentation I am not able to find answer to one scenario Let's say we have explicitly disable predictive optimization for a Catalog named "CatalogXYZ" and then after that we have enabled this at Account level. Later a us...

  • 1 kudos
1 More Replies
NelsonE
by New Contributor II
  • 814 Views
  • 2 replies
  • 1 kudos

Terraform Failed to get oauth access token. Please retry after logout and login again. with GCP

Hi I'm having trouble creating a databricks_mws_vpc_endpoint with Terraform.I already created 2 Private Service Connect (PSC) and I'm trying to create the vpc endpoint for Databricks but I'm getting this error:BAD_REQUEST: Failed to get oauth access ...

  • 814 Views
  • 2 replies
  • 1 kudos
Latest Reply
wlam1
New Contributor II
  • 1 kudos

Thank you @NelsonE ! This helped me as well. Tried messing around with all kinds of authentication methods but this was what worked.For the record, I am also using service account impersonation to register VPC endpoints on Terraform / GCP for Databri...

  • 1 kudos
1 More Replies
vieiralaura
by New Contributor
  • 202 Views
  • 0 replies
  • 0 kudos

Help with cookiecutter Prompts in Databricks Notebooks

Hi everyone! I’m working on using cookiecutter to help us set up consistent project templates on Databricks. So far, it’s mostly working, but I’m struggling with the prompts – they’re not displaying well in the Databricks environment.I’ve tried some ...

  • 202 Views
  • 0 replies
  • 0 kudos

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels