cancel
Showing results for 
Search instead for 
Did you mean: 
Administration & Architecture
Explore discussions on Databricks administration, deployment strategies, and architectural best practices. Connect with administrators and architects to optimize your Databricks environment for performance, scalability, and security.
cancel
Showing results for 
Search instead for 
Did you mean: 
Data + AI Summit 2024 - Data Lakehouse Architecture


Forum Posts

Phani1
by Valued Contributor II
  • 175 Views
  • 1 replies
  • 0 kudos

Downstream usage control on Serverless

Hi All,We've noticed a significant increase in our Databricks  Serverless usage due to downstream system activity. We would like to reduce overall consumption by serverless. Please suggest us the possible ways and  best practices we can implement to ...

  • 175 Views
  • 1 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hello @Phani1, You might want to review this document: https://docs.databricks.com/en/compute/serverless/best-practices.html Let me know if you have any further question.

  • 0 kudos
amberleong
by New Contributor
  • 147 Views
  • 1 replies
  • 0 kudos

How to only allow one git branch, one folder?

Users are able to switch branches in the git UI. How to restrict or only allow one branch?Also, for the sparse checkout, how to view only one folder (without files from root)

  • 147 Views
  • 1 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hello @amberleong, To restrict users from switching branches in the Git UI and only allow one branch, you can implement branch protection rules in your Git repository, directly from your source code git tool.

  • 0 kudos
dtb_usr
by New Contributor II
  • 474 Views
  • 2 replies
  • 0 kudos

Okta SSO Unified login in GCP

Hi,There are versions of this question posted already but they seem to refer to legacy features. Our organisation uses google workspace IP provisioned via Okta as the first landing point and all apps are secured behind this. We have purchased Databri...

  • 474 Views
  • 2 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hello @dtb_usr, It is possible to use OKTA IdP to log into Databricks in GCP, please refer to: https://docs.gcp.databricks.com/en/admin/users-groups/scim/okta.html

  • 0 kudos
1 More Replies
as-emp
by New Contributor
  • 457 Views
  • 1 replies
  • 0 kudos

Datadog, OpenTelemetry, and Databricks container service

We have successfully gotten Datadog agent(s) installed and running on databricks clusters via init script - this part seems to be working fine. We are working on instrumenting our jobs using the OpenTelemetry endpoint feature of the Datadog agent, wh...

  • 457 Views
  • 1 replies
  • 0 kudos
Latest Reply
arjun_kr
Databricks Employee
  • 0 kudos

The agent installations via the init script would install the agents in the Spark containers (All user workloads + spark processes run in the container). The users don't have direct access to the host machine and can't install any agents. You may nee...

  • 0 kudos
LMe
by New Contributor
  • 190 Views
  • 3 replies
  • 0 kudos

Get a static IP for my Databricks App

Hello,I'm trying to find how to set-up a static IP for a Azure Databricks App. I tried to set-up a NAT gateway to have a static IP for the workspace, but it doesn't change anything, I still can't access my OpenAI ressource even if I authorize the NaT...

  • 190 Views
  • 3 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

I found the following guide which might be what you are actually looking to achieve https://docs.databricks.com/en/archive/admin-guide/nat-gateway-proxying.html   

  • 0 kudos
2 More Replies
Ravichcv85
by New Contributor
  • 227 Views
  • 1 replies
  • 0 kudos

Community edition login

Hi Am not able to login to community edition, its saying not a member , can someone please help?

  • 227 Views
  • 1 replies
  • 0 kudos
Latest Reply
Rjdudley
Contributor II
  • 0 kudos

I'll ask the dumb question first--did you sign up for it?  Although both Databricks Community and Databricks Community Cloud Edition have similar names and are run by Databricks, they do not share a login.  You need to register separately for each.

  • 0 kudos
AnkurMittal008
by New Contributor III
  • 383 Views
  • 2 replies
  • 1 kudos

Resolved! Enable Predictive optimization

 In case to use predictive optimization we should first enable this at account level ? If this is the case then by doing this each of the catalogue/schema/table in Account will start using predictive optimization by default? should we first disable t...

  • 383 Views
  • 2 replies
  • 1 kudos
Latest Reply
AnkurMittal008
New Contributor III
  • 1 kudos

Thanks a lot @SparkJun , In documentation I am not able to find answer to one scenario Let's say we have explicitly disable predictive optimization for a Catalog named "CatalogXYZ" and then after that we have enabled this at Account level. Later a us...

  • 1 kudos
1 More Replies
NelsonE
by New Contributor II
  • 721 Views
  • 2 replies
  • 1 kudos

Terraform Failed to get oauth access token. Please retry after logout and login again. with GCP

Hi I'm having trouble creating a databricks_mws_vpc_endpoint with Terraform.I already created 2 Private Service Connect (PSC) and I'm trying to create the vpc endpoint for Databricks but I'm getting this error:BAD_REQUEST: Failed to get oauth access ...

  • 721 Views
  • 2 replies
  • 1 kudos
Latest Reply
wlam1
New Contributor II
  • 1 kudos

Thank you @NelsonE ! This helped me as well. Tried messing around with all kinds of authentication methods but this was what worked.For the record, I am also using service account impersonation to register VPC endpoints on Terraform / GCP for Databri...

  • 1 kudos
1 More Replies
vieiralaura
by New Contributor
  • 167 Views
  • 0 replies
  • 0 kudos

Help with cookiecutter Prompts in Databricks Notebooks

Hi everyone! I’m working on using cookiecutter to help us set up consistent project templates on Databricks. So far, it’s mostly working, but I’m struggling with the prompts – they’re not displaying well in the Databricks environment.I’ve tried some ...

  • 167 Views
  • 0 replies
  • 0 kudos
takak
by New Contributor II
  • 271 Views
  • 3 replies
  • 1 kudos

Cannot downgrade workspace object permissions using API

Hi!I'd like to restrict some users' permissions using REST API and got an issue while trying to update a permission on 'directories'.I'm trying to set a user's permission on their default username folder in the workspace to 'can edit' so that they ca...

  • 271 Views
  • 3 replies
  • 1 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 1 kudos

Hi @takak, Greetings from Databricks! What is the REST API you are making the call to? Looks like this might not be supported programmatically, but will try to test it internally. it appears that the CAN_MANAGE permission is a higher-level permission...

  • 1 kudos
2 More Replies
debal
by New Contributor
  • 200 Views
  • 1 replies
  • 0 kudos

Terraform Databricks Integration - specially for Unity Catalog in AWS S3

We are attempting to provision Unity Catalog using Terraform, but we're encountering issues with establishing authentication with AWS through IAM Roles and Policies.For EC2/Cluster instances, the instance profile works fine with a trust relationship ...

  • 200 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

To provision Unity Catalog using Terraform and authenticate with AWS through IAM Roles and Policies, you'll need to follow these steps: Create an IAM Role for Unity Catalog: First, you need to create an IAM role that Unity Catalog can assume to acces...

  • 0 kudos
Kutbuddin
by New Contributor II
  • 278 Views
  • 2 replies
  • 0 kudos

[INTERNAL_ERROR] Query could not be scheduled: HTTP Response code: 503. Please try again later

We have a databricks job configured to run a dbt project. The dbt cli compute cluster being used is serverless with a serverless sql warehouse. We encountered this error during a run. SQLSTATE: XX000Any idea why this occurred?

  • 278 Views
  • 2 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hello @Kutbuddin, The error you encountered, SQLSTATE: XX000 [INTERNAL_ERROR] Query could not be scheduled: HTTP Response code: 503. Please try again later, typically indicates an issue with the scheduling of the query due to a temporary unavailabili...

  • 0 kudos
1 More Replies
rpl
by Contributor
  • 669 Views
  • 8 replies
  • 2 kudos

Resolved! Which API to use to list groups in which a given user is a member

Is there an API that can be used to list groups in which a given user is a member? Specifically, I’d be interested in account (not workspace) groups.It seems there used to be a workspace-level list-parents API referred to in the answers to this quest...

  • 669 Views
  • 8 replies
  • 2 kudos
Latest Reply
rpl
Contributor
  • 2 kudos

@Alberto_Umana, do i understand correctly that there isn’t currently an endpoint that returns the info I need? May I request this feature be added on the roadmap? A bit of context: we have tens of catalogs in Unity Catalog and several workspaces in o...

  • 2 kudos
7 More Replies
k1t3k
by New Contributor II
  • 387 Views
  • 2 replies
  • 0 kudos

Global init script fails on Databricks 16.0

#!/bin/bashpip install package1 --index-url https://link-to-indexpip install package2 --index-url https://link-to-indexThis init script fails witherror: externally-managed-environment× This environment is externally managed╰─> To install Python packa...

  • 387 Views
  • 2 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hi @k1t3k, Are you installing a custom package? could you please share the package name you are installing to validate? The error you are encountering, "externally-managed-environment," when running your global init script with Databricks Runtime 16....

  • 0 kudos
1 More Replies
MDV
by New Contributor III
  • 452 Views
  • 2 replies
  • 1 kudos

Resolved! Compute configuration : single user with service principal of azure data datafactory ?

Is it possible to have the service principal (ID) of an Azure data factory as the Single user access on an databricks cluster ?Reason I'm asking is because we are starting to use unity catalog , but would still have the need to execute stored procedu...

MDV_0-1730973817325.png
  • 452 Views
  • 2 replies
  • 1 kudos
Latest Reply
BigRoux
Databricks Employee
  • 1 kudos

Yes, this is possible. First, create a new service principal in Azure or use an existing one. This could be either a managed identity from Azure Data Factory or a manually created service principal in Microsoft Entra ID (formerly Azure AD). Next, in ...

  • 1 kudos
1 More Replies

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels