cancel
Showing results for 
Search instead for 
Did you mean: 
Administration & Architecture
Explore discussions on Databricks administration, deployment strategies, and architectural best practices. Connect with administrators and architects to optimize your Databricks environment for performance, scalability, and security.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

mannepk85
by New Contributor III
  • 1141 Views
  • 1 replies
  • 0 kudos

Resolved! Attach a databricks_instance_pool to databricks_cluster_policy via terraform

Hello Team,I am trying to create a databricks instance pool and attach it to a cluster policy in our terraform code. But I am having hard time finding a good documentation. Has any one done it? Below is my sample code and I am getting errorI keep get...

  • 1141 Views
  • 1 replies
  • 0 kudos
Latest Reply
mannepk85
New Contributor III
  • 0 kudos

Fixed it! "instance_pool_id" : { type = "fixed" values = "databricks_instance_pool.dev_test_cluster_pool.id"}

  • 0 kudos
SanthoshKumarK
by New Contributor II
  • 1269 Views
  • 2 replies
  • 0 kudos

Lakehouse federation -' on behalf of ' queries

Is it possible to achieve the following in a lake-house federation setup using Azure Databricks?1. Establish an external connection (EC1) to an external data source (EDS) using the credentials of user U1.2. Create a foreign catalog (FC1) utilizing EC...

  • 1269 Views
  • 2 replies
  • 0 kudos
Latest Reply
SanthoshKumarK
New Contributor II
  • 0 kudos

Thanks for explaining the authorization flow, @rangu . In the example mentioned, does Databricks support passing the user’s credentials to an external data source? For instance, can it pass the OAuth token for the user along with the externalID crede...

  • 0 kudos
1 More Replies
noorbasha534
by Valued Contributor II
  • 2437 Views
  • 4 replies
  • 0 kudos

Libraries installation governance

Dear allI like to know the best practices around libraries installation on Databricks compute - all-purpose, job.The need is to screen the libraries, conduct vulnerability tests, and then let them be installed through a centralized CI/CD process. How...

  • 2437 Views
  • 4 replies
  • 0 kudos
Latest Reply
noorbasha534
Valued Contributor II
  • 0 kudos

@filipniziol thanks again for your time. The thing is we like to block access to these URLs as at times we found developers & data scientists downloading packages that were marked as vulnerable by Maven.

  • 0 kudos
3 More Replies
VJ5
by New Contributor
  • 1357 Views
  • 2 replies
  • 0 kudos

Azure Databricks Serverless Compute

Hello,Looking for documents related to Azure Databricks Serverless Compute. What are the things we need to consider for security point of view when we decide to use serverless compute? 

  • 1357 Views
  • 2 replies
  • 0 kudos
Latest Reply
David-jono123
New Contributor II
  • 0 kudos

These steps are really helpful. I especially appreciate the reminder to check my credentials and consider browser-related issues, as those are often overlooked. I'll make sure to clear my cache and cookies first, and if that doesn't work, I’ll try us...

  • 0 kudos
1 More Replies
matthiasjg
by New Contributor II
  • 981 Views
  • 1 replies
  • 0 kudos

How to NOT install or disable or uninstall Databricks Delta Live Tables dlt module on jobs cluster?

I need to NOT have the Databricks Delta Live Tables (DLT) Python stub installed on job cluster b/c of naming conflict w/ pip library dlt (and I also don't need delta live tables).There is no "simple" way of uninstalling. It's not installed via pip as...

  • 981 Views
  • 1 replies
  • 0 kudos
Latest Reply
matthiasjg
New Contributor II
  • 0 kudos

For anyone facing a similar problem: I've addressed the issue of the dlt module conflict on my job cluster, by using an init script to remove the dlt module from the cluster's Python environment.Simply by doing:%bash #!/bin/bash rm -rf /databricks/sp...

  • 0 kudos
JessieWen
by Databricks Employee
  • 1149 Views
  • 1 replies
  • 0 kudos

legacy repo error fetching git status files over 200MB

Working directory contains files that exceed the allowed limit of 200 MB.  how to solve this?

  • 1149 Views
  • 1 replies
  • 0 kudos
Latest Reply
filipniziol
Esteemed Contributor
  • 0 kudos

Hi @JessieWen ,What you can do besides removing some files from the repo, is to use "Sparce mode" and select only certain paths to be synchronized with Databricks repos. Hope it helps

  • 0 kudos
harripy
by New Contributor III
  • 7110 Views
  • 8 replies
  • 0 kudos

Databricks SQL connectivity in Python with Service Principals

Tried to use M2M OAuth connectivity on Databricks SQL Warehouse in Python:from databricks.sdk.core import Config, oauth_service_principal from databricks import sql .... config = Config(host=f"https://{host}", client_...

  • 7110 Views
  • 8 replies
  • 0 kudos
Latest Reply
Mat_Conquest
New Contributor II
  • 0 kudos

Did anyone get this to work? I have tried the code above but I get a slightly different error but I don't see the same level of details from the logs2024-10-04 14:59:25,508 [databricks.sdk][DEBUG] Attempting to configure auth: pat2024-10-04 14:59:25,...

  • 0 kudos
7 More Replies
OU_Professor
by New Contributor II
  • 13289 Views
  • 1 replies
  • 0 kudos

Connect Community Edition to Power BI Desktop

I have submitted this question several times to Databricks over the past few weeks, and I have gotten no response at all, not even an acknowledgement that my request was received.Please help.How can I connect a certain dataset in Databricks Community...

  • 13289 Views
  • 1 replies
  • 0 kudos
Latest Reply
Knguyen
New Contributor II
  • 0 kudos

Hi @Retired_mod,It seams the Commnity Edition doesn't let us to generate the personal-access-token any more. Could you let us know some where we can get the token in the Comminity Edition?Thanks.

  • 0 kudos
echiro
by New Contributor II
  • 924 Views
  • 1 replies
  • 0 kudos

cluster administrator

Is individual cluster more cost effective or shared group cluster?

  • 924 Views
  • 1 replies
  • 0 kudos
Latest Reply
rangu
New Contributor III
  • 0 kudos

This is very generic, it depends upon use case. If you have a bunch of users trying to read data from catalogs, and perform data analysis or analytics creating a common cluster will be more cost effective and provided better performance. Also, largel...

  • 0 kudos
JKR
by Contributor
  • 1799 Views
  • 1 replies
  • 0 kudos

How to assign user group for email notification in databricks Alerts

How can I assign a azure databricks user group to an alert for notification?Current scenario is whenever we need to add a user for alert email notification we are manually adding that user email address to each we setup (more than 100) which is very ...

JKR_0-1723550146638.png
  • 1799 Views
  • 1 replies
  • 0 kudos
Latest Reply
rangu
New Contributor III
  • 0 kudos

One option is to handle the logic inside the python notebook to trigger alerts using emali and smtp lib which accepts databricks local groups and AD groups that are synched.

  • 0 kudos
cgrass
by New Contributor III
  • 2002 Views
  • 1 replies
  • 0 kudos

Resolved! Resource organization in a large company

Hello,We are using Azure Databricks in a single tenant. We will have many teams working in multiple (Unity Enabled) Workspaces using a variety of Catalogs, External Locations, Storage Credentials, ect. Some of those resources will be shared (e.g., an...

Administration & Architecture
Architecture
azure
catalogs
design
  • 2002 Views
  • 1 replies
  • 0 kudos
Latest Reply
cgrass
New Contributor III
  • 0 kudos

We are using Azure Databricks in a single tenant. We will have many teams working in multiple (Unity Enabled) Workspaces using a variety of Catalogs, External Locations, Storage Credentials, etc. Some of those resources will be shared (e.g., an Exter...

  • 0 kudos
SunilSamal
by New Contributor II
  • 2866 Views
  • 3 replies
  • 0 kudos

HTTPSConnectionPool(host='sandvik.peakon.com', port=443): Max retries exceeded with url: /api/v1/seg

while connecting to an api from databricks notebook with the bearer token I am getting the below errorHTTPSConnectionPool(host='sandvik.peakon.com', port=443): Max retries exceeded with url: /api/v1/segments?page=1 (Caused by SSLError(SSLCertVerifica...

  • 2866 Views
  • 3 replies
  • 0 kudos
Latest Reply
saikumar246
Databricks Employee
  • 0 kudos

Hi @SunilSamal  The error you are encountering, SSLCertVerificationError, indicates that the SSL certificate verification failed because the local issuer certificate could not be obtained. This is a common issue when the SSL certificate chain is inco...

  • 0 kudos
2 More Replies
FlukeStarbucker
by New Contributor III
  • 2011 Views
  • 4 replies
  • 1 kudos

Resolved! Unable to get S3 connection working

I can't get past the error below. I've read and reread the instructions several times at the URL below and for the life of me cannot figure out what I'm missing in my AWS setup. Any tips on how to track down my issue? https://docs.databricks.com/en/c...

  • 2011 Views
  • 4 replies
  • 1 kudos
Latest Reply
FlukeStarbucker
New Contributor III
  • 1 kudos

I got it working, there was a weird typo where the role ARN was duplicated. Thanks.

  • 1 kudos
3 More Replies
mx2323
by New Contributor
  • 969 Views
  • 1 replies
  • 0 kudos

samples catalog doesnt have an information schema

 we are looking to do an integration with databricks, and i've noticed that the samples database doesn't have an INFORMATION_SCHEMA  we rely on the existence of the information_schema to help us understand what views / tables exist in each catalog. w...

mx2323_0-1726607557549.png
  • 969 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

The "samples" catalog in Databricks does not have an INFORMATION_SCHEMA because it is designed primarily for demonstration and educational purposes, rather than for production use. This schema is typically included in catalogs created on Unity Catalo...

  • 0 kudos
NetRider
by New Contributor
  • 2005 Views
  • 1 replies
  • 0 kudos

Issue when connecting to Databricks using OAuth 2.0 token connection string

Following the instructions on Authentication settings for the Databricks ODBC Driver | Databricks on AWS I have created both Personal access token and OAuth 2.0 token connection options in an application. However, I realized that, when I use OAuth 2....

  • 2005 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

This will require further analysis, based on the information it seems there is no direct way to restrict this. Are you able to submit a support case so this can be analyzed? 

  • 0 kudos