cancel
Showing results for 
Search instead for 
Did you mean: 
Administration & Architecture
Explore discussions on Databricks administration, deployment strategies, and architectural best practices. Connect with administrators and architects to optimize your Databricks environment for performance, scalability, and security.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

mkieszek
by New Contributor II
  • 1947 Views
  • 4 replies
  • 1 kudos

set access control list for job cluster

Hello,currently, we are using an init script that calls the DBW API to add "can_attach_to" permissions for a specific group to initialize the job cluster. How can we set an access control list for job clusters? Is it possible to add it to a policy?

  • 1947 Views
  • 4 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @mkieszek ,You can use REST API to set access control list for job clusters. The endpoint is for doing that you can find below/api/2.0/permissions/clusters/{cluster_id}  (https://docs.databricks.com/api/workspace/clusters/setpermissions)And it's n...

  • 1 kudos
3 More Replies
AlbertWang
by Valued Contributor
  • 2042 Views
  • 5 replies
  • 2 kudos

Resolved! Databricks bundle custom variables feature does not work for workspace host

Hi all,I am not sure whether this is a bug. I followed the document.My databricks.yml file: bundle: name: test variables: DATABRICKS_HOST: description: The host of the Databricks workspace. DATABRICKS_JOB_RUNNING_SERVICE_PRINCIPAL_CLIENT_I...

  • 2042 Views
  • 5 replies
  • 2 kudos
Latest Reply
filipniziol
Esteemed Contributor
  • 2 kudos

Hi @AlbertWang ,The documentation states you cannot map workspace host: 

  • 2 kudos
4 More Replies
Janpie
by New Contributor
  • 1008 Views
  • 1 replies
  • 0 kudos

Installation of python module nevergrad with R package reticulate fails

I'm trying to run Robyn in Databricks but it fails since the python module nevergrad fails to install properly using the r package reticulate. This code still worked properly about 1,5 months ago so maybe the issue is on some new update in databricks...

  • 1008 Views
  • 1 replies
  • 0 kudos
Latest Reply
macromill
New Contributor II
  • 0 kudos

My same issue soleve this answer, thank you Kaniz!

  • 0 kudos
6502
by New Contributor III
  • 3563 Views
  • 5 replies
  • 0 kudos

In Shared cluster with unity catalog Python process can't connect

The story is that I can access a service listening on port 80 using a single node cluster, but can't do the same using a shared node cluster. I read about the `spark.databricks.pyspark.iptable.outbound.whitelisted.ports` , however, setting that : `sp...

  • 3563 Views
  • 5 replies
  • 0 kudos
Latest Reply
adriennn
Valued Contributor
  • 0 kudos

@6502 you first need to allow the script in the metastore configuration (navgiate to catalog, on the top of the page there's a small cogwheel, click on your catalog name), then:

  • 0 kudos
4 More Replies
PabloCSD
by Valued Contributor II
  • 1698 Views
  • 1 replies
  • 0 kudos

Resolved! Databricks Asset Bundle deploy a job: Python wheel could not be found

I created a repository for testing DAB, and I'm getting this error: run failed with error message Python wheel with name dab_test_repo could not be found. Please check the driver logs for more details This is my databricks.yml: bundle: name: dab_t...

  • 1698 Views
  • 1 replies
  • 0 kudos
Latest Reply
PabloCSD
Valued Contributor II
  • 0 kudos

Fixed it:Uncommenting out: # libraries: # - whl: ./dist/*.whl 

  • 0 kudos
Behwar
by New Contributor III
  • 2996 Views
  • 1 replies
  • 0 kudos

Databricks Azure Cross-Tenant connection to storage account

I'm currently facing a challenge with establishing a cross-tenant connection between Azure Databricks in Tenant A and a Storage Account in Tenant B. Below is the detailed setup of both tenants:| Tenant A | Tenant B || -------- | -------- || Azure Dat...

  • 2996 Views
  • 1 replies
  • 0 kudos
Latest Reply
Kaib
New Contributor II
  • 0 kudos

Have you tried connecting to the storage account from the compute plane itself?I've used this in a notebook to confirm that i can access a storage account over the private endpoint, even when i can't add it as an external data source in the UI.from a...

  • 0 kudos
DenJackson
by New Contributor II
  • 1327 Views
  • 1 replies
  • 1 kudos

Resolved! Can't Link Git and Databricks Community Edition

I am trying to link Git and Databricks but keep running into the error: "Error fetching credentials." I am on the Settings > Linked Accounts tab, and I cannot select anything except the "Retry" button at the bottom of the page, which does not resolve...

  • 1327 Views
  • 1 replies
  • 1 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 1 kudos

Hi @DenJackson ,Community edition it pretty limite and it seems that the option to integrate repository is disabled. If you need more capabalities you can sign up for 14-day free trial. 

  • 1 kudos
eric-c
by New Contributor
  • 1105 Views
  • 1 replies
  • 0 kudos

Alert ACL's not available

The list of available ACL's for alerts suggests there's a way to set "can run" and "can manage" permissions but neither the REST API nor the Databricks CLI show options for setting permissions. Is there a way to set acl's on alerts? If so, how?Alert ...

  • 1105 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

The API call needed to set the ACLs for an alert and other object types as (alerts | dashboards | data_sources | queries) is:https://docs.databricks.com/api/workspace/dbsqlpermissions/set

  • 0 kudos
Databrixer
by New Contributor II
  • 1499 Views
  • 2 replies
  • 4 kudos

Changes in Databricks rest api

Hi all, Just wanted to ask if there is an announcement regarding changes on Databricks REST API?  Because I had weird experiences in using the REST API. Specifically for Query History API Last Aug 29, 2024. I created a script to pull get request in t...

  • 1499 Views
  • 2 replies
  • 4 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 4 kudos

Yeah, totally agree with you. It should be documented/mentioned somewhere. Or maybe the API should be versioned, so if they introduce a new version, it won't break existing workflows.Anyway, thanks for sharing!

  • 4 kudos
1 More Replies
jacobholland
by New Contributor II
  • 1301 Views
  • 2 replies
  • 0 kudos

Removing Admin Users

Hello, I am trying to deal with something I thought would be straightforward, but am hitting some walls. Basically the original user associated with my Databricks account has left my organisation and, when trying to remove / disable this user, I am m...

  • 1301 Views
  • 2 replies
  • 0 kudos
Latest Reply
jacobholland
New Contributor II
  • 0 kudos

Just to follow-up; I can administer & disable OTHER account admins and do the needful in every other case, but this one user account is particularly privileged.

  • 0 kudos
1 More Replies
ragonzalez
by New Contributor
  • 1073 Views
  • 2 replies
  • 0 kudos

CLI jobs create issue

I am trying to create a job using the CLI and getting the below error attachedThe command itself was grabbed from the UI so I know it is ok 

ragonzalez_0-1726182063032.png
  • 1073 Views
  • 2 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @ragonzalez ,You can try to escape quotation marks and see if that helps. Something similar to below:{ \"custom_tags.test_tag\": { \"type\": \"fixed\", \"value\": \"test_value\" } }\n"And you don't have to provide whole request body to command lin...

  • 0 kudos
1 More Replies
PiotrU
by Contributor II
  • 1099 Views
  • 2 replies
  • 1 kudos

No possibility to schedule DLT once per minute

HelloI wanted to setup DLT to run every minute, previously it was possible with setting up in JSON "schedule": "1 * * * *"  - but now I see that is not accepted - is there any other workaround to make it working ?

  • 1099 Views
  • 2 replies
  • 1 kudos
Latest Reply
PiotrU
Contributor II
  • 1 kudos

in that case we can attach dlt to job ... but in json for dlt - I was not able to do that still

  • 1 kudos
1 More Replies
bubbybedi
by New Contributor II
  • 791 Views
  • 1 replies
  • 0 kudos

Databricks Workspace deleted - Need to recover

Hi ,Have created Databricks workspace on GCP. It was somehow deleted which Iam not sure . Is there anyway to recover database.It was deleted on 30-Jun-2025. Is there any retention period to recover.Thanks,Balbeer

  • 791 Views
  • 1 replies
  • 0 kudos
Latest Reply
bubbybedi
New Contributor II
  • 0 kudos

sorry typo 30-jun-2024

  • 0 kudos
jannikj
by New Contributor III
  • 6045 Views
  • 3 replies
  • 0 kudos

Resolved! Notebook cells stuck on "waiting to run" when using Cluster Libraries

Hey,we're observing the following problem when trying to run a notebook on a cluster with libraries installed:All notebook cells are stuck in "Waiting to run" (also ones containing only a '1+1' statement).When running the cluster without installing t...

  • 6045 Views
  • 3 replies
  • 0 kudos
Latest Reply
BhawaniD
New Contributor II
  • 0 kudos

Did you manage to fix this issue? I am facing a similar situation while running a notebook to read the XML files from the storage account. 

  • 0 kudos
2 More Replies
BhawaniD
by New Contributor II
  • 1584 Views
  • 3 replies
  • 2 kudos

Notebook is stuck and cluster goes into waiting state while using spark libraries

Hey,We have installed the com.databricks:spark-xml_2.12:0.18.0 library in our VNET-injected Databricks workspace to read XML files from a storage account. The notebook runs successfully for text files when the cluster is started without the library i...

  • 1584 Views
  • 3 replies
  • 2 kudos
Latest Reply
Witold
Honored Contributor
  • 2 kudos

Since it's a maven dependency it should be simply HTTP and port 80/443.Besides, are you aware that native XML support is included since runtime 14.3? This replaces the library spark-xml.

  • 2 kudos
2 More Replies