- 231 Views
- 4 replies
- 1 kudos
set access control list for job cluster
Hello,currently, we are using an init script that calls the DBW API to add "can_attach_to" permissions for a specific group to initialize the job cluster. How can we set an access control list for job clusters? Is it possible to add it to a policy?
- 231 Views
- 4 replies
- 1 kudos
- 1 kudos
Hi @mkieszek ,You can use REST API to set access control list for job clusters. The endpoint is for doing that you can find below/api/2.0/permissions/clusters/{cluster_id} (https://docs.databricks.com/api/workspace/clusters/setpermissions)And it's n...
- 1 kudos
- 179 Views
- 0 replies
- 5 kudos
Azure Databricks Multi Tenant Solution
Hello Everyone,For the past few months, we’ve been extensively exploring the use of Databricks as the core of our data warehousing product. We provide analytics dashboards to other organizations and are particularly interested in the Column-Level Sec...
- 179 Views
- 0 replies
- 5 kudos
- 124 Views
- 0 replies
- 0 kudos
Issue when connecting to Databricks using OAuth 2.0 token connection string
Following the instructions on Authentication settings for the Databricks ODBC Driver | Databricks on AWS I have created both Personal access token and OAuth 2.0 token connection options in an application. However, I realized that, when I use OAuth 2....
- 124 Views
- 0 replies
- 0 kudos
- 240 Views
- 5 replies
- 2 kudos
Resolved! Databricks bundle custom variables feature does not work for workspace host
Hi all,I am not sure whether this is a bug. I followed the document.My databricks.yml file: bundle: name: test variables: DATABRICKS_HOST: description: The host of the Databricks workspace. DATABRICKS_JOB_RUNNING_SERVICE_PRINCIPAL_CLIENT_I...
- 240 Views
- 5 replies
- 2 kudos
- 2 kudos
Hi @AlbertWang ,The documentation states you cannot map workspace host:
- 2 kudos
- 67 Views
- 0 replies
- 0 kudos
Troubleshooting Cluster
We had a failure on a previously running fact table load (our biggest one) and it looked like an executor was failing due to a timeout error. As a test we upped the cluster size and changed the spark.executor.heartbeatinterval to 300s and the spark....
- 67 Views
- 0 replies
- 0 kudos
- 77 Views
- 0 replies
- 0 kudos
Any Databricks system tables contain info of the saved/pre-defined queries
How can I find the saved/pre-defined queries in Databricks system tables?system.query.history seems NOT having the info, like query-id or query-name
- 77 Views
- 0 replies
- 0 kudos
- 317 Views
- 2 replies
- 1 kudos
Installation of python module nevergrad with R package reticulate fails
I'm trying to run Robyn in Databricks but it fails since the python module nevergrad fails to install properly using the r package reticulate. This code still worked properly about 1,5 months ago so maybe the issue is on some new update in databricks...
- 317 Views
- 2 replies
- 1 kudos
- 1 kudos
My same issue soleve this answer, thank you Kaniz!
- 1 kudos
- 69 Views
- 0 replies
- 0 kudos
samples catalog doesnt have an information schema
we are looking to do an integration with databricks, and i've noticed that the samples database doesn't have an INFORMATION_SCHEMA we rely on the existence of the information_schema to help us understand what views / tables exist in each catalog. w...
- 69 Views
- 0 replies
- 0 kudos
- 1478 Views
- 5 replies
- 0 kudos
In Shared cluster with unity catalog Python process can't connect
The story is that I can access a service listening on port 80 using a single node cluster, but can't do the same using a shared node cluster. I read about the `spark.databricks.pyspark.iptable.outbound.whitelisted.ports` , however, setting that : `sp...
- 1478 Views
- 5 replies
- 0 kudos
- 0 kudos
@6502 you first need to allow the script in the metastore configuration (navgiate to catalog, on the top of the page there's a small cogwheel, click on your catalog name), then:
- 0 kudos
- 225 Views
- 1 replies
- 0 kudos
Resolved! Databricks Asset Bundle deploy a job: Python wheel could not be found
I created a repository for testing DAB, and I'm getting this error: run failed with error message Python wheel with name dab_test_repo could not be found. Please check the driver logs for more details This is my databricks.yml: bundle: name: dab_t...
- 225 Views
- 1 replies
- 0 kudos
- 476 Views
- 1 replies
- 0 kudos
Databricks Azure Cross-Tenant connection to storage account
I'm currently facing a challenge with establishing a cross-tenant connection between Azure Databricks in Tenant A and a Storage Account in Tenant B. Below is the detailed setup of both tenants:| Tenant A | Tenant B || -------- | -------- || Azure Dat...
- 476 Views
- 1 replies
- 0 kudos
- 0 kudos
Have you tried connecting to the storage account from the compute plane itself?I've used this in a notebook to confirm that i can access a storage account over the private endpoint, even when i can't add it as an external data source in the UI.from a...
- 0 kudos
- 197 Views
- 1 replies
- 1 kudos
Connecting to Databricks from Workato. JDBCDriver 500593 Communication Link failure
I'm trying to connect to Databricks from Workato, to pull data in as part of a Workato Recipe. I'm getting the following error when I test the connection:"Database bridge error: Failed since could not connect to the database - Failed to initialize p...
- 197 Views
- 1 replies
- 1 kudos
- 1 kudos
This turned out to be location-related. It looks as though the location of the Workato instance, along with the particular way that our application is integrated with Workato, combine to cause the connection to fail. When I switch to a US host, fro...
- 1 kudos
- 200 Views
- 1 replies
- 1 kudos
Resolved! Can't Link Git and Databricks Community Edition
I am trying to link Git and Databricks but keep running into the error: "Error fetching credentials." I am on the Settings > Linked Accounts tab, and I cannot select anything except the "Retry" button at the bottom of the page, which does not resolve...
- 200 Views
- 1 replies
- 1 kudos
- 1 kudos
Hi @DenJackson ,Community edition it pretty limite and it seems that the option to integrate repository is disabled. If you need more capabalities you can sign up for 14-day free trial.
- 1 kudos
- 424 Views
- 1 replies
- 0 kudos
Alert ACL's not available
The list of available ACL's for alerts suggests there's a way to set "can run" and "can manage" permissions but neither the REST API nor the Databricks CLI show options for setting permissions. Is there a way to set acl's on alerts? If so, how?Alert ...
- 424 Views
- 1 replies
- 0 kudos
- 0 kudos
The API call needed to set the ACLs for an alert and other object types as (alerts | dashboards | data_sources | queries) is:https://docs.databricks.com/api/workspace/dbsqlpermissions/set
- 0 kudos
- 567 Views
- 2 replies
- 4 kudos
Changes in Databricks rest api
Hi all, Just wanted to ask if there is an announcement regarding changes on Databricks REST API? Because I had weird experiences in using the REST API. Specifically for Query History API Last Aug 29, 2024. I created a script to pull get request in t...
- 567 Views
- 2 replies
- 4 kudos
- 4 kudos
Yeah, totally agree with you. It should be documented/mentioned somewhere. Or maybe the API should be versioned, so if they introduce a new version, it won't break existing workflows.Anyway, thanks for sharing!
- 4 kudos
Connect with Databricks Users in Your Area
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group-
Access control
1 -
ActiveDirectory
1 -
AmazonKMS
1 -
Apache spark
1 -
App
1 -
Availability
1 -
Availability Zone
1 -
AWS
5 -
Aws databricks
1 -
AZ
1 -
Azure
8 -
Azure Data Lake Storage
1 -
Azure databricks
6 -
Azure databricks workspace
1 -
Best practice
1 -
Best Practices
2 -
Billing
2 -
Bucket
1 -
Cache
1 -
Change
1 -
Checkpoint
1 -
Checkpoint Path
1 -
Cluster
1 -
Cluster Pools
1 -
Clusters
1 -
ClustersJob
1 -
Compliance
1 -
Compute Instances
1 -
Cost
1 -
Credential passthrough
1 -
Data
1 -
Data Ingestion & connectivity
6 -
Data Plane
1 -
Databricks Account
1 -
Databricks Control Plane
1 -
Databricks Error Message
2 -
Databricks Partner
1 -
Databricks Repos
1 -
Databricks Runtime
1 -
Databricks SQL
3 -
Databricks SQL Dashboard
1 -
Databricks workspace
1 -
DatabricksJobs
1 -
DatabricksLTS
1 -
DBFS
1 -
DBR
3 -
Dbt
1 -
Dbu
3 -
Deep learning
1 -
DeleteTags Permissions
1 -
Delta
4 -
Delta Sharing
1 -
Delta table
1 -
Dev
1 -
Different Instance Types
1 -
Disaster recovery
1 -
DisasterRecoveryPlan
1 -
DLT Pipeline
1 -
EBS
1 -
Email
2 -
External Data Sources
1 -
Feature
1 -
GA
1 -
Ganglia
3 -
Ganglia Metrics
2 -
GangliaMetrics
1 -
GCP
1 -
GCP Support
1 -
Gdpr
1 -
Gpu
2 -
Group Entitlements
1 -
HIPAA
1 -
Hyperopt
1 -
Init script
1 -
InstanceType
1 -
Integrations
1 -
IP Addresses
1 -
IPRange
1 -
Job
1 -
Job Cluster
1 -
Job clusters
1 -
Job Run
1 -
JOBS
1 -
Key
1 -
KMS
1 -
KMSKey
1 -
Lakehouse
1 -
Limit
1 -
Live Table
1 -
Log
2 -
LTS
3 -
Metrics
1 -
MFA
1 -
ML
1 -
Model Serving
1 -
Multiple workspaces
1 -
Notebook Results
1 -
Okta
1 -
On-premises
1 -
Partner
41 -
Pools
1 -
Premium Workspace
1 -
Public Preview
1 -
Redis
1 -
Repos
1 -
Rest API
1 -
Root Bucket
2 -
SCIM API
1 -
Security
1 -
Security Group
1 -
Security Patch
1 -
Service principal
1 -
Service Principals
1 -
Single User Access Permission
1 -
Sns
1 -
Spark
1 -
Spark-submit
1 -
Spot instances
1 -
SQL
1 -
Sql Warehouse
1 -
Sql Warehouse Endpoints
1 -
Ssh
1 -
Sso
2 -
Streaming Data
1 -
Subnet
1 -
Sync Users
1 -
Tags
1 -
Team Members
1 -
Thrift
1 -
TODAY
1 -
Track Costs
1 -
Unity Catalog
1 -
Use
1 -
User
1 -
Version
1 -
Vulnerability Issue
1 -
Welcome Email
1 -
Workspace
2 -
Workspace Access
1
- « Previous
- Next »
User | Count |
---|---|
34 | |
9 | |
8 | |
8 | |
8 |