- 253 Views
- 1 replies
- 1 kudos
Connecting to Databricks from Workato. JDBCDriver 500593 Communication Link failure
I'm trying to connect to Databricks from Workato, to pull data in as part of a Workato Recipe. I'm getting the following error when I test the connection:"Database bridge error: Failed since could not connect to the database - Failed to initialize p...
- 253 Views
- 1 replies
- 1 kudos
- 1 kudos
This turned out to be location-related. It looks as though the location of the Workato instance, along with the particular way that our application is integrated with Workato, combine to cause the connection to fail. When I switch to a US host, fro...
- 1 kudos
- 287 Views
- 1 replies
- 1 kudos
Resolved! Can't Link Git and Databricks Community Edition
I am trying to link Git and Databricks but keep running into the error: "Error fetching credentials." I am on the Settings > Linked Accounts tab, and I cannot select anything except the "Retry" button at the bottom of the page, which does not resolve...
- 287 Views
- 1 replies
- 1 kudos
- 1 kudos
Hi @DenJackson ,Community edition it pretty limite and it seems that the option to integrate repository is disabled. If you need more capabalities you can sign up for 14-day free trial.
- 1 kudos
- 437 Views
- 1 replies
- 0 kudos
Alert ACL's not available
The list of available ACL's for alerts suggests there's a way to set "can run" and "can manage" permissions but neither the REST API nor the Databricks CLI show options for setting permissions. Is there a way to set acl's on alerts? If so, how?Alert ...
- 437 Views
- 1 replies
- 0 kudos
- 0 kudos
The API call needed to set the ACLs for an alert and other object types as (alerts | dashboards | data_sources | queries) is:https://docs.databricks.com/api/workspace/dbsqlpermissions/set
- 0 kudos
- 594 Views
- 2 replies
- 4 kudos
Changes in Databricks rest api
Hi all, Just wanted to ask if there is an announcement regarding changes on Databricks REST API? Because I had weird experiences in using the REST API. Specifically for Query History API Last Aug 29, 2024. I created a script to pull get request in t...
- 594 Views
- 2 replies
- 4 kudos
- 4 kudos
Yeah, totally agree with you. It should be documented/mentioned somewhere. Or maybe the API should be versioned, so if they introduce a new version, it won't break existing workflows.Anyway, thanks for sharing!
- 4 kudos
- 228 Views
- 2 replies
- 0 kudos
Removing Admin Users
Hello, I am trying to deal with something I thought would be straightforward, but am hitting some walls. Basically the original user associated with my Databricks account has left my organisation and, when trying to remove / disable this user, I am m...
- 228 Views
- 2 replies
- 0 kudos
- 0 kudos
Just to follow-up; I can administer & disable OTHER account admins and do the needful in every other case, but this one user account is particularly privileged.
- 0 kudos
- 375 Views
- 2 replies
- 0 kudos
CLI jobs create issue
I am trying to create a job using the CLI and getting the below error attachedThe command itself was grabbed from the UI so I know it is ok
- 375 Views
- 2 replies
- 0 kudos
- 0 kudos
Hi @ragonzalez ,You can try to escape quotation marks and see if that helps. Something similar to below:{ \"custom_tags.test_tag\": { \"type\": \"fixed\", \"value\": \"test_value\" } }\n"And you don't have to provide whole request body to command lin...
- 0 kudos
- 183 Views
- 2 replies
- 1 kudos
No possibility to schedule DLT once per minute
HelloI wanted to setup DLT to run every minute, previously it was possible with setting up in JSON "schedule": "1 * * * *" - but now I see that is not accepted - is there any other workaround to make it working ?
- 183 Views
- 2 replies
- 1 kudos
- 1 kudos
in that case we can attach dlt to job ... but in json for dlt - I was not able to do that still
- 1 kudos
- 5708 Views
- 23 replies
- 22 kudos
Resolved! Unable to login to Azure Databricks Account Console
I have a personal Azure pay-as-you-go subscription in which I have the 'Global Administrator' role. I am also the databricks account administrator.Until two weeks ago, I was able to access the databricks account console without any issues, but I am f...
- 5708 Views
- 23 replies
- 22 kudos
- 22 kudos
Thanks @dustinvannoy your solution worked for me. I had posted my issue in another post I will refer to this article if anyone is having the same issue.CheersPJ
- 22 kudos
- 258 Views
- 1 replies
- 0 kudos
Databricks Workspace deleted - Need to recover
Hi ,Have created Databricks workspace on GCP. It was somehow deleted which Iam not sure . Is there anyway to recover database.It was deleted on 30-Jun-2025. Is there any retention period to recover.Thanks,Balbeer
- 258 Views
- 1 replies
- 0 kudos
- 515 Views
- 0 replies
- 0 kudos
We are getting AttributeError: module 'numpy.typing' has no attribute 'NDArray' with tidal jobs
File "/databricks/python/lib/python3.8/site-packages/PIL/_typing.py", line 10, in <module>NumpyArray = npt.NDArray[Any]AttributeError: module 'numpy.typing' has no attribute 'NDArray' we are using 10.4 runtime version , please suggest
- 515 Views
- 0 replies
- 0 kudos
- 3284 Views
- 3 replies
- 0 kudos
Resolved! Notebook cells stuck on "waiting to run" when using Cluster Libraries
Hey,we're observing the following problem when trying to run a notebook on a cluster with libraries installed:All notebook cells are stuck in "Waiting to run" (also ones containing only a '1+1' statement).When running the cluster without installing t...
- 3284 Views
- 3 replies
- 0 kudos
- 0 kudos
Did you manage to fix this issue? I am facing a similar situation while running a notebook to read the XML files from the storage account.
- 0 kudos
- 358 Views
- 3 replies
- 2 kudos
Notebook is stuck and cluster goes into waiting state while using spark libraries
Hey,We have installed the com.databricks:spark-xml_2.12:0.18.0 library in our VNET-injected Databricks workspace to read XML files from a storage account. The notebook runs successfully for text files when the cluster is started without the library i...
- 358 Views
- 3 replies
- 2 kudos
- 2 kudos
Since it's a maven dependency it should be simply HTTP and port 80/443.Besides, are you aware that native XML support is included since runtime 14.3? This replaces the library spark-xml.
- 2 kudos
- 381 Views
- 1 replies
- 1 kudos
Resolved! Do Databricks Update the Default Python Libraries in Cluster Runtimes?
Hi all,I’ve been trying to find information about whether Databricks regularly updates the default Python libraries in their cluster runtimes. I checked two different sources but didn’t find clear details.Default python libraries in runtime 11.3 LTS ...
- 381 Views
- 1 replies
- 1 kudos
- 1 kudos
Yes, Databricks does that when releasing new versions of the runtime. Just compare the libraries of the other runtimes.
- 1 kudos
- 175 Views
- 0 replies
- 0 kudos
jar cluster install invalid authority
Hi allFirst time user - hope this is the right spot.My goal is to import and export to an onPrem Oracle database. To this end, I am attempting to upload the ojdbc8.jar to the cluster. This does not, however, go well. The error is:The error text: "Lib...
- 175 Views
- 0 replies
- 0 kudos
- 555 Views
- 4 replies
- 0 kudos
Resolved! NAT Gateway with Azure Databricks
Hi all, what are the steps to enable external communication through a NAT Gateway? Our Databricks instance was created through VNET injection and did not have a NAT gateway created by default. We now want to pass external traffic through NAT gateway ...
- 555 Views
- 4 replies
- 0 kudos
- 0 kudos
I made the changes and it works! I was able to communicate to the NAT Gateway IP in our external on-prem application.I noticed though that Databricks instance is not showing the NAT Gateway within Parameters. Is this an issue?
- 0 kudos
Connect with Databricks Users in Your Area
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group-
Access control
1 -
ActiveDirectory
1 -
AmazonKMS
1 -
Apache spark
1 -
App
1 -
Availability
1 -
Availability Zone
1 -
AWS
5 -
Aws databricks
1 -
AZ
1 -
Azure
8 -
Azure Data Lake Storage
1 -
Azure databricks
6 -
Azure databricks workspace
1 -
Best practice
1 -
Best Practices
2 -
Billing
2 -
Bucket
1 -
Cache
1 -
Change
1 -
Checkpoint
1 -
Checkpoint Path
1 -
Cluster
1 -
Cluster Pools
1 -
Clusters
1 -
ClustersJob
1 -
Compliance
1 -
Compute Instances
1 -
Cost
1 -
Credential passthrough
1 -
Data
1 -
Data Ingestion & connectivity
6 -
Data Plane
1 -
Databricks Account
1 -
Databricks Control Plane
1 -
Databricks Error Message
2 -
Databricks Partner
1 -
Databricks Repos
1 -
Databricks Runtime
1 -
Databricks SQL
3 -
Databricks SQL Dashboard
1 -
Databricks workspace
1 -
DatabricksJobs
1 -
DatabricksLTS
1 -
DBFS
1 -
DBR
3 -
Dbt
1 -
Dbu
3 -
Deep learning
1 -
DeleteTags Permissions
1 -
Delta
4 -
Delta Sharing
1 -
Delta table
1 -
Dev
1 -
Different Instance Types
1 -
Disaster recovery
1 -
DisasterRecoveryPlan
1 -
DLT Pipeline
1 -
EBS
1 -
Email
2 -
External Data Sources
1 -
Feature
1 -
GA
1 -
Ganglia
3 -
Ganglia Metrics
2 -
GangliaMetrics
1 -
GCP
1 -
GCP Support
1 -
Gdpr
1 -
Gpu
2 -
Group Entitlements
1 -
HIPAA
1 -
Hyperopt
1 -
Init script
1 -
InstanceType
1 -
Integrations
1 -
IP Addresses
1 -
IPRange
1 -
Job
1 -
Job Cluster
1 -
Job clusters
1 -
Job Run
1 -
JOBS
1 -
Key
1 -
KMS
1 -
KMSKey
1 -
Lakehouse
1 -
Limit
1 -
Live Table
1 -
Log
2 -
LTS
3 -
Metrics
1 -
MFA
1 -
ML
1 -
Model Serving
1 -
Multiple workspaces
1 -
Notebook Results
1 -
Okta
1 -
On-premises
1 -
Partner
43 -
Pools
1 -
Premium Workspace
1 -
Public Preview
1 -
Redis
1 -
Repos
1 -
Rest API
1 -
Root Bucket
2 -
SCIM API
1 -
Security
1 -
Security Group
1 -
Security Patch
1 -
Service principal
1 -
Service Principals
1 -
Single User Access Permission
1 -
Sns
1 -
Spark
1 -
Spark-submit
1 -
Spot instances
1 -
SQL
1 -
Sql Warehouse
1 -
Sql Warehouse Endpoints
1 -
Ssh
1 -
Sso
2 -
Streaming Data
1 -
Subnet
1 -
Sync Users
1 -
Tags
1 -
Team Members
1 -
Thrift
1 -
TODAY
1 -
Track Costs
1 -
Unity Catalog
1 -
Use
1 -
User
1 -
Version
1 -
Vulnerability Issue
1 -
Welcome Email
1 -
Workspace
2 -
Workspace Access
1
- « Previous
- Next »
User | Count |
---|---|
36 | |
9 | |
9 | |
8 | |
8 |