Hi All,Recipeint encountering an issue while trying to access my organizational data (providers data) in a Delta Sharing scenario(databricks to databricks), and I'm hoping to get some guidance on how to resolve it. Here is the error message recipient...
Hi @Learnit, To resolve this issue, consider running the OPTIMIZE command on the shared Delta table to reduce the number of active files. After optimization, you may encounter the RemoveFiles limit if the OPTIMIZE command removed more than 100K files
We have successfully set up Azure Databricks with standard private link (front-end and back-end).The front-end uses the authentication workspace as prescribed in the documentation.Suppose we use a "custom DNS" for the configuration below. Can we set...
Hi @m997al, Your understanding is correct: The Azure Databricks authentication workspace is unique per Azure region, and its uniqueness extends to each Azure-region-per DNS zone. If you have further questions or need additional clarification, feel fr...
Hi,I am using terraform to create a databricks metastore using this.Now the issue is that I want to use this metastore reference in another terraform project where I create the databricks workspaces but there is no data source for metastore. Can any...
Hi @ducng, When working with Databricks metastores in Terraform, you can create a metastore using the databricks_metastore resource. This metastore acts as a top-level container for objects in Unity Catalog, including data assets like tables and view...
Hello Team,I understand that as the Job Owner, they can grant additional permissions to other users to manage/run/view the job.If "Can Manage" permission is given to the other users, that user can Edit the job including the Run-As parameter to themse...
Hi @VJ3, Let’s delve into the permissions and behaviour of jobs in Databricks when it comes to managing and running them.
“Can Manage” Permission:
When a secondary user is granted the “Can Manage” permission on a job owned by the primary user, th...
Hello Team,We are currently integrating Databricks Rest APIs into our in-house application for managing access permissions. While testing with curl and Postman, we've successfully accessed certain APIs like listing cluster permission. However, we're ...
Hi @Chaitanya07, Dealing with CORS (Cross-Origin Resource Sharing) issues can be a bit tricky, but I’ll provide some guidance to help you resolve this issue when integrating Databricks REST APIs into your in-house application.
Understanding CORS:...
Hi! I have a Job running to process multiple streaming tables. In the beginning, it was working fine, but now I have 80 tables running in this job, the problem is that all the runs are trying to run at the same time throwing an error. Is there a way ...
Hi @Maxi1693, It appears that you’re encountering issues with parallel execution of tasks in your Databricks job.
Let’s address this by considering a few strategies:
Concurrency Limit for Tasks:
Databricks allows a maximum of 1000 concurrent tas...
Hi,I know that with single clusters, theres a single driver node and one driver per cluster. With shared clusters, multiple jobs can run concurrently. Does this still run on a single driver container or multiple driver containers run per application?...
Hi @SirCrayon,
In the context of shared clusters, where multiple jobs can run concurrently, the behavior regarding driver containers differs from that of single clusters.
Single Clusters:
In a single cluster, there is indeed a single driver node...
Hi, I have created a new UC enabled metastore using an Azure storage account and container connected to a Databricks workspace using an access connector. At first glance everything seems to work. I encounter a problem, however, when I try to query UC...
Hi @alm, It appears that you’re encountering issues related to the Hive metastore client when querying UC (Unified Catalog) using a shared access mode cluster in your Databricks workspace.
Let’s troubleshoot this step by step:
Hive Metastore ...
Hi all,In my cluster, some of the PyPI cluster libraries started failing to get installed. It is weird because some of them get installed and some of are constantly failing. In every failed one, the error message is the same (just a package name is d...
I've had this issue myself. What ended up to be the problem is I had windows line endings in my .sh script. You need to convert them to Linux line endings.
Hello, i have been trying to get a pipeline in Azure DevOps to communicate with a Databricks Workspace that uses private link connection. I have tried setting up a service connection using a service principle that is also attached to the workspace, h...
Hi @JJ11 , Integrating Azure DevOps with an Azure Databricks Workspace that uses a private link connection can be a bit tricky, but I’ll guide you through the process.
Here are some steps and suggestions to help you achieve this:
Azure Private Li...
Hi @Kroy, To determine the Databricks Units (DBU) consumption in Azure Databricks, you can follow these steps:
Understanding DBUs:
DBUs represent a unit of processing capability in Azure Databricks.They are billed on a per-second usage basis.The ac...
Able to connect using personal access token and view the data in Hive; but when I publish to dashboard to share-point it results in :- PERMISSION_DENIED: You do not have permissions to autostart 1127-******-*********
Hi @test-fall,
The PERMISSION_DENIED error typically occurs when your user account lacks the necessary permissions to perform a specific action. In your case, it seems related to publishing a dashboard to SharePoint.
Let’s troubleshoot this issue:...
Done, needed to enable the system schema -- Run this command in notebook- to get list of available schema %shcurl -v -X GET -H "Authorization: Bearer <PAT Token>" "https://adb-<xxx>.azuredatabricks.net/api/2.0/unity-catalog/metastores/<metastore-id>...
I would like to programmatically delete some MLflow runs based on a given run id. I am interested in the best practices on how to do this in Databricks workspaces.I know that MLflow cli has gc command which seems quite useful since it also deletes as...
Hello Databricks I wanted to ask a couple questions regarding switching SSO from onelogin to OKTA and turning on user provisioning. We have a total of 4 workspaces ( 1 sandbox , 2 dev and 1 prod) within our account. We have unified login enabled for ...
Hi @RoyRoger711, Let’s break down your questions regarding switching SSO from OneLogin to Okta and enabling user provisioning for Databricks workspaces.
Switching SSO from OneLogin to Okta:
When transitioning from OneLogin to Okta for Single Sign...