- 4301 Views
- 2 replies
- 3 kudos
Resolved! Databricks Asset Bundles + Artifacts + Poetry
Hello,I've configured the DABs on our project successfully. Moreover, I could switch from setuptools to poetry almost successfully. In the project's databricks.yml I configured it as the documentation suggested, I've just changed the name of the arti...
- 4301 Views
- 2 replies
- 3 kudos
- 3 kudos
Hi @Fiabane ,Could you first check:Do you see your .whl file in your artifacts folder?Could you try to install the package by running the code in your notebook : %pip install <path to your wheel>As far as I understand you want to have a job ...
- 3 kudos
- 1933 Views
- 2 replies
- 0 kudos
Databricks Kryo setup
I would like to consolidate all our Spark jobs in Databricks. One of those jobs that are currently running in Azure HDInsight is not properly working using a Databricks JAR job.It uses Spark 3.3 RDDs and requires configuring Kryo serialisation. There...
- 1933 Views
- 2 replies
- 0 kudos
- 0 kudos
Integrating Spark tasks with Databricks can greatly improve your workflow. For tasks that require Kryo serialization, make sure you configure your Spark session correctly. You may need to adjust the serialization settings in your Spark configuration....
- 0 kudos
- 945 Views
- 1 replies
- 2 kudos
Azure Databricks Classic Compute Plane Firewall
I’m designing a compute plane configuration that will align our data platform with internal policies from a security perspective. As part of this exercise I'm documenting how the permissible traffic inbound and outbound is controlled using NSG rules,...
- 945 Views
- 1 replies
- 2 kudos
- 2 kudos
@Jim-Shady wrote:I’m designing a compute plane configuration that will align our data platform with internal policies from a security perspective. As part of this exercise I'm documenting how the permissible traffic inbound and outbound is controlled...
- 2 kudos
- 5445 Views
- 1 replies
- 0 kudos
Resolved! How to deploy to Databricks Assets Bundle from Azure DevOps using Service Principal?
I have a CI/CD process that after a Pull Request (PR) to main it deploys to staging.It works using a Personal Access Token using Azure Pipelines.From local, deploying using Service Principal works (https://community.databricks.com/t5/administration-a...
- 5445 Views
- 1 replies
- 0 kudos
- 0 kudos
I needed to deploy a job using CI/CD Azure Pipelines without using the OAuth, this is the way:First you need to have configured the Service Principal, for that you need to generate it in your workspace with this you will have:A host: Which is your wo...
- 0 kudos
- 2725 Views
- 5 replies
- 0 kudos
Resolved! Unable to Create Job Task Using Git Provider Invalid Path
I am attempting to create a task in a job using the Git Provider as a source and GitHub is the provider. The repo is a private repo. Regardless of how I enter the path to the notebook I receive the same error that the notebook path is invalid and o...
- 2725 Views
- 5 replies
- 0 kudos
- 0 kudos
Like I said in a previous response. This started working automatically a few days ago with no changes on our end. The developer who was working on this decided to try it one more time and it just worked, no error this time. I don't know if Databri...
- 0 kudos
- 1450 Views
- 1 replies
- 0 kudos
Bring data from databricks to sharepoint list using the Power Automate
Good afternoon to all and I am new to this community.We are trying to bring data from databricks to sharepoint list using the Power Automate app (create workflow and trigger it when there is new record or exising record is modified in source table in...
- 1450 Views
- 1 replies
- 0 kudos
- 0 kudos
Hi all, Can anyone assist me with this request ?Thanks in advance
- 0 kudos
- 915 Views
- 1 replies
- 0 kudos
Tabs for notebooks
Browsing this page of the documentation, the displayed GIF shows a notebook that is opened in its own tab. I've been looking for how to enable this feature in my own workspace, but cannot find it.Does anyone know how to enable this feature?
- 915 Views
- 1 replies
- 0 kudos
- 0 kudos
Nope.It seems that is some kind of new version of the UI.In the SQL editor one can open multiple tabs. But for python notebooks I have no idea.
- 0 kudos
- 1217 Views
- 1 replies
- 0 kudos
How to generate an Azure Subscription from a Databricks Generated Service Principal?
Hello, I currently have a Service Principal (SP) Client_Id and its associated secret, I generated it directly from my workspace in Databricks, i was following this post: https://github.com/databricks/cli/issues/1722, but I don't know how to generate ...
- 1217 Views
- 1 replies
- 0 kudos
- 0 kudos
Learn to summon an Azure Subscription from a Databricks-generated Service Principal. Harness the power of data with this vital step in Azure infrastructure management. Mastering it is as crucial as surviving Fnaf
- 0 kudos
- 1964 Views
- 1 replies
- 0 kudos
Resolved! Restrictions on setting environment variables in Compute Policies
As recommended by Databricks, we are trying to use Compute Policies to set environment variables, which are used by our notebooks, across clusters.However, when specifying a JSON string as env var, we are getting this error upon applying the policy t...
- 1964 Views
- 1 replies
- 0 kudos
- 0 kudos
This is because you use Shared access mode.This enables multiple users to use the cluster simultaneously.However, there are features that do not work on these Shared access mode clusters:https://docs.databricks.com/en/compute/access-mode-limitations....
- 0 kudos
- 3314 Views
- 1 replies
- 0 kudos
One Azure Tenant with Multiple Azure Databricks Accounts
Hi there,We have one Azure tenant with multiple subscriptions. Each subscription is a project for itself.At this moment, we have only one Azure Databricks account, and all workspaces (created under different subscriptions) are associated with it.Can ...
- 3314 Views
- 1 replies
- 0 kudos
- 0 kudos
hello @stevanovic ,as far as I understand, in Azure, you can create one databricks account per tenant, meaning for example unity catalog is also tenant-level resource.There is a fantastic blog post available here:https://community.databricks.com/t5/t...
- 0 kudos
- 2479 Views
- 3 replies
- 2 kudos
Resolved! Silly question-Easy way to show full notebook path or owner in UI?
We have a few people working in Databricks right now in different clones of the same repository. Occasionally we'll have multiple people with the same branch open- one working, another just has it open to see what it looks like, sort of deal.This has...
- 2479 Views
- 3 replies
- 2 kudos
- 2 kudos
hi @Kayla ,I think the easiest way to check the current notebook location when opened is just hover the mouse cursor over the name of the notebook (top left, "ADE 3.1 - Streaming Deduplication" in this case) and wait for about 1-2 seconds; after that...
- 2 kudos
- 1771 Views
- 1 replies
- 1 kudos
Delta Lake: Running Delete and writes concurrently
Is it safe to run a delete query when there are active writes to a delta lake table? Next question : Is it safe to run a vacuum when writes are being done actively?
- 1771 Views
- 1 replies
- 1 kudos
- 1 kudos
Hello @sharat_n ,Yes, it is generally safe to run a DELETE query on a Delta Lake table while active writes are happening.Delta Lake is designed with ACID transactions, meaning operations like DELETE, UPDATE, and MERGE are atomic and isolated.In other...
- 1 kudos
- 2530 Views
- 2 replies
- 1 kudos
Resolved! Missing 'Permissions' settings for Delta Live Tables Pipelines
Context: Azure Databricks, I am account admin, workspace admin, and pipeline owner (as confirmed via the API and visually in the Pipelines screen). When attempting to grant CAN_MANAGE access to developers for our DLT pipelines via the Databricks web ...
- 2530 Views
- 2 replies
- 1 kudos
- 1 kudos
If you click on the link of the Pipeline and then you access the kebab menu you will see the Permissions page
- 1 kudos
- 1169 Views
- 1 replies
- 0 kudos
Unable to create a databricks workspace
I am unable to create a databricks workspace with vnet injection. I get this error:│ { │ "status": "Failed", │ "error": { │ "code": "InternalServerError", │ "message": "INTERNAL_ERROR: Unexpected error: Cannot call getCertifiedMetastoreFo...
- 1169 Views
- 1 replies
- 0 kudos
- 0 kudos
The issue resolved itself when I tried to create a new resource group, dedicated to just Databricks. I don't know why that worked. If anyone know what went wrong, I would appreciate feedback!
- 0 kudos
- 6734 Views
- 5 replies
- 1 kudos
Resolved! Unable to access Live Ganglia UI
I’m having a trouble accessing Ganglia UI on a job run cluster. The job run is executed by a service principal and I would like to enable the Ganglia UI view to an user/admin group in Databricks. The error I get is: HTTP ERROR 403 PERMISSION_DENIED: ...
- 6734 Views
- 5 replies
- 1 kudos
- 1 kudos
Hi @arkadiuszr - can you please try the following steps and let us know? To resolve this error, first you want to add your service principal to the workspace: Go to admin console as a workspace adminSelect “service principal tab”Click “Add service ...
- 1 kudos
Join Us as a Local Community Builder!
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up Now-
Access control
1 -
Apache spark
1 -
Azure
7 -
Azure databricks
5 -
Billing
2 -
Cluster
1 -
Compliance
1 -
Data Ingestion & connectivity
5 -
Databricks Runtime
1 -
Databricks SQL
2 -
DBFS
1 -
Dbt
1 -
Delta Sharing
1 -
DLT Pipeline
1 -
GA
1 -
Gdpr
1 -
Github
1 -
Partner
59 -
Public Preview
1 -
Service Principals
1 -
Unity Catalog
1 -
Workspace
2
- « Previous
- Next »
| User | Count |
|---|---|
| 119 | |
| 39 | |
| 37 | |
| 28 | |
| 25 |