- 528 Views
- 1 replies
- 0 kudos
Resolved! Restrictions on setting environment variables in Compute Policies
As recommended by Databricks, we are trying to use Compute Policies to set environment variables, which are used by our notebooks, across clusters.However, when specifying a JSON string as env var, we are getting this error upon applying the policy t...
- 528 Views
- 1 replies
- 0 kudos
- 0 kudos
This is because you use Shared access mode.This enables multiple users to use the cluster simultaneously.However, there are features that do not work on these Shared access mode clusters:https://docs.databricks.com/en/compute/access-mode-limitations....
- 0 kudos
- 575 Views
- 1 replies
- 0 kudos
One Azure Tenant with Multiple Azure Databricks Accounts
Hi there,We have one Azure tenant with multiple subscriptions. Each subscription is a project for itself.At this moment, we have only one Azure Databricks account, and all workspaces (created under different subscriptions) are associated with it.Can ...
- 575 Views
- 1 replies
- 0 kudos
- 0 kudos
hello @stevanovic ,as far as I understand, in Azure, you can create one databricks account per tenant, meaning for example unity catalog is also tenant-level resource.There is a fantastic blog post available here:https://community.databricks.com/t5/t...
- 0 kudos
- 534 Views
- 3 replies
- 2 kudos
Resolved! Silly question-Easy way to show full notebook path or owner in UI?
We have a few people working in Databricks right now in different clones of the same repository. Occasionally we'll have multiple people with the same branch open- one working, another just has it open to see what it looks like, sort of deal.This has...
- 534 Views
- 3 replies
- 2 kudos
- 2 kudos
hi @Kayla ,I think the easiest way to check the current notebook location when opened is just hover the mouse cursor over the name of the notebook (top left, "ADE 3.1 - Streaming Deduplication" in this case) and wait for about 1-2 seconds; after that...
- 2 kudos
- 519 Views
- 1 replies
- 0 kudos
Delta Lake: Running Delete and writes concurrently
Is it safe to run a delete query when there are active writes to a delta lake table? Next question : Is it safe to run a vacuum when writes are being done actively?
- 519 Views
- 1 replies
- 0 kudos
- 0 kudos
Hello @sharat_n ,Yes, it is generally safe to run a DELETE query on a Delta Lake table while active writes are happening.Delta Lake is designed with ACID transactions, meaning operations like DELETE, UPDATE, and MERGE are atomic and isolated.In other...
- 0 kudos
- 371 Views
- 0 replies
- 0 kudos
Databricks report error: unexpected end of stream, read 0 bytes from 4 (socket was closed by server)
Has anyone encountered this error and knows how to resolve it?"Unexpected end of stream, read 0 bytes from 4 (socket was closed by server)."This occurs in Databricks while generating reports.I've already adjusted the wait_timeout to 28,800, and both ...
- 371 Views
- 0 replies
- 0 kudos
- 583 Views
- 2 replies
- 1 kudos
Resolved! Missing 'Permissions' settings for Delta Live Tables Pipelines
Context: Azure Databricks, I am account admin, workspace admin, and pipeline owner (as confirmed via the API and visually in the Pipelines screen). When attempting to grant CAN_MANAGE access to developers for our DLT pipelines via the Databricks web ...
- 583 Views
- 2 replies
- 1 kudos
- 1 kudos
If you click on the link of the Pipeline and then you access the kebab menu you will see the Permissions page
- 1 kudos
- 497 Views
- 1 replies
- 0 kudos
Unable to create a databricks workspace
I am unable to create a databricks workspace with vnet injection. I get this error:│ { │ "status": "Failed", │ "error": { │ "code": "InternalServerError", │ "message": "INTERNAL_ERROR: Unexpected error: Cannot call getCertifiedMetastoreFo...
- 497 Views
- 1 replies
- 0 kudos
- 0 kudos
The issue resolved itself when I tried to create a new resource group, dedicated to just Databricks. I don't know why that worked. If anyone know what went wrong, I would appreciate feedback!
- 0 kudos
- 5080 Views
- 5 replies
- 1 kudos
Resolved! Unable to access Live Ganglia UI
I’m having a trouble accessing Ganglia UI on a job run cluster. The job run is executed by a service principal and I would like to enable the Ganglia UI view to an user/admin group in Databricks. The error I get is: HTTP ERROR 403 PERMISSION_DENIED: ...
- 5080 Views
- 5 replies
- 1 kudos
- 1 kudos
Hi @arkadiuszr - can you please try the following steps and let us know? To resolve this error, first you want to add your service principal to the workspace: Go to admin console as a workspace adminSelect “service principal tab”Click “Add service ...
- 1 kudos
- 10486 Views
- 24 replies
- 22 kudos
Resolved! Unable to login to Azure Databricks Account Console
I have a personal Azure pay-as-you-go subscription in which I have the 'Global Administrator' role. I am also the databricks account administrator.Until two weeks ago, I was able to access the databricks account console without any issues, but I am f...
- 10486 Views
- 24 replies
- 22 kudos
- 22 kudos
Thanks, @dustinvannoy and @RameshRetnasamy!When I first started the Formula1 training course and couldn’t find the Account Console, I assumed Databricks must have changed something. Thanks for clarifying—I’m now able to follow along with the instruct...
- 22 kudos
- 464 Views
- 0 replies
- 0 kudos
VPAT Form
How do I find a Voluntary Product Accessibility Template (VPAT) from Databricks?
- 464 Views
- 0 replies
- 0 kudos
- 3208 Views
- 7 replies
- 5 kudos
Resolved! How to get rid of a pesky gen AI feature in the editor ?
Hi,The editor interface has that gen AI feature following empty lines with a cursor. I find that very distracting and irritating. More over, once a line is deleted that unsolicited thing is interfering with code (snapshots included).How to get rid of...
- 3208 Views
- 7 replies
- 5 kudos
- 5 kudos
Summary:I was not able to solve the UI/UX artifact on my own (on user side)The UI/UX issue was resolved somewhere on Databricks side. The UI/AX artifact is no longer interfering with work.
- 5 kudos
- 356 Views
- 1 replies
- 0 kudos
Resolved! Attach a databricks_instance_pool to databricks_cluster_policy via terraform
Hello Team,I am trying to create a databricks instance pool and attach it to a cluster policy in our terraform code. But I am having hard time finding a good documentation. Has any one done it? Below is my sample code and I am getting errorI keep get...
- 356 Views
- 1 replies
- 0 kudos
- 0 kudos
Fixed it! "instance_pool_id" : { type = "fixed" values = "databricks_instance_pool.dev_test_cluster_pool.id"}
- 0 kudos
- 327 Views
- 1 replies
- 0 kudos
Removal of account admin
Hi, I'm having issues with removing account admin (probably the first one, to which databricks account was related to). Under user management, when I hit the delete user button, it prompts:Either missing permissions to delete <user_email> or deleting...
- 327 Views
- 1 replies
- 0 kudos
- 0 kudos
I have also tried removing account using databricks CLI: databricks account users delete <ID>Error: Either missing permissions to delete <account> or deleting own account.
- 0 kudos
- 325 Views
- 0 replies
- 0 kudos
Databricks Bundles - Terraform state management
Hello,I had a look at DABs today and it seems they are using Terraform under the hood. The state is stored in Databricks Workspace, in the bundle deployment directory. Is it possible to use just the state management functionality that DABs must have ...
- 325 Views
- 0 replies
- 0 kudos
- 590 Views
- 2 replies
- 0 kudos
Lakehouse federation -' on behalf of ' queries
Is it possible to achieve the following in a lake-house federation setup using Azure Databricks?1. Establish an external connection (EC1) to an external data source (EDS) using the credentials of user U1.2. Create a foreign catalog (FC1) utilizing EC...
- 590 Views
- 2 replies
- 0 kudos
- 0 kudos
Thanks for explaining the authorization flow, @rangu . In the example mentioned, does Databricks support passing the user’s credentials to an external data source? For instance, can it pass the OAuth token for the user along with the externalID crede...
- 0 kudos
Connect with Databricks Users in Your Area
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group-
Access control
1 -
Access Delta Tables
2 -
ActiveDirectory
1 -
AmazonKMS
1 -
Apache spark
1 -
App
1 -
Availability
1 -
Availability Zone
1 -
AWS
5 -
Aws databricks
1 -
AZ
1 -
Azure
8 -
Azure Data Lake Storage
1 -
Azure databricks
6 -
Azure databricks workspace
1 -
Best practice
1 -
Best Practices
2 -
Billing
2 -
Bucket
1 -
Cache
1 -
Change
1 -
Checkpoint
1 -
Checkpoint Path
1 -
Cluster
1 -
Cluster Pools
1 -
Clusters
1 -
ClustersJob
1 -
Compliance
1 -
Compute Instances
1 -
Cost
1 -
Credential passthrough
1 -
Data
1 -
Data Ingestion & connectivity
6 -
Data Plane
1 -
Databricks Account
1 -
Databricks Control Plane
1 -
Databricks Error Message
2 -
Databricks Partner
1 -
Databricks Repos
1 -
Databricks Runtime
1 -
Databricks SQL
3 -
Databricks SQL Dashboard
1 -
Databricks workspace
1 -
DatabricksJobs
1 -
DatabricksLTS
1 -
DBFS
1 -
DBR
3 -
Dbt
1 -
Dbu
3 -
Deep learning
1 -
DeleteTags Permissions
1 -
Delta
4 -
Delta Sharing
1 -
Delta table
1 -
Dev
1 -
Different Instance Types
1 -
Disaster recovery
1 -
DisasterRecoveryPlan
1 -
DLT Pipeline
1 -
EBS
1 -
Email
2 -
External Data Sources
1 -
Feature
1 -
GA
1 -
Ganglia
3 -
Ganglia Metrics
2 -
GangliaMetrics
1 -
GCP
1 -
GCP Support
1 -
Gdpr
1 -
Gpu
2 -
Group Entitlements
1 -
HIPAA
1 -
Hyperopt
1 -
Init script
1 -
InstanceType
1 -
Integrations
1 -
IP Addresses
1 -
IPRange
1 -
Job
1 -
Job Cluster
1 -
Job clusters
1 -
Job Run
1 -
JOBS
1 -
Key
1 -
KMS
1 -
KMSKey
1 -
Lakehouse
1 -
Limit
1 -
Live Table
1 -
Log
2 -
LTS
3 -
Metrics
1 -
MFA
1 -
ML
1 -
Model Serving
1 -
Multiple workspaces
1 -
Notebook Results
1 -
Okta
1 -
On-premises
1 -
Partner
62 -
Pools
1 -
Premium Workspace
1 -
Public Preview
1 -
Redis
1 -
Repos
1 -
Rest API
1 -
Root Bucket
2 -
SCIM API
1 -
Security
1 -
Security Group
1 -
Security Patch
1 -
Service principal
1 -
Service Principals
1 -
Single User Access Permission
1 -
Sns
1 -
Spark
1 -
Spark-submit
1 -
Spot instances
1 -
SQL
1 -
Sql Warehouse
1 -
Sql Warehouse Endpoints
1 -
Ssh
1 -
Sso
2 -
Streaming Data
1 -
Subnet
1 -
Sync Users
1 -
Tags
1 -
Team Members
1 -
Thrift
1 -
TODAY
1 -
Track Costs
1 -
Unity Catalog
1 -
Use
1 -
User
1 -
Version
1 -
Vulnerability Issue
1 -
Welcome Email
1 -
Workspace
2 -
Workspace Access
1
- « Previous
- Next »
User | Count |
---|---|
40 | |
13 | |
9 | |
9 | |
9 |