- 475 Views
- 1 replies
- 0 kudos
Assistance Required for Enabling Unity Catalog in Databricks Workspace
Hi,I hope this message finds you well.I am reaching out regarding a concern with Databricks Administrator privileges. I have an Azure subscription and I use Azure Databricks for my tutorials, but I currently do not have Global Administrator access, w...
- 475 Views
- 1 replies
- 0 kudos
- 0 kudos
Hello, each account should have at least one account administrator that is the one that has the permissions on the managed account and that can give you this access if applicable or assist you with the set up of the UC, you might need to ask internal...
- 0 kudos
- 678 Views
- 1 replies
- 1 kudos
Disabling Default Personal Compute Policy in Terraform (AWS)
I'm looking to disable the Personal Compute Policy in Terraform, however I don't see an option to do this. According to this documentation here https://docs.databricks.com/en/admin/clusters/personal-compute.html#customize-the-personal-compute-policy ...
- 678 Views
- 1 replies
- 1 kudos
- 1 kudos
I have the same issue in terraform-provider-google.I found how to do it from Admin UI (UI -> Settings -> Features Enablement -> disable Personal compute). At the same time I didn't find how to have the same behavior from Terraform.
- 1 kudos
- 8722 Views
- 1 replies
- 0 kudos
Resolved! Delete Databricks account
Hi everyone, as in the topic, I would like to delete unnnecesarily created account. I have found outdated solutions (e.g. https://community.databricks.com/t5/data-engineering/how-to-delete-databricks-account/m-p/6323#M2501), but they do not work anym...
- 8722 Views
- 1 replies
- 0 kudos
- 0 kudos
Hi @Edyta ,FOR AWS:Manage your subscription | Databricks on AWS" Before you delete a Databricks account, you must first cancel your Databricks subscription and delete all Unity Catalog metastores in the account. After you delete all metastores associ...
- 0 kudos
- 1905 Views
- 3 replies
- 2 kudos
Resolved! Message queue to directly trigger job
Hi All,I'm very new to Databricks and trying to understand my use case a bit better.I have a databricks script / job that I want to be reactive to events outside of my databricks environment. A best case scenario would be if my script / job could aut...
- 1905 Views
- 3 replies
- 2 kudos
- 2 kudos
Hi @AmandaOhare ,You can use AWS Lambda to achieve that. You can setup queue trigger that will activate AWS lambda function. In that function you can call datbricks rest API that will launch workflow/job.
- 2 kudos
- 699 Views
- 1 replies
- 3 kudos
Refresh permission on Lakeview Dashboard
Hi folks!I'm sure I'm not the only one, but our users have the tendency to click the big Refresh button on all dashboards every time they open them.Using resources efficiently is something I value deeply, so our team came up with a schedule policy - ...
- 699 Views
- 1 replies
- 3 kudos
- 3 kudos
Hi @leo-machado this is a very reasonable request. It's not feasible right now, but I'll pass it on to the product team as I'm sure you're not the only one.
- 3 kudos
- 929 Views
- 0 replies
- 0 kudos
Databricks Community - Cannot See Reply To My Posts
Databricks Community - Cannot See Reply To My Posts.Am I the only one facing this issue or others are also facing the same.
- 929 Views
- 0 replies
- 0 kudos
- 353 Views
- 1 replies
- 0 kudos
Unable to access databricls workspaces
Hi Team, Previously we are able to access databricks workspaces using username and password but now its asking for token, But we dont have access to mail account with username that we are using while login. Could you please assist on the issue how ca...
- 353 Views
- 1 replies
- 0 kudos
- 240 Views
- 0 replies
- 0 kudos
Issue with updating email with SCIM Provisioning
Hi all,For our set-up we have configured SCIM provisioning using Entra ID, group assignment on Azure is dealt with by IdentityIQ Sailpoint, and have enabled SSO for Databricks. It has/is working fine apart from one scenario. The original email assign...
- 240 Views
- 0 replies
- 0 kudos
- 238 Views
- 0 replies
- 0 kudos
How to add 'additionallyAllowedTenants' in Databricks config or PySpark config?
I have a multi-tenant Azure app. I am using this app's credentials to read ADLS container files from Databricks cluster using PySpark dataframe.I need to set this 'additionallyAllowedTenants' flag value to '*' or a specific tenant_id of the multi-ten...
- 238 Views
- 0 replies
- 0 kudos
- 946 Views
- 1 replies
- 0 kudos
Resolved! Using Databricks Connect with PyCharm behind Proxy
I am using Databricks connect with PyCharm in order to locally develop my pipelines before deploying to Databricks. I set up authentication for databricks-connect using the following guide https://learn.microsoft.com/en-us/azure/databricks/dev-tools/...
- 946 Views
- 1 replies
- 0 kudos
- 0 kudos
I managed to solve the problem by including my login and password in the HTTPS_PROXY environment variable.I set it to something like this HTTPS_PROXY = "http://username:password@proxy.company.com:port" and databricks-connect can create spark sessions...
- 0 kudos
- 480 Views
- 3 replies
- 1 kudos
Can't login to new AWS databricks workspace
Just get our first AWS Databricks workspace created by following the instructions: https://docs.databricks.com/en/admin/workspace/create-uc-workspace.htmladopted the customer-managed VPC and private link for the setup. Upon reviewing the network conf...
- 480 Views
- 3 replies
- 1 kudos
- 1 kudos
the problem resolved after networking team currently set the DNS to point to the endpoint. We can login through the workspace link now.
- 1 kudos
- 1512 Views
- 2 replies
- 0 kudos
Databricks REST API to fetch mount points
Is there a way to fetch workspace mount points ( mount infos) through REST API or SQL-query ? ( similar to the python API "display(dbutils.fs.mounts())" ) I couldn't find any REST API for the mounts in the official databricks API documentation ( ...
- 1512 Views
- 2 replies
- 0 kudos
- 0 kudos
How to create mount points using aws Databricks with shared cluster?
- 0 kudos
- 990 Views
- 4 replies
- 3 kudos
Resolved! Does REST API usage is chargeable ?
If we use REST API to perform some operation on the Databricks cluster- like running some SQL Statements, will that be chargeable, will it costs us.
- 990 Views
- 4 replies
- 3 kudos
- 3 kudos
Yep, to sum up - anything that requires compute power and makes the cluster to spin up will be charged.
- 3 kudos
- 330 Views
- 0 replies
- 0 kudos
Custom Runtime marketplace
Hi! Is there a possibility to share the solution accelerator on the custom runtime via the databricks marketplace?
- 330 Views
- 0 replies
- 0 kudos
- 950 Views
- 3 replies
- 1 kudos
Resolved! Confusion about SCC, NAT, and Internet connection (Azure)
Hi all,I created an Azure Databricks workspace with the below featuresDeployed in my Azure virtual network (VNet injection)Enabled Secure cluster connectivityEnabled the Back-end Private LinkI did not create and use a NAT. I found that I still can ac...
- 950 Views
- 3 replies
- 1 kudos
- 1 kudos
Hi @AlbertWang ,If you choose SCC option, there shouldn't be NAT gateway created. You can check that yourself if you go the managed resource group that is created when you deploying databricks workspace.When you use SCC then if you would like to have...
- 1 kudos
Connect with Databricks Users in Your Area
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group-
Access control
1 -
Access Delta Tables
2 -
ActiveDirectory
1 -
AmazonKMS
1 -
Apache spark
1 -
App
1 -
Availability
1 -
Availability Zone
1 -
AWS
5 -
Aws databricks
1 -
AZ
1 -
Azure
8 -
Azure Data Lake Storage
1 -
Azure databricks
6 -
Azure databricks workspace
1 -
Best practice
1 -
Best Practices
2 -
Billing
2 -
Bucket
1 -
Cache
1 -
Change
1 -
Checkpoint
1 -
Checkpoint Path
1 -
Cluster
1 -
Cluster Pools
1 -
Clusters
1 -
ClustersJob
1 -
Compliance
1 -
Compute Instances
1 -
Cost
1 -
Credential passthrough
1 -
Data
1 -
Data Ingestion & connectivity
6 -
Data Plane
1 -
Databricks Account
1 -
Databricks Control Plane
1 -
Databricks Error Message
2 -
Databricks Partner
1 -
Databricks Repos
1 -
Databricks Runtime
1 -
Databricks SQL
3 -
Databricks SQL Dashboard
1 -
Databricks workspace
1 -
DatabricksJobs
1 -
DatabricksLTS
1 -
DBFS
1 -
DBR
3 -
Dbt
1 -
Dbu
3 -
Deep learning
1 -
DeleteTags Permissions
1 -
Delta
4 -
Delta Sharing
1 -
Delta table
1 -
Dev
1 -
Different Instance Types
1 -
Disaster recovery
1 -
DisasterRecoveryPlan
1 -
DLT Pipeline
1 -
EBS
1 -
Email
2 -
External Data Sources
1 -
Feature
1 -
GA
1 -
Ganglia
3 -
Ganglia Metrics
2 -
GangliaMetrics
1 -
GCP
1 -
GCP Support
1 -
Gdpr
1 -
Gpu
2 -
Group Entitlements
1 -
HIPAA
1 -
Hyperopt
1 -
Init script
1 -
InstanceType
1 -
Integrations
1 -
IP Addresses
1 -
IPRange
1 -
Job
1 -
Job Cluster
1 -
Job clusters
1 -
Job Run
1 -
JOBS
1 -
Key
1 -
KMS
1 -
KMSKey
1 -
Lakehouse
1 -
Limit
1 -
Live Table
1 -
Log
2 -
LTS
3 -
Metrics
1 -
MFA
1 -
ML
1 -
Model Serving
1 -
Multiple workspaces
1 -
Notebook Results
1 -
Okta
1 -
On-premises
1 -
Partner
62 -
Pools
1 -
Premium Workspace
1 -
Public Preview
1 -
Redis
1 -
Repos
1 -
Rest API
1 -
Root Bucket
2 -
SCIM API
1 -
Security
1 -
Security Group
1 -
Security Patch
1 -
Service principal
1 -
Service Principals
1 -
Single User Access Permission
1 -
Sns
1 -
Spark
1 -
Spark-submit
1 -
Spot instances
1 -
SQL
1 -
Sql Warehouse
1 -
Sql Warehouse Endpoints
1 -
Ssh
1 -
Sso
2 -
Streaming Data
1 -
Subnet
1 -
Sync Users
1 -
Tags
1 -
Team Members
1 -
Thrift
1 -
TODAY
1 -
Track Costs
1 -
Unity Catalog
1 -
Use
1 -
User
1 -
Version
1 -
Vulnerability Issue
1 -
Welcome Email
1 -
Workspace
2 -
Workspace Access
1
- « Previous
- Next »
User | Count |
---|---|
40 | |
13 | |
9 | |
9 | |
9 |