- 4170 Views
- 2 replies
- 0 kudos
Can I configure Notebook Result Downloads with Databricks CLI , API or Terraform provider ?
I'm Databricks Admin and I'm looking for a solution to automate some Security Workspace settings.Those are:Notebook result downloadSQL result downloadNotebook table clipboard featuresI can't find these options in the Databricks terraform provider, Da...
- 4170 Views
- 2 replies
- 0 kudos
- 0 kudos
Hi @Redford, With the Databricks API, you have the capability to toggle the following features: Enable/Disable Features Notebook result download (key name: enableResultsDownloading)Notebook table clipboard features (key name: enableNotebo...
- 0 kudos
- 2790 Views
- 2 replies
- 1 kudos
Not able to edit the user group entitlement
I am workspace admin , but when editing the entitlement (allow unrestricted cluster creation) of user group, it is not happening, it gives a prompt to confirm the remove, but evening after confirming it does not get removes. after clicking remove , i...
- 2790 Views
- 2 replies
- 1 kudos
- 1459 Views
- 0 replies
- 0 kudos
"Forgot my password" does not send mail to users in my organization.
Hi, we are using databricks in my organization. SSO logins stopped working for us about a week ago. As a workaround, we asked our users to use the admin login page, and use their organization passwords. This didn't work because the passwords were inc...
- 1459 Views
- 0 replies
- 0 kudos
- 2166 Views
- 0 replies
- 0 kudos
HIPAA compliance "pay-as-you-go" option
Hi does Databricks offer HIPAA compliance with the "pay-as-you-go" option? I know that Snowflake has the ability to do that, what about Databricks?
- 2166 Views
- 0 replies
- 0 kudos
- 7175 Views
- 2 replies
- 1 kudos
Resolved! Unity Catalog - Lakehouse Federation: Permission to read data from foreign catalogs
I have seup connection "SQL-SV-conn" to SQL Server and based on that connection I have created foreign catalog "FC-SQL-SV".I have granted All permission on CATALOG to developers:Use CatalogUse SchemaSelectBut they can not query table (e.g. by running...
- 7175 Views
- 2 replies
- 1 kudos
- 1 kudos
OK, I have found out the answer in below docummentation:https://learn.microsoft.com/en-us/azure/databricks/query-federation/#limitationsSingle-user access mode is only available for users that own the connection.So when I use e.g. Job Cluster that ru...
- 1 kudos
- 3441 Views
- 0 replies
- 1 kudos
Workspace Model Registry with Unity Catalog
if the "default catalog for the workspace" is to Unity Catalog, how can we access a model from the workspace model registry?I have already tried mlflow.set_tracking_uri("databricks") but still try to find catalog in UC.
- 3441 Views
- 0 replies
- 1 kudos
- 2595 Views
- 0 replies
- 0 kudos
Specify cidr when using aws quick start template
Hello,I want to create a new databricks workspace using aws quick start template. However, I want to specify a cidr address that is specified by our network administration. Is it possible to do that. From what I read, the quick start template does no...
- 2595 Views
- 0 replies
- 0 kudos
- 3805 Views
- 1 replies
- 1 kudos
Resolved! Security and vulnerability for Azure databricks clusters in Data Plane
Microsoft defender is not supported for azure databricks clusters. Can someone point me to a document which describe how the security vulnerabilities are reported and fixed for azure databricks clusters in data plane.
- 3805 Views
- 1 replies
- 1 kudos
- 1 kudos
Hi Aria,can you check this link?Basically Databricks checks for vulnerabilities and does pentesting etc."Databricks will use commercially reasonable efforts to address critical vulnerabilities within 14 days, high severity within 30 days, and medium ...
- 1 kudos
- 3181 Views
- 1 replies
- 0 kudos
Automatic UC enablement with existing Unity Catalog
Hello,The environment I'm working in has an existing account and unity catalog in our primary region. These have been primarily PoC, and have existed for a year or so.Now we're building out our IaC and getting ready to move towards production. In new...
- 3181 Views
- 1 replies
- 0 kudos
- 0 kudos
- 0 kudos
- 2738 Views
- 1 replies
- 0 kudos
Connect to Salesforce
Curious if there's a Databricks connector for Salesforce on AWS?
- 2738 Views
- 1 replies
- 0 kudos
- 0 kudos
There is no "databricks" connector like the on you have in Unity Fedarating e.g. for Snowflake.You can use partner ecosystem e.g. Fivetran https://www.fivetran.com/connectors/salesforceto integrate Salesforce data to your Lakehouse. You also have spa...
- 0 kudos
- 3480 Views
- 1 replies
- 1 kudos
How to allocate costs per SQL query?
By using System Tables (systen.billing.usage) I'm able to identity DBU usage per query, but I'm not able to identify who ran each query because is not part of the table. I'm also aware of query history where all the queries and who ran them is listed...
- 3480 Views
- 1 replies
- 1 kudos
- 1 kudos
thanks @Retired_mod for the reply, however query_id is not part of the system.billing.usage table, so no way to join them by IDs. What my Databricks account team suggested me is to join them by timestamps since both tables contain a column like that....
- 1 kudos
- 4517 Views
- 2 replies
- 0 kudos
NPIP tunnel setup failure during launch
In AWS with this current error when spinning up the SQL warehouse or personal compute.Backend private link is enabled.Error: NPIP tunnel setup failure during launch. Please try again later and contact Databricks if the problem persists. Instance boot...
- 4517 Views
- 2 replies
- 0 kudos
- 0 kudos
Hello, Thanks for contacting Databricks Support. Based on the error message: NPIP_TUNNEL_SETUP_FAILURE. It indicates that bootstrap failed due to network connectivity issues between the data plane and control plane. Seems like you have already dow...
- 0 kudos
- 1500 Views
- 0 replies
- 0 kudos
Error running history command twice using DeltaTable
Hi I'm using Unity Catalog on Azure with a Managed Identity connected to my Storage Account. I can read and write data without issues, and interact with data using pyspark and SQL.I can display the history by running a cell with the following code:fr...
- 1500 Views
- 0 replies
- 0 kudos
- 4310 Views
- 2 replies
- 1 kudos
How to monitor a python wheel job with Prometheus?
Hi Community,We have a Databricks job with a single Python wheel task that runs our streaming pyspark job. The job runs on a single-node compute cluster and consumes from Kafka.Our monitoring stack is Prometheus + Grafana.I want the job's metrics to ...
- 4310 Views
- 2 replies
- 1 kudos
- 1 kudos
Hi I'm trying to use the metrics registry object inside an UDF function, but I can't because it's not serializable due to Lock. Our goal is to be able to count the number of messages parsed, and the number of messages we can't parsed (due to exceptio...
- 1 kudos
- 1050 Views
- 0 replies
- 0 kudos
Handling Kafka Topics with Avro Schema
Our input data resides in a Kafka topic, and we utilize the Kafka schema registry with Avro schemas. While I can retrieve the schema from the registry, I am facing challenges creating a Spark DataFrame that correctly serializes data for streaming rea...
- 1050 Views
- 0 replies
- 0 kudos
Connect with Databricks Users in Your Area
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group-
Access control
1 -
Access Delta Tables
2 -
ActiveDirectory
1 -
AmazonKMS
1 -
Apache spark
1 -
App
1 -
Availability
1 -
Availability Zone
1 -
AWS
5 -
Aws databricks
1 -
AZ
1 -
Azure
8 -
Azure Data Lake Storage
1 -
Azure databricks
6 -
Azure databricks workspace
1 -
Best practice
1 -
Best Practices
2 -
Billing
2 -
Bucket
1 -
Cache
1 -
Change
1 -
Checkpoint
1 -
Checkpoint Path
1 -
Cluster
1 -
Cluster Pools
1 -
Clusters
1 -
ClustersJob
1 -
Compliance
1 -
Compute Instances
1 -
Cost
1 -
Credential passthrough
1 -
Data
1 -
Data Ingestion & connectivity
6 -
Data Plane
1 -
Databricks Account
1 -
Databricks Control Plane
1 -
Databricks Error Message
2 -
Databricks Partner
1 -
Databricks Repos
1 -
Databricks Runtime
1 -
Databricks SQL
3 -
Databricks SQL Dashboard
1 -
Databricks workspace
1 -
DatabricksJobs
1 -
DatabricksLTS
1 -
DBFS
1 -
DBR
3 -
Dbt
1 -
Dbu
3 -
Deep learning
1 -
DeleteTags Permissions
1 -
Delta
4 -
Delta Sharing
1 -
Delta table
1 -
Dev
1 -
Different Instance Types
1 -
Disaster recovery
1 -
DisasterRecoveryPlan
1 -
DLT Pipeline
1 -
EBS
1 -
Email
2 -
External Data Sources
1 -
Feature
1 -
GA
1 -
Ganglia
3 -
Ganglia Metrics
2 -
GangliaMetrics
1 -
GCP
1 -
GCP Support
1 -
Gdpr
1 -
Gpu
2 -
Group Entitlements
1 -
HIPAA
1 -
Hyperopt
1 -
Init script
1 -
InstanceType
1 -
Integrations
1 -
IP Addresses
1 -
IPRange
1 -
Job
1 -
Job Cluster
1 -
Job clusters
1 -
Job Run
1 -
JOBS
1 -
Key
1 -
KMS
1 -
KMSKey
1 -
Lakehouse
1 -
Limit
1 -
Live Table
1 -
Log
2 -
LTS
3 -
Metrics
1 -
MFA
1 -
ML
1 -
Model Serving
1 -
Multiple workspaces
1 -
Notebook Results
1 -
Okta
1 -
On-premises
1 -
Partner
62 -
Pools
1 -
Premium Workspace
1 -
Public Preview
1 -
Redis
1 -
Repos
1 -
Rest API
1 -
Root Bucket
2 -
SCIM API
1 -
Security
1 -
Security Group
1 -
Security Patch
1 -
Service principal
1 -
Service Principals
1 -
Single User Access Permission
1 -
Sns
1 -
Spark
1 -
Spark-submit
1 -
Spot instances
1 -
SQL
1 -
Sql Warehouse
1 -
Sql Warehouse Endpoints
1 -
Ssh
1 -
Sso
2 -
Streaming Data
1 -
Subnet
1 -
Sync Users
1 -
Tags
1 -
Team Members
1 -
Thrift
1 -
TODAY
1 -
Track Costs
1 -
Unity Catalog
1 -
Use
1 -
User
1 -
Version
1 -
Vulnerability Issue
1 -
Welcome Email
1 -
Workspace
2 -
Workspace Access
1
- « Previous
- Next »
User | Count |
---|---|
40 | |
13 | |
9 | |
9 | |
9 |