- 1237 Views
- 3 replies
- 2 kudos
Databricks UI error
I receive this respone when trying to download a csv file: adb-737684297741988.8.azuredatabricks.net/ajax-api/2.0/workspace/export?path=/Repos/jhu_admin%40siemensgamesa.com/sgre-dfp-lifestrategy-python/cpo/output/Rattlesnake Flats/Rdata/AssetCheck_Ra...
- 1237 Views
- 3 replies
- 2 kudos
- 2 kudos
@szymon_dybczak I am using the GUI (see the screenshot in the original post): File marked -> Three dots -> Export
- 2 kudos
- 1244 Views
- 1 replies
- 0 kudos
JDBC Error with OpenJDK 21
Connecting to Azure Databricks using latest JDBC driver and OpenJDK 21 I got this error. I already set:--add-opens=java.base/java.nio=ALL-UNNAMED-Dio.netty.tryReflectionSetAccessible=true java.sql.SQLException: [Databricks][JDBCDriver](500618) Error ...
- 1244 Views
- 1 replies
- 0 kudos
- 0 kudos
Version 02.06.38.1068Using SQLSquirrel with ibm-semeru-certified-17-jdk (17.0.11+9)With JRE options --add-opens=jdk.unsupported/sun.misc=ALL-UNNAMED failsShould be --add-opens=java.base/java.nio=ALL-UNNAMEDcaused by: java.lang.reflect.InaccessibleOb...
- 0 kudos
- 287 Views
- 0 replies
- 0 kudos
Using Databricks CLI for generating Notebooks not supported or not implemented
Hi I'm a Data engineer and recently developed a Notebook analytics template for general purposes that I would like to be the standard on my company. Continuing, I created another notebook with a text widget that uses the user input to map the folder ...
- 287 Views
- 0 replies
- 0 kudos
- 460 Views
- 1 replies
- 7 kudos
Databricks Unity Catalog Bug - Reset of Network Connectivity Configuration not possible
The following use case is strange regarding the Network Connectivity Configuration (NCC):I create a Workspace (the NCC is empty)I create a NCCI attach the NCC to the WorkspaceI want to remove the NCC from the Workspace -> not possibleTherefore, I can...
- 460 Views
- 1 replies
- 7 kudos
- 415 Views
- 1 replies
- 1 kudos
[Metastore] can't assign to workspace in cross Azure Account envirronment
Hi all,hope you are doing well. We have a metastore in tenant A, and workspaces in tenant B. Tenant B has Contributor Role to tenant A, to access the metastore. For some reasons, we can not assign the metastore from tenant A to the workspace in tenan...
- 415 Views
- 1 replies
- 1 kudos
- 1 kudos
- 1 kudos
- 821 Views
- 3 replies
- 4 kudos
dbutils secrets break matplotlib in recent Databricks runtimes
Using a Databricks cluster running Databricks Runtime version 15.4 LTS Beta, and with no init-scripts, the following code works fine:import numpy as npimport pandas as pddf = pd.DataFrame(np.random.randn(2000, 10))df.plot(subplots=True, figsize=(10, ...
- 821 Views
- 3 replies
- 4 kudos
- 4 kudos
OK, I understand the error now:This specific secret is a small alphanumeric string. The probability of it being part of the base64 of a big PNG is quite large.Is there any workaround for this? I know I can't turn off the redaction of secrets. Can I d...
- 4 kudos
- 828 Views
- 1 replies
- 0 kudos
Obtain access of Azure metastore storage account to configure Lifecycle management
I recently set up an Azure Databricks workspace with an automatically created metastore and metastore-level root storage within the metastore blob storage account. All the catalogs, schemas, and tables/volumes have been created without a specified or...
- 828 Views
- 1 replies
- 0 kudos
- 0 kudos
- 0 kudos
- 929 Views
- 1 replies
- 0 kudos
Databricks connection pool(ing) using python
I have a python application which I need to connect with databricks (Databricks SQL Connector). The process for that is quite simple -1. connect with dbr2. execute your query3. get the result4. close the connectionExample from - https://docs.databric...
- 929 Views
- 1 replies
- 0 kudos
- 0 kudos
Hi @Retired_mod, thanks for your reply. I have 2 follow-up questions for you - When you say "consider managing connection lifetimes by closing and recreating them periodically" - the problem I face with this is when my API request comes in and I crea...
- 0 kudos
- 2446 Views
- 5 replies
- 4 kudos
Resolved! UCX- installation error -
I'm getting the following error when trying to install UCX on a specific workspace.I have already installed ucx from my machine on other workspace and works, now on this workspace i'm getting this error ERROR [d.l.blueprint.parallel][installing_das...
- 2446 Views
- 5 replies
- 4 kudos
- 4 kudos
Issue was resolved in lsql v0.7.4, a ucx dependency: https://github.com/databrickslabs/lsql/releases/tag/v0.7.41. Upgrade the Databricks cli: https://github.com/databricks/cli/releases/2. Upgrade ucx: `databricks labs install ucx`
- 4 kudos
- 1474 Views
- 5 replies
- 2 kudos
Promoting assets from Dev workspace to Prod
What is the best way to promote Databricks objects from one workspace to another? For example, DABs are great for code and models. However, SQL alerts do not seem to fall into DABs. The API allows for retrieving queries and creating them via API in t...
- 1474 Views
- 5 replies
- 2 kudos
- 2 kudos
https://docs.databricks.com/api/workspace/queries https://docs.databricks.com/en/sql/dbsql-api-latest.html#changes-to-the-alerts-api The API path is now api/2.0/sql/alerts, replacing the legacy path of /api/2.0/preview/sql/alerts. https://docs.databr...
- 2 kudos
- 615 Views
- 1 replies
- 1 kudos
System Tables
Hi,I've recently enabled the billing schema in system tables, but still not able to view it in catalog. I am both an account admin and workspace admin, as well as an owner of the workspace in Azure. A colleague (who used to be Azure AD admin but no l...
- 615 Views
- 1 replies
- 1 kudos
- 1 kudos
- 1 kudos
- 418 Views
- 1 replies
- 1 kudos
How to list my groups in unitity catalog via an API
I'm trying to list my groups in unity catalog via an API call, but it keeps giving the error "Bad Target"My GET request is:-https://adb-3676685197829242.2.azuredatabricks.net/api/2.1/accounts/cfb89269-0d9e-4235-b7985-455481b501a/scim/v2/GroupsI'm abl...
- 418 Views
- 1 replies
- 1 kudos
- 1 kudos
- 1 kudos
- 620 Views
- 1 replies
- 1 kudos
Databricks + Apache Iceberg = advantageous or wasted effort due to duplicate functionality ?
Trying to design a Lakehouse. Spark is at the base layer. Now wondering if adding Apache Iceberg sitting below Spark will be of help, or, not ? Preferring Iceberg for its auto indexing, ACID query facilities over big hetergenous datasets. Wonder if i...
- 620 Views
- 1 replies
- 1 kudos
- 1 kudos
Hello, if you're planning on building your own open source stack of spark+iceberg, it can be a good choice. If you're on Databricks, however, you're going to miss out a *lot* on delta features that are baked into the platform. Specifically compute +...
- 1 kudos
- 1723 Views
- 4 replies
- 3 kudos
Resolved! [VNET injection] Container and container subnet
Hi,I was researching everywhere and could not find the answer. I understand that when workspace is created, it has 2 subnets, host and container. The VM, which runs the Databricks container, is in host subnet, which logically means the container is a...
- 1723 Views
- 4 replies
- 3 kudos
- 3 kudos
Hi @data_bricklayer ,Public Subnet (host): The public subnet is typically used for resources that need to communicate with the internet or other Azure services. In Azure Databricks, this subnet is used for driver nodes of the clusters that require ou...
- 3 kudos
- 318 Views
- 0 replies
- 0 kudos
vocareum lab showing no workspace found.
Hi,I bought Generative AI Engineering with Databricks lab subscription by paying $75 just to get the hands on experience.The labs worked for one day and then they stop working. Please active the labs ASAP, I am preparing for the certification exam an...
- 318 Views
- 0 replies
- 0 kudos
Connect with Databricks Users in Your Area
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.
If there isn’t a group near you, start one and help create a community that brings people together.
Request a New Group-
Access control
1 -
Access Delta Tables
2 -
ActiveDirectory
1 -
AmazonKMS
1 -
Apache spark
1 -
App
1 -
Availability
1 -
Availability Zone
1 -
AWS
5 -
Aws databricks
1 -
AZ
1 -
Azure
8 -
Azure Data Lake Storage
1 -
Azure databricks
6 -
Azure databricks workspace
1 -
Best practice
1 -
Best Practices
2 -
Billing
2 -
Bucket
1 -
Cache
1 -
Change
1 -
Checkpoint
1 -
Checkpoint Path
1 -
Cluster
1 -
Cluster Pools
1 -
Clusters
1 -
ClustersJob
1 -
Compliance
1 -
Compute Instances
1 -
Cost
1 -
Credential passthrough
1 -
Data
1 -
Data Ingestion & connectivity
6 -
Data Plane
1 -
Databricks Account
1 -
Databricks Control Plane
1 -
Databricks Error Message
2 -
Databricks Partner
1 -
Databricks Repos
1 -
Databricks Runtime
1 -
Databricks SQL
3 -
Databricks SQL Dashboard
1 -
Databricks workspace
1 -
DatabricksJobs
1 -
DatabricksLTS
1 -
DBFS
1 -
DBR
3 -
Dbt
1 -
Dbu
3 -
Deep learning
1 -
DeleteTags Permissions
1 -
Delta
4 -
Delta Sharing
1 -
Delta table
1 -
Dev
1 -
Different Instance Types
1 -
Disaster recovery
1 -
DisasterRecoveryPlan
1 -
DLT Pipeline
1 -
EBS
1 -
Email
2 -
External Data Sources
1 -
Feature
1 -
GA
1 -
Ganglia
3 -
Ganglia Metrics
2 -
GangliaMetrics
1 -
GCP
1 -
GCP Support
1 -
Gdpr
1 -
Gpu
2 -
Group Entitlements
1 -
HIPAA
1 -
Hyperopt
1 -
Init script
1 -
InstanceType
1 -
Integrations
1 -
IP Addresses
1 -
IPRange
1 -
Job
1 -
Job Cluster
1 -
Job clusters
1 -
Job Run
1 -
JOBS
1 -
Key
1 -
KMS
1 -
KMSKey
1 -
Lakehouse
1 -
Limit
1 -
Live Table
1 -
Log
2 -
LTS
3 -
Metrics
1 -
MFA
1 -
ML
1 -
Model Serving
1 -
Multiple workspaces
1 -
Notebook Results
1 -
Okta
1 -
On-premises
1 -
Partner
58 -
Pools
1 -
Premium Workspace
1 -
Public Preview
1 -
Redis
1 -
Repos
1 -
Rest API
1 -
Root Bucket
2 -
SCIM API
1 -
Security
1 -
Security Group
1 -
Security Patch
1 -
Service principal
1 -
Service Principals
1 -
Single User Access Permission
1 -
Sns
1 -
Spark
1 -
Spark-submit
1 -
Spot instances
1 -
SQL
1 -
Sql Warehouse
1 -
Sql Warehouse Endpoints
1 -
Ssh
1 -
Sso
2 -
Streaming Data
1 -
Subnet
1 -
Sync Users
1 -
Tags
1 -
Team Members
1 -
Thrift
1 -
TODAY
1 -
Track Costs
1 -
Unity Catalog
1 -
Use
1 -
User
1 -
Version
1 -
Vulnerability Issue
1 -
Welcome Email
1 -
Workspace
2 -
Workspace Access
1
- « Previous
- Next »
User | Count |
---|---|
40 | |
11 | |
9 | |
9 | |
9 |