- 143 Views
- 3 replies
- 0 kudos
Running jobs as service principal, while pulling code from Azure DevOps
In our Dataplatform, our jobs are defined in a dataplatform_jobs.yml within a Databricks Asset Bundle, and then pushed to Databricks via an Azure Devops Pipeline (Azure Devops is where our codebase resides). Currently, this results in workflows looki...
- 143 Views
- 3 replies
- 0 kudos
- 0 kudos
Hello @LuukDSL ,Could you share a snippet of your CI/CD YAML file so we can give more specific advice?I’ve connected Azure DevOps to Databricks using ARM credentials, and it was set the pipeline’s "Run as" user to a service principal, no extra steps ...
- 0 kudos
- 173 Views
- 1 replies
- 1 kudos
VS Code - ipynb vs py execution - spark issue
Databricks Connect works inside VS Code notebook but the same code fails in a standalone script withValueError: default auth: cannot configure default credentialsI’m developing locally with **Databricks Connect 16.1.6** and VS Code.Inside a Jupyter n...
- 173 Views
- 1 replies
- 1 kudos
- 1 kudos
Hi @Sisi ,I think what's happening here is when you debug with option "Debug current file with Databricks Connect" then VS Code is using Databricks extension, which automatically handles authentication and sets up proper configuration.The regular Pyt...
- 1 kudos
- 185 Views
- 1 replies
- 1 kudos
Lakebase use cases
1. What are the use cases for Lakebase? When should I use the Lakebase Postgres over delta tables?2. What are the differences between open-source Postgres and Lakebase?3. Should I utilize Lakebase for all OLTP requirements?
- 185 Views
- 1 replies
- 1 kudos
- 1 kudos
Hi @Sharanya13 ,1. Use Lakebase whenever you have application workload (OLTP) and you require low latency. For analytical workloads use Lakehouse. Here you have couple of example use cases from documentation:Serving data and/or features from the lake...
- 1 kudos
- 304 Views
- 6 replies
- 2 kudos
Out of memory error when installing environment dependencies of UC Python UDF
Hi,I've created a small UC Python UDF to test whether it works with custom dependencies (new PP feature), and every time I'm getting OOM errors with this message: [UDF_ENVIRONMENT_USER_ERROR.OUT_OF_MEMORY] Failed to install UDF dependencies for <cata...
- 304 Views
- 6 replies
- 2 kudos
- 2 kudos
I tried with cluster, spent some couple of hours to load some libraries but unable to do. may be someone else can help you on this.
- 2 kudos
- 676 Views
- 2 replies
- 3 kudos
Resolved! Markdown Cells Do Not Render Consistently
When I am creating a notebook in the UI editor on DataBricks, markdown cells do not always render after I run them. They still appear in 'editing mode'. See the screenshot below, it should have rendered a H1.Again, this behavior is not consistent. So...
- 676 Views
- 2 replies
- 3 kudos
- 3 kudos
Hi @bdanielatl, Thank you for reporting the issue with markdown cells not rendering consistently. This appears to be a known issue that has been encountered by other users as well. I will report it internally.
- 3 kudos
- 611 Views
- 5 replies
- 4 kudos
Metastore deletion issues
Good afternoon, I have an issue with my metastore in North Europe.All my workspaces got detached:If I go to Databricks console, I can see the metastore in North Europe I created.However, when I select the metastore in North Europe, I get the followin...
- 611 Views
- 5 replies
- 4 kudos
- 4 kudos
I solved the issue by deleting all the asignments before deleting the metastore.1. Access to Databricks CLI and authenticate 2. List metastores>> databricks account metastores list 3. List wotrkspaces and check assignments>> databricks account worksp...
- 4 kudos
- 351 Views
- 1 replies
- 1 kudos
Drop schema or catalog using cascade function
Hello In Databricks (non-Unity Catalog), I have two schemas (schema_a and schema_b) that both use the same root location in DBFS or external storage like ADLS.Example:abfss://container@storage_account.dfs.core.windows.net/data/project/schema_aabfss:/...
- 351 Views
- 1 replies
- 1 kudos
- 1 kudos
Hello @EjB For the given example, here is the response:Will DROP SCHEMA schema_a CASCADE remove or affect tables in schema_b?No, unless:1. The tables in schema_a are managed tables, AND2. Tables in schema_b store their data physically inside /schema_...
- 1 kudos
- 217 Views
- 0 replies
- 0 kudos
Asset Bundle Include Glob paths not resolving recursive directories
Hello,When trying to include resource definitions in nested yaml files, the recursive paths I am specifying in the include section are not resolving as would be expected.With the include path resources/**/*.yml and a directory structure structure as ...
- 217 Views
- 0 replies
- 0 kudos
- 1475 Views
- 3 replies
- 0 kudos
[INTERNAL_ERROR] Query could not be scheduled: HTTP Response code: 503. Please try again later
We have a databricks job configured to run a dbt project. The dbt cli compute cluster being used is serverless with a serverless sql warehouse. We encountered this error during a run. SQLSTATE: XX000Any idea why this occurred?
- 1475 Views
- 3 replies
- 0 kudos
- 0 kudos
Hi,We are experiencing same recurring HTTP errors throughout the day when executing queries on Databricks. The specific error message we receive is: "[INTERNAL_ERROR] Query could not be scheduled: HTTP Response code: 503. Please try again later. SQLS...
- 0 kudos
- 582 Views
- 2 replies
- 3 kudos
Resolved! System tables performance optimization
Hi AllAre there any Databricks lab projects or GitHub repositories that leverage system tables to provide dashboards or code for monitoring, and more importantly, for optimizing workflows and clusters based on usage?
- 582 Views
- 2 replies
- 3 kudos
- 3 kudos
+1 to @szymon_dybczak. I would also add the dashboards for DB SQL Warehouse monitoring.
- 3 kudos
- 3789 Views
- 7 replies
- 1 kudos
Resolved! Creating Azure Databricks Workspace Without NAT Gateway
Hello,Recently, when I create a new Databricks workspace on Azure, it automatically create a NAT Gateway which incurs additional cost !When creating the workspace, I don't choose secure cluster connectivity, so I'm expecting not to have a NAT Gateway...
- 3789 Views
- 7 replies
- 1 kudos
- 1 kudos
Hello @OmarSEKkOUTI The solution is simple: try to create a vnet-injected workspace. like here: Azure Vnet injection This is to avoid creating a NAT gateway. Also, Databricks says:If you enable secure cluster connectivity on your workspace that use...
- 1 kudos
- 445 Views
- 2 replies
- 0 kudos
Databricks Apps with FastAPI
I have a FastAPI Databricks App deployed to a hosted workspace. How can I call the API from external tools like Postman?P.S. I was able to call the API within the same browser.
- 445 Views
- 2 replies
- 0 kudos
- 0 kudos
Hi @AndyK ,External tools don’t share browser’s login context, so you’ll need to add a bearer token in the Authorization header manually.Start by generating an access token. You can either use the Databricks CLI (databricks auth token) to get an OAut...
- 0 kudos
- 526 Views
- 6 replies
- 0 kudos
Unable to see sample data in Hive Metastore after moving to GCE
Hi,We have recently moved from GKE to GCE, it is taking forever to load the sample data in the manged delta tables.Even running simple select sql statements are taking forever. Totally clueless here, any help will be appreciatedThanks
- 526 Views
- 6 replies
- 0 kudos
- 0 kudos
Hi All,Strangely after struggle for 2 days we figured out that we can't run the cluster in scalable mode, so after selecting single node mode we are able to execute queries and job. It seems there is a bug in the Databrick's GKE to GCE migration. Won...
- 0 kudos
- 291 Views
- 1 replies
- 0 kudos
Best Approach to Retrieve Policy IDs Across Multiple Workspaces
Hi, I’m aware of the API endpoint api/2.0/policies/clusters/list to fetch a list of policy IDs and names. However, we have 50 different workspaces, and I need to retrieve the specific policy ID and name for each one.Could you advise on the most effic...
- 291 Views
- 1 replies
- 0 kudos
- 0 kudos
Hi @Chinu ,Databricks does not provide a global API to query all workspaces in a single call. I guess your only option for now is to use scripting approach.
- 0 kudos
- 381 Views
- 1 replies
- 0 kudos
Oauth Token federation
Dear allHas anyone tried oauth token federation for authentication with Databricks REST APIs?appreciate if there is a re-usable code snippet to achieve the same.
- 381 Views
- 1 replies
- 0 kudos
- 0 kudos
Documentation says - '''An account federation policy enables all users and service principals in your Azure Databricks account to access Databricks APIs using tokens from your identity provider. '''' But, in the examples that are provided, I do not s...
- 0 kudos
Join Us as a Local Community Builder!
Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!
Sign Up Now-
Access control
1 -
Apache spark
1 -
AWS
5 -
Azure
7 -
Azure databricks
5 -
Billing
2 -
Cluster
1 -
Compliance
1 -
Data Ingestion & connectivity
5 -
Databricks Runtime
1 -
Databricks SQL
2 -
DBFS
1 -
Dbt
1 -
Delta
4 -
Delta Sharing
1 -
DLT Pipeline
1 -
GA
1 -
Gdpr
1 -
Github
1 -
Partner
24 -
Public Preview
1 -
Service Principals
1 -
Unity Catalog
1 -
Workspace
2
- « Previous
- Next »
User | Count |
---|---|
52 | |
34 | |
25 | |
17 | |
10 |