cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

mgcasas-aws
by New Contributor
  • 1879 Views
  • 1 replies
  • 1 kudos

Resolved! Azure Databricks Serverless private connection to S3 bucket

I'm looking for technical references to connect an Azure Databricks serverless workspace to an S3 bucket over a private site-to-site VPN connection. Found the following to connect AWS (consumer) to Azure (provider), but I'm looking for the other way....

  • 1879 Views
  • 1 replies
  • 1 kudos
Latest Reply
Sai_Ponugoti
Databricks Employee
  • 1 kudos

Hello @mgcasas-aws Thank you for your question! We’re currently working on a solution for private cross-cloud Delta Sharing (Azure → AWS). In the meantime, here’s a possible approach: Update your Azure Storage Account network settings from private e...

  • 1 kudos
Rainer
by New Contributor
  • 589 Views
  • 2 replies
  • 0 kudos

pyspark.testing.assertSchemaEqual() ignoreColumnOrder parameter exists in 3.5.0 only on Databricks

Hi, I am using the pyspark.testing.assertSchemaEqual() function in my code using the ignoreColumnOrder parameter that is available since pyspark 4.0.0. https://spark.apache.org/docs/4.0.0/api/python/reference/api/pyspark.testing.assertSchemaEqual.htm...

  • 589 Views
  • 2 replies
  • 0 kudos
Latest Reply
saurabh18cs
Honored Contributor II
  • 0 kudos

Hi @Rainer When you use Databricks Connect, your local code is executed against the Databricks cluster, which uses the Databricks Runtime’s PySpark, not your local PySpark installation. meaning your master driver node is also running on remote comput...

  • 0 kudos
1 More Replies
karthikmani
by New Contributor
  • 314 Views
  • 1 replies
  • 0 kudos

git actions to deploy dabs

HI All,I am trying to create a dabs deployment via git actions. However I am getting below error constantly. Requesting your suggestion on what I am doing wrong here? Thanks.Note we are using OIDC authentication from git. Our company has disabled tok...

  • 314 Views
  • 1 replies
  • 0 kudos
Latest Reply
saurabh18cs
Honored Contributor II
  • 0 kudos

Hi @karthikmani Databricks CLI is not able to find or use the right authentication method/environment variables for OIDC. can you try with this : In your databricks.yaml, you have auth_type: github-oidc and in your workflow you use DATABRICKS_AUTH_TY...

  • 0 kudos
rohith_23
by New Contributor II
  • 429 Views
  • 3 replies
  • 2 kudos

Resolved! org.apache.hadoop.hive.ql.metadata.HiveException: MetaException

Hi Data Enthusiasts,I have been facing few errors in SQL warehouse for quiet a long time and its happening pretty randomly.We checked query runs and captured the errors below.I believe this is something to do with hive. And I am facing this when ther...

  • 429 Views
  • 3 replies
  • 2 kudos
Latest Reply
NandiniN
Databricks Employee
  • 2 kudos

Hi @rohith_23 , These errors all relate to problems communicating with the Hive Metastore Service (HMS), which is the central component to store metadata (schemas, table locations, column types, etc.) about your tables. The core of the issue in all t...

  • 2 kudos
2 More Replies
GastonClynhens
by New Contributor III
  • 469 Views
  • 1 replies
  • 1 kudos

Resolved! Power BI refresh history info is different from ADF monitor info

In Azure Data Factory, I have a pipeline defined with Bronze - Silver - Gold layers + the final step 4 entails the refresh of a Power BI semantic model.This final step is executed via a Databricks notebook and contains the following tasks:# getting p...

  • 469 Views
  • 1 replies
  • 1 kudos
Latest Reply
GastonClynhens
New Contributor III
  • 1 kudos

considering the logging of the executed notebook:Failed to refresh dataset: {"error":{"code":"ItemNotFound","message":"Dataset \"xxxxxxxxxxxxxxxx\" is not found!Please verify datasetId is correct and user have sufficient permissions."}}The dataset wa...

  • 1 kudos
paulchen
by New Contributor
  • 1919 Views
  • 1 replies
  • 1 kudos

Resolved! Service principle used in Bitbucket CICD pipelines not working

The Databricks Asset Bundle is used for the Bitbucket CICD pipelines.The service principle is used in both of the local Databricks configuration and Bitbucket CICD environment.The service principle is only working in the local environment for deploym...

  • 1919 Views
  • 1 replies
  • 1 kudos
Latest Reply
sarahbhord
Databricks Employee
  • 1 kudos

Hey PaulChen Having your Databricks service principal (SP) work locally but fail in Bitbucket CI/CD usually means environment variables aren’t set up right, or the pipeline is falling back to an unexpected config.   Quick checklist: Set all SP creden...

  • 1 kudos
cz0
by New Contributor
  • 2353 Views
  • 2 replies
  • 1 kudos

Monitoring structured streaming and Log4J properties

Hi guys, I would like to monitor streaming job on metrics like delay, processing time and more. I found this documentation but I get message on starting and terminating phase and not while I process a record. The job is a pretty easy streaming which ...

  • 2353 Views
  • 2 replies
  • 1 kudos
Latest Reply
saurabh18cs
Honored Contributor II
  • 1 kudos

Hi @cz0  The StreamingQueryListener in Spark is designed to give you metrics at the micro-batch level (not per individual record), which is typical for Spark Structured StreamingonQueryStarted: Called when the streaming job starts.onQueryProgress: Ca...

  • 1 kudos
1 More Replies
jeff2
by New Contributor
  • 1500 Views
  • 1 replies
  • 1 kudos

Resolved! When embedding redash, is it possible to make it visible without an account?

Same as the title.I created a redash in Databricks and want to embed it to show it in another portal. However, other users have accounts in the portal but not in Databricks.In this case, is it possible to show Redash to all these users?

  • 1500 Views
  • 1 replies
  • 1 kudos
Latest Reply
mark_ott
Databricks Employee
  • 1 kudos

It is possible to embed a Redash dashboard created in Databricks into another portal so that users without Databricks accounts can view it, but this requires specific setup and permission management. How Embedding Works To show the dashboard to users...

  • 1 kudos
TinaDouglass
by New Contributor
  • 345 Views
  • 3 replies
  • 1 kudos

Resolved! Summarized Data from Source system into Bronze

Hello,We are just starting with Databricks. Quick question.  We have a table in our legacy source system that summarizes values that are used on legacy reports and used for payment in our legacy system.  The business wants a dashboard on our new plat...

  • 345 Views
  • 3 replies
  • 1 kudos
Latest Reply
Advika
Databricks Employee
  • 1 kudos

Hello @TinaDouglass! Did the suggestions shared above help address your concern? If so, please consider marking one as the accepted solution.

  • 1 kudos
2 More Replies
juanjomendez96
by Contributor
  • 1016 Views
  • 16 replies
  • 4 kudos

Resolved! Control Databricks Platform version

Hello there! I have noticed that my Databricks UI has been changed from time to time, I was wondering how can I control the Databricks Platform version so I don't keep having new changes and new ways/names in my UI. I have found a release page https:...

  • 1016 Views
  • 16 replies
  • 4 kudos
Latest Reply
georgeb
New Contributor II
  • 4 kudos

Hi,  Can we get an official feedback when the issue with adding new users/groups in Databricks apps UI (not working) will be fixed? I tried with python SDK as well and it does not work.  The issue was posted previously and applies to my case as well....

  • 4 kudos
15 More Replies
bhargavabasava
by New Contributor III
  • 297 Views
  • 1 replies
  • 0 kudos

Support for JDBC writes from serverless compute

Hi team,Are there any plans in place to support JDBC writes using serverless compute.

  • 297 Views
  • 1 replies
  • 0 kudos
Latest Reply
saurabh18cs
Honored Contributor II
  • 0 kudos

Databricks Serverless SQL supports JDBC reads (queries) via the Databricks SQL endpoints, but JDBC writes (inserts/updates) directly from serverless compute are not generally supported. This is a common limitation because serverless SQL endpoints are...

  • 0 kudos
y_sanjay
by New Contributor
  • 2360 Views
  • 2 replies
  • 2 kudos

Temporary view

Hi,I wrote a query to create temp view in my catalog, query execution was successful and returned the result as 'OK' in SQL editor window. However, when I executed the command 'Show Tables' and' Select * {temp_view}', it's not identifying the view. W...

  • 2360 Views
  • 2 replies
  • 2 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 2 kudos

Hi @y_sanjay ,I guess that is somehow related to how sessions are managed within SQL Editor. For instance when I ran following queries in SQL Editor all at once it worked and I've got 3 result sets:1) First result set with OK status - which means tha...

  • 2 kudos
1 More Replies
Data_NXT
by New Contributor III
  • 487 Views
  • 3 replies
  • 5 kudos

Resolved! Databricks Business dashboards - Interactive cluster Total dollar spent

I'm working on Databricks Business Dashboards and trying to calculate interactive cluster compute time and total dollar spend per workspace.As per standard understanding, the total dollar spent = Interactive Clusters + Job Clusters + SQL Warehouses.I...

  • 487 Views
  • 3 replies
  • 5 kudos
Latest Reply
nayan_wylde
Honored Contributor III
  • 5 kudos

Also the system table will not provide you the  exact dollar amount that you spend in an interactive compute. Here is the cost breakdown for running interactive compute:ComponentDescriptionCost SourceDBU CostBased on workload type and tierDatabricksV...

  • 5 kudos
2 More Replies
Sunil_Poluri
by New Contributor
  • 1356 Views
  • 1 replies
  • 1 kudos

Resolved! Unexpected Schema ID Folder Creation in Unity Catalog External Location

I've set up Unity Catalog with an external location pointing to a storage account. For each schema, I’ve configured a dedicated container path. For example:abfss://schemas@<storage_account>.dfs.core.windows.net/_unityStorage/schemas/<schema_id>When I...

  • 1356 Views
  • 1 replies
  • 1 kudos
Latest Reply
Louis_Frolio
Databricks Employee
  • 1 kudos

Hey @Sunil_Poluri , I did some research (learned a few things) and here is what I found.  Unity Catalog manages cloud storage mapping for schemas using internal IDs (schema_id) to ensure data isolation, governance, and uniqueness within a metastore—e...

  • 1 kudos
Anubhav2011
by New Contributor II
  • 798 Views
  • 5 replies
  • 4 kudos

What is the Power of DLT Pipeline to read streaming data

I am getting thousands of records every second in my bronze table from Qlik and every second the bronze table is getting truncated and load with new data by Qlik itself. How do I process this much data every second to my silver streaming table before...

  • 798 Views
  • 5 replies
  • 4 kudos
Latest Reply
Krishna_S
Databricks Employee
  • 4 kudos

The Apply Changes API is getting deprecated. The AUTO CDC APIs replace the APPLY CHANGES APIs, and have the same syntax. The APPLY CHANGES APIs are still available, but Databricks recommends using the AUTO CDC APIs in their place. Please refer to the...

  • 4 kudos
4 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels