cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

Niil
by New Contributor II
  • 2099 Views
  • 1 replies
  • 2 kudos

Resolved! AI Agents in ETL

Hi, I recently found a blog online about Databricks using AI Agents to automate ETL, but I can't find where these capabilities are located in Databricks. Does anyone know?Here is the blog,https://www.heliverse.com/blog/databricks-ai-agents-streamlini...

  • 2099 Views
  • 1 replies
  • 2 kudos
Latest Reply
santhakumar11
New Contributor III
  • 2 kudos

Hi NiilDatabrick has introduced AI Agent sub-categories as part of its Generative AI capabilities. We can now automate tasks such as Extract, Transformation, and Load (ETL).Ex,  Information Extraction Agent - we can able to transform large a volume  ...

  • 2 kudos
marcelhfm
by New Contributor II
  • 3028 Views
  • 7 replies
  • 0 kudos

Internal GRPC errors when using databricks connect

Hey there, in our local development flow we heavily rely on databricks asset bundles and databricks connect. Recently, locally run workflows (i.e. just pyspark python files) have begun to frequently fail with the following grpc error:pyspark.errors.e...

  • 3028 Views
  • 7 replies
  • 0 kudos
Latest Reply
ChrisChieu
Databricks Employee
  • 0 kudos

@marcelhfm it might be a Spark Connect issueI would say it is the same for the rest of you, guys Nothing much to do until the situation is fixed by Databricks 

  • 0 kudos
6 More Replies
sparkycloud
by New Contributor II
  • 1743 Views
  • 3 replies
  • 1 kudos

Observability and monitoring accross multiple workspaces(both job clusters and serverless compute)

Hi all, Today what are the best option available today for observability and monitoring databricks jobs accross all workspaces. We have 100 of workspaces and it hard to do monitoring to check failed and successeded jobs.We tried using: 1. Team webhoo...

  • 1743 Views
  • 3 replies
  • 1 kudos
Latest Reply
sparkycloud
New Contributor II
  • 1 kudos

Hey Brahmareddy, Thanks so much for responding. Sorry I forgot to mention we are in Azure. Lets go through one by one.1. Audit logs(Azure Monitor) : AFAIK this requires init scripts  and jar build that will not support in serverless or its not the ca...

  • 1 kudos
2 More Replies
ashraf1395
by Honored Contributor
  • 1840 Views
  • 2 replies
  • 1 kudos

Resolved! Not readable format of event_logs traceback in dlt pipeline

This is my dlt pipeline event_log - why is it not in readable foramt how can i correct it.This is my pipeline code :   import logging logger = logging.getLogger(__name__) logger.info("Error") raise "Error is error"    

ashraf1395_0-1740044260033.png
  • 1840 Views
  • 2 replies
  • 1 kudos
Latest Reply
jorperort
Contributor
  • 1 kudos

Hi @ashraf1395 ,I'm working with Delta Live Tables (DLT) and the event_log table. I would like to know if it is possible to access the event handler that DLT uses to write custom logs and send them to this table when events are published.If this is n...

  • 1 kudos
1 More Replies
susanne
by Contributor
  • 1106 Views
  • 2 replies
  • 1 kudos

How to write event_log destination into DLT Settings JSON via Asset Bundles

Hi all,I would like to publish the event_log of my DLT Pipeline to a specific schema in Unity Catalog.Following this article (https://docs.databricks.com/gcp/en/dlt/observability#query-the-event-log) this can be done by writing this into the DLTs set...

  • 1106 Views
  • 2 replies
  • 1 kudos
Latest Reply
ashraf1395
Honored Contributor
  • 1 kudos

Hi @susanne , indeed , i tried to create it using dabs as well. This feature is not available using dabs I guess, maybe they will add it once event_logs moves to ga from public preview.databricks API will be a good alternative but if you try it using...

  • 1 kudos
1 More Replies
21f3001806
by New Contributor III
  • 1060 Views
  • 1 replies
  • 1 kudos

Resolved! Creating event_log for dlt pipeline using dabs

Hi there, i have dlt pipeline and I recently came to know about the event_log feature, i want to deploy my dlt pipeline along with the event_log using databricks asset bundles but i am not able to find any resources for it.If anyone has tried it , yo...

  • 1060 Views
  • 1 replies
  • 1 kudos
Latest Reply
ashraf1395
Honored Contributor
  • 1 kudos

Hi there @21f3001806 ,I guess you are talking about the this : https://docs.databricks.com/api/workspace/pipelines/create#event_logIts still in public preview I tried creating it through UI or by chaing the pipeline settings it worked. But when I imp...

  • 1 kudos
DBStudent
by New Contributor II
  • 902 Views
  • 3 replies
  • 0 kudos

Data migration from S3 to Databricks

I currently have an S3 bucket with around ~80 tables, each of which has hive-style partition columns  S3RootFolder/Table1Name/Year=2024/Month=12/Day=1/xxx.parquetS3RootFolder/Table1Name/Year=2024/Month=12/Day=2/xxx.parquetS3RootFolder/Table2Name/Year...

  • 902 Views
  • 3 replies
  • 0 kudos
Latest Reply
Brahmareddy
Esteemed Contributor
  • 0 kudos

Hi DBStudent,You’re absolutely right—coalescing during write won’t help with the initial bottleneck, since the issue is really with the read side, where Spark has to list and open 110k small files. If you can’t pre-compact them before reading, then o...

  • 0 kudos
2 More Replies
Volker
by Contributor
  • 2322 Views
  • 1 replies
  • 0 kudos

From Partitioning to Liquid Clustering

We had some delta tables that where previously partitioned on year, month, day, and hour. This resulted in quite small partitions and we now switched to liquid clustering.We followed these steps:Remove partitioning by doing REPLACEALTER TABLE --- CLU...

  • 2322 Views
  • 1 replies
  • 0 kudos
Latest Reply
Isi
Honored Contributor III
  • 0 kudos

Hey @Volker ,First of all, I’d recommend considering Auto Liquid Clustering, as it can simplify the process of defining clustering keys.You can read more about it in the Databricks documentation (it’s currently in Public Preview, but you can probably...

  • 0 kudos
PavloR
by New Contributor III
  • 3237 Views
  • 10 replies
  • 6 kudos

ISOLATION_STARTUP_FAILURE for Serverless

Hi, I'm getting an error in my job which is using Serverless cluster: `[ISOLATION_STARTUP_FAILURE] Failed to start isolated execution environment. Please contact Databricks support. SQLSTATE: XXKSS`It was working fine and then, this error appeared on...

  • 3237 Views
  • 10 replies
  • 6 kudos
Latest Reply
dileeshm
New Contributor II
  • 6 kudos

@Alberto_Umana Looks like the issue resurfaced again. We started seeing "[ISOLATION_STARTUP_FAILURE] Failed to start isolated execution environment. Please contact Databricks support. SQLSTATE: XXKSS`" error with serverless compute from last night.Co...

  • 6 kudos
9 More Replies
tingwei
by New Contributor II
  • 7263 Views
  • 5 replies
  • 5 kudos

ISOLATION_STARTUP_FAILURE

Hi I'm getting error in my data pipeline[ISOLATION_STARTUP_FAILURE] Failed to start isolated execution environment. Please contact Databricks support. SQLSTATE: XXKSSit was working fine and suddenly it is keep failing. Please advice. 

  • 7263 Views
  • 5 replies
  • 5 kudos
Latest Reply
Jean-Boutros
New Contributor II
  • 5 kudos

I am running on serverless and only yesterday I started seeing this error. Any thoughts?[ISOLATION_STARTUP_FAILURE] Failed to start isolated execution environment. Please contact Databricks support. SQLSTATE: XXKSS

  • 5 kudos
4 More Replies
jeffn
by New Contributor
  • 758 Views
  • 1 replies
  • 0 kudos

Unable to create Cluster via REST API

I have tried all 3 of the example payloads listed here to create a Compute Cluster: https://docs.databricks.com/api/azure/workspace/clusters/createAll return the same error: Invalid JSON given in the body of the request - expected a mapOther Compute ...

  • 758 Views
  • 1 replies
  • 0 kudos
Latest Reply
Panda
Valued Contributor
  • 0 kudos

@jeffn  - I was able to successfully create a cluster using the same Azure documentation, both via the REST API and the Databricks SDK. Hopefully it helps.from databricks.sdk import WorkspaceClient from databricks.sdk.service.compute import CreateClu...

  • 0 kudos
sriramnedunuri
by New Contributor III
  • 872 Views
  • 1 replies
  • 0 kudos

Resolved! Regexp_replaces pattern only once

select  regexp_replace('asdfhsdf&&1&&asdfasdf&&2&&asdf','&&[0-100]&&','') output here it replaces first pattern but not the 2nd patterno/poutputasdfhsdfasdfasdf&&2&&asdf  even if we use position its not working  something like  both wont work select ...

  • 872 Views
  • 1 replies
  • 0 kudos
Latest Reply
sriramnedunuri
New Contributor III
  • 0 kudos

changing to [0-9] works this can be closed.

  • 0 kudos
dkxxx-rc
by Contributor
  • 7542 Views
  • 3 replies
  • 2 kudos

Resolved! CREATE TEMP TABLE

The Databricks assistant tells me (sometimes) that `CREATE TEMP TABLE` is a valid SQL operation.  And other sources (e.g., https://www.freecodecamp.org/news/sql-temp-table-how-to-create-a-temporary-sql-table/) say the same.But in actual practice, thi...

  • 7542 Views
  • 3 replies
  • 2 kudos
Latest Reply
dkxxx-rc
Contributor
  • 2 kudos

In addition to accepting KaranamS's answer, I will note a longer and useful discussion, with caveats, at https://community.databricks.com/t5/data-engineering/how-do-temp-views-actually-work/m-p/20137/highlight/true#M13584.

  • 2 kudos
2 More Replies
Upendra_Dwivedi
by Contributor
  • 1667 Views
  • 5 replies
  • 0 kudos

Remote SQL Server Instance Connection using JDBC

Hi All,I am connecting Remote SQL Server Instance using JDBC Driver, I have enabled TCP/IP and Setup the Firewall Rule. When i am querying the instance i am getting this error:(com.microsoft.sqlserver.jdbc.SQLServerException) The TCP/IP connection to...

  • 1667 Views
  • 5 replies
  • 0 kudos
Latest Reply
turagittech
Contributor
  • 0 kudos

If you want to access a local SQL server, you'll need a Private Link to access the server. If it's on your own local machine, that's likely not possible. Creating a VPN to your machine is a unique problem, and you would be better off using a VM or a ...

  • 0 kudos
4 More Replies
sanq
by New Contributor II
  • 7075 Views
  • 3 replies
  • 7 kudos

what formatter is used to format SQL cell in databricks

Databricks launched formatter Black which formats python cells, I can also see SQL cell getting formatted, but not sure which formatter is being used for SQL cell formatting. No clarity given on docs.

  • 7075 Views
  • 3 replies
  • 7 kudos
Latest Reply
mitch_DE
New Contributor II
  • 7 kudos

The formatter is mentioned here: Develop code in Databricks notebooks - Azure Databricks | Microsoft LearnIt is this npm package: @gethue/sql-formatter - npm

  • 7 kudos
2 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels