cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

harvey-c
by New Contributor III
  • 1768 Views
  • 0 replies
  • 0 kudos

Wrong FS: abfss://....., expected: dbfs:/ Error in DLT pipeline

Dear Databricks community members:SymptomReceived the error for a delta load, after a successful initial load with a  Unity Catalog Volume as a data source.org.apache.spark.sql.streaming.StreamingQueryException: [STREAM_FAILED] Query [id = xxx, runId...

  • 1768 Views
  • 0 replies
  • 0 kudos
GijsM
by New Contributor
  • 4050 Views
  • 1 replies
  • 0 kudos

Thousands of ETL pipelines with long execution times and small dataset sizes

Hi,I work for a small company, we're mostly focussing on small retail and e-commerce customers. We provides data analysis and automated data connections between their platforms. Most of our datasets are things like order data, google ads click data, ...

  • 4050 Views
  • 1 replies
  • 0 kudos
Latest Reply
brockb
Databricks Employee
  • 0 kudos

Hi, Thanks for the information, there is a lot to unpack and some assumptions that need to be made without fully understanding the details, so here are a few thoughts: If the cluster start times longer because of the libraries you're installing, can ...

  • 0 kudos
Phani1
by Databricks MVP
  • 3338 Views
  • 1 replies
  • 0 kudos

Query Delta table from .net

Hi Team,How can expose data stored in delta table through API like exposing sql data through .net api?

Data Engineering
delta
dotnet
  • 3338 Views
  • 1 replies
  • 0 kudos
Latest Reply
BjarkeM
New Contributor III
  • 0 kudos

You can use the SQL Statement Execution API.At energinet.dk we have created this open-source .NET client, which we use internally in the company.

  • 0 kudos
-werners-
by Esteemed Contributor III
  • 8281 Views
  • 2 replies
  • 3 kudos

Resolved! best way to store config files in a Unity workspace (Scala/typesafe)

We use typesafe (scala) to read configuration values from hocon files.When not using Unity, we read the configuration files from /dbfs/...  works fine.However, with Unity, usage of dbfs is frowned upon.So I started looking into alternatives.And unfor...

  • 8281 Views
  • 2 replies
  • 3 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 3 kudos

In the end we will continue to use dbfs.  Maybe in the future when volumes are supported by scala io we can re-evaluate, but for now dbfs seems the way to go.

  • 3 kudos
1 More Replies
mudholkar
by New Contributor III
  • 3876 Views
  • 1 replies
  • 6 kudos

I am getting an SSLError: HTTPSConnectionPool while making a call to https restapis from azure databricks I have tried to set a verify=false parameter in the call too.

response = requests.request("POST", url, verify=False, headers=headers, data=payload)   SSLError: HTTPSConnectionPool(host='dcs.adobedc.net', port=443): Max retries exceeded with url: /collection/d99e6dfcffb0b5aeaec2cf76cd3bc2b9e9c414b0c74a528d13dd39...

  • 3876 Views
  • 1 replies
  • 6 kudos
Latest Reply
JFG
New Contributor II
  • 6 kudos

Any luck with this?

  • 6 kudos
mkrish28
by Databricks Partner
  • 2515 Views
  • 2 replies
  • 0 kudos

Resolved! Regarding Exam got suspended

Hello Team,I had a disappointing experience while attempting my first DataBricks certification. Abruptly, the proctor asked me to show my desk, and after complying. Eventually, they suspended my exam, citing excessive eye movement and other practices...

  • 2515 Views
  • 2 replies
  • 0 kudos
Latest Reply
Cert-Team
Databricks Employee
  • 0 kudos

@mkrish28 I'm sorry to hear you had this experience. Thank you for logging at ticket with the support team. They have informed me they have rescheduled your exam. Good luck!

  • 0 kudos
1 More Replies
samur
by New Contributor II
  • 2299 Views
  • 1 replies
  • 1 kudos

DBR 14.1 - foreachBatch in Spark Connect Shared Clusters are not supported in Unity Catalog.

I am getting this error on DBR 14.1AnalysisException: [UC_COMMAND_NOT_SUPPORTED.WITHOUT_RECOMMENDATION] The command(s): foreachBatch in Spark Connect Shared Clusters are not supported in Unity Catalog.This is the code: wstream = df.writeStream.foreac...

  • 2299 Views
  • 1 replies
  • 1 kudos
Iam
by New Contributor II
  • 2231 Views
  • 1 replies
  • 0 kudos

CANNOT_RENAME_ACROSS_SCHEMA message error

Hello...We enabled Unity Catalog and we are migrating schemas. When I ran the command sync schema catalog01.schema01 FROM hive_metastore.schema01 dry run  I got the error CANNOT_RENAME_ACROSS_CATALOG, reviewing your documentation it only said   CANNO...

  • 2231 Views
  • 1 replies
  • 0 kudos
PetitLepton
by New Contributor II
  • 9338 Views
  • 1 replies
  • 0 kudos

List parameter in Python SQL connector 3.0.1

Hi,up to recently in version of the Python SQL connector 2.9.3, I was using a list as a parameter in the cursor.execute(operation, parameters) method without any trouble. It seems that it is not possible anymore in version 3.0.1 as the parsing of par...

  • 9338 Views
  • 1 replies
  • 0 kudos
Latest Reply
PetitLepton
New Contributor II
  • 0 kudos

I should better read the documentation : https://github.com/databricks/databricks-sql-python/blob/v3.0.0/docs/parameters.md. 

  • 0 kudos
dcardenas
by New Contributor
  • 978 Views
  • 0 replies
  • 0 kudos

Retrieving Logs with Job API Get-outputs service

Hello,I would like to retrieve the logs of some job that where launched using the Job Rest Api 2.0. I see in the doc that can be done with the service get-ouputs, however each time I check the service I just get the metadata part of the response but ...

  • 978 Views
  • 0 replies
  • 0 kudos
ken2
by Databricks Partner
  • 3247 Views
  • 3 replies
  • 0 kudos

How to convert entity_id to notebook name or job

Hi, Databricks developers!I use system.access.table_lineage refering to this page.It's difficult for us to recognize which notebook was indicated by the entity_id.How do I get the table to convert entity_ids to Job names or Notebook names?

  • 3247 Views
  • 3 replies
  • 0 kudos
Latest Reply
mlamairesse
Databricks Employee
  • 0 kudos

Workflows system tables are coming very soon. 

  • 0 kudos
2 More Replies
cg3
by New Contributor
  • 1129 Views
  • 0 replies
  • 0 kudos

Define VIEW in Databricks Asset Bundles?

Is it possible to define a Unity Catalog VIEW in a Databricks Asset Bundle, or specify in the bundle that a specific notebook gets run once per deployment?

  • 1129 Views
  • 0 replies
  • 0 kudos
Kishan1003
by New Contributor
  • 3991 Views
  • 1 replies
  • 0 kudos

Merge Operation is very slow for S/4 Table ACDOCA

Hello,we have a scenario in Databricks where every day  we get 60-70 million records  and it takes a lot of time to merge the data into 28 billion records which is already sitting there . The time taken to rewrite the files which are affected is too ...

  • 3991 Views
  • 1 replies
  • 0 kudos
Latest Reply
177991
New Contributor II
  • 0 kudos

Hi @Kishan1003  did you find something helpful? Im dealing with a similar situation, acdoca table on my side is around 300M (fairly smaller), and incoming daily data is usually around 1M. I have try partition using period, like fiscyearper column, zo...

  • 0 kudos
costi9992
by Databricks Partner
  • 7242 Views
  • 6 replies
  • 0 kudos

Resolved! Add policy init_scripts.*.volumes.destination for dlt not working

Hi,I tried to create a policy to use it for DLTs that are ran with shared clusters, but when i run the DLT with this policy I have an error. Init-script is added to Allowed JARs/Init Scripts.DLT events error: Cluster scoped init script /Volumes/main/...

  • 7242 Views
  • 6 replies
  • 0 kudos
Latest Reply
ayush007
New Contributor II
  • 0 kudos

@costi9992I am facing same issue with UC enabled cluster with 13.3 Databricks Runtime.I have uploaded the init shell script in Volume with particular init script allowed by metastore admin.But I get the same error as you stated .When I looked in clus...

  • 0 kudos
5 More Replies
shivam-singh
by New Contributor
  • 1548 Views
  • 1 replies
  • 0 kudos

Databricks-Autoloader-S3-KMS

Hi, I am working on a requirement where I am using autoloader in a DLT pipeline to ingest new files as they come.This flow is working fine. However I am facing an issue, when we have the source bucket an s3 location, since the bucket is having a SSE-...

  • 1548 Views
  • 1 replies
  • 0 kudos
Latest Reply
kulkpd
Contributor
  • 0 kudos

Can you please paste the exact errors and check below things:check following if its related to KMS:1. IAM role policy and KMS policy should have allow permissions2. Did you use extraConfig while mounting the source-s3 bucket:If you have used IAM role...

  • 0 kudos
Labels