cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

lawrence009
by Contributor
  • 4786 Views
  • 5 replies
  • 1 kudos

Contact Support re Billing Error

How do I contact billing support? I am billed through AWS Marketplace and noticed last month the SQL Pro discount is not being reflected in my statement.

  • 4786 Views
  • 5 replies
  • 1 kudos
Latest Reply
santiagortiiz
Databricks Partner
  • 1 kudos

Hi, could anybody provide a contact email? I have sent emails to many contacts described in the support page here and in AWS, but no response from any channel. My problem is that databricks charged me by the resources used during a free trial, what i...

  • 1 kudos
4 More Replies
LukeD
by New Contributor II
  • 3267 Views
  • 3 replies
  • 1 kudos

Billing support contact

Hi,What is the best way to contact Databricks support? I see the differences between AWS billing and Databricks report and I'm looking for explanation of that. I've send 3 messages last week by this form https://www.databricks.com/company/contact but...

  • 3267 Views
  • 3 replies
  • 1 kudos
Latest Reply
santiagortiiz
Databricks Partner
  • 1 kudos

Hi, I'm facing the same issue with signing in my workspace, and I have a billing error, databricks charged me for a free trial, and I have sent a lot of emails, posted a topic in the community, I contacted people in AWS and they said that it must be ...

  • 1 kudos
2 More Replies
MCosta
by New Contributor III
  • 16957 Views
  • 10 replies
  • 19 kudos

Resolved! Debugging!

Hi ML folks, We are using Databricks to train deep learning models. The code, however, has a complex structure of classes. This would work fine in a perfect bug-free world like Alice in Wonderland. Debugging in Databricks is awkward. We ended up do...

  • 16957 Views
  • 10 replies
  • 19 kudos
Latest Reply
petern
New Contributor II
  • 19 kudos

Has this been solved yet; a mature way to debug code on databricks. I'm running in the same kind of issue.Variable explorer can be used and pdb, but not the same really..

  • 19 kudos
9 More Replies
DatBoi
by Contributor
  • 7032 Views
  • 2 replies
  • 2 kudos

Resolved! How big should a delta table be to benefit from liquid clustering?

My questions is pretty straightforward - how big should a delta table be to benefit from liquid clustering? I know the answer will most likely depend on the details of how you are querying the data, but what is the recommendation?I know Databricks re...

  • 7032 Views
  • 2 replies
  • 2 kudos
Latest Reply
daniel_sahal
Databricks MVP
  • 2 kudos

@DatBoi Once you watch this video you'll understand more about Liquid Clustering https://www.youtube.com/watch?v=5t6wX28JC_M&ab_channel=DeltaLakeLong story short:I know Databricks recommends not partitioning on tables less than 1 TB and aim for 1 GB ...

  • 2 kudos
1 More Replies
demost11
by New Contributor II
  • 1077 Views
  • 0 replies
  • 0 kudos

Tracking DBMS CDC

We're using Databricks to incrementally extract data from SQL Server tables into S3. The data contains a timestamp column. We need a place to store the maximum retrieved timestamp per table so it can retrieved during the next run.Does Databricks cont...

  • 1077 Views
  • 0 replies
  • 0 kudos
Jagan_etl
by New Contributor II
  • 2446 Views
  • 3 replies
  • 0 kudos

Avro file format generation

Hi All,We are using cluster with 9.1 run time version, I'm getting "incompatible schema exception" error while writing the data into avro file. Fields in Avro schema are more compared to dataframe output Fields. I tried the same in community edition ...

  • 2446 Views
  • 3 replies
  • 0 kudos
Latest Reply
Jagan_etl
New Contributor II
  • 0 kudos

Hi All,Any suggestions on this.

  • 0 kudos
2 More Replies
BhaveshPatel
by New Contributor
  • 2162 Views
  • 1 replies
  • 1 kudos

Auto loader

Suppose I have 1000's of historical .csv files stored from Jan, 2022 in a folder of my azure blob storage container. I want to use auto loader to read files beginning only on 1st, Oct, 2023 and ignoring all the files before this date to build a pipel...

  • 2162 Views
  • 1 replies
  • 1 kudos
Latest Reply
daniel_sahal
Databricks MVP
  • 1 kudos

@BhaveshPatel Three things that you can do:- Move the files to the separate folder,- Use a filter on metadata fields to filter out the unnecessary files,- Use a pathGlobFilter to filter only on the files you need

  • 1 kudos
Bharathi7
by Databricks Partner
  • 2726 Views
  • 3 replies
  • 0 kudos

Python UDF fails with UNAVAILABLE: Channel shutdownNow invoked

I'm using a Python UDF to apply OCR to each row of a dataframe which contains the URL to a PDF document. This is how I define my UDF:  def extract_text(url: str): ocr = MyOcr(url) extracted_text = ocr.get_text() return json.dumps(extracte...

  • 2726 Views
  • 3 replies
  • 0 kudos
Latest Reply
daniel_sahal
Databricks MVP
  • 0 kudos

@Bharathi7 It's really hard to determine what's going on without knowing what acutally MyOcr function does.Maybe there's some kind of timeout on the service side? To many parallell connections?

  • 0 kudos
2 More Replies
Poovarasan
by Databricks Partner
  • 2955 Views
  • 1 replies
  • 0 kudos

com.databricks.sql.transaction.tahoe.ColumnMappingException: Found duplicated column id `2` in colum

Hi,Currently, I am using the below-mentioned query to create a materialized view. It was working fine until yesterday in the DLT pipeline, but from today on, the below-provided code throws an error (com.databricks.sql.transaction.tahoe.ColumnMappingE...

Data Engineering
ColumnMapping
dlt
  • 2955 Views
  • 1 replies
  • 0 kudos
elgeo
by Valued Contributor II
  • 28754 Views
  • 3 replies
  • 2 kudos

Data type length enforcement

Hello. Is there a way to enforce the length of a column in SQL? For example that a column has to be exactly 18 characters? Thank you!

  • 28754 Views
  • 3 replies
  • 2 kudos
Latest Reply
databricks31
Databricks Partner
  • 2 kudos

we are facing similar issues while write into adls location delta format, after that we created on top delta location unity catalog tables. below format of data type length should be possible to change spark sql supported ?Azure SQL Spark            ...

  • 2 kudos
2 More Replies
Ajay-Pandey
by Databricks MVP
  • 2790 Views
  • 2 replies
  • 7 kudos

docs.databricks.com

Rename and drop columns with Delta Lake column mapping. Hi all,Now databricks started supporting column rename and drop.Column mapping requires the following Delta protocols:Reader version 2 or above.Writer version 5 or above.Blog URL##Available in D...

  • 2790 Views
  • 2 replies
  • 7 kudos
Latest Reply
Poovarasan
Databricks Partner
  • 7 kudos

Above mentioned feature is not working in the DLT pipeline. if the scrip has more than 4 columns 

  • 7 kudos
1 More Replies
seydouHR
by New Contributor III
  • 5131 Views
  • 3 replies
  • 0 kudos

Resolved! CLONE not supported on delta table with Liquid Clustering

Hello all,We are building a data warehouse on Unity Catalog and we use the SHALLOW CLONE command to allow folks to spin up their own dev environments by light copying the prod tables. We also started using Liquid Clustering on our feature tables, tho...

  • 5131 Views
  • 3 replies
  • 0 kudos
Latest Reply
seydouHR
New Contributor III
  • 0 kudos

Thanks Kaniz for your reply. I was able to get it make it work using runtime 14.0.Regards, 

  • 0 kudos
2 More Replies
caldempsey
by New Contributor
  • 4110 Views
  • 0 replies
  • 0 kudos

Delta Lake Spark fails to write _delta_log via a Notebook without granting the Notebook data access

I have set up a Jupyter Notebook w/ PySpark connected to a Spark cluster, where the Spark instance is intended to perform writes to a Delta table.I'm observing that the Spark instance fails to complete the writes if the Jupyter Notebook doesn't have ...

Data Engineering
deltalake
Docker
spark
  • 4110 Views
  • 0 replies
  • 0 kudos
chrisf_sts
by Databricks Partner
  • 2149 Views
  • 0 replies
  • 0 kudos

Can I generate a uuid4 column when I do a COPY INTO command?

I have raw call log data and the logs don't have a unique id number so I generate a uuid4 number when i load them using spark.  Now I want to save the records to a table, and run a COPY INTO command every day to ingest new records.  I am only appendi...

  • 2149 Views
  • 0 replies
  • 0 kudos
Labels