cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

jeremy98
by Honored Contributor
  • 808 Views
  • 1 replies
  • 0 kudos

Resolved! Is there a INTERVAL data type?

Hi community,I was using a column in postgresSQL that is a DATETIME.TIMEDELTA, is it possible to have the same data type also in Databricks?

  • 808 Views
  • 1 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hi @jeremy98, You can use the TIMESTAMP and TIMESTAMP_NTZ data types to handle date and time values, similar to the DATETIME type in PostgreSQL. However, Databricks does not have a direct equivalent to PostgreSQL's TIMEDELTA type https://docs.databri...

  • 0 kudos
pavan_yndpl
by New Contributor
  • 1838 Views
  • 1 replies
  • 0 kudos

How to resolve SSL_connect error when VPN is enabled

I am trying to connect to Databricks using ODBC protocol with Simba Driver DSN. I am able to successfully connect and access the data when our corporate VPN is turned OFF. but when it's turned ON , I am getting the following error "[Simba][ThriftExte...

  • 1838 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

The error you are encountering, " (14) Unexpected response from server during a HTTP connection: SSL_connect:", when trying to connect to Databricks using the ODBC protocol with the Simba Driver DSN while the corporate VPN is turned on, is likely rel...

  • 0 kudos
raghu2
by New Contributor III
  • 1501 Views
  • 2 replies
  • 1 kudos

Liquid Cluster enabled table - concurrent writes

I am trying to insert rows into a Liquid cluster enabled delta table using multiple threads. This link, states that liquid clustering is used for : Tables with concurrent write requirements.I get this error: [DELTA_CONCURRENT_APPEND] ConcurrentAppend...

  • 1501 Views
  • 2 replies
  • 1 kudos
Latest Reply
TejeshS
Contributor
  • 1 kudos

We encountered a similar issue as well, and the workaround we tried was partitioning those columns, as Liquid clustering can sometimes trigger this error.

  • 1 kudos
1 More Replies
shanisolomon
by New Contributor II
  • 1117 Views
  • 2 replies
  • 0 kudos

Databricks inconsistent count and select

Hi, I have a table with 2 versions:1. Add txn: path = "a.parquet" numRecords = 10 deletionVector = null2. Add txn: path = "a.parquet" numRecords = 10 deletionVector = (..., cardinality = 2)Please note both transactions point to the same physical path...

  • 1117 Views
  • 2 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

Hello the behavior observed indeed seems to be inconsistent with the expected behavior in delta, do you have a support contract to open a support ticket so this can be further analyzed?

  • 0 kudos
1 More Replies
Tej_04
by New Contributor
  • 2565 Views
  • 1 replies
  • 0 kudos

Avoid scientific values

I am trying to insert data into catalog tables on data bricks but the values are being displayed in scientific notation which i am trying to avoid how do I view the data in standard formatfor example - 0.0000000 is being displayed as 0E-7

Data Engineering
catalogtables scientificnotation
  • 2565 Views
  • 1 replies
  • 0 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 0 kudos

Hi @Tej_04, Can you try with format_number Function. SELECT format_number(column_name, decimal_places) AS column_name FROM table_name; https://docs.databricks.com/en/sql/language-manual/functions/format_number.html

  • 0 kudos
noimeta
by Contributor III
  • 17688 Views
  • 15 replies
  • 12 kudos

Resolved! Error when create an external location using code

I'm trying to create an external location from notebook, and I got this kind of error[PARSE_SYNTAX_ERROR] Syntax error at or near 'LOCATION'(line 1, pos 16)   == SQL == CREATE EXTERNAL LOCATION IF NOT EXISTS test_location URL 's3://test-bronze/db/tes...

  • 17688 Views
  • 15 replies
  • 12 kudos
Latest Reply
Lokeshv
New Contributor II
  • 12 kudos

Hey everyone,I'm facing an issue with retrieving data from a volume or table that contains a string with a symbol, for example, 'databricks+'. Whenever I try to retrieve this data, I encounter a syntax error. Can anyone help me resolve this issue?

  • 12 kudos
14 More Replies
aliacovella
by Contributor
  • 2915 Views
  • 8 replies
  • 5 kudos

Resolved! DLT Pipleline with only views

I'm trying to create a pipeline containing a view from a federated source. In this case, I'd like to just create materialized views from the federation and and schedule the pipeline for execution. If I define a pipeline  with only something like the ...

  • 2915 Views
  • 8 replies
  • 5 kudos
Latest Reply
Alberto_Umana
Databricks Employee
  • 5 kudos

No problem, if you have any other questions let me know!

  • 5 kudos
7 More Replies
TejeshS
by Contributor
  • 3024 Views
  • 7 replies
  • 0 kudos

How to enable row tracking on Delta Live Tables?

We are encountering a scenario where we need to enable support for Incremental Processing on Materialized views having DLT base tables. However, we have observed that the compute is being executed with the COMPLETE_RECOMPUTE mode instead of INCREMENT...

  • 3024 Views
  • 7 replies
  • 0 kudos
Latest Reply
TejeshS
Contributor
  • 0 kudos

Moreover, we have CDF enabled DLT tables, but as per documentation we see a limitation if CDF is enabled then row Tracking won't be possible. Use row tracking for Delta tables | Databricks on AWSBut as per our use case we need incremental processing ...

  • 0 kudos
6 More Replies
David_Billa
by New Contributor III
  • 1791 Views
  • 1 replies
  • 1 kudos

Create table from json and flatten in the same SQL

Any help in writing SQL to create the table using JSON file and flatten in the same step? As I'm new to JSON it would be nice if someone can give the heads up by referencing to any document or help provide the recommended the solution.Sample JSON fil...

  • 1791 Views
  • 1 replies
  • 1 kudos
Latest Reply
Walter_C
Databricks Employee
  • 1 kudos

To create a table directly from a JSON file and flatten it using SQL in Databricks, you can use the CREATE TABLE statement with the USING JSON clause. However, SQL alone does not provide a direct way to flatten nested JSON structures. You would typic...

  • 1 kudos
Erik
by Valued Contributor III
  • 1329 Views
  • 3 replies
  • 1 kudos

What does durationMs.commitBatch measure?

With a structured streamin job from Kafka, we have a metric in durationMs called commitBatch. There is also an example of this in this databricks documentation. I can not find any description of what this measures, and how it relates to the other met...

  • 1329 Views
  • 3 replies
  • 1 kudos
Latest Reply
Walter_C
Databricks Employee
  • 1 kudos

The commitBatch metric is a part of the overall triggerExecution time, which encompasses all stages of planning and executing the microbatch, including committing the batch data and updating offsets. The commitBatch metric may not always be present i...

  • 1 kudos
2 More Replies
Tahseen0354
by Valued Contributor
  • 8339 Views
  • 5 replies
  • 3 kudos

Resolved! Why I am not receiving any mail sent to the Azure AD Group mailbox when databricks job fails ?

I have created an Azure AD Group in "Microsoft 365" type with its own email address, which being added to the Notification of a Databricks Job (on failure). But there is no mail sent to the Azure Group mailbox when the job fails.I am able to send a d...

  • 8339 Views
  • 5 replies
  • 3 kudos
Latest Reply
Lanky
New Contributor II
  • 3 kudos

Hello Guys, I have setup ses receive email for databricks notification. When i send email message from google mail or yahoo mail, it gets to the SES email receiving rule. However, notification from databricks doesn't get to the same SES email receivi...

  • 3 kudos
4 More Replies
meghana_tulla
by New Contributor III
  • 1179 Views
  • 1 replies
  • 1 kudos

Automating Admin Consent for Azure Databricks SCIM App Creation Using Terraform.

I am trying to automate the creation of an Azure AD application (specifically, an Azure Databricks SCIM app) and grant admin consent for its API permissions using Terraform. The required API permissions include Application.ReadWrite.All, Application....

  • 1179 Views
  • 1 replies
  • 1 kudos
Latest Reply
Walter_C
Databricks Employee
  • 1 kudos

The Service Principal you're using for authentication may not have sufficient permissions to grant admin consent. Ensure that the Service Principal has the necessary roles assigned, such as "Global Administrator" or "Privileged Role Administrator" . ...

  • 1 kudos
KishanDaxini
by New Contributor
  • 1010 Views
  • 1 replies
  • 0 kudos

Handling non notebook file types in repos

Hi, I have ,py, .txt, .yml, .json files in my repo, when I merging the feature branch to master branch, the file type of these files are getting changed to notebook, which is causing error while importing these files into my different notebooks.P.S. ...

  • 1010 Views
  • 1 replies
  • 0 kudos
Latest Reply
Walter_C
Databricks Employee
  • 0 kudos

IPYNB notebooks are the default format when creating a new notebook on Databricks. To change the default to the Databricks source format, log into your Databricks workspace, click your profile in the upper-right of the page, then click Settings and n...

  • 0 kudos
Abishrp
by Contributor
  • 794 Views
  • 1 replies
  • 1 kudos

Issue in finding OS in which my cluster runs

During the configuration of job compute cluster, I didn't mention any OS details. How can i find which OS my cluster is running?Also, Is there any way to get pricing details of all instances with different category  Job Compute - premium  , All Purpo...

  • 794 Views
  • 1 replies
  • 1 kudos
Latest Reply
Walter_C
Databricks Employee
  • 1 kudos

You can find the OS for your runtime in the system environment information in release notes, for example for Runtime 16.1 it can be found in: https://docs.databricks.com/en/release-notes/runtime/16.1.html#system-environment which is Ubuntu 24.04.1 LT...

  • 1 kudos
anshi_t_k
by New Contributor III
  • 1993 Views
  • 4 replies
  • 0 kudos

Practice question for data engineer exam

A data engineer, User A, has promoted a pipeline to production by using the REST API to programmatically create several jobs. A DevOps engineer, User B, has configured an external orchestration tool to trigger job runs through the REST API. Both user...

  • 1993 Views
  • 4 replies
  • 0 kudos
Latest Reply
rakeshdey
New Contributor II
  • 0 kudos

The answer should be answer B, when you try to get job run information, creator_user_email id always populated as 'Run As' in workflow , so which credential used to trigger job.. if you get workflow infor through rest api then Ans A correct

  • 0 kudos
3 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels