cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

flashmav1
by New Contributor II
  • 372 Views
  • 6 replies
  • 3 kudos

Resolved! [NUMERIC_VALUE_OUT_OF_RANGE.WITHOUT_SUGGESTION] The -12874815911431.6200000000 rounded half up from

I am using dataricks version 15.4 and getting below error whicle reading from jdbc and writing to aws S3 location: [NUMERIC_VALUE_OUT_OF_RANGE.WITHOUT_SUGGESTION] The -12874815911431.6200000000 rounded half up from -12874815911431.6200000000 cannot b...

  • 372 Views
  • 6 replies
  • 3 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 3 kudos

 

  • 3 kudos
5 More Replies
Lenoq
by New Contributor II
  • 266 Views
  • 4 replies
  • 2 kudos

Are there SQL linters for Databricks GUI queries and notebook %sql cells?

 I'm looking for SQL linters in two different contexts within Databricks GUI:SQL queries in Databricks SQL Editor (GUI) - Is there a built-in linter for writing SQL queries in the Databricks SQL workspace?%sql magic cells in Databricks notebooks (GUI...

  • 266 Views
  • 4 replies
  • 2 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 2 kudos

in databricks notebook is possible to connect own linters but I don't think it will work in SQL editor.

  • 2 kudos
3 More Replies
thatsawinner
by New Contributor II
  • 397 Views
  • 4 replies
  • 1 kudos

Resolved! Notebook Session Has Crashed

I am getting a pop-up error message in the right hand corner of my Databricks session, "Your notebook session has crashed."  This is a notebook I've been working in for a while.The only line of code I am running is pip listThe error at the bottom of ...

  • 397 Views
  • 4 replies
  • 1 kudos
Latest Reply
Hubert-Dudek
Esteemed Contributor III
  • 1 kudos

From the screen, it looks like an interactive cluster—maybe try rebooting it.

  • 1 kudos
3 More Replies
Lenoq
by New Contributor II
  • 173 Views
  • 1 replies
  • 1 kudos

Resolved! Looking for CLI-based SQL formatter for Databricks: Alternative to gethue/sql-formatter for .

I'm looking for a SQL formatter (CLI) for two different contexts within Databricks.1. SQL queries in .sql files - What tool does Databricks use by default to format SQL?2. %sql magic cells in notebooks (.ipynb files) - What tool does Databricks use b...

  • 173 Views
  • 1 replies
  • 1 kudos
Latest Reply
ilir_nuredini
Honored Contributor
  • 1 kudos

Hi @Lenoq ,As far as I know, currently there is no tool that meets your given criteria by databricks/labs. The closest that I can see is the tool you gave as an example, but as you said its not that good. what I would do maybe is an intermedial optio...

  • 1 kudos
Prasanna_N
by New Contributor
  • 3373 Views
  • 2 replies
  • 2 kudos

Inference table Monitoring

i have data from march1 to march 14 in the final inference table and i have given 1 week granularity. after that profile and drift table is generated and i see the window start time as like this objectstart: "2025-02-24T00:00:00.000Z"end: "2025-03-03...

  • 3373 Views
  • 2 replies
  • 2 kudos
Latest Reply
AbhayPSingh
Databricks Employee
  • 2 kudos

More or less repeating what Mark said and adding some additional thoughts. Why the Window Starts from February 24 The reason you're seeing a window starting from February 24 (even though your data starts March 1) is because monitoring systems align t...

  • 2 kudos
1 More Replies
Ronis
by New Contributor
  • 9277 Views
  • 6 replies
  • 1 kudos

SSRS Connect to Databricks

Hi ,I need to connect databricks query from microsoft SSRS.is it possible ? How do you make the connection?

  • 9277 Views
  • 6 replies
  • 1 kudos
Latest Reply
-werners-
Esteemed Contributor III
  • 1 kudos

SSRS has limited auth methods.  It also is EOL so my answer is: No.This is not a limitation of Databricks but SSRS.You could define the connection as a linked server on sql server, that might open some extra (MS native) options.PS. it is best not to ...

  • 1 kudos
5 More Replies
ManojkMohan
by Honored Contributor II
  • 203 Views
  • 1 replies
  • 0 kudos

Exposing Databricks API in Salesforce

Use Case:I want to expose a data bricks API URL in Salesforce, Salesforce will hit that exposed end point every time a record is created and data will be transferred from Salesforce to DatabricksWhen i try creating a serving end pointI am unable to s...

ManojkMohan_0-1761252978733.png
  • 203 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

When integrating Salesforce with Databricks to push data upon record creation, using a serving endpoint is not the most common or optimal approach. Although Databricks Feature Serving endpoints can be used for model or feature APIs, they are primaril...

  • 0 kudos
JuliandaCruz
by New Contributor II
  • 475 Views
  • 4 replies
  • 0 kudos

Access to Databricks Volumes via Databricks Connect not working anymore

Hi all, I use the extension to debug my python code regularly and since yesterday accessing files in the Databricks Volume isn't working anymore. The situation in the UI of Databricks is as follows:When I execute a glob statement to list all zip-file...

  • 475 Views
  • 4 replies
  • 0 kudos
Latest Reply
mmayorga
Databricks Employee
  • 0 kudos

  hi @JuliandaCruz  Thank you for reaching out! I was able to reproduce your case while using Databricks Connect. The "Upload and Run file" option worked fine and returned results, which is essentially the same as running from the Databricks UI. Thou...

  • 0 kudos
3 More Replies
AgusBudianto
by Contributor
  • 464 Views
  • 5 replies
  • 2 kudos

Resolved! Why am I getting NameError name _all_timezones_unchecked' is not defined

I defined the following local time function get datetime: def get_sysdate():    jkt_tz = pytz.timezone('Asia/Jakarta')     sysdate = datetime.now(jkt_tz).strftime('%Y-%m-%d %H:%M:%S')    return sysdatespark.udf.register("get_sysdate", get_sysdate)But...

  • 464 Views
  • 5 replies
  • 2 kudos
Latest Reply
AgusBudianto
Contributor
  • 2 kudos

Hi @Khaja_Zaffer I have connected with MS Support and explained: pytz is no thread safe package, I believe it will have some issue when executor init it parallelly. Second, this is a 3rd party lib, and suggest using the built-in library from ZoneInfo...

  • 2 kudos
4 More Replies
turagittech
by Contributor
  • 2447 Views
  • 2 replies
  • 0 kudos

Batch reading from sql server tables with cdc on ssql server tables

Hi all,I need to do a batch load from sql server into Databricks. I have CC enabled on some tables. The simple appears to be union CDC and regular table to get a single set of records to load, but this appears to be fraught with risk of out of sequen...

  • 2447 Views
  • 2 replies
  • 0 kudos
Latest Reply
Krishna_S
Databricks Employee
  • 0 kudos

Yes, you can use TVFs on Databricks. Please check the following link: https://docs.databricks.com/aws/en/sql/language-manual/sql-ref-syntax-qry-select-tvf#gsc.tab=0Can you please elaborate on how you are loading the SQL Server Data into Databricks? H...

  • 0 kudos
1 More Replies
AshMod
by New Contributor II
  • 209 Views
  • 2 replies
  • 1 kudos

Job runs on serverless eventhough Job config has cluster definitions

Hi,I am defining the job along with job cluster specification using python sdk. But when the job runs it is using the serverless compute, instead of the defined cluster. I can say the job uses serverless from the job_run log and also from the system....

AshMod_0-1761230207138.png AshMod_1-1761230258750.png AshMod_2-1761231900872.png
  • 209 Views
  • 2 replies
  • 1 kudos
Latest Reply
AshMod
New Contributor II
  • 1 kudos

Thanks for checking @ManojkMohan. I found the issue in the job task definition. There is a job_clusters list in the job definition, where I provide the cluster config details. But this alone is not sufficient to have the task use the cluster. The job...

  • 1 kudos
1 More Replies
saab123
by New Contributor II
  • 3544 Views
  • 1 replies
  • 0 kudos

Not able to connect to Neo4j Aura Db from databricks

I am trying to connect to a Neo4j AuraDb instance-f9374927. Created a free professional instance of Neo4j. I am able to connect to this instance, add nodes and relationships.   Created a Databricks shared cluster 14.3 LTS (includes Apache Spark 3.5.0...

  • 3544 Views
  • 1 replies
  • 0 kudos
Latest Reply
mark_ott
Databricks Employee
  • 0 kudos

The connection issue between your Databricks cluster and Neo4j AuraDB instance (f9374927) with the ServiceUnavailableException: No routing server available message is tied to network-level SSL configuration and connectivity rather than incorrect code...

  • 0 kudos
dbxlearner
by New Contributor II
  • 3167 Views
  • 3 replies
  • 1 kudos

Resolved! Deploying using Databricks asset bundles (DABs) in a closed network

Hello, I'm trying to deploy DBX workflows using DABs using an Azure DevOps pipeline, in a network that cannot download the required terraform databricks provider package online, due to firewall/network restrictions.I have followed this post: https://...

  • 3167 Views
  • 3 replies
  • 1 kudos
Latest Reply
dbxlearner
New Contributor II
  • 1 kudos

Another thing I noticed is, when running the 'databricks bundle debug terraform' command, it mentions these variables:I have tried setting these variables as environment variables in my ADO pipeline, specially the databricks terraform provider variab...

  • 1 kudos
2 More Replies
ticuss
by New Contributor
  • 204 Views
  • 1 replies
  • 1 kudos

Resolved! Lakebase / Feature Store error: “Failed to get identity details for username” (service principal)

Hello,I’m running into a Lakebase / Feature Store issue related to service principal authentication when trying to log or read from the Databricks Feature Store. Migrating from the legacy online tables.  Here’s the exact error:psycopg2.OperationalErr...

  • 204 Views
  • 1 replies
  • 1 kudos
Latest Reply
mark_ott
Databricks Employee
  • 1 kudos

The error you’re encountering —psycopg2.OperationalError: FATAL: Failed to get identity details for username: "user_uuid" — typically arises from an OAuth identity mismatch or invalid token scope when a Databricks service principal is used to authent...

  • 1 kudos
janglais
by New Contributor
  • 221 Views
  • 2 replies
  • 0 kudos

Resolved! DLT Pipeline with unknown deleted source data

Hello.. I need help. So the context is : - ERP data for company in my group is stored in sql tables - Currently, once per day we copy the last 2 months of data (creation date) from each table into our datalake landing zone (we can however do full cop...

  • 221 Views
  • 2 replies
  • 0 kudos
Latest Reply
madams
Contributor III
  • 0 kudos

Your solution #1 is very frustrating to me as well, for a number of reasons.  Simply put, we have to be able to compare incoming data to target data for normal ETL operations. One way around this is to create a view of your target silver table, outsi...

  • 0 kudos
1 More Replies

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local community—sign up today to get started!

Sign Up Now
Labels