cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Forum Posts

AbdulBasit
by New Contributor II
  • 378 Views
  • 1 replies
  • 0 kudos

Autloader error for assuming a role

Hi @Retired_mod I have seen numerous post by you. Thanks for continuously providing support. Can you or your colleagues help on this. We have a basic user which assumes a role with S3 policy to a specific bucket. When we try to read the bucket from D...

  • 378 Views
  • 1 replies
  • 0 kudos
Latest Reply
AbdulBasit
New Contributor II
  • 0 kudos

Py4JJavaError: An error occurred while calling o503.json. : java.nio.file.AccessDeniedException: s3a://xxxxxx.json: shaded.databricks.org.apache.hadoop.fs.s3a.auth.NoAuthWithAWSException: No AWS Credentials provided by AwsCredentialContextTokenProvid...

  • 0 kudos
manoj_
by New Contributor II
  • 608 Views
  • 1 replies
  • 0 kudos

Databricks view error

Data source errorDataSource.Error: ODBC: ERROR [42000] [Microsoft][Hardy] (80) Syntax or semantic analysis error thrown in server while executing query.Error message from server: org.apache.hive.service.cli.HiveSQLException: Error running query: [DEL...

  • 608 Views
  • 1 replies
  • 0 kudos
Latest Reply
manoj_
New Contributor II
  • 0 kudos

This view used to run till last week and suddenly started giving this error. So need to check what can be the reason for this issue

  • 0 kudos
ksilva
by New Contributor
  • 3792 Views
  • 4 replies
  • 1 kudos

Incorrect secret value when loaded as environment variable

I recently faced an issue that took good hours to identify. I'm loading an environment variable with a secretENVVAR: {{secrets/scope/key}}The secret is loaded in my application, I could verify it's there, but its value is not correct. I realised tha...

  • 3792 Views
  • 4 replies
  • 1 kudos
Latest Reply
danmlopsmaz
New Contributor II
  • 1 kudos

Hi team, is there an update or fix for this?

  • 1 kudos
3 More Replies
marcuskw
by Contributor II
  • 1948 Views
  • 5 replies
  • 5 kudos

Resolved! IDENTIFIER not working in UPDATE

The following code works perfectly fine: df = spark.createDataFrame([('A', 1), ('B', 2)]) df.createOrReplaceTempView('temp') spark.sql(""" SELECT IDENTIFIER(:col) FROM temp """, args={ "col": "_1" } ).display(...

  • 1948 Views
  • 5 replies
  • 5 kudos
Latest Reply
marcuskw
Contributor II
  • 5 kudos

If it helps anyone else I found this article that described a few limitations:https://community.databricks.com/t5/technical-blog/how-not-to-build-an-execute-immediate-demo/ba-p/82167 

  • 5 kudos
4 More Replies
leireroman
by New Contributor III
  • 1463 Views
  • 3 replies
  • 0 kudos

Resolved! RESOURCE_EXHAUSTED dbutils.jobs.taskValues.get

I've a job in Databricks running multiple tasks in parallel. Those tasks read parameters of the job using the utility of dbutils. I'm getting the following error when trying to read parameters in my different tasks:com.databricks.common.client.Databr...

image.png
  • 1463 Views
  • 3 replies
  • 0 kudos
Latest Reply
leireroman
New Contributor III
  • 0 kudos

Hi all,Our solution has been to use job parameters and dynamic value references. These are read using dbutils.widgets.get() instead of dbutils.jobs.taskValues.get(). Now, our ETL is working well again.Pass context about job runs into job tasks - Azur...

  • 0 kudos
2 More Replies
4kb_nick
by New Contributor III
  • 1410 Views
  • 3 replies
  • 0 kudos

Unity Catalog Lineage Not Working on GCP

Hello,We have set up a lakehouse in Databricks for one of our clients. One of the features our client would like to use is the Unity Catalog data lineage view. This is a handy feature that we have used with other clients (in both AWS and Azure) witho...

  • 1410 Views
  • 3 replies
  • 0 kudos
Latest Reply
4kb_nick
New Contributor III
  • 0 kudos

Hello,It's been a few months since this exchange. The feature limitation is not documented anywhere - documents imply that this should be working in GCP:https://docs.gcp.databricks.com/en/data-governance/unity-catalog/data-lineage.htmlIs this feature...

  • 0 kudos
2 More Replies
Valentin14
by New Contributor II
  • 7505 Views
  • 5 replies
  • 4 kudos

Import module never ends on random branches

Hello,Since a week ago, our notebook are stuck in running on the firsts cells which import python module from our github repository which is cloned in databricks.The cells stays in running state and when we try to manually cancel the jobs in databric...

  • 7505 Views
  • 5 replies
  • 4 kudos
Latest Reply
timo199
New Contributor II
  • 4 kudos

@Retired_mod 

  • 4 kudos
4 More Replies
SebastianCar28
by New Contributor
  • 271 Views
  • 0 replies
  • 0 kudos

How to implement Lifecycle of Data When Use ADLS

Hello everyone, nice to greet you. I have a question about the data lifecycle in ADLS. I know ADLS has its own rules, but they aren't working properly because I have two ADLS accounts: one for hot data and another for cool storage where the informati...

  • 271 Views
  • 0 replies
  • 0 kudos
weldermartins
by Honored Contributor
  • 7486 Views
  • 6 replies
  • 10 kudos

Resolved! Spark - API Jira

Hello guys. I use pyspark in my daily life. A demand has arisen to collect information in Jira. I was able to do this via Talend ESB, but I wouldn't want to use different tools to get the job done. Do you have any example of how to extract data from ...

  • 7486 Views
  • 6 replies
  • 10 kudos
Latest Reply
Marty73
New Contributor II
  • 10 kudos

Hi,There is also a new Databricks for Jira add-on on the Atlassian Marketplace. It is easy to setup and exports are directly created within Jira. They can be one-time, scheduled, or real-time. It can also export additional Jira data such as Assets, C...

  • 10 kudos
5 More Replies
lurban
by New Contributor
  • 4607 Views
  • 1 replies
  • 0 kudos

[INTERNAL_ERROR] The Spark SQL phase analysis failed with an internal error

Hello,I am currently working through an issue I am seeing when querying a Lakehouse Federation UC table in a workflow. I am using pyspark to query a table through Lakehouse Federation which returns a result based on the query. When running it in a ce...

  • 4607 Views
  • 1 replies
  • 0 kudos
Latest Reply
LindasonUk
New Contributor II
  • 0 kudos

I hit a similar error when trying to view FC data using a cluster with a lower Databricks Runtime.Needs to be DBR 13.1 or higher.

  • 0 kudos
pgrandjean
by New Contributor III
  • 13116 Views
  • 6 replies
  • 2 kudos

How to transfer ownership of a database and/or table?

We created a new Service Principal (SP) on Azure and would like to transfer the ownership of the databases and tables created with the old SP. The issue is that these databases and tables are not visible to the users using the new SP.I am using a Hiv...

  • 13116 Views
  • 6 replies
  • 2 kudos
Latest Reply
VivekChandran
New Contributor II
  • 2 kudos

Regarding the [PARSE_SYNTAX_ERROR] Syntax error at or near 'OWNER'.Remember to wrap the new owner name in the SQL statement with the Grave Accent (`) as the below sample. ALTER SCHEMA schema_name OWNER TO `new_oner_name`;  

  • 2 kudos
5 More Replies
jacovangelder
by Honored Contributor
  • 1176 Views
  • 1 replies
  • 0 kudos

Resolved! Hanging/frozen cancelling/running cells in notebook

I'm constantly getting frozen "Cancelling" or "Running" cells since a few days when prototyping some stuff in notebooks. Examples:And the same for instantiating functions (should be instant), neither finishing or able to cancelThis button doesn't wor...

jacovangelder_1-1723638718925.png jacovangelder_2-1723638736905.png jacovangelder_3-1723638759584.png
  • 1176 Views
  • 1 replies
  • 0 kudos
Latest Reply
jacovangelder
Honored Contributor
  • 0 kudos

Seems to have resolved itself today. Can't reproduce it anymore

  • 0 kudos
guangyi
by Contributor III
  • 1025 Views
  • 1 replies
  • 0 kudos

How to create a single CSV file with specified file name Spark in Databricks?

I know how to use Spark in Databricks to create a CSV, but it always has lots of side effects.For example, here is my code:file_path = “dbfs:/mnt/target_folder/file.csv”df.write.mode("overwrite").csv(file_path, header=True)Then what I got isA folder ...

  • 1025 Views
  • 1 replies
  • 0 kudos
Latest Reply
szymon_dybczak
Esteemed Contributor III
  • 0 kudos

Hi @guangyi ,To disable _commited_xxx, _started_xxx and _SUCCSSS you must set below spark options:  spark.conf.set("spark.databricks.io.directoryCommit.createSuccessFile","false") spark.conf.set("mapreduce.fileoutputcommitter.marksuccessfuljobs", "f...

  • 0 kudos
TylerTamasaucka
by New Contributor II
  • 28596 Views
  • 5 replies
  • 2 kudos

org.apache.spark.sql.AnalysisException: Undefined function: 'MAX'

I am trying to create a JAR for a Azure Databricks job but some code that works when using the notebook interface does not work when calling the library through a job. The weird part is that the job will complete the first run successfully but on an...

  • 28596 Views
  • 5 replies
  • 2 kudos
Latest Reply
skaja
New Contributor II
  • 2 kudos

I am facing similar issue when trying to use from_utc_timestamp function. I am able to call the function from databricks notebook but when I use the same function inside my java jar and running as a job in databricks, it is giving below error. Analys...

  • 2 kudos
4 More Replies

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group
Labels