โ10-24-2024 12:31 AM
Hi,
We have a workflow which calls 3 notebooks and when we run this workflow using shared cluster it runs fine, but when ran with job cluster , one of the notebooks fail.This notebook uses sql function Lpad and looks like it errors because of it. Has anyone experienced anything similar.
Thanks,
Shalini
โ10-24-2024 11:26 AM
Hi @ShaliniC ,
The possible causes are:
1. The Databricks Runtime version of the job cluster is different than the one of the the shared cluster.
2. The libraries installed on the shared cluster are missing in the job cluster.
Could you check the what are the above configuration changes and adjust the job cluster to the configuration of the shared cluster?
If it does not help, could you please share the details of the error message?
โ10-31-2024 02:20 AM
Hi,
The databricks runtime version and the installed libraries are the same on shared and job cluster. We had also raised a case with databricks support for the same. where they identified the error to be a syntax error in lpad function which we have used.
Here is the explanation for reason for difference between clusters , which we have received from databricks
Single access mode compute uses Classic Pyspark and Shared clusters use the secure SparkConnect interface for python API, So likely the behaviour is inconsistent between SparkConnect and Classic PySpark.
Thanks,Shalini
โ10-25-2024 12:06 AM
notebooks are executing sequentially or parallel in this workflow?
โ10-31-2024 02:22 AM
Hi,
the error occured in the parallel run.
We had also raised a case with databricks support for the same. where they identified the error to be a syntax fail in lpad function which we have used.
Here is the explanation for reason for difference between clusters , which we have received from databricks
Single access mode compute uses Classic Pyspark and Shared clusters use the secure SparkConnect interface for python API, So likely the behaviour is inconsistent between SparkConnect and Classic PySpark.
Thanks,Shalini
Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you wonโt want to miss the chance to attend and share knowledge.
If there isnโt a group near you, start one and help create a community that brings people together.
Request a New Group