cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

ISSUE: PySpark task exception handling on "Shared Compute" cluster

geronimo_signol
New Contributor

I am experiencing an issue with a PySpark job that behaves differently depending on the compute environment in Databricks. And this is blocking us from deploying the job into the PROD environment for our planned release.

Specifically:

- When running the job on a personal cluster, everything works as expected. All exceptions within the try/catch blocks are successfully caught and handled.
- However, when I run the same job on a shared cluster, it fails, and no exceptions are being caught by the try/catch blocks.


Any guidance or insights you could provide would be greatly appreciated.

Example (running piece of code in a workspace notebook): https://github.com/user-attachments/assets/78b38c5a-98f6-4bb0-82c3-45946d6c5500

 

Any ideas? @andrews 

 

1 REPLY 1

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group