Fatal error: The Python kernel is unresponsive.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-07-2023 09:59 PM
I am using MultiThread in this job which creates 8 parallel jobs. It fails for few times in a day and sometimes stuck in any of the Python notebook cell process. Here
The Python process exited with an unknown exit code.
The last 10 KB of the process's stderr and stdout can be found below. See driver logs for full logs.
---------------------------------------------------------------------------
Last messages on stderr:
Tue Feb 7 17:10:18 2023 Connection to spark from PID 24461
Tue Feb 7 17:10:18 2023 Initialized gateway on port 34499
Tue Feb 7 17:10:19 2023 Connected to spark.
---------------------------------------------------------------------------
Last messages on stdout:
NOTE: When using the `ipython kernel` entry point, Ctrl-C will not work.
To exit, you will have to explicitly quit this process, by either sending
"quit" from a client, or using Ctrl-\ in UNIX-like environments.
To read more about this, see https://github.com/ipython/ipython/issues/2049
- Labels:
-
DAIS2023
-
Fatal Error
-
KB
-
Python
-
Python Kernel
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-08-2023 05:31 AM
Hi @Mahesh Chahareโ , What is the DBR version you are using?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-08-2023 05:33 AM
Hi @Lakshay Goelโ , 11.3 LTS (includes Apache Spark 3.3.0, Scala 2.12)
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-08-2023 05:39 AM
please try using this configuration:
spark.databricks.python.defaultPythonRepl pythonshell
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-08-2023 05:54 AM
@Lakshay Goelโ , I have updated it on the cluster. I'll update you whether it works for me or not. Thanks !
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-08-2023 06:02 AM
sure
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-09-2023 11:31 AM
โ@Lakshay Goel
Am also facing same issues.โ
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-10-2023 02:52 AM
Hi @Manoj Kumarโ , If you are facing this issue on DBR 11.x, you can enable the same configuration as above i.e. spark.databricks.python.defaultPythonRepl pythonshell
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-16-2023 08:58 PM
Yeah it's working thanks @Lakshay Goelโ โ
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-17-2023 03:10 AM
Happy to help!
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-12-2023 10:22 PM
Hello We are also facing the same issue, did check by enabling the spark.databricks.python.defaultPythonRepl pythonshell , the issue has not been solved yet
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-13-2023 04:58 AM
Could you please try running the job on DBR 10.4?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ02-24-2023 05:44 AM
If you do face cluster issues. Make sure to first, restart the cluster if this one was running for a long time to get the patches, then if you can use the latest runtime. If the issue is still persistent open a ticket.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ04-08-2023 12:33 AM
Hi @Mahesh Chahareโ
Hope everything is going great.
Just wanted to check in if you were able to resolve your issue. If yes, would you be happy to mark an answer as best so that other members can find the solution more quickly? If not, please tell us so we can help you.
Cheers!
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
โ04-24-2023 08:52 AM
Hello,
I have the same problem. After upgrading from 10.5 to 12.2 we started to face this issue. Setting
spark.databricks.python.defaultPythonRepl pythonshell didn't help