Failed job with "A fatal error has been detected by the Java Runtime Environment"
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
11-20-2024 01:24 AM
Hi community,
I have a question regarding an error that I get sometimes when running a job.
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGSEGV (0xb) at pc=0x00007fc941e74996, pid=940, tid=0x00007fc892dff640
#
# JRE version: OpenJDK Runtime Environment (Zulu 8.72.0.17-CA-linux64) (8.0_382-b05) (build 1.8.0_382-b05)
# Java VM: OpenJDK 64-Bit Server VM (25.382-b05 mixed mode linux-amd64 compressed oops)
# Problematic frame:
# J 33375 C2 org.apache.spark.unsafe.types.UTF8String.hashCode()I (18 bytes) @ 0x00007fc941e74996 [0x00007fc941e748a0+0xf6]
#
# Failed to write core dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again
#
# An error report file with more information is saved as:
# /databricks/driver/hs_err_pid940.log
#
# If you would like to submit a bug report, please visit:
# http://www.azul.com/support/
#
Why do I get this error and why does the job run when I restart it? Is there a possibility to remediate this issue?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
11-21-2024 06:10 AM
Is this happening regularly? Is this using a job cluster or all purpose cluster?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
11-21-2024 06:16 AM
It happened twice now yesterday but I first noticed this yesterday so I cannot tell if this is happening regularly. This is using a job cluster
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
11-26-2024 11:32 PM
It happend now multiple times in the last week. Do you have an idea what could cause this problem?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
11-26-2024 11:34 PM
In the last run there has been additional information in the error message:
#
# A fatal error has been detected by the Java Runtime Environment:
#
# SIGSEGV (0xb) at pc=0x00007f168e094210, pid=1002, tid=0x00007f15dd1ff640
#
# JRE version: OpenJDK Runtime Environment (Zulu 8.72.0.17-CA-linux64) (8.0_382-b05) (build 1.8.0_382-b05)
# Java VM: OpenJDK 64-Bit Server VM (25.382-b05 mixed mode linux-amd64 compressed oops)
# Problematic frame:
# J 33949 C2 org.apache.spark.unsafe.types.UTF8String.hashCode()I (18 bytes) @ 0x00007f168e094210 [0x00007f168e0940e0+0x130]
#
# Failed to write core dump. Core dumps have been disabled. To enable core dumping, try "ulimit -c unlimited" before starting Java again
#
# An error report file with more information is saved as:
# /databricks/driver/hs_err_pid1002.log
Compiled method (c2) 335210 34051 4 org.apache.spark.sql.catalyst.util.ArrayData::foreach (64 bytes)
total in heap [0x00007f168d6b4610,0x00007f168d6b52b0] = 3232
relocation [0x00007f168d6b4738,0x00007f168d6b47b8] = 128
main code [0x00007f168d6b47c0,0x00007f168d6b4ca0] = 1248
stub code [0x00007f168d6b4ca0,0x00007f168d6b4ce8] = 72
oops [0x00007f168d6b4ce8,0x00007f168d6b4d00] = 24
metadata [0x00007f168d6b4d00,0x00007f168d6b4dc0] = 192
scopes data [0x00007f168d6b4dc0,0x00007f168d6b50f8] = 824
scopes pcs [0x00007f168d6b50f8,0x00007f168d6b51d8] = 224
dependencies [0x00007f168d6b51d8,0x00007f168d6b51e8] = 16
handler table [0x00007f168d6b51e8,0x00007f168d6b5290] = 168
nul chk table [0x00007f168d6b5290,0x00007f168d6b52b0] = 32
#
# If you would like to submit a bug report, please visit:
# http://www.azul.com/support/
#

