Databricks resets notebook all the time
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
08-09-2024 12:52 AM
Whenever I run my script it resets the notebook state:
"The spark driver has stopped unexpectedly and is restarting. Your notebook will be automatically reattached.
This means I cannot even figure out efficiently what is causing the trouble. I am trying to do some data manipulation in R on the driver node. The amount of data is relatively small (a few GB) but I constantly run into this issue.
- Labels:
-
Spark
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
08-09-2024 12:56 AM
It seems to have to do with some form of memory limit of the R session, but I cannot figure out:
1. Which parameter needs to be adjusted here
2. Why any sane person would setup an R session with only 2 GB of memory in the first place
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
08-09-2024 01:31 AM
To add insult to injury it says this:
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
08-09-2024 02:28 AM
Alternatively if I try to write it as RDS I get this error:
Error in saveRDS(get(object), file = file_path): error writing to connection
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
08-09-2024 03:15 AM
To get closer to the error:
There is same mystical size limit.

