Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
Whenever I run my script it resets the notebook state: "The spark driver has stopped unexpectedly and is restarting. Your notebook will be automatically reattached.
at com.databricks.spark.chauffeur.Chauffeur.onDriverStateChange(Chauffeur.scala:1467)" This means I cannot even figure out efficiently what is causing the trouble. I am trying to do some data manipulation in R on the driver node. The amount of data is relatively small (a few GB) but I constantly run into this issue.
It seems to have to do with some form of memory limit of the R session, but I cannot figure out: 1. Which parameter needs to be adjusted here 2. Why any sane person would setup an R session with only 2 GB of memory in the first place