I'm attempting to create a tableau extract on tableau server with a connection to databricks large sql warehouse. The extract process fails due to spark.driver.maxResultSize error.
Using a databricks interactive cluster in the data science & engineering workspace, I can edit the spark config to change spark.driver.maxResultSize and resolve this error.
Is there a way to change the spark.driver.maxResultSize on a databricks sql warehouse?
Is there a way to reduce the size of the data collected? The full table selected is just 987 MB, but when the sql warehouse tries to read > collect > send to tableau the collect process is resulting in over 32GB.
Any other ideas on how to solve? I've got a bunch of tableau extract processes that suffer from this spark.driver.maxResultSize error.
Driver error message:
java.lang.RuntimeException: [Simba][Hardy] (35) Error from server: error code: '0' error message: 'Error running query: org.apache.spark.SparkException: Job aborted due to stage failure: Total size of serialized results of 196 tasks (32.8 GB) is bigger than spark.driver.maxResultSize (32.0 GB)'.