cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

Databricks SQL Warehouse, Tableau and spark.driver.maxResultSize error

Jpeterson
New Contributor III

I'm attempting to create a tableau extract on tableau server with a connection to databricks large sql warehouse. The extract process fails due to spark.driver.maxResultSize error.

Using a databricks interactive cluster in the data science & engineering workspace, I can edit the spark config to change spark.driver.maxResultSize and resolve this error.

Is there a way to change the spark.driver.maxResultSize on a databricks sql warehouse?

Is there a way to reduce the size of the data collected? The full table selected is just 987 MB, but when the sql warehouse tries to read > collect > send to tableau the collect process is resulting in over 32GB.

Any other ideas on how to solve? I've got a bunch of tableau extract processes that suffer from this spark.driver.maxResultSize error.

Driver error message:

java.lang.RuntimeException: [Simba][Hardy] (35) Error from server: error code: '0' error message: 'Error running query: org.apache.spark.SparkException: Job aborted due to stage failure: Total size of serialized results of 196 tasks (32.8 GB) is bigger than spark.driver.maxResultSize (32.0 GB)'.

7 REPLIES 7

karthik_p
Esteemed Contributor

@Josh Peterson​ did you tried to add under data access config in sql, few of config parameters are supported. Data access configuration | Databricks on AWS

Jpeterson
New Contributor III

I've tried to add that config under data access, but it doesn't seem to be a supported key.

Error message:

Line 1 "spark.driver.maxResultSize": Illegal key

dvm
New Contributor II

Hi,

SQL clusters share the same Spark configuration. Ask the workspace administrator - to add the spark config under SQL Admin/SQL Workspace.

Instructions at https://docs.databricks.com/sql/admin/sql-configuration-parameters.html

Jpeterson
New Contributor III

spark.driver.maxResultSize isn't a supported parameter on SQL Warehouses https://docs.databricks.com/sql/language-manual/sql-ref-parameters.html

Maybe this can be added to roadmap, or for BI integrations that ask for select * or full collects to operate differently on the sql warehouses?

Marty18
New Contributor II

We are having the same issue on our end with Tableau Data Extracts It would be great to be able to change these options at the individual sql warehouse level and not more broadly.

DBUser2
New Contributor III

Wondering if any solution to this problem was found. We are having the same issue on our end with using Sql Warehouse.

Gilbertson13
New Contributor II

Despite these hurdles, it's a great experience to work with these tools. For a bit of fun and relaxation, I enjoy playing Wordle Unlimited to unwind after a day of tackling data issues.

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group