cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results for 
Search instead for 
Did you mean: 

How to get rid of "Command result size exceeds limit"

AmineHY
Contributor

I am working on Databricks Notebook and trying to display a map using Floium and I keep getting this error

> Command result size exceeds limit: Exceeded 20971520 bytes (current = 20973510)

How can I get increase the memory limit?

I already reduced the size of my pandas dataframe.

1 ACCEPTED SOLUTION

Accepted Solutions

daniel_sahal
Esteemed Contributor

Notebook has got a default 20MB limit for outputs. There's nothing you can do about that.

https://kb.databricks.com/en_US/jobs/job-cluster-limit-nb-output

View solution in original post

4 REPLIES 4

daniel_sahal
Esteemed Contributor

Notebook has got a default 20MB limit for outputs. There's nothing you can do about that.

https://kb.databricks.com/en_US/jobs/job-cluster-limit-nb-output

AmineHY
Contributor

Thanks, Daniel,

I have read this article, and I guess this problem occurs only while rendering the output, I'll try to save the result to disk in HTML format and visualize the results through the browser.

Have a good day.

Amine.

sher
Valued Contributor II

you can try with another module like geopandas / pygal

labromb
Contributor

Hi,

I have the same problem with keplergl, and the save to disk option, whilst helpful isn't super practical... So how does one plot large datasets in kepler?

Any thought welcome

Connect with Databricks Users in Your Area

Join a Regional User Group to connect with local Databricks users. Events will be happening in your city, and you won’t want to miss the chance to attend and share knowledge.

If there isn’t a group near you, start one and help create a community that brings people together.

Request a New Group