cancel
Showing results for 
Search instead for 
Did you mean: 
Data Engineering
cancel
Showing results for 
Search instead for 
Did you mean: 

Spark Configuration Parameter for Cluster Downscaling

Hubert-Dudek
Esteemed Contributor III

spark.databricks.aggressiveWindowDownS This parameter is designed to determine the frequency, in seconds, at which the cluster decides to downscale.

By adjusting this setting, you can fine-tune how rapidly clusters release workers. A higher value will result in the cluster holding onto workers longer before releasing them. The maximum limit for this parameter is set at 600 seconds.

1 REPLY 1

Haiyangl104
New Contributor III

I wish there was a configuration to toggle upscaling behavior. I want the clusters to scale up only if the bottleneck is approaching 70% memory usage. Currently the autoscaling is only based on CPU not Memory (RAM).

Welcome to Databricks Community: Lets learn, network and celebrate together

Join our fast-growing data practitioner and expert community of 80K+ members, ready to discover, help and collaborate together while making meaningful connections. 

Click here to register and join today! 

Engage in exciting technical discussions, join a group with your peers and meet our Featured Members.