How does Job Cluster knows how many resources to assign to an Application ?
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
09-27-2023 12:47 AM
Hi All Enthusiasts !
As per my understanding when a user submits an application in spark cluster it specifies how much memory, executors etc. it would need .
But in Data bricks notebooks we never specify that anywhere. If we have submitted the notebook in a Job cluster how does Data Bricks Resource Manager decides how much it will allocate resources to this one
In a cluster having pool I understand we have idle resources which can be allocated as a cluster but still don't understand how much on notebook will be assigned resources
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
09-28-2023 06:55 PM
When you create a Job, you specify a Cluster Configuration with the amount of memory, CPU, nodes, etc, for the cluster: https://docs.databricks.com/en/workflows/jobs/create-run-jobs.html
The notebook will run on the cluster following those configurations.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
09-29-2023 02:04 AM
@DBEnthusiast great question! Today, with Job Clusters, you have to specify this. As @btafur note, you do this by setting CPU, memory etc. We are in early preview of Serverless Job Clusters where you no longer specify this configuration, instead Databricks figures this out using the workload's requirements.

