cancel
Showing results forย 
Search instead forย 
Did you mean:ย 
Data Engineering
Join discussions on data engineering best practices, architectures, and optimization strategies within the Databricks Community. Exchange insights and solutions with fellow data engineers.
cancel
Showing results forย 
Search instead forย 
Did you mean:ย 

Databricks Asset Bundle conditional job cluster size?

Bart_DE
New Contributor II

Hey folks,

Can someone please suggest if there is a way to spawn a job cluster of a given size if a parameter of the job invocation (e.g file_name) contains a desired value? I have a job which 90% of the time deals with very small files, but the remaining 10% requires much bigger machines. How can i define job.yml file in such a way? Is this even possible to be done without some pre-process that checks the file name?

 

Thanks!

1 ACCEPTED SOLUTION

Accepted Solutions

lingareddy_Alva
Honored Contributor II

Hi @Bart_DE 

No โ€” a single job.yml file canโ€™t โ€œlook insideโ€ a parameter like file_name and then decide to spin up a different job-cluster size on the fly.
Job-cluster definitions in Databricks Workflows (Jobs) are static. All the heavy-lifting has to be done outside the cluster specification:

 

LR

View solution in original post

1 REPLY 1

lingareddy_Alva
Honored Contributor II

Hi @Bart_DE 

No โ€” a single job.yml file canโ€™t โ€œlook insideโ€ a parameter like file_name and then decide to spin up a different job-cluster size on the fly.
Job-cluster definitions in Databricks Workflows (Jobs) are static. All the heavy-lifting has to be done outside the cluster specification:

 

LR

Join Us as a Local Community Builder!

Passionate about hosting events and connecting people? Help us grow a vibrant local communityโ€”sign up today to get started!

Sign Up Now