<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Programmatically set minimum workers for a job cluster based on file size? in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/programmatically-set-minimum-workers-for-a-job-cluster-based-on/m-p/127517#M47996</link>
    <description>&lt;P&gt;Hi Alena,&amp;nbsp;&lt;/P&gt;&lt;P&gt;Jobs API has update functionality to be able to do that:&amp;nbsp;&lt;A href="https://docs.databricks.com/api/workspace/jobs_21/update" target="_blank"&gt;https://docs.databricks.com/api/workspace/jobs_21/update&lt;/A&gt;&lt;/P&gt;&lt;P&gt;If for some reason you can’t update your pipeline before you trigger it you can also consider creating a new job with desired configuration every time you run a trigger (&lt;SPAN&gt;POST&amp;nbsp;&lt;/SPAN&gt;&lt;SPAN&gt;/api/2.2/jobs/create).&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Kerem Durak&lt;/SPAN&gt;&lt;/P&gt;</description>
    <pubDate>Tue, 05 Aug 2025 23:27:56 GMT</pubDate>
    <dc:creator>kerem</dc:creator>
    <dc:date>2025-08-05T23:27:56Z</dc:date>
    <item>
      <title>Programmatically set minimum workers for a job cluster based on file size?</title>
      <link>https://community.databricks.com/t5/data-engineering/programmatically-set-minimum-workers-for-a-job-cluster-based-on/m-p/127498#M47987</link>
      <description>&lt;P&gt;I’m running an ingestion pipeline with a Databricks job:&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;&lt;P&gt;A file lands in S3&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;A Lambda is triggered&lt;/P&gt;&lt;/LI&gt;&lt;LI&gt;&lt;P&gt;The Lambda runs a Databricks job&lt;/P&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;The incoming files vary a lot in size, which makes processing times vary as well. My job cluster has autoscaling enabled, but scaling up takes time.&lt;/P&gt;&lt;P&gt;Ideally, if a 10 GB file comes in, I’d like the job to start with more than one worker immediately, instead of waiting for autoscaling to kick in.&lt;/P&gt;&lt;P&gt;I’m currently using the run-now API to trigger the job, but I don’t see a way to adjust the job cluster configuration at runtime.&lt;/P&gt;&lt;P&gt;Is there a way to programmatically set the minimum number of workers for a job cluster depending on the incoming file size?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Tue, 05 Aug 2025 20:41:55 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/programmatically-set-minimum-workers-for-a-job-cluster-based-on/m-p/127498#M47987</guid>
      <dc:creator>Alena</dc:creator>
      <dc:date>2025-08-05T20:41:55Z</dc:date>
    </item>
    <item>
      <title>Re: Programmatically set minimum workers for a job cluster based on file size?</title>
      <link>https://community.databricks.com/t5/data-engineering/programmatically-set-minimum-workers-for-a-job-cluster-based-on/m-p/127517#M47996</link>
      <description>&lt;P&gt;Hi Alena,&amp;nbsp;&lt;/P&gt;&lt;P&gt;Jobs API has update functionality to be able to do that:&amp;nbsp;&lt;A href="https://docs.databricks.com/api/workspace/jobs_21/update" target="_blank"&gt;https://docs.databricks.com/api/workspace/jobs_21/update&lt;/A&gt;&lt;/P&gt;&lt;P&gt;If for some reason you can’t update your pipeline before you trigger it you can also consider creating a new job with desired configuration every time you run a trigger (&lt;SPAN&gt;POST&amp;nbsp;&lt;/SPAN&gt;&lt;SPAN&gt;/api/2.2/jobs/create).&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Kerem Durak&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 05 Aug 2025 23:27:56 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/programmatically-set-minimum-workers-for-a-job-cluster-based-on/m-p/127517#M47996</guid>
      <dc:creator>kerem</dc:creator>
      <dc:date>2025-08-05T23:27:56Z</dc:date>
    </item>
  </channel>
</rss>

