<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Can we parameterize the compute in job cluster in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/can-we-parameterize-the-compute-in-job-cluster/m-p/76351#M35204</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/55535"&gt;@NarenderKumar&lt;/a&gt;&amp;nbsp;, &lt;BR /&gt;&lt;BR /&gt;Have you considered leveraging&amp;nbsp;&lt;A href="https://docs.databricks.com/en/compute/configure.html#enable-autoscaling" target="_self"&gt;autoscaling&lt;/A&gt;&amp;nbsp;for the existing cluster?&lt;BR /&gt;&lt;BR /&gt;If this does not meet your needs, are the differing volume/workloads known in advance? If so, could different compute be provisioned using Infrastructure as Code based on the differing workload characteristics? Here's a doc on using Terraform with Databricks:&amp;nbsp;&lt;A href="https://docs.databricks.com/en/dev-tools/terraform/index.html" target="_blank"&gt;https://docs.databricks.com/en/dev-tools/terraform/index.html&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;Thank you.&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Mon, 01 Jul 2024 17:30:22 GMT</pubDate>
    <dc:creator>brockb</dc:creator>
    <dc:date>2024-07-01T17:30:22Z</dc:date>
    <item>
      <title>Can we parameterize the compute in job cluster</title>
      <link>https://community.databricks.com/t5/data-engineering/can-we-parameterize-the-compute-in-job-cluster/m-p/76252#M35169</link>
      <description>&lt;P&gt;I have created a workflow job in databricks with job parameters.&lt;/P&gt;&lt;P&gt;I want to run the job same with different workloads and data volume.&lt;/P&gt;&lt;P&gt;So I want the compute cluster to be parametrized so that I can pass the compute requirements(driver, executor size and number of nodes) dynamically when I run the job.&lt;/P&gt;&lt;P&gt;Is this possible in databricks?&lt;/P&gt;</description>
      <pubDate>Mon, 01 Jul 2024 10:32:05 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/can-we-parameterize-the-compute-in-job-cluster/m-p/76252#M35169</guid>
      <dc:creator>NarenderKumar</dc:creator>
      <dc:date>2024-07-01T10:32:05Z</dc:date>
    </item>
    <item>
      <title>Re: Can we parameterize the compute in job cluster</title>
      <link>https://community.databricks.com/t5/data-engineering/can-we-parameterize-the-compute-in-job-cluster/m-p/76348#M35202</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/55535"&gt;@NarenderKumar&lt;/a&gt;&amp;nbsp;, If you want to change an existing job compute you would have to update the job settings before triggering a new run. Feel free to open a feature request with your idea through the Databricks&amp;nbsp;&lt;A class="reference external" href="https://ideas.databricks.com/?_ga=2.165990520.773921462.1719851023-704541007.1707435477" target="_blank" rel="noopener"&gt;Ideas Portal&lt;/A&gt;.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 01 Jul 2024 17:26:45 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/can-we-parameterize-the-compute-in-job-cluster/m-p/76348#M35202</guid>
      <dc:creator>raphaelblg</dc:creator>
      <dc:date>2024-07-01T17:26:45Z</dc:date>
    </item>
    <item>
      <title>Re: Can we parameterize the compute in job cluster</title>
      <link>https://community.databricks.com/t5/data-engineering/can-we-parameterize-the-compute-in-job-cluster/m-p/76351#M35204</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/55535"&gt;@NarenderKumar&lt;/a&gt;&amp;nbsp;, &lt;BR /&gt;&lt;BR /&gt;Have you considered leveraging&amp;nbsp;&lt;A href="https://docs.databricks.com/en/compute/configure.html#enable-autoscaling" target="_self"&gt;autoscaling&lt;/A&gt;&amp;nbsp;for the existing cluster?&lt;BR /&gt;&lt;BR /&gt;If this does not meet your needs, are the differing volume/workloads known in advance? If so, could different compute be provisioned using Infrastructure as Code based on the differing workload characteristics? Here's a doc on using Terraform with Databricks:&amp;nbsp;&lt;A href="https://docs.databricks.com/en/dev-tools/terraform/index.html" target="_blank"&gt;https://docs.databricks.com/en/dev-tools/terraform/index.html&lt;/A&gt;&lt;/P&gt;
&lt;P&gt;Thank you.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 01 Jul 2024 17:30:22 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/can-we-parameterize-the-compute-in-job-cluster/m-p/76351#M35204</guid>
      <dc:creator>brockb</dc:creator>
      <dc:date>2024-07-01T17:30:22Z</dc:date>
    </item>
  </channel>
</rss>

