<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic why doesn't databricks allow setting executor metrics in Administration &amp; Architecture</title>
    <link>https://community.databricks.com/t5/administration-architecture/why-doesn-t-databricks-allow-setting-executor-metrics/m-p/108312#M2912</link>
    <description>&lt;P&gt;I have an all-purpose compute cluster that processes different data sets for various jobs. I am struggling to optimize executor metrics like below.&lt;BR /&gt;spark.executor.memory 4g&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;Is it allowed to override default executor metrics and specify such configurations at the cluster level for an all-purpose compute cluster? (in&amp;nbsp;&lt;SPAN&gt;Spark config section under Advance cluster options&lt;/SPAN&gt;)&lt;BR /&gt;How do I specify such configurations at runtime while submitting a job to a job-compute cluster?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Sat, 01 Feb 2025 19:19:22 GMT</pubDate>
    <dc:creator>marc88</dc:creator>
    <dc:date>2025-02-01T19:19:22Z</dc:date>
    <item>
      <title>why doesn't databricks allow setting executor metrics</title>
      <link>https://community.databricks.com/t5/administration-architecture/why-doesn-t-databricks-allow-setting-executor-metrics/m-p/108312#M2912</link>
      <description>&lt;P&gt;I have an all-purpose compute cluster that processes different data sets for various jobs. I am struggling to optimize executor metrics like below.&lt;BR /&gt;spark.executor.memory 4g&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;Is it allowed to override default executor metrics and specify such configurations at the cluster level for an all-purpose compute cluster? (in&amp;nbsp;&lt;SPAN&gt;Spark config section under Advance cluster options&lt;/SPAN&gt;)&lt;BR /&gt;How do I specify such configurations at runtime while submitting a job to a job-compute cluster?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Sat, 01 Feb 2025 19:19:22 GMT</pubDate>
      <guid>https://community.databricks.com/t5/administration-architecture/why-doesn-t-databricks-allow-setting-executor-metrics/m-p/108312#M2912</guid>
      <dc:creator>marc88</dc:creator>
      <dc:date>2025-02-01T19:19:22Z</dc:date>
    </item>
    <item>
      <title>Re: why doesn't databricks allow setting executor metrics</title>
      <link>https://community.databricks.com/t5/administration-architecture/why-doesn-t-databricks-allow-setting-executor-metrics/m-p/108329#M2915</link>
      <description>&lt;P&gt;Hello&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/93587"&gt;@marc88&lt;/a&gt;,&lt;/P&gt;
&lt;P&gt;As you mentioned in Spark config under Advance cluster options you can do it &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt; once cluster boots up it will be set at run level. Or you can draft a cluster policy and apply it across for job computes when creating your workflow.&lt;/P&gt;</description>
      <pubDate>Sun, 02 Feb 2025 02:27:33 GMT</pubDate>
      <guid>https://community.databricks.com/t5/administration-architecture/why-doesn-t-databricks-allow-setting-executor-metrics/m-p/108329#M2915</guid>
      <dc:creator>Alberto_Umana</dc:creator>
      <dc:date>2025-02-02T02:27:33Z</dc:date>
    </item>
  </channel>
</rss>

