<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Running a jar on Databricks shared cluster using Airflow in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/running-a-jar-on-databricks-shared-cluster-using-airflow/m-p/99972#M40160</link>
    <description>&lt;P&gt;Hello&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/112661"&gt;@ayush19&lt;/a&gt;,&lt;/P&gt;
&lt;P class="p1"&gt;Here are some suggestions, but would need to check how are the parameters configured.&lt;/P&gt;
&lt;P class="p1"&gt;&lt;STRONG&gt;Use an Existing Cluster&lt;/STRONG&gt;: Instead of creating a new cluster each time, configure the DatabricksSubmitRunOperator to use an existing cluster. This can be done by specifying the existing_cluster_id parameter in the operator. This way, the cluster will not restart, and the jar file will not be reinstalled.&lt;/P&gt;
&lt;P class="p1"&gt;&lt;STRONG&gt;Cluster Configuration&lt;/STRONG&gt;: Ensure that the cluster configuration does not force instance replacement upon restart. According to the context, one way to achieve this is by disabling multi-AZ (Availability Zone) selection in the cluster configuration. This can help in reusing the same instances rather than creating new ones&lt;/P&gt;</description>
    <pubDate>Mon, 25 Nov 2024 15:53:07 GMT</pubDate>
    <dc:creator>Alberto_Umana</dc:creator>
    <dc:date>2024-11-25T15:53:07Z</dc:date>
    <item>
      <title>Running a jar on Databricks shared cluster using Airflow</title>
      <link>https://community.databricks.com/t5/data-engineering/running-a-jar-on-databricks-shared-cluster-using-airflow/m-p/99928#M40145</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;I have a requirement to run a jar already installed on a Databricks cluster. It needs to be orchestrated using Apache Airflow.&amp;nbsp;&lt;/P&gt;&lt;P&gt;I followed the docs for the operator which can be used to do so&amp;nbsp;&lt;A href="https://airflow.apache.org/docs/apache-airflow-providers-databricks/1.0.0/operators.html" target="_blank"&gt;https://airflow.apache.org/docs/apache-airflow-providers-databricks/1.0.0/operators.html&lt;/A&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The issue is, that every time I run this DAG, the cluster restarts and the jar file is installed on the cluster again. The file is already stored in Volume and installed on cluster, yet it restarts and re-installs jar on cluster.&amp;nbsp;&lt;/P&gt;&lt;P&gt;How can I avoid this?&lt;/P&gt;</description>
      <pubDate>Mon, 25 Nov 2024 07:05:16 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-a-jar-on-databricks-shared-cluster-using-airflow/m-p/99928#M40145</guid>
      <dc:creator>ayush19</dc:creator>
      <dc:date>2024-11-25T07:05:16Z</dc:date>
    </item>
    <item>
      <title>Re: Running a jar on Databricks shared cluster using Airflow</title>
      <link>https://community.databricks.com/t5/data-engineering/running-a-jar-on-databricks-shared-cluster-using-airflow/m-p/99972#M40160</link>
      <description>&lt;P&gt;Hello&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/112661"&gt;@ayush19&lt;/a&gt;,&lt;/P&gt;
&lt;P class="p1"&gt;Here are some suggestions, but would need to check how are the parameters configured.&lt;/P&gt;
&lt;P class="p1"&gt;&lt;STRONG&gt;Use an Existing Cluster&lt;/STRONG&gt;: Instead of creating a new cluster each time, configure the DatabricksSubmitRunOperator to use an existing cluster. This can be done by specifying the existing_cluster_id parameter in the operator. This way, the cluster will not restart, and the jar file will not be reinstalled.&lt;/P&gt;
&lt;P class="p1"&gt;&lt;STRONG&gt;Cluster Configuration&lt;/STRONG&gt;: Ensure that the cluster configuration does not force instance replacement upon restart. According to the context, one way to achieve this is by disabling multi-AZ (Availability Zone) selection in the cluster configuration. This can help in reusing the same instances rather than creating new ones&lt;/P&gt;</description>
      <pubDate>Mon, 25 Nov 2024 15:53:07 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-a-jar-on-databricks-shared-cluster-using-airflow/m-p/99972#M40160</guid>
      <dc:creator>Alberto_Umana</dc:creator>
      <dc:date>2024-11-25T15:53:07Z</dc:date>
    </item>
    <item>
      <title>Re: Running a jar on Databricks shared cluster using Airflow</title>
      <link>https://community.databricks.com/t5/data-engineering/running-a-jar-on-databricks-shared-cluster-using-airflow/m-p/100043#M40176</link>
      <description>&lt;P&gt;Hi Alberto,&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am using an existing cluster for it and not creating new cluster. I am using an all purpose cluster and which is used by multiple people in different regions so I'm not sure if I can disable Multi AZ. Is there a solution in which I can use an existing instance of cluster?&amp;nbsp;&lt;BR /&gt;Also if you could please explain why is it restarting exactly? the Jar file is already installed on cluster, then what's the need to install it again?&lt;/P&gt;</description>
      <pubDate>Tue, 26 Nov 2024 09:18:00 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/running-a-jar-on-databricks-shared-cluster-using-airflow/m-p/100043#M40176</guid>
      <dc:creator>ayush19</dc:creator>
      <dc:date>2024-11-26T09:18:00Z</dc:date>
    </item>
  </channel>
</rss>

