<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Why configure a job timeout? in Community Articles</title>
    <link>https://community.databricks.com/t5/community-articles/why-configure-a-job-timeout/m-p/71339#M102</link>
    <description>&lt;P&gt;If you use Databricks Jobs for your workloads, it is possible you might have run into a situation where you find your jobs to be in "hung" state.&lt;/P&gt;
&lt;P&gt;Before cancelling the job it is important to collect the thread dump as I described &lt;A href="https://community.databricks.com/t5/community-discussions/how-to-collect-a-thread-dump-from-databricks-spark-ui/td-p/68979&amp;nbsp;" target="_self"&gt;here&lt;/A&gt;&amp;nbsp;to be able to find the root cause.&lt;/P&gt;
&lt;P&gt;But how do we not run into a situation where our jobs are in the "hung" state for a prolonged time?&lt;/P&gt;
&lt;P&gt;If you know the expected time of the job completion, with a buffer you should always set Job timeouts as described &lt;A href="https://docs.databricks.com/en/workflows/jobs/settings.html#configure-an-expected-completion-time-or-a-timeout-for-a-job" target="_self"&gt;here&lt;/A&gt;.&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;To configure a maximum completion time for a job, enter the maximum duration in the&amp;nbsp;&lt;/SPAN&gt;&lt;STRONG&gt;Timeout&lt;/STRONG&gt;&lt;SPAN&gt;&amp;nbsp;field. If the job does not complete in this time, Databricks sets its status to “Timed Out” and the job is stopped.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;For Streaming Jobs check my post &lt;A href="https://community.databricks.com/t5/knowledge-sharing-hub/monitoring-a-streaming-job/m-p/71340#M103" target="_self"&gt;here&lt;/A&gt;.&lt;/P&gt;</description>
    <pubDate>Sat, 01 Jun 2024 21:46:59 GMT</pubDate>
    <dc:creator>NandiniN</dc:creator>
    <dc:date>2024-06-01T21:46:59Z</dc:date>
    <item>
      <title>Why configure a job timeout?</title>
      <link>https://community.databricks.com/t5/community-articles/why-configure-a-job-timeout/m-p/71339#M102</link>
      <description>&lt;P&gt;If you use Databricks Jobs for your workloads, it is possible you might have run into a situation where you find your jobs to be in "hung" state.&lt;/P&gt;
&lt;P&gt;Before cancelling the job it is important to collect the thread dump as I described &lt;A href="https://community.databricks.com/t5/community-discussions/how-to-collect-a-thread-dump-from-databricks-spark-ui/td-p/68979&amp;nbsp;" target="_self"&gt;here&lt;/A&gt;&amp;nbsp;to be able to find the root cause.&lt;/P&gt;
&lt;P&gt;But how do we not run into a situation where our jobs are in the "hung" state for a prolonged time?&lt;/P&gt;
&lt;P&gt;If you know the expected time of the job completion, with a buffer you should always set Job timeouts as described &lt;A href="https://docs.databricks.com/en/workflows/jobs/settings.html#configure-an-expected-completion-time-or-a-timeout-for-a-job" target="_self"&gt;here&lt;/A&gt;.&amp;nbsp;&lt;/P&gt;
&lt;P&gt;&lt;SPAN&gt;To configure a maximum completion time for a job, enter the maximum duration in the&amp;nbsp;&lt;/SPAN&gt;&lt;STRONG&gt;Timeout&lt;/STRONG&gt;&lt;SPAN&gt;&amp;nbsp;field. If the job does not complete in this time, Databricks sets its status to “Timed Out” and the job is stopped.&lt;/SPAN&gt;&lt;/P&gt;
&lt;P&gt;For Streaming Jobs check my post &lt;A href="https://community.databricks.com/t5/knowledge-sharing-hub/monitoring-a-streaming-job/m-p/71340#M103" target="_self"&gt;here&lt;/A&gt;.&lt;/P&gt;</description>
      <pubDate>Sat, 01 Jun 2024 21:46:59 GMT</pubDate>
      <guid>https://community.databricks.com/t5/community-articles/why-configure-a-job-timeout/m-p/71339#M102</guid>
      <dc:creator>NandiniN</dc:creator>
      <dc:date>2024-06-01T21:46:59Z</dc:date>
    </item>
  </channel>
</rss>

