<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: The driver is temporarily unavailable in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/the-driver-is-temporarily-unavailable/m-p/19096#M12753</link>
    <description>&lt;P&gt;The error messages mean the Spark driver is having insufficient resources.  Although increasing the driver memory or using a bigger instance type is a quick workaround, identifying the issue is the key thing. &lt;/P&gt;&lt;P&gt;On the application side , review if there are operations which is driver intensive. Say, collect() or toPandas() etc. &lt;/P&gt;&lt;P&gt;Also check if the instance type used is adequate for the workload. &lt;/P&gt;&lt;P&gt;&lt;/P&gt;</description>
    <pubDate>Fri, 25 Jun 2021 18:47:58 GMT</pubDate>
    <dc:creator>brickster_2018</dc:creator>
    <dc:date>2021-06-25T18:47:58Z</dc:date>
    <item>
      <title>The driver is temporarily unavailable</title>
      <link>https://community.databricks.com/t5/data-engineering/the-driver-is-temporarily-unavailable/m-p/19095#M12752</link>
      <description>&lt;P&gt;My job fails with Driver is temporarily unavailable. Apparently, it's permanently unavailable, because the job is not pausing but failing. &lt;/P&gt;</description>
      <pubDate>Fri, 25 Jun 2021 18:43:48 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/the-driver-is-temporarily-unavailable/m-p/19095#M12752</guid>
      <dc:creator>brickster_2018</dc:creator>
      <dc:date>2021-06-25T18:43:48Z</dc:date>
    </item>
    <item>
      <title>Re: The driver is temporarily unavailable</title>
      <link>https://community.databricks.com/t5/data-engineering/the-driver-is-temporarily-unavailable/m-p/19096#M12753</link>
      <description>&lt;P&gt;The error messages mean the Spark driver is having insufficient resources.  Although increasing the driver memory or using a bigger instance type is a quick workaround, identifying the issue is the key thing. &lt;/P&gt;&lt;P&gt;On the application side , review if there are operations which is driver intensive. Say, collect() or toPandas() etc. &lt;/P&gt;&lt;P&gt;Also check if the instance type used is adequate for the workload. &lt;/P&gt;&lt;P&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 25 Jun 2021 18:47:58 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/the-driver-is-temporarily-unavailable/m-p/19096#M12753</guid>
      <dc:creator>brickster_2018</dc:creator>
      <dc:date>2021-06-25T18:47:58Z</dc:date>
    </item>
    <item>
      <title>Re: The driver is temporarily unavailable</title>
      <link>https://community.databricks.com/t5/data-engineering/the-driver-is-temporarily-unavailable/m-p/39891#M27087</link>
      <description>&lt;P&gt;I am facing the same issues .&amp;nbsp; I am writing in batches using a simple for loop. I don't have any collect statements inside the loop. I am rewriting the partitions with partition overwrite dynamic mode in a huge wide delta table - several tb. The incremental load is sometimes 1-2 tb&lt;/P&gt;</description>
      <pubDate>Mon, 14 Aug 2023 20:10:17 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/the-driver-is-temporarily-unavailable/m-p/39891#M27087</guid>
      <dc:creator>Chalki</dc:creator>
      <dc:date>2023-08-14T20:10:17Z</dc:date>
    </item>
  </channel>
</rss>

