<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Re: Fatal error when writing a big pandas dF in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/fatal-error-when-writing-a-big-pandas-df/m-p/59703#M31493</link>
    <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/92566"&gt;@chari&lt;/a&gt;&amp;nbsp;,Thanks for bringing up your concerns, always happy to help&amp;nbsp;&lt;span class="lia-unicode-emoji" title=":beaming_face_with_smiling_eyes:"&gt;😁&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;We understand that you are facing the following error while you are writing a pandas dataframe containing 100000rows in excel.&lt;/P&gt;
&lt;P&gt;As per the Error &amp;gt;&amp;gt;&amp;gt; Fatal error: The Python kernel is unresponsive. The Python process exited with exit code 137 (SIGKILL: Killed). This may have been caused by an OOM error. Check your command's memory usage.&lt;/P&gt;
&lt;P&gt;The driver node is OOM leading to this error. This can be fixed by:&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Choosing a higher driver instance size depending on the workload.&lt;/LI&gt;
&lt;LI&gt;Splitting workloads into multiple clusters&lt;/LI&gt;
&lt;LI&gt;Moving to a jobs cluster&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;Please let me know if this helps and leave a like if this helps, followups are appreciated.&lt;BR /&gt;Kudos&lt;BR /&gt;Ayushi&lt;/P&gt;</description>
    <pubDate>Thu, 08 Feb 2024 13:33:53 GMT</pubDate>
    <dc:creator>Ayushi_Suthar</dc:creator>
    <dc:date>2024-02-08T13:33:53Z</dc:date>
    <item>
      <title>Fatal error when writing a big pandas dF</title>
      <link>https://community.databricks.com/t5/data-engineering/fatal-error-when-writing-a-big-pandas-df/m-p/59696#M31486</link>
      <description>&lt;P&gt;Hello DB community,&lt;/P&gt;&lt;P&gt;I was trying to write a pandas dataframe containing 100000 rows as excel. Moments in the execution I received a fatal error : "Python kernel is unresponsive."&lt;/P&gt;&lt;P&gt;However, I am constrained from increasing the number of clusters or other relevant infrastructure change. So my only option is to fix the code.&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am looking to implement more than one option I would get from this forum. pls suggest.&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;</description>
      <pubDate>Thu, 08 Feb 2024 12:47:46 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/fatal-error-when-writing-a-big-pandas-df/m-p/59696#M31486</guid>
      <dc:creator>chari</dc:creator>
      <dc:date>2024-02-08T12:47:46Z</dc:date>
    </item>
    <item>
      <title>Re: Fatal error when writing a big pandas dF</title>
      <link>https://community.databricks.com/t5/data-engineering/fatal-error-when-writing-a-big-pandas-df/m-p/59703#M31493</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.databricks.com/t5/user/viewprofilepage/user-id/92566"&gt;@chari&lt;/a&gt;&amp;nbsp;,Thanks for bringing up your concerns, always happy to help&amp;nbsp;&lt;span class="lia-unicode-emoji" title=":beaming_face_with_smiling_eyes:"&gt;😁&lt;/span&gt;&lt;/P&gt;
&lt;P&gt;We understand that you are facing the following error while you are writing a pandas dataframe containing 100000rows in excel.&lt;/P&gt;
&lt;P&gt;As per the Error &amp;gt;&amp;gt;&amp;gt; Fatal error: The Python kernel is unresponsive. The Python process exited with exit code 137 (SIGKILL: Killed). This may have been caused by an OOM error. Check your command's memory usage.&lt;/P&gt;
&lt;P&gt;The driver node is OOM leading to this error. This can be fixed by:&lt;/P&gt;
&lt;UL&gt;
&lt;LI&gt;Choosing a higher driver instance size depending on the workload.&lt;/LI&gt;
&lt;LI&gt;Splitting workloads into multiple clusters&lt;/LI&gt;
&lt;LI&gt;Moving to a jobs cluster&lt;/LI&gt;
&lt;/UL&gt;
&lt;P&gt;Please let me know if this helps and leave a like if this helps, followups are appreciated.&lt;BR /&gt;Kudos&lt;BR /&gt;Ayushi&lt;/P&gt;</description>
      <pubDate>Thu, 08 Feb 2024 13:33:53 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/fatal-error-when-writing-a-big-pandas-df/m-p/59703#M31493</guid>
      <dc:creator>Ayushi_Suthar</dc:creator>
      <dc:date>2024-02-08T13:33:53Z</dc:date>
    </item>
    <item>
      <title>Re: Fatal error when writing a big pandas dF</title>
      <link>https://community.databricks.com/t5/data-engineering/fatal-error-when-writing-a-big-pandas-df/m-p/60002#M31560</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;Unfortunately, I cant update my cluster until six months.&amp;nbsp;&lt;/P&gt;&lt;P&gt;But I want to use a spark dataframe to write as a CSV. Does it help?&lt;/P&gt;</description>
      <pubDate>Tue, 13 Feb 2024 07:19:11 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/fatal-error-when-writing-a-big-pandas-df/m-p/60002#M31560</guid>
      <dc:creator>chari</dc:creator>
      <dc:date>2024-02-13T07:19:11Z</dc:date>
    </item>
  </channel>
</rss>

