<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic How can process  3.5GB GZ (~90GB) nested JSON and convert them to tabular formats with less processing time and optimized cost in Azure Databricks? in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/how-can-process-3-5gb-gz-90gb-nested-json-and-convert-them-to/m-p/34054#M24848</link>
    <description>&lt;P&gt;I have a total of 5000 files (Nested JSON ~ 3.5 GB). I have written a code which converts the json to Table in minutes (for JSON size till 1 GB) but when I am trying to process 3.5GB GZ json it is mostly getting failed because of Garbage collection. I have tried multiple clusters as well, still it is taking 18 minutes to read the file only and since it is a nested json it reads it only as a single record.&lt;/P&gt;&lt;P&gt;Please find the sample JSON Structure.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="Data frame structure"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/1582i96FBF6A4B4D8BB8C/image-size/large?v=v2&amp;amp;px=999" role="button" title="Data frame structure" alt="Data frame structure" /&gt;&lt;/span&gt;Code snippet:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="Code"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/1576i9468D0CE9962A6F0/image-size/large?v=v2&amp;amp;px=999" role="button" title="Code" alt="Code" /&gt;&lt;/span&gt;Reading Code:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="Reading Code"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/1584iDEC8AC05168273BB/image-size/large?v=v2&amp;amp;px=999" role="button" title="Reading Code" alt="Reading Code" /&gt;&lt;/span&gt;I am looking a way first to process one 3.5 GZ file and after that my focus will be working on 5000 similar files. I am searching for a way which will be more optimized and cost effective. Currently I am using Azure Databricks but I am open you use any other alternate technology as well.&lt;/P&gt;</description>
    <pubDate>Mon, 22 Aug 2022 21:43:15 GMT</pubDate>
    <dc:creator>sanchit_popli</dc:creator>
    <dc:date>2022-08-22T21:43:15Z</dc:date>
    <item>
      <title>How can process  3.5GB GZ (~90GB) nested JSON and convert them to tabular formats with less processing time and optimized cost in Azure Databricks?</title>
      <link>https://community.databricks.com/t5/data-engineering/how-can-process-3-5gb-gz-90gb-nested-json-and-convert-them-to/m-p/34054#M24848</link>
      <description>&lt;P&gt;I have a total of 5000 files (Nested JSON ~ 3.5 GB). I have written a code which converts the json to Table in minutes (for JSON size till 1 GB) but when I am trying to process 3.5GB GZ json it is mostly getting failed because of Garbage collection. I have tried multiple clusters as well, still it is taking 18 minutes to read the file only and since it is a nested json it reads it only as a single record.&lt;/P&gt;&lt;P&gt;Please find the sample JSON Structure.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="Data frame structure"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/1582i96FBF6A4B4D8BB8C/image-size/large?v=v2&amp;amp;px=999" role="button" title="Data frame structure" alt="Data frame structure" /&gt;&lt;/span&gt;Code snippet:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="Code"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/1576i9468D0CE9962A6F0/image-size/large?v=v2&amp;amp;px=999" role="button" title="Code" alt="Code" /&gt;&lt;/span&gt;Reading Code:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper" image-alt="Reading Code"&gt;&lt;img src="https://community.databricks.com/t5/image/serverpage/image-id/1584iDEC8AC05168273BB/image-size/large?v=v2&amp;amp;px=999" role="button" title="Reading Code" alt="Reading Code" /&gt;&lt;/span&gt;I am looking a way first to process one 3.5 GZ file and after that my focus will be working on 5000 similar files. I am searching for a way which will be more optimized and cost effective. Currently I am using Azure Databricks but I am open you use any other alternate technology as well.&lt;/P&gt;</description>
      <pubDate>Mon, 22 Aug 2022 21:43:15 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/how-can-process-3-5gb-gz-90gb-nested-json-and-convert-them-to/m-p/34054#M24848</guid>
      <dc:creator>sanchit_popli</dc:creator>
      <dc:date>2022-08-22T21:43:15Z</dc:date>
    </item>
  </channel>
</rss>

