<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>topic Importing JSON files when format is subject to evolution in Data Engineering</title>
    <link>https://community.databricks.com/t5/data-engineering/importing-json-files-when-format-is-subject-to-evolution/m-p/68832#M33753</link>
    <description>&lt;P&gt;Hi there,&lt;/P&gt;&lt;P&gt;I'm reaching out for some assistance with importing JSON files into Databricks. Still a beginner even if I've gained experience working with various data import batches (CSV/JSON) for application monitoring:&amp;nbsp; I'm currently facing a challenge with a specific JSON data set.&lt;/P&gt;&lt;P&gt;The structure of this JSON file has evolved over the past year, with new fields being added. As a result, our current Python code using inferSchema to automatically detect the format is encountering errors during import.&lt;/P&gt;&lt;P&gt;I've explored several approaches to handle schema evolution, including concepts like schema merging, but haven't yet found a solution that works consistently. I believe I'm close to a solution, but I'm running into some roadblocks.&lt;/P&gt;&lt;P&gt;Any insights or suggestions you might have on handling evolving JSON schemas during import would be greatly appreciated.&lt;/P&gt;&lt;P&gt;Thanks in advance for your help!&lt;/P&gt;</description>
    <pubDate>Sun, 12 May 2024 21:31:06 GMT</pubDate>
    <dc:creator>etum</dc:creator>
    <dc:date>2024-05-12T21:31:06Z</dc:date>
    <item>
      <title>Importing JSON files when format is subject to evolution</title>
      <link>https://community.databricks.com/t5/data-engineering/importing-json-files-when-format-is-subject-to-evolution/m-p/68832#M33753</link>
      <description>&lt;P&gt;Hi there,&lt;/P&gt;&lt;P&gt;I'm reaching out for some assistance with importing JSON files into Databricks. Still a beginner even if I've gained experience working with various data import batches (CSV/JSON) for application monitoring:&amp;nbsp; I'm currently facing a challenge with a specific JSON data set.&lt;/P&gt;&lt;P&gt;The structure of this JSON file has evolved over the past year, with new fields being added. As a result, our current Python code using inferSchema to automatically detect the format is encountering errors during import.&lt;/P&gt;&lt;P&gt;I've explored several approaches to handle schema evolution, including concepts like schema merging, but haven't yet found a solution that works consistently. I believe I'm close to a solution, but I'm running into some roadblocks.&lt;/P&gt;&lt;P&gt;Any insights or suggestions you might have on handling evolving JSON schemas during import would be greatly appreciated.&lt;/P&gt;&lt;P&gt;Thanks in advance for your help!&lt;/P&gt;</description>
      <pubDate>Sun, 12 May 2024 21:31:06 GMT</pubDate>
      <guid>https://community.databricks.com/t5/data-engineering/importing-json-files-when-format-is-subject-to-evolution/m-p/68832#M33753</guid>
      <dc:creator>etum</dc:creator>
      <dc:date>2024-05-12T21:31:06Z</dc:date>
    </item>
  </channel>
</rss>

