Schema Evolution in Azure databricks
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
02-28-2024 10:10 AM - edited 02-28-2024 10:15 AM
Hi All -
In my scenario, Loading data from 100 of Json files.
Problem is, fields/columns are missing when JSON file contains new fields.
Full Load:
while writing JSON to delta use the option ("mergeschema", "true") so that we do not miss new columns
Incremental Load:
Problem is here as schema does not match with existing schema.
Could you please assist with schema comparison while doing incremental load.
New JSON files schema should compare with existing JSON files schema.
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
01-13-2025 02:03 PM
For these scenarios, you can use schema evolution capabilities like mergeSchema or opt to use the new VariantType to avoid requiring a schema at time of ingest.

