I am curious to get the community's thoughts on this. Is it generally preferrable to load raw data based on its inferred columns or not? And is it preferred to keep the raw data in its original structure or to flatten it into a more tabular structure? If it's better to keep the original structure, what is generally validated at the raw data table versus the subsequent "base" (flattened) data?
For example, assuming we already have a data extraction process to land .json files in a landing path, we can leverage AutoLoader to incrementally load data into a target raw table. If I leverage schema inferral, I can more easily flatten the data in subsequent steps, but the inferred values may not always be accurate, so I may want to either use `schemaHints` or maybe force typing via Spark SQL or something.
