One of the source systems generates from time to time a parquet file which is only 220kb in size.
But reading it fails.
"java.io.IOException: Could not read or convert schema for file: 1-2022-00-51-56.parquet
Caused by: org.apache.spark.sql.AnalysisException: Parquet type not supported: INT32 (UINT_32);
"
I tried to use a schema and mergeSchema option
df =spark.read.options(mergeSchema=True).schema(mdd_schema_struct).parquet(target)
This is able to read the file and display but if you run count or merge it it would fail with
"Caused by: java.lang.RuntimeException: Illegal row group of 0 rows"
Does anyone know what could be the issue.