Spark read Parquet files of different versions
问题 I have parquet files generated for over a year with a Version1 schema. And with a recent schema change the newer parquet files have Version2 schema extra columns. So when i load parquet files from the old version and new version together and try to filter on the changed columns i get an exception. I would like for spark to read old and new files and fill in null values where the column is not present.Is there a workaround for this where spark fills null values when the column is not found?