fix: Spark Schema Evolution Fix for nested columns #14075
+191
−6
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Describe the issue this Pull Request addresses
Currently when we are trying to read a Parquet file through Spark and the file has a different schema than the requested schema we can have runtime errors. These errors happen specifically when a column contains a record and the query selects a subset of columns from that record. If the schema of the record for that particular data file does not have the any of the requested fields, then there will be an error.
Summary and Changelog
The Spark Parquet reader is updated to remove any fields from the selected fields if the those fields result in records with no sub-fields.
This is validated by enabling the FileGroupReader path in one of the existing tests.
Impact
Fixes a bug in the FileGroupReader path for Spark reads.
Risk Level
Low
Documentation Update
Contributor's checklist