Hi,

I am attempting to read a number of smaller parquet files and merge them into a 
larger parquet file.

The files are created by Spark jobs that run periodically throughout the day.

The issue I have is that the small parquet files can have slightly different 
schemas and when I create the Dataset it complains that the schemas aren’t the 
same. 

Spark handles this by merging the schemas together, is there functionality in 
pyarrow that can do the same?

Thanks,
Dan

Reply via email to