Hi guys, We're testing parquet performance for our big data environment. Parquet is better than orc, but we believe that the parquet has more potential. Any comments and suggestions are welcomed. The test environment is as follows: 1. Server 48 cores + 256gb memory. 2. Spark 2.1.0 + hdfs 2.6.0 + parquet-mr-1.8.1 +parquet-format-2.3.0-incubating. 3. The size of hdfs file is 3MB. 4. Parquet-me sets default values, row group size 128MB, data page size 1MB.