Hi guys,
We're testing parquet performance for our big data environment. Parquet is 
better than orc, but we believe that the parquet has more potential. Any 
comments and suggestions are welcomed. The test environment is as follows:
1. Server 48 cores + 256gb memory.
2. Spark 2.1.0 + hdfs 2.6.0 + parquet-mr-1.8.1 +parquet-format-2.3.0-incubating.
3. The size of hdfs file is 3MB.
4. Parquet-me sets default values, row group size 128MB, data page size 1MB.

发自我的 iPhone

Reply via email to