The old statistics have many problems. The sorting order was not defined
properly and by specification it does not care about the logical type which can
modify the order (e.g. UTF8 vs. DECIMAL for the primitive type BINARY.) See
PARQUET-686 <https://issues.apache.org/jira/browse/PARQUET-686> for more
details. If you use parquet 1.9.0 you can use the configuration
"parquet.strings.signed-min-max.enabled” to read/write statistics for
string-ish BINARY logical types (none, UTF8, ENUM, JSON). If you use it, be
sure that all of your related BINARY values use only the lower 7 bits of their
bytes so the signed comparison would produce the same results as the proper
Because of these problems with the already existing incorrect min/max values we
decided to specify new ones. It is already implemented but not yet released.
See PARQUET-1025 <https://issues.apache.org/jira/browse/PARQUET-1025> for
> On 13 Feb 2018, at 19:04, Siva Gudavalli <gudavalli.s...@yahoo.com.INVALID>
> 3 MB files are too small for parquet, Try to increase the size.
> Keep an eye on statistics. In our case, we haven’t seen statistics being
> generated for string data types and will perform a Scan.
>> On Feb 12, 2018, at 9:24 PM, ilegend <511618...@qq.com> wrote:
>> Hi guys,
>> We're testing parquet performance for our big data environment. Parquet is
>> better than orc, but we believe that the parquet has more potential. Any
>> comments and suggestions are welcomed. The test environment is as follows:
>> 1. Server 48 cores + 256gb memory.
>> 2. Spark 2.1.0 + hdfs 2.6.0 + parquet-mr-1.8.1
>> 3. The size of hdfs file is 3MB.
>> 4. Parquet-me sets default values, row group size 128MB, data page size 1MB.
>> 发自我的 iPhone