wangyum opened a new pull request #24346: [SPARK-27176][FOLLOW-UP][SQL] Upgrade 
Hive parquet to 1.10.1
URL: https://github.com/apache/spark/pull/24346
 
 
   ## What changes were proposed in this pull request?
   
   When we compile and test Hadoop 3.2, we will hint the following two issues:
   1. JobSummaryLevel is not a member of object 
org.apache.parquet.hadoop.ParquetOutputFormat. Fixed by 
[PARQUET-381](https://issues.apache.org/jira/browse/PARQUET-381)(Parquet 1.9.0)
   2. java.lang.NoSuchFieldError: BROTLI
       at 
org.apache.parquet.hadoop.metadata.CompressionCodecName.<clinit>(CompressionCodecName.java:31).
 Fixed by 
[PARQUET-1143](https://issues.apache.org/jira/browse/PARQUET-1143)(Parquet 
1.10.0)
   
   The reason is that the `parquet-hadoop-bundle-1.8.1.jar` conflicts with 
Parquet 1.10.1.
   I think it would be safe to upgrade Hive's parquet to 1.10.1 to workaround 
this issue.
   
   This is what Hive did when upgrading Parquet 1.8.1 to 1.10.0: 
[HIVE-17000](https://issues.apache.org/jira/browse/HIVE-17000) and 
[HIVE-19464](https://issues.apache.org/jira/browse/HIVE-19464). We can see that 
all changes are related to vectors, and vectors are disabled by default: see 
[HIVE-14826](https://issues.apache.org/jira/browse/HIVE-14826) and 
[HiveConf.java#L2723](https://github.com/apache/hive/blob/rel/release-2.3.4/common/src/java/org/apache/hadoop/hive/conf/HiveConf.java#L2723).
   
   ## How was this patch tested?
   
   1. manual tests
   2. [upgrade Hive Parquet to 1.10.1 annd run Hadoop 3.2 test on 
jenkins](https://github.com/apache/spark/pull/24044#commits-pushed-0c3f962)
   

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to