Github user rxin commented on the issue:
https://github.com/apache/spark/pull/16281
We haven't really added much to Hive though, and as a matter of fact the
dependency on Hive is decreasing. Parquet is a much more manageable piece of
code to fork. In the past we have seen fairly critical bugs with almost every
upgrade. and coupled with the fact that Parquet cannot always make releases
fast enough (yes it happened in the past when we asked to have released but
didn't get them), or have proper testing, it has always been very risky to just
upgrade a major version of Parquet. In addition, we already have a forked
Parquet reader in Spark that is vectorized (that is different from the one in
parquet mr).
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]