Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/22343
What I asked was the following, wasn't it?
> In case-insensitive mode, when converting hive parquet table to parquet
data source, we switch the duplicated fields resolution mode to ask parquet
data source to pick the first matched field - the same behavior as hive parquet
table - to keep behaviors consistent.
Spark should not pick up the first matched field in any cases because it's
considered as a correctness behavior in previous PR which is backported to
`branch-2.3` https://github.com/apache/spark/pull/22183. I don't think we need
to follow incorrect Hive behavior.
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]