Github user dongjoon-hyun commented on the issue:

    https://github.com/apache/spark/pull/22343
  
    What I asked was the following, wasn't it? 
    > In case-insensitive mode, when converting hive parquet table to parquet 
data source, we switch the duplicated fields resolution mode to ask parquet 
data source to pick the first matched field - the same behavior as hive parquet 
table - to keep behaviors consistent.
    
    Spark should not pick up the first matched field in any cases because it's 
considered as a correctness behavior in previous PR which is backported to 
`branch-2.3` https://github.com/apache/spark/pull/22183. I don't think we need 
to follow incorrect Hive behavior.


---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to