Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/22262
@seancxmao . Could you explain why we need this PR? Apache Spark 2.3.1
already shows exceptions like the following for the case (for both ORC and
Parquet), doesn't it?
```scala
scala> spark.version
res5: String = 2.3.1
scala> sql("set spark.sql.caseSensitive=true")
scala> spark.read.orc("/tmp/o").printSchema
root
|-- a: integer (nullable = true)
|-- A: integer (nullable = true)
scala> sql("set spark.sql.caseSensitive=false")
scala> spark.read.orc("/tmp/o").printSchema
18/09/01 20:06:05 WARN DataSource: Found duplicate column(s) in the data
schema and the partition schema: `a`;
org.apache.spark.sql.AnalysisException: Found duplicate column(s) in the
data schema: `a`;
```
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]