Github user dongjoon-hyun commented on the issue:

    https://github.com/apache/spark/pull/22262
  
    @seancxmao . Could you explain why we need this PR? Apache Spark 2.3.1 
already shows exceptions like the following for the case (for both ORC and 
Parquet), doesn't it?
    
    ```scala
    scala> spark.version
    res5: String = 2.3.1
    
    scala> sql("set spark.sql.caseSensitive=true")
    scala> spark.read.orc("/tmp/o").printSchema
    root
     |-- a: integer (nullable = true)
     |-- A: integer (nullable = true)
    
    scala> sql("set spark.sql.caseSensitive=false")
    scala> spark.read.orc("/tmp/o").printSchema
    18/09/01 20:06:05 WARN DataSource: Found duplicate column(s) in the data 
schema and the partition schema: `a`;
    org.apache.spark.sql.AnalysisException: Found duplicate column(s) in the 
data schema: `a`;
    ```


---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to