Github user dongjoon-hyun commented on the issue:

    https://github.com/apache/spark/pull/22262
  
    The following is the sequence.
    
    ```scala
    scala> sql("insert overwrite local directory '/tmp/parquet' stored as 
parquet select 1 id, 2 id")
    ```
    
    ```
    $ parquet-tools schema /tmp/parquet
    message hive_schema {
      optional int32 id;
      optional int32 id;
    }
    ```
    
    ```scala
    scala> sql("create table parquet(id int) USING parquet LOCATION 
'/tmp/parquet'")
    res3: org.apache.spark.sql.DataFrame = []
    
    scala> sql("select * from parquet")
    res4: org.apache.spark.sql.DataFrame = [id: int]
    
    scala> sql("select * from parquet").show
    18/09/07 23:31:03 ERROR Executor: Exception in task 0.0 in stage 2.0 (TID 2)
    java.lang.RuntimeException: [id] INT32 was added twice
    ```


---

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to