Github user dongjoon-hyun commented on the issue:
https://github.com/apache/spark/pull/22262
The following is the sequence.
```scala
scala> sql("insert overwrite local directory '/tmp/parquet' stored as
parquet select 1 id, 2 id")
```
```
$ parquet-tools schema /tmp/parquet
message hive_schema {
optional int32 id;
optional int32 id;
}
```
```scala
scala> sql("create table parquet(id int) USING parquet LOCATION
'/tmp/parquet'")
res3: org.apache.spark.sql.DataFrame = []
scala> sql("select * from parquet")
res4: org.apache.spark.sql.DataFrame = [id: int]
scala> sql("select * from parquet").show
18/09/07 23:31:03 ERROR Executor: Exception in task 0.0 in stage 2.0 (TID 2)
java.lang.RuntimeException: [id] INT32 was added twice
```
---
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]