[ 
https://issues.apache.org/jira/browse/SPARK-2890?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14095224#comment-14095224
 ] 

Jianshi Huang commented on SPARK-2890:
--------------------------------------

I think the fault is on my side. I should've changed project the duplicated 
columns into different names.

So the current behavior makes sense. I'll close this issue.

Jianshi

> Spark SQL should allow SELECT with duplicated columns
> -----------------------------------------------------
>
>                 Key: SPARK-2890
>                 URL: https://issues.apache.org/jira/browse/SPARK-2890
>             Project: Spark
>          Issue Type: Bug
>          Components: SQL
>    Affects Versions: 1.1.0
>            Reporter: Jianshi Huang
>
> Spark reported error java.lang.IllegalArgumentException with messages:
> java.lang.IllegalArgumentException: requirement failed: Found fields with the 
> same name.
>         at scala.Predef$.require(Predef.scala:233)
>         at 
> org.apache.spark.sql.catalyst.types.StructType.<init>(dataTypes.scala:317)
>         at 
> org.apache.spark.sql.catalyst.types.StructType$.fromAttributes(dataTypes.scala:310)
>         at 
> org.apache.spark.sql.parquet.ParquetTypesConverter$.convertToString(ParquetTypes.scala:306)
>         at 
> org.apache.spark.sql.parquet.ParquetTableScan.execute(ParquetTableOperations.scala:83)
>         at 
> org.apache.spark.sql.execution.Filter.execute(basicOperators.scala:57)
>         at 
> org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:85)
>         at org.apache.spark.sql.SchemaRDD.collect(SchemaRDD.scala:433)
> After trial and error, it seems it's caused by duplicated columns in my 
> select clause.
> I made the duplication on purpose for my code to parse correctly. I think we 
> should allow users to specify duplicated columns as return value.
> Jianshi



--
This message was sent by Atlassian JIRA
(v6.2#6252)

---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org
For additional commands, e-mail: issues-h...@spark.apache.org

Reply via email to