[ https://issues.apache.org/jira/browse/SPARK-15029?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Apache Spark reassigned SPARK-15029: ------------------------------------ Assignee: Apache Spark > Bad error message for two generators in the project clause > ---------------------------------------------------------- > > Key: SPARK-15029 > URL: https://issues.apache.org/jira/browse/SPARK-15029 > Project: Spark > Issue Type: Bug > Components: SQL > Reporter: Reynold Xin > Assignee: Apache Spark > > {code} > scala> spark.range(1000).map(i => (Array[Long](i), > Array[Long](i))).selectExpr("explode(_1)", "explode(_2)").explain(true) > org.apache.spark.sql.AnalysisException: Only one generator allowed per select > but Generate and and Explode found.; > at > org.apache.spark.sql.catalyst.analysis.CheckAnalysis$class.failAnalysis(CheckAnalysis.scala:39) > at > org.apache.spark.sql.catalyst.analysis.Analyzer.failAnalysis(Analyzer.scala:54) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveGenerate$$anonfun$apply$21$$anonfun$53.apply(Analyzer.scala:1275) > at > org.apache.spark.sql.catalyst.analysis.Analyzer$ResolveGenerate$$anonfun$apply$21$$anonfun$53.apply(Analyzer.scala:1272) > at > scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241) > at > scala.collection.TraversableLike$$anonfun$flatMap$1.apply(TraversableLike.scala:241) > at > scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59) > at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:48) > at scala.collection.TraversableLike$class.flatMap(TraversableLike.scala:241) > at scala.collection.AbstractTraversable.flatMap(Traversable.scala:104) > {code} > It's confusing to call one "Generator" and the other "Explode". There is also > two "and"s. -- This message was sent by Atlassian JIRA (v6.3.4#6332) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org