[ https://issues.apache.org/jira/browse/SPARK-29134?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Jungtaek Lim resolved SPARK-29134. ---------------------------------- Resolution: Invalid Just affected by problematic commit (had been reverted) [[https://github.com/apache/spark/commit/850833fa177ec1f265e143fc383e40ec2c8341a6]] See more: [[https://github.com/apache/spark/pull/25790#issuecomment-531517800]] > Flaky test: org.apache.spark.sql.execution.SQLExecutionSuite.concurrent query > execution with fork-join pool (SPARK-13747) > ------------------------------------------------------------------------------------------------------------------------- > > Key: SPARK-29134 > URL: https://issues.apache.org/jira/browse/SPARK-29134 > Project: Spark > Issue Type: Bug > Components: SQL, Tests > Affects Versions: 3.0.0 > Reporter: Jungtaek Lim > Priority: Major > > [https://amplab.cs.berkeley.edu/jenkins/view/Spark%20QA%20Test/job/spark-master-test-maven-hadoop-2.7-jdk-11-ubuntu-testing/1544/testReport/] > {code:java} > scala.ScalaReflectionException: type T1 is not a class > at scala.reflect.api.Symbols$SymbolApi.asClass(Symbols.scala:284) > at scala.reflect.api.Symbols$SymbolApi.asClass$(Symbols.scala:284) > at > scala.reflect.internal.Symbols$SymbolContextApiImpl.asClass(Symbols.scala:106) > at > org.apache.spark.sql.catalyst.ScalaReflection$.getClassFromType(ScalaReflection.scala:617) > at > org.apache.spark.sql.catalyst.ScalaReflection$.$anonfun$dataTypeFor$1(ScalaReflection.scala:105) > at > scala.reflect.internal.tpe.TypeConstraints$UndoLog.undo(TypeConstraints.scala:68) > at > org.apache.spark.sql.catalyst.ScalaReflection.cleanUpReflectionObjects(ScalaReflection.scala:848) > at > org.apache.spark.sql.catalyst.ScalaReflection.cleanUpReflectionObjects$(ScalaReflection.scala:847) > at > org.apache.spark.sql.catalyst.ScalaReflection$.cleanUpReflectionObjects(ScalaReflection.scala:47) > at > org.apache.spark.sql.catalyst.ScalaReflection$.dataTypeFor(ScalaReflection.scala:86) > at > org.apache.spark.sql.catalyst.ScalaReflection$.$anonfun$serializerFor$6(ScalaReflection.scala:551) > at > scala.collection.TraversableLike.$anonfun$map$1(TraversableLike.scala:237) > at scala.collection.immutable.List.foreach(List.scala:392) > at scala.collection.TraversableLike.map(TraversableLike.scala:237) > at scala.collection.TraversableLike.map$(TraversableLike.scala:230) > at scala.collection.immutable.List.map(List.scala:298) > at > org.apache.spark.sql.catalyst.ScalaReflection$.$anonfun$serializerFor$1(ScalaReflection.scala:541) > at > scala.reflect.internal.tpe.TypeConstraints$UndoLog.undo(TypeConstraints.scala:68) > at > org.apache.spark.sql.catalyst.ScalaReflection.cleanUpReflectionObjects(ScalaReflection.scala:848) > at > org.apache.spark.sql.catalyst.ScalaReflection.cleanUpReflectionObjects$(ScalaReflection.scala:847) > at > org.apache.spark.sql.catalyst.ScalaReflection$.cleanUpReflectionObjects(ScalaReflection.scala:47) > at > org.apache.spark.sql.catalyst.ScalaReflection$.serializerFor(ScalaReflection.scala:410) > at > org.apache.spark.sql.catalyst.ScalaReflection$.$anonfun$serializerForType$1(ScalaReflection.scala:399) > at > scala.reflect.internal.tpe.TypeConstraints$UndoLog.undo(TypeConstraints.scala:68) > at > org.apache.spark.sql.catalyst.ScalaReflection.cleanUpReflectionObjects(ScalaReflection.scala:848) > at > org.apache.spark.sql.catalyst.ScalaReflection.cleanUpReflectionObjects$(ScalaReflection.scala:847) > at > org.apache.spark.sql.catalyst.ScalaReflection$.cleanUpReflectionObjects(ScalaReflection.scala:47) > at > org.apache.spark.sql.catalyst.ScalaReflection$.serializerForType(ScalaReflection.scala:391) > at > org.apache.spark.sql.catalyst.encoders.ExpressionEncoder$.apply(ExpressionEncoder.scala:54) > at org.apache.spark.sql.Encoders$.product(Encoders.scala:285) > at > org.apache.spark.sql.LowPrioritySQLImplicits.newProductEncoder(SQLImplicits.scala:251) > at > org.apache.spark.sql.LowPrioritySQLImplicits.newProductEncoder$(SQLImplicits.scala:251) > at > org.apache.spark.sql.SQLImplicits.newProductEncoder(SQLImplicits.scala:32) > at > org.apache.spark.sql.execution.SQLExecutionSuite.$anonfun$new$3(SQLExecutionSuite.scala:50) > at > scala.runtime.java8.JFunction1$mcJI$sp.apply(JFunction1$mcJI$sp.java:23) > at scala.Function1.apply$mcVI$sp(Function1.scala:41) > at > scala.collection.parallel.immutable.ParRange$ParRangeIterator.foreach(ParRange.scala:94) > at > scala.collection.parallel.ParIterableLike$Foreach.leaf(ParIterableLike.scala:974) > at scala.collection.parallel.Task.$anonfun$tryLeaf$1(Tasks.scala:53) > at > scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23) > at scala.util.control.Breaks$$anon$1.catchBreak(Breaks.scala:67) > at scala.collection.parallel.Task.tryLeaf(Tasks.scala:56) > at scala.collection.parallel.Task.tryLeaf$(Tasks.scala:50) > at > scala.collection.parallel.ParIterableLike$Foreach.tryLeaf(ParIterableLike.scala:971) > at > scala.collection.parallel.AdaptiveWorkStealingTasks$WrappedTask.internal(Tasks.scala:160) > at > scala.collection.parallel.AdaptiveWorkStealingTasks$WrappedTask.internal$(Tasks.scala:157) > at > scala.collection.parallel.AdaptiveWorkStealingForkJoinTasks$WrappedTask.internal(Tasks.scala:440) > at > scala.collection.parallel.AdaptiveWorkStealingTasks$WrappedTask.compute(Tasks.scala:150) > at > scala.collection.parallel.AdaptiveWorkStealingTasks$WrappedTask.compute$(Tasks.scala:149) > at > scala.collection.parallel.AdaptiveWorkStealingForkJoinTasks$WrappedTask.compute(Tasks.scala:440) > at > java.base/java.util.concurrent.RecursiveAction.exec(RecursiveAction.java:189) > at > java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290) > at > java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020) > at > java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656) > at > java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594) > at > java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:177) > {code} -- This message was sent by Atlassian Jira (v8.3.4#803005) --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@spark.apache.org For additional commands, e-mail: issues-h...@spark.apache.org