[
https://issues.apache.org/jira/browse/SPARK-35838?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Dongjoon Hyun resolved SPARK-35838.
-----------------------------------
Fix Version/s: 3.2.0
Resolution: Fixed
Issue resolved by pull request 32994
[https://github.com/apache/spark/pull/32994]
> Ensure kafka-0-10-sql module can maven test independently in Scala 2.13
> -----------------------------------------------------------------------
>
> Key: SPARK-35838
> URL: https://issues.apache.org/jira/browse/SPARK-35838
> Project: Spark
> Issue Type: Sub-task
> Components: Build
> Affects Versions: 3.2.0
> Reporter: Yang Jie
> Assignee: Yang Jie
> Priority: Minor
> Fix For: 3.2.0
>
>
>
> Execute
>
> {code:java}
> mvn clean install -Phadoop-3.2 -Phive-2.3 -Phadoop-cloud -Pmesos -Pyarn
> -Pkinesis-asl -Phive-thriftserver -Pspark-ganglia-lgpl -Pkubernetes -Phive
> -Pscala-2.13 -pl external/kafka-0-10-sql
> {code}
>
> 1 scala test aborted, the error message is
> {code:java}
> Discovery starting.
> Discovery completed in 857 milliseconds.
> Run starting. Expected test count is: 464
> ...
> KafkaRelationSuiteV2:
> - explicit earliest to latest offsets
> - default starting and ending offsets
> - explicit offsets
> - default starting and ending offsets with headers
> - timestamp provided for starting and ending
> - timestamp provided for starting, offset provided for ending
> - timestamp provided for ending, offset provided for starting
> - timestamp provided for starting, ending not provided
> - timestamp provided for ending, starting not provided
> - global timestamp provided for starting and ending
> - no matched offset for timestamp - startingOffsets
> - preferences on offset related options
> - no matched offset for timestamp - endingOffsets
> *** RUN ABORTED ***
> java.lang.NoClassDefFoundError: scala/collection/parallel/TaskSupport
> at org.apache.spark.SparkContext.$anonfun$union$1(SparkContext.scala:1411)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
> at org.apache.spark.SparkContext.withScope(SparkContext.scala:788)
> at org.apache.spark.SparkContext.union(SparkContext.scala:1405)
> at
> org.apache.spark.sql.execution.UnionExec.doExecute(basicPhysicalOperators.scala:697)
> at
> org.apache.spark.sql.execution.SparkPlan.$anonfun$execute$1(SparkPlan.scala:182)
> at
> org.apache.spark.sql.execution.SparkPlan.$anonfun$executeQuery$1(SparkPlan.scala:220)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
> at
> org.apache.spark.sql.execution.SparkPlan.executeQuery(SparkPlan.scala:217)
> ...
> Cause: java.lang.ClassNotFoundException:
> scala.collection.parallel.TaskSupport
> at java.net.URLClassLoader.findClass(URLClassLoader.java:382)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:418)
> at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352)
> at java.lang.ClassLoader.loadClass(ClassLoader.java:351)
> at org.apache.spark.SparkContext.$anonfun$union$1(SparkContext.scala:1411)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
> at org.apache.spark.SparkContext.withScope(SparkContext.scala:788)
> at org.apache.spark.SparkContext.union(SparkContext.scala:1405)
> at
> org.apache.spark.sql.execution.UnionExec.doExecute(basicPhysicalOperators.scala:697)
> ...
> {code}
>
--
This message was sent by Atlassian Jira
(v8.3.4#803005)
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]