damccorm opened a new issue, #20293:
URL: https://github.com/apache/beam/issues/20293

   Gradle task *_:runners:spark:test_* fails during Java 11 Precommit job
   
    
   
   Example stack trace:
   ```
   
   > Task :runners:spark:test
       20/05/26 07:26:31 INFO 
org.apache.beam.runners.spark.metrics.MetricsAccumulator:
   Instantiated metrics accumulator: {
         "metrics": {
         }
       }
   
   
org.apache.beam.runners.spark.structuredstreaming.StructuredStreamingPipelineStateTest
   > testBatchPipelineRunningState STANDARD_ERROR
       20/05/26 07:26:32 INFO 
org.apache.beam.runners.spark.structuredstreaming.SparkStructuredStreamingRunner:
   *** SparkStructuredStreamingRunner is based on spark structured streaming 
framework and is no more 
   
       based on RDD/DStream API. See
        
https://spark.apache.org/docs/latest/structured-streaming-programming-guide.html
   
       It is still experimental, its coverage of the Beam model is partial. ***
   
   org.apache.beam.runners.spark.SparkPortableExecutionTest
   > testExecution STANDARD_ERROR
       20/05/26 07:26:33 WARN 
org.apache.beam.runners.spark.translation.GroupNonMergingWindowsFunctions:
   Either coder LengthPrefixCoder(ByteArrayCoder) or GlobalWindow$Coder is not 
consistent with equals.
   That might cause issues on some runners.
   
   
org.apache.beam.runners.spark.structuredstreaming.translation.batch.FlattenTest
   > testFlatten STANDARD_ERROR
       20/05/26 07:26:34 WARN org.apache.hadoop.util.NativeCodeLoader: Unable
   to load native-hadoop library for your platform... using builtin-java 
classes where applicable
   
   org.apache.beam.runners.spark.SparkPortableExecutionTest
   > testExecution STANDARD_ERROR
       20/05/26 07:26:34 ERROR 
org.apache.beam.runners.jobsubmission.JobInvocation:
   Error during job invocation fakeId.
       java.lang.IllegalArgumentException: Unsupported class file
   major version 55
        at org.apache.xbean.asm6.ClassReader.<init>(ClassReader.java:166)
        at org.apache.xbean.asm6.ClassReader.<init>(ClassReader.java:148)
   
        at org.apache.xbean.asm6.ClassReader.<init>(ClassReader.java:136)
        at org.apache.xbean.asm6.ClassReader.<init>(ClassReader.java:237)
   
        at 
org.apache.spark.util.ClosureCleaner$.getClassReader(ClosureCleaner.scala:49)
        at 
org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:517)
   
        at 
org.apache.spark.util.FieldAccessFinder$$anon$3$$anonfun$visitMethodInsn$2.apply(ClosureCleaner.scala:500)
   
        at 
scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
   
        at 
scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:134)
        at
   
scala.collection.mutable.HashMap$$anon$1$$anonfun$foreach$2.apply(HashMap.scala:134)
        at 
scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:236)
   
        at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40)
        at scala.collection.mutable.HashMap$$anon$1.foreach(HashMap.scala:134)
   
        at 
scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
        at 
org.apache.spark.util.FieldAccessFinder$$anon$3.visitMethodInsn(ClosureCleaner.scala:500)
   
        at org.apache.xbean.asm6.ClassReader.readCode(ClassReader.java:2175)
        at org.apache.xbean.asm6.ClassReader.readMethod(ClassReader.java:1238)
   
        at org.apache.xbean.asm6.ClassReader.accept(ClassReader.java:631)
        at org.apache.xbean.asm6.ClassReader.accept(ClassReader.java:355)
   
        at 
org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:307)
   
        at 
org.apache.spark.util.ClosureCleaner$$anonfun$org$apache$spark$util$ClosureCleaner$$clean$14.apply(ClosureCleaner.scala:306)
   
        at scala.collection.immutable.List.foreach(List.scala:392)
        at 
org.apache.spark.util.ClosureCleaner$.org$apache$spark$util$ClosureCleaner$$clean(ClosureCleaner.scala:306)
   
        at org.apache.spark.util.ClosureCleaner$.clean(ClosureCleaner.scala:162)
        at org.apache.spark.SparkContext.clean(SparkContext.scala:2326)
   
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2100)
        at org.apache.spark.SparkContext.runJob(SparkContext.scala:2126)
   
        at org.apache.spark.rdd.RDD$$anonfun$collect$1.apply(RDD.scala:990)
        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
   
        at 
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
        at org.apache.spark.rdd.RDD.withScope(RDD.scala:385)
   
        at org.apache.spark.rdd.RDD.collect(RDD.scala:989)
        at 
org.apache.spark.api.java.JavaRDDLike$class.collect(JavaRDDLike.scala:361)
   
        at 
org.apache.spark.api.java.AbstractJavaRDDLike.collect(JavaRDDLike.scala:45)
        at 
org.apache.beam.runners.spark.translation.BoundedDataset.getBytes(BoundedDataset.java:76)
   
        at 
org.apache.beam.runners.spark.translation.SparkBatchPortablePipelineTranslator.broadcastSideInput(SparkBatchPortablePipelineTranslator.java:354)
   
        at 
org.apache.beam.runners.spark.translation.SparkBatchPortablePipelineTranslator.broadcastSideInputs(SparkBatchPortablePipelineTranslator.java:338)
   
        at 
org.apache.beam.runners.spark.translation.SparkBatchPortablePipelineTranslator.translateExecutableStage(SparkBatchPortablePipelineTranslator.java:216)
   
        at 
org.apache.beam.runners.spark.translation.SparkBatchPortablePipelineTranslator.translate(SparkBatchPortablePipelineTranslator.java:138)
   
        at 
org.apache.beam.runners.spark.SparkPipelineRunner.lambda$run$1(SparkPipelineRunner.java:122)
   
        at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
        at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
   
        at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
    
        at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
     
        at java.base/java.lang.Thread.run(Thread.java:834)
   
   org.apache.beam.runners.spark.SparkPortableExecutionTest
   > testExecution FAILED
       java.lang.AssertionError: expected:<DONE> but was:<FAILED>
           at org.junit.Assert.fail(Assert.java:89)
   
          at org.junit.Assert.failNotEquals(Assert.java:835)
           at org.junit.Assert.assertEquals(Assert.java:120)
   
          at org.junit.Assert.assertEquals(Assert.java:146)
           at 
org.apache.beam.runners.spark.SparkPortableExecutionTest.testExecution(SparkPortableExecutionTest.java:159)
   
   org.apache.beam.runners.spark.SparkPortableExecutionTest
   > testExecStageWithMultipleOutputs STANDARD_ERROR
       20/05/26 07:26:35 INFO 
org.apache.beam.runners.jobsubmission.JobInvocation:
   Starting job invocation testExecStageWithMultipleOutputs
       20/05/26 07:26:36 INFO org.apache.beam.runners.spark.SparkPipelineRunner:
   PipelineOptions.filesToStage was not specified. Defaulting to files from the 
classpath
       20/05/26
   07:26:36 INFO org.apache.beam.runners.spark.SparkPipelineRunner: Will stage 
289 files. (Enable logging
   at DEBUG level to see which files will be staged.)
       20/05/26 07:26:36 INFO org.apache.beam.runners.spark.SparkPipelineRunner:
   Running job testExecStageWithMultipleOutputs on Spark master local[4]
       20/05/26 07:26:37 INFO org.apache.beam.runners.spark.SparkPipelineRunner:
   Job testExecStageWithMultipleOutputs: Pipeline translated successfully. 
Computing outputs
   
   Gradle
   Test Executor 114 started executing tests.
   
   ```
   
    
   
   Imported from Jira 
[BEAM-10083](https://issues.apache.org/jira/browse/BEAM-10083). Original Jira 
may contain additional context.
   Reported by: pawel.pasterz.
   Subtask of issue #20290


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to