See
<https://ci-beam.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/2171/display/redirect>
Changes:
------------------------------------------
[...truncated 56.81 KB...]
org.apache.beam.sdk.transforms.FlattenTest > testFlattenPCollectionsEmpty FAILED
java.lang.RuntimeException at FlattenTest.java:138
org.apache.beam.sdk.transforms.FlattenTest >
testFlattenPCollectionsSingletonList FAILED
java.lang.RuntimeException at FlattenTest.java:112
org.apache.beam.sdk.transforms.FlattenTest > testFlattenPCollections FAILED
java.lang.RuntimeException at FlattenTest.java:100
org.apache.beam.sdk.transforms.FlattenTest > testFlattenPCollectionsThenParDo
FAILED
java.lang.RuntimeException at FlattenTest.java:126
org.apache.beam.sdk.transforms.FlattenTest >
testFlattenMultiplePCollectionsHavingMultipleConsumers FAILED
java.lang.RuntimeException at FlattenTest.java:355
org.apache.beam.sdk.transforms.FlattenTest >
testFlattenWithDifferentInputAndOutputCoders FAILED
java.lang.RuntimeException at FlattenTest.java:377
org.apache.beam.sdk.transforms.FlattenTest > testFlattenIterables FAILED
java.lang.RuntimeException at FlattenTest.java:260
org.apache.beam.sdk.transforms.FlattenTest > testFlattenIterablesSets FAILED
java.lang.RuntimeException at FlattenTest.java:288
org.apache.beam.sdk.transforms.FlattenTest >
testFlattenPCollectionsEmptyThenParDo FAILED
java.lang.RuntimeException at FlattenTest.java:232
org.apache.beam.sdk.transforms.KvSwapTest > testKvSwap FAILED
java.lang.RuntimeException at KvSwapTest.java:80
org.apache.beam.sdk.transforms.ParDoTest$BasicTests > testParDo FAILED
java.lang.RuntimeException at ParDoTest.java:404
org.apache.beam.sdk.transforms.ParDoTest$BasicTests >
testPipelineOptionsParameter FAILED
java.lang.RuntimeException at ParDoTest.java:646
org.apache.beam.sdk.transforms.ParDoTest$BasicTests > testParDoEmpty FAILED
java.lang.RuntimeException at ParDoTest.java:420
org.apache.beam.sdk.transforms.ParDoTest$BasicTests >
testParDoInCustomTransform FAILED
java.lang.RuntimeException at ParDoTest.java:466
org.apache.beam.sdk.transforms.ParDoTest$BasicTests > testParDoEmptyOutputs
FAILED
java.lang.RuntimeException at ParDoTest.java:436
org.apache.beam.sdk.transforms.ParDoTest$LifecycleTests >
testWindowingInStartAndFinishBundle FAILED
java.lang.RuntimeException at ParDoTest.java:1800
org.apache.beam.sdk.transforms.ParDoTest$LifecycleTests >
testParDoWithErrorInStartBatch FAILED
java.lang.AssertionError at MatcherAssert.java:18
org.apache.beam.sdk.transforms.ParDoTest$LifecycleTests >
testParDoWithErrorInProcessElement FAILED
java.lang.AssertionError at MatcherAssert.java:18
org.apache.beam.sdk.transforms.ParDoTest$LifecycleTests >
testParDoWithErrorInFinishBatch FAILED
java.lang.AssertionError at MatcherAssert.java:18
org.apache.beam.sdk.transforms.ParDoTest$StateCoderInferenceTests >
testCombiningStateCoderInference FAILED
java.lang.RuntimeException at ParDoTest.java:4326
org.apache.beam.sdk.transforms.ParDoTest$StateCoderInferenceTests >
testBagStateCoderInference FAILED
java.lang.RuntimeException at ParDoTest.java:3915
org.apache.beam.sdk.transforms.ParDoTest$TimerTests >
testOutputTimestampDefaultBounded FAILED
java.lang.RuntimeException at ParDoTest.java:4862
org.apache.beam.sdk.transforms.ParDoTest$TimerTests >
testEventTimeTimerAbsolute FAILED
java.lang.RuntimeException at ParDoTest.java:4643
org.apache.beam.sdk.transforms.ParDoTest$TimerTests >
testNoOutputTimestampDefaultBounded FAILED
java.lang.RuntimeException at ParDoTest.java:4912
org.apache.beam.sdk.transforms.ParDoTest$TimerTests >
testTimerReceivedInOriginalWindow FAILED
java.lang.RuntimeException at ParDoTest.java:4609
org.apache.beam.sdk.transforms.ParDoTest$TimerTests > testEventTimeTimerBounded
FAILED
java.lang.RuntimeException at ParDoTest.java:4490
org.apache.beam.sdk.transforms.ParDoTest$TimerTests >
testOutOfBoundsProcessingTimeTimerHold FAILED
java.lang.RuntimeException at ParDoTest.java:4997
org.apache.beam.sdk.transforms.ParDoTest$TimerTests >
testEventTimeTimerAlignBounded FAILED
java.lang.RuntimeException at ParDoTest.java:4564
org.apache.beam.sdk.transforms.ParDoTest$TimerTests >
testNoOutputTimestampDefaultStreaming FAILED
java.lang.RuntimeException at ParDoTest.java:4912
org.apache.beam.sdk.transforms.ParDoTest$TimerTests >
testGbkFollowedByUserTimers FAILED
java.lang.RuntimeException at ParDoTest.java:4529
org.apache.beam.sdk.transforms.ParDoTest$TimerTests >
testOutOfBoundsEventTimeTimerHold FAILED
java.lang.RuntimeException at ParDoTest.java:4954
org.apache.beam.sdk.transforms.ParDoTest$TimerTests >
testOutOfBoundsEventTimeTimer FAILED
java.lang.RuntimeException at ParDoTest.java:4818
org.apache.beam.sdk.transforms.ParDoTest$TimerTests >
testEventTimeTimerMultipleKeys FAILED
java.lang.RuntimeException at ParDoTest.java:4778
org.apache.beam.sdk.transforms.ParDoTest$TimerTests >
testPipelineOptionsParameterOnTimer FAILED
java.lang.RuntimeException at ParDoTest.java:5491
org.apache.beam.sdk.transforms.ReifyTimestampsTest >
extractFromValuesWhenValueTimestampedLaterSucceeds FAILED
java.lang.RuntimeException at ReifyTimestampsTest.java:128
org.apache.beam.sdk.transforms.ReifyTimestampsTest > extractFromValuesSucceeds
FAILED
java.lang.RuntimeException at ReifyTimestampsTest.java:92
org.apache.beam.sdk.transforms.ReifyTimestampsTest > inValuesSucceeds FAILED
java.lang.RuntimeException at ReifyTimestampsTest.java:60
org.apache.beam.sdk.transforms.join.CoGroupByKeyTest > testCoGroupByKey FAILED
java.lang.RuntimeException at CoGroupByKeyTest.java:296
org.apache.beam.sdk.transforms.join.CoGroupByKeyTest > testCoGroupByKeyGetOnly
FAILED
java.lang.RuntimeException at CoGroupByKeyTest.java:139
org.apache.beam.sdk.transforms.join.CoGroupByKeyTest >
testCoGroupByKeyHandleResults FAILED
java.lang.RuntimeException at CoGroupByKeyTest.java:466
org.apache.beam.sdk.transforms.join.CoGroupByKeyTest >
testCoGroupByKeyWithWindowing FAILED
java.lang.RuntimeException at CoGroupByKeyTest.java:494
org.apache.beam.sdk.values.PCollectionRowTupleTest >
testComposePCollectionRowTuple FAILED
java.lang.RuntimeException at PCollectionRowTupleTest.java:104
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Invoking stop() from
shutdown hook
[DefaultSparkRunner-thread] INFO org.apache.spark.scheduler.DAGScheduler - Job
0 failed: foreach at BoundedDataset.java:127, took 1432.858501 s
[shutdown-hook-0] INFO org.apache.spark.scheduler.DAGScheduler -
ShuffleMapStage 0 (mapToPair at GroupNonMergingWindowsFunctions.java:273)
failed in 1432.855 s due to Stage cancelled because SparkContext was shut down
[spark-runner-job-invoker-0] INFO org.apache.spark.SparkContext - SparkContext
already stopped.
[spark-runner-job-invoker-1] ERROR
org.apache.beam.runners.jobsubmission.JobInvocation - Error during job
invocation
flattentest0testflatteninputmultiplecopies-jenkins-0623222616-f3ae127e_c66cfdf4-7490-4314-861d-3f93011d43a3.
[dispatcher-event-loop-3] INFO org.apache.spark.MapOutputTrackerMasterEndpoint
- MapOutputTrackerMasterEndpoint stopped!
java.lang.RuntimeException: org.apache.spark.SparkException: Job 0 cancelled
because SparkContext was shut down
at
org.apache.beam.runners.spark.SparkPipelineResult.runtimeExceptionFrom(SparkPipelineResult.java:60)
at
org.apache.beam.runners.spark.SparkPipelineResult.beamExceptionFrom(SparkPipelineResult.java:77)
at
org.apache.beam.runners.spark.SparkPipelineResult.waitUntilFinish(SparkPipelineResult.java:104)
at
org.apache.beam.runners.spark.SparkPipelineResult.waitUntilFinish(SparkPipelineResult.java:92)
at
org.apache.beam.runners.spark.SparkPipelineRunner.run(SparkPipelineRunner.java:187)
at
org.apache.beam.runners.jobsubmission.JobInvocation.runPipeline(JobInvocation.java:86)
at
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask$TrustedFutureInterruptibleTask.runInterruptibly(TrustedListenableFutureTask.java:125)
at
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.InterruptibleTask.run(InterruptibleTask.java:57)
at
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.TrustedListenableFutureTask.run(TrustedListenableFutureTask.java:78)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
Caused by: org.apache.spark.SparkException: Job 0 cancelled because
SparkContext was shut down
at
org.apache.spark.scheduler.DAGScheduler.$anonfun$cleanUpAfterSchedulerStop$1(DAGScheduler.scala:1085)
at
org.apache.spark.scheduler.DAGScheduler.$anonfun$cleanUpAfterSchedulerStop$1$adapted(DAGScheduler.scala:1083)
at scala.collection.mutable.HashSet.foreach(HashSet.scala:79)
at
org.apache.spark.scheduler.DAGScheduler.cleanUpAfterSchedulerStop(DAGScheduler.scala:1083)
at
org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onStop(DAGScheduler.scala:2463)
at org.apache.spark.util.EventLoop.stop(EventLoop.scala:84)
at org.apache.spark.scheduler.DAGScheduler.stop(DAGScheduler.scala:2369)
at
org.apache.spark.SparkContext.$anonfun$stop$12(SparkContext.scala:2069)
at org.apache.spark.util.Utils$.tryLogNonFatalError(Utils.scala:1419)
at org.apache.spark.SparkContext.stop(SparkContext.scala:2069)
at org.apache.spark.SparkContext.$anonfun$new$37(SparkContext.scala:661)
at
org.apache.spark.util.SparkShutdownHook.run(ShutdownHookManager.scala:214)
at
org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$2(ShutdownHookManager.scala:188)
at
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at org.apache.spark.util.Utils$.logUncaughtExceptions(Utils.scala:1996)
at
org.apache.spark.util.SparkShutdownHookManager.$anonfun$runAll$1(ShutdownHookManager.scala:188)
at
scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at scala.util.Try$.apply(Try.scala:213)
at
org.apache.spark.util.SparkShutdownHookManager.runAll(ShutdownHookManager.scala:188)
at
org.apache.spark.util.SparkShutdownHookManager$$anon$2.run(ShutdownHookManager.scala:178)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
at
org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:868)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2196)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2217)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2236)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:2261)
at org.apache.spark.rdd.RDD.$anonfun$foreach$1(RDD.scala:1012)
at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
at
org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
at org.apache.spark.rdd.RDD.withScope(RDD.scala:414)
at org.apache.spark.rdd.RDD.foreach(RDD.scala:1010)
at org.apache.spark.api.java.JavaRDDLike.foreach(JavaRDDLike.scala:352)
at org.apache.spark.api.java.JavaRDDLike.foreach$(JavaRDDLike.scala:351)
at
org.apache.spark.api.java.AbstractJavaRDDLike.foreach(JavaRDDLike.scala:45)
at
org.apache.beam.runners.spark.translation.BoundedDataset.action(BoundedDataset.java:127)
at
org.apache.beam.runners.spark.translation.SparkTranslationContext.computeOutputs(SparkTranslationContext.java:90)
at
org.apache.beam.runners.spark.SparkPipelineRunner.lambda$run$2(SparkPipelineRunner.java:181)
at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
... 3 more
[spark-runner-job-invoker-1] WARN
org.apache.beam.runners.jobsubmission.InMemoryJobService - Failed to remove job
staging directory for token
flattentest0testflatteninputmultiplecopies-jenkins-0623222616-f3ae127e_725773c5-a2bf-4bc1-8f6b-df8243dcd428.
java.lang.NullPointerException
at
org.apache.beam.runners.jobsubmission.JobServerDriver.lambda$createJobService$0(JobServerDriver.java:67)
at
org.apache.beam.runners.jobsubmission.InMemoryJobService.lambda$run$0(InMemoryJobService.java:261)
at
org.apache.beam.runners.jobsubmission.JobInvocation.setState(JobInvocation.java:249)
at
org.apache.beam.runners.jobsubmission.JobInvocation.access$200(JobInvocation.java:51)
at
org.apache.beam.runners.jobsubmission.JobInvocation$1.onFailure(JobInvocation.java:153)
at
org.apache.beam.vendor.guava.v26_0_jre.com.google.common.util.concurrent.Futures$CallbackListener.run(Futures.java:1052)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:750)
[shutdown-hook-0] INFO org.apache.spark.storage.memory.MemoryStore -
MemoryStore cleared
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManager - BlockManager
stopped
[shutdown-hook-0] INFO org.apache.spark.storage.BlockManagerMaster -
BlockManagerMaster stopped
[dispatcher-event-loop-3] INFO
org.apache.spark.scheduler.OutputCommitCoordinator$OutputCommitCoordinatorEndpoint
- OutputCommitCoordinator stopped!
[shutdown-hook-0] INFO org.apache.spark.SparkContext - Successfully stopped
SparkContext
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown
hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting
directory /tmp/spark-250cb028-45c5-4820-b182-5e2a6f148779
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown
hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting
directory /tmp/spark-38531552-085b-4054-b72d-3621a71476c1
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown
hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting
directory /tmp/spark-c9798649-a215-4a43-bf84-b1843ae99e3a
234 tests completed, 43 failed, 1 skipped
> Task :runners:spark:3:job-server:validatesPortableRunnerBatch FAILED
> Task :runners:spark:3:job-server:validatesPortableRunnerDocker
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Shutdown
hook called
[shutdown-hook-0] INFO org.apache.spark.util.ShutdownHookManager - Deleting
directory /tmp/spark-be700623-a738-4a1f-bbb0-68cb3e0f86ad
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task
':runners:spark:3:job-server:validatesPortableRunnerBatch'.
> There were failing tests. See the report at:
> file://<https://ci-beam.apache.org/job/beam_PostCommit_Java_PVR_Spark_Batch/ws/src/runners/spark/3/job-server/build/reports/tests/validatesPortableRunnerBatch/index.html>
* Try:
> Run with --stacktrace option to get the stack trace.
> Run with --info or --debug option to get more log output.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with
Gradle 8.0.
You can use '--warning-mode all' to show the individual deprecation warnings
and determine if they come from your own scripts or plugins.
See
https://docs.gradle.org/7.5.1/userguide/command_line_interface.html#sec:command_line_warnings
Execution optimizations have been disabled for 1 invalid unit(s) of work during
this build to ensure correctness.
Please consult deprecation warnings for more details.
BUILD FAILED in 2h 4m
118 actionable tasks: 78 executed, 36 from cache, 4 up-to-date
Publishing build scan...
https://ge.apache.org/s/4biuhyqi6bgfk
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]