[
https://issues.apache.org/jira/browse/BEAM-11485?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
Kenneth Knowles updated BEAM-11485:
-----------------------------------
Status: Open (was: Triage Needed)
> Spark test failure:
> org.apache.beam.sdk.transforms.CombineFnsTest.testComposedCombineNullValues
> -----------------------------------------------------------------------------------------------
>
> Key: BEAM-11485
> URL: https://issues.apache.org/jira/browse/BEAM-11485
> Project: Beam
> Issue Type: Sub-task
> Components: runner-spark, test-failures
> Reporter: Tyson Hamilton
> Priority: P1
> Labels: flake, portability-spark
>
> h1.
> From:
> [https://ci-beam.apache.org/job/beam_PostCommit_Java_PVR_Spark_Streaming/466/testReport/org.apache.beam.sdk.transforms/CombineFnsTest/testComposedCombineNullValues/]
>
> {code:java}
> Regression
> org.apache.beam.sdk.transforms.CombineFnsTest.testComposedCombineNullValues
> Failing for the past 1 build (Since #466 ) Took 41 sec. Error Message
> java.lang.AssertionError: Expected 1 successful assertions, but found 0.
> Expected: is <1L> but: was <0L>
> Stacktrace
> java.lang.AssertionError: Expected 1 successful assertions, but found 0.
> Expected: is <1L> but: was <0L> at
> org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18) at
> org.apache.beam.sdk.testing.TestPipeline.verifyPAssertsSucceeded(TestPipeline.java:516)
> at org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:354) at
> org.apache.beam.sdk.testing.TestPipeline.run(TestPipeline.java:334) at
> org.apache.beam.sdk.transforms.CombineFnsTest.testComposedCombineNullValues(CombineFnsTest.java:254)
> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498) at
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
> at
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
> at
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
> at
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
> at
> org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:266)
> at
> org.apache.beam.sdk.testing.TestPipeline$1.evaluate(TestPipeline.java:322) at
> org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:305) at
> org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100)
> at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:365) at
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103)
> at
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63)
> at org.junit.runners.ParentRunner$4.run(ParentRunner.java:330) at
> org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:78) at
> org.junit.runners.ParentRunner.runChildren(ParentRunner.java:328) at
> org.junit.runners.ParentRunner.access$100(ParentRunner.java:65) at
> org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:292) at
> org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:305) at
> org.junit.runners.ParentRunner.run(ParentRunner.java:412) at
> org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.runTestClass(JUnitTestClassExecutor.java:110)
> at
> org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:58)
> at
> org.gradle.api.internal.tasks.testing.junit.JUnitTestClassExecutor.execute(JUnitTestClassExecutor.java:38)
> at
> org.gradle.api.internal.tasks.testing.junit.AbstractJUnitTestClassProcessor.processTestClass(AbstractJUnitTestClassProcessor.java:62)
> at
> org.gradle.api.internal.tasks.testing.SuiteTestClassProcessor.processTestClass(SuiteTestClassProcessor.java:51)
> at sun.reflect.GeneratedMethodAccessor155.invoke(Unknown Source) at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498) at
> org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
> at
> org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
> at
> org.gradle.internal.dispatch.ContextClassLoaderDispatch.dispatch(ContextClassLoaderDispatch.java:33)
> at
> org.gradle.internal.dispatch.ProxyDispatchAdapter$DispatchingInvocationHandler.invoke(ProxyDispatchAdapter.java:94)
> at com.sun.proxy.$Proxy2.processTestClass(Unknown Source) at
> org.gradle.api.internal.tasks.testing.worker.TestWorker.processTestClass(TestWorker.java:119)
> at sun.reflect.GeneratedMethodAccessor154.invoke(Unknown Source) at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
> at java.lang.reflect.Method.invoke(Method.java:498) at
> org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:36)
> at
> org.gradle.internal.dispatch.ReflectionDispatch.dispatch(ReflectionDispatch.java:24)
> at
> org.gradle.internal.remote.internal.hub.MessageHubBackedObjectConnection$DispatchWrapper.dispatch(MessageHubBackedObjectConnection.java:182)
> at
> org.gradle.internal.remote.internal.hub.MessageHubBackedObjectConnection$DispatchWrapper.dispatch(MessageHubBackedObjectConnection.java:164)
> at
> org.gradle.internal.remote.internal.hub.MessageHub$Handler.run(MessageHub.java:414)
> at
> org.gradle.internal.concurrent.ExecutorPolicy$CatchAndRecordFailures.onExecute(ExecutorPolicy.java:64)
> at
> org.gradle.internal.concurrent.ManagedExecutorImpl$1.run(ManagedExecutorImpl.java:48)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at
> org.gradle.internal.concurrent.ThreadFactoryImpl$ManagedThreadRunnable.run(ThreadFactoryImpl.java:56)
> at java.lang.Thread.run(Thread.java:748)
> Standard Error
> 20/12/17 00:25:03 INFO org.apache.beam.runners.jobsubmission.JobServerDriver:
> ArtifactStagingService started on localhost:39215 20/12/17 00:25:03 INFO
> org.apache.beam.runners.jobsubmission.JobServerDriver: Java ExpansionService
> started on localhost:34547 20/12/17 00:25:03 INFO
> org.apache.beam.runners.jobsubmission.JobServerDriver: JobService started on
> localhost:36361 20/12/17 00:25:07 INFO
> org.apache.beam.runners.portability.PortableRunner: Using job server
> endpoint: localhost:36361 20/12/17 00:25:07 INFO
> org.apache.beam.runners.portability.PortableRunner: PrepareJobResponse:
> preparation_id:
> "combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_d818a0e3-681f-4c5a-9f67-915fa230821a"
> artifact_staging_endpoint { url: "localhost:39215" } staging_session_token:
> "combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_d818a0e3-681f-4c5a-9f67-915fa230821a"
> 20/12/17 00:25:07 INFO
> org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Staging
> artifacts for
> combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_d818a0e3-681f-4c5a-9f67-915fa230821a.
> 20/12/17 00:25:07 INFO
> org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService:
> Resolving artifacts for
> combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_d818a0e3-681f-4c5a-9f67-915fa230821a.EMBEDDED.
> 20/12/17 00:25:07 INFO
> org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService: Getting
> 313 artifacts for
> combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_d818a0e3-681f-4c5a-9f67-915fa230821a.null.
> 20/12/17 00:25:09 INFO
> org.apache.beam.runners.fnexecution.artifact.ArtifactStagingService:
> Artifacts fully staged for
> combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_d818a0e3-681f-4c5a-9f67-915fa230821a.
> Dec 17, 2020 12:25:09 AM
> org.apache.beam.vendor.grpc.v1p26p0.io.grpc.netty.NettyServerHandler
> onStreamError WARNING: Stream Error
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.http2.Http2Exception$StreamException:
> Received DATA frame for an unknown stream 3 at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.http2.Http2Exception.streamError(Http2Exception.java:147)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$FrameReadListener.shouldIgnoreHeadersOrDataFrame(DefaultHttp2ConnectionDecoder.java:591)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder$FrameReadListener.onDataRead(DefaultHttp2ConnectionDecoder.java:239)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.http2.Http2InboundFrameLogger$1.onDataRead(Http2InboundFrameLogger.java:48)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.http2.DefaultHttp2FrameReader.readDataFrame(DefaultHttp2FrameReader.java:422)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.http2.DefaultHttp2FrameReader.processPayloadState(DefaultHttp2FrameReader.java:251)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.http2.DefaultHttp2FrameReader.readFrame(DefaultHttp2FrameReader.java:160)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.http2.Http2InboundFrameLogger.readFrame(Http2InboundFrameLogger.java:41)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.http2.DefaultHttp2ConnectionDecoder.decodeFrame(DefaultHttp2ConnectionDecoder.java:174)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.http2.Http2ConnectionHandler$FrameDecoder.decode(Http2ConnectionHandler.java:378)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.http2.Http2ConnectionHandler.decode(Http2ConnectionHandler.java:438)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.ByteToMessageDecoder.decodeRemovalReentryProtection(ByteToMessageDecoder.java:505)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.ByteToMessageDecoder.callDecode(ByteToMessageDecoder.java:444)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:283)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1422)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:931)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:700)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:635)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:552)
> 20/12/17 00:25:09 INFO org.apache.beam.runners.spark.SparkJobInvoker:
> Invoking job
> combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_01ebc07f-41a4-4913-b308-f1eeb377e453
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:514)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.SingleThreadEventExecutor$6.run(SingleThreadEventExecutor.java:1044)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
> at
> org.apache.beam.vendor.grpc.v1p26p0.io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> at java.lang.Thread.run(Thread.java:748) 20/12/17 00:25:09 INFO
> org.apache.beam.runners.jobsubmission.JobInvocation: Starting job invocation
> combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_01ebc07f-41a4-4913-b308-f1eeb377e453
> 20/12/17 00:25:09 INFO org.apache.beam.runners.portability.PortableRunner:
> RunJobResponse: job_id:
> "combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_01ebc07f-41a4-4913-b308-f1eeb377e453"
> 20/12/17 00:25:09 INFO org.apache.beam.runners.spark.SparkPipelineRunner:
> Will stage 313 files. (Enable logging at DEBUG level to see which files will
> be staged.) 20/12/17 00:25:09 INFO
> org.apache.beam.runners.spark.translation.SparkContextFactory: Creating a
> brand new Spark Context. 20/12/17 00:25:09 INFO
> org.apache.beam.runners.spark.SparkPipelineRunner: Running job
> combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_01ebc07f-41a4-4913-b308-f1eeb377e453
> on Spark master local[4] 20/12/17 00:25:09 INFO
> org.apache.beam.runners.spark.SparkPipelineRunner: Running job
> combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_01ebc07f-41a4-4913-b308-f1eeb377e453
> on Spark master local[4] 20/12/17 00:25:10 INFO
> org.apache.beam.runners.spark.SparkPipelineRunner: Job
> combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_01ebc07f-41a4-4913-b308-f1eeb377e453:
> Pipeline translated successfully. Computing outputs 20/12/17 00:25:10 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:10 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:10 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:10 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:11 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:11 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:11 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:11 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:12 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:12 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:12 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:12 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:13 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:13 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:13 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:13 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:14 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:14 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:14 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:14 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:15 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:15 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:15 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:15 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:16 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:16 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:16 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:16 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:17 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:17 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:17 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:17 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:18 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:18 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:18 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:18 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:19 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:19 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:19 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:19 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:20 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:20 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:20 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:20 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:21 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:21 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:21 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:21 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:22 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:22 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:22 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:22 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:23 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:23 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:23 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:23 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:24 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:24 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:24 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:24 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:25 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:25 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:25 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:25 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:26 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:26 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:26 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:26 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:27 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:27 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:27 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:27 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:28 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:28 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:28 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:28 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:29 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:29 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:29 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:29 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:30 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:30 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:30 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:30 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:31 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:31 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:31 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:31 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:32 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:32 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:32 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:32 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:33 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:33 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:33 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:33 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:34 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:34 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:34 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:34 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:35 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:35 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:35 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:35 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:36 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:36 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:36 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:36 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:37 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:37 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:37 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:37 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:38 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:38 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:38 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:38 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:39 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:39 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:39 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:39 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:40 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:40 WARN
> org.apache.spark.streaming.dstream.QueueInputDStream: queueStream doesn't
> support checkpointing 20/12/17 00:25:40 WARN
> org.apache.spark.streaming.util.BatchedWriteAheadLog: BatchedWriteAheadLog
> Writer queue interrupted. Exception in thread "streaming-job-executor-0"
> java.lang.Error: java.lang.InterruptedException at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1155)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748) Caused by:
> java.lang.InterruptedException at
> java.util.concurrent.locks.AbstractQueuedSynchronizer.doAcquireSharedInterruptibly(AbstractQueuedSynchronizer.java:998)
> at
> java.util.concurrent.locks.AbstractQueuedSynchronizer.acquireSharedInterruptibly(AbstractQueuedSynchronizer.java:1304)
> at scala.concurrent.impl.Promise$DefaultPromise.tryAwait(Promise.scala:206)
> at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:222) at
> scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:157) at
> org.apache.spark.util.ThreadUtils$.awaitReady(ThreadUtils.scala:243) at
> org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:750) at
> org.apache.spark.SparkContext.runJob(SparkContext.scala:2061) at
> org.apache.spark.SparkContext.runJob(SparkContext.scala:2082) at
> org.apache.spark.SparkContext.runJob(SparkContext.scala:2101) at
> org.apache.spark.SparkContext.runJob(SparkContext.scala:2126) at
> org.apache.spark.rdd.RDD$$anonfun$foreach$1.apply(RDD.scala:972) at
> org.apache.spark.rdd.RDD$$anonfun$foreach$1.apply(RDD.scala:970) at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:151)
> at
> org.apache.spark.rdd.RDDOperationScope$.withScope(RDDOperationScope.scala:112)
> at org.apache.spark.rdd.RDD.withScope(RDD.scala:385) at
> org.apache.spark.rdd.RDD.foreach(RDD.scala:970) at
> org.apache.spark.api.java.JavaRDDLike$class.foreach(JavaRDDLike.scala:351) at
> org.apache.spark.api.java.AbstractJavaRDDLike.foreach(JavaRDDLike.scala:45)
> at
> org.apache.beam.runners.spark.translation.streaming.UnboundedDataset.lambda$action$e3b46054$1(UnboundedDataset.java:79)
> at
> org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$1.apply(JavaDStreamLike.scala:272)
> at
> org.apache.spark.streaming.api.java.JavaDStreamLike$$anonfun$foreachRDD$1.apply(JavaDStreamLike.scala:272)
> at
> org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:628)
> at
> org.apache.spark.streaming.dstream.DStream$$anonfun$foreachRDD$1$$anonfun$apply$mcV$sp$3.apply(DStream.scala:628)
> at
> org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(ForEachDStream.scala:51)
> at
> org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:51)
> at
> org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1$$anonfun$apply$mcV$sp$1.apply(ForEachDStream.scala:51)
> at
> org.apache.spark.streaming.dstream.DStream.createRDDWithLocalProperties(DStream.scala:416)
> at
> org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply$mcV$sp(ForEachDStream.scala:50)
> at
> org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:50)
> at
> org.apache.spark.streaming.dstream.ForEachDStream$$anonfun$1.apply(ForEachDStream.scala:50)
> at scala.util.Try$.apply(Try.scala:192) at
> org.apache.spark.streaming.scheduler.Job.run(Job.scala:39) at
> org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply$mcV$sp(JobScheduler.scala:257)
> at
> org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:257)
> at
> org.apache.spark.streaming.scheduler.JobScheduler$JobHandler$$anonfun$run$1.apply(JobScheduler.scala:257)
> at scala.util.DynamicVariable.withValue(DynamicVariable.scala:58) at
> org.apache.spark.streaming.scheduler.JobScheduler$JobHandler.run(JobScheduler.scala:256)
> at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> ... 2 more 20/12/17 00:25:42 ERROR org.apache.spark.util.Utils: Aborting
> task java.io.IOException: Failed to connect to localhost/127.0.0.1:33183 at
> org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:245)
> at
> org.apache.spark.network.client.TransportClientFactory.createClient(TransportClientFactory.java:187)
> at
> org.apache.spark.rpc.netty.NettyRpcEnv.org$apache$spark$rpc$netty$NettyRpcEnv$$downloadClient(NettyRpcEnv.scala:368)
> at
> org.apache.spark.rpc.netty.NettyRpcEnv$$anonfun$openChannel$1.apply$mcV$sp(NettyRpcEnv.scala:336)
> at
> org.apache.spark.rpc.netty.NettyRpcEnv$$anonfun$openChannel$1.apply(NettyRpcEnv.scala:335)
> at
> org.apache.spark.rpc.netty.NettyRpcEnv$$anonfun$openChannel$1.apply(NettyRpcEnv.scala:335)
> at
> org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1394)
> at org.apache.spark.rpc.netty.NettyRpcEnv.openChannel(NettyRpcEnv.scala:339)
> at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:693) at
> org.apache.spark.util.Utils$.fetchFile(Utils.scala:509) at
> org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:816)
> at
> org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:808)
> at
> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
> at
> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
> at
> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
> at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:236)
> at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40) at
> scala.collection.mutable.HashMap.foreach(HashMap.scala:130) at
> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
> at
> org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:808)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:375) at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748) Caused by:
> io.netty.channel.AbstractChannel$AnnotatedConnectException: Connection
> refused: localhost/127.0.0.1:33183 Caused by: java.net.ConnectException:
> Connection refused at sun.nio.ch.SocketChannelImpl.checkConnect(Native
> Method) at
> sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:716) at
> io.netty.channel.socket.nio.NioSocketChannel.doFinishConnect(NioSocketChannel.java:330)
> at
> io.netty.channel.nio.AbstractNioChannel$AbstractNioUnsafe.finishConnect(AbstractNioChannel.java:334)
> at
> io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:702)
> at
> io.netty.channel.nio.NioEventLoop.processSelectedKeysOptimized(NioEventLoop.java:650)
> at
> io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:576)
> at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:493) at
> io.netty.util.concurrent.SingleThreadEventExecutor$4.run(SingleThreadEventExecutor.java:989)
> at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)
> at
> io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
> at java.lang.Thread.run(Thread.java:748) 20/12/17 00:25:42 WARN
> org.apache.spark.util.Utils: Suppressing exception in catch: null
> java.lang.NullPointerException at
> org.apache.spark.util.Utils$.tryWithSafeFinallyAndFailureCallbacks(Utils.scala:1402)
> at org.apache.spark.rpc.netty.NettyRpcEnv.openChannel(NettyRpcEnv.scala:339)
> at org.apache.spark.util.Utils$.doFetchFile(Utils.scala:693) at
> org.apache.spark.util.Utils$.fetchFile(Utils.scala:509) at
> org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:816)
> at
> org.apache.spark.executor.Executor$$anonfun$org$apache$spark$executor$Executor$$updateDependencies$5.apply(Executor.scala:808)
> at
> scala.collection.TraversableLike$WithFilter$$anonfun$foreach$1.apply(TraversableLike.scala:733)
> at
> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
> at
> scala.collection.mutable.HashMap$$anonfun$foreach$1.apply(HashMap.scala:130)
> at scala.collection.mutable.HashTable$class.foreachEntry(HashTable.scala:236)
> at scala.collection.mutable.HashMap.foreachEntry(HashMap.scala:40) at
> scala.collection.mutable.HashMap.foreach(HashMap.scala:130) at
> scala.collection.TraversableLike$WithFilter.foreach(TraversableLike.scala:732)
> at
> org.apache.spark.executor.Executor.org$apache$spark$executor$Executor$$updateDependencies(Executor.scala:808)
> at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:375) at
> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
> at
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
> at java.lang.Thread.run(Thread.java:748) 20/12/17 00:25:42 ERROR
> org.apache.spark.executor.Executor: Exception in task 0.0 in stage 0.0 (TID
> 0): Failed to connect to localhost/127.0.0.1:33183 20/12/17 00:25:42 INFO
> org.apache.beam.runners.spark.SparkPipelineRunner: Job
> combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_01ebc07f-41a4-4913-b308-f1eeb377e453
> finished. 20/12/17 00:25:42 WARN
> org.apache.spark.streaming.StreamingContext: StreamingContext has already
> been stopped 20/12/17 00:25:43 ERROR org.apache.spark.executor.Executor:
> Exception in task 3.0 in stage 0.0 (TID 3): null 20/12/17 00:25:42 ERROR
> org.apache.spark.executor.Executor: Exception in task 2.0 in stage 0.0 (TID
> 2): Cannot retrieve files with 'spark' scheme without an active SparkEnv.
> 20/12/17 00:25:43 ERROR org.apache.spark.executor.Executor: Exception in task
> 1.0 in stage 0.0 (TID 1): null 20/12/17 00:25:45 INFO
> org.apache.beam.runners.jobsubmission.InMemoryJobService: Getting job metrics
> for
> combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_01ebc07f-41a4-4913-b308-f1eeb377e453
> 20/12/17 00:25:45 INFO
> org.apache.beam.runners.jobsubmission.InMemoryJobService: Finished getting
> job metrics for
> combinefnstest0testcomposedcombinenullvalues-jenkins-1217002507-53b86688_01ebc07f-41a4-4913-b308-f1eeb377e453
> 20/12/17 00:25:45 INFO
> org.apache.beam.runners.jobsubmission.JobServerDriver: JobServer stopped on
> localhost:36361 20/12/17 00:25:45 INFO
> org.apache.beam.runners.jobsubmission.JobServerDriver: ArtifactStagingServer
> stopped on localhost:39215 20/12/17 00:25:45 INFO
> org.apache.beam.runners.jobsubmission.JobServerDriver: Expansion stopped on
> localhost:34547
> {code}
>
--
This message was sent by Atlassian Jira
(v8.3.4#803005)