See
<https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/2716/display/redirect?page=changes>
Changes:
[12602502+Ardagan] [BEAM-9431] Remove ReadFromPubSub/Read-out0-ElementCount
from the
------------------------------------------
[...truncated 773.02 KB...]
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Mar 11, 2020 4:05:24 PM
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor run
SEVERE: Exception while executing runnable
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@6e6bec9a
java.lang.IllegalStateException: call already closed
at
org.apache.beam.vendor.grpc.v1p26p0.com.google.common.base.Preconditions.checkState(Preconditions.java:511)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl.closeInternal(ServerCallImpl.java:209)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl.close(ServerCallImpl.java:202)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.stub.ServerCalls$ServerCallStreamObserverImpl.onError(ServerCalls.java:360)
at
org.apache.beam.runners.fnexecution.artifact.AbstractArtifactStagingService$PutArtifactStreamObserver.onCompleted(AbstractArtifactStagingService.java:226)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.stub.ServerCalls$StreamingServerCallHandler$StreamingServerCallListener.onHalfClose(ServerCalls.java:262)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.PartialForwardingServerCallListener.onHalfClose(PartialForwardingServerCallListener.java:35)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.ForwardingServerCallListener.onHalfClose(ForwardingServerCallListener.java:23)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.ForwardingServerCallListener$SimpleForwardingServerCallListener.onHalfClose(ForwardingServerCallListener.java:40)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.Contexts$ContextualizedServerCallListener.onHalfClose(Contexts.java:86)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:331)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:817)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
[grpc-default-executor-1] ERROR
org.apache.beam.runners.fnexecution.artifact.AbstractArtifactStagingService -
Failed to write chunk of artifact worker to
/tmp/beam-artifact-staging/go-job-7-1583942724100167279_2340a45b-e9c0-408e-ae7a-2265141dcbd6/artifacts/artifact_87eba76e7f3164534045ba922e7770fb58bbd14ad732bbf5ba6f11cc56989e6e
java.io.IOException: No space left on device
at java.io.FileOutputStream.writeBytes(Native Method)
at java.io.FileOutputStream.write(FileOutputStream.java:326)
at java.io.BufferedOutputStream.write(BufferedOutputStream.java:122)
at
java.nio.channels.Channels$WritableByteChannelImpl.write(Channels.java:458)
at
org.apache.beam.runners.fnexecution.artifact.AbstractArtifactStagingService$PutArtifactStreamObserver.onNext(AbstractArtifactStagingService.java:170)
at
org.apache.beam.runners.fnexecution.artifact.AbstractArtifactStagingService$PutArtifactStreamObserver.onNext(AbstractArtifactStagingService.java:129)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.stub.ServerCalls$StreamingServerCallHandler$StreamingServerCallListener.onMessage(ServerCalls.java:251)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.ForwardingServerCallListener.onMessage(ForwardingServerCallListener.java:33)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.Contexts$ContextualizedServerCallListener.onMessage(Contexts.java:76)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.messagesAvailableInternal(ServerCallImpl.java:309)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.messagesAvailable(ServerCallImpl.java:292)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1MessagesAvailable.runInContext(ServerImpl.java:782)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Mar 11, 2020 4:05:28 PM
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor run
SEVERE: Exception while executing runnable
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed@7ee6fd48
java.lang.IllegalStateException: call already closed
at
org.apache.beam.vendor.grpc.v1p26p0.com.google.common.base.Preconditions.checkState(Preconditions.java:511)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl.closeInternal(ServerCallImpl.java:209)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl.close(ServerCallImpl.java:202)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.stub.ServerCalls$ServerCallStreamObserverImpl.onError(ServerCalls.java:360)
at
org.apache.beam.runners.fnexecution.artifact.AbstractArtifactStagingService$PutArtifactStreamObserver.onCompleted(AbstractArtifactStagingService.java:226)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.stub.ServerCalls$StreamingServerCallHandler$StreamingServerCallListener.onHalfClose(ServerCalls.java:262)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.PartialForwardingServerCallListener.onHalfClose(PartialForwardingServerCallListener.java:35)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.ForwardingServerCallListener.onHalfClose(ForwardingServerCallListener.java:23)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.ForwardingServerCallListener$SimpleForwardingServerCallListener.onHalfClose(ForwardingServerCallListener.java:40)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.Contexts$ContextualizedServerCallListener.onHalfClose(Contexts.java:86)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.halfClosed(ServerCallImpl.java:331)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1HalfClosed.runInContext(ServerImpl.java:817)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
[grpc-default-executor-1] ERROR
org.apache.beam.runners.fnexecution.artifact.AbstractArtifactStagingService -
Failed to write chunk of artifact worker to
/tmp/beam-artifact-staging/go-job-7-1583942724100167279_2340a45b-e9c0-408e-ae7a-2265141dcbd6/artifacts/artifact_87eba76e7f3164534045ba922e7770fb58bbd14ad732bbf5ba6f11cc56989e6e
java.io.IOException: No space left on device
at java.io.FileOutputStream.writeBytes(Native Method)
at java.io.FileOutputStream.write(FileOutputStream.java:326)
at java.io.BufferedOutputStream.write(BufferedOutputStream.java:122)
at
java.nio.channels.Channels$WritableByteChannelImpl.write(Channels.java:458)
at
org.apache.beam.runners.fnexecution.artifact.AbstractArtifactStagingService$PutArtifactStreamObserver.onNext(AbstractArtifactStagingService.java:170)
at
org.apache.beam.runners.fnexecution.artifact.AbstractArtifactStagingService$PutArtifactStreamObserver.onNext(AbstractArtifactStagingService.java:129)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.stub.ServerCalls$StreamingServerCallHandler$StreamingServerCallListener.onMessage(ServerCalls.java:251)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.ForwardingServerCallListener.onMessage(ForwardingServerCallListener.java:33)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.Contexts$ContextualizedServerCallListener.onMessage(Contexts.java:76)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.messagesAvailableInternal(ServerCallImpl.java:309)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.messagesAvailable(ServerCallImpl.java:292)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1MessagesAvailable.runInContext(ServerImpl.java:782)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
[grpc-default-executor-1] ERROR
org.apache.beam.runners.fnexecution.artifact.AbstractArtifactStagingService -
Failed to write chunk of artifact worker to
/tmp/beam-artifact-staging/go-job-7-1583942724100167279_2340a45b-e9c0-408e-ae7a-2265141dcbd6/artifacts/artifact_87eba76e7f3164534045ba922e7770fb58bbd14ad732bbf5ba6f11cc56989e6e
java.io.IOException: No space left on device
at java.io.FileOutputStream.writeBytes(Native Method)
at java.io.FileOutputStream.write(FileOutputStream.java:326)
at java.io.BufferedOutputStream.write(BufferedOutputStream.java:122)
at
java.nio.channels.Channels$WritableByteChannelImpl.write(Channels.java:458)
at
org.apache.beam.runners.fnexecution.artifact.AbstractArtifactStagingService$PutArtifactStreamObserver.onNext(AbstractArtifactStagingService.java:170)
at
org.apache.beam.runners.fnexecution.artifact.AbstractArtifactStagingService$PutArtifactStreamObserver.onNext(AbstractArtifactStagingService.java:129)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.stub.ServerCalls$StreamingServerCallHandler$StreamingServerCallListener.onMessage(ServerCalls.java:251)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.ForwardingServerCallListener.onMessage(ForwardingServerCallListener.java:33)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.Contexts$ContextualizedServerCallListener.onMessage(Contexts.java:76)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.messagesAvailableInternal(ServerCallImpl.java:309)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerCallImpl$ServerStreamListenerImpl.messagesAvailable(ServerCallImpl.java:292)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ServerImpl$JumpToApplicationThreadServerStreamListener$1MessagesAvailable.runInContext(ServerImpl.java:782)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.ContextRunnable.run(ContextRunnable.java:37)
at
org.apache.beam.vendor.grpc.v1p26p0.io.grpc.internal.SerializingExecutor.run(SerializingExecutor.java:123)
at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
2020/03/11 16:05:36 Test flatten:flatten failed: staging artifacts:
failed to stage ./sdks/go/test/build/bin/linux-amd64/worker in 3 attempts:
failed to close stream for ./sdks/go/test/build/bin/linux-amd64/worker
caused by:
rpc error: code = DataLoss desc = Failed to write chunk of artifact worker to
/tmp/beam-artifact-staging/go-job-7-1583942724100167279_2340a45b-e9c0-408e-ae7a-2265141dcbd6/artifacts/artifact_87eba76e7f3164534045ba922e7770fb58bbd14ad732bbf5ba6f11cc56989e6e;
failed to close stream for ./sdks/go/test/build/bin/linux-amd64/worker
caused by:
rpc error: code = DataLoss desc = Failed to write chunk of artifact worker to
/tmp/beam-artifact-staging/go-job-7-1583942724100167279_2340a45b-e9c0-408e-ae7a-2265141dcbd6/artifacts/artifact_87eba76e7f3164534045ba922e7770fb58bbd14ad732bbf5ba6f11cc56989e6e;
failed to close stream for ./sdks/go/test/build/bin/linux-amd64/worker
caused by:
rpc error: code = DataLoss desc = Failed to write chunk of artifact worker to
/tmp/beam-artifact-staging/go-job-7-1583942724100167279_2340a45b-e9c0-408e-ae7a-2265141dcbd6/artifacts/artifact_87eba76e7f3164534045ba922e7770fb58bbd14ad732bbf5ba6f11cc56989e6e;
failed to send chunks for ./sdks/go/test/build/bin/linux-amd64/worker
caused by:
chunk send failed
caused by:
EOF
2020/03/11 16:05:36 Result: 2 tests failed
if [[ ! -z "$JOB_PORT" ]]; then
# Shut down the job server
kill %1 || echo "Failed to shut down job server"
fi
# Delete the container locally and remotely
docker rmi $CONTAINER:$TAG || echo "Failed to remove container"
Error response from daemon: write
/var/lib/docker/image/overlay2/.tmp-repositories.json618424199: no space left
on device
Failed to remove container
gcloud --quiet container images delete $CONTAINER:$TAG || echo "Failed to
delete container"
Traceback (most recent call last):
File "/usr/lib/python2.7/logging/__init__.py", line 872, in emit
stream.write(ufs % msg)
File "/usr/lib/python2.7/codecs.py", line 706, in write
return self.writer.write(data)
File "/usr/lib/python2.7/codecs.py", line 370, in write
self.stream.write(data)
IOError: [Errno 28] No space left on device
Logged from file backend.py, line 107
Traceback (most recent call last):
File "/usr/lib/python2.7/logging/__init__.py", line 872, in emit
stream.write(ufs % msg)
File "/usr/lib/python2.7/codecs.py", line 706, in write
return self.writer.write(data)
File "/usr/lib/python2.7/codecs.py", line 370, in write
self.stream.write(data)
IOError: [Errno 28] No space left on device
Logged from file backend.py, line 107
Traceback (most recent call last):
File "/usr/lib/python2.7/logging/__init__.py", line 872, in emit
stream.write(ufs % msg)
File "/usr/lib/python2.7/codecs.py", line 706, in write
return self.writer.write(data)
File "/usr/lib/python2.7/codecs.py", line 370, in write
self.stream.write(data)
IOError: [Errno 28] No space left on device
Logged from file backend.py, line 107
Traceback (most recent call last):
File "/usr/lib/python2.7/logging/__init__.py", line 872, in emit
stream.write(ufs % msg)
File "/usr/lib/python2.7/codecs.py", line 706, in write
return self.writer.write(data)
File "/usr/lib/python2.7/codecs.py", line 370, in write
self.stream.write(data)
IOError: [Errno 28] No space left on device
Logged from file base.py, line 916
Traceback (most recent call last):
File "/usr/lib/python2.7/logging/__init__.py", line 872, in emit
stream.write(ufs % msg)
File "/usr/lib/python2.7/codecs.py", line 706, in write
return self.writer.write(data)
File "/usr/lib/python2.7/codecs.py", line 370, in write
self.stream.write(data)
IOError: [Errno 28] No space left on device
Logged from file log.py, line 186
Digests:
Traceback (most recent call last):
File "/usr/lib/python2.7/logging/__init__.py", line 872, in emit
stream.write(ufs % msg)
File "/usr/lib/python2.7/codecs.py", line 706, in write
return self.writer.write(data)
File "/usr/lib/python2.7/codecs.py", line 370, in write
self.stream.write(data)
IOError: [Errno 28] No space left on device
Logged from file log.py, line 186
-
us.gcr.io/apache-beam-testing/jenkins/beam_go_sdk@sha256:d68135088f2011d64e579387f2fb3388083500be5d177db2f151d25cb4c286be
Associated tags:
- 20200311-160238
Tags:
- us.gcr.io/apache-beam-testing/jenkins/beam_go_sdk:20200311-160238
Deleted [us.gcr.io/apache-beam-testing/jenkins/beam_go_sdk:20200311-160238].
Deleted
[us.gcr.io/apache-beam-testing/jenkins/beam_go_sdk@sha256:d68135088f2011d64e579387f2fb3388083500be5d177db2f151d25cb4c286be].
# Clean up tempdir
rm -rf $TMPDIR
if [[ "$TEST_EXIT_CODE" -eq 0 ]]; then
echo ">>> SUCCESS"
else
echo ">>> FAILURE"
fi
>>> FAILURE
exit $TEST_EXIT_CODE
> Task :sdks:go:test:flinkValidatesRunner FAILED
FAILURE: Build failed with an exception.
* Where:
Build file
'<https://builds.apache.org/job/beam_PostCommit_Go_VR_Flink/ws/src/sdks/go/test/build.gradle'>
line: 59
* What went wrong:
Execution failed for task ':sdks:go:test:flinkValidatesRunner'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug
option to get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
Deprecated Gradle features were used in this build, making it incompatible with
Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See
https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings
BUILD FAILED in 4m 49s
67 actionable tasks: 49 executed, 18 from cache
A build scan cannot be produced as an error occurred spooling the build data.
Please report this problem via https://gradle.com/scans/help/plugin and include
the following via copy/paste:
----------
Gradle version: 5.2.1
Plugin version: 2.3
java.lang.IllegalStateException: java.io.IOException: No space left on device
at com.gradle.scan.a.e.b.flush(SourceFile:233)
at com.gradle.scan.a.e.b.close(SourceFile:244)
at com.gradle.scan.a.e.c.close(SourceFile:66)
at com.gradle.scan.plugin.internal.n.a(SourceFile:78)
at com.gradle.scan.plugin.internal.e.c.c.b(SourceFile:125)
at com.gradle.scan.plugin.internal.e.c.c.a(SourceFile:117)
at com.gradle.scan.plugin.internal.o.a$a.a(SourceFile:30)
at com.gradle.scan.plugin.internal.o.a$a.a(SourceFile:19)
at com.gradle.scan.plugin.internal.o.a.c(SourceFile:60)
Caused by: java.io.IOException: No space left on device
at com.gradle.scan.a.e.b.flush(SourceFile:231)
... 8 more
----------
Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]