See
<https://builds.apache.org/job/beam_PostCommit_Python_Verify/4181/display/redirect>
------------------------------------------
[...truncated 1.54 MB...]
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
}
]
},
"output_name": "out",
"user_name":
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0).output"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s14"
},
"user_name":
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0)"
}
},
{
"kind": "ParallelDo",
"name": "s17",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "_finalize_write"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {
"SideInput-s15": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "SideInput-s15"
},
"SideInput-s16": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "SideInput-s16"
}
},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "write/Write/WriteImpl/FinalizeWrite.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s7"
},
"serialized_fn":
"ref_AppliedPTransform_write/Write/WriteImpl/FinalizeWrite_20",
"user_name": "write/Write/WriteImpl/FinalizeWrite/FinalizeWrite"
}
}
],
"type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
createTime: u'2018-02-08T20:41:20.713075Z'
currentStateTime: u'1970-01-01T00:00:00Z'
id: u'2018-02-08_12_41_19-4541629610088221710'
location: u'us-central1'
name: u'beamapp-jenkins-0208204118-252128'
projectId: u'apache-beam-testing'
stageStates: []
steps: []
tempFiles: []
type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2018-02-08_12_41_19-4541629610088221710]
root: INFO: To access the Dataflow monitoring console, please navigate to
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-02-08_12_41_19-4541629610088221710?project=apache-beam-testing
root: INFO: Job 2018-02-08_12_41_19-4541629610088221710 is in state
JOB_STATE_PENDING
root: INFO: 2018-02-08T20:41:19.903Z: JOB_MESSAGE_WARNING: (3f071c115ab441d2):
Job 2018-02-08_12_41_19-4541629610088221710 might autoscale up to 30 workers.
root: INFO: 2018-02-08T20:41:19.923Z: JOB_MESSAGE_DETAILED: (3f071c115ab44f43):
Autoscaling is enabled for job 2018-02-08_12_41_19-4541629610088221710. The
number of workers will be between 1 and 30.
root: INFO: 2018-02-08T20:41:19.960Z: JOB_MESSAGE_DETAILED: (3f071c115ab44cb4):
Autoscaling was automatically enabled for job
2018-02-08_12_41_19-4541629610088221710.
root: INFO: 2018-02-08T20:41:22.277Z: JOB_MESSAGE_DETAILED: (92eee085f775f008):
Checking required Cloud APIs are enabled.
root: INFO: 2018-02-08T20:41:22.506Z: JOB_MESSAGE_DETAILED: (92eee085f775f8d8):
Expanding CoGroupByKey operations into optimizable parts.
root: INFO: 2018-02-08T20:41:22.523Z: JOB_MESSAGE_DEBUG: (92eee085f775ff7f):
Combiner lifting skipped for step write/Write/WriteImpl/GroupByKey: GroupByKey
not followed by a combiner.
root: INFO: 2018-02-08T20:41:22.554Z: JOB_MESSAGE_DEBUG: (92eee085f775fe99):
Combiner lifting skipped for step group: GroupByKey not followed by a combiner.
root: INFO: 2018-02-08T20:41:22.585Z: JOB_MESSAGE_DETAILED: (92eee085f775fdb3):
Expanding GroupByKey operations into optimizable parts.
root: INFO: 2018-02-08T20:41:22.609Z: JOB_MESSAGE_DETAILED: (92eee085f775fccd):
Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2018-02-08T20:41:22.642Z: JOB_MESSAGE_DEBUG: (92eee085f775fa1b):
Annotating graph with Autotuner information.
root: INFO: 2018-02-08T20:41:22.668Z: JOB_MESSAGE_DETAILED: (92eee085f775f84f):
Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2018-02-08T20:41:22.696Z: JOB_MESSAGE_DETAILED: (92eee085f775f769):
Fusing consumer split into read/Read
root: INFO: 2018-02-08T20:41:22.729Z: JOB_MESSAGE_DETAILED: (92eee085f775f683):
Fusing consumer group/Write into group/Reify
root: INFO: 2018-02-08T20:41:22.762Z: JOB_MESSAGE_DETAILED: (92eee085f775f59d):
Fusing consumer group/GroupByWindow into group/Read
root: INFO: 2018-02-08T20:41:22.794Z: JOB_MESSAGE_DETAILED: (92eee085f775f4b7):
Fusing consumer write/Write/WriteImpl/GroupByKey/GroupByWindow into
write/Write/WriteImpl/GroupByKey/Read
root: INFO: 2018-02-08T20:41:22.819Z: JOB_MESSAGE_DETAILED: (92eee085f775f3d1):
Fusing consumer write/Write/WriteImpl/GroupByKey/Write into
write/Write/WriteImpl/GroupByKey/Reify
root: INFO: 2018-02-08T20:41:22.848Z: JOB_MESSAGE_DETAILED: (92eee085f775f2eb):
Fusing consumer write/Write/WriteImpl/WindowInto(WindowIntoFn) into
write/Write/WriteImpl/Pair
root: INFO: 2018-02-08T20:41:22.878Z: JOB_MESSAGE_DETAILED: (92eee085f775f205):
Fusing consumer write/Write/WriteImpl/GroupByKey/Reify into
write/Write/WriteImpl/WindowInto(WindowIntoFn)
root: INFO: 2018-02-08T20:41:22.904Z: JOB_MESSAGE_DETAILED: (92eee085f775f11f):
Fusing consumer pair_with_one into split
root: INFO: 2018-02-08T20:41:22.915Z: JOB_MESSAGE_DETAILED: (92eee085f775f039):
Fusing consumer group/Reify into pair_with_one
root: INFO: 2018-02-08T20:41:22.942Z: JOB_MESSAGE_DETAILED: (92eee085f775ff53):
Fusing consumer write/Write/WriteImpl/WriteBundles/WriteBundles into format
root: INFO: 2018-02-08T20:41:22.968Z: JOB_MESSAGE_DETAILED: (92eee085f775fe6d):
Fusing consumer write/Write/WriteImpl/Pair into
write/Write/WriteImpl/WriteBundles/WriteBundles
root: INFO: 2018-02-08T20:41:22.995Z: JOB_MESSAGE_DETAILED: (92eee085f775fd87):
Fusing consumer format into count
root: INFO: 2018-02-08T20:41:23.018Z: JOB_MESSAGE_DETAILED: (92eee085f775fca1):
Fusing consumer write/Write/WriteImpl/Extract into
write/Write/WriteImpl/GroupByKey/GroupByWindow
root: INFO: 2018-02-08T20:41:23.052Z: JOB_MESSAGE_DETAILED: (92eee085f775fbbb):
Fusing consumer count into group/GroupByWindow
root: INFO: 2018-02-08T20:41:23.086Z: JOB_MESSAGE_DETAILED: (92eee085f775fad5):
Fusing consumer write/Write/WriteImpl/InitializeWrite into
write/Write/WriteImpl/DoOnce/Read
root: INFO: 2018-02-08T20:41:23.114Z: JOB_MESSAGE_DEBUG: (92eee085f775f9ef):
Workflow config is missing a default resource spec.
root: INFO: 2018-02-08T20:41:23.145Z: JOB_MESSAGE_DEBUG: (92eee085f775f909):
Adding StepResource setup and teardown to workflow graph.
root: INFO: 2018-02-08T20:41:23.169Z: JOB_MESSAGE_DEBUG: (92eee085f775f823):
Adding workflow start and stop steps.
root: INFO: 2018-02-08T20:41:23.199Z: JOB_MESSAGE_DEBUG: (92eee085f775f73d):
Assigning stage ids.
root: INFO: 2018-02-08T20:41:23.311Z: JOB_MESSAGE_DEBUG: (e23224b7e92083d6):
Executing wait step start25
root: INFO: 2018-02-08T20:41:23.377Z: JOB_MESSAGE_BASIC: (a2e00ca6b90abb30):
Executing operation
write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
root: INFO: 2018-02-08T20:41:23.400Z: JOB_MESSAGE_BASIC: (e23224b7e9208d7c):
Executing operation group/Create
root: INFO: 2018-02-08T20:41:23.411Z: JOB_MESSAGE_DEBUG: (bec9a3401cb34e21):
Starting worker pool setup.
root: INFO: 2018-02-08T20:41:23.447Z: JOB_MESSAGE_BASIC: (bec9a3401cb34d7b):
Starting 1 workers in us-central1-f...
root: INFO: 2018-02-08T20:41:23.513Z: JOB_MESSAGE_DEBUG: (e23224b7e92082e6):
Value "group/Session" materialized.
root: INFO: 2018-02-08T20:41:23.572Z: JOB_MESSAGE_BASIC: (fef37066d90d0532):
Executing operation read/Read+split+pair_with_one+group/Reify+group/Write
root: INFO: Job 2018-02-08_12_41_19-4541629610088221710 is in state
JOB_STATE_RUNNING
root: INFO: 2018-02-08T20:41:31.955Z: JOB_MESSAGE_DETAILED: (12e782f1271849f):
Autoscaling: Raised the number of workers to 0 based on the rate of progress in
the currently running step(s).
root: INFO: 2018-02-08T20:41:47.648Z: JOB_MESSAGE_DETAILED: (12e782f12718976):
Autoscaling: Raised the number of workers to 1 based on the rate of progress in
the currently running step(s).
root: INFO: 2018-02-08T20:43:33.843Z: JOB_MESSAGE_DETAILED: (d05db0f41ce9d2f):
Workers have started successfully.
root: INFO: 2018-02-08T20:46:15.827Z: JOB_MESSAGE_ERROR: (8204ecdead94b6d4):
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py",
line 609, in do_work
work_executor.execute()
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py",
line 167, in execute
op.start()
File "apache_beam/runners/worker/operations.py", line 340, in
apache_beam.runners.worker.operations.DoOperation.start
def start(self):
File "apache_beam/runners/worker/operations.py", line 341, in
apache_beam.runners.worker.operations.DoOperation.start
with self.scoped_start_state:
File "apache_beam/runners/worker/operations.py", line 346, in
apache_beam.runners.worker.operations.DoOperation.start
pickler.loads(self.spec.serialized_fn))
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line
217, in loads
s = zlib.decompress(c)
error: Error -3 while decompressing data: incorrect header check
root: INFO: 2018-02-08T20:46:19.174Z: JOB_MESSAGE_ERROR: (8204ecdead94bf48):
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py",
line 609, in do_work
work_executor.execute()
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py",
line 167, in execute
op.start()
File "apache_beam/runners/worker/operations.py", line 340, in
apache_beam.runners.worker.operations.DoOperation.start
def start(self):
File "apache_beam/runners/worker/operations.py", line 341, in
apache_beam.runners.worker.operations.DoOperation.start
with self.scoped_start_state:
File "apache_beam/runners/worker/operations.py", line 346, in
apache_beam.runners.worker.operations.DoOperation.start
pickler.loads(self.spec.serialized_fn))
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line
217, in loads
s = zlib.decompress(c)
error: Error -3 while decompressing data: incorrect header check
root: INFO: 2018-02-08T20:46:22.505Z: JOB_MESSAGE_ERROR: (8204ecdead94b7bc):
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py",
line 609, in do_work
work_executor.execute()
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py",
line 167, in execute
op.start()
File "apache_beam/runners/worker/operations.py", line 340, in
apache_beam.runners.worker.operations.DoOperation.start
def start(self):
File "apache_beam/runners/worker/operations.py", line 341, in
apache_beam.runners.worker.operations.DoOperation.start
with self.scoped_start_state:
File "apache_beam/runners/worker/operations.py", line 346, in
apache_beam.runners.worker.operations.DoOperation.start
pickler.loads(self.spec.serialized_fn))
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line
217, in loads
s = zlib.decompress(c)
error: Error -3 while decompressing data: incorrect header check
root: INFO: 2018-02-08T20:46:25.858Z: JOB_MESSAGE_ERROR: (8204ecdead94b030):
Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py",
line 609, in do_work
work_executor.execute()
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py",
line 167, in execute
op.start()
File "apache_beam/runners/worker/operations.py", line 340, in
apache_beam.runners.worker.operations.DoOperation.start
def start(self):
File "apache_beam/runners/worker/operations.py", line 341, in
apache_beam.runners.worker.operations.DoOperation.start
with self.scoped_start_state:
File "apache_beam/runners/worker/operations.py", line 346, in
apache_beam.runners.worker.operations.DoOperation.start
pickler.loads(self.spec.serialized_fn))
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line
217, in loads
s = zlib.decompress(c)
error: Error -3 while decompressing data: incorrect header check
root: INFO: 2018-02-08T20:46:25.897Z: JOB_MESSAGE_DEBUG: (a28958f0997a32ba):
Executing failure step failure24
root: INFO: 2018-02-08T20:46:25.935Z: JOB_MESSAGE_ERROR: (a28958f0997a30d8):
Workflow failed. Causes: (fef37066d90d0d53):
S05:read/Read+split+pair_with_one+group/Reify+group/Write failed.,
(43fb6c8d75e9ff90): A work item was attempted 4 times without success. Each
time the worker eventually lost contact with the service. The work item was
attempted on:
beamapp-jenkins-020820411-02081241-4461-harness-hm94,
beamapp-jenkins-020820411-02081241-4461-harness-hm94,
beamapp-jenkins-020820411-02081241-4461-harness-hm94,
beamapp-jenkins-020820411-02081241-4461-harness-hm94
root: INFO: 2018-02-08T20:46:26.057Z: JOB_MESSAGE_DETAILED: (92eee085f775fb32):
Cleaning up.
root: INFO: 2018-02-08T20:46:26.105Z: JOB_MESSAGE_DEBUG: (92eee085f775f966):
Starting worker pool teardown.
root: INFO: 2018-02-08T20:46:26.130Z: JOB_MESSAGE_BASIC: (92eee085f775f880):
Stopping worker pool...
root: INFO: 2018-02-08T20:48:48.012Z: JOB_MESSAGE_DETAILED: (12e782f12718a2f):
Autoscaling: Reduced the number of workers to 0 based on the rate of progress
in the currently running step(s).
root: INFO: 2018-02-08T20:48:48.052Z: JOB_MESSAGE_BASIC: (92eee085f775f4e8):
Worker pool stopped.
root: INFO: 2018-02-08T20:48:48.084Z: JOB_MESSAGE_DEBUG: (92eee085f775f31c):
Tearing down pending resources...
root: INFO: Job 2018-02-08_12_41_19-4541629610088221710 is in state
JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------
----------------------------------------------------------------------
Ran 3 tests in 465.602s
FAILED (errors=3)
Build step 'Execute shell' marked build as failure