See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/1075/changes>
Changes:
[chamikara] Increments major version used by Dataflow runner to 5
[robertwb] Remove dataflow_test.py
[robertwb] Code cleanup now that all runners support windowed side inputs.
------------------------------------------
[...truncated 8425 lines...]
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "assert:even/UnKey.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s15"
},
"serialized_fn": "<string of 1052 bytes>",
"user_name": "assert:even/UnKey"
}
},
{
"kind": "ParallelDo",
"name": "s17",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "_equal"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "assert:even/Match.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s16"
},
"serialized_fn": "<string of 1212 bytes>",
"user_name": "assert:even/Match"
}
}
],
"type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
id: u'2017-01-23_15_30_54-12011182360156094633'
projectId: u'apache-beam-testing'
steps: []
tempFiles: []
type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2017-01-23_15_30_54-12011182360156094633]
root: INFO: To access the Dataflow monitoring console, please navigate to
https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2017-01-23_15_30_54-12011182360156094633
root: INFO: Job 2017-01-23_15_30_54-12011182360156094633 is in state
JOB_STATE_RUNNING
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab092b:
2017-01-23T23:30:55.403Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312046): Checking
required Cloud APIs are enabled.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b0b:
2017-01-23T23:30:55.883Z: JOB_MESSAGE_DEBUG: (3fadbfc3db312195): Combiner
lifting skipped for step assert_that/Group: GroupByKey not followed by a
combiner.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b0d:
2017-01-23T23:30:55.885Z: JOB_MESSAGE_DEBUG: (3fadbfc3db312ac3): Combiner
lifting skipped for step assert:even/Group: GroupByKey not followed by a
combiner.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b0f:
2017-01-23T23:30:55.887Z: JOB_MESSAGE_DEBUG: (3fadbfc3db3123f1): Combiner
lifting skipped for step assert:odd/Group: GroupByKey not followed by a
combiner.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b12:
2017-01-23T23:30:55.890Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312d1f): Expanding
GroupByKey operations into optimizable parts.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b14:
2017-01-23T23:30:55.892Z: JOB_MESSAGE_DETAILED: (3fadbfc3db31264d): Lifting
ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b1e:
2017-01-23T23:30:55.902Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312b05): Annotating
graph with Autotuner information.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b29:
2017-01-23T23:30:55.913Z: JOB_MESSAGE_DETAILED: (3fadbfc3db3129f8): Fusing
adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b2f:
2017-01-23T23:30:55.919Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312c54): Fusing
consumer assert:odd/ToVoidKey into assert:odd/WindowInto
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b32:
2017-01-23T23:30:55.922Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312582): Fusing
consumer assert:odd/UnKey into assert:odd/Group/GroupByWindow
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b34:
2017-01-23T23:30:55.924Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312eb0): Fusing
consumer assert:even/UnKey into assert:even/Group/GroupByWindow
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b37:
2017-01-23T23:30:55.927Z: JOB_MESSAGE_DETAILED: (3fadbfc3db3127de): Fusing
consumer assert:even/Group/GroupByWindow into assert:even/Group/Read
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b39:
2017-01-23T23:30:55.929Z: JOB_MESSAGE_DETAILED: (3fadbfc3db31210c): Fusing
consumer assert_that/Match into assert_that/UnKey
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b3b:
2017-01-23T23:30:55.931Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312a3a): Fusing
consumer assert_that/UnKey into assert_that/Group/GroupByWindow
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b3d:
2017-01-23T23:30:55.933Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312368): Fusing
consumer assert_that/Group/GroupByWindow into assert_that/Group/Read
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b3f:
2017-01-23T23:30:55.935Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312c96): Fusing
consumer assert_that/Group/Write into assert_that/Group/Reify
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b42:
2017-01-23T23:30:55.938Z: JOB_MESSAGE_DETAILED: (3fadbfc3db3125c4): Fusing
consumer assert_that/Group/Reify into assert_that/ToVoidKey
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b44:
2017-01-23T23:30:55.940Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312ef2): Fusing
consumer assert_that/ToVoidKey into assert_that/WindowInto
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b46:
2017-01-23T23:30:55.942Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312820): Fusing
consumer assert:odd/Group/GroupByWindow into assert:odd/Group/Read
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b48:
2017-01-23T23:30:55.944Z: JOB_MESSAGE_DETAILED: (3fadbfc3db31214e): Fusing
consumer assert:even/Group/Write into assert:even/Group/Reify
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b4a:
2017-01-23T23:30:55.946Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312a7c): Fusing
consumer assert:even/Match into assert:even/UnKey
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b4c:
2017-01-23T23:30:55.948Z: JOB_MESSAGE_DETAILED: (3fadbfc3db3123aa): Fusing
consumer assert:even/Group/Reify into assert:even/ToVoidKey
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b4f:
2017-01-23T23:30:55.951Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312cd8): Fusing
consumer assert:odd/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b51:
2017-01-23T23:30:55.953Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312606): Fusing
consumer assert:odd/Group/Write into assert:odd/Group/Reify
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b53:
2017-01-23T23:30:55.955Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312f34): Fusing
consumer assert:even/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b55:
2017-01-23T23:30:55.957Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312862): Fusing
consumer assert:even/ToVoidKey into assert:even/WindowInto
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b57:
2017-01-23T23:30:55.959Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312190): Fusing
consumer assert_that/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b59:
2017-01-23T23:30:55.961Z: JOB_MESSAGE_DETAILED: (3fadbfc3db312abe): Fusing
consumer assert:odd/Match into assert:odd/UnKey
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0b5b:
2017-01-23T23:30:55.963Z: JOB_MESSAGE_DETAILED: (3fadbfc3db3123ec): Fusing
consumer assert:odd/Group/Reify into assert:odd/ToVoidKey
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0bb1:
2017-01-23T23:30:56.049Z: JOB_MESSAGE_DEBUG: (3fadbfc3db3120c0): Workflow
config is missing a default resource spec.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0bb3:
2017-01-23T23:30:56.051Z: JOB_MESSAGE_DETAILED: (3fadbfc3db3129ee): Adding
StepResource setup and teardown to workflow graph.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0bf6:
2017-01-23T23:30:56.118Z: JOB_MESSAGE_DEBUG: (e31afff0dc916c6a): Adding
workflow start and stop steps.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0c28:
2017-01-23T23:30:56.168Z: JOB_MESSAGE_DEBUG: (dd9172f79e564817): Assigning
stage ids.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0ccc:
2017-01-23T23:30:56.332Z: JOB_MESSAGE_DEBUG: (aaac8d07a592185a): Executing wait
step start2
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0cd8:
2017-01-23T23:30:56.344Z: JOB_MESSAGE_DEBUG: (e5b879d3fc421531): Executing
operation Some Numbers
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0ce6:
2017-01-23T23:30:56.358Z: JOB_MESSAGE_DEBUG: (e31afff0dc916fc1): Value "Some
Numbers.out" materialized.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0cf1:
2017-01-23T23:30:56.369Z: JOB_MESSAGE_BASIC: S01: (aaac8d07a5921fae): Executing
operation assert:odd/Group/Create
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0cf4:
2017-01-23T23:30:56.372Z: JOB_MESSAGE_BASIC: S02: (3fadbfc3db3128e1): Executing
operation assert:even/Group/Create
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0cf5:
2017-01-23T23:30:56.373Z: JOB_MESSAGE_BASIC: S03: (6e37547a5e2c658a): Executing
operation assert_that/Group/Create
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0dbe:
2017-01-23T23:30:56.574Z: JOB_MESSAGE_DEBUG: (e9a91083e1434023): Starting
worker pool setup.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0dc0:
2017-01-23T23:30:56.576Z: JOB_MESSAGE_BASIC: (e9a91083e14348bd): Starting 1
workers...
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0dd2:
2017-01-23T23:30:56.594Z: JOB_MESSAGE_DEBUG: (708a4aa4ce81981a): Value
"assert:odd/Group/Session" materialized.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0dd4:
2017-01-23T23:30:56.596Z: JOB_MESSAGE_DEBUG: (dd9172f79e564c8e): Value
"assert_that/Group/Session" materialized.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0dea:
2017-01-23T23:30:56.618Z: JOB_MESSAGE_DEBUG: (3bd4ccbdb25e4eac): Value
"assert:even/Group/Session" materialized.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdab0df5:
2017-01-23T23:30:56.629Z: JOB_MESSAGE_BASIC: S04: (e5b879d3fc4213da): Executing
operation
ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdabed2f:
2017-01-23T23:31:53.775Z: JOB_MESSAGE_DETAILED: (d3212f834d43fd2a): Workers
have started successfully.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd6e4:
2017-01-23T23:33:59.140Z: JOB_MESSAGE_ERROR: (d3f0897736e1c274): Traceback
(most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py",
line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 899, in
dataflow_worker.executor.MapTaskExecutor.execute
(dataflow_worker/executor.c:26452)
op.start()
File "dataflow_worker/executor.py", line 464, in
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
def start(self):
File "dataflow_worker/executor.py", line 469, in
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
pickler.loads(self.spec.serialized_fn))
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line
212, in loads
return dill.loads(s)
File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
return load(file)
File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
obj = pik.load()
File "/usr/lib/python2.7/pickle.py", line 858, in load
dispatch[key](self)
File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
value = func(*args)
File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in
_import_module
return getattr(__import__(module, None, None, [obj]), obj)
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py",
line 26, in <module>
import hamcrest as hc
ImportError: No module named hamcrest
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd7cb:
2017-01-23T23:33:59.371Z: JOB_MESSAGE_ERROR: (d3f0897736e1c8b9): Traceback
(most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py",
line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 899, in
dataflow_worker.executor.MapTaskExecutor.execute
(dataflow_worker/executor.c:26452)
op.start()
File "dataflow_worker/executor.py", line 464, in
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
def start(self):
File "dataflow_worker/executor.py", line 469, in
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
pickler.loads(self.spec.serialized_fn))
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line
212, in loads
return dill.loads(s)
File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
return load(file)
File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
obj = pik.load()
File "/usr/lib/python2.7/pickle.py", line 858, in load
dispatch[key](self)
File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
value = func(*args)
File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in
_import_module
return getattr(__import__(module, None, None, [obj]), obj)
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py",
line 26, in <module>
import hamcrest as hc
ImportError: No module named hamcrest
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd882:
2017-01-23T23:33:59.554Z: JOB_MESSAGE_ERROR: (d3f0897736e1c3b3): Traceback
(most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py",
line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 899, in
dataflow_worker.executor.MapTaskExecutor.execute
(dataflow_worker/executor.c:26452)
op.start()
File "dataflow_worker/executor.py", line 464, in
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
def start(self):
File "dataflow_worker/executor.py", line 469, in
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
pickler.loads(self.spec.serialized_fn))
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line
212, in loads
return dill.loads(s)
File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
return load(file)
File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
obj = pik.load()
File "/usr/lib/python2.7/pickle.py", line 858, in load
dispatch[key](self)
File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
value = func(*args)
File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in
_import_module
return getattr(__import__(module, None, None, [obj]), obj)
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py",
line 26, in <module>
import hamcrest as hc
ImportError: No module named hamcrest
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd965:
2017-01-23T23:33:59.781Z: JOB_MESSAGE_ERROR: (d3f0897736e1c9f8): Traceback
(most recent call last):
File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py",
line 514, in do_work
work_executor.execute()
File "dataflow_worker/executor.py", line 899, in
dataflow_worker.executor.MapTaskExecutor.execute
(dataflow_worker/executor.c:26452)
op.start()
File "dataflow_worker/executor.py", line 464, in
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
def start(self):
File "dataflow_worker/executor.py", line 469, in
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
pickler.loads(self.spec.serialized_fn))
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line
212, in loads
return dill.loads(s)
File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
return load(file)
File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
obj = pik.load()
File "/usr/lib/python2.7/pickle.py", line 858, in load
dispatch[key](self)
File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
value = func(*args)
File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in
_import_module
return getattr(__import__(module, None, None, [obj]), obj)
File
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py",
line 26, in <module>
import hamcrest as hc
ImportError: No module named hamcrest
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd98f:
2017-01-23T23:33:59.823Z: JOB_MESSAGE_DEBUG: (3fadbfc3db312ff5): Executing
failure step failure1
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd991:
2017-01-23T23:33:59.825Z: JOB_MESSAGE_ERROR: (3fadbfc3db312923): Workflow
failed. Causes: (e5b879d3fc42115d):
S04:ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write
failed.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadd9cf:
2017-01-23T23:33:59.887Z: JOB_MESSAGE_DETAILED: (580466590c7e466a): Cleaning up.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadda4c:
2017-01-23T23:34:00.012Z: JOB_MESSAGE_DEBUG: (580466590c7e4304): Starting
worker pool teardown.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdadda4e:
2017-01-23T23:34:00.014Z: JOB_MESSAGE_BASIC: (580466590c7e4f9e): Stopping
worker pool...
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdaf12d6:
2017-01-23T23:35:20.022Z: JOB_MESSAGE_BASIC: (580466590c7e4285): Worker pool
stopped.
root: INFO: 2017-01-23_15_30_54-12011182360156094633_00000159cdaf16d8:
2017-01-23T23:35:21.048Z: JOB_MESSAGE_DEBUG: (580466590c7e4853): Tearing down
pending resources...
root: INFO: Job 2017-01-23_15_30_54-12011182360156094633 is in state
JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------
----------------------------------------------------------------------
Ran 14 tests in 1186.872s
FAILED (errors=4)
Build step 'Execute shell' marked build as failure