See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/1077/>

------------------------------------------
[...truncated 8519 lines...]
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa8b: 
2017-01-24T09:18:30.027Z: JOB_MESSAGE_DETAILED: (87856bc85f4e273e): Fusing 
consumer assert:even/Group/GroupByWindow into assert:even/Group/Read
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa8d: 
2017-01-24T09:18:30.029Z: JOB_MESSAGE_DETAILED: (87856bc85f4e24e0): Fusing 
consumer assert_that/Match into assert_that/UnKey
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa8f: 
2017-01-24T09:18:30.031Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2282): Fusing 
consumer assert_that/UnKey into assert_that/Group/GroupByWindow
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa91: 
2017-01-24T09:18:30.033Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2024): Fusing 
consumer assert_that/Group/GroupByWindow into assert_that/Group/Read
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa93: 
2017-01-24T09:18:30.035Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2dc6): Fusing 
consumer assert_that/Group/Write into assert_that/Group/Reify
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa95: 
2017-01-24T09:18:30.037Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2b68): Fusing 
consumer assert_that/Group/Reify into assert_that/ToVoidKey
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa97: 
2017-01-24T09:18:30.039Z: JOB_MESSAGE_DETAILED: (87856bc85f4e290a): Fusing 
consumer assert_that/ToVoidKey into assert_that/WindowInto
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa99: 
2017-01-24T09:18:30.041Z: JOB_MESSAGE_DETAILED: (87856bc85f4e26ac): Fusing 
consumer assert:odd/Group/GroupByWindow into assert:odd/Group/Read
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa9b: 
2017-01-24T09:18:30.043Z: JOB_MESSAGE_DETAILED: (87856bc85f4e244e): Fusing 
consumer assert:even/Group/Write into assert:even/Group/Reify
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa9d: 
2017-01-24T09:18:30.045Z: JOB_MESSAGE_DETAILED: (87856bc85f4e21f0): Fusing 
consumer assert:even/Match into assert:even/UnKey
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fa9f: 
2017-01-24T09:18:30.047Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2f92): Fusing 
consumer assert:even/Group/Reify into assert:even/ToVoidKey
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faa1: 
2017-01-24T09:18:30.049Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2d34): Fusing 
consumer assert:odd/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faa3: 
2017-01-24T09:18:30.051Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2ad6): Fusing 
consumer assert:odd/Group/Write into assert:odd/Group/Reify
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faa5: 
2017-01-24T09:18:30.053Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2878): Fusing 
consumer assert:even/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faa7: 
2017-01-24T09:18:30.055Z: JOB_MESSAGE_DETAILED: (87856bc85f4e261a): Fusing 
consumer assert:even/ToVoidKey into assert:even/WindowInto
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faa9: 
2017-01-24T09:18:30.057Z: JOB_MESSAGE_DETAILED: (87856bc85f4e23bc): Fusing 
consumer assert_that/WindowInto into ClassifyNumbers/ClassifyNumbers
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faab: 
2017-01-24T09:18:30.059Z: JOB_MESSAGE_DETAILED: (87856bc85f4e215e): Fusing 
consumer assert:odd/Match into assert:odd/UnKey
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4faad: 
2017-01-24T09:18:30.061Z: JOB_MESSAGE_DETAILED: (87856bc85f4e2f00): Fusing 
consumer assert:odd/Group/Reify into assert:odd/ToVoidKey
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fb00: 
2017-01-24T09:18:30.144Z: JOB_MESSAGE_DEBUG: (87856bc85f4e250c): Workflow 
config is missing a default resource spec.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fb03: 
2017-01-24T09:18:30.147Z: JOB_MESSAGE_DETAILED: (87856bc85f4e22ae): Adding 
StepResource setup and teardown to workflow graph.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fb21: 
2017-01-24T09:18:30.177Z: JOB_MESSAGE_DEBUG: (31cd4de23232ef81): Adding 
workflow start and stop steps.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fb53: 
2017-01-24T09:18:30.227Z: JOB_MESSAGE_DEBUG: (f93568740c05d1b1): Assigning 
stage ids.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fba9: 
2017-01-24T09:18:30.313Z: JOB_MESSAGE_DEBUG: (f93568740c05d098): Executing wait 
step start2
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fbb9: 
2017-01-24T09:18:30.329Z: JOB_MESSAGE_DEBUG: (ac2138444fc23d7d): Executing 
operation Some Numbers
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fbee: 
2017-01-24T09:18:30.382Z: JOB_MESSAGE_DEBUG: (a5eb5b5ec15467df): Value "Some 
Numbers.out" materialized.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fbf7: 
2017-01-24T09:18:30.391Z: JOB_MESSAGE_BASIC: S02: (f93568740c05d91c): Executing 
operation assert:even/Group/Create
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fbfa: 
2017-01-24T09:18:30.394Z: JOB_MESSAGE_BASIC: S03: (d90ef3f76c48fa2a): Executing 
operation assert_that/Group/Create
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fc1b: 
2017-01-24T09:18:30.427Z: JOB_MESSAGE_BASIC: S01: (af4267f01a05b233): Executing 
operation assert:odd/Group/Create
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fcc3: 
2017-01-24T09:18:30.595Z: JOB_MESSAGE_DEBUG: (b9d847bf036a3ee3): Starting 
worker pool setup.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fcc5: 
2017-01-24T09:18:30.597Z: JOB_MESSAGE_BASIC: (b9d847bf036a36e5): Starting 1 
workers...
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fcd6: 
2017-01-24T09:18:30.614Z: JOB_MESSAGE_DEBUG: (f1cc619595f671fd): Value 
"assert_that/Group/Session" materialized.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fcd8: 
2017-01-24T09:18:30.616Z: JOB_MESSAGE_DEBUG: (f93568740c05d2a8): Value 
"assert:even/Group/Session" materialized.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc4fce3: 
2017-01-24T09:18:30.627Z: JOB_MESSAGE_BASIC: S04: (87856bc85f4e2cc3): Executing 
operation 
ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc6b696: 
2017-01-24T09:20:23.702Z: JOB_MESSAGE_DETAILED: (43b1b18f7b1c6b33): Workers 
have started successfully.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7ca91: 
2017-01-24T09:21:34.353Z: JOB_MESSAGE_ERROR: (72f067c206330db1): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 
212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in 
_import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py",
 line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cb72: 
2017-01-24T09:21:34.578Z: JOB_MESSAGE_ERROR: (72f067c2063301ae): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 
212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in 
_import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py",
 line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cc0f: 
2017-01-24T09:21:34.735Z: JOB_MESSAGE_ERROR: (72f067c206330e98): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 
212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in 
_import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py",
 line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7ccaf: 
2017-01-24T09:21:34.895Z: JOB_MESSAGE_ERROR: (72f067c206330b82): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 
212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in 
_import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py",
 line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cd46: 
2017-01-24T09:21:35.046Z: JOB_MESSAGE_ERROR: (72f067c20633086c): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 
212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in 
_import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py",
 line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7ce14: 
2017-01-24T09:21:35.252Z: JOB_MESSAGE_ERROR: (72f067c206330c69): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 
212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in 
_import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py",
 line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cedc: 
2017-01-24T09:21:35.452Z: JOB_MESSAGE_ERROR: (72f067c206330066): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 
212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in 
_import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py",
 line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cfad: 
2017-01-24T09:21:35.661Z: JOB_MESSAGE_ERROR: (72f067c206330463): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 899, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:26452)
    op.start()
  File "dataflow_worker/executor.py", line 464, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:15269)
    def start(self):
  File "dataflow_worker/executor.py", line 469, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14434)
    pickler.loads(self.spec.serialized_fn))
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/internal/pickler.py", line 
212, in loads
    return dill.loads(s)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 260, in loads
    return load(file)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 250, in load
    obj = pik.load()
  File "/usr/lib/python2.7/pickle.py", line 858, in load
    dispatch[key](self)
  File "/usr/lib/python2.7/pickle.py", line 1133, in load_reduce
    value = func(*args)
  File "/usr/local/lib/python2.7/dist-packages/dill/dill.py", line 726, in 
_import_module
    return getattr(__import__(module, None, None, [obj]), obj)
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/ptransform_test.py",
 line 26, in <module>
    import hamcrest as hc
ImportError: No module named hamcrest

root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cfd6: 
2017-01-24T09:21:35.702Z: JOB_MESSAGE_DEBUG: (5129b9404837122d): Executing 
failure step failure1
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7cfd9: 
2017-01-24T09:21:35.705Z: JOB_MESSAGE_ERROR: (5129b94048371343): Workflow 
failed. Causes: (87856bc85f4e2d60): 
S04:ClassifyNumbers/ClassifyNumbers+assert:odd/WindowInto+assert:odd/ToVoidKey+assert:even/WindowInto+assert:even/ToVoidKey+assert:even/Group/Reify+assert:even/Group/Write+assert_that/WindowInto+assert_that/ToVoidKey+assert_that/Group/Reify+assert_that/Group/Write+assert:odd/Group/Reify+assert:odd/Group/Write
 failed.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7d016: 
2017-01-24T09:21:35.766Z: JOB_MESSAGE_DETAILED: (90c2002d0fc71cca): Cleaning up.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7d094: 
2017-01-24T09:21:35.892Z: JOB_MESSAGE_DEBUG: (90c2002d0fc71e84): Starting 
worker pool teardown.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc7d096: 
2017-01-24T09:21:35.894Z: JOB_MESSAGE_BASIC: (90c2002d0fc7103e): Stopping 
worker pool...
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc8f596: 
2017-01-24T09:22:50.902Z: JOB_MESSAGE_BASIC: (90c2002d0fc71ad5): Worker pool 
stopped.
root: INFO: 2017-01-24_01_18_28-12399986343756263108_00000159cfc8f99a: 
2017-01-24T09:22:51.930Z: JOB_MESSAGE_DEBUG: (90c2002d0fc71003): Tearing down 
pending resources...
root: INFO: Job 2017-01-24_01_18_28-12399986343756263108 is in state 
JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 14 tests in 1040.088s

FAILED (errors=4)
Build step 'Execute shell' marked build as failure

Reply via email to