See
<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/1644/display/redirect>
------------------------------------------
[...truncated 72.20 KB...]
print('Found: %s.' % self.build_console_url(pipeline.options))
test_read_metrics (apache_beam.transforms.ptransform_test.PTransformTest) ... ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded
/tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_undeclared_outputs (apache_beam.transforms.ptransform_test.PTransformTest)
... ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded
/tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_multiple_empty_outputs
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded
/tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_par_do_with_multiple_outputs_and_using_yield
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded
/tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_as_dict_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ...
ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded
/tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_as_singleton_with_different_defaults
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ...
ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded
/tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_as_list_and_as_dict_side_inputs
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded
/tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded
/tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_as_singleton_without_unique_labels
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded
/tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_empty_singleton_side_input
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz
Collecting mock (from -r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz
Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded
/tmp/dataflow-requirements-cache/setuptools-39.1.0.zip
Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1))
File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz
Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded
/tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz
Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2))
File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.0.3.tar.gz
Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr
test_flattened_side_input
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_default_value_singleton_side_input
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_multi_valued_singleton_side_input
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_iterable_side_input
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
======================================================================
ERROR: test_par_do_with_multiple_outputs_and_using_return
(apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/transforms/ptransform_test.py",>
line 244, in test_par_do_with_multiple_outputs_and_using_return
pipeline.run()
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",>
line 102, in run
result = super(TestPipeline, self).run(test_runner_api)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",>
line 389, in run
self.to_runner_api(), self.runner, self._options).run(False)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",>
line 402, in run
return self.runner.run_pipeline(self)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",>
line 48, in run_pipeline
self.result = super(TestDataflowRunner, self).run_pipeline(pipeline)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",>
line 362, in run_pipeline
self.dataflow_client.create_job(self.job), self)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/utils/retry.py",>
line 180, in wrapper
return fun(*args, **kwargs)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",>
line 470, in create_job
self.create_job_description(job)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",>
line 500, in create_job_description
job.options, file_copy=self._gcs_file_copy)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",>
line 329, in stage_job_resources
setup_options.requirements_file, requirements_cache_path)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",>
line 263, in _populate_requirements_cache
processes.check_call(cmd_args)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/utils/processes.py",>
line 44, in check_call
return subprocess.check_call(*args, **kwargs)
File "/usr/lib/python2.7/subprocess.py", line 541, in check_call
raise CalledProcessError(retcode, cmd)
CalledProcessError: Command
'['<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/bin/python',>
'-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r',
'postcommit_requirements.txt', '--no-binary', ':all:']' returned non-zero exit
status 2
-------------------- >> begin captured logging << --------------------
root: DEBUG: Connecting using Google Application Default Credentials.
root: INFO: Starting GCS upload to
gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0515153229-354108.1526398349.354278/pipeline.pb...
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
root: INFO: Completed GCS upload to
gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0515153229-354108.1526398349.354278/pipeline.pb
root: INFO: Starting GCS upload to
gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0515153229-354108.1526398349.354278/requirements.txt...
root: INFO: Completed GCS upload to
gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0515153229-354108.1526398349.354278/requirements.txt
root: INFO: Executing command:
['<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/bin/python',>
'-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r',
'postcommit_requirements.txt', '--no-binary', ':all:']
--------------------- >> end captured logging << ---------------------
======================================================================
ERROR: test_flatten_multiple_pcollections_having_multiple_consumers
(apache_beam.transforms.ptransform_test.PTransformTest)
----------------------------------------------------------------------
Traceback (most recent call last):
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/transforms/ptransform_test.py",>
line 558, in test_flatten_multiple_pcollections_having_multiple_consumers
pipeline.run()
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",>
line 102, in run
result = super(TestPipeline, self).run(test_runner_api)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",>
line 389, in run
self.to_runner_api(), self.runner, self._options).run(False)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",>
line 402, in run
return self.runner.run_pipeline(self)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",>
line 48, in run_pipeline
self.result = super(TestDataflowRunner, self).run_pipeline(pipeline)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",>
line 362, in run_pipeline
self.dataflow_client.create_job(self.job), self)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/utils/retry.py",>
line 180, in wrapper
return fun(*args, **kwargs)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",>
line 470, in create_job
self.create_job_description(job)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",>
line 500, in create_job_description
job.options, file_copy=self._gcs_file_copy)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",>
line 329, in stage_job_resources
setup_options.requirements_file, requirements_cache_path)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/dependency.py",>
line 263, in _populate_requirements_cache
processes.check_call(cmd_args)
File
"<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/apache_beam/utils/processes.py",>
line 44, in check_call
return subprocess.check_call(*args, **kwargs)
File "/usr/lib/python2.7/subprocess.py", line 541, in check_call
raise CalledProcessError(retcode, cmd)
CalledProcessError: Command
'['<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/bin/python',>
'-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r',
'postcommit_requirements.txt', '--no-binary', ':all:']' returned non-zero exit
status 2
-------------------- >> begin captured logging << --------------------
root: DEBUG: Connecting using Google Application Default Credentials.
root: INFO: Starting GCS upload to
gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0515153229-331318.1526398349.331494/pipeline.pb...
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
root: INFO: Completed GCS upload to
gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0515153229-331318.1526398349.331494/pipeline.pb
root: INFO: Starting GCS upload to
gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0515153229-331318.1526398349.331494/requirements.txt...
root: INFO: Completed GCS upload to
gs://temp-storage-for-end-to-end-tests/staging-validatesrunner-test/beamapp-jenkins-0515153229-331318.1526398349.331494/requirements.txt
root: INFO: Executing command:
['<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/bin/python',>
'-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r',
'postcommit_requirements.txt', '--no-binary', ':all:']
--------------------- >> end captured logging << ---------------------
----------------------------------------------------------------------
XML:
<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 16 tests in 1747.997s
FAILED (errors=2)
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_32_37-14403948038456206230?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_39_45-9721084800465220991?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_47_17-10213914719350645389?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_54_04-3494677997403759866?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_32_35-7576714796728505092?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_40_34-11401563495647624484?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_47_46-9536776248332820208?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_32_35-12839307802128079638?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_40_14-1432642210495004185?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_47_41-9434956304677883381?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_54_18-14332173872153237871?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_32_38-2250451714965257441?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_40_07-2945879026317262988?project=apache-beam-testing.
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-05-15_08_47_45-18057529855950065625?project=apache-beam-testing.
Build step 'Execute shell' marked build as failure