See <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/789/display/redirect?page=changes>
Changes: [lukasz.gajowy] [BEAM-3359] Remove setting Flink master to [auto] in TestFlinkRunner [lukasz.gajowy] [BEAM-3370] Allow to run IO Performance tests using Flink runner [lukasz.gajowy] [BEAM-3370] Fix problem with directories not being staged from classpath ------------------------------------------ [...truncated 163.31 KB...] File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.0.0.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.0.0.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py>:54: DeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported print('Found: %s.' % self.build_console_url(pipeline.options)) <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py>:54: DeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported print('Found: %s.' % self.build_console_url(pipeline.options)) <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py>:54: DeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported print('Found: %s.' % self.build_console_url(pipeline.options)) <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py>:54: DeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported print('Found: %s.' % self.build_console_url(pipeline.options)) <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py>:54: DeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported print('Found: %s.' % self.build_console_url(pipeline.options)) <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py>:54: DeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported print('Found: %s.' % self.build_console_url(pipeline.options)) <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py>:54: DeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported print('Found: %s.' % self.build_console_url(pipeline.options)) <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py>:54: DeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported print('Found: %s.' % self.build_console_url(pipeline.options)) test_flatten_multiple_pcollections_having_multiple_consumers (apache_beam.transforms.ptransform_test.PTransformTest) ... ok test_par_do_with_multiple_outputs_and_using_return (apache_beam.transforms.ptransform_test.PTransformTest) ... ok test_undeclared_outputs (apache_beam.transforms.ptransform_test.PTransformTest) ... ok Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.0.0.zip File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.0.0.zip File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.0.0.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_multiple_empty_outputs (apache_beam.transforms.ptransform_test.PTransformTest) ... ok Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) test_as_singleton_with_different_defaults (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.0.0.zip File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.0.0.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr test_as_dict_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_as_list_and_as_dict_side_inputs (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_default_value_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_empty_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_as_singleton_without_unique_labels (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_flattened_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok ====================================================================== ERROR: test_par_do_with_multiple_outputs_and_using_yield (apache_beam.transforms.ptransform_test.PTransformTest) ---------------------------------------------------------------------- Traceback (most recent call last): File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/transforms/ptransform_test.py",> line 234, in test_par_do_with_multiple_outputs_and_using_yield pipeline.run() File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 104, in run result = super(TestPipeline, self).run(test_runner_api) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 394, in run self.to_runner_api(), self.runner, self._options).run(False) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 407, in run return self.runner.run_pipeline(self) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 50, in run_pipeline self.result = super(TestDataflowRunner, self).run_pipeline(pipeline) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 371, in run_pipeline self.dataflow_client.create_job(self.job), self) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/utils/retry.py",> line 184, in wrapper return fun(*args, **kwargs) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 490, in create_job self.create_job_description(job) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 519, in create_job_description resources = self._stage_resources(job.options) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 452, in _stage_resources staging_location=google_cloud_options.staging_location) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/portability/stager.py",> line 161, in stage_job_resources requirements_cache_path) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/portability/stager.py",> line 411, in _populate_requirements_cache processes.check_call(cmd_args) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/utils/processes.py",> line 46, in check_call return subprocess.check_call(*args, **kwargs) File "/usr/lib/python2.7/subprocess.py", line 541, in check_call raise CalledProcessError(retcode, cmd) CalledProcessError: Command '['<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--exists-action', 'i', '--no-binary', ':all:']' returned non-zero exit status 1 -------------------- >> begin captured logging << -------------------- root: DEBUG: Connecting using Google Application Default Credentials. root: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0813115930-448044.1534161570.448324/pipeline.pb... oauth2client.transport: INFO: Attempting refresh to obtain initial access_token root: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0813115930-448044.1534161570.448324/pipeline.pb root: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0813115930-448044.1534161570.448324/requirements.txt... root: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0813115930-448044.1534161570.448324/requirements.txt root: INFO: Executing command: ['<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--exists-action', 'i', '--no-binary', ':all:'] --------------------- >> end captured logging << --------------------- ---------------------------------------------------------------------- XML: <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/nosetests.xml> ---------------------------------------------------------------------- Ran 14 tests in 873.117s FAILED (errors=1) Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_04_59_37-909708137737926610?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_05_06_20-12949425862417260542?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_04_59_37-18389084043616097374?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_05_06_11-16428788496756542186?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_04_59_39-17556613244496050811?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_04_59_37-1198438949341459674?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_05_06_11-10415261373663772309?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_04_59_37-15703246070004095749?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_04_59_37-6876405194871583298?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_05_06_11-16650152713650764768?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_04_59_37-16986436523297547999?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_05_06_20-17088524865063370468?project=apache-beam-testing. Found: https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-08-13_04_59_37-7600653780860345840?project=apache-beam-testing. > Task :beam-sdks-python:validatesRunnerStreamingTests FAILED :beam-sdks-python:validatesRunnerStreamingTests (Thread[Task worker for ':',5,main]) completed. Took 14 mins 34.846 secs. FAILURE: Build failed with an exception. * Where: Build file '<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build.gradle'> line: 246 * What went wrong: Execution failed for task ':beam-sdks-python:validatesRunnerStreamingTests'. > Process 'command 'sh'' finished with non-zero exit value 1 * Try: Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0. See https://docs.gradle.org/4.8/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 28m 54s 4 actionable tasks: 4 executed Publishing build scan... https://gradle.com/s/rqkamcvtlkquk Build step 'Invoke Gradle script' changed build result to FAILURE Build step 'Invoke Gradle script' marked build as failure
