See <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/1095/display/redirect?page=changes>
Changes: [github] Revert "[BEAM-4780] Updating to DockerJobBundleFactory in ------------------------------------------ [...truncated 82.21 KB...] Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.4.1.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr WARNING:root:Waiting indefinitely for streaming job. WARNING:root:Waiting indefinitely for streaming job. WARNING:root:Waiting indefinitely for streaming job. WARNING:root:Waiting indefinitely for streaming job. WARNING:root:Waiting indefinitely for streaming job. WARNING:root:Waiting indefinitely for streaming job. WARNING:root:Waiting indefinitely for streaming job. <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py>:54: DeprecationWarning: options is deprecated since First stable release. References to <pipeline>.options will not be supported print('Found: %s.' % self.build_console_url(pipeline.options)) WARNING:root:Waiting indefinitely for streaming job. test_par_do_with_multiple_outputs_and_using_yield (apache_beam.transforms.ptransform_test.PTransformTest) ... ok Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.4.1.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr test_as_singleton_with_different_defaults (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.4.1.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr WARNING:root:Waiting indefinitely for streaming job. test_undeclared_outputs (apache_beam.transforms.ptransform_test.PTransformTest) ... ok Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.4.1.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr WARNING:root:Waiting indefinitely for streaming job. test_as_dict_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.4.1.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr test_flatten_multiple_pcollections_having_multiple_consumers (apache_beam.transforms.ptransform_test.PTransformTest) ... ok test_par_do_with_multiple_outputs_and_using_return (apache_beam.transforms.ptransform_test.PTransformTest) ... ok WARNING:root:Waiting indefinitely for streaming job. Collecting pyhamcrest (from -r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/PyHamcrest-1.9.0.tar.gz Collecting mock (from -r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/mock-2.0.0.tar.gz Collecting setuptools (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/setuptools-40.4.1.zip Collecting six (from pyhamcrest->-r postcommit_requirements.txt (line 1)) File was already downloaded /tmp/dataflow-requirements-cache/six-1.11.0.tar.gz Collecting funcsigs>=1 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/funcsigs-1.0.2.tar.gz Collecting pbr>=0.11 (from mock->-r postcommit_requirements.txt (line 2)) File was already downloaded /tmp/dataflow-requirements-cache/pbr-4.2.0.tar.gz Successfully downloaded pyhamcrest mock setuptools six funcsigs pbr WARNING:root:Waiting indefinitely for streaming job. WARNING:root:Waiting indefinitely for streaming job. test_as_list_and_as_dict_side_inputs (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_as_singleton_without_unique_labels (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_empty_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_default_value_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_flattened_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ERROR ====================================================================== ERROR: test_multiple_empty_outputs (apache_beam.transforms.ptransform_test.PTransformTest) ---------------------------------------------------------------------- Traceback (most recent call last): File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/transforms/ptransform_test.py",> line 277, in test_multiple_empty_outputs pipeline.run() File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 104, in run result = super(TestPipeline, self).run(test_runner_api) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 394, in run self.to_runner_api(), self.runner, self._options).run(False) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 407, in run return self.runner.run_pipeline(self) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 50, in run_pipeline self.result = super(TestDataflowRunner, self).run_pipeline(pipeline) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 383, in run_pipeline self.dataflow_client.create_job(self.job), self) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/utils/retry.py",> line 184, in wrapper return fun(*args, **kwargs) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 504, in create_job return self.submit_job_description(job) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/utils/retry.py",> line 184, in wrapper return fun(*args, **kwargs) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 551, in submit_job_description response = self._client.projects_locations_jobs.Create(request) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/clients/dataflow/dataflow_v1b3_client.py",> line 646, in Create config, request, global_params=global_params) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/apitools/base/py/base_api.py",> line 722, in _RunMethod return self.ProcessHttpResponse(method_config, http_response, request) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/apitools/base/py/base_api.py",> line 728, in ProcessHttpResponse self.__ProcessHttpResponse(method_config, http_response, request)) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/apitools/base/py/base_api.py",> line 599, in __ProcessHttpResponse http_response, method_config=method_config, request=request) HttpConflictError: HttpError accessing <https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>: response: <{'status': '409', 'content-length': '318', 'x-xss-protection': '1; mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding': 'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF', '-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Thu, 20 Sep 2018 01:57:55 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type': 'application/json; charset=UTF-8'}>, content <{ "error": { "code": 409, "message": "(f5cbebf399ecd889): The workflow could not be created. Causes: (637e9adddfb0d443): There is already an active job named beamapp-jenkins-0920015745-784722. If you want to submit a second job, try again by setting a different name.", "status": "ALREADY_EXISTS" } } > ====================================================================== ERROR: test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ---------------------------------------------------------------------- Traceback (most recent call last): File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/nose/plugins/multiprocess.py",> line 812, in run test(orig) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/nose/case.py",> line 45, in __call__ return self.run(*arg, **kwarg) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/nose/case.py",> line 133, in run self.runTest(result) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/nose/case.py",> line 151, in runTest test(result) File "/usr/lib/python2.7/unittest/case.py", line 393, in __call__ return self.run(*args, **kwds) File "/usr/lib/python2.7/unittest/case.py", line 329, in run testMethod() File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/transforms/sideinputs_test.py",> line 274, in test_as_list_twice pipeline.run() File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 104, in run result = super(TestPipeline, self).run(test_runner_api) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 394, in run self.to_runner_api(), self.runner, self._options).run(False) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/pipeline.py",> line 407, in run return self.runner.run_pipeline(self) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 68, in run_pipeline self.result.cancel() File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 1175, in cancel self.job_id(), 'JOB_STATE_CANCELLED'): File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/utils/retry.py",> line 184, in wrapper return fun(*args, **kwargs) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 600, in modify_job_state self._client.projects_locations_jobs.Update(request) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/apache_beam/runners/dataflow/internal/clients/dataflow/dataflow_v1b3_client.py",> line 698, in Update config, request, global_params=global_params) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/apitools/base/py/base_api.py",> line 720, in _RunMethod http, http_request, **opts) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/apitools/base/py/http_wrapper.py",> line 346, in MakeRequest check_response_func=check_response_func) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/apitools/base/py/http_wrapper.py",> line 396, in _MakeRequestNoRetry redirections=redirections, connection_type=connection_type) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/oauth2client/transport.py",> line 175, in new_request redirections, connection_type) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/oauth2client/transport.py",> line 282, in request connection_type=connection_type) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1694, in request (response, content) = self._request(conn, authority, uri, request_uri, method, body, headers, redirections, cachekey) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1434, in _request (response, content) = self._conn_request(conn, request_uri, method, body, headers) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/httplib2/__init__.py",> line 1390, in _conn_request response = conn.getresponse() File "/usr/lib/python2.7/httplib.py", line 1136, in getresponse response.begin() File "/usr/lib/python2.7/httplib.py", line 453, in begin version, status, reason = self._read_status() File "/usr/lib/python2.7/httplib.py", line 409, in _read_status line = self.fp.readline(_MAXLINE + 1) File "/usr/lib/python2.7/socket.py", line 480, in readline data = self._sock.recv(self._rbufsize) File "/usr/lib/python2.7/ssl.py", line 756, in recv return self.read(buflen) File "/usr/lib/python2.7/ssl.py", line 643, in read v = self._sslobj.read(len) File "<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build/gradleenv/local/lib/python2.7/site-packages/nose/plugins/multiprocess.py",> line 276, in signalhandler raise TimedOutException() TimedOutException: 'test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest)' ---------------------------------------------------------------------- XML: <https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/nosetests.xml> ---------------------------------------------------------------------- Ran 14 tests in 3001.675s FAILED (errors=2) > Task :beam-sdks-python:validatesRunnerStreamingTests FAILED :beam-sdks-python:validatesRunnerStreamingTests (Thread[Task worker for ':',5,main]) completed. Took 50 mins 3.219 secs. FAILURE: Build failed with an exception. * Where: Build file '<https://builds.apache.org/job/beam_PostCommit_Py_VR_Dataflow/ws/src/sdks/python/build.gradle'> line: 315 * What went wrong: Execution failed for task ':beam-sdks-python:validatesRunnerStreamingTests'. > Process 'command 'sh'' finished with non-zero exit value 1 * Try: Run with --stacktrace option to get the stack trace. Run with --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 5.0. See https://docs.gradle.org/4.8/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 1h 5m 3s 4 actionable tasks: 4 executed Publishing build scan... https://gradle.com/s/qmnxgh4p5obga Build step 'Invoke Gradle script' changed build result to FAILURE Build step 'Invoke Gradle script' marked build as failure
