See <https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/1721/display/redirect?page=changes>
Changes: [Kyle Weaver] [BEAM-11848] Fix Docker images list. [Kyle Weaver] [BEAM-11848] Store Docker images in a variable for consistency. ------------------------------------------ [...truncated 6.90 MB...] INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:30.967Z: JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Match into assert_that/Unkey INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:31.005Z: JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:31.031Z: JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:31.055Z: JOB_MESSAGE_DEBUG: Adding workflow start and stop steps. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:31.106Z: JOB_MESSAGE_DEBUG: Assigning stage ids. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:32.216Z: JOB_MESSAGE_DEBUG: Starting worker pool setup. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:32.249Z: JOB_MESSAGE_BASIC: Starting 1 workers in us-central1-f... INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:32.281Z: JOB_MESSAGE_DEBUG: Starting worker pool setup. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:49.858Z: JOB_MESSAGE_BASIC: Your project already contains 100 Dataflow-created metric descriptors, so new user metrics of the form custom.googleapis.com/* will not be created. However, all user metrics are also available in the metric dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, you can delete old / unused metric descriptors. See https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list and https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:54.318Z: JOB_MESSAGE_DETAILED: Autoscaling: Reduced the number of workers to 0 based on low average worker CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:54.383Z: JOB_MESSAGE_BASIC: Worker pool stopped. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:54.442Z: JOB_MESSAGE_DEBUG: Tearing down pending resources... INFO:apache_beam.runners.dataflow.dataflow_runner:Job 2021-03-02_11_14_17-3877378555004050267 is in state JOB_STATE_DONE INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:27:11.927Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 1 so that the pipeline can catch up with its backlog and keep up with its input rate. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:27:53.040Z: JOB_MESSAGE_DETAILED: Workers have started successfully. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:27:53.069Z: JOB_MESSAGE_DETAILED: Workers have started successfully. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:34:34.292Z: JOB_MESSAGE_DETAILED: Cleaning up. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:34:34.357Z: JOB_MESSAGE_DEBUG: Starting worker pool teardown. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:34:34.486Z: JOB_MESSAGE_BASIC: Stopping worker pool... INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:34:34.524Z: JOB_MESSAGE_DEBUG: Starting worker pool teardown. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:34:34.544Z: JOB_MESSAGE_BASIC: Stopping worker pool... INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:35:24.473Z: JOB_MESSAGE_DETAILED: Autoscaling: Reduced the number of workers to 0 based on low average worker CPU utilization, and the pipeline having sufficiently low backlog and keeping up with input rate. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:35:24.569Z: JOB_MESSAGE_BASIC: Worker pool stopped. INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:35:24.620Z: JOB_MESSAGE_DEBUG: Tearing down pending resources... INFO:apache_beam.runners.dataflow.dataflow_runner:Job 2021-03-02_11_26_23-12279737927248559039 is in state JOB_STATE_DONE test_gbk_many_values (apache_beam.runners.portability.fn_api_runner.fn_runner_test.FnApiBasedStateBackedCoderTest) ... ok Test TimestampCombiner with EARLIEST. ... ok Test TimestampCombiner with LATEST. ... ok Runs streaming Dataflow job and verifies that user metrics are reported ... ok test_user_counter_using_pardo (apache_beam.metrics.metric_test.MetricsTest) ... ok test_flatten_one_single_pcollection (apache_beam.transforms.ptransform_test.PTransformTest) ... ok test_dofn_lifecycle (apache_beam.transforms.dofn_lifecycle_test.DoFnLifecycleTest) ... ok test_run_packable_combine_per_key (apache_beam.runners.portability.fn_api_runner.translations_test.TranslationsTest) ... ok test_flatten_multiple_pcollections_having_multiple_consumers (apache_beam.transforms.ptransform_test.PTransformTest) ... ok test_flatten_a_flattened_pcollection (apache_beam.transforms.ptransform_test.PTransformTest) ... ok test_flatten_pcollections (apache_beam.transforms.ptransform_test.PTransformTest) ... ok test_multiple_empty_outputs (apache_beam.transforms.ptransform_test.PTransformTest) ... ok test_par_do_with_multiple_outputs_and_using_return (apache_beam.transforms.ptransform_test.PTransformTest) ... ok test_par_do_with_multiple_outputs_and_using_yield (apache_beam.transforms.ptransform_test.PTransformTest) ... ok test_combine (apache_beam.transforms.combinefn_lifecycle_test.CombineFnLifecycleTest) ... ok test_combining_value_state (apache_beam.transforms.combinefn_lifecycle_test.CombineFnLifecycleTest) ... SKIP: BEAM-11793 test_non_liftable_combine (apache_beam.transforms.combinefn_lifecycle_test.CombineFnLifecycleTest) ... ok test_impulse (apache_beam.transforms.ptransform_test.PTransformTest) ... ok test_as_dict_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_undeclared_outputs (apache_beam.transforms.ptransform_test.PTransformTest) ... ok test_empty_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ERROR test_as_singleton_without_unique_labels (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_default_value_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_as_list_and_as_dict_side_inputs (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_as_singleton_with_different_defaults (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_flattened_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_iterable_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok test_reshuffle_preserves_timestamps (apache_beam.transforms.util_test.ReshuffleTest) ... ok test_element_param (apache_beam.pipeline_test.DoFnTest) ... ok test_key_param (apache_beam.pipeline_test.DoFnTest) ... ok ====================================================================== ERROR: test_empty_singleton_side_input (apache_beam.transforms.sideinputs_test.SideInputsTest) ---------------------------------------------------------------------- Traceback (most recent call last): File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/transforms/sideinputs_test.py",> line 164, in test_empty_singleton_side_input pipeline.run() File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/testing/test_pipeline.py",> line 110, in run result = super(TestPipeline, self).run( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/pipeline.py",> line 527, in run return Pipeline.from_runner_api( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/pipeline.py",> line 559, in run return self.runner.run_pipeline(self, self._options) File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py",> line 56, in run_pipeline self.result = super(TestDataflowRunner, File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py",> line 619, in run_pipeline self.dataflow_client.create_job(self.job), self) File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/utils/retry.py",> line 260, in wrapper return fun(*args, **kwargs) File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 655, in create_job self.create_job_description(job) File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 711, in create_job_description resources = self._stage_resources(job.proto_pipeline, job.options) File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 607, in _stage_resources staged_resources = resource_stager.stage_job_resources( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/portability/stager.py",> line 384, in stage_job_resources self.stage_artifact( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 975, in stage_artifact self._dataflow_application_client._gcs_file_copy( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/utils/retry.py",> line 260, in wrapper return fun(*args, **kwargs) File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 572, in _gcs_file_copy self.stage_file(to_folder, to_name, f, total_size=total_size) File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py",> line 633, in stage_file response = self._storage_client.objects.Insert(request, upload=upload) File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/io/gcp/internal/clients/storage/storage_v1_client.py",> line 1154, in Insert return self._RunMethod( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/base_api.py",> line 714, in _RunMethod http_response = upload.InitializeUpload( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/transfer.py",> line 908, in InitializeUpload return self.StreamInChunks() File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/transfer.py",> line 1018, in StreamInChunks return self.__StreamMedia( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/transfer.py",> line 957, in __StreamMedia response = send_func(self.stream.tell()) File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/transfer.py",> line 942, in CallSendChunk return self.__SendChunk( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/transfer.py",> line 1120, in __SendChunk return self.__SendMediaRequest(request, end) File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/transfer.py",> line 1031, in __SendMediaRequest response = http_wrapper.MakeRequest( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/http_wrapper.py",> line 348, in MakeRequest return _MakeRequestNoRetry( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/http_wrapper.py",> line 397, in _MakeRequestNoRetry info, content = http.request( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/oauth2client/transport.py",> line 173, in new_request resp, content = request(orig_request_method, uri, method, body, File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/oauth2client/transport.py",> line 280, in request return http_callable(uri, method=method, body=body, headers=headers, File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/oauth2client/transport.py",> line 173, in new_request resp, content = request(orig_request_method, uri, method, body, File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/oauth2client/transport.py",> line 280, in request return http_callable(uri, method=method, body=body, headers=headers, File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/httplib2/__init__.py",> line 1982, in request (response, content) = self._request( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/httplib2/__init__.py",> line 1650, in _request (response, content) = self._conn_request( File "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/httplib2/__init__.py",> line 1589, in _conn_request response = conn.getresponse() File "/usr/lib/python3.8/http/client.py", line 1347, in getresponse response.begin() File "/usr/lib/python3.8/http/client.py", line 307, in begin version, status, reason = self._read_status() File "/usr/lib/python3.8/http/client.py", line 268, in _read_status line = str(self.fp.readline(_MAXLINE + 1), "iso-8859-1") File "/usr/lib/python3.8/socket.py", line 669, in readinto return self._sock.recv_into(b) File "/usr/lib/python3.8/ssl.py", line 1241, in recv_into return self.read(nbytes, buffer) File "/usr/lib/python3.8/ssl.py", line 1099, in read return self._sslobj.read(len, buffer) socket.timeout: The read operation timed out -------------------- >> begin captured logging << -------------------- root: DEBUG: Unhandled type_constraint: Union[] root: DEBUG: Unhandled type_constraint: Union[] root: DEBUG: Unhandled type_constraint: Union[] root: DEBUG: Unhandled type_constraint: Union[] apache_beam.runners.portability.stager: INFO: Executing command: ['<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/bin/python',> '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 'postcommit_requirements.txt', '--exists-action', 'i', '--no-binary', ':all:'] apache_beam.runners.portability.stager: INFO: Copying Beam SDK "<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/build/apache-beam.tar.gz"> to staging location. root: WARNING: Make sure that locally built Python SDK docker image has Python 3.8 interpreter. root: INFO: Default Python SDK image for environment is apache/beam_python3.8_sdk:2.29.0.dev root: INFO: Using provided Python SDK container image: gcr.io/cloud-dataflow/v1beta3/python38-fnapi:beam-master-20210216 root: INFO: Python SDK container image set to "gcr.io/cloud-dataflow/v1beta3/python38-fnapi:beam-master-20210216" for Docker environment apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pipeline.pb... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pipeline.pb in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/requirements.txt... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/requirements.txt in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/parameterized-0.7.4.tar.gz... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/parameterized-0.7.4.tar.gz in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/six-1.14.0.tar.gz... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/six-1.14.0.tar.gz in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pbr-5.5.1.tar.gz... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pbr-5.5.1.tar.gz in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/mock-2.0.0.tar.gz... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/mock-2.0.0.tar.gz in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/parameterized-0.7.5.tar.gz... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/parameterized-0.7.5.tar.gz in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/six-1.15.0.tar.gz... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/six-1.15.0.tar.gz in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/funcsigs-1.0.2.tar.gz... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/funcsigs-1.0.2.tar.gz in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pbr-5.4.5.tar.gz... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pbr-5.4.5.tar.gz in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pbr-5.5.0.tar.gz... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pbr-5.5.0.tar.gz in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/PyHamcrest-1.10.1.tar.gz... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/PyHamcrest-1.10.1.tar.gz in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/dataflow_python_sdk.tar... apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/dataflow_python_sdk.tar in 0 seconds. apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/dataflow-worker.jar... root: DEBUG: Caught socket error, retrying: The read operation timed out root: DEBUG: Retrying request to url https://www.googleapis.com/resumable/upload/storage/v1/b/temp-storage-for-end-to-end-tests/o?alt=json&name=staging-it%2Fbeamapp-jenkins-0302190829-715299.1614712109.715453%2Fdataflow-worker.jar&uploadType=resumable&upload_id=ABg5-UyCaBvGqIigg4oUBdGTvSguosdAVwkpTem0uGSUVG3dXKcrj1fpJRlmLqJsNsIXuTU8rsPJoh6t0Xu5H09XF0k after exception The read operation timed out root: DEBUG: Caught socket error, retrying: The read operation timed out root: DEBUG: Retrying request to url https://www.googleapis.com/resumable/upload/storage/v1/b/temp-storage-for-end-to-end-tests/o?alt=json&name=staging-it%2Fbeamapp-jenkins-0302190829-715299.1614712109.715453%2Fdataflow-worker.jar&uploadType=resumable&upload_id=ABg5-UyCaBvGqIigg4oUBdGTvSguosdAVwkpTem0uGSUVG3dXKcrj1fpJRlmLqJsNsIXuTU8rsPJoh6t0Xu5H09XF0k after exception The read operation timed out root: DEBUG: Caught socket error, retrying: The read operation timed out root: DEBUG: Retrying request to url https://www.googleapis.com/resumable/upload/storage/v1/b/temp-storage-for-end-to-end-tests/o?alt=json&name=staging-it%2Fbeamapp-jenkins-0302190829-715299.1614712109.715453%2Fdataflow-worker.jar&uploadType=resumable&upload_id=ABg5-UyCaBvGqIigg4oUBdGTvSguosdAVwkpTem0uGSUVG3dXKcrj1fpJRlmLqJsNsIXuTU8rsPJoh6t0Xu5H09XF0k after exception The read operation timed out root: DEBUG: Caught socket error, retrying: The read operation timed out root: DEBUG: Retrying request to url https://www.googleapis.com/resumable/upload/storage/v1/b/temp-storage-for-end-to-end-tests/o?alt=json&name=staging-it%2Fbeamapp-jenkins-0302190829-715299.1614712109.715453%2Fdataflow-worker.jar&uploadType=resumable&upload_id=ABg5-UyCaBvGqIigg4oUBdGTvSguosdAVwkpTem0uGSUVG3dXKcrj1fpJRlmLqJsNsIXuTU8rsPJoh6t0Xu5H09XF0k after exception The read operation timed out --------------------- >> end captured logging << --------------------- Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_14-13943366067081075099?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_57_33-16623453239515669342?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_14_17-3877378555004050267?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_14-4419599924360388386?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_56_15-14842964048563698116?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_05_39-8455836401246472975?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_14-15943533369302786209?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_57_43-10617591512122183475?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_07_17-11864484398572268062?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_20-1830250065812661291?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_57_50-5368258546993500429?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_07_24-5702133584715318823?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_16_53-15367884531952662520?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_26_23-12279737927248559039?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_16-17010943174437930586?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_57_41-14061954277959098189?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_06_45-6582233596818101775?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_15-10535986145323142695?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_57_44-1980392533697876033?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_07_29-408221025644516323?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_17-5659882029872609228?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_59_40-15030047714684468717?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_09_15-16212178204408819893?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_20-428486447617181518?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_58_50-14707847996887189360?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_13_31-9008142121648237298?project=apache-beam-testing ---------------------------------------------------------------------- XML: nosetests-validatesRunnerStreamingTests-df-py38.xml ---------------------------------------------------------------------- XML: <https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/nosetests.xml> ---------------------------------------------------------------------- Ran 31 tests in 2919.361s FAILED (SKIP=1, errors=1) > Task :sdks:python:test-suites:dataflow:py38:validatesRunnerStreamingTests > FAILED FAILURE: Build failed with an exception. * Where: Script '<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 175 * What went wrong: Execution failed for task ':sdks:python:test-suites:dataflow:py38:validatesRunnerStreamingTests'. > Process 'command 'sh'' finished with non-zero exit value 1 * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.8/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 1h 35m 4s 78 actionable tasks: 50 executed, 28 from cache Publishing build scan... https://gradle.com/s/vkc473ziirhws Build step 'Invoke Gradle script' changed build result to FAILURE Build step 'Invoke Gradle script' marked build as failure --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
