See 
<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/1721/display/redirect?page=changes>

Changes:

[Kyle Weaver] [BEAM-11848] Fix Docker images list.

[Kyle Weaver] [BEAM-11848] Store Docker images in a variable for consistency.


------------------------------------------
[...truncated 6.90 MB...]
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:30.967Z: 
JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Match into assert_that/Unkey
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:31.005Z: 
JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:31.031Z: 
JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:31.055Z: 
JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:31.106Z: 
JOB_MESSAGE_DEBUG: Assigning stage ids.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:32.216Z: 
JOB_MESSAGE_DEBUG: Starting worker pool setup.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:32.249Z: 
JOB_MESSAGE_BASIC: Starting 1 workers in us-central1-f...
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:32.281Z: 
JOB_MESSAGE_DEBUG: Starting worker pool setup.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:49.858Z: 
JOB_MESSAGE_BASIC: Your project already contains 100 Dataflow-created metric 
descriptors, so new user metrics of the form custom.googleapis.com/* will not 
be created. However, all user metrics are also available in the metric 
dataflow.googleapis.com/job/user_counter. If you rely on the custom metrics, 
you can delete old / unused metric descriptors. See 
https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list
 and 
https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:54.318Z: 
JOB_MESSAGE_DETAILED: Autoscaling: Reduced the number of workers to 0 based on 
low average worker CPU utilization, and the pipeline having sufficiently low 
backlog and keeping up with input rate.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:54.383Z: 
JOB_MESSAGE_BASIC: Worker pool stopped.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:26:54.442Z: 
JOB_MESSAGE_DEBUG: Tearing down pending resources...
INFO:apache_beam.runners.dataflow.dataflow_runner:Job 
2021-03-02_11_14_17-3877378555004050267 is in state JOB_STATE_DONE
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:27:11.927Z: 
JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 1 so that 
the pipeline can catch up with its backlog and keep up with its input rate.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:27:53.040Z: 
JOB_MESSAGE_DETAILED: Workers have started successfully.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:27:53.069Z: 
JOB_MESSAGE_DETAILED: Workers have started successfully.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:34:34.292Z: 
JOB_MESSAGE_DETAILED: Cleaning up.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:34:34.357Z: 
JOB_MESSAGE_DEBUG: Starting worker pool teardown.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:34:34.486Z: 
JOB_MESSAGE_BASIC: Stopping worker pool...
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:34:34.524Z: 
JOB_MESSAGE_DEBUG: Starting worker pool teardown.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:34:34.544Z: 
JOB_MESSAGE_BASIC: Stopping worker pool...
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:35:24.473Z: 
JOB_MESSAGE_DETAILED: Autoscaling: Reduced the number of workers to 0 based on 
low average worker CPU utilization, and the pipeline having sufficiently low 
backlog and keeping up with input rate.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:35:24.569Z: 
JOB_MESSAGE_BASIC: Worker pool stopped.
INFO:apache_beam.runners.dataflow.dataflow_runner:2021-03-02T19:35:24.620Z: 
JOB_MESSAGE_DEBUG: Tearing down pending resources...
INFO:apache_beam.runners.dataflow.dataflow_runner:Job 
2021-03-02_11_26_23-12279737927248559039 is in state JOB_STATE_DONE
test_gbk_many_values 
(apache_beam.runners.portability.fn_api_runner.fn_runner_test.FnApiBasedStateBackedCoderTest)
 ... ok
Test TimestampCombiner with EARLIEST. ... ok
Test TimestampCombiner with LATEST. ... ok
Runs streaming Dataflow job and verifies that user metrics are reported ... ok
test_user_counter_using_pardo (apache_beam.metrics.metric_test.MetricsTest) ... 
ok
test_flatten_one_single_pcollection 
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_dofn_lifecycle 
(apache_beam.transforms.dofn_lifecycle_test.DoFnLifecycleTest) ... ok
test_run_packable_combine_per_key 
(apache_beam.runners.portability.fn_api_runner.translations_test.TranslationsTest)
 ... ok
test_flatten_multiple_pcollections_having_multiple_consumers 
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_flatten_a_flattened_pcollection 
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_flatten_pcollections 
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_multiple_empty_outputs 
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_par_do_with_multiple_outputs_and_using_return 
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_par_do_with_multiple_outputs_and_using_yield 
(apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_combine 
(apache_beam.transforms.combinefn_lifecycle_test.CombineFnLifecycleTest) ... ok
test_combining_value_state 
(apache_beam.transforms.combinefn_lifecycle_test.CombineFnLifecycleTest) ... 
SKIP: BEAM-11793
test_non_liftable_combine 
(apache_beam.transforms.combinefn_lifecycle_test.CombineFnLifecycleTest) ... ok
test_impulse (apache_beam.transforms.ptransform_test.PTransformTest) ... ok
test_as_dict_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... 
ok
test_undeclared_outputs (apache_beam.transforms.ptransform_test.PTransformTest) 
... ok
test_empty_singleton_side_input 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ERROR
test_as_singleton_without_unique_labels 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_default_value_singleton_side_input 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_as_list_twice (apache_beam.transforms.sideinputs_test.SideInputsTest) ... 
ok
test_as_list_and_as_dict_side_inputs 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_as_singleton_with_different_defaults 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_flattened_side_input 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_iterable_side_input 
(apache_beam.transforms.sideinputs_test.SideInputsTest) ... ok
test_reshuffle_preserves_timestamps 
(apache_beam.transforms.util_test.ReshuffleTest) ... ok
test_element_param (apache_beam.pipeline_test.DoFnTest) ... ok
test_key_param (apache_beam.pipeline_test.DoFnTest) ... ok

======================================================================
ERROR: test_empty_singleton_side_input 
(apache_beam.transforms.sideinputs_test.SideInputsTest)
----------------------------------------------------------------------
Traceback (most recent call last):
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/transforms/sideinputs_test.py";,>
 line 164, in test_empty_singleton_side_input
    pipeline.run()
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/testing/test_pipeline.py";,>
 line 110, in run
    result = super(TestPipeline, self).run(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/pipeline.py";,>
 line 527, in run
    return Pipeline.from_runner_api(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/pipeline.py";,>
 line 559, in run
    return self.runner.run_pipeline(self, self._options)
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py";,>
 line 56, in run_pipeline
    self.result = super(TestDataflowRunner,
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py";,>
 line 619, in run_pipeline
    self.dataflow_client.create_job(self.job), self)
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/utils/retry.py";,>
 line 260, in wrapper
    return fun(*args, **kwargs)
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py";,>
 line 655, in create_job
    self.create_job_description(job)
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py";,>
 line 711, in create_job_description
    resources = self._stage_resources(job.proto_pipeline, job.options)
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py";,>
 line 607, in _stage_resources
    staged_resources = resource_stager.stage_job_resources(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/portability/stager.py";,>
 line 384, in stage_job_resources
    self.stage_artifact(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py";,>
 line 975, in stage_artifact
    self._dataflow_application_client._gcs_file_copy(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/utils/retry.py";,>
 line 260, in wrapper
    return fun(*args, **kwargs)
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py";,>
 line 572, in _gcs_file_copy
    self.stage_file(to_folder, to_name, f, total_size=total_size)
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py";,>
 line 633, in stage_file
    response = self._storage_client.objects.Insert(request, upload=upload)
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/apache_beam/io/gcp/internal/clients/storage/storage_v1_client.py";,>
 line 1154, in Insert
    return self._RunMethod(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/base_api.py";,>
 line 714, in _RunMethod
    http_response = upload.InitializeUpload(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/transfer.py";,>
 line 908, in InitializeUpload
    return self.StreamInChunks()
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/transfer.py";,>
 line 1018, in StreamInChunks
    return self.__StreamMedia(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/transfer.py";,>
 line 957, in __StreamMedia
    response = send_func(self.stream.tell())
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/transfer.py";,>
 line 942, in CallSendChunk
    return self.__SendChunk(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/transfer.py";,>
 line 1120, in __SendChunk
    return self.__SendMediaRequest(request, end)
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/transfer.py";,>
 line 1031, in __SendMediaRequest
    response = http_wrapper.MakeRequest(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/http_wrapper.py";,>
 line 348, in MakeRequest
    return _MakeRequestNoRetry(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/apitools/base/py/http_wrapper.py";,>
 line 397, in _MakeRequestNoRetry
    info, content = http.request(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/oauth2client/transport.py";,>
 line 173, in new_request
    resp, content = request(orig_request_method, uri, method, body,
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/oauth2client/transport.py";,>
 line 280, in request
    return http_callable(uri, method=method, body=body, headers=headers,
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/oauth2client/transport.py";,>
 line 173, in new_request
    resp, content = request(orig_request_method, uri, method, body,
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/oauth2client/transport.py";,>
 line 280, in request
    return http_callable(uri, method=method, body=body, headers=headers,
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/httplib2/__init__.py";,>
 line 1982, in request
    (response, content) = self._request(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/httplib2/__init__.py";,>
 line 1650, in _request
    (response, content) = self._conn_request(
  File 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/lib/python3.8/site-packages/httplib2/__init__.py";,>
 line 1589, in _conn_request
    response = conn.getresponse()
  File "/usr/lib/python3.8/http/client.py", line 1347, in getresponse
    response.begin()
  File "/usr/lib/python3.8/http/client.py", line 307, in begin
    version, status, reason = self._read_status()
  File "/usr/lib/python3.8/http/client.py", line 268, in _read_status
    line = str(self.fp.readline(_MAXLINE + 1), "iso-8859-1")
  File "/usr/lib/python3.8/socket.py", line 669, in readinto
    return self._sock.recv_into(b)
  File "/usr/lib/python3.8/ssl.py", line 1241, in recv_into
    return self.read(nbytes, buffer)
  File "/usr/lib/python3.8/ssl.py", line 1099, in read
    return self._sslobj.read(len, buffer)
socket.timeout: The read operation timed out
-------------------- >> begin captured logging << --------------------
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
root: DEBUG: Unhandled type_constraint: Union[]
apache_beam.runners.portability.stager: INFO: Executing command: 
['<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/build/gradleenv/-1734967051/bin/python',>
 '-m', 'pip', 'download', '--dest', '/tmp/dataflow-requirements-cache', '-r', 
'postcommit_requirements.txt', '--exists-action', 'i', '--no-binary', ':all:']
apache_beam.runners.portability.stager: INFO: Copying Beam SDK 
"<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/build/apache-beam.tar.gz";>
 to staging location.
root: WARNING: Make sure that locally built Python SDK docker image has Python 
3.8 interpreter.
root: INFO: Default Python SDK image for environment is 
apache/beam_python3.8_sdk:2.29.0.dev
root: INFO: Using provided Python SDK container image: 
gcr.io/cloud-dataflow/v1beta3/python38-fnapi:beam-master-20210216
root: INFO: Python SDK container image set to 
"gcr.io/cloud-dataflow/v1beta3/python38-fnapi:beam-master-20210216" for Docker 
environment
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pipeline.pb...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pipeline.pb
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/requirements.txt...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/requirements.txt
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/parameterized-0.7.4.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/parameterized-0.7.4.tar.gz
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/six-1.14.0.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/six-1.14.0.tar.gz
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pbr-5.5.1.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pbr-5.5.1.tar.gz
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/mock-2.0.0.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/mock-2.0.0.tar.gz
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/parameterized-0.7.5.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/parameterized-0.7.5.tar.gz
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/six-1.15.0.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/six-1.15.0.tar.gz
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/funcsigs-1.0.2.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/funcsigs-1.0.2.tar.gz
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pbr-5.4.5.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pbr-5.4.5.tar.gz
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pbr-5.5.0.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/pbr-5.5.0.tar.gz
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/PyHamcrest-1.10.1.tar.gz...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/PyHamcrest-1.10.1.tar.gz
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/dataflow_python_sdk.tar...
apache_beam.runners.dataflow.internal.apiclient: INFO: Completed GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/dataflow_python_sdk.tar
 in 0 seconds.
apache_beam.runners.dataflow.internal.apiclient: INFO: Starting GCS upload to 
gs://temp-storage-for-end-to-end-tests/staging-it/beamapp-jenkins-0302190829-715299.1614712109.715453/dataflow-worker.jar...
root: DEBUG: Caught socket error, retrying: The read operation timed out
root: DEBUG: Retrying request to url 
https://www.googleapis.com/resumable/upload/storage/v1/b/temp-storage-for-end-to-end-tests/o?alt=json&name=staging-it%2Fbeamapp-jenkins-0302190829-715299.1614712109.715453%2Fdataflow-worker.jar&uploadType=resumable&upload_id=ABg5-UyCaBvGqIigg4oUBdGTvSguosdAVwkpTem0uGSUVG3dXKcrj1fpJRlmLqJsNsIXuTU8rsPJoh6t0Xu5H09XF0k
 after exception The read operation timed out
root: DEBUG: Caught socket error, retrying: The read operation timed out
root: DEBUG: Retrying request to url 
https://www.googleapis.com/resumable/upload/storage/v1/b/temp-storage-for-end-to-end-tests/o?alt=json&name=staging-it%2Fbeamapp-jenkins-0302190829-715299.1614712109.715453%2Fdataflow-worker.jar&uploadType=resumable&upload_id=ABg5-UyCaBvGqIigg4oUBdGTvSguosdAVwkpTem0uGSUVG3dXKcrj1fpJRlmLqJsNsIXuTU8rsPJoh6t0Xu5H09XF0k
 after exception The read operation timed out
root: DEBUG: Caught socket error, retrying: The read operation timed out
root: DEBUG: Retrying request to url 
https://www.googleapis.com/resumable/upload/storage/v1/b/temp-storage-for-end-to-end-tests/o?alt=json&name=staging-it%2Fbeamapp-jenkins-0302190829-715299.1614712109.715453%2Fdataflow-worker.jar&uploadType=resumable&upload_id=ABg5-UyCaBvGqIigg4oUBdGTvSguosdAVwkpTem0uGSUVG3dXKcrj1fpJRlmLqJsNsIXuTU8rsPJoh6t0Xu5H09XF0k
 after exception The read operation timed out
root: DEBUG: Caught socket error, retrying: The read operation timed out
root: DEBUG: Retrying request to url 
https://www.googleapis.com/resumable/upload/storage/v1/b/temp-storage-for-end-to-end-tests/o?alt=json&name=staging-it%2Fbeamapp-jenkins-0302190829-715299.1614712109.715453%2Fdataflow-worker.jar&uploadType=resumable&upload_id=ABg5-UyCaBvGqIigg4oUBdGTvSguosdAVwkpTem0uGSUVG3dXKcrj1fpJRlmLqJsNsIXuTU8rsPJoh6t0Xu5H09XF0k
 after exception The read operation timed out
--------------------- >> end captured logging << ---------------------
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_14-13943366067081075099?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_57_33-16623453239515669342?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_14_17-3877378555004050267?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_14-4419599924360388386?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_56_15-14842964048563698116?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_05_39-8455836401246472975?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_14-15943533369302786209?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_57_43-10617591512122183475?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_07_17-11864484398572268062?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_20-1830250065812661291?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_57_50-5368258546993500429?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_07_24-5702133584715318823?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_16_53-15367884531952662520?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_26_23-12279737927248559039?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_16-17010943174437930586?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_57_41-14061954277959098189?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_06_45-6582233596818101775?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_15-10535986145323142695?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_57_44-1980392533697876033?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_07_29-408221025644516323?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_17-5659882029872609228?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_59_40-15030047714684468717?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_09_15-16212178204408819893?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_48_20-428486447617181518?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_10_58_50-14707847996887189360?project=apache-beam-testing
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2021-03-02_11_13_31-9008142121648237298?project=apache-beam-testing

----------------------------------------------------------------------
XML: nosetests-validatesRunnerStreamingTests-df-py38.xml
----------------------------------------------------------------------
XML: 
<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 31 tests in 2919.361s

FAILED (SKIP=1, errors=1)

> Task :sdks:python:test-suites:dataflow:py38:validatesRunnerStreamingTests 
> FAILED

FAILURE: Build failed with an exception.

* Where:
Script 
'<https://ci-beam.apache.org/job/beam_PostCommit_Py_VR_Dataflow_V2/ws/src/sdks/python/test-suites/dataflow/common.gradle'>
 line: 175

* What went wrong:
Execution failed for task 
':sdks:python:test-suites:dataflow:py38:validatesRunnerStreamingTests'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug 
option to get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 7.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See 
https://docs.gradle.org/6.8/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1h 35m 4s
78 actionable tasks: 50 executed, 28 from cache

Publishing build scan...
https://gradle.com/s/vkc473ziirhws

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to