See 
<https://builds.apache.org/job/beam_PostCommit_Python2/1097/display/redirect?page=changes>

Changes:

[piotr.szczepanik] [BEAM-8819] Fix AvroCoder serialisation by introduction of

[piotr.szczepanik] Added missing license header for AvroGenericCoder

[piotr.szczepanik] Fixed code style violations

[piotr.szczepanik] Fixed missing AvroCoder -> AvroGenericCoder in python tests


------------------------------------------
[...truncated 2.18 MB...]
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "assert_that/Match.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s22"
        }, 
        "serialized_fn": "<string of 1340 bytes>", 
        "user_name": "assert_that/Match"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
apache_beam.runners.dataflow.internal.apiclient: INFO: Create job: <Job
 createTime: u'2019-11-28T14:08:22.797195Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2019-11-28_06_08_20-5333801521925908286'
 location: u'us-central1'
 name: u'beamapp-jenkins-1128140144-236709'
 projectId: u'apache-beam-testing'
 stageStates: []
 startTime: u'2019-11-28T14:08:22.797195Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
apache_beam.runners.dataflow.internal.apiclient: INFO: Created job with id: 
[2019-11-28_06_08_20-5333801521925908286]
apache_beam.runners.dataflow.internal.apiclient: INFO: To access the Dataflow 
monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-11-28_06_08_20-5333801521925908286?project=apache-beam-testing
apache_beam.runners.dataflow.dataflow_runner: INFO: Job 
2019-11-28_06_08_20-5333801521925908286 is in state JOB_STATE_RUNNING
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:20.539Z: 
JOB_MESSAGE_DETAILED: Autoscaling is enabled for job 
2019-11-28_06_08_20-5333801521925908286. The number of workers will be between 
1 and 1000.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:20.539Z: 
JOB_MESSAGE_DETAILED: Autoscaling was automatically enabled for job 
2019-11-28_06_08_20-5333801521925908286.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:24.461Z: 
JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service 
Account.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:25.481Z: 
JOB_MESSAGE_BASIC: Worker configuration: n1-standard-1 in us-central1-f.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.258Z: 
JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.285Z: 
JOB_MESSAGE_DEBUG: Combiner lifting skipped for step 
assert_that/Group/GroupByKey: GroupByKey not followed by a combiner.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.333Z: 
JOB_MESSAGE_DEBUG: Combiner lifting skipped for step read from 
datastore/GroupByKey: GroupByKey not followed by a combiner.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.360Z: 
JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into optimizable parts.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.382Z: 
JOB_MESSAGE_DETAILED: Lifting ValueCombiningMappingFns into 
MergeBucketsMappingFns
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.473Z: 
JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.503Z: 
JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.521Z: 
JOB_MESSAGE_DETAILED: Fusing consumer read from datastore/SplitQuery into read 
from datastore/UserQuery/Read
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.544Z: 
JOB_MESSAGE_DETAILED: Fusing consumer read from datastore/GroupByKey/Reify into 
read from datastore/SplitQuery
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.568Z: 
JOB_MESSAGE_DETAILED: Fusing consumer read from datastore/GroupByKey/Write into 
read from datastore/GroupByKey/Reify
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.593Z: 
JOB_MESSAGE_DETAILED: Fusing consumer read from 
datastore/GroupByKey/GroupByWindow into read from datastore/GroupByKey/Read
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.618Z: 
JOB_MESSAGE_DETAILED: Fusing consumer read from datastore/Values into read from 
datastore/GroupByKey/GroupByWindow
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.641Z: 
JOB_MESSAGE_DETAILED: Fusing consumer read from datastore/Flatten into read 
from datastore/Values
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.667Z: 
JOB_MESSAGE_DETAILED: Fusing consumer read from datastore/Read into read from 
datastore/Flatten
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.689Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
Globally/CombineGlobally(CountCombineFn)/KeyWithVoid into read from 
datastore/Read
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.717Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
Globally/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey+Globally/CombineGlobally(CountCombineFn)/CombinePerKey/Combine/Partial
 into Globally/CombineGlobally(CountCombineFn)/KeyWithVoid
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.740Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
Globally/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Reify into 
Globally/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey+Globally/CombineGlobally(CountCombineFn)/CombinePerKey/Combine/Partial
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.766Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
Globally/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Write into 
Globally/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Reify
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.787Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
Globally/CombineGlobally(CountCombineFn)/CombinePerKey/Combine into 
Globally/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Read
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.815Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
Globally/CombineGlobally(CountCombineFn)/CombinePerKey/Combine/Extract into 
Globally/CombineGlobally(CountCombineFn)/CombinePerKey/Combine
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.837Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
Globally/CombineGlobally(CountCombineFn)/UnKey into 
Globally/CombineGlobally(CountCombineFn)/CombinePerKey/Combine/Extract
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.859Z: 
JOB_MESSAGE_DETAILED: Unzipping flatten s19 for input s17.out
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.880Z: 
JOB_MESSAGE_DETAILED: Fusing unzipped copy of 
assert_that/Group/GroupByKey/Reify, through flatten assert_that/Group/Flatten, 
into producer assert_that/Group/pair_with_0
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.907Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
assert_that/Group/GroupByKey/GroupByWindow into 
assert_that/Group/GroupByKey/Read
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.932Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
assert_that/Group/Map(_merge_tagged_vals_under_key) into 
assert_that/Group/GroupByKey/GroupByWindow
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.955Z: 
JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Unkey into 
assert_that/Group/Map(_merge_tagged_vals_under_key)
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.978Z: 
JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Match into assert_that/Unkey
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:26.999Z: 
JOB_MESSAGE_DETAILED: Unzipping flatten s19-u40 for input s20-reify-value9-c38
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.027Z: 
JOB_MESSAGE_DETAILED: Fusing unzipped copy of 
assert_that/Group/GroupByKey/Write, through flatten 
assert_that/Group/Flatten/Unzipped-1, into producer 
assert_that/Group/GroupByKey/Reify
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.049Z: 
JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/pair_with_0 into 
assert_that/Create/Read
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.076Z: 
JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/GroupByKey/Reify into 
assert_that/Group/pair_with_1
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.096Z: 
JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/GroupByKey/Write into 
assert_that/Group/GroupByKey/Reify
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.120Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
Globally/CombineGlobally(CountCombineFn)/InjectDefault/InjectDefault into 
Globally/CombineGlobally(CountCombineFn)/DoOnce/Read
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.141Z: 
JOB_MESSAGE_DETAILED: Fusing consumer assert_that/WindowInto(WindowIntoFn) into 
Globally/CombineGlobally(CountCombineFn)/InjectDefault/InjectDefault
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.165Z: 
JOB_MESSAGE_DETAILED: Fusing consumer assert_that/ToVoidKey into 
assert_that/WindowInto(WindowIntoFn)
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.189Z: 
JOB_MESSAGE_DETAILED: Fusing consumer assert_that/Group/pair_with_1 into 
assert_that/ToVoidKey
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.215Z: 
JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.236Z: 
JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.263Z: 
JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.287Z: 
JOB_MESSAGE_DEBUG: Assigning stage ids.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.403Z: 
JOB_MESSAGE_DEBUG: Executing wait step start51
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.450Z: 
JOB_MESSAGE_BASIC: Executing operation assert_that/Group/GroupByKey/Create
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.480Z: 
JOB_MESSAGE_DEBUG: Starting worker pool setup.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.480Z: 
JOB_MESSAGE_BASIC: Executing operation read from datastore/GroupByKey/Create
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.501Z: 
JOB_MESSAGE_BASIC: Executing operation 
Globally/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Create
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.505Z: 
JOB_MESSAGE_BASIC: Starting 1 workers in us-central1-f...
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.534Z: 
JOB_MESSAGE_BASIC: Finished operation assert_that/Group/GroupByKey/Create
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.548Z: 
JOB_MESSAGE_BASIC: Finished operation read from datastore/GroupByKey/Create
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.548Z: 
JOB_MESSAGE_BASIC: Finished operation 
Globally/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Create
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.586Z: 
JOB_MESSAGE_DEBUG: Value "assert_that/Group/GroupByKey/Session" materialized.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.610Z: 
JOB_MESSAGE_DEBUG: Value "read from datastore/GroupByKey/Session" materialized.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.639Z: 
JOB_MESSAGE_DEBUG: Value 
"Globally/CombineGlobally(CountCombineFn)/CombinePerKey/GroupByKey/Session" 
materialized.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.666Z: 
JOB_MESSAGE_BASIC: Executing operation 
assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:27.692Z: 
JOB_MESSAGE_BASIC: Executing operation read from datastore/UserQuery/Read+read 
from datastore/SplitQuery+read from datastore/GroupByKey/Reify+read from 
datastore/GroupByKey/Write
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:08:57.507Z: 
JOB_MESSAGE_WARNING: Your project already contains 100 Dataflow-created metric 
descriptors and Stackdriver will not create new Dataflow custom metrics for 
this job. Each unique user-defined metric name (independent of the DoFn in 
which it is defined) produces a new metric descriptor. To delete old / unused 
metric descriptors see 
https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list
 and 
https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:09:01.287Z: 
JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 1 based on 
the rate of progress in the currently running step(s).
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:10:07.903Z: 
JOB_MESSAGE_DETAILED: Workers have started successfully.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:10:07.927Z: 
JOB_MESSAGE_DETAILED: Workers have started successfully.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:11:29.190Z: 
JOB_MESSAGE_ERROR: Workflow failed.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:11:29.243Z: 
JOB_MESSAGE_BASIC: Finished operation 
assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:11:29.243Z: 
JOB_MESSAGE_BASIC: Finished operation read from datastore/UserQuery/Read+read 
from datastore/SplitQuery+read from datastore/GroupByKey/Reify+read from 
datastore/GroupByKey/Write
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:11:29.353Z: 
JOB_MESSAGE_DETAILED: Cleaning up.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:11:29.417Z: 
JOB_MESSAGE_DEBUG: Starting worker pool teardown.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:11:29.448Z: 
JOB_MESSAGE_BASIC: Stopping worker pool...
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:13:34.066Z: 
JOB_MESSAGE_DETAILED: Autoscaling: Resized worker pool from 1 to 0.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:13:34.101Z: 
JOB_MESSAGE_BASIC: Worker pool stopped.
apache_beam.runners.dataflow.dataflow_runner: INFO: 2019-11-28T14:13:34.127Z: 
JOB_MESSAGE_DEBUG: Tearing down pending resources...
apache_beam.runners.dataflow.dataflow_runner: INFO: Job 
2019-11-28_06_08_20-5333801521925908286 is in state JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

======================================================================
ERROR: test_multiple_destinations_transform 
(apache_beam.io.gcp.bigquery_test.BigQueryStreamingInsertTransformIntegrationTests)
----------------------------------------------------------------------
Traceback (most recent call last):
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/nose/plugins/multiprocess.py";,>
 line 812, in run
    test(orig)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/nose/case.py";,>
 line 45, in __call__
    return self.run(*arg, **kwarg)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/nose/case.py";,>
 line 133, in run
    self.runTest(result)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/nose/case.py";,>
 line 151, in runTest
    test(result)
  File "/usr/lib/python2.7/unittest/case.py", line 393, in __call__
    return self.run(*args, **kwds)
  File "/usr/lib/python2.7/unittest/case.py", line 329, in run
    testMethod()
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/io/gcp/bigquery_test.py";,>
 line 740, in test_multiple_destinations_transform
    equal_to([(full_output_table_1, bad_record)]))
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/pipeline.py";,>
 line 436, in __exit__
    self.run().wait_until_finish()
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/pipeline.py";,>
 line 416, in run
    self._options).run(False)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/pipeline.py";,>
 line 429, in run
    return self.runner.run_pipeline(self, self._options)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/runners/dataflow/test_dataflow_runner.py";,>
 line 73, in run_pipeline
    self.result.cancel()
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py";,>
 line 1464, in cancel
    return self.state
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py";,>
 line 1404, in state
    self._update_job()
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/runners/dataflow/dataflow_runner.py";,>
 line 1360, in _update_job
    self._job = self._runner.dataflow_client.get_job(self.job_id())
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/utils/retry.py";,>
 line 209, in wrapper
    return fun(*args, **kwargs)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/apiclient.py";,>
 line 673, in get_job
    response = self._client.projects_locations_jobs.Get(request)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/apache_beam/runners/dataflow/internal/clients/dataflow/dataflow_v1b3_client.py";,>
 line 661, in Get
    config, request, global_params=global_params)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/apitools/base/py/base_api.py";,>
 line 729, in _RunMethod
    http, http_request, **opts)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/apitools/base/py/http_wrapper.py";,>
 line 346, in MakeRequest
    check_response_func=check_response_func)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/apitools/base/py/http_wrapper.py";,>
 line 396, in _MakeRequestNoRetry
    redirections=redirections, connection_type=connection_type)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/oauth2client/transport.py";,>
 line 169, in new_request
    redirections, connection_type)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/oauth2client/transport.py";,>
 line 169, in new_request
    redirections, connection_type)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/httplib2/__init__.py";,>
 line 2133, in request
    cachekey,
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/httplib2/__init__.py";,>
 line 1796, in _request
    conn, request_uri, method, body, headers
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/httplib2/__init__.py";,>
 line 1737, in _conn_request
    response = conn.getresponse()
  File "/usr/lib/python2.7/httplib.py", line 1152, in getresponse
    response.begin()
  File "/usr/lib/python2.7/httplib.py", line 463, in begin
    version, status, reason = self._read_status()
  File "/usr/lib/python2.7/httplib.py", line 419, in _read_status
    line = self.fp.readline(_MAXLINE + 1)
  File "/usr/lib/python2.7/socket.py", line 480, in readline
    data = self._sock.recv(self._rbufsize)
  File "/usr/lib/python2.7/ssl.py", line 756, in recv
    return self.read(buflen)
  File "/usr/lib/python2.7/ssl.py", line 643, in read
    v = self._sslobj.read(len)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/build/gradleenv/-194514014/local/lib/python2.7/site-packages/nose/plugins/multiprocess.py";,>
 line 276, in signalhandler
    raise TimedOutException()
TimedOutException: 'test_multiple_destinations_transform 
(apache_beam.io.gcp.bigquery_test.BigQueryStreamingInsertTransformIntegrationTests)'

----------------------------------------------------------------------
XML: nosetests-postCommitIT-df.xml
----------------------------------------------------------------------
XML: 
<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 45 tests in 5465.254s

FAILED (SKIP=5, errors=3)

> Task :sdks:python:test-suites:dataflow:py2:postCommitIT FAILED

FAILURE: Build completed with 2 failures.

1: Task failed with an exception.
-----------
* Where:
Build file 
'<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/test-suites/direct/py2/build.gradle'>
 line: 70

* What went wrong:
Execution failed for task ':sdks:python:test-suites:direct:py2:directRunnerIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug 
option to get more log output. Run with --scan to get full insights.
==============================================================================

2: Task failed with an exception.
-----------
* Where:
Build file 
'<https://builds.apache.org/job/beam_PostCommit_Python2/ws/src/sdks/python/test-suites/dataflow/py2/build.gradle'>
 line: 85

* What went wrong:
Execution failed for task ':sdks:python:test-suites:dataflow:py2:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --info or --debug 
option to get more log output. Run with --scan to get full insights.
==============================================================================

* Get more help at https://help.gradle.org

Deprecated Gradle features were used in this build, making it incompatible with 
Gradle 6.0.
Use '--warning-mode all' to show the individual deprecation warnings.
See 
https://docs.gradle.org/5.2.1/userguide/command_line_interface.html#sec:command_line_warnings

BUILD FAILED in 1h 32m 10s
120 actionable tasks: 95 executed, 22 from cache, 3 up-to-date

Publishing build scan...

Publishing failed.

The response from https://scans-in.gradle.com/in/5.2.1/2.3 was not from the 
build scan server.
Your network environment may be interfering, or the service may be unavailable.

If you believe this to be in error, please report this problem via 
https://gradle.com/scans/help/plugin and include the following via copy/paste:

----------
Gradle version: 5.2.1
Plugin version: 2.3
Request URL: https://scans-in.gradle.com/in/5.2.1/2.3
Request ID: 300953ec-5f67-4d10-8165-8701436f3586
Response status code: 503
Response content type: text/html; charset=utf-8
Response server type: cloudflare
----------

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to