See
<https://builds.apache.org/job/beam_PostCommit_Python_Verify/2098/display/redirect?page=changes>
Changes:
[dhalperi] Remove IoChannelUtils from PackageUtil
------------------------------------------
[...truncated 634.27 KB...]
{
"kind": "ParallelDo",
"name": "s30",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "_merge_tagged_vals_under_key"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name":
"assert:even/Group/Map(_merge_tagged_vals_under_key).out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s29"
},
"serialized_fn": "<string of 1340 bytes>",
"user_name": "assert:even/Group/Map(_merge_tagged_vals_under_key)"
}
},
{
"kind": "ParallelDo",
"name": "s31",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "<lambda>"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "assert:even/Unkey.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s30"
},
"serialized_fn": "<string of 960 bytes>",
"user_name": "assert:even/Unkey"
}
},
{
"kind": "ParallelDo",
"name": "s32",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "_equal"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "assert:even/Match.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s31"
},
"serialized_fn": "<string of 1112 bytes>",
"user_name": "assert:even/Match"
}
}
],
"type": "JOB_TYPE_BATCH"
}
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url
https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json
after exception HttpError accessing
<https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>:
response: <{'status': '429', 'content-length': '440', 'x-xss-protection': '1;
mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding':
'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF',
'-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 03 May
2017 16:27:49 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type':
'application/json; charset=UTF-8'}>, content <{
"error": {
"code": 429,
"message": "(fcc21ced99e37205): The workflow could not be created. Causes:
(34d3115f49f2cda): Too many running jobs. Project apache-beam-testing is
running 25 jobs and project limit for active jobs is 25. To fix this, cancel an
existing workflow via the UI, wait for a workflow to finish or contact
[email protected] to request an increase in quota.",
"status": "RESOURCE_EXHAUSTED"
}
}
>
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url
https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json
after exception HttpError accessing
<https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>:
response: <{'status': '429', 'content-length': '440', 'x-xss-protection': '1;
mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding':
'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF',
'-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 03 May
2017 16:27:50 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type':
'application/json; charset=UTF-8'}>, content <{
"error": {
"code": 429,
"message": "(e55eb6629418a40): The workflow could not be created. Causes:
(bb3bda853f817482): Too many running jobs. Project apache-beam-testing is
running 25 jobs and project limit for active jobs is 25. To fix this, cancel an
existing workflow via the UI, wait for a workflow to finish or contact
[email protected] to request an increase in quota.",
"status": "RESOURCE_EXHAUSTED"
}
}
>
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url
https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json
after exception HttpError accessing
<https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>:
response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1;
mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding':
'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF',
'-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 03 May
2017 16:27:56 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type':
'application/json; charset=UTF-8'}>, content <{
"error": {
"code": 429,
"message": "(3207d0af2c897be2): The workflow could not be created. Causes:
(edc5b13bde512365): Too many running jobs. Project apache-beam-testing is
running 25 jobs and project limit for active jobs is 25. To fix this, cancel an
existing workflow via the UI, wait for a workflow to finish or contact
[email protected] to request an increase in quota.",
"status": "RESOURCE_EXHAUSTED"
}
}
>
root: DEBUG: Response returned status 429, retrying
root: DEBUG: Retrying request to url
https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json
after exception HttpError accessing
<https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>:
response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1;
mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding':
'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF',
'-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 03 May
2017 16:28:02 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type':
'application/json; charset=UTF-8'}>, content <{
"error": {
"code": 429,
"message": "(126b0b61974710b6): The workflow could not be created. Causes:
(1e602abefc8096d3): Too many running jobs. Project apache-beam-testing is
running 25 jobs and project limit for active jobs is 25. To fix this, cancel an
existing workflow via the UI, wait for a workflow to finish or contact
[email protected] to request an increase in quota.",
"status": "RESOURCE_EXHAUSTED"
}
}
>
root: ERROR: HTTP status 429 trying to create job at dataflow service endpoint
https://dataflow.googleapis.com
root: CRITICAL: details of server error: HttpError accessing
<https://dataflow.googleapis.com/v1b3/projects/apache-beam-testing/locations/us-central1/jobs?alt=json>:
response: <{'status': '429', 'content-length': '441', 'x-xss-protection': '1;
mode=block', 'x-content-type-options': 'nosniff', 'transfer-encoding':
'chunked', 'vary': 'Origin, X-Origin, Referer', 'server': 'ESF',
'-content-encoding': 'gzip', 'cache-control': 'private', 'date': 'Wed, 03 May
2017 16:28:17 GMT', 'x-frame-options': 'SAMEORIGIN', 'content-type':
'application/json; charset=UTF-8'}>, content <{
"error": {
"code": 429,
"message": "(6eeb594ace23805d): The workflow could not be created. Causes:
(4ee3f02c20009b02): Too many running jobs. Project apache-beam-testing is
running 25 jobs and project limit for active jobs is 25. To fix this, cancel an
existing workflow via the UI, wait for a workflow to finish or contact
[email protected] to request an increase in quota.",
"status": "RESOURCE_EXHAUSTED"
}
}
>
--------------------- >> end captured logging << ---------------------
----------------------------------------------------------------------
Ran 15 tests in 1427.368s
FAILED (errors=1)
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_27_48-8852677359240508171?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_34_33-7960319692552477396?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_40_32-5169337114158688529?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_45_57-41850976304989017?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_27_47-16888918878427213762?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_35_17-9742794674994404992?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_40_56-11113268926944442522?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_28_31-2772219258342997150?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_35_05-9282004339487295910?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_40_49-7836355371819706203?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_27_47-12446233105389338038?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_33_26-11443180852716378722?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_39_01-11918176939402246945?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/job/2017-05-03_09_45_02-8259414461937193857?project=apache-beam-testing
Build step 'Execute shell' marked build as failure