See
<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/1148/display/redirect?page=changes>
Changes:
[iemejia] Add missing ASF license to ExecutableStageTranslation file
------------------------------------------
[...truncated 316.52 KB...]
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name":
"assert_that/Group/Map(_merge_tagged_vals_under_key).out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s11"
},
"serialized_fn": "<string of 1380 bytes>",
"user_name": "assert_that/Group/Map(_merge_tagged_vals_under_key)"
}
},
{
"kind": "ParallelDo",
"name": "s13",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "<lambda>"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "assert_that/Unkey.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s12"
},
"serialized_fn": "<string of 980 bytes>",
"user_name": "assert_that/Unkey"
}
},
{
"kind": "ParallelDo",
"name": "s14",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "_equal"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "assert_that/Match.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s13"
},
"serialized_fn": "<string of 1156 bytes>",
"user_name": "assert_that/Match"
}
}
],
"type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
createTime: u'2018-03-20T13:26:01.056777Z'
currentStateTime: u'1970-01-01T00:00:00Z'
id: u'2018-03-20_06_26_00-5512300713639124978'
location: u'us-central1'
name: u'beamapp-jenkins-0320132546-521356'
projectId: u'apache-beam-testing'
stageStates: []
steps: []
tempFiles: []
type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2018-03-20_06_26_00-5512300713639124978]
root: INFO: To access the Dataflow monitoring console, please navigate to
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_26_00-5512300713639124978?project=apache-beam-testing
root: INFO: Job 2018-03-20_06_26_00-5512300713639124978 is in state
JOB_STATE_PENDING
root: INFO: 2018-03-20T13:26:00.193Z: JOB_MESSAGE_WARNING: Job
2018-03-20_06_26_00-5512300713639124978 might autoscale up to 250 workers.
root: INFO: 2018-03-20T13:26:00.210Z: JOB_MESSAGE_DETAILED: Autoscaling is
enabled for job 2018-03-20_06_26_00-5512300713639124978. The number of workers
will be between 1 and 250.
root: INFO: 2018-03-20T13:26:00.232Z: JOB_MESSAGE_DETAILED: Autoscaling was
automatically enabled for job 2018-03-20_06_26_00-5512300713639124978.
root: INFO: 2018-03-20T13:26:03.426Z: JOB_MESSAGE_DETAILED: Checking required
Cloud APIs are enabled.
root: INFO: 2018-03-20T13:26:03.596Z: JOB_MESSAGE_DETAILED: Checking
permissions granted to controller Service Account.
root: INFO: 2018-03-20T13:26:05.983Z: JOB_MESSAGE_DETAILED: Expanding
CoGroupByKey operations into optimizable parts.
root: INFO: 2018-03-20T13:26:06.017Z: JOB_MESSAGE_DEBUG: Combiner lifting
skipped for step assert_that/Group/GroupByKey: GroupByKey not followed by a
combiner.
root: INFO: 2018-03-20T13:26:06.038Z: JOB_MESSAGE_DETAILED: Expanding
GroupByKey operations into optimizable parts.
root: INFO: 2018-03-20T13:26:06.069Z: JOB_MESSAGE_DETAILED: Lifting
ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2018-03-20T13:26:06.095Z: JOB_MESSAGE_DEBUG: Annotating graph with
Autotuner information.
root: INFO: 2018-03-20T13:26:06.134Z: JOB_MESSAGE_DETAILED: Fusing adjacent
ParDo, Read, Write, and Flatten operations
root: INFO: 2018-03-20T13:26:06.156Z: JOB_MESSAGE_DETAILED: Fusing consumer
assert_that/Group/Map(_merge_tagged_vals_under_key) into
assert_that/Group/GroupByKey/GroupByWindow
root: INFO: 2018-03-20T13:26:06.177Z: JOB_MESSAGE_DETAILED: Fusing consumer
assert_that/Match into assert_that/Unkey
root: INFO: 2018-03-20T13:26:06.209Z: JOB_MESSAGE_DETAILED: Fusing consumer
assert_that/Unkey into assert_that/Group/Map(_merge_tagged_vals_under_key)
root: INFO: 2018-03-20T13:26:06.233Z: JOB_MESSAGE_DETAILED: Unzipping flatten
s10 for input s8.out
root: INFO: 2018-03-20T13:26:06.262Z: JOB_MESSAGE_DETAILED: Fusing unzipped
copy of assert_that/Group/GroupByKey/Reify, through flatten
assert_that/Group/Flatten, into producer assert_that/Group/pair_with_0
root: INFO: 2018-03-20T13:26:06.294Z: JOB_MESSAGE_DETAILED: Unzipping flatten
s10-u13 for input s11-reify-value0-c11
root: INFO: 2018-03-20T13:26:06.308Z: JOB_MESSAGE_DETAILED: Fusing unzipped
copy of assert_that/Group/GroupByKey/Write, through flatten s10-u13, into
producer assert_that/Group/GroupByKey/Reify
root: INFO: 2018-03-20T13:26:06.338Z: JOB_MESSAGE_DETAILED: Fusing consumer
assert_that/Group/GroupByKey/GroupByWindow into
assert_that/Group/GroupByKey/Read
root: INFO: 2018-03-20T13:26:06.362Z: JOB_MESSAGE_DETAILED: Fusing consumer
assert_that/Group/GroupByKey/Reify into assert_that/Group/pair_with_1
root: INFO: 2018-03-20T13:26:06.388Z: JOB_MESSAGE_DETAILED: Fusing consumer
assert_that/Group/GroupByKey/Write into assert_that/Group/GroupByKey/Reify
root: INFO: 2018-03-20T13:26:06.418Z: JOB_MESSAGE_DETAILED: Fusing consumer
assert_that/Group/pair_with_1 into assert_that/ToVoidKey
root: INFO: 2018-03-20T13:26:06.442Z: JOB_MESSAGE_DETAILED: Fusing consumer
assert_that/Group/pair_with_0 into assert_that/Create/Read
root: INFO: 2018-03-20T13:26:06.473Z: JOB_MESSAGE_DETAILED: Fusing consumer
assert_that/ToVoidKey into assert_that/WindowInto(WindowIntoFn)
root: INFO: 2018-03-20T13:26:06.497Z: JOB_MESSAGE_DETAILED: Fusing consumer
assert_that/WindowInto(WindowIntoFn) into FlatMap(<lambda at
sideinputs_test.py:165>)/FlatMap(<lambda at sideinputs_test.py:165>)
root: INFO: 2018-03-20T13:26:06.520Z: JOB_MESSAGE_DETAILED: Fusing consumer
FlatMap(<lambda at sideinputs_test.py:165>)/FlatMap(<lambda at
sideinputs_test.py:165>) into start/Read
root: INFO: 2018-03-20T13:26:06.552Z: JOB_MESSAGE_DEBUG: Workflow config is
missing a default resource spec.
root: INFO: 2018-03-20T13:26:06.576Z: JOB_MESSAGE_DEBUG: Adding StepResource
setup and teardown to workflow graph.
root: INFO: 2018-03-20T13:26:06.608Z: JOB_MESSAGE_DEBUG: Adding workflow start
and stop steps.
root: INFO: 2018-03-20T13:26:06.635Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2018-03-20T13:26:06.768Z: JOB_MESSAGE_DEBUG: Executing wait step
start21
root: INFO: 2018-03-20T13:26:06.835Z: JOB_MESSAGE_BASIC: Executing operation
assert_that/Group/GroupByKey/Create
root: INFO: 2018-03-20T13:26:06.869Z: JOB_MESSAGE_BASIC: Executing operation
side/Read
root: INFO: 2018-03-20T13:26:06.881Z: JOB_MESSAGE_DEBUG: Starting worker pool
setup.
root: INFO: 2018-03-20T13:26:06.903Z: JOB_MESSAGE_BASIC: Starting 1 workers in
us-central1-f...
root: INFO: 2018-03-20T13:26:06.925Z: JOB_MESSAGE_DEBUG: Value "side/Read.out"
materialized.
root: INFO: 2018-03-20T13:26:06.990Z: JOB_MESSAGE_BASIC: Executing operation
FlatMap(<lambda at sideinputs_test.py:165>)/_UnpickledSideInput(Read.out.0)
root: INFO: 2018-03-20T13:26:07.045Z: JOB_MESSAGE_DEBUG: Value
"assert_that/Group/GroupByKey/Session" materialized.
root: INFO: 2018-03-20T13:26:07.077Z: JOB_MESSAGE_DEBUG: Value "FlatMap(<lambda
at sideinputs_test.py:165>)/_UnpickledSideInput(Read.out.0).output"
materialized.
root: INFO: 2018-03-20T13:26:07.112Z: JOB_MESSAGE_BASIC: Executing operation
assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2018-03-20T13:26:07.143Z: JOB_MESSAGE_BASIC: Executing operation
start/Read+FlatMap(<lambda at sideinputs_test.py:165>)/FlatMap(<lambda at
sideinputs_test.py:165>)+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: Job 2018-03-20_06_26_00-5512300713639124978 is in state
JOB_STATE_RUNNING
root: INFO: 2018-03-20T13:26:14.771Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised
the number of workers to 0 based on the rate of progress in the currently
running step(s).
root: INFO: 2018-03-20T13:27:05.026Z: JOB_MESSAGE_ERROR: Startup of the worker
pool in zone us-central1-f failed to bring up any of the desired 1 workers.
QUOTA_EXCEEDED: Quota 'DISKS_TOTAL_GB' exceeded. Limit: 21000.0 in region
us-central1.
root: INFO: 2018-03-20T13:27:05.057Z: JOB_MESSAGE_ERROR: Workflow failed.
root: INFO: 2018-03-20T13:27:05.440Z: JOB_MESSAGE_DETAILED: Cleaning up.
root: INFO: 2018-03-20T13:27:05.482Z: JOB_MESSAGE_DEBUG: Starting worker pool
teardown.
root: INFO: 2018-03-20T13:27:05.504Z: JOB_MESSAGE_BASIC: Stopping worker pool...
root: INFO: 2018-03-20T13:27:21.017Z: JOB_MESSAGE_DEBUG: Tearing down pending
resources...
root: INFO: Job 2018-03-20_06_26_00-5512300713639124978 is in state
JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------
----------------------------------------------------------------------
Ran 16 tests in 1401.324s
FAILED (errors=4)
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_11_27-10310204475851904369?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_18_48-10932415101164771928?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_25_39-9916507116655823063?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_26_58-6347546351124095199?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_11_27-15580952472524332547?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_18_28-15654329671840705487?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_25_15-7441854485727004513?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_26_41-15746069443069673455?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_11_27-3059862061204436450?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_18_38-1199791738937302827?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_25_53-3548584222256344285?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_27_33-16046020936014067874?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_11_27-7846900568132959614?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_18_34-18154866645975277529?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_26_00-5512300713639124978?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-20_06_27_40-3058362086335794835?project=apache-beam-testing
Build step 'Execute shell' marked build as failure