See
<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/432/display/redirect?page=changes>
Changes:
[altay] Adding debug server to sdk worker to get threaddumps
------------------------------------------
[...truncated 1.07 MB...]
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name":
"assert_that/Group/Map(_merge_tagged_vals_under_key).out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s13"
},
"serialized_fn": "<string of 1380 bytes>",
"user_name": "assert_that/Group/Map(_merge_tagged_vals_under_key)"
}
},
{
"kind": "ParallelDo",
"name": "s15",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "<lambda>"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "assert_that/Unkey.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s14"
},
"serialized_fn": "<string of 980 bytes>",
"user_name": "assert_that/Unkey"
}
},
{
"kind": "ParallelDo",
"name": "s16",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "_equal"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "assert_that/Match.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s15"
},
"serialized_fn": "<string of 1160 bytes>",
"user_name": "assert_that/Match"
}
}
],
"type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
createTime: u'2017-12-11T18:18:20.790421Z'
currentStateTime: u'1970-01-01T00:00:00Z'
id: u'2017-12-11_10_18_20-331615697938870470'
location: u'us-central1'
name: u'beamapp-jenkins-1211181759-019285'
projectId: u'apache-beam-testing'
stageStates: []
steps: []
tempFiles: []
type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2017-12-11_10_18_20-331615697938870470]
root: INFO: To access the Dataflow monitoring console, please navigate to
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_18_20-331615697938870470?project=apache-beam-testing
root: INFO: Job 2017-12-11_10_18_20-331615697938870470 is in state
JOB_STATE_PENDING
root: INFO: 2017-12-11T18:18:20.085Z: JOB_MESSAGE_WARNING: (49a22b86596fed2):
Setting the number of workers (1) disables autoscaling for this job. If you are
trying to cap autoscaling, consider only setting max_num_workers. If you want
to disable autoscaling altogether, the documented way is to explicitly use
autoscalingAlgorithm=NONE.
root: INFO: 2017-12-11T18:18:22.765Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb22a):
Checking required Cloud APIs are enabled.
root: INFO: 2017-12-11T18:18:25.547Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbc42):
Expanding CoGroupByKey operations into optimizable parts.
root: INFO: 2017-12-11T18:18:25.570Z: JOB_MESSAGE_DEBUG: (a72caaa2580bb603):
Combiner lifting skipped for step assert_that/Group/GroupByKey: GroupByKey not
followed by a combiner.
root: INFO: 2017-12-11T18:18:25.586Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb1d9):
Expanding GroupByKey operations into optimizable parts.
root: INFO: 2017-12-11T18:18:25.606Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbdaf):
Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2017-12-11T18:18:25.632Z: JOB_MESSAGE_DEBUG: (a72caaa2580bb55b):
Annotating graph with Autotuner information.
root: INFO: 2017-12-11T18:18:25.693Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbd07):
Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2017-12-11T18:18:25.722Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb8dd):
Unzipping flatten s12 for input s10.out
root: INFO: 2017-12-11T18:18:25.756Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb4b3):
Fusing unzipped copy of assert_that/Group/GroupByKey/Reify, through flatten
assert_that/Group/Flatten, into producer assert_that/Group/pair_with_0
root: INFO: 2017-12-11T18:18:25.783Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb089):
Fusing consumer assert_that/Group/Map(_merge_tagged_vals_under_key) into
assert_that/Group/GroupByKey/GroupByWindow
root: INFO: 2017-12-11T18:18:25.803Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbc5f):
Fusing consumer assert_that/Unkey into
assert_that/Group/Map(_merge_tagged_vals_under_key)
root: INFO: 2017-12-11T18:18:25.822Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb835):
Fusing consumer assert_that/Group/GroupByKey/GroupByWindow into
assert_that/Group/GroupByKey/Read
root: INFO: 2017-12-11T18:18:25.849Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb40b):
Fusing consumer assert_that/Match into assert_that/Unkey
root: INFO: 2017-12-11T18:18:25.871Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbfe1):
Unzipping flatten s12-u13 for input s13-reify-value0-c11
root: INFO: 2017-12-11T18:18:25.894Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbbb7):
Fusing unzipped copy of assert_that/Group/GroupByKey/Write, through flatten
s12-u13, into producer assert_that/Group/GroupByKey/Reify
root: INFO: 2017-12-11T18:18:25.911Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb78d):
Fusing consumer assert_that/Group/pair_with_0 into assert_that/Create/Read
root: INFO: 2017-12-11T18:18:25.932Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb363):
Fusing consumer assert_that/Group/GroupByKey/Write into
assert_that/Group/GroupByKey/Reify
root: INFO: 2017-12-11T18:18:25.952Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbf39):
Fusing consumer assert_that/Group/GroupByKey/Reify into
assert_that/Group/pair_with_1
root: INFO: 2017-12-11T18:18:25.972Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbb0f):
Fusing consumer FlatMap(<lambda at sideinputs_test.py:302>)/FlatMap(<lambda at
sideinputs_test.py:302>) into main input/Read
root: INFO: 2017-12-11T18:18:25.991Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb6e5):
Fusing consumer assert_that/Group/pair_with_1 into assert_that/ToVoidKey
root: INFO: 2017-12-11T18:18:26.015Z: JOB_MESSAGE_DETAILED: (a72caaa2580bb2bb):
Fusing consumer assert_that/WindowInto(WindowIntoFn) into FlatMap(<lambda at
sideinputs_test.py:302>)/FlatMap(<lambda at sideinputs_test.py:302>)
root: INFO: 2017-12-11T18:18:26.033Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbe91):
Fusing consumer assert_that/ToVoidKey into assert_that/WindowInto(WindowIntoFn)
root: INFO: 2017-12-11T18:18:26.054Z: JOB_MESSAGE_DEBUG: (a72caaa2580bba67):
Workflow config is missing a default resource spec.
root: INFO: 2017-12-11T18:18:26.074Z: JOB_MESSAGE_DEBUG: (a72caaa2580bb63d):
Adding StepResource setup and teardown to workflow graph.
root: INFO: 2017-12-11T18:18:26.099Z: JOB_MESSAGE_DEBUG: (a72caaa2580bb213):
Adding workflow start and stop steps.
root: INFO: 2017-12-11T18:18:26.122Z: JOB_MESSAGE_DEBUG: (a72caaa2580bbde9):
Assigning stage ids.
root: INFO: Job 2017-12-11_10_18_20-331615697938870470 is in state
JOB_STATE_RUNNING
root: INFO: 2017-12-11T18:18:26.316Z: JOB_MESSAGE_DEBUG: (62475f0455e388be):
Executing wait step start22
root: INFO: 2017-12-11T18:18:26.374Z: JOB_MESSAGE_BASIC: (8dd961acd8e88041):
Executing operation side1/Read
root: INFO: 2017-12-11T18:18:26.409Z: JOB_MESSAGE_BASIC: (62475f0455e38fe0):
Executing operation side2/Read
root: INFO: 2017-12-11T18:18:26.441Z: JOB_MESSAGE_DEBUG: (8dd961acd8e88167):
Value "side1/Read.out" materialized.
root: INFO: 2017-12-11T18:18:26.472Z: JOB_MESSAGE_DEBUG: (62475f0455e38702):
Value "side2/Read.out" materialized.
root: INFO: 2017-12-11T18:18:26.537Z: JOB_MESSAGE_BASIC: (62475f0455e38249):
Executing operation Flatten
root: INFO: 2017-12-11T18:18:26.624Z: JOB_MESSAGE_DEBUG: (62475f0455e388d7):
Value "Flatten.out" materialized.
root: INFO: 2017-12-11T18:18:26.684Z: JOB_MESSAGE_BASIC: (8dd961acd8e8825c):
Executing operation FlatMap(<lambda at
sideinputs_test.py:302>)/_UnpickledSideInput(Flatten.out.0)
root: INFO: 2017-12-11T18:18:26.756Z: JOB_MESSAGE_DEBUG: (8dd961acd8e8850a):
Value "FlatMap(<lambda at
sideinputs_test.py:302>)/_UnpickledSideInput(Flatten.out.0).output"
materialized.
root: INFO: 2017-12-11T18:18:26.819Z: JOB_MESSAGE_BASIC: (62475f0455e38262):
Executing operation assert_that/Group/GroupByKey/Create
root: INFO: 2017-12-11T18:18:26.861Z: JOB_MESSAGE_DEBUG: (cd1119aa2b5c67b6):
Starting worker pool setup.
root: INFO: 2017-12-11T18:18:26.890Z: JOB_MESSAGE_BASIC: (cd1119aa2b5c68d0):
Starting 1 workers in us-central1-f...
root: INFO: 2017-12-11T18:18:26.953Z: JOB_MESSAGE_DEBUG: (9f404f4a3c7dd91b):
Value "assert_that/Group/GroupByKey/Session" materialized.
root: INFO: 2017-12-11T18:18:27.018Z: JOB_MESSAGE_BASIC: (9f404f4a3c7dda0c):
Executing operation main input/Read+FlatMap(<lambda at
sideinputs_test.py:302>)/FlatMap(<lambda at
sideinputs_test.py:302>)+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2017-12-11T18:18:27.044Z: JOB_MESSAGE_BASIC: (62475f0455e388f0):
Executing operation
assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2017-12-11T18:18:33.851Z: JOB_MESSAGE_DETAILED: (4bd4c805d4422ebb):
Autoscaling: Raised the number of workers to 0 based on the rate of progress in
the currently running step(s).
root: INFO: 2017-12-11T18:18:57.734Z: JOB_MESSAGE_ERROR: (4bd4c805d4422b8c):
Startup of the worker pool in zone us-central1-f failed to bring up any of the
desired 1 workers. ZONE_RESOURCE_POOL_EXHAUSTED: The zone
'projects/apache-beam-testing/zones/us-central1-f' does not have enough
resources available to fulfill the request. Try a different zone, or try again
later.
root: INFO: 2017-12-11T18:18:57.783Z: JOB_MESSAGE_ERROR: (4bd4c805d4422e16):
Workflow failed.
root: INFO: 2017-12-11T18:18:58.229Z: JOB_MESSAGE_DETAILED: (a72caaa2580bbb2c):
Cleaning up.
root: INFO: 2017-12-11T18:18:58.315Z: JOB_MESSAGE_DEBUG: (a72caaa2580bb2d8):
Starting worker pool teardown.
root: INFO: 2017-12-11T18:18:58.348Z: JOB_MESSAGE_BASIC: (a72caaa2580bbeae):
Stopping worker pool...
root: INFO: 2017-12-11T18:20:06.426Z: JOB_MESSAGE_BASIC: (a72caaa2580bbe06):
Worker pool stopped.
root: INFO: 2017-12-11T18:20:06.461Z: JOB_MESSAGE_DEBUG: (a72caaa2580bb5b2):
Tearing down pending resources...
root: INFO: Job 2017-12-11_10_18_20-331615697938870470 is in state
JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------
----------------------------------------------------------------------
Ran 15 tests in 924.728s
FAILED (errors=10)
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_11_17-2050635262786868742?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_14_12-5347517398133682125?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_19_08-960612389348443737?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_11_17-15962750727635144905?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_13_39-4091447656855636797?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_16_30-14639992260540082984?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_11_15-9665064188856117708?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_12_49-11904819970812392070?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_15_37-13733329581878086187?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_18_20-331615697938870470?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_20_38-88395723950014942?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_11_18-2561225770366590499?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_14_01-1920037737539727064?project=apache-beam-testing
Found:
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-12-11_10_16_49-8350530028329332064?project=apache-beam-testing
Build step 'Execute shell' marked build as failure