See
<https://builds.apache.org/job/beam_PerformanceTests_Python/769/display/redirect?page=changes>
Changes:
[jb] Exclude go from rat configuration, use .gitignore for excludes and align
[iemejia] [BEAM-3422] Update reproducible-build-maven-plugin to version 0.4
[jbonofre] [BEAM-3282] MqttIO reader now use receive for timeout in order to
return
------------------------------------------
[...truncated 90.29 KB...]
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
}
]
},
"output_name": "out",
"user_name":
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0).output"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s8"
},
"user_name":
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0)"
}
},
{
"kind": "CollectionToSingleton",
"name": "SideInput-s16",
"properties": {
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
}
]
},
"output_name": "out",
"user_name":
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0).output"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s14"
},
"user_name":
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0)"
}
},
{
"kind": "ParallelDo",
"name": "s17",
"properties": {
"display_data": [
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.CallableWrapperDoFn",
"type": "STRING",
"value": "_finalize_write"
},
{
"key": "fn",
"label": "Transform Function",
"namespace": "apache_beam.transforms.core.ParDo",
"shortValue": "CallableWrapperDoFn",
"type": "STRING",
"value": "apache_beam.transforms.core.CallableWrapperDoFn"
}
],
"non_parallel_inputs": {
"SideInput-s15": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "SideInput-s15"
},
"SideInput-s16": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "SideInput-s16"
}
},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "write/Write/WriteImpl/FinalizeWrite.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s7"
},
"serialized_fn": "<string of 2420 bytes>",
"user_name": "write/Write/WriteImpl/FinalizeWrite/FinalizeWrite"
}
}
],
"type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
createTime: u'2018-01-09T12:10:07.115182Z'
currentStateTime: u'1970-01-01T00:00:00Z'
id: u'2018-01-09_04_10_06-6324618345595965919'
location: u'us-central1'
name: u'beamapp-jenkins-0109121005-033215'
projectId: u'apache-beam-testing'
stageStates: []
steps: []
tempFiles: []
type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2018-01-09_04_10_06-6324618345595965919]
root: INFO: To access the Dataflow monitoring console, please navigate to
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-01-09_04_10_06-6324618345595965919?project=apache-beam-testing
root: INFO: Job 2018-01-09_04_10_06-6324618345595965919 is in state
JOB_STATE_PENDING
root: INFO: 2018-01-09T12:10:06.421Z: JOB_MESSAGE_DETAILED: (57c58ee0452a9fdb):
Autoscaling is enabled for job 2018-01-09_04_10_06-6324618345595965919. The
number of workers will be between 1 and 15.
root: INFO: 2018-01-09T12:10:06.450Z: JOB_MESSAGE_DETAILED: (57c58ee0452a9f5a):
Autoscaling was automatically enabled for job
2018-01-09_04_10_06-6324618345595965919.
root: INFO: 2018-01-09T12:10:08.843Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd205):
Checking required Cloud APIs are enabled.
root: INFO: 2018-01-09T12:10:09.706Z: JOB_MESSAGE_DETAILED: (a8542d126dbbdc7f):
Expanding CoGroupByKey operations into optimizable parts.
root: INFO: 2018-01-09T12:10:09.736Z: JOB_MESSAGE_DEBUG: (a8542d126dbbd350):
Combiner lifting skipped for step write/Write/WriteImpl/GroupByKey: GroupByKey
not followed by a combiner.
root: INFO: 2018-01-09T12:10:09.763Z: JOB_MESSAGE_DEBUG: (a8542d126dbbd286):
Combiner lifting skipped for step group: GroupByKey not followed by a combiner.
root: INFO: 2018-01-09T12:10:09.787Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd1bc):
Expanding GroupByKey operations into optimizable parts.
root: INFO: 2018-01-09T12:10:09.820Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd0f2):
Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2018-01-09T12:10:09.855Z: JOB_MESSAGE_DEBUG: (a8542d126dbbde94):
Annotating graph with Autotuner information.
root: INFO: 2018-01-09T12:10:09.891Z: JOB_MESSAGE_DETAILED: (a8542d126dbbdd00):
Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2018-01-09T12:10:09.922Z: JOB_MESSAGE_DETAILED: (a8542d126dbbdc36):
Fusing consumer split into read/Read
root: INFO: 2018-01-09T12:10:09.952Z: JOB_MESSAGE_DETAILED: (a8542d126dbbdb6c):
Fusing consumer group/Write into group/Reify
root: INFO: 2018-01-09T12:10:09.982Z: JOB_MESSAGE_DETAILED: (a8542d126dbbdaa2):
Fusing consumer group/GroupByWindow into group/Read
root: INFO: 2018-01-09T12:10:10.012Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd9d8):
Fusing consumer write/Write/WriteImpl/GroupByKey/GroupByWindow into
write/Write/WriteImpl/GroupByKey/Read
root: INFO: 2018-01-09T12:10:10.043Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd90e):
Fusing consumer write/Write/WriteImpl/GroupByKey/Write into
write/Write/WriteImpl/GroupByKey/Reify
root: INFO: 2018-01-09T12:10:10.074Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd844):
Fusing consumer write/Write/WriteImpl/WindowInto(WindowIntoFn) into
write/Write/WriteImpl/Pair
root: INFO: 2018-01-09T12:10:10.101Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd77a):
Fusing consumer write/Write/WriteImpl/GroupByKey/Reify into
write/Write/WriteImpl/WindowInto(WindowIntoFn)
root: INFO: 2018-01-09T12:10:10.131Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd6b0):
Fusing consumer pair_with_one into split
root: INFO: 2018-01-09T12:10:10.161Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd5e6):
Fusing consumer group/Reify into pair_with_one
root: INFO: 2018-01-09T12:10:10.191Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd51c):
Fusing consumer write/Write/WriteImpl/WriteBundles/WriteBundles into format
root: INFO: 2018-01-09T12:10:10.220Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd452):
Fusing consumer write/Write/WriteImpl/Pair into
write/Write/WriteImpl/WriteBundles/WriteBundles
root: INFO: 2018-01-09T12:10:10.250Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd388):
Fusing consumer format into count
root: INFO: 2018-01-09T12:10:10.274Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd2be):
Fusing consumer write/Write/WriteImpl/Extract into
write/Write/WriteImpl/GroupByKey/GroupByWindow
root: INFO: 2018-01-09T12:10:10.301Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd1f4):
Fusing consumer count into group/GroupByWindow
root: INFO: 2018-01-09T12:10:10.326Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd12a):
Fusing consumer write/Write/WriteImpl/InitializeWrite into
write/Write/WriteImpl/DoOnce/Read
root: INFO: 2018-01-09T12:10:10.359Z: JOB_MESSAGE_DEBUG: (a8542d126dbbd060):
Workflow config is missing a default resource spec.
root: INFO: 2018-01-09T12:10:10.380Z: JOB_MESSAGE_DEBUG: (a8542d126dbbdf96):
Adding StepResource setup and teardown to workflow graph.
root: INFO: 2018-01-09T12:10:10.410Z: JOB_MESSAGE_DEBUG: (a8542d126dbbdecc):
Adding workflow start and stop steps.
root: INFO: 2018-01-09T12:10:10.440Z: JOB_MESSAGE_DEBUG: (a8542d126dbbde02):
Assigning stage ids.
root: INFO: 2018-01-09T12:10:10.570Z: JOB_MESSAGE_DEBUG: (91a883546c333725):
Executing wait step start25
root: INFO: 2018-01-09T12:10:10.617Z: JOB_MESSAGE_BASIC: (b6315d4e85b45476):
Executing operation
write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
root: INFO: 2018-01-09T12:10:10.648Z: JOB_MESSAGE_BASIC: (93450b0d75dcbf61):
Executing operation group/Create
root: INFO: 2018-01-09T12:10:10.659Z: JOB_MESSAGE_DEBUG: (788fd71de50cea40):
Starting worker pool setup.
root: INFO: 2018-01-09T12:10:10.681Z: JOB_MESSAGE_BASIC: (788fd71de50ce992):
Starting 1 workers in us-central1-f...
root: INFO: 2018-01-09T12:10:10.756Z: JOB_MESSAGE_DEBUG: (93450b0d75dcb1a3):
Value "group/Session" materialized.
root: INFO: 2018-01-09T12:10:10.870Z: JOB_MESSAGE_BASIC: (93450b0d75dcb878):
Executing operation read/Read+split+pair_with_one+group/Reify+group/Write
root: INFO: Job 2018-01-09_04_10_06-6324618345595965919 is in state
JOB_STATE_RUNNING
root: INFO: 2018-01-09T12:10:19.589Z: JOB_MESSAGE_DETAILED: (de52740bc6606924):
Autoscaling: Raised the number of workers to 0 based on the rate of progress in
the currently running step(s).
root: INFO: 2018-01-09T12:10:59.099Z: JOB_MESSAGE_ERROR: (de52740bc66060c9):
Startup of the worker pool in zone us-central1-f failed to bring up any of the
desired 1 workers. QUOTA_EXCEEDED: Quota 'DISKS_TOTAL_GB' exceeded. Limit:
21000.0 in region us-central1.
root: INFO: 2018-01-09T12:10:59.125Z: JOB_MESSAGE_ERROR: (de52740bc660699f):
Workflow failed.
root: INFO: 2018-01-09T12:10:59.367Z: JOB_MESSAGE_DETAILED: (a8542d126dbbd275):
Cleaning up.
root: INFO: 2018-01-09T12:10:59.409Z: JOB_MESSAGE_DEBUG: (a8542d126dbbd0e1):
Starting worker pool teardown.
root: INFO: 2018-01-09T12:10:59.432Z: JOB_MESSAGE_BASIC: (a8542d126dbbd017):
Stopping worker pool...
root: INFO: 2018-01-09T12:11:08.694Z: JOB_MESSAGE_DEBUG: (a8542d126dbbdc25):
Tearing down pending resources...
root: INFO: Job 2018-01-09_04_10_06-6324618345595965919 is in state
JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------
----------------------------------------------------------------------
Ran 2 tests in 194.751s
FAILED (errors=2)
2018-01-09 12:11:17,339 5a7302f7 MainThread beam_integration_benchmark(1/1)
ERROR Error during benchmark beam_integration_benchmark
Traceback (most recent call last):
File
"<https://builds.apache.org/job/beam_PerformanceTests_Python/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",>
line 601, in RunBenchmark
DoRunPhase(spec, collector, detailed_timer)
File
"<https://builds.apache.org/job/beam_PerformanceTests_Python/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",>
line 504, in DoRunPhase
samples = spec.BenchmarkRun(spec)
File
"<https://builds.apache.org/job/beam_PerformanceTests_Python/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",>
line 159, in Run
job_type=job_type)
File
"<https://builds.apache.org/job/beam_PerformanceTests_Python/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",>
line 90, in SubmitJob
assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2018-01-09 12:11:17,340 5a7302f7 MainThread beam_integration_benchmark(1/1)
INFO Cleaning up benchmark beam_integration_benchmark
2018-01-09 12:11:17,352 5a7302f7 MainThread beam_integration_benchmark(1/1)
ERROR Benchmark 1/1 beam_integration_benchmark (UID:
beam_integration_benchmark0) failed. Execution will continue.
2018-01-09 12:11:17,395 5a7302f7 MainThread INFO Benchmark run statuses:
---------------------------------------------------------------
Name UID Status
---------------------------------------------------------------
beam_integration_benchmark beam_integration_benchmark0 FAILED
---------------------------------------------------------------
Success rate: 0.00% (0/1)
2018-01-09 12:11:17,396 5a7302f7 MainThread INFO Complete logs can be found
at: /tmp/perfkitbenchmarker/runs/5a7302f7/pkb.log
2018-01-09 12:11:17,396 5a7302f7 MainThread INFO Completion statuses can be
found at: /tmp/perfkitbenchmarker/runs/5a7302f7/completion_statuses.json
Build step 'Execute shell' marked build as failure
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]