See
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/524/display/redirect>
Changes:
------------------------------------------
[...truncated 156.23 KB...]
"value": "_finalize_write"
}
],
"non_parallel_inputs": {
"side0-write/Write/WriteImpl/FinalizeWrite": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "SideInput-s18"
},
"side1-write/Write/WriteImpl/FinalizeWrite": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "SideInput-s19"
},
"side2-write/Write/WriteImpl/FinalizeWrite": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "SideInput-s20"
}
},
"output_info": [
{
"encoding": {
"@type": "kind:windowed_value",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
"component_encodings": [
{
"@type":
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
"component_encodings": []
},
{
"@type":
"FastPrimitivesCoder$eNprYE5OLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqYIfgYGhvi0xJycpMTk7HiwlkJ8pgVkJmfnpEJNYQGawlpbyJZUnKQHACYlLgM=",
"component_encodings": []
}
],
"is_pair_like": true
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "out",
"user_name": "write/Write/WriteImpl/FinalizeWrite.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s7"
},
"serialized_fn":
"eNrNV/l/G8UV10rOwZKmAUogpIdISbumSEpIHBKXQlIlAaNYcdcGb4+wjHZHmo13d/bNzFoxtSiQynGhd8vZu/S+6X0ff0f/mb4ZSU4FNs1v6edj767ezHsz7/u+7zu7T5ecgGQkYNRvUZJUlSCpbHORyGrABbXrJI5JK6aLgmQZFaf52dSGwuQzYPWg6HilQqHgt1MojQWJOP61iKS2345SEkdPUr8rIkVtmPB2GZccw4qOzGgA25reDrRlggdUStjOJrwb9Ry1klGfRamSsGN8jzhg7NWQ4iaJ4kLaM+cX0PywNtuwEzd4Q7MHtmNCRWmWKxNPwo1NswOeq6u2Xc18Dd7WylsXYLfzxkwUvaQibvsLeJ+P0iUb3o7R9/TgJsfbg5HaUUz9jCjmZ4K2o0tw81gEnqF3KqvLJM5xnuDLUUiFPa+IioLHtHFuZINbMPA7enCr49kYeOChdwh7gzCK46qvr7YfcxIauw23mQpIJeD2Puxz4Q5v28gT9nvn8bkjp2s1RZOsIhEn0qEVrG6FpmFF8cGNSiVr2UolUpUg5nlYG2BTOzx17MTxo/cdPzE1dfRITVCZx1iKd+atq2mnJKG+zNs67Xcxvf139+A9DtvL7mD7PZxVgDLO19mkeeJLRkQo4c5GwWwz4Jg1HBiDy9jk8GYv8HklorRT5waf9+ICd/XgYMu7SWetow22oAlLFLzPuwvtlYPOcChPJg9NhRXeRtPV9bUN3u/tHQ/RiXlrFMcxg5W7N/WcNOkHPMFyS4m1HZTobm8CzSRXHD7g3YCPSZQMq3ePAUDzqJbFJEqh4u3TBNcdlWIlBYliTNJPaRfvyMfquukIn1GiEao189Y4K1NFBfZVNVdRbJ8SnTyhqZqLSUAZjw1UhxCqw0xf721Y5n6E9uDoBZhyvEmdOTLuUMU0ZW3x6nUmyeLa2WHPGgscG1s6M9yy/UfTLAqWYhrOY6AZ3V823NeD446BISSKwInNHDemn8YpNkzjxj7Yg/sdb7dGL9AKoJtJ5wcf8vajUXtP6+36po2ncU9CK9L08mF4wBChG6Uh7/oJ4qlhRDl6cCtN03FMGGmbipPYf5O3DScN/MsR7epgp8aaLxCUKGRcnga6sW34sMNu83bifE1SrUFQ78NpF844DatRwP9S45a6vV5YLaxbl4vzBTjb7MNDk8ZllAs83IcZT6ClxnhCaxdpuhSlcnSvyJgs01qXiyWJedGaTsufo8LwNQ3ogu5hf5GLsM7zVM0s+HMrR47VpAhqMlzS3a0YT2v/BUptUI9qtgKPmK3cH5OkFZIHoDF73qoX4Jx3q+5ywRNfYEhN5o3dzhoJNfgM5Q2aa3B+UsGcCx8Zg6tDlY/lREa6ZplWHsUKU4J5gzEO61FYWINHXXhszDVKMi6Un/Awj1HqFr2bNUXeRCnw+vBRFz5mwvvoGyjfh4+vwSdcuMAazc2qF1D8AY8zrB2WqYhlKjV2NOr1J1RBWReLq4UQS3XZ6hVXi/LAqnWxFBbVxOWCLp/atoojYWmpKI6tFsOJfQW0bQ+3Dexqx8A2fCrpp3ZpL16ftUKcCb7TbBRN6iFtE1RUeELTxHsKLXN1HsfU0KrM22WJLVg+GJa7kWLlBA/jsmIER1JapjHVHV8eNAwNy0SWCTqknZgq9NalqZbPRkKqsury0XxZpmmgCUKF9sGIBw7KA/eYaxVQ+FqD9o0jqSAwKqZrrjiPJYTedv1bxlFAgZpzB6sHbdO5Z5JMrWz0NnTMcExTYEb7zDl3RgguIGK3K7joFU1sWDJQjEgUm9W16kFyBVL2iNFhf3MS8tnXrfqeQnGPNWHttnZZO62iVSxCNok0BBcEu8B4E98BZFOBciHvw7ILXfZ4Dy5toRor7KTsw5MufHINVnvQQ8enXPhU3mIPspP5FXh6QzkPX5NyPsNQEZ912AmmZe5yDz7tsP+taX29Gjs15OZAQqzGfH1iVfNnDcXjyiRDvVi/XnrxmXG9eG723xY7p1F/3oXPIurPadQ/h+B93oUvbID3xQ3w7r0m8L6kwfvyCLyv9OCr1wLeC5uA90b9fREhfMlA+PL1gvCVcQhfRcll59gsQxX9GgL5dRe+gUC+2mTXQ6K+qSWK/d/I0rcUfNthAQsZZW3WYYwZCfkOW2IxQ5n4LnvlrWTitc1l4nuasN934QeI82uasD9Ewv7IhR/34Scu/FTLxM+2kImfG5n4hQu/XINf9eB1dPy1C7/ZYPpv6Vt+US2agBjLht8hs3/fgz84hg6DlXCBP27lP5hhP2ReXAZx8GPnTxjlz0as8V2500GKpvCXrUIMp9inB6fPwvAn/BWD/M0kjMXLkzwmut768KXw94Zlzl4NrFQkyXx8+23hm6qAf+CQ+ViLpD860P65nrcU/Kv6H4PT0vg=",
"user_name": "write/Write/WriteImpl/FinalizeWrite/FinalizeWrite"
}
}
],
"type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
createTime: '2019-09-30T12:51:39.615163Z'
currentStateTime: '1970-01-01T00:00:00Z'
id: '2019-09-30_05_51_38-14277359654686163704'
location: 'us-central1'
name: 'beamapp-jenkins-0930125136-596838'
projectId: 'apache-beam-testing'
stageStates: []
startTime: '2019-09-30T12:51:39.615163Z'
steps: []
tempFiles: []
type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2019-09-30_05_51_38-14277359654686163704]
root: INFO: To access the Dataflow monitoring console, please navigate to
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2019-09-30_05_51_38-14277359654686163704?project=apache-beam-testing
root: INFO: Job 2019-09-30_05_51_38-14277359654686163704 is in state
JOB_STATE_RUNNING
root: INFO: 2019-09-30T12:51:41.708Z: JOB_MESSAGE_DETAILED: Checking
permissions granted to controller Service Account.
root: INFO: 2019-09-30T12:51:42.226Z: JOB_MESSAGE_BASIC: Worker configuration:
n1-standard-1 in us-central1-a.
root: INFO: 2019-09-30T12:51:42.949Z: JOB_MESSAGE_DETAILED: Expanding
CoGroupByKey operations into optimizable parts.
root: INFO: 2019-09-30T12:51:42.989Z: JOB_MESSAGE_DEBUG: Combiner lifting
skipped for step write/Write/WriteImpl/GroupByKey: GroupByKey not followed by a
combiner.
root: INFO: 2019-09-30T12:51:43.022Z: JOB_MESSAGE_DEBUG: Combiner lifting
skipped for step group: GroupByKey not followed by a combiner.
root: INFO: 2019-09-30T12:51:43.062Z: JOB_MESSAGE_DETAILED: Expanding
GroupByKey operations into optimizable parts.
root: INFO: 2019-09-30T12:51:43.093Z: JOB_MESSAGE_DETAILED: Lifting
ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2019-09-30T12:51:43.179Z: JOB_MESSAGE_DEBUG: Annotating graph with
Autotuner information.
root: INFO: 2019-09-30T12:51:43.230Z: JOB_MESSAGE_DETAILED: Fusing adjacent
ParDo, Read, Write, and Flatten operations
root: INFO: 2019-09-30T12:51:43.267Z: JOB_MESSAGE_DETAILED: Fusing consumer
split into read/Read
root: INFO: 2019-09-30T12:51:43.303Z: JOB_MESSAGE_DETAILED: Fusing consumer
pair_with_one into split
root: INFO: 2019-09-30T12:51:43.338Z: JOB_MESSAGE_DETAILED: Fusing consumer
group/Reify into pair_with_one
root: INFO: 2019-09-30T12:51:43.376Z: JOB_MESSAGE_DETAILED: Fusing consumer
group/Write into group/Reify
root: INFO: 2019-09-30T12:51:43.412Z: JOB_MESSAGE_DETAILED: Fusing consumer
group/GroupByWindow into group/Read
root: INFO: 2019-09-30T12:51:43.446Z: JOB_MESSAGE_DETAILED: Fusing consumer
count into group/GroupByWindow
root: INFO: 2019-09-30T12:51:43.482Z: JOB_MESSAGE_DETAILED: Fusing consumer
format into count
root: INFO: 2019-09-30T12:51:43.522Z: JOB_MESSAGE_DETAILED: Fusing consumer
write/Write/WriteImpl/WriteBundles/WriteBundles into format
root: INFO: 2019-09-30T12:51:43.544Z: JOB_MESSAGE_DETAILED: Fusing consumer
write/Write/WriteImpl/Pair into write/Write/WriteImpl/WriteBundles/WriteBundles
root: INFO: 2019-09-30T12:51:43.575Z: JOB_MESSAGE_DETAILED: Fusing consumer
write/Write/WriteImpl/WindowInto(WindowIntoFn) into write/Write/WriteImpl/Pair
root: INFO: 2019-09-30T12:51:43.611Z: JOB_MESSAGE_DETAILED: Fusing consumer
write/Write/WriteImpl/GroupByKey/Reify into
write/Write/WriteImpl/WindowInto(WindowIntoFn)
root: INFO: 2019-09-30T12:51:43.649Z: JOB_MESSAGE_DETAILED: Fusing consumer
write/Write/WriteImpl/GroupByKey/Write into
write/Write/WriteImpl/GroupByKey/Reify
root: INFO: 2019-09-30T12:51:43.683Z: JOB_MESSAGE_DETAILED: Fusing consumer
write/Write/WriteImpl/GroupByKey/GroupByWindow into
write/Write/WriteImpl/GroupByKey/Read
root: INFO: 2019-09-30T12:51:43.721Z: JOB_MESSAGE_DETAILED: Fusing consumer
write/Write/WriteImpl/Extract into
write/Write/WriteImpl/GroupByKey/GroupByWindow
root: INFO: 2019-09-30T12:51:43.755Z: JOB_MESSAGE_DETAILED: Fusing consumer
write/Write/WriteImpl/InitializeWrite into write/Write/WriteImpl/DoOnce/Read
root: INFO: 2019-09-30T12:51:43.795Z: JOB_MESSAGE_DEBUG: Workflow config is
missing a default resource spec.
root: INFO: 2019-09-30T12:51:43.829Z: JOB_MESSAGE_DEBUG: Adding StepResource
setup and teardown to workflow graph.
root: INFO: 2019-09-30T12:51:43.866Z: JOB_MESSAGE_DEBUG: Adding workflow start
and stop steps.
root: INFO: 2019-09-30T12:51:43.903Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2019-09-30T12:51:44.046Z: JOB_MESSAGE_DEBUG: Executing wait step
start26
root: INFO: 2019-09-30T12:51:44.122Z: JOB_MESSAGE_BASIC: Executing operation
write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
root: INFO: 2019-09-30T12:51:44.159Z: JOB_MESSAGE_BASIC: Executing operation
write/Write/WriteImpl/GroupByKey/Create
root: INFO: 2019-09-30T12:51:44.171Z: JOB_MESSAGE_DEBUG: Starting worker pool
setup.
root: INFO: 2019-09-30T12:51:44.186Z: JOB_MESSAGE_BASIC: Executing operation
group/Create
root: INFO: 2019-09-30T12:51:44.207Z: JOB_MESSAGE_BASIC: Starting 10 workers in
us-central1-a...
root: INFO: 2019-09-30T12:51:44.279Z: JOB_MESSAGE_BASIC: Finished operation
group/Create
root: INFO: 2019-09-30T12:51:44.279Z: JOB_MESSAGE_BASIC: Finished operation
write/Write/WriteImpl/GroupByKey/Create
root: INFO: 2019-09-30T12:51:44.357Z: JOB_MESSAGE_DEBUG: Value
"write/Write/WriteImpl/GroupByKey/Session" materialized.
root: INFO: 2019-09-30T12:51:44.388Z: JOB_MESSAGE_DEBUG: Value "group/Session"
materialized.
root: INFO: 2019-09-30T12:51:44.454Z: JOB_MESSAGE_BASIC: Executing operation
read/Read+split+pair_with_one+group/Reify+group/Write
root: INFO: 2019-09-30T12:52:18.773Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised
the number of workers to 10 based on the rate of progress in the currently
running step(s).
root: INFO: 2019-09-30T12:52:53.891Z: JOB_MESSAGE_DETAILED: Workers have
started successfully.
root: INFO: 2019-09-30T12:52:53.919Z: JOB_MESSAGE_DETAILED: Workers have
started successfully.
root: INFO: 2019-09-30T12:57:15.180Z: JOB_MESSAGE_ERROR: Traceback (most recent
call last):
File "/usr/local/lib/python3.6/site-packages/dataflow_worker/batchworker.py",
line 773, in run
self._load_main_session(self.local_staging_directory)
File "/usr/local/lib/python3.6/site-packages/dataflow_worker/batchworker.py",
line 489, in _load_main_session
pickler.load_session(session_file)
File
"/usr/local/lib/python3.6/site-packages/apache_beam/internal/pickler.py", line
287, in load_session
return dill.load_session(file_path)
File "/usr/local/lib/python3.6/site-packages/dill/_dill.py", line 410, in
load_session
module = unpickler.load()
TypeError: _create_function() takes from 2 to 6 positional arguments but 7 were
given
root: INFO: 2019-09-30T12:57:17.261Z: JOB_MESSAGE_ERROR: Traceback (most recent
call last):
File "/usr/local/lib/python3.6/site-packages/dataflow_worker/batchworker.py",
line 773, in run
self._load_main_session(self.local_staging_directory)
File "/usr/local/lib/python3.6/site-packages/dataflow_worker/batchworker.py",
line 489, in _load_main_session
pickler.load_session(session_file)
File
"/usr/local/lib/python3.6/site-packages/apache_beam/internal/pickler.py", line
287, in load_session
return dill.load_session(file_path)
File "/usr/local/lib/python3.6/site-packages/dill/_dill.py", line 410, in
load_session
module = unpickler.load()
TypeError: _create_function() takes from 2 to 6 positional arguments but 7 were
given
root: INFO: 2019-09-30T12:57:19.335Z: JOB_MESSAGE_ERROR: Traceback (most recent
call last):
File "/usr/local/lib/python3.6/site-packages/dataflow_worker/batchworker.py",
line 773, in run
self._load_main_session(self.local_staging_directory)
File "/usr/local/lib/python3.6/site-packages/dataflow_worker/batchworker.py",
line 489, in _load_main_session
pickler.load_session(session_file)
File
"/usr/local/lib/python3.6/site-packages/apache_beam/internal/pickler.py", line
287, in load_session
return dill.load_session(file_path)
File "/usr/local/lib/python3.6/site-packages/dill/_dill.py", line 410, in
load_session
module = unpickler.load()
TypeError: _create_function() takes from 2 to 6 positional arguments but 7 were
given
root: INFO: 2019-09-30T12:57:21.402Z: JOB_MESSAGE_ERROR: Traceback (most recent
call last):
File "/usr/local/lib/python3.6/site-packages/dataflow_worker/batchworker.py",
line 773, in run
self._load_main_session(self.local_staging_directory)
File "/usr/local/lib/python3.6/site-packages/dataflow_worker/batchworker.py",
line 489, in _load_main_session
pickler.load_session(session_file)
File
"/usr/local/lib/python3.6/site-packages/apache_beam/internal/pickler.py", line
287, in load_session
return dill.load_session(file_path)
File "/usr/local/lib/python3.6/site-packages/dill/_dill.py", line 410, in
load_session
module = unpickler.load()
TypeError: _create_function() takes from 2 to 6 positional arguments but 7 were
given
root: INFO: 2019-09-30T12:57:21.611Z: JOB_MESSAGE_ERROR: Traceback (most recent
call last):
File "/usr/local/lib/python3.6/site-packages/dataflow_worker/batchworker.py",
line 773, in run
self._load_main_session(self.local_staging_directory)
File "/usr/local/lib/python3.6/site-packages/dataflow_worker/batchworker.py",
line 489, in _load_main_session
pickler.load_session(session_file)
File
"/usr/local/lib/python3.6/site-packages/apache_beam/internal/pickler.py", line
287, in load_session
return dill.load_session(file_path)
File "/usr/local/lib/python3.6/site-packages/dill/_dill.py", line 410, in
load_session
module = unpickler.load()
TypeError: _create_function() takes from 2 to 6 positional arguments but 7 were
given
root: INFO: 2019-09-30T12:57:21.640Z: JOB_MESSAGE_BASIC: Finished operation
read/Read+split+pair_with_one+group/Reify+group/Write
root: INFO: 2019-09-30T12:57:21.703Z: JOB_MESSAGE_DEBUG: Executing failure step
failure25
root: INFO: 2019-09-30T12:57:21.732Z: JOB_MESSAGE_ERROR: Workflow failed.
Causes: S06:read/Read+split+pair_with_one+group/Reify+group/Write failed.,
Internal Issue (92ba1b7412fbabf6): 63963027:24514
root: INFO: 2019-09-30T12:57:22.074Z: JOB_MESSAGE_WARNING:
S01:write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
failed.
root: INFO: 2019-09-30T12:57:22.113Z: JOB_MESSAGE_BASIC: Finished operation
write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
root: INFO: 2019-09-30T12:57:22.228Z: JOB_MESSAGE_DETAILED: Cleaning up.
root: INFO: 2019-09-30T12:57:22.294Z: JOB_MESSAGE_DEBUG: Starting worker pool
teardown.
root: INFO: 2019-09-30T12:57:22.328Z: JOB_MESSAGE_BASIC: Stopping worker pool...
root: INFO: 2019-09-30T13:03:40.394Z: JOB_MESSAGE_DETAILED: Autoscaling:
Reduced the number of workers to 0 based on the rate of progress in the
currently running step(s).
root: INFO: 2019-09-30T13:03:40.448Z: JOB_MESSAGE_BASIC: Worker pool stopped.
root: INFO: 2019-09-30T13:03:40.485Z: JOB_MESSAGE_DEBUG: Tearing down pending
resources...
root: INFO: Job 2019-09-30_05_51_38-14277359654686163704 is in state
JOB_STATE_FAILED
apache_beam.io.filesystem: DEBUG: Listing files in
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1569847895543/results'
apache_beam.io.filesystem: DEBUG: translate_pattern:
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1569847895543/results*'
->
'gs\\:\\/\\/temp\\-storage\\-for\\-end\\-to\\-end\\-tests\\/py\\-it\\-cloud\\/output\\/1569847895543\\/results[^/\\\\]*'
root: INFO: Starting the size estimation of the input
root: INFO: Finished listing 0 files in 0.06242084503173828 seconds.
--------------------- >> end captured logging << ---------------------
----------------------------------------------------------------------
XML: nosetests-integrationTest-perf.xml
----------------------------------------------------------------------
XML:
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 1 test in 735.106s
FAILED (errors=1)
FAILURE: Build failed with an exception.
* What went wrong:
Execution failed for task
':sdks:python:test-suites:dataflow:py36:integrationTest'.
> Process 'command 'sh'' finished with non-zero exit value 1
* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to
get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
BUILD FAILED in 13m 2s
2019-09-30 13:03:55,553 afee3381 MainThread beam_integration_benchmark(1/1)
ERROR Error during benchmark beam_integration_benchmark
Traceback (most recent call last):
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",>
line 841, in RunBenchmark
DoRunPhase(spec, collector, detailed_timer)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",>
line 687, in DoRunPhase
samples = spec.BenchmarkRun(spec)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",>
line 160, in Run
job_type=job_type)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",>
line 91, in SubmitJob
assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2019-09-30 13:03:55,566 afee3381 MainThread beam_integration_benchmark(1/1)
INFO Cleaning up benchmark beam_integration_benchmark
2019-09-30 13:03:55,569 afee3381 MainThread beam_integration_benchmark(1/1)
ERROR Exception running benchmark
Traceback (most recent call last):
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",>
line 984, in RunBenchmarkTask
RunBenchmark(spec, collector)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",>
line 841, in RunBenchmark
DoRunPhase(spec, collector, detailed_timer)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",>
line 687, in DoRunPhase
samples = spec.BenchmarkRun(spec)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",>
line 160, in Run
job_type=job_type)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",>
line 91, in SubmitJob
assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2019-09-30 13:03:55,574 afee3381 MainThread beam_integration_benchmark(1/1)
ERROR Benchmark 1/1 beam_integration_benchmark (UID:
beam_integration_benchmark0) failed. Execution will continue.
2019-09-30 13:03:55,575 afee3381 MainThread beam_integration_benchmark(1/1)
INFO Benchmark run statuses:
---------------------------------------------------------------------------------
Name UID Status Failed
Substatus
---------------------------------------------------------------------------------
beam_integration_benchmark beam_integration_benchmark0 FAILED
---------------------------------------------------------------------------------
Success rate: 0.00% (0/1)
2019-09-30 13:03:55,586 afee3381 MainThread beam_integration_benchmark(1/1)
INFO Complete logs can be found at:
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/runs/afee3381/pkb.log>
2019-09-30 13:03:55,587 afee3381 MainThread beam_integration_benchmark(1/1)
INFO Completion statuses can be found at:
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/runs/afee3381/completion_statuses.json>
Build step 'Execute shell' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]