See
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/1326/display/redirect?page=changes>
Changes:
[xhan] Documentation bug fix for FlatMapElements#via() SimpleFunction in the
[github] [BEAM-7923] Screendiff Integration Tests (#11338)
[github] [BEAM-9642] Add SDF execution units. (#11327)
------------------------------------------
[...truncated 155.05 KB...]
"pipeline_proto_coder_id": "ref_Coder_FastPrimitivesCoder_5"
},
{
"@type": "kind:global_window"
}
],
"is_wrapper": true
},
"output_name": "None",
"user_name": "write/Write/WriteImpl/FinalizeWrite.out"
}
],
"parallel_input": {
"@type": "OutputReference",
"output_name": "out",
"step_name": "s7"
},
"serialized_fn":
"eNrNWHd8HMUV3ruTbXltg0twcEi5GAwr4O4sGzvGIRA426AcPisngTfFWfZ2527W2vZmZiWLeDG2I8kxSQhphIT0QnoP6b0XSA+kJ6SH9Eqq82buTvIpsuL/4Cdp56a8t++97733zen6nOHYse1QYtWJHRQFs0PeiFjAi07EiF62fd+u+2Q3s+OYsG3RjlAHre8gZFLIGmZO0zSrEUKuS4kX4W/d5kS3Gl5o+961xBpjniA69JhLpIgYj4lFvVBwWOC4nu8XLfnULYcRWxArtAPiiiT2UWShEhnYNYwyVyiRRWrFC+NEKE0ces2luBIlYmZpsblQLjGv6YWgT8ESM4/zLl/xnLKh6BJ01hYR47B0CpbV4JRqdQ+cOgXLa7DCXIRyMYscwjmspD3qVVYjwcCwJo+JA6uqSX0PPMiYHQRB9gkv0q1hHIe8cESH0zByq1N4sGGuQCWc2sxVvloy5LaA082zcL2wzmhvJUHf+k1uIWrgEk4stczlGqwxV3eraPpRvaPnIWqzcO6ckmeYi3Ez8FBGhgAeauo4l8aWYt/GaD1MuWxRYruEwcOr5nKcNjy/hYvFk0bD2weP6HI3itHVkBdHbT8hFoZr1ENhfUjYwnOulouDnTXIYxQemcJaw1yAipUEnGniRw3OUra0lCjj1nWlhx/ZrlrX4WyVfFwwOGcCjBr0JXVzjTRbJmqIp5jt+V7YtEIyhiNmxLlH1PucSHp1Xpf1ao23B304umxcEF6OlLXno7WFFIp1FQYnCmKGmYDOtgwsmT0yrxIRwfqZQMW2oBgFIgPVT6XDG1LYaNAzzV14pMm3lkqCBHGBY9LZTVJA3ApodUFErYFwwUvxeMETBcePErfUyu1S/6Ytm7dcuOGCTZs3b+wvoSGJjwVxAV1Hpf8ydjNgw6aKltS7kzIUhGFFFhPh+fqlrJkEJBSDvu0QGvnK3c1o66OofG6pZNR4IUlh6x54tGFulpUwLij6zhFKS5Xg+oKq7dLumedAEPulHe3SVytwUZcdsUJYt64KY88Z8Yk7hOoGpDYdHpPCxYYKqmsLGy6ZS3D6+DY8osNj0cpLU7jMUIk76pEx2ZTKc3WWRhI6MlN12GbQs81ePC9xls0Ftk/AjhpcblQyFQ3/cpVV5WVHNG2/ph3JaIez2hBcUZ2AgT4lhU4x2RnhcRNQMRmulGgUkNJeEo54Ie+MBe7bo6Q0FrERjn6QknTDGiRMVWrokGEJtbU7Ym45SkIxMGwNjm/cVOLMKXF3RCaBjHfpuCCUWkEoxuNwpTLlIt8O6q59MezceTRT1qBqnibzkEWBxVClLPRpa3ep9qVC1C5ZGJyEx/cJqNVgqCtiTSIsWwjMiWH1mnri+QJdgqtUmHFb7sLVk7C7BmaXqBfEERNWELmJbOFPMFfO6r0tF+CJE/CkGjxZqbdQ1hGWBXsm4Sk1sOjO6lwAOgQncA1F+BCpLCKVqyyqlMuHBPawjLY3K/FyFV6HM1qK06zG+7X9uJXT3KwmerTDmkJTLJB7eMjNaSNZjVXk1O3R1gyJhZq7QJveF4s6OzOTnJo0ctpqHA6hDg3TwzaqlawKjksaNpYm1GUumdfhymA58n2ici8fNfIcKyO/zs2PeYLmA6TavKA27oQkT3wiqzJvO5JyiJu3ed5GgbDpE4HSErxifofHuMiLsahznudJ6MgUIkzKoMa16/ja89WzCI4At1VVvscFEMUBMitEFPkcGootPe57DoGmaq2IL1DzFPy0PYjF+HTJgae2fRLCXtVxVHffzljEYISeI8A3s0o3BCoUnTQL1dslqUA0BTG9UnGUNXeaws77MuXlWvb0zMLMqszKTG8ml8llgfVhovIaYHOlUK2mkFQFjNZgbAL21WCcXpPCtcZxXf6p83T53V7oRmPEVea3e/1+bCRpCtcZ5irUMaYuPa7VoqOWxgPzaNxhczHIvMAT3ug0fVyPKg+mcMg8VXmL96m67Yy0tR2eR9ug6oxtLU9DLRMpTNZ5XSmS0eLCDuK2oqn56Kxztq3rCOp6egpH66oVjKkwtNXcMI+ay/F+YfutqLU1PQM1PTOFZ6nkaYUJbjzRfbL1oi41OjwbNdyUwnNUblhIffDcLnnJU7w47e2MLzo8D0Wfn8IL1MsDz2ERh5uP9t59q3HPvceOHeB1+YOkuGLGyQDxlFcCJIcXnshMSWyK17iurlS2b/2PtA63JBPwohq8eBJuTeElmIQvrcHLVLG0qlZeACTVwsvNM2TrxJdsnaHMrZ083zraD684gYWvpLckdXoelTn5qhRebdADVCbTa1J4LT1MZUbclsLr0Ek6RSWmr0/hDXV6A5WovDGFN9Eb6U0yAlRqmoI3z03f/SdF32+hSMtvNeglVHLt21J4u0HL7R7cYstMZajcs192wXcgSb6zjyIvvuv+4sV3d/Pi7TvXZmlV9o731OC92Dtul73jfQjb+2vwAfr/MfogPR6ND6Xw4Q4aH0nho4jGbdMofCyFj3dQ+EQKn5yFwqfmRmHDSaHwaYnCZzoofDaFz81GYY47y+cRji8oOL54f8FxRzccd+I1hVbpLoo3jy8hKF+uwVcQlDur9IFJ61+VtE4fMFT+NQFfNyihDdqklHp0L1W0+w0a0JAitX6T3jEftd41N7XeLcvjWzX4NiJxlyyP72B5fLcG35uA79fgB5Jaf3g8tf5IlcI9Kfy4Uwo/SeGn7cb0sxR+PtOYfpHCLzsl8asU7u2UxAka36+p7K6/qcFvJ+F3Kfwe7fhDDf54EmX6p64y/XMKf+nY9tcU/tZVpvel8PeOTf9I4Z+zyvRfZN5/i7ToC+3V4d+o4D8pHDPUjUowr9nEqgmZlsFInUBH+5S+rXVLHG5PWQZl+g6yLA7qxnw8ywd1/BLLWA73KpnWf1E8brXvmawHl4+oKmvFEw1YMI8Bc3Ax19nC1usXydfLb7J4V5EXJMvHezd+geasd7bK2fS8LWG2+nbFFrd06TikbAk+6c0Vjbcgx2pIgsRXJ+V3BMKWKq+SumDL8FPxvxRVZNo=",
"user_name": "write/Write/WriteImpl/FinalizeWrite/FinalizeWrite"
}
}
],
"type": "JOB_TYPE_BATCH"
}
INFO:apache_beam.runners.dataflow.internal.apiclient:Create job: <Job
createTime: '2020-04-14T19:27:41.995875Z'
currentStateTime: '1970-01-01T00:00:00Z'
id: '2020-04-14_12_27_40-6423619403806197362'
location: 'us-central1'
name: 'beamapp-jenkins-0414192738-103970'
projectId: 'apache-beam-testing'
stageStates: []
startTime: '2020-04-14T19:27:41.995875Z'
steps: []
tempFiles: []
type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
INFO:apache_beam.runners.dataflow.internal.apiclient:Created job with id:
[2020-04-14_12_27_40-6423619403806197362]
INFO:apache_beam.runners.dataflow.internal.apiclient:To access the Dataflow
monitoring console, please navigate to
https://console.cloud.google.com/dataflow/jobs/us-central1/2020-04-14_12_27_40-6423619403806197362?project=apache-beam-testing
INFO:apache_beam.runners.dataflow.dataflow_runner:Job
2020-04-14_12_27_40-6423619403806197362 is in state JOB_STATE_RUNNING
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:44.233Z:
JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service
Account.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:44.742Z:
JOB_MESSAGE_BASIC: Worker configuration: n1-standard-1 in us-central1-f.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.408Z:
JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.435Z:
JOB_MESSAGE_DEBUG: Combiner lifting skipped for step
write/Write/WriteImpl/GroupByKey: GroupByKey not followed by a combiner.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.464Z:
JOB_MESSAGE_DEBUG: Combiner lifting skipped for step group: GroupByKey not
followed by a combiner.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.505Z:
JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into optimizable parts.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.539Z:
JOB_MESSAGE_DETAILED: Lifting ValueCombiningMappingFns into
MergeBucketsMappingFns
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.626Z:
JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.683Z:
JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.704Z:
JOB_MESSAGE_DETAILED: Fusing consumer split into read/Read
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.728Z:
JOB_MESSAGE_DETAILED: Fusing consumer pair_with_one into split
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.756Z:
JOB_MESSAGE_DETAILED: Fusing consumer group/Reify into pair_with_one
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.781Z:
JOB_MESSAGE_DETAILED: Fusing consumer group/Write into group/Reify
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.808Z:
JOB_MESSAGE_DETAILED: Fusing consumer group/GroupByWindow into group/Read
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.831Z:
JOB_MESSAGE_DETAILED: Fusing consumer count into group/GroupByWindow
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.863Z:
JOB_MESSAGE_DETAILED: Fusing consumer format into count
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.887Z:
JOB_MESSAGE_DETAILED: Fusing consumer
write/Write/WriteImpl/WriteBundles/WriteBundles into format
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.915Z:
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/Pair into
write/Write/WriteImpl/WriteBundles/WriteBundles
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.944Z:
JOB_MESSAGE_DETAILED: Fusing consumer
write/Write/WriteImpl/WindowInto(WindowIntoFn) into write/Write/WriteImpl/Pair
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.974Z:
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/GroupByKey/Reify
into write/Write/WriteImpl/WindowInto(WindowIntoFn)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.005Z:
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/GroupByKey/Write
into write/Write/WriteImpl/GroupByKey/Reify
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.032Z:
JOB_MESSAGE_DETAILED: Fusing consumer
write/Write/WriteImpl/GroupByKey/GroupByWindow into
write/Write/WriteImpl/GroupByKey/Read
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.061Z:
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/Extract into
write/Write/WriteImpl/GroupByKey/GroupByWindow
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.085Z:
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/InitializeWrite
into write/Write/WriteImpl/DoOnce/Read
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.108Z:
JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.135Z:
JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.166Z:
JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.192Z:
JOB_MESSAGE_DEBUG: Assigning stage ids.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.388Z:
JOB_MESSAGE_DEBUG: Executing wait step start26
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.450Z:
JOB_MESSAGE_BASIC: Executing operation
write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.471Z:
JOB_MESSAGE_BASIC: Executing operation write/Write/WriteImpl/GroupByKey/Create
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.491Z:
JOB_MESSAGE_DEBUG: Starting worker pool setup.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.496Z:
JOB_MESSAGE_BASIC: Executing operation group/Create
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.511Z:
JOB_MESSAGE_BASIC: Starting 10 workers in us-central1-f...
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.561Z:
JOB_MESSAGE_BASIC: Finished operation group/Create
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.561Z:
JOB_MESSAGE_BASIC: Finished operation write/Write/WriteImpl/GroupByKey/Create
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.619Z:
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/GroupByKey/Session"
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.641Z:
JOB_MESSAGE_DEBUG: Value "group/Session" materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.695Z:
JOB_MESSAGE_BASIC: Executing operation
read/Read+split+pair_with_one+group/Reify+group/Write
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:28:12.207Z:
JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 3 based on
the rate of progress in the currently running stage(s).
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:28:12.231Z:
JOB_MESSAGE_DETAILED: Resized worker pool to 3, though goal was 10. This could
be a quota issue.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:28:17.677Z:
JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 9 based on
the rate of progress in the currently running stage(s).
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:28:17.709Z:
JOB_MESSAGE_DETAILED: Resized worker pool to 9, though goal was 10. This could
be a quota issue.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:28:23.150Z:
JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 10 based on
the rate of progress in the currently running stage(s).
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:28:23.792Z:
JOB_MESSAGE_WARNING: Your project already contains 100 Dataflow-created metric
descriptors and Stackdriver will not create new Dataflow custom metrics for
this job. Each unique user-defined metric name (independent of the DoFn in
which it is defined) produces a new metric descriptor. To delete old / unused
metric descriptors see
https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list
and
https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:29:35.460Z:
JOB_MESSAGE_DETAILED: Workers have started successfully.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:29:35.494Z:
JOB_MESSAGE_DETAILED: Workers have started successfully.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.266Z:
JOB_MESSAGE_BASIC: Finished operation
write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.411Z:
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/DoOnce/Read.out" materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.446Z:
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/InitializeWrite.out"
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.525Z:
JOB_MESSAGE_BASIC: Executing operation
write/Write/WriteImpl/WriteBundles/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.556Z:
JOB_MESSAGE_BASIC: Executing operation
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.577Z:
JOB_MESSAGE_BASIC: Finished operation
write/Write/WriteImpl/WriteBundles/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.587Z:
JOB_MESSAGE_BASIC: Executing operation
write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.602Z:
JOB_MESSAGE_BASIC: Finished operation
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.631Z:
JOB_MESSAGE_BASIC: Finished operation
write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.709Z:
JOB_MESSAGE_DEBUG: Value
"write/Write/WriteImpl/WriteBundles/_UnpickledSideInput(InitializeWrite.out.0).output"
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.739Z:
JOB_MESSAGE_DEBUG: Value
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0).output"
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.775Z:
JOB_MESSAGE_DEBUG: Value
"write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(InitializeWrite.out.0).output"
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:33:46.366Z:
JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service
Account.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:34:55.189Z:
JOB_MESSAGE_BASIC: Finished operation
read/Read+split+pair_with_one+group/Reify+group/Write
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:34:55.249Z:
JOB_MESSAGE_BASIC: Executing operation group/Close
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:34:55.291Z:
JOB_MESSAGE_BASIC: Finished operation group/Close
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:34:55.362Z:
JOB_MESSAGE_BASIC: Executing operation
group/Read+group/GroupByWindow+count+format+write/Write/WriteImpl/WriteBundles/WriteBundles+write/Write/WriteImpl/Pair+write/Write/WriteImpl/WindowInto(WindowIntoFn)+write/Write/WriteImpl/GroupByKey/Reify+write/Write/WriteImpl/GroupByKey/Write
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:49.249Z:
JOB_MESSAGE_BASIC: Finished operation
group/Read+group/GroupByWindow+count+format+write/Write/WriteImpl/WriteBundles/WriteBundles+write/Write/WriteImpl/Pair+write/Write/WriteImpl/WindowInto(WindowIntoFn)+write/Write/WriteImpl/GroupByKey/Reify+write/Write/WriteImpl/GroupByKey/Write
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:49.320Z:
JOB_MESSAGE_BASIC: Executing operation write/Write/WriteImpl/GroupByKey/Close
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:49.364Z:
JOB_MESSAGE_BASIC: Finished operation write/Write/WriteImpl/GroupByKey/Close
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:49.420Z:
JOB_MESSAGE_BASIC: Executing operation
write/Write/WriteImpl/GroupByKey/Read+write/Write/WriteImpl/GroupByKey/GroupByWindow+write/Write/WriteImpl/Extract
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:51.723Z:
JOB_MESSAGE_BASIC: Finished operation
write/Write/WriteImpl/GroupByKey/Read+write/Write/WriteImpl/GroupByKey/GroupByWindow+write/Write/WriteImpl/Extract
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:51.895Z:
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/Extract.out" materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:51.960Z:
JOB_MESSAGE_BASIC: Executing operation
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:51.983Z:
JOB_MESSAGE_BASIC: Executing operation
write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(Extract.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:52Z:
JOB_MESSAGE_BASIC: Finished operation
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:52.026Z:
JOB_MESSAGE_BASIC: Finished operation
write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(Extract.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:52.058Z:
JOB_MESSAGE_DEBUG: Value
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0).output"
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:52.090Z:
JOB_MESSAGE_DEBUG: Value
"write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(Extract.out.0).output"
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:52.155Z:
JOB_MESSAGE_BASIC: Executing operation
write/Write/WriteImpl/PreFinalize/PreFinalize
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:54.987Z:
JOB_MESSAGE_BASIC: Finished operation
write/Write/WriteImpl/PreFinalize/PreFinalize
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:55.043Z:
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/PreFinalize.out" materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:55.100Z:
JOB_MESSAGE_BASIC: Executing operation
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(PreFinalize.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:55.145Z:
JOB_MESSAGE_BASIC: Finished operation
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(PreFinalize.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:55.205Z:
JOB_MESSAGE_DEBUG: Value
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(PreFinalize.out.0).output"
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:55.267Z:
JOB_MESSAGE_BASIC: Executing operation
write/Write/WriteImpl/FinalizeWrite/FinalizeWrite
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:57.745Z:
JOB_MESSAGE_BASIC: Finished operation
write/Write/WriteImpl/FinalizeWrite/FinalizeWrite
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:57.808Z:
JOB_MESSAGE_DEBUG: Executing success step success24
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:57.918Z:
JOB_MESSAGE_DETAILED: Cleaning up.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:58.009Z:
JOB_MESSAGE_DEBUG: Starting worker pool teardown.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:58.029Z:
JOB_MESSAGE_BASIC: Stopping worker pool...
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:38:51.873Z:
JOB_MESSAGE_DETAILED: Autoscaling: Resized worker pool from 10 to 0.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:38:51.915Z:
JOB_MESSAGE_BASIC: Worker pool stopped.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:38:51.955Z:
JOB_MESSAGE_DEBUG: Tearing down pending resources...
INFO:apache_beam.runners.dataflow.dataflow_runner:Job
2020-04-14_12_27_40-6423619403806197362 is in state JOB_STATE_DONE
DEBUG:apache_beam.io.filesystem:Listing files in
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results'
DEBUG:apache_beam.io.filesystem:translate_pattern:
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results*-of-*'
->
'gs\\:\\/\\/temp\\-storage\\-for\\-end\\-to\\-end\\-tests\\/py\\-it\\-cloud\\/output\\/1586892456631\\/results[^/\\\\]*\\-of\\-[^/\\\\]*'
INFO:apache_beam.io.gcp.gcsio:Starting the size estimation of the input
INFO:apache_beam.io.gcp.gcsio:Finished listing 30 files in 0.06600570678710938
seconds.
INFO:apache_beam.testing.pipeline_verifiers:Find 30 files in
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results*-of-*:
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00023-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00012-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00028-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00029-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00003-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00002-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00025-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00015-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00026-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00006-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00008-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00010-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00016-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00019-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00013-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00011-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00018-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00021-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00000-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00005-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00024-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00017-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00027-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00009-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00022-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00020-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00014-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00001-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00007-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00004-of-00030
INFO:apache_beam.testing.pipeline_verifiers:Read from given path
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results*-of-*,
26186927 lines, checksum: ea0ca2e5ee4ea5f218790f28d0b9fe7d09d8d710.
INFO:root:average word length: 19
DEBUG:apache_beam.io.filesystem:Listing files in
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results'
DEBUG:apache_beam.io.filesystem:translate_pattern:
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results*'
->
'gs\\:\\/\\/temp\\-storage\\-for\\-end\\-to\\-end\\-tests\\/py\\-it\\-cloud\\/output\\/1586892456631\\/results[^/\\\\]*'
INFO:apache_beam.io.gcp.gcsio:Starting the size estimation of the input
INFO:apache_beam.io.gcp.gcsio:Finished listing 30 files in 0.07889842987060547
seconds.
error: [Errno 111] Connection refused
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/src/sdks/python/scripts/run_integration_test.sh>:
line 278: 20406 Terminated python setup.py nosetests
--test-pipeline-options="$PIPELINE_OPTS" --with-xunitmp
--xunitmp-file=$XUNIT_FILE --ignore-files '.*py3\d?\.py$' $TEST_OPTS
The message received from the daemon indicates that the daemon has disappeared.
Build request sent: Build{id=784b86ac-0811-4204-b98e-d368e6c6a074,
currentDir=<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/src}>
Attempting to read last messages from the daemon log...
Daemon pid: 24086
log file: /home/jenkins/.gradle/daemon/5.2.1/daemon-24086.out.log
----- Last 20 lines from daemon log file - daemon-24086.out.log -----
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00017-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00027-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00009-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00022-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00020-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00014-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00001-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00007-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00004-of-00030
INFO:apache_beam.testing.pipeline_verifiers:Read from given path
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results*-of-*,
26186927 lines, checksum: ea0ca2e5ee4ea5f218790f28d0b9fe7d09d8d710.
INFO:root:average word length: 19
DEBUG:apache_beam.io.filesystem:Listing files in
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results'
DEBUG:apache_beam.io.filesystem:translate_pattern:
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results*'
->
'gs\\:\\/\\/temp\\-storage\\-for\\-end\\-to\\-end\\-tests\\/py\\-it\\-cloud\\/output\\/1586892456631\\/results[^/\\\\]*'
INFO:apache_beam.io.gcp.gcsio:Starting the size estimation of the input
INFO:apache_beam.io.gcp.gcsio:Finished listing 30 files in 0.07889842987060547
seconds.
Worker logs:
https://console.cloud.google.com/dataflow/jobs/us-central1/2020-04-14_12_27_40-6423619403806197362?project=apache-beam-testing
error: [Errno 111] Connection refused
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/src/sdks/python/scripts/run_integration_test.sh>:
line 278: 20406 Terminated python setup.py nosetests
--test-pipeline-options="$PIPELINE_OPTS" --with-xunitmp
--xunitmp-file=$XUNIT_FILE --ignore-files '.*py3\d?\.py$' $TEST_OPTS
:sdks:python:test-suites:dataflow:py35:integrationTest (Thread[Execution worker
for ':',5,main]) completed. Took 13 mins 15.713 secs.
Daemon vm is shutting down... The daemon has exited normally or was terminated
in response to a user interrupt.
----- End of the daemon log -----
FAILURE: Build failed with an exception.
* What went wrong:
Gradle build daemon disappeared unexpectedly (it may have been killed or may
have crashed)
* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to
get more log output. Run with --scan to get full insights.
* Get more help at https://help.gradle.org
2020-04-14 19:40:51,170 3541c250 MainThread beam_integration_benchmark(1/1)
ERROR Error during benchmark beam_integration_benchmark
Traceback (most recent call last):
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",>
line 846, in RunBenchmark
DoRunPhase(spec, collector, detailed_timer)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",>
line 689, in DoRunPhase
samples = spec.BenchmarkRun(spec)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",>
line 161, in Run
job_type=job_type)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",>
line 96, in SubmitJob
assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2020-04-14 19:40:51,172 3541c250 MainThread beam_integration_benchmark(1/1)
INFO Cleaning up benchmark beam_integration_benchmark
2020-04-14 19:40:51,174 3541c250 MainThread beam_integration_benchmark(1/1)
ERROR Exception running benchmark
Traceback (most recent call last):
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",>
line 995, in RunBenchmarkTask
RunBenchmark(spec, collector)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",>
line 846, in RunBenchmark
DoRunPhase(spec, collector, detailed_timer)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",>
line 689, in DoRunPhase
samples = spec.BenchmarkRun(spec)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",>
line 161, in Run
job_type=job_type)
File
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",>
line 96, in SubmitJob
assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2020-04-14 19:40:51,175 3541c250 MainThread beam_integration_benchmark(1/1)
ERROR Benchmark 1/1 beam_integration_benchmark (UID:
beam_integration_benchmark0) failed. Execution will continue.
2020-04-14 19:40:51,175 3541c250 MainThread beam_integration_benchmark(1/1)
INFO Benchmark run statuses:
---------------------------------------------------------------------------------
Name UID Status Failed
Substatus
---------------------------------------------------------------------------------
beam_integration_benchmark beam_integration_benchmark0 FAILED
---------------------------------------------------------------------------------
Success rate: 0.00% (0/1)
2020-04-14 19:40:51,175 3541c250 MainThread beam_integration_benchmark(1/1)
INFO Complete logs can be found at:
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/runs/3541c250/pkb.log>
2020-04-14 19:40:51,176 3541c250 MainThread beam_integration_benchmark(1/1)
INFO Completion statuses can be found at:
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/runs/3541c250/completion_statuses.json>
Build step 'Execute shell' marked build as failure
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]