See 
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/1326/display/redirect?page=changes>

Changes:

[xhan] Documentation bug fix for FlatMapElements#via() SimpleFunction in the

[github] [BEAM-7923] Screendiff Integration Tests (#11338)

[github] [BEAM-9642] Add SDF execution units. (#11327)


------------------------------------------
[...truncated 155.05 KB...]
                  "pipeline_proto_coder_id": "ref_Coder_FastPrimitivesCoder_5"
                },
                {
                  "@type": "kind:global_window"
                }
              ],
              "is_wrapper": true
            },
            "output_name": "None",
            "user_name": "write/Write/WriteImpl/FinalizeWrite.out"
          }
        ],
        "parallel_input": {
          "@type": "OutputReference",
          "output_name": "out",
          "step_name": "s7"
        },
        "serialized_fn": 
"eNrNWHd8HMUV3ruTbXltg0twcEi5GAwr4O4sGzvGIRA426AcPisngTfFWfZ2527W2vZmZiWLeDG2I8kxSQhphIT0QnoP6b0XSA+kJ6SH9Eqq82buTvIpsuL/4Cdp56a8t++97733zen6nOHYse1QYtWJHRQFs0PeiFjAi07EiF62fd+u+2Q3s+OYsG3RjlAHre8gZFLIGmZO0zSrEUKuS4kX4W/d5kS3Gl5o+961xBpjniA69JhLpIgYj4lFvVBwWOC4nu8XLfnULYcRWxArtAPiiiT2UWShEhnYNYwyVyiRRWrFC+NEKE0ces2luBIlYmZpsblQLjGv6YWgT8ESM4/zLl/xnLKh6BJ01hYR47B0CpbV4JRqdQ+cOgXLa7DCXIRyMYscwjmspD3qVVYjwcCwJo+JA6uqSX0PPMiYHQRB9gkv0q1hHIe8cESH0zByq1N4sGGuQCWc2sxVvloy5LaA082zcL2wzmhvJUHf+k1uIWrgEk4stczlGqwxV3eraPpRvaPnIWqzcO6ckmeYi3Ez8FBGhgAeauo4l8aWYt/GaD1MuWxRYruEwcOr5nKcNjy/hYvFk0bD2weP6HI3itHVkBdHbT8hFoZr1ENhfUjYwnOulouDnTXIYxQemcJaw1yAipUEnGniRw3OUra0lCjj1nWlhx/ZrlrX4WyVfFwwOGcCjBr0JXVzjTRbJmqIp5jt+V7YtEIyhiNmxLlH1PucSHp1Xpf1ao23B304umxcEF6OlLXno7WFFIp1FQYnCmKGmYDOtgwsmT0yrxIRwfqZQMW2oBgFIgPVT6XDG1LYaNAzzV14pMm3lkqCBHGBY9LZTVJA3ApodUFErYFwwUvxeMETBcePErfUyu1S/6Ytm7dcuOGCTZs3b+wvoSGJjwVxAV1Hpf8ydjNgw6aKltS7kzIUhGFFFhPh+fqlrJkEJBSDvu0QGvnK3c1o66OofG6pZNR4IUlh6x54tGFulpUwLij6zhFKS5Xg+oKq7dLumedAEPulHe3SVytwUZcdsUJYt64KY88Z8Yk7hOoGpDYdHpPCxYYKqmsLGy6ZS3D6+DY8osNj0cpLU7jMUIk76pEx2ZTKc3WWRhI6MlN12GbQs81ePC9xls0Ftk/AjhpcblQyFQ3/cpVV5WVHNG2/ph3JaIez2hBcUZ2AgT4lhU4x2RnhcRNQMRmulGgUkNJeEo54Ie+MBe7bo6Q0FrERjn6QknTDGiRMVWrokGEJtbU7Ym45SkIxMGwNjm/cVOLMKXF3RCaBjHfpuCCUWkEoxuNwpTLlIt8O6q59MezceTRT1qBqnibzkEWBxVClLPRpa3ep9qVC1C5ZGJyEx/cJqNVgqCtiTSIsWwjMiWH1mnri+QJdgqtUmHFb7sLVk7C7BmaXqBfEERNWELmJbOFPMFfO6r0tF+CJE/CkGjxZqbdQ1hGWBXsm4Sk1sOjO6lwAOgQncA1F+BCpLCKVqyyqlMuHBPawjLY3K/FyFV6HM1qK06zG+7X9uJXT3KwmerTDmkJTLJB7eMjNaSNZjVXk1O3R1gyJhZq7QJveF4s6OzOTnJo0ctpqHA6hDg3TwzaqlawKjksaNpYm1GUumdfhymA58n2ici8fNfIcKyO/zs2PeYLmA6TavKA27oQkT3wiqzJvO5JyiJu3ed5GgbDpE4HSErxifofHuMiLsahznudJ6MgUIkzKoMa16/ja89WzCI4At1VVvscFEMUBMitEFPkcGootPe57DoGmaq2IL1DzFPy0PYjF+HTJgae2fRLCXtVxVHffzljEYISeI8A3s0o3BCoUnTQL1dslqUA0BTG9UnGUNXeaws77MuXlWvb0zMLMqszKTG8ml8llgfVhovIaYHOlUK2mkFQFjNZgbAL21WCcXpPCtcZxXf6p83T53V7oRmPEVea3e/1+bCRpCtcZ5irUMaYuPa7VoqOWxgPzaNxhczHIvMAT3ug0fVyPKg+mcMg8VXmL96m67Yy0tR2eR9ug6oxtLU9DLRMpTNZ5XSmS0eLCDuK2oqn56Kxztq3rCOp6egpH66oVjKkwtNXcMI+ay/F+YfutqLU1PQM1PTOFZ6nkaYUJbjzRfbL1oi41OjwbNdyUwnNUblhIffDcLnnJU7w47e2MLzo8D0Wfn8IL1MsDz2ERh5uP9t59q3HPvceOHeB1+YOkuGLGyQDxlFcCJIcXnshMSWyK17iurlS2b/2PtA63JBPwohq8eBJuTeElmIQvrcHLVLG0qlZeACTVwsvNM2TrxJdsnaHMrZ083zraD684gYWvpLckdXoelTn5qhRebdADVCbTa1J4LT1MZUbclsLr0Ek6RSWmr0/hDXV6A5WovDGFN9Eb6U0yAlRqmoI3z03f/SdF32+hSMtvNeglVHLt21J4u0HL7R7cYstMZajcs192wXcgSb6zjyIvvuv+4sV3d/Pi7TvXZmlV9o731OC92Dtul73jfQjb+2vwAfr/MfogPR6ND6Xw4Q4aH0nho4jGbdMofCyFj3dQ+EQKn5yFwqfmRmHDSaHwaYnCZzoofDaFz81GYY47y+cRji8oOL54f8FxRzccd+I1hVbpLoo3jy8hKF+uwVcQlDur9IFJ61+VtE4fMFT+NQFfNyihDdqklHp0L1W0+w0a0JAitX6T3jEftd41N7XeLcvjWzX4NiJxlyyP72B5fLcG35uA79fgB5Jaf3g8tf5IlcI9Kfy4Uwo/SeGn7cb0sxR+PtOYfpHCLzsl8asU7u2UxAka36+p7K6/qcFvJ+F3Kfwe7fhDDf54EmX6p64y/XMKf+nY9tcU/tZVpvel8PeOTf9I4Z+zyvRfZN5/i7ToC+3V4d+o4D8pHDPUjUowr9nEqgmZlsFInUBH+5S+rXVLHG5PWQZl+g6yLA7qxnw8ywd1/BLLWA73KpnWf1E8brXvmawHl4+oKmvFEw1YMI8Bc3Ax19nC1usXydfLb7J4V5EXJMvHezd+geasd7bK2fS8LWG2+nbFFrd06TikbAk+6c0Vjbcgx2pIgsRXJ+V3BMKWKq+SumDL8FPxvxRVZNo=",
        "user_name": "write/Write/WriteImpl/FinalizeWrite/FinalizeWrite"
      }
    }
  ],
  "type": "JOB_TYPE_BATCH"
}
INFO:apache_beam.runners.dataflow.internal.apiclient:Create job: <Job
 createTime: '2020-04-14T19:27:41.995875Z'
 currentStateTime: '1970-01-01T00:00:00Z'
 id: '2020-04-14_12_27_40-6423619403806197362'
 location: 'us-central1'
 name: 'beamapp-jenkins-0414192738-103970'
 projectId: 'apache-beam-testing'
 stageStates: []
 startTime: '2020-04-14T19:27:41.995875Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
INFO:apache_beam.runners.dataflow.internal.apiclient:Created job with id: 
[2020-04-14_12_27_40-6423619403806197362]
INFO:apache_beam.runners.dataflow.internal.apiclient:To access the Dataflow 
monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobs/us-central1/2020-04-14_12_27_40-6423619403806197362?project=apache-beam-testing
INFO:apache_beam.runners.dataflow.dataflow_runner:Job 
2020-04-14_12_27_40-6423619403806197362 is in state JOB_STATE_RUNNING
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:44.233Z: 
JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service 
Account.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:44.742Z: 
JOB_MESSAGE_BASIC: Worker configuration: n1-standard-1 in us-central1-f.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.408Z: 
JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.435Z: 
JOB_MESSAGE_DEBUG: Combiner lifting skipped for step 
write/Write/WriteImpl/GroupByKey: GroupByKey not followed by a combiner.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.464Z: 
JOB_MESSAGE_DEBUG: Combiner lifting skipped for step group: GroupByKey not 
followed by a combiner.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.505Z: 
JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into optimizable parts.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.539Z: 
JOB_MESSAGE_DETAILED: Lifting ValueCombiningMappingFns into 
MergeBucketsMappingFns
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.626Z: 
JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.683Z: 
JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.704Z: 
JOB_MESSAGE_DETAILED: Fusing consumer split into read/Read
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.728Z: 
JOB_MESSAGE_DETAILED: Fusing consumer pair_with_one into split
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.756Z: 
JOB_MESSAGE_DETAILED: Fusing consumer group/Reify into pair_with_one
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.781Z: 
JOB_MESSAGE_DETAILED: Fusing consumer group/Write into group/Reify
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.808Z: 
JOB_MESSAGE_DETAILED: Fusing consumer group/GroupByWindow into group/Read
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.831Z: 
JOB_MESSAGE_DETAILED: Fusing consumer count into group/GroupByWindow
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.863Z: 
JOB_MESSAGE_DETAILED: Fusing consumer format into count
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.887Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
write/Write/WriteImpl/WriteBundles/WriteBundles into format
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.915Z: 
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/Pair into 
write/Write/WriteImpl/WriteBundles/WriteBundles
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.944Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
write/Write/WriteImpl/WindowInto(WindowIntoFn) into write/Write/WriteImpl/Pair
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:45.974Z: 
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/GroupByKey/Reify 
into write/Write/WriteImpl/WindowInto(WindowIntoFn)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.005Z: 
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/GroupByKey/Write 
into write/Write/WriteImpl/GroupByKey/Reify
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.032Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
write/Write/WriteImpl/GroupByKey/GroupByWindow into 
write/Write/WriteImpl/GroupByKey/Read
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.061Z: 
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/Extract into 
write/Write/WriteImpl/GroupByKey/GroupByWindow
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.085Z: 
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/InitializeWrite 
into write/Write/WriteImpl/DoOnce/Read
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.108Z: 
JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.135Z: 
JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.166Z: 
JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.192Z: 
JOB_MESSAGE_DEBUG: Assigning stage ids.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.388Z: 
JOB_MESSAGE_DEBUG: Executing wait step start26
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.450Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.471Z: 
JOB_MESSAGE_BASIC: Executing operation write/Write/WriteImpl/GroupByKey/Create
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.491Z: 
JOB_MESSAGE_DEBUG: Starting worker pool setup.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.496Z: 
JOB_MESSAGE_BASIC: Executing operation group/Create
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.511Z: 
JOB_MESSAGE_BASIC: Starting 10 workers in us-central1-f...
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.561Z: 
JOB_MESSAGE_BASIC: Finished operation group/Create
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.561Z: 
JOB_MESSAGE_BASIC: Finished operation write/Write/WriteImpl/GroupByKey/Create
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.619Z: 
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/GroupByKey/Session" 
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.641Z: 
JOB_MESSAGE_DEBUG: Value "group/Session" materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:27:46.695Z: 
JOB_MESSAGE_BASIC: Executing operation 
read/Read+split+pair_with_one+group/Reify+group/Write
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:28:12.207Z: 
JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 3 based on 
the rate of progress in the currently running stage(s).
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:28:12.231Z: 
JOB_MESSAGE_DETAILED: Resized worker pool to 3, though goal was 10.  This could 
be a quota issue.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:28:17.677Z: 
JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 9 based on 
the rate of progress in the currently running stage(s).
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:28:17.709Z: 
JOB_MESSAGE_DETAILED: Resized worker pool to 9, though goal was 10.  This could 
be a quota issue.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:28:23.150Z: 
JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 10 based on 
the rate of progress in the currently running stage(s).
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:28:23.792Z: 
JOB_MESSAGE_WARNING: Your project already contains 100 Dataflow-created metric 
descriptors and Stackdriver will not create new Dataflow custom metrics for 
this job. Each unique user-defined metric name (independent of the DoFn in 
which it is defined) produces a new metric descriptor. To delete old / unused 
metric descriptors see 
https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list
 and 
https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:29:35.460Z: 
JOB_MESSAGE_DETAILED: Workers have started successfully.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:29:35.494Z: 
JOB_MESSAGE_DETAILED: Workers have started successfully.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.266Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.411Z: 
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/DoOnce/Read.out" materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.446Z: 
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/InitializeWrite.out" 
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.525Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/WriteBundles/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.556Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.577Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/WriteBundles/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.587Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.602Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.631Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.709Z: 
JOB_MESSAGE_DEBUG: Value 
"write/Write/WriteImpl/WriteBundles/_UnpickledSideInput(InitializeWrite.out.0).output"
 materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.739Z: 
JOB_MESSAGE_DEBUG: Value 
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0).output"
 materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:32:44.775Z: 
JOB_MESSAGE_DEBUG: Value 
"write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(InitializeWrite.out.0).output"
 materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:33:46.366Z: 
JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service 
Account.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:34:55.189Z: 
JOB_MESSAGE_BASIC: Finished operation 
read/Read+split+pair_with_one+group/Reify+group/Write
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:34:55.249Z: 
JOB_MESSAGE_BASIC: Executing operation group/Close
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:34:55.291Z: 
JOB_MESSAGE_BASIC: Finished operation group/Close
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:34:55.362Z: 
JOB_MESSAGE_BASIC: Executing operation 
group/Read+group/GroupByWindow+count+format+write/Write/WriteImpl/WriteBundles/WriteBundles+write/Write/WriteImpl/Pair+write/Write/WriteImpl/WindowInto(WindowIntoFn)+write/Write/WriteImpl/GroupByKey/Reify+write/Write/WriteImpl/GroupByKey/Write
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:49.249Z: 
JOB_MESSAGE_BASIC: Finished operation 
group/Read+group/GroupByWindow+count+format+write/Write/WriteImpl/WriteBundles/WriteBundles+write/Write/WriteImpl/Pair+write/Write/WriteImpl/WindowInto(WindowIntoFn)+write/Write/WriteImpl/GroupByKey/Reify+write/Write/WriteImpl/GroupByKey/Write
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:49.320Z: 
JOB_MESSAGE_BASIC: Executing operation write/Write/WriteImpl/GroupByKey/Close
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:49.364Z: 
JOB_MESSAGE_BASIC: Finished operation write/Write/WriteImpl/GroupByKey/Close
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:49.420Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/GroupByKey/Read+write/Write/WriteImpl/GroupByKey/GroupByWindow+write/Write/WriteImpl/Extract
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:51.723Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/GroupByKey/Read+write/Write/WriteImpl/GroupByKey/GroupByWindow+write/Write/WriteImpl/Extract
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:51.895Z: 
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/Extract.out" materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:51.960Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:51.983Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(Extract.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:52Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:52.026Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(Extract.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:52.058Z: 
JOB_MESSAGE_DEBUG: Value 
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0).output" 
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:52.090Z: 
JOB_MESSAGE_DEBUG: Value 
"write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(Extract.out.0).output" 
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:52.155Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/PreFinalize/PreFinalize
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:54.987Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/PreFinalize/PreFinalize
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:55.043Z: 
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/PreFinalize.out" materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:55.100Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(PreFinalize.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:55.145Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(PreFinalize.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:55.205Z: 
JOB_MESSAGE_DEBUG: Value 
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(PreFinalize.out.0).output"
 materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:55.267Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/FinalizeWrite/FinalizeWrite
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:57.745Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/FinalizeWrite/FinalizeWrite
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:57.808Z: 
JOB_MESSAGE_DEBUG: Executing success step success24
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:57.918Z: 
JOB_MESSAGE_DETAILED: Cleaning up.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:58.009Z: 
JOB_MESSAGE_DEBUG: Starting worker pool teardown.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:36:58.029Z: 
JOB_MESSAGE_BASIC: Stopping worker pool...
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:38:51.873Z: 
JOB_MESSAGE_DETAILED: Autoscaling: Resized worker pool from 10 to 0.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:38:51.915Z: 
JOB_MESSAGE_BASIC: Worker pool stopped.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:38:51.955Z: 
JOB_MESSAGE_DEBUG: Tearing down pending resources...
INFO:apache_beam.runners.dataflow.dataflow_runner:Job 
2020-04-14_12_27_40-6423619403806197362 is in state JOB_STATE_DONE
DEBUG:apache_beam.io.filesystem:Listing files in 
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results'
DEBUG:apache_beam.io.filesystem:translate_pattern: 
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results*-of-*'
 -> 
'gs\\:\\/\\/temp\\-storage\\-for\\-end\\-to\\-end\\-tests\\/py\\-it\\-cloud\\/output\\/1586892456631\\/results[^/\\\\]*\\-of\\-[^/\\\\]*'
INFO:apache_beam.io.gcp.gcsio:Starting the size estimation of the input
INFO:apache_beam.io.gcp.gcsio:Finished listing 30 files in 0.06600570678710938 
seconds.
INFO:apache_beam.testing.pipeline_verifiers:Find 30 files in 
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results*-of-*:
 
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00023-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00012-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00028-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00029-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00003-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00002-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00025-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00015-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00026-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00006-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00008-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00010-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00016-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00019-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00013-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00011-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00018-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00021-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00000-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00005-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00024-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00017-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00027-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00009-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00022-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00020-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00014-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00001-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00007-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00004-of-00030
INFO:apache_beam.testing.pipeline_verifiers:Read from given path 
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results*-of-*,
 26186927 lines, checksum: ea0ca2e5ee4ea5f218790f28d0b9fe7d09d8d710.
INFO:root:average word length: 19
DEBUG:apache_beam.io.filesystem:Listing files in 
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results'
DEBUG:apache_beam.io.filesystem:translate_pattern: 
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results*'
 -> 
'gs\\:\\/\\/temp\\-storage\\-for\\-end\\-to\\-end\\-tests\\/py\\-it\\-cloud\\/output\\/1586892456631\\/results[^/\\\\]*'
INFO:apache_beam.io.gcp.gcsio:Starting the size estimation of the input
INFO:apache_beam.io.gcp.gcsio:Finished listing 30 files in 0.07889842987060547 
seconds.
error: [Errno 111] Connection refused
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/src/sdks/python/scripts/run_integration_test.sh>:
 line 278: 20406 Terminated              python setup.py nosetests 
--test-pipeline-options="$PIPELINE_OPTS" --with-xunitmp 
--xunitmp-file=$XUNIT_FILE --ignore-files '.*py3\d?\.py$' $TEST_OPTS
The message received from the daemon indicates that the daemon has disappeared.
Build request sent: Build{id=784b86ac-0811-4204-b98e-d368e6c6a074, 
currentDir=<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/src}>
Attempting to read last messages from the daemon log...
Daemon pid: 24086
  log file: /home/jenkins/.gradle/daemon/5.2.1/daemon-24086.out.log
----- Last  20 lines from daemon log file - daemon-24086.out.log -----
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00017-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00027-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00009-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00022-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00020-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00014-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00001-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00007-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results-00004-of-00030
INFO:apache_beam.testing.pipeline_verifiers:Read from given path 
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results*-of-*,
 26186927 lines, checksum: ea0ca2e5ee4ea5f218790f28d0b9fe7d09d8d710.
INFO:root:average word length: 19
DEBUG:apache_beam.io.filesystem:Listing files in 
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results'
DEBUG:apache_beam.io.filesystem:translate_pattern: 
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586892456631/results*'
 -> 
'gs\\:\\/\\/temp\\-storage\\-for\\-end\\-to\\-end\\-tests\\/py\\-it\\-cloud\\/output\\/1586892456631\\/results[^/\\\\]*'
INFO:apache_beam.io.gcp.gcsio:Starting the size estimation of the input
INFO:apache_beam.io.gcp.gcsio:Finished listing 30 files in 0.07889842987060547 
seconds.
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2020-04-14_12_27_40-6423619403806197362?project=apache-beam-testing
error: [Errno 111] Connection refused
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/src/sdks/python/scripts/run_integration_test.sh>:
 line 278: 20406 Terminated              python setup.py nosetests 
--test-pipeline-options="$PIPELINE_OPTS" --with-xunitmp 
--xunitmp-file=$XUNIT_FILE --ignore-files '.*py3\d?\.py$' $TEST_OPTS
:sdks:python:test-suites:dataflow:py35:integrationTest (Thread[Execution worker 
for ':',5,main]) completed. Took 13 mins 15.713 secs.
Daemon vm is shutting down... The daemon has exited normally or was terminated 
in response to a user interrupt.
----- End of the daemon log -----


FAILURE: Build failed with an exception.

* What went wrong:
Gradle build daemon disappeared unexpectedly (it may have been killed or may 
have crashed)

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to 
get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

2020-04-14 19:40:51,170 3541c250 MainThread beam_integration_benchmark(1/1) 
ERROR    Error during benchmark beam_integration_benchmark
Traceback (most recent call last):
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py";,>
 line 846, in RunBenchmark
    DoRunPhase(spec, collector, detailed_timer)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py";,>
 line 689, in DoRunPhase
    samples = spec.BenchmarkRun(spec)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py";,>
 line 161, in Run
    job_type=job_type)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py";,>
 line 96, in SubmitJob
    assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2020-04-14 19:40:51,172 3541c250 MainThread beam_integration_benchmark(1/1) 
INFO     Cleaning up benchmark beam_integration_benchmark
2020-04-14 19:40:51,174 3541c250 MainThread beam_integration_benchmark(1/1) 
ERROR    Exception running benchmark
Traceback (most recent call last):
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py";,>
 line 995, in RunBenchmarkTask
    RunBenchmark(spec, collector)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py";,>
 line 846, in RunBenchmark
    DoRunPhase(spec, collector, detailed_timer)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py";,>
 line 689, in DoRunPhase
    samples = spec.BenchmarkRun(spec)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py";,>
 line 161, in Run
    job_type=job_type)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py";,>
 line 96, in SubmitJob
    assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2020-04-14 19:40:51,175 3541c250 MainThread beam_integration_benchmark(1/1) 
ERROR    Benchmark 1/1 beam_integration_benchmark (UID: 
beam_integration_benchmark0) failed. Execution will continue.
2020-04-14 19:40:51,175 3541c250 MainThread beam_integration_benchmark(1/1) 
INFO     Benchmark run statuses:
---------------------------------------------------------------------------------
Name                        UID                          Status  Failed 
Substatus
---------------------------------------------------------------------------------
beam_integration_benchmark  beam_integration_benchmark0  FAILED                 
 
---------------------------------------------------------------------------------
Success rate: 0.00% (0/1)
2020-04-14 19:40:51,175 3541c250 MainThread beam_integration_benchmark(1/1) 
INFO     Complete logs can be found at: 
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/runs/3541c250/pkb.log>
2020-04-14 19:40:51,176 3541c250 MainThread beam_integration_benchmark(1/1) 
INFO     Completion statuses can be found at: 
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py35/ws/runs/3541c250/completion_statuses.json>
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to