See 
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/1303/display/redirect?page=changes>

Changes:

[xhan] Documentation bug fix for FlatMapElements#via() SimpleFunction in the

[github] [BEAM-7923] Screendiff Integration Tests (#11338)


------------------------------------------
[...truncated 155.34 KB...]
                  ],
                  "is_pair_like": true,
                  "pipeline_proto_coder_id": "ref_Coder_FastPrimitivesCoder_5"
                },
                {
                  "@type": "kind:global_window"
                }
              ],
              "is_wrapper": true
            },
            "output_name": "None",
            "user_name": "write/Write/WriteImpl/FinalizeWrite.out"
          }
        ],
        "parallel_input": {
          "@type": "OutputReference",
          "output_name": "out",
          "step_name": "s7"
        },
        "serialized_fn": 
"eNrNWHd8HEcVvjvJbe04LsRgUriYGE4hd2c7WLFNSEjOlmMOn5WVEi/FbPZ2527W2vZmZiUraJM4jgomQIAQmukl9A6h914SOqZD6IReQ0mcN7N3kk/Iwv+Zn6Tdvbfzvp33vvfet6cbugq2FVk2JWadWH5JMCvgjZD5vGSHjGgVy/Osukf2MiuKCNse9gUaZHoOQjaBXMHoymQyZiOArg4QN8TfusWJZjbcwPLca4k5wlxBNOg2limXGGFZk0fEhgU1YxHaIhbahHNYSLuNpXKNGI2ISd1AcFhkO67nlUx51EybEUsQM7B84og48hB2sXLZtWcQfa5QLkuUxQ2iWCgkDpp6dBiLGdNSY6E0MbfpBrBsEk4z8vi5Ix+4Tu2h5BBMiCVCxmH5JJyuw4pabR+snIRVOqyO6/vgQYXZSRDkgHBDzRzE84AbDGlwBmZuTQIPLhgr8EEN1yNmZAlqRow03APwkA6EMELvgJeGLS/GdSwcdh3CtAFhCde+Whr72zZYi8APTeDMgqEhcOoh9w5ndaTOCy1H2TU4W5HHBYNzxuFhOuSNBW1PONfYg9dNvq1cFsSPihzDtpqkiIVRJIFTFGF6IlzwcjRadEXR9sLYKafZLW/cvKV3y9YNW3s3brloU5kRHntIybq4PhO2ZM/kcUOG/XAqt39eAusL9Cyap+cauCoDj8D1Mpog9k1OLeZweGQ1o7Zphxg1dCZc2XjrpA2Gl4/i/iqhSk8P4p+fwKPqxkoZtARLdyBL3RJwgXEe2ovrC61bsd+zYbNTDBtomnm8tEHRWNMJ0fTCehunpG4Wz5/Ts6yit0Mf2eYcqU0Z2mB0y6KLRQgbjSV46bt+i7xNKn5ZRuXIs7BGLzTWytaQvRggkcxyPTdomgEZwTMW9KOnVC+ZlFgyQZtrcb2zKANBGHZkKRaup13GmrFPAtHvWTahoadS1YupuojK45ZqVp23kgS27YPHFIxe2aejguLWOdadqdprQ1H1dnnvzHGXH3nlvlbrKwtc3LGPSNWZZl4VRK495BFnAOF2STQNHpvAJQWVE8cSFlw6l+P08u24RIPH4S4vS+DygrFcptKWg0Q2lgwWKsaZaJTe22Y2vQ33xORg2za8EbarqhhxAyccMX1MrswpTrUdJxqNEkfBcE3Rb3nmf3lr0Ke4GHbJiATbOdcMa8SBLZtcgysK9GxjMa6XFSvHGOwah8frUC1Us9UM/nVVV1e0qcxYZip7KDeQgSfUxmF3j3JpxwJo2mMwtJRp6JPyfhIMuQFvn4vcs4ZJeSRkQxzjImUZltlPmCrewCaDsp/NvSFzKmEciF2DZv/ohb1lzuwyd4Zkp0vqy8clpZzyUYpGoV9t5WLP8uuOdQlcuftwtpIB3ThDdjwLfZMhpKzs6d0OqIGs8tMadTA4AVf1CLhah70d6WoSYSKdWJ6Gekw9dj2BIcETVY7xtrwLT5qAJ+vwlA5X149CJkw/dGKpFPuMVbNGfBoCPHUcTB2uUfAm+trCNMGagLoONr2yNhd7NsEP4FDkDmnKIU1d1UXVSuUakRHZ/bmxjINUHcomubEcXzeW3d/l5ET3oYykTywYwztO11CO9Y7lnO61GbQtdBakdrEotbWuuuRVo2sNHm/MOrgSSKFWzanQHdKwcLpCQ5aJcR1a+iuh5xFVVvmwkefYgvn1Tn7EFTTvo6bnBbXwTkDyxCOy/fNpwxAnb/G8hQ5B0yMCvSU1pXyfy7jIi5GwvZ7nSWDLAiFM+iDiuvV83QXqWIKmAJq2r+dyAa4aaZJzEYYeh/1Kcl3uuTaBIaVByB54qnN3+JEYne5t8NVtjwQQqEGoNG8HYyGDkJ4jIDJyChtApaJdREw9XY5A4JMgaL8ayubcRRjvvjdbWZHJrch2Z5dnl2UXZ3PZXA6Ge7AMR3Q4QG0a12oJjNYEXKvD08ZhTIeEOglcVzjB2Lie9h0nUjfMI1J7lS9xVGAtqTqIs+zGBA4VjNUSXb13OWaq6CniTfMg9llc9DPXd4U7PK1+4wg5kcCkcbrKA77S1S17qIU2NQ9avxrOLZSnI8rhBJ5R53UFJPPIheVHLaCb51Pj9toW1jMR61kJPLuuRkArhSnMLfPA7FSzNs1aC+k5iPTcBJ6nyipNE9x6ormdPqgDRoPnI8JtCbxAVY2Jugov7PCXUslL09HOxKLBi9D1xQm8RD3cd20WcjhyePHRI4W77zl27Hpelz9xPR6Hl+rwsgl4eQKvwDp6pQ6viut0B+2jBSr5fnUCrynQm6gk6rUJvI5OUZnt2xN4PQLQm6nM1xsSeGOd3kJlxG9K4M30VnqbQp+Et8ytzBtPSpnfSlFx31agl1Ipo29P4B0F+r81851q+ztbsy+VqGx1oNI9JufTu1CJ3t1D8fieU6VH7+3Uozt2r8tRXXb1+3R4P3b1HbKrP4BsfFCHD3Ww8eEEPtJm46MJfAzZuH2ahY8n8Ik2C59M4FPHsfDpuVnYdFIsfEay8Nk2C59L4PMnw8IX5mBh9ovCF5GFLykuvnyquLizk4u78N2A6nSAotx/BRn5qg5fQ0buqtFToaVfl1pK/2/08xsCvlmgLt1Ph6hHfRpQpXXfokAZRT37Nr1zPj07OreefUdW/nd1+B7m+ais/O9j5f9Ahx+Ow490+LHUs5+cSM/u7tCzn6oe+VkCP2/3yC8S+GVrYv0qgV/PTKzfJHBPu1d+m8DvpntlHH6vwx8m4I8J/An38Wcd/tLRgX9N4G9t9L8n8I+ODrw3gX+2Uf+VwL+P68D/kHn/mZFOfIxKg/vQ+f4EjhVUZaYxNwKWyWKYJy8eXGNZ9Og5yHJ4Um86grnNJrZNwLrmwWqt0ranb2+DrY+sO0VbINFWpl9lYj/2LFmM8hWWsIV4q5pV3yRRx+XLg4m3CX7542zR7CfOlq7tMVNQGlucPmgJnhKm4ZEeqWZQ2FfNEna/jt8rGVuaPlX9X8blZuulky1D81RcF+w0vCg9AE1gYIM=",
        "user_name": "write/Write/WriteImpl/FinalizeWrite/FinalizeWrite"
      }
    }
  ],
  "type": "JOB_TYPE_BATCH"
}
INFO:apache_beam.runners.dataflow.internal.apiclient:Create job: <Job
 createTime: '2020-04-14T19:03:06.696053Z'
 currentStateTime: '1970-01-01T00:00:00Z'
 id: '2020-04-14_12_02_57-7999963810299807213'
 location: 'us-central1'
 name: 'beamapp-jenkins-0414190249-754840'
 projectId: 'apache-beam-testing'
 stageStates: []
 startTime: '2020-04-14T19:03:06.696053Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
INFO:apache_beam.runners.dataflow.internal.apiclient:Created job with id: 
[2020-04-14_12_02_57-7999963810299807213]
INFO:apache_beam.runners.dataflow.internal.apiclient:To access the Dataflow 
monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobs/us-central1/2020-04-14_12_02_57-7999963810299807213?project=apache-beam-testing
INFO:apache_beam.runners.dataflow.dataflow_runner:Job 
2020-04-14_12_02_57-7999963810299807213 is in state JOB_STATE_RUNNING
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:08.931Z: 
JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service 
Account.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:09.425Z: 
JOB_MESSAGE_BASIC: Worker configuration: n1-standard-1 in us-central1-f.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.080Z: 
JOB_MESSAGE_DETAILED: Expanding CoGroupByKey operations into optimizable parts.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.120Z: 
JOB_MESSAGE_DEBUG: Combiner lifting skipped for step 
write/Write/WriteImpl/GroupByKey: GroupByKey not followed by a combiner.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.153Z: 
JOB_MESSAGE_DEBUG: Combiner lifting skipped for step group: GroupByKey not 
followed by a combiner.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.194Z: 
JOB_MESSAGE_DETAILED: Expanding GroupByKey operations into optimizable parts.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.233Z: 
JOB_MESSAGE_DETAILED: Lifting ValueCombiningMappingFns into 
MergeBucketsMappingFns
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.346Z: 
JOB_MESSAGE_DEBUG: Annotating graph with Autotuner information.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.404Z: 
JOB_MESSAGE_DETAILED: Fusing adjacent ParDo, Read, Write, and Flatten operations
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.447Z: 
JOB_MESSAGE_DETAILED: Fusing consumer split into read/Read
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.469Z: 
JOB_MESSAGE_DETAILED: Fusing consumer pair_with_one into split
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.500Z: 
JOB_MESSAGE_DETAILED: Fusing consumer group/Reify into pair_with_one
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.535Z: 
JOB_MESSAGE_DETAILED: Fusing consumer group/Write into group/Reify
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.568Z: 
JOB_MESSAGE_DETAILED: Fusing consumer group/GroupByWindow into group/Read
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.603Z: 
JOB_MESSAGE_DETAILED: Fusing consumer count into group/GroupByWindow
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.635Z: 
JOB_MESSAGE_DETAILED: Fusing consumer format into count
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.668Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
write/Write/WriteImpl/WriteBundles/WriteBundles into format
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.703Z: 
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/Pair into 
write/Write/WriteImpl/WriteBundles/WriteBundles
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.734Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
write/Write/WriteImpl/WindowInto(WindowIntoFn) into write/Write/WriteImpl/Pair
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.773Z: 
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/GroupByKey/Reify 
into write/Write/WriteImpl/WindowInto(WindowIntoFn)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.795Z: 
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/GroupByKey/Write 
into write/Write/WriteImpl/GroupByKey/Reify
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.834Z: 
JOB_MESSAGE_DETAILED: Fusing consumer 
write/Write/WriteImpl/GroupByKey/GroupByWindow into 
write/Write/WriteImpl/GroupByKey/Read
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.868Z: 
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/Extract into 
write/Write/WriteImpl/GroupByKey/GroupByWindow
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.901Z: 
JOB_MESSAGE_DETAILED: Fusing consumer write/Write/WriteImpl/InitializeWrite 
into write/Write/WriteImpl/DoOnce/Read
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.937Z: 
JOB_MESSAGE_DEBUG: Workflow config is missing a default resource spec.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.963Z: 
JOB_MESSAGE_DEBUG: Adding StepResource setup and teardown to workflow graph.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:10.997Z: 
JOB_MESSAGE_DEBUG: Adding workflow start and stop steps.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:11.031Z: 
JOB_MESSAGE_DEBUG: Assigning stage ids.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:11.222Z: 
JOB_MESSAGE_DEBUG: Executing wait step start26
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:11.318Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:11.356Z: 
JOB_MESSAGE_BASIC: Executing operation write/Write/WriteImpl/GroupByKey/Create
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:11.359Z: 
JOB_MESSAGE_DEBUG: Starting worker pool setup.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:11.390Z: 
JOB_MESSAGE_BASIC: Starting 10 workers in us-central1-f...
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:11.390Z: 
JOB_MESSAGE_BASIC: Executing operation group/Create
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:11.444Z: 
JOB_MESSAGE_BASIC: Finished operation write/Write/WriteImpl/GroupByKey/Create
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:11.445Z: 
JOB_MESSAGE_BASIC: Finished operation group/Create
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:11.501Z: 
JOB_MESSAGE_DEBUG: Value "group/Session" materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:11.539Z: 
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/GroupByKey/Session" 
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:11.576Z: 
JOB_MESSAGE_BASIC: Executing operation 
read/Read+split+pair_with_one+group/Reify+group/Write
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:21.103Z: 
JOB_MESSAGE_WARNING: Your project already contains 100 Dataflow-created metric 
descriptors and Stackdriver will not create new Dataflow custom metrics for 
this job. Each unique user-defined metric name (independent of the DoFn in 
which it is defined) produces a new metric descriptor. To delete old / unused 
metric descriptors see 
https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.list
 and 
https://developers.google.com/apis-explorer/#p/monitoring/v3/monitoring.projects.metricDescriptors.delete
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:35.738Z: 
JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 7 based on 
the rate of progress in the currently running stage(s).
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:35.775Z: 
JOB_MESSAGE_DETAILED: Resized worker pool to 7, though goal was 10.  This could 
be a quota issue.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:03:41.156Z: 
JOB_MESSAGE_DETAILED: Autoscaling: Raised the number of workers to 10 based on 
the rate of progress in the currently running stage(s).
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:04:53.725Z: 
JOB_MESSAGE_DETAILED: Workers have started successfully.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:04:53.755Z: 
JOB_MESSAGE_DETAILED: Workers have started successfully.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:08:06.927Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:08:07.008Z: 
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/DoOnce/Read.out" materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:08:07.054Z: 
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/InitializeWrite.out" 
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:08:07.134Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/WriteBundles/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:08:07.171Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:08:07.202Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/WriteBundles/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:08:07.207Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:08:07.231Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:08:07.268Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(InitializeWrite.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:08:07.299Z: 
JOB_MESSAGE_DEBUG: Value 
"write/Write/WriteImpl/WriteBundles/_UnpickledSideInput(InitializeWrite.out.0).output"
 materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:08:07.342Z: 
JOB_MESSAGE_DEBUG: Value 
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0).output"
 materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:08:07.384Z: 
JOB_MESSAGE_DEBUG: Value 
"write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(InitializeWrite.out.0).output"
 materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:09:11.190Z: 
JOB_MESSAGE_DETAILED: Checking permissions granted to controller Service 
Account.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:10:17.226Z: 
JOB_MESSAGE_BASIC: Finished operation 
read/Read+split+pair_with_one+group/Reify+group/Write
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:10:17.351Z: 
JOB_MESSAGE_BASIC: Executing operation group/Close
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:10:17.420Z: 
JOB_MESSAGE_BASIC: Finished operation group/Close
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:10:17.502Z: 
JOB_MESSAGE_BASIC: Executing operation 
group/Read+group/GroupByWindow+count+format+write/Write/WriteImpl/WriteBundles/WriteBundles+write/Write/WriteImpl/Pair+write/Write/WriteImpl/WindowInto(WindowIntoFn)+write/Write/WriteImpl/GroupByKey/Reify+write/Write/WriteImpl/GroupByKey/Write
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:08.686Z: 
JOB_MESSAGE_BASIC: Finished operation 
group/Read+group/GroupByWindow+count+format+write/Write/WriteImpl/WriteBundles/WriteBundles+write/Write/WriteImpl/Pair+write/Write/WriteImpl/WindowInto(WindowIntoFn)+write/Write/WriteImpl/GroupByKey/Reify+write/Write/WriteImpl/GroupByKey/Write
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:08.776Z: 
JOB_MESSAGE_BASIC: Executing operation write/Write/WriteImpl/GroupByKey/Close
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:08.870Z: 
JOB_MESSAGE_BASIC: Finished operation write/Write/WriteImpl/GroupByKey/Close
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:08.957Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/GroupByKey/Read+write/Write/WriteImpl/GroupByKey/GroupByWindow+write/Write/WriteImpl/Extract
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:11.473Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/GroupByKey/Read+write/Write/WriteImpl/GroupByKey/GroupByWindow+write/Write/WriteImpl/Extract
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:11.811Z: 
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/Extract.out" materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:12.030Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:12.091Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(Extract.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:12.131Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:12.168Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(Extract.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:12.220Z: 
JOB_MESSAGE_DEBUG: Value 
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0).output" 
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:12.258Z: 
JOB_MESSAGE_DEBUG: Value 
"write/Write/WriteImpl/PreFinalize/_UnpickledSideInput(Extract.out.0).output" 
materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:12.344Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/PreFinalize/PreFinalize
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:14.912Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/PreFinalize/PreFinalize
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:15.103Z: 
JOB_MESSAGE_DEBUG: Value "write/Write/WriteImpl/PreFinalize.out" materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:15.220Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(PreFinalize.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:15.288Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(PreFinalize.out.0)
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:15.379Z: 
JOB_MESSAGE_DEBUG: Value 
"write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(PreFinalize.out.0).output"
 materialized.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:15.468Z: 
JOB_MESSAGE_BASIC: Executing operation 
write/Write/WriteImpl/FinalizeWrite/FinalizeWrite
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:17.987Z: 
JOB_MESSAGE_BASIC: Finished operation 
write/Write/WriteImpl/FinalizeWrite/FinalizeWrite
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:18.074Z: 
JOB_MESSAGE_DEBUG: Executing success step success24
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:18.241Z: 
JOB_MESSAGE_DETAILED: Cleaning up.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:18.303Z: 
JOB_MESSAGE_DEBUG: Starting worker pool teardown.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:12:18.338Z: 
JOB_MESSAGE_BASIC: Stopping worker pool...
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:14:34.820Z: 
JOB_MESSAGE_DETAILED: Autoscaling: Resized worker pool from 10 to 0.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:14:34.872Z: 
JOB_MESSAGE_BASIC: Worker pool stopped.
INFO:apache_beam.runners.dataflow.dataflow_runner:2020-04-14T19:14:34.919Z: 
JOB_MESSAGE_DEBUG: Tearing down pending resources...
INFO:apache_beam.runners.dataflow.dataflow_runner:Job 
2020-04-14_12_02_57-7999963810299807213 is in state JOB_STATE_DONE
DEBUG:apache_beam.io.filesystem:Listing files in 
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results'
DEBUG:apache_beam.io.filesystem:translate_pattern: 
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results*-of-*'
 -> 
'gs\\:\\/\\/temp\\-storage\\-for\\-end\\-to\\-end\\-tests\\/py\\-it\\-cloud\\/output\\/1586890961872\\/results[^/\\\\]*\\-of\\-[^/\\\\]*'
INFO:apache_beam.io.gcp.gcsio:Starting the size estimation of the input
INFO:apache_beam.io.gcp.gcsio:Finished listing 30 files in 0.05901288986206055 
seconds.
INFO:apache_beam.testing.pipeline_verifiers:Find 30 files in 
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results*-of-*:
 
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00000-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00001-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00002-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00003-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00004-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00005-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00006-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00007-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00008-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00009-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00010-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00011-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00012-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00013-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00014-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00015-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00016-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00017-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00018-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00019-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00020-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00021-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00022-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00023-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00024-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00025-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00026-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00027-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00028-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00029-of-00030
INFO:apache_beam.testing.pipeline_verifiers:Read from given path 
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results*-of-*,
 26186927 lines, checksum: ea0ca2e5ee4ea5f218790f28d0b9fe7d09d8d710.
INFO:root:average word length: 19
DEBUG:apache_beam.io.filesystem:Listing files in 
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results'
DEBUG:apache_beam.io.filesystem:translate_pattern: 
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results*'
 -> 
'gs\\:\\/\\/temp\\-storage\\-for\\-end\\-to\\-end\\-tests\\/py\\-it\\-cloud\\/output\\/1586890961872\\/results[^/\\\\]*'
INFO:apache_beam.io.gcp.gcsio:Starting the size estimation of the input
INFO:apache_beam.io.gcp.gcsio:Finished listing 30 files in 0.05802750587463379 
seconds.
error: [Errno 111] Connection refused
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/src/sdks/python/scripts/run_integration_test.sh>:
 line 278:  5878 Terminated              python setup.py nosetests 
--test-pipeline-options="$PIPELINE_OPTS" --with-xunitmp 
--xunitmp-file=$XUNIT_FILE --ignore-files '.*py3\d?\.py$' $TEST_OPTS
The message received from the daemon indicates that the daemon has disappeared.
Build request sent: Build{id=5537ee4a-69c5-4afb-9438-0876a19e14f1, 
currentDir=<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/src}>
Attempting to read last messages from the daemon log...
Daemon pid: 12096
  log file: /home/jenkins/.gradle/daemon/5.2.1/daemon-12096.out.log
----- Last  20 lines from daemon log file - daemon-12096.out.log -----
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00021-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00022-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00023-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00024-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00025-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00026-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00027-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00028-of-00030
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results-00029-of-00030
INFO:apache_beam.testing.pipeline_verifiers:Read from given path 
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results*-of-*,
 26186927 lines, checksum: ea0ca2e5ee4ea5f218790f28d0b9fe7d09d8d710.
INFO:root:average word length: 19
DEBUG:apache_beam.io.filesystem:Listing files in 
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results'
DEBUG:apache_beam.io.filesystem:translate_pattern: 
'gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1586890961872/results*'
 -> 
'gs\\:\\/\\/temp\\-storage\\-for\\-end\\-to\\-end\\-tests\\/py\\-it\\-cloud\\/output\\/1586890961872\\/results[^/\\\\]*'
INFO:apache_beam.io.gcp.gcsio:Starting the size estimation of the input
INFO:apache_beam.io.gcp.gcsio:Finished listing 30 files in 0.05802750587463379 
seconds.
Worker logs: 
https://console.cloud.google.com/dataflow/jobs/us-central1/2020-04-14_12_02_57-7999963810299807213?project=apache-beam-testing
error: [Errno 111] Connection refused
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/src/sdks/python/scripts/run_integration_test.sh>:
 line 278:  5878 Terminated              python setup.py nosetests 
--test-pipeline-options="$PIPELINE_OPTS" --with-xunitmp 
--xunitmp-file=$XUNIT_FILE --ignore-files '.*py3\d?\.py$' $TEST_OPTS
:sdks:python:test-suites:dataflow:py36:integrationTest (Thread[Execution worker 
for ':' Thread 15,5,main]) completed. Took 14 mins 6.821 secs.
Daemon vm is shutting down... The daemon has exited normally or was terminated 
in response to a user interrupt.
----- End of the daemon log -----


FAILURE: Build failed with an exception.

* What went wrong:
Gradle build daemon disappeared unexpectedly (it may have been killed or may 
have crashed)

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to 
get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

2020-04-14 19:16:38,784 99d11f71 MainThread beam_integration_benchmark(1/1) 
ERROR    Error during benchmark beam_integration_benchmark
Traceback (most recent call last):
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py";,>
 line 846, in RunBenchmark
    DoRunPhase(spec, collector, detailed_timer)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py";,>
 line 689, in DoRunPhase
    samples = spec.BenchmarkRun(spec)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py";,>
 line 161, in Run
    job_type=job_type)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py";,>
 line 96, in SubmitJob
    assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2020-04-14 19:16:38,785 99d11f71 MainThread beam_integration_benchmark(1/1) 
INFO     Cleaning up benchmark beam_integration_benchmark
2020-04-14 19:16:38,788 99d11f71 MainThread beam_integration_benchmark(1/1) 
ERROR    Exception running benchmark
Traceback (most recent call last):
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py";,>
 line 995, in RunBenchmarkTask
    RunBenchmark(spec, collector)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py";,>
 line 846, in RunBenchmark
    DoRunPhase(spec, collector, detailed_timer)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py";,>
 line 689, in DoRunPhase
    samples = spec.BenchmarkRun(spec)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py";,>
 line 161, in Run
    job_type=job_type)
  File 
"<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py";,>
 line 96, in SubmitJob
    assert retcode == 0, "Integration Test Failed."
AssertionError: Integration Test Failed.
2020-04-14 19:16:38,788 99d11f71 MainThread beam_integration_benchmark(1/1) 
ERROR    Benchmark 1/1 beam_integration_benchmark (UID: 
beam_integration_benchmark0) failed. Execution will continue.
2020-04-14 19:16:38,789 99d11f71 MainThread beam_integration_benchmark(1/1) 
INFO     Benchmark run statuses:
---------------------------------------------------------------------------------
Name                        UID                          Status  Failed 
Substatus
---------------------------------------------------------------------------------
beam_integration_benchmark  beam_integration_benchmark0  FAILED                 
 
---------------------------------------------------------------------------------
Success rate: 0.00% (0/1)
2020-04-14 19:16:38,789 99d11f71 MainThread beam_integration_benchmark(1/1) 
INFO     Complete logs can be found at: 
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/runs/99d11f71/pkb.log>
2020-04-14 19:16:38,789 99d11f71 MainThread beam_integration_benchmark(1/1) 
INFO     Completion statuses can be found at: 
<https://builds.apache.org/job/beam_PerformanceTests_WordCountIT_Py36/ws/runs/99d11f71/completion_statuses.json>
Build step 'Execute shell' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to