See <https://builds.apache.org/job/beam_PostCommit_PythonVerify/134/changes>

Changes:

[robertwb] Remove expensive per-element-step logging context.

[robertwb] Cache dofn.proces method.

[robertwb] Restore (faster) logging context.

------------------------------------------
[...truncated 2482 lines...]
        "serialized_fn": "<string of 1340 bytes>", 
        "user_name": "write/WriteImpl/write_bundles"
      }
    }, 
    {
      "kind": "CollectionToSingleton", 
      "name": "s11", 
      "properties": {
        "output_info": [
          {
            "encoding": {
              "@type": "WindowedValueCoder$<string of 736 bytes>", 
              "component_encodings": [
                {
                  "@type": "WindowedValueCoder$<string of 736 bytes>", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$<string of 172 bytes>", 
                      "component_encodings": [
                        {
                          "@type": "FastPrimitivesCoder$<string of 172 bytes>", 
                          "component_encodings": []
                        }, 
                        {
                          "@type": "FastPrimitivesCoder$<string of 172 bytes>", 
                          "component_encodings": []
                        }
                      ], 
                      "is_pair_like": true
                    }, 
                    {
                      "@type": 
"TimestampCoder$gAJjYXBhY2hlX2JlYW0uY29kZXJzLmNvZGVycwpUaW1lc3RhbXBDb2RlcgpxACmBcQF9cQJiLg==",
 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "SingletonCoder$<string of 344 bytes>", 
                      "component_encodings": []
                    }
                  ], 
                  "is_wrapper": true
                }
              ]
            }, 
            "output_name": "out", 
            "user_name": 
"write/WriteImpl/ViewAsIterable(write|WriteImpl|write_bundles.None)/CreatePCollectionView.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s10"
        }, 
        "user_name": 
"write/WriteImpl/ViewAsIterable(write|WriteImpl|write_bundles.None)/CreatePCollectionView"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s12", 
      "properties": {
        "non_parallel_inputs": {
          "s11": {
            "@type": "OutputReference", 
            "output_name": "out", 
            "step_name": "s11"
          }, 
          "s9": {
            "@type": "OutputReference", 
            "output_name": "out", 
            "step_name": "s9"
          }
        }, 
        "output_info": [
          {
            "encoding": {
              "@type": "WindowedValueCoder$<string of 736 bytes>", 
              "component_encodings": [
                {
                  "@type": "FastPrimitivesCoder$<string of 172 bytes>", 
                  "component_encodings": [
                    {
                      "@type": "FastPrimitivesCoder$<string of 172 bytes>", 
                      "component_encodings": []
                    }, 
                    {
                      "@type": "FastPrimitivesCoder$<string of 172 bytes>", 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": 
"TimestampCoder$gAJjYXBhY2hlX2JlYW0uY29kZXJzLmNvZGVycwpUaW1lc3RhbXBDb2RlcgpxACmBcQF9cQJiLg==",
 
                  "component_encodings": []
                }, 
                {
                  "@type": "SingletonCoder$<string of 344 bytes>", 
                  "component_encodings": []
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "write/WriteImpl/finalize_write.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s7"
        }, 
        "serialized_fn": "<string of 1496 bytes>", 
        "user_name": "write/WriteImpl/finalize_write"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
INFO:root:Create job: <Job
 id: u'2016-07-21_17_39_34-11919126913628582743'
 projectId: u'apache-beam-testing'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
INFO:root:Created job with id: [2016-07-21_17_39_34-11919126913628582743]
INFO:root:To access the Dataflow monitoring console, please navigate to 
https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2016-07-21_17_39_34-11919126913628582743
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b0e27: 
2016-07-22T00:39:35.207Z: JOB_MESSAGE_DETAILED: (66e966f9c1475bc3): Checking 
required Cloud APIs are enabled.
INFO:root:Job 2016-07-21_17_39_34-11919126913628582743 is in state 
JOB_STATE_RUNNING
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b0fcc: 
2016-07-22T00:39:35.628Z: JOB_MESSAGE_DEBUG: (66e966f9c1475d0a): Combiner 
lifting skipped for step group: GroupByKey not followed by a combiner.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b0fcf: 
2016-07-22T00:39:35.631Z: JOB_MESSAGE_DETAILED: (66e966f9c14752a8): Expanding 
GroupByKey operations into optimizable parts.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b0fd2: 
2016-07-22T00:39:35.634Z: JOB_MESSAGE_DETAILED: (66e966f9c1475846): Lifting 
ValueCombiningMappingFns into MergeBucketsMappingFns
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b0fd8: 
2016-07-22T00:39:35.640Z: JOB_MESSAGE_DETAILED: (66e966f9c1475382): Annotating 
graph with Autotuner information.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1005: 
2016-07-22T00:39:35.685Z: JOB_MESSAGE_DETAILED: (66e966f9c14759fa): Fusing 
adjacent ParDo, Read, Write, and Flatten operations
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1008: 
2016-07-22T00:39:35.688Z: JOB_MESSAGE_DETAILED: (66e966f9c1475f98): Fusing 
consumer split into read
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1009: 
2016-07-22T00:39:35.689Z: JOB_MESSAGE_DETAILED: (66e966f9c1475536): Fusing 
consumer group/Reify into pair_with_one
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b100c: 
2016-07-22T00:39:35.692Z: JOB_MESSAGE_DETAILED: (66e966f9c1475ad4): Fusing 
consumer format into count
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b100e: 
2016-07-22T00:39:35.694Z: JOB_MESSAGE_DETAILED: (66e966f9c1475072): Fusing 
consumer count into group/GroupByWindow
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1014: 
2016-07-22T00:39:35.700Z: JOB_MESSAGE_DETAILED: (66e966f9c1475bae): Fusing 
consumer pair_with_one into split
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1017: 
2016-07-22T00:39:35.703Z: JOB_MESSAGE_DETAILED: (66e966f9c147514c): Fusing 
consumer group/GroupByWindow into group/Read
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1019: 
2016-07-22T00:39:35.705Z: JOB_MESSAGE_DETAILED: (66e966f9c14756ea): Fusing 
consumer write/WriteImpl/write_bundles into format
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b101c: 
2016-07-22T00:39:35.708Z: JOB_MESSAGE_DETAILED: (66e966f9c1475c88): Fusing 
consumer group/Write into group/Reify
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1052: 
2016-07-22T00:39:35.762Z: JOB_MESSAGE_DEBUG: (66e966f9c147581c): Workflow 
config is missing a default resource spec.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1055: 
2016-07-22T00:39:35.765Z: JOB_MESSAGE_DETAILED: (66e966f9c1475dba): Adding 
StepResource setup and teardown to workflow graph.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1086: 
2016-07-22T00:39:35.814Z: JOB_MESSAGE_DEBUG: (1607a536476d0625): Adding 
workflow start and stop steps.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1092: 
2016-07-22T00:39:35.826Z: JOB_MESSAGE_DEBUG: (964bb1306ba3ade2): Assigning 
stage ids.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b10f7: 
2016-07-22T00:39:35.927Z: JOB_MESSAGE_DEBUG: (191004cfedb4f4be): Executing wait 
step start2
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1103: 
2016-07-22T00:39:35.939Z: JOB_MESSAGE_DEBUG: (f9b92bf05eb3131b): Executing 
operation write/WriteImpl/DoOnce
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1111: 
2016-07-22T00:39:35.953Z: JOB_MESSAGE_DEBUG: (66e966f9c1475e94): Value 
"write/WriteImpl/DoOnce.out" materialized.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b111d: 
2016-07-22T00:39:35.965Z: JOB_MESSAGE_BASIC: S01: (ca37e07305f6e896): Executing 
operation write/WriteImpl/initialize_write
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1120: 
2016-07-22T00:39:35.968Z: JOB_MESSAGE_BASIC: S02: (946f41bda0d8987f): Executing 
operation group/Create
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b11ea: 
2016-07-22T00:39:36.170Z: JOB_MESSAGE_DEBUG: (a80265b0f93b1617): Starting 
worker pool setup.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b11ec: 
2016-07-22T00:39:36.172Z: JOB_MESSAGE_BASIC: (a80265b0f93b1289): Starting 1 
workers...
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b1200: 
2016-07-22T00:39:36.192Z: JOB_MESSAGE_DEBUG: (a14a60ddd9585b02): Value 
"group/Session" materialized.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100b120d: 
2016-07-22T00:39:36.205Z: JOB_MESSAGE_BASIC: S03: (e85bfb9a3ce83b61): Executing 
operation read+split+pair_with_one+group/Reify+group/Write
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100c6cfa: 
2016-07-22T00:41:05.018Z: JOB_MESSAGE_DETAILED: (77d5029a39d1af44): Workers 
have started successfully.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100d12b0: 
2016-07-22T00:41:47.440Z: JOB_MESSAGE_ERROR: (a562237edf71b227): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 462, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 891, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24041)
    op.start()
  File "dataflow_worker/executor.py", line 477, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14398)
    def start(self):
  File "dataflow_worker/executor.py", line 508, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14333)
    self.dofn_runner.start()
  File "apache_beam/runners/common.py", line 92, in 
apache_beam.runners.common.DoFnRunner.start (apache_beam/runners/common.c:3261)
    self.logging_context.exit()
AttributeError: 'PerThreadLoggingContext' object has no attribute 'exit'

INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100d1cbb: 
2016-07-22T00:41:50.011Z: JOB_MESSAGE_ERROR: (69e9bbf70c56e239): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 462, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 891, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24041)
    op.start()
  File "dataflow_worker/executor.py", line 477, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14398)
    def start(self):
  File "dataflow_worker/executor.py", line 508, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14333)
    self.dofn_runner.start()
  File "apache_beam/runners/common.py", line 92, in 
apache_beam.runners.common.DoFnRunner.start (apache_beam/runners/common.c:3261)
    self.logging_context.exit()
AttributeError: 'PerThreadLoggingContext' object has no attribute 'exit'

INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100d20e2: 
2016-07-22T00:41:51.074Z: JOB_MESSAGE_ERROR: (37d7ddc04502f29d): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 462, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 891, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24041)
    op.start()
  File "dataflow_worker/executor.py", line 477, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14398)
    def start(self):
  File "dataflow_worker/executor.py", line 508, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14333)
    self.dofn_runner.start()
  File "apache_beam/runners/common.py", line 92, in 
apache_beam.runners.common.DoFnRunner.start (apache_beam/runners/common.c:3261)
    self.logging_context.exit()
AttributeError: 'PerThreadLoggingContext' object has no attribute 'exit'

INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100d250c: 
2016-07-22T00:41:52.140Z: JOB_MESSAGE_ERROR: (6a6c8d9ff1dfb2ba): Traceback 
(most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 462, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 891, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24041)
    op.start()
  File "dataflow_worker/executor.py", line 477, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14398)
    def start(self):
  File "dataflow_worker/executor.py", line 508, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14333)
    self.dofn_runner.start()
  File "apache_beam/runners/common.py", line 92, in 
apache_beam.runners.common.DoFnRunner.start (apache_beam/runners/common.c:3261)
    self.logging_context.exit()
AttributeError: 'PerThreadLoggingContext' object has no attribute 'exit'

INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100d252b: 
2016-07-22T00:41:52.171Z: JOB_MESSAGE_WARNING: S01: (3a4ddfa9bb933898): Unable 
to delete temp files: 
"gs://temp-storage-for-end-to-end-tests/temp-wordcount/py-wordcount-1469147972.1469147973.379150/dax-tmp-2016-07-21_17_39_34-11919126913628582743-S01-1-ca37e07305f6e388/[email protected]."
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100d253c: 
2016-07-22T00:41:52.188Z: JOB_MESSAGE_DEBUG: (a14a60ddd9585d64): Executing 
failure step failure1
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100d253e: 
2016-07-22T00:41:52.190Z: JOB_MESSAGE_ERROR: (a14a60ddd95858da): Workflow 
failed. Causes: (ca37e07305f6e415): S01:write/WriteImpl/initialize_write failed.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100d2579: 
2016-07-22T00:41:52.249Z: JOB_MESSAGE_DETAILED: (f9b92bf05eb313be): Cleaning up.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100d25f6: 
2016-07-22T00:41:52.374Z: JOB_MESSAGE_DEBUG: (f9b92bf05eb3164d): Starting 
worker pool teardown.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100d25f8: 
2016-07-22T00:41:52.376Z: JOB_MESSAGE_BASIC: (f9b92bf05eb31d57): Stopping 
worker pool...
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100d294c: 
2016-07-22T00:41:53.228Z: JOB_MESSAGE_DETAILED: (75c8f82f1135ba1c): Failed to 
publish the result of the work update. Causes: (75c8f82f1135b137): Failed to 
update work status. Causes: (9380ca2d8863cec8): Work "1405345380650175154" not 
leased (or the lease was lost).
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100e4af8: 
2016-07-22T00:43:07.384Z: JOB_MESSAGE_BASIC: (f9b92bf05eb31fe6): Worker pool 
stopped.
INFO:root:2016-07-21_17_39_34-11919126913628582743_00000156100e5068: 
2016-07-22T00:43:08.776Z: JOB_MESSAGE_DEBUG: (f9b92bf05eb313bb): Tearing down 
pending resources...
INFO:root:Job 2016-07-21_17_39_34-11919126913628582743 is in state 
JOB_STATE_FAILED
Traceback (most recent call last):
  File "/usr/lib/python2.7/runpy.py", line 162, in _run_module_as_main
    "__main__", fname, loader, pkg_name)
  File "/usr/lib/python2.7/runpy.py", line 72, in _run_code
    exec code in run_globals
  File 
"<https://builds.apache.org/job/beam_PostCommit_PythonVerify/ws/sdks/python/apache_beam/examples/wordcount.py";,>
 line 102, in <module>
    run()
  File 
"<https://builds.apache.org/job/beam_PostCommit_PythonVerify/ws/sdks/python/apache_beam/examples/wordcount.py";,>
 line 93, in run
    result = p.run()
  File "apache_beam/pipeline.py", line 159, in run
    return self.runner.run(self)
  File "apache_beam/runners/dataflow_runner.py", line 188, in run
    % getattr(self, 'last_error_msg', None), self.result)
apache_beam.runners.dataflow_runner.DataflowRuntimeException: Dataflow pipeline 
failed:
(6a6c8d9ff1dfb2ba): Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 462, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 891, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24041)
    op.start()
  File "dataflow_worker/executor.py", line 477, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14398)
    def start(self):
  File "dataflow_worker/executor.py", line 508, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:14333)
    self.dofn_runner.start()
  File "apache_beam/runners/common.py", line 92, in 
apache_beam.runners.common.DoFnRunner.start (apache_beam/runners/common.c:3261)
    self.logging_context.exit()
AttributeError: 'PerThreadLoggingContext' object has no attribute 'exit'


# Grep will exit with status 1 if success message was not found.
echo ">>> CHECKING JOB SUCCESS"
>>> CHECKING JOB SUCCESS
grep JOB_STATE_DONE job_output
Build step 'Execute shell' marked build as failure

Reply via email to