See <https://builds.apache.org/job/beam_PostCommit_Python_Verify/857/>

------------------------------------------
[...truncated 2746 lines...]
                    {
                      "@type": 
"PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
 
                      "component_encodings": [
                        {
                          "@type": 
"PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
 
                          "component_encodings": []
                        }, 
                        {
                          "@type": 
"PickleCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwBmcnZOanOIA5XIYNmYyFjbSFTkh4A9DYR/g==",
 
                          "component_encodings": []
                        }
                      ], 
                      "is_pair_like": true
                    }
                  ], 
                  "is_wrapper": true
                }
              ]
            }, 
            "output_name": "out", 
            "user_name": "ViewAsSingleton(side.None)/CreatePCollectionView.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s2"
        }, 
        "user_name": "ViewAsSingleton(side.None)/CreatePCollectionView"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s4", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "<lambda>"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {
          "s3": {
            "@type": "OutputReference", 
            "output_name": "out", 
            "step_name": "s3"
          }
        }, 
        "output_info": [
          {
            "encoding": {
              "@type": "WindowedValueCoder$<string of 408 bytes>", 
              "component_encodings": [
                {
                  "@type": 
"FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                  "component_encodings": [
                    {
                      "@type": 
"FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                      "component_encodings": []
                    }, 
                    {
                      "@type": 
"FastPrimitivesCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": 
"TimestampCoder$eJxrYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlwhmbmpxSWJuQXOID5XIYNmYyFjbSFTkh4ANWETWg==",
 
                  "component_encodings": []
                }, 
                {
                  "@type": "SingletonCoder$<string of 256 bytes>", 
                  "component_encodings": []
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "compute.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s1"
        }, 
        "serialized_fn": "<string of 1116 bytes>", 
        "user_name": "compute"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 id: u'2016-12-10_13_49_18-3339530418630653316'
 projectId: u'apache-beam-testing'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2016-12-10_13_49_18-3339530418630653316]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.developers.google.com/project/apache-beam-testing/dataflow/job/2016-12-10_13_49_18-3339530418630653316
root: INFO: Job 2016-12-10_13_49_18-3339530418630653316 is in state 
JOB_STATE_RUNNING
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab63465: 
2016-12-10T21:49:19.333Z: JOB_MESSAGE_DETAILED: (6ef52d1b87a86642): Checking 
required Cloud APIs are enabled.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab635b8: 
2016-12-10T21:49:19.672Z: JOB_MESSAGE_DETAILED: (6ef52d1b87a86e19): Expanding 
GroupByKey operations into optimizable parts.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab635ba: 
2016-12-10T21:49:19.674Z: JOB_MESSAGE_DETAILED: (6ef52d1b87a86fd7): Lifting 
ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab635bd: 
2016-12-10T21:49:19.677Z: JOB_MESSAGE_DETAILED: (6ef52d1b87a86195): Annotating 
graph with Autotuner information.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab635c6: 
2016-12-10T21:49:19.686Z: JOB_MESSAGE_DETAILED: (6ef52d1b87a866cf): Fusing 
adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab635d6: 
2016-12-10T21:49:19.702Z: JOB_MESSAGE_DEBUG: (6ef52d1b87a86301): Workflow 
config is missing a default resource spec.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab635d9: 
2016-12-10T21:49:19.705Z: JOB_MESSAGE_DETAILED: (6ef52d1b87a864bf): Adding 
StepResource setup and teardown to workflow graph.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab63615: 
2016-12-10T21:49:19.765Z: JOB_MESSAGE_DEBUG: (9283d660b09d6ceb): Adding 
workflow start and stop steps.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab63620: 
2016-12-10T21:49:19.776Z: JOB_MESSAGE_DEBUG: (18957f9ce3127afa): Assigning 
stage ids.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab636b7: 
2016-12-10T21:49:19.927Z: JOB_MESSAGE_DEBUG: (a1f224083d18fed4): Executing wait 
step start2
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab636c4: 
2016-12-10T21:49:19.940Z: JOB_MESSAGE_DEBUG: (683e319ea146a34f): Executing 
operation start
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab636d1: 
2016-12-10T21:49:19.953Z: JOB_MESSAGE_DEBUG: (18957f9ce3127746): Value 
"side.out" materialized.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab636dc: 
2016-12-10T21:49:19.964Z: JOB_MESSAGE_BASIC: S01: (e0125a6196999a26): Executing 
operation ViewAsSingleton(side.None)/CreatePCollectionView
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab636e9: 
2016-12-10T21:49:19.977Z: JOB_MESSAGE_DEBUG: (58a227220900858f): Value 
"ViewAsSingleton(side.None)/CreatePCollectionView.out" materialized.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab636f3: 
2016-12-10T21:49:19.987Z: JOB_MESSAGE_BASIC: S02: (b53a5faa9097e1a5): Executing 
operation compute
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab637bf: 
2016-12-10T21:49:20.191Z: JOB_MESSAGE_DEBUG: (1a38e6d40e151659): Starting 
worker pool setup.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab637c1: 
2016-12-10T21:49:20.193Z: JOB_MESSAGE_BASIC: (1a38e6d40e1515f3): Starting 1 
workers...
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab801fe: 
2016-12-10T21:51:17.502Z: JOB_MESSAGE_DETAILED: (a024fdc9eba91da5): Workers 
have started successfully.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab8ae0a: 
2016-12-10T21:52:01.546Z: JOB_MESSAGE_ERROR: (f303cdc53c0dd1e): Traceback (most 
recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 892, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24008)
    op.start()
  File "dataflow_worker/executor.py", line 456, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
    def start(self):
  File "dataflow_worker/executor.py", line 483, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 89, in 
apache_beam.runners.common.DoFnRunner.__init__ 
(apache_beam/runners/common.c:3469)
    args, kwargs, [side_input[global_window]
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 192, in __getitem__
    _FilteringIterable(self._iterable, target_window), self._view_options)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 
279, in _from_runtime_iterable
    'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.

root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab8ae73: 
2016-12-10T21:52:01.651Z: JOB_MESSAGE_ERROR: (f303cdc53c0d0ee): Traceback (most 
recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 892, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24008)
    op.start()
  File "dataflow_worker/executor.py", line 456, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
    def start(self):
  File "dataflow_worker/executor.py", line 483, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 89, in 
apache_beam.runners.common.DoFnRunner.__init__ 
(apache_beam/runners/common.c:3469)
    args, kwargs, [side_input[global_window]
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 192, in __getitem__
    _FilteringIterable(self._iterable, target_window), self._view_options)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 
279, in _from_runtime_iterable
    'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.

root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab8aed5: 
2016-12-10T21:52:01.749Z: JOB_MESSAGE_ERROR: (f303cdc53c0d4be): Traceback (most 
recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 892, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24008)
    op.start()
  File "dataflow_worker/executor.py", line 456, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
    def start(self):
  File "dataflow_worker/executor.py", line 483, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 89, in 
apache_beam.runners.common.DoFnRunner.__init__ 
(apache_beam/runners/common.c:3469)
    args, kwargs, [side_input[global_window]
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 192, in __getitem__
    _FilteringIterable(self._iterable, target_window), self._view_options)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 
279, in _from_runtime_iterable
    'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.

root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab8af3a: 
2016-12-10T21:52:01.850Z: JOB_MESSAGE_ERROR: (f303cdc53c0d88e): Traceback (most 
recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 892, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24008)
    op.start()
  File "dataflow_worker/executor.py", line 456, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
    def start(self):
  File "dataflow_worker/executor.py", line 483, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 89, in 
apache_beam.runners.common.DoFnRunner.__init__ 
(apache_beam/runners/common.c:3469)
    args, kwargs, [side_input[global_window]
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 192, in __getitem__
    _FilteringIterable(self._iterable, target_window), self._view_options)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 
279, in _from_runtime_iterable
    'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.

root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab8afc8: 
2016-12-10T21:52:01.992Z: JOB_MESSAGE_ERROR: (f303cdc53c0dc5e): Traceback (most 
recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 892, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24008)
    op.start()
  File "dataflow_worker/executor.py", line 456, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
    def start(self):
  File "dataflow_worker/executor.py", line 483, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 89, in 
apache_beam.runners.common.DoFnRunner.__init__ 
(apache_beam/runners/common.c:3469)
    args, kwargs, [side_input[global_window]
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 192, in __getitem__
    _FilteringIterable(self._iterable, target_window), self._view_options)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 
279, in _from_runtime_iterable
    'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.

root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab8b027: 
2016-12-10T21:52:02.087Z: JOB_MESSAGE_ERROR: (f303cdc53c0d02e): Traceback (most 
recent call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 514, in do_work
    work_executor.execute()
  File "dataflow_worker/executor.py", line 892, in 
dataflow_worker.executor.MapTaskExecutor.execute 
(dataflow_worker/executor.c:24008)
    op.start()
  File "dataflow_worker/executor.py", line 456, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13870)
    def start(self):
  File "dataflow_worker/executor.py", line 483, in 
dataflow_worker.executor.DoOperation.start (dataflow_worker/executor.c:13685)
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 89, in 
apache_beam.runners.common.DoFnRunner.__init__ 
(apache_beam/runners/common.c:3469)
    args, kwargs, [side_input[global_window]
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 192, in __getitem__
    _FilteringIterable(self._iterable, target_window), self._view_options)
  File "/usr/local/lib/python2.7/dist-packages/apache_beam/pvalue.py", line 
279, in _from_runtime_iterable
    'PCollection with more than one element accessed as '
ValueError: PCollection with more than one element accessed as a singleton view.

root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab8b054: 
2016-12-10T21:52:02.132Z: JOB_MESSAGE_DEBUG: (d84ad0e359732359): Executing 
failure step failure1
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab8b057: 
2016-12-10T21:52:02.135Z: JOB_MESSAGE_ERROR: (d84ad0e359732bfb): Workflow 
failed. Causes: (b53a5faa9097e16e): S02:compute failed.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab8b08e: 
2016-12-10T21:52:02.190Z: JOB_MESSAGE_DETAILED: (c3165e279e60eef4): Cleaning up.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab8b10e: 
2016-12-10T21:52:02.318Z: JOB_MESSAGE_DEBUG: (c3165e279e60e7ab): Starting 
worker pool teardown.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab8b111: 
2016-12-10T21:52:02.321Z: JOB_MESSAGE_BASIC: (c3165e279e60e825): Stopping 
worker pool...
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab9d610: 
2016-12-10T21:53:17.328Z: JOB_MESSAGE_BASIC: (c3165e279e60e0dc): Worker pool 
stopped.
root: INFO: 2016-12-10_13_49_18-3339530418630653316_00000158eab9d624: 
2016-12-10T21:53:17.348Z: JOB_MESSAGE_DEBUG: (c3165e279e60ea0d): Tearing down 
pending resources...
root: INFO: Job 2016-12-10_13_49_18-3339530418630653316 is in state 
JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 21 tests in 5126.617s

FAILED (errors=1)
Build step 'Execute shell' marked build as failure

Reply via email to