See 
<https://builds.apache.org/job/beam_PostCommit_Python_ValidatesRunner_Dataflow/1158/display/redirect?page=changes>

Changes:

[Pablo] Fixing check for sideinput_io_metrics experiment flag.

[iemejia] Remove testing package-info from main package for GCP IO

[iemejia] Update maven failsafe/surefire plugin to version 2.21.0

[iemejia] [BEAM-3873] Update commons-compress to version 1.16.1 (fix

[iemejia] Remove maven warnings

[tgroh] Add Side Inputs to ExecutableStage

------------------------------------------
[...truncated 758.22 KB...]
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "<lambda>"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "kind:pair", 
                  "component_encodings": [
                    {
                      "@type": "kind:bytes"
                    }, 
                    {
                      "@type": 
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                      "component_encodings": [
                        {
                          "@type": 
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                          "component_encodings": []
                        }, 
                        {
                          "@type": 
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                          "component_encodings": []
                        }
                      ], 
                      "is_pair_like": true
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "compute/MapToVoidKey0.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s2"
        }, 
        "serialized_fn": "<string of 968 bytes>", 
        "user_name": "compute/MapToVoidKey0"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 createTime: u'2018-03-21T22:47:39.744165Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2018-03-21_15_47_38-5872482624041273330'
 location: u'us-central1'
 name: u'beamapp-jenkins-0321224730-414664'
 projectId: u'apache-beam-testing'
 stageStates: []
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2018-03-21_15_47_38-5872482624041273330]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_47_38-5872482624041273330?project=apache-beam-testing
root: INFO: Job 2018-03-21_15_47_38-5872482624041273330 is in state 
JOB_STATE_PENDING
root: INFO: 2018-03-21T22:47:38.816Z: JOB_MESSAGE_WARNING: Job 
2018-03-21_15_47_38-5872482624041273330 might autoscale up to 250 workers.
root: INFO: 2018-03-21T22:47:38.844Z: JOB_MESSAGE_DETAILED: Autoscaling is 
enabled for job 2018-03-21_15_47_38-5872482624041273330. The number of workers 
will be between 1 and 250.
root: INFO: 2018-03-21T22:47:38.873Z: JOB_MESSAGE_DETAILED: Autoscaling was 
automatically enabled for job 2018-03-21_15_47_38-5872482624041273330.
root: INFO: 2018-03-21T22:47:41.640Z: JOB_MESSAGE_DETAILED: Checking required 
Cloud APIs are enabled.
root: INFO: 2018-03-21T22:47:41.817Z: JOB_MESSAGE_DETAILED: Checking 
permissions granted to controller Service Account.
root: INFO: 2018-03-21T22:47:42.618Z: JOB_MESSAGE_DETAILED: Expanding 
CoGroupByKey operations into optimizable parts.
root: INFO: 2018-03-21T22:47:42.648Z: JOB_MESSAGE_DEBUG: Combiner lifting 
skipped for step assert_that/Group/GroupByKey: GroupByKey not followed by a 
combiner.
root: INFO: 2018-03-21T22:47:42.670Z: JOB_MESSAGE_DETAILED: Expanding 
GroupByKey operations into optimizable parts.
root: INFO: 2018-03-21T22:47:42.694Z: JOB_MESSAGE_DETAILED: Lifting 
ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2018-03-21T22:47:42.724Z: JOB_MESSAGE_DEBUG: Annotating graph with 
Autotuner information.
root: INFO: 2018-03-21T22:47:42.758Z: JOB_MESSAGE_DETAILED: Fusing adjacent 
ParDo, Read, Write, and Flatten operations
root: INFO: 2018-03-21T22:47:42.787Z: JOB_MESSAGE_DETAILED: Fusing consumer 
compute/MapToVoidKey0 into side/Read
root: INFO: 2018-03-21T22:47:42.819Z: JOB_MESSAGE_DETAILED: Fusing consumer 
compute/MapToVoidKey0 into side/Read
root: INFO: 2018-03-21T22:47:42.856Z: JOB_MESSAGE_DETAILED: Fusing consumer 
assert_that/Group/GroupByKey/GroupByWindow into 
assert_that/Group/GroupByKey/Read
root: INFO: 2018-03-21T22:47:42.884Z: JOB_MESSAGE_DETAILED: Fusing consumer 
assert_that/Unkey into assert_that/Group/Map(_merge_tagged_vals_under_key)
root: INFO: 2018-03-21T22:47:42.918Z: JOB_MESSAGE_DETAILED: Fusing consumer 
assert_that/Match into assert_that/Unkey
root: INFO: 2018-03-21T22:47:42.942Z: JOB_MESSAGE_DETAILED: Fusing consumer 
assert_that/Group/Map(_merge_tagged_vals_under_key) into 
assert_that/Group/GroupByKey/GroupByWindow
root: INFO: 2018-03-21T22:47:42.968Z: JOB_MESSAGE_DETAILED: Unzipping flatten 
s11 for input s10.out
root: INFO: 2018-03-21T22:47:42.999Z: JOB_MESSAGE_DETAILED: Fusing unzipped 
copy of assert_that/Group/GroupByKey/Reify, through flatten 
assert_that/Group/Flatten, into producer assert_that/Group/pair_with_1
root: INFO: 2018-03-21T22:47:43.013Z: JOB_MESSAGE_DETAILED: Unzipping flatten 
s11-u13 for input s12-reify-value0-c11
root: INFO: 2018-03-21T22:47:43.038Z: JOB_MESSAGE_DETAILED: Fusing unzipped 
copy of assert_that/Group/GroupByKey/Write, through flatten s11-u13, into 
producer assert_that/Group/GroupByKey/Reify
root: INFO: 2018-03-21T22:47:43.070Z: JOB_MESSAGE_DETAILED: Fusing consumer 
assert_that/Group/GroupByKey/Reify into assert_that/Group/pair_with_0
root: INFO: 2018-03-21T22:47:43.098Z: JOB_MESSAGE_DETAILED: Fusing consumer 
assert_that/Group/GroupByKey/Write into assert_that/Group/GroupByKey/Reify
root: INFO: 2018-03-21T22:47:43.132Z: JOB_MESSAGE_DETAILED: Fusing consumer 
assert_that/Group/pair_with_1 into assert_that/ToVoidKey
root: INFO: 2018-03-21T22:47:43.157Z: JOB_MESSAGE_DETAILED: Fusing consumer 
assert_that/WindowInto(WindowIntoFn) into compute/compute
root: INFO: 2018-03-21T22:47:43.191Z: JOB_MESSAGE_DETAILED: Fusing consumer 
compute/compute into start/Read
root: INFO: 2018-03-21T22:47:43.224Z: JOB_MESSAGE_DETAILED: Fusing consumer 
assert_that/Group/pair_with_0 into assert_that/Create/Read
root: INFO: 2018-03-21T22:47:43.250Z: JOB_MESSAGE_DETAILED: Fusing consumer 
assert_that/ToVoidKey into assert_that/WindowInto(WindowIntoFn)
root: INFO: 2018-03-21T22:47:43.284Z: JOB_MESSAGE_DEBUG: Workflow config is 
missing a default resource spec.
root: INFO: 2018-03-21T22:47:43.313Z: JOB_MESSAGE_DEBUG: Adding StepResource 
setup and teardown to workflow graph.
root: INFO: 2018-03-21T22:47:43.369Z: JOB_MESSAGE_DEBUG: Adding workflow start 
and stop steps.
root: INFO: 2018-03-21T22:47:43.401Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2018-03-21T22:47:43.519Z: JOB_MESSAGE_DEBUG: Executing wait step 
start22
root: INFO: 2018-03-21T22:47:43.573Z: JOB_MESSAGE_BASIC: Executing operation 
side/Read+compute/MapToVoidKey0+compute/MapToVoidKey0
root: INFO: 2018-03-21T22:47:43.589Z: JOB_MESSAGE_BASIC: Executing operation 
assert_that/Group/GroupByKey/Create
root: INFO: 2018-03-21T22:47:43.602Z: JOB_MESSAGE_DEBUG: Starting worker pool 
setup.
root: INFO: 2018-03-21T22:47:43.622Z: JOB_MESSAGE_BASIC: Starting 1 workers in 
us-central1-f...
root: INFO: 2018-03-21T22:47:43.720Z: JOB_MESSAGE_DEBUG: Value 
"assert_that/Group/GroupByKey/Session" materialized.
root: INFO: 2018-03-21T22:47:43.791Z: JOB_MESSAGE_BASIC: Executing operation 
assert_that/Create/Read+assert_that/Group/pair_with_0+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: Job 2018-03-21_15_47_38-5872482624041273330 is in state 
JOB_STATE_RUNNING
root: INFO: 2018-03-21T22:47:53.479Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised 
the number of workers to 0 based on the rate of progress in the currently 
running step(s).
root: INFO: 2018-03-21T22:48:09.183Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised 
the number of workers to 1 based on the rate of progress in the currently 
running step(s).
root: INFO: 2018-03-21T22:50:01.400Z: JOB_MESSAGE_DETAILED: Workers have 
started successfully.
root: INFO: 2018-03-21T22:53:14.697Z: JOB_MESSAGE_DEBUG: Value 
"compute/MapToVoidKey0.out" materialized.
root: INFO: 2018-03-21T22:53:14.762Z: JOB_MESSAGE_BASIC: Executing operation 
compute/_DataflowIterableSideInput(MapToVoidKey0.out.0)
root: INFO: 2018-03-21T22:53:14.872Z: JOB_MESSAGE_DEBUG: Value 
"compute/_DataflowIterableSideInput(MapToVoidKey0.out.0).output" materialized.
root: INFO: 2018-03-21T22:53:14.929Z: JOB_MESSAGE_BASIC: Executing operation 
start/Read+compute/compute+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
root: INFO: 2018-03-21T22:53:23.795Z: JOB_MESSAGE_ERROR: Traceback (most recent 
call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 609, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", 
line 167, in execute
    op.start()
  File "apache_beam/runners/worker/operations.py", line 339, in 
apache_beam.runners.worker.operations.DoOperation.start
    def start(self):
  File "apache_beam/runners/worker/operations.py", line 340, in 
apache_beam.runners.worker.operations.DoOperation.start
    with self.scoped_start_state:
  File "apache_beam/runners/worker/operations.py", line 372, in 
apache_beam.runners.worker.operations.DoOperation.start
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 483, in 
apache_beam.runners.common.DoFnRunner.__init__
    self.do_fn_invoker = DoFnInvoker.create_invoker(
  File "apache_beam/runners/common.py", line 203, in 
apache_beam.runners.common.DoFnInvoker.create_invoker
    return PerWindowInvoker(
  File "apache_beam/runners/common.py", line 313, in 
apache_beam.runners.common.PerWindowInvoker.__init__
    input_args, input_kwargs, [si[global_window] for si in side_inputs])
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 62, in __getitem__
    self._cache[window] = self._view_class._from_runtime_iterable(
AttributeError: type object '_DataflowIterableSideInput' has no attribute 
'_from_runtime_iterable'

root: INFO: 2018-03-21T22:53:27.180Z: JOB_MESSAGE_ERROR: Traceback (most recent 
call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 609, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", 
line 167, in execute
    op.start()
  File "apache_beam/runners/worker/operations.py", line 339, in 
apache_beam.runners.worker.operations.DoOperation.start
    def start(self):
  File "apache_beam/runners/worker/operations.py", line 340, in 
apache_beam.runners.worker.operations.DoOperation.start
    with self.scoped_start_state:
  File "apache_beam/runners/worker/operations.py", line 372, in 
apache_beam.runners.worker.operations.DoOperation.start
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 483, in 
apache_beam.runners.common.DoFnRunner.__init__
    self.do_fn_invoker = DoFnInvoker.create_invoker(
  File "apache_beam/runners/common.py", line 203, in 
apache_beam.runners.common.DoFnInvoker.create_invoker
    return PerWindowInvoker(
  File "apache_beam/runners/common.py", line 313, in 
apache_beam.runners.common.PerWindowInvoker.__init__
    input_args, input_kwargs, [si[global_window] for si in side_inputs])
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 62, in __getitem__
    self._cache[window] = self._view_class._from_runtime_iterable(
AttributeError: type object '_DataflowIterableSideInput' has no attribute 
'_from_runtime_iterable'

root: INFO: 2018-03-21T22:53:30.569Z: JOB_MESSAGE_ERROR: Traceback (most recent 
call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 609, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", 
line 167, in execute
    op.start()
  File "apache_beam/runners/worker/operations.py", line 339, in 
apache_beam.runners.worker.operations.DoOperation.start
    def start(self):
  File "apache_beam/runners/worker/operations.py", line 340, in 
apache_beam.runners.worker.operations.DoOperation.start
    with self.scoped_start_state:
  File "apache_beam/runners/worker/operations.py", line 372, in 
apache_beam.runners.worker.operations.DoOperation.start
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 483, in 
apache_beam.runners.common.DoFnRunner.__init__
    self.do_fn_invoker = DoFnInvoker.create_invoker(
  File "apache_beam/runners/common.py", line 203, in 
apache_beam.runners.common.DoFnInvoker.create_invoker
    return PerWindowInvoker(
  File "apache_beam/runners/common.py", line 313, in 
apache_beam.runners.common.PerWindowInvoker.__init__
    input_args, input_kwargs, [si[global_window] for si in side_inputs])
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 62, in __getitem__
    self._cache[window] = self._view_class._from_runtime_iterable(
AttributeError: type object '_DataflowIterableSideInput' has no attribute 
'_from_runtime_iterable'

root: INFO: 2018-03-21T22:53:33.934Z: JOB_MESSAGE_ERROR: Traceback (most recent 
call last):
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/batchworker.py", 
line 609, in do_work
    work_executor.execute()
  File "/usr/local/lib/python2.7/dist-packages/dataflow_worker/executor.py", 
line 167, in execute
    op.start()
  File "apache_beam/runners/worker/operations.py", line 339, in 
apache_beam.runners.worker.operations.DoOperation.start
    def start(self):
  File "apache_beam/runners/worker/operations.py", line 340, in 
apache_beam.runners.worker.operations.DoOperation.start
    with self.scoped_start_state:
  File "apache_beam/runners/worker/operations.py", line 372, in 
apache_beam.runners.worker.operations.DoOperation.start
    self.dofn_runner = common.DoFnRunner(
  File "apache_beam/runners/common.py", line 483, in 
apache_beam.runners.common.DoFnRunner.__init__
    self.do_fn_invoker = DoFnInvoker.create_invoker(
  File "apache_beam/runners/common.py", line 203, in 
apache_beam.runners.common.DoFnInvoker.create_invoker
    return PerWindowInvoker(
  File "apache_beam/runners/common.py", line 313, in 
apache_beam.runners.common.PerWindowInvoker.__init__
    input_args, input_kwargs, [si[global_window] for si in side_inputs])
  File 
"/usr/local/lib/python2.7/dist-packages/apache_beam/transforms/sideinputs.py", 
line 62, in __getitem__
    self._cache[window] = self._view_class._from_runtime_iterable(
AttributeError: type object '_DataflowIterableSideInput' has no attribute 
'_from_runtime_iterable'

root: INFO: 2018-03-21T22:53:33.980Z: JOB_MESSAGE_DEBUG: Executing failure step 
failure21
root: INFO: 2018-03-21T22:53:34.008Z: JOB_MESSAGE_ERROR: Workflow failed. 
Causes: 
S05:start/Read+compute/compute+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write
 failed., A work item was attempted 4 times without success. Each time the 
worker eventually lost contact with the service. The work item was attempted 
on: 
  beamapp-jenkins-032122473-03211547-4579-harness-nf9m,
  beamapp-jenkins-032122473-03211547-4579-harness-nf9m,
  beamapp-jenkins-032122473-03211547-4579-harness-nf9m,
  beamapp-jenkins-032122473-03211547-4579-harness-nf9m
root: INFO: 2018-03-21T22:53:34.125Z: JOB_MESSAGE_DETAILED: Cleaning up.
root: INFO: 2018-03-21T22:53:34.172Z: JOB_MESSAGE_DEBUG: Starting worker pool 
teardown.
root: INFO: 2018-03-21T22:53:34.205Z: JOB_MESSAGE_BASIC: Stopping worker pool...
root: INFO: 2018-03-21T22:55:01.815Z: JOB_MESSAGE_DETAILED: Autoscaling: 
Reduced the number of workers to 0 based on the rate of progress in the 
currently running step(s).
root: INFO: 2018-03-21T22:55:01.867Z: JOB_MESSAGE_DEBUG: Tearing down pending 
resources...
root: INFO: Job 2018-03-21_15_47_38-5872482624041273330 is in state 
JOB_STATE_FAILED
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 16 tests in 1803.189s

FAILED (errors=9)
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_25_24-4448359334982537282?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_32_53-565034839070729173?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_40_53-4972656836611472700?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_47_38-5872482624041273330?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_25_24-5145634561768741218?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_33_26-17951261812159407481?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_40_51-2658859845955245725?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_47_50-13995110321908084770?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_25_23-12639694579356003338?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_33_22-415921566333085740?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_40_32-17113019966122211081?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_47_57-553940328731293000?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_25_23-17606291834145499369?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_33_23-5491383758640923186?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_40_36-16168717961000468113?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-03-21_15_48_00-15660015474431706626?project=apache-beam-testing
Build step 'Execute shell' marked build as failure
Not sending mail to unregistered user [email protected]
Not sending mail to unregistered user [email protected]

Reply via email to