See 
<https://builds.apache.org/job/beam_PostCommit_Python_Verify/2886/display/redirect>

------------------------------------------
[...truncated 638.62 KB...]
      "name": "SideInput-s16", 
      "properties": {
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": "kind:windowed_value", 
                  "component_encodings": [
                    {
                      "@type": 
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                      "component_encodings": [
                        {
                          "@type": 
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                          "component_encodings": []
                        }, 
                        {
                          "@type": 
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                          "component_encodings": []
                        }
                      ], 
                      "is_pair_like": true
                    }, 
                    {
                      "@type": "kind:global_window"
                    }
                  ], 
                  "is_wrapper": true
                }
              ]
            }, 
            "output_name": "out", 
            "user_name": 
"write/Write/WriteImpl/FinalizeWrite/AsIter(Extract.out.0).output"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s14"
        }, 
        "user_name": "write/Write/WriteImpl/FinalizeWrite/AsIter(Extract.out.0)"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s17", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "_finalize_write"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {
          "SideInput-s15": {
            "@type": "OutputReference", 
            "output_name": "out", 
            "step_name": "SideInput-s15"
          }, 
          "SideInput-s16": {
            "@type": "OutputReference", 
            "output_name": "out", 
            "step_name": "SideInput-s16"
          }
        }, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": 
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                  "component_encodings": [
                    {
                      "@type": 
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                      "component_encodings": []
                    }, 
                    {
                      "@type": 
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "write/Write/WriteImpl/FinalizeWrite.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s7"
        }, 
        "serialized_fn": "<string of 1056 bytes>", 
        "user_name": "write/Write/WriteImpl/FinalizeWrite/FinalizeWrite"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
root: INFO: Create job: <Job
 createTime: u'2017-08-08T20:48:32.174437Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2017-08-08_13_48_31-14349531698595810259'
 location: u'global'
 name: u'beamapp-jenkins-0808204830-470453'
 projectId: u'apache-beam-testing'
 stageStates: []
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2017-08-08_13_48_31-14349531698595810259]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-08-08_13_48_31-14349531698595810259?project=apache-beam-testing
root: INFO: Job 2017-08-08_13_48_31-14349531698595810259 is in state 
JOB_STATE_RUNNING
root: INFO: 2017-08-08T20:48:31.651Z: JOB_MESSAGE_WARNING: (c723c70d8402c0ba): 
Setting the number of workers (1) disables autoscaling for this job. If you are 
trying to cap autoscaling, consider only setting max_num_workers. If you want 
to disable autoscaling altogether, the documented way is to explicitly use 
autoscalingAlgorithm=NONE.
root: INFO: 2017-08-08T20:48:34.006Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6afd3): 
Checking required Cloud APIs are enabled.
root: INFO: 2017-08-08T20:48:35.050Z: JOB_MESSAGE_DEBUG: (e2839b1a0fb6aaee): 
Combiner lifting skipped for step write/Write/WriteImpl/GroupByKey: GroupByKey 
not followed by a combiner.
root: INFO: 2017-08-08T20:48:35.052Z: JOB_MESSAGE_DEBUG: (e2839b1a0fb6ab58): 
Combiner lifting skipped for step group: GroupByKey not followed by a combiner.
root: INFO: 2017-08-08T20:48:35.055Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6abc2): 
Expanding GroupByKey operations into optimizable parts.
root: INFO: 2017-08-08T20:48:35.058Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6ac2c): 
Lifting ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2017-08-08T20:48:35.066Z: JOB_MESSAGE_DEBUG: (e2839b1a0fb6ad6a): 
Annotating graph with Autotuner information.
root: INFO: 2017-08-08T20:48:35.078Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6ae3e): 
Fusing adjacent ParDo, Read, Write, and Flatten operations
root: INFO: 2017-08-08T20:48:35.082Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6aea8): 
Fusing consumer split into read/Read
root: INFO: 2017-08-08T20:48:35.084Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6af12): 
Fusing consumer group/Write into group/Reify
root: INFO: 2017-08-08T20:48:35.087Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6af7c): 
Fusing consumer group/GroupByWindow into group/Read
root: INFO: 2017-08-08T20:48:35.089Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6afe6): 
Fusing consumer write/Write/WriteImpl/GroupByKey/GroupByWindow into 
write/Write/WriteImpl/GroupByKey/Read
root: INFO: 2017-08-08T20:48:35.092Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a050): 
Fusing consumer write/Write/WriteImpl/GroupByKey/Write into 
write/Write/WriteImpl/GroupByKey/Reify
root: INFO: 2017-08-08T20:48:35.094Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a0ba): 
Fusing consumer write/Write/WriteImpl/WindowInto(WindowIntoFn) into 
write/Write/WriteImpl/Pair
root: INFO: 2017-08-08T20:48:35.096Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a124): 
Fusing consumer write/Write/WriteImpl/GroupByKey/Reify into 
write/Write/WriteImpl/WindowInto(WindowIntoFn)
root: INFO: 2017-08-08T20:48:35.099Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a18e): 
Fusing consumer pair_with_one into split
root: INFO: 2017-08-08T20:48:35.101Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a1f8): 
Fusing consumer group/Reify into pair_with_one
root: INFO: 2017-08-08T20:48:35.103Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a262): 
Fusing consumer write/Write/WriteImpl/WriteBundles/WriteBundles into format
root: INFO: 2017-08-08T20:48:35.106Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a2cc): 
Fusing consumer write/Write/WriteImpl/Pair into 
write/Write/WriteImpl/WriteBundles/WriteBundles
root: INFO: 2017-08-08T20:48:35.109Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a336): 
Fusing consumer format into count
root: INFO: 2017-08-08T20:48:35.111Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a3a0): 
Fusing consumer write/Write/WriteImpl/Extract into 
write/Write/WriteImpl/GroupByKey/GroupByWindow
root: INFO: 2017-08-08T20:48:35.115Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a40a): 
Fusing consumer count into group/GroupByWindow
root: INFO: 2017-08-08T20:48:35.117Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a474): 
Fusing consumer write/Write/WriteImpl/InitializeWrite into 
write/Write/WriteImpl/DoOnce/Read
root: INFO: 2017-08-08T20:48:35.130Z: JOB_MESSAGE_DEBUG: (e2839b1a0fb6a686): 
Workflow config is missing a default resource spec.
root: INFO: 2017-08-08T20:48:35.133Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a6f0): 
Adding StepResource setup and teardown to workflow graph.
root: INFO: 2017-08-08T20:48:35.136Z: JOB_MESSAGE_DEBUG: (e2839b1a0fb6a75a): 
Adding workflow start and stop steps.
root: INFO: 2017-08-08T20:48:35.139Z: JOB_MESSAGE_DEBUG: (e2839b1a0fb6a7c4): 
Assigning stage ids.
root: INFO: 2017-08-08T20:48:35.193Z: JOB_MESSAGE_DEBUG: (769653ea69118890): 
Executing wait step start25
root: INFO: 2017-08-08T20:48:35.205Z: JOB_MESSAGE_BASIC: (6ee333daa69492a8): 
Executing operation 
write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite
root: INFO: 2017-08-08T20:48:35.213Z: JOB_MESSAGE_BASIC: (769653ea6911872e): 
Executing operation group/Create
root: INFO: 2017-08-08T20:48:35.409Z: JOB_MESSAGE_DEBUG: (46bab0729da20e7f): 
Starting worker pool setup.
root: INFO: 2017-08-08T20:48:35.411Z: JOB_MESSAGE_BASIC: (46bab0729da20bb9): 
Starting 1 workers in us-central1-f...
root: INFO: 2017-08-08T20:48:35.434Z: JOB_MESSAGE_DEBUG: (14ff05f036c64a10): 
Value "group/Session" materialized.
root: INFO: 2017-08-08T20:48:35.448Z: JOB_MESSAGE_BASIC: (14ff05f036c64f21): 
Executing operation read/Read+split+pair_with_one+group/Reify+group/Write
root: INFO: 2017-08-08T20:49:16.588Z: JOB_MESSAGE_DETAILED: (8692f1679c5de6aa): 
Workers have started successfully.
root: INFO: 2017-08-08T20:51:54.088Z: JOB_MESSAGE_DEBUG: (a7612a5edc2da54d): 
Value "write/Write/WriteImpl/DoOnce/Read.out" materialized.
root: INFO: 2017-08-08T20:51:54.099Z: JOB_MESSAGE_DEBUG: (6ee333daa694930b): 
Value "write/Write/WriteImpl/InitializeWrite.out" materialized.
root: INFO: 2017-08-08T20:51:54.112Z: JOB_MESSAGE_BASIC: (a7612a5edc2dac95): 
Executing operation 
write/Write/WriteImpl/FinalizeWrite/AsSingleton(InitializeWrite.out.0)
root: INFO: 2017-08-08T20:51:54.119Z: JOB_MESSAGE_BASIC: (6ee333daa6949fe8): 
Executing operation 
write/Write/WriteImpl/WriteBundles/AsSingleton(InitializeWrite.out.0)
root: INFO: 2017-08-08T20:51:54.177Z: JOB_MESSAGE_DEBUG: (dcfdc09735e96b58): 
Value 
"write/Write/WriteImpl/WriteBundles/AsSingleton(InitializeWrite.out.0).output" 
materialized.
root: INFO: 2017-08-08T20:51:54.187Z: JOB_MESSAGE_DEBUG: (a7612a5edc2daf81): 
Value 
"write/Write/WriteImpl/FinalizeWrite/AsSingleton(InitializeWrite.out.0).output" 
materialized.
root: INFO: 2017-08-08T20:52:08.029Z: JOB_MESSAGE_BASIC: (a7612a5edc2da5b2): 
Executing operation group/Close
root: INFO: 2017-08-08T20:52:08.042Z: JOB_MESSAGE_BASIC: (14ff05f036c64f56): 
Executing operation write/Write/WriteImpl/GroupByKey/Create
root: INFO: 2017-08-08T20:52:08.231Z: JOB_MESSAGE_DEBUG: (a7612a5edc2dacfa): 
Value "write/Write/WriteImpl/GroupByKey/Session" materialized.
root: INFO: 2017-08-08T20:52:08.245Z: JOB_MESSAGE_BASIC: (a7612a5edc2da442): 
Executing operation 
group/Read+group/GroupByWindow+count+format+write/Write/WriteImpl/WriteBundles/WriteBundles+write/Write/WriteImpl/Pair+write/Write/WriteImpl/WindowInto(WindowIntoFn)+write/Write/WriteImpl/GroupByKey/Reify+write/Write/WriteImpl/GroupByKey/Write
root: INFO: 2017-08-08T20:52:21.076Z: JOB_MESSAGE_BASIC: (14ff05f036c640c2): 
Executing operation write/Write/WriteImpl/GroupByKey/Close
root: INFO: 2017-08-08T20:52:21.090Z: JOB_MESSAGE_BASIC: (a7612a5edc2da4a7): 
Executing operation 
write/Write/WriteImpl/GroupByKey/Read+write/Write/WriteImpl/GroupByKey/GroupByWindow+write/Write/WriteImpl/Extract
root: INFO: 2017-08-08T20:52:29.217Z: JOB_MESSAGE_DEBUG: (14ff05f036c64365): 
Value "write/Write/WriteImpl/Extract.out" materialized.
root: INFO: 2017-08-08T20:52:29.235Z: JOB_MESSAGE_BASIC: (a7612a5edc2da1c7): 
Executing operation write/Write/WriteImpl/FinalizeWrite/AsIter(Extract.out.0)
root: INFO: 2017-08-08T20:52:29.295Z: JOB_MESSAGE_DEBUG: (14ff05f036c649ad): 
Value "write/Write/WriteImpl/FinalizeWrite/AsIter(Extract.out.0).output" 
materialized.
root: INFO: 2017-08-08T20:52:29.326Z: JOB_MESSAGE_BASIC: (a7612a5edc2da5ca): 
Executing operation write/Write/WriteImpl/FinalizeWrite/FinalizeWrite
root: INFO: 2017-08-08T20:52:32.026Z: JOB_MESSAGE_DEBUG: (14ff05f036c64ebe): 
Executing success step success23
root: INFO: 2017-08-08T20:52:32.299Z: JOB_MESSAGE_DETAILED: (e2839b1a0fb6a9d6): 
Cleaning up.
root: INFO: 2017-08-08T20:52:32.302Z: JOB_MESSAGE_DEBUG: (e2839b1a0fb6aa40): 
Starting worker pool teardown.
root: INFO: 2017-08-08T20:52:32.304Z: JOB_MESSAGE_BASIC: (e2839b1a0fb6aaaa): 
Stopping worker pool...
root: INFO: 2017-08-08T20:53:40.913Z: JOB_MESSAGE_DEBUG: (e2839b1a0fb6ad26): 
Tearing down pending resources...
root: INFO: Job 2017-08-08_13_48_31-14349531698595810259 is in state 
JOB_STATE_DONE
root: INFO: Wait 20 seconds...
root: DEBUG: Connecting using Google Application Default Credentials.
root: INFO: Starting the size estimation of the input
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
root: INFO: Finished the size estimation of the input at 3 files. Estimation 
took 0.0699858665466 seconds
root: INFO: Find 3 files in 
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1502225310/results*-of-*:
 
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1502225310/results-00000-of-00003
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1502225310/results-00001-of-00003
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1502225310/results-00002-of-00003
root: DEBUG: Connecting using Google Application Default Credentials.
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
root: DEBUG: Connecting using Google Application Default Credentials.
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
root: DEBUG: Connecting using Google Application Default Credentials.
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
root: INFO: Read from given path 
gs://temp-storage-for-end-to-end-tests/py-it-cloud/output/1502225310/results*-of-*,
 4784 lines, checksum: 33535a832b7db6d78389759577d4ff495980b9c0.
root: INFO: number of empty lines: 1663
root: INFO: average word length: 4
root: WARNING: Retry with exponential backoff: waiting for 4.99814241251 
seconds before retrying delete_files because we caught exception: BeamIOError: 
Delete operation failed with exceptions {'*': IOError(OSError(2, 'No such file 
or directory'),), '-': IOError(OSError(2, 'No such file or directory'),), '/': 
IOError(OSError(13, 'Permission denied'),), '1': IOError(OSError(2, 'No such 
file or directory'),), '0': IOError(OSError(2, 'No such file or directory'),), 
'3': IOError(OSError(2, 'No such file or directory'),), '2': IOError(OSError(2, 
'No such file or directory'),), '5': IOError(OSError(2, 'No such file or 
directory'),), ':': IOError(OSError(2, 'No such file or directory'),), 'a': 
IOError(OSError(2, 'No such file or directory'),), 'c': IOError(OSError(2, 'No 
such file or directory'),), 'e': IOError(OSError(2, 'No such file or 
directory'),), 'd': IOError(OSError(2, 'No such file or directory'),), 'g': 
IOError(OSError(2, 'No such file or directory'),), 'f': IOError(OSError(2, 'No 
such file or directory'),), 'i': IOError(OSError(2, 'No such file or 
directory'),), 'm': IOError(OSError(2, 'No such file or directory'),), 'l': 
IOError(OSError(2, 'No such file or directory'),), 'o': IOError(OSError(2, 'No 
such file or directory'),), 'n': IOError(OSError(2, 'No such file or 
directory'),), 'p': IOError(OSError(2, 'No such file or directory'),), 's': 
IOError(OSError(2, 'No such file or directory'),), 'r': IOError(OSError(2, 'No 
such file or directory'),), 'u': IOError(OSError(2, 'No such file or 
directory'),), 't': IOError(OSError(2, 'No such file or directory'),), 'y': 
IOError(OSError(2, 'No such file or directory'),)}
 Traceback for above exception (most recent call last):
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/utils/retry.py";,>
 line 175, in wrapper
    return fun(*args, **kwargs)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/testing/test_utils.py";,>
 line 91, in delete_files
    FileSystems.delete(file_paths)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/io/filesystems.py";,>
 line 224, in delete
    return filesystem.delete(paths)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/io/localfilesystem.py";,>
 line 267, in delete
    raise BeamIOError("Delete operation failed", exceptions)

root: WARNING: Retry with exponential backoff: waiting for 7.93785468515 
seconds before retrying delete_files because we caught exception: BeamIOError: 
Delete operation failed with exceptions {'*': IOError(OSError(2, 'No such file 
or directory'),), '-': IOError(OSError(2, 'No such file or directory'),), '/': 
IOError(OSError(13, 'Permission denied'),), '1': IOError(OSError(2, 'No such 
file or directory'),), '0': IOError(OSError(2, 'No such file or directory'),), 
'3': IOError(OSError(2, 'No such file or directory'),), '2': IOError(OSError(2, 
'No such file or directory'),), '5': IOError(OSError(2, 'No such file or 
directory'),), ':': IOError(OSError(2, 'No such file or directory'),), 'a': 
IOError(OSError(2, 'No such file or directory'),), 'c': IOError(OSError(2, 'No 
such file or directory'),), 'e': IOError(OSError(2, 'No such file or 
directory'),), 'd': IOError(OSError(2, 'No such file or directory'),), 'g': 
IOError(OSError(2, 'No such file or directory'),), 'f': IOError(OSError(2, 'No 
such file or directory'),), 'i': IOError(OSError(2, 'No such file or 
directory'),), 'm': IOError(OSError(2, 'No such file or directory'),), 'l': 
IOError(OSError(2, 'No such file or directory'),), 'o': IOError(OSError(2, 'No 
such file or directory'),), 'n': IOError(OSError(2, 'No such file or 
directory'),), 'p': IOError(OSError(2, 'No such file or directory'),), 's': 
IOError(OSError(2, 'No such file or directory'),), 'r': IOError(OSError(2, 'No 
such file or directory'),), 'u': IOError(OSError(2, 'No such file or 
directory'),), 't': IOError(OSError(2, 'No such file or directory'),), 'y': 
IOError(OSError(2, 'No such file or directory'),)}
 Traceback for above exception (most recent call last):
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/utils/retry.py";,>
 line 175, in wrapper
    return fun(*args, **kwargs)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/testing/test_utils.py";,>
 line 91, in delete_files
    FileSystems.delete(file_paths)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/io/filesystems.py";,>
 line 224, in delete
    return filesystem.delete(paths)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/io/localfilesystem.py";,>
 line 267, in delete
    raise BeamIOError("Delete operation failed", exceptions)

root: WARNING: Retry with exponential backoff: waiting for 12.6419520595 
seconds before retrying delete_files because we caught exception: BeamIOError: 
Delete operation failed with exceptions {'*': IOError(OSError(2, 'No such file 
or directory'),), '-': IOError(OSError(2, 'No such file or directory'),), '/': 
IOError(OSError(13, 'Permission denied'),), '1': IOError(OSError(2, 'No such 
file or directory'),), '0': IOError(OSError(2, 'No such file or directory'),), 
'3': IOError(OSError(2, 'No such file or directory'),), '2': IOError(OSError(2, 
'No such file or directory'),), '5': IOError(OSError(2, 'No such file or 
directory'),), ':': IOError(OSError(2, 'No such file or directory'),), 'a': 
IOError(OSError(2, 'No such file or directory'),), 'c': IOError(OSError(2, 'No 
such file or directory'),), 'e': IOError(OSError(2, 'No such file or 
directory'),), 'd': IOError(OSError(2, 'No such file or directory'),), 'g': 
IOError(OSError(2, 'No such file or directory'),), 'f': IOError(OSError(2, 'No 
such file or directory'),), 'i': IOError(OSError(2, 'No such file or 
directory'),), 'm': IOError(OSError(2, 'No such file or directory'),), 'l': 
IOError(OSError(2, 'No such file or directory'),), 'o': IOError(OSError(2, 'No 
such file or directory'),), 'n': IOError(OSError(2, 'No such file or 
directory'),), 'p': IOError(OSError(2, 'No such file or directory'),), 's': 
IOError(OSError(2, 'No such file or directory'),), 'r': IOError(OSError(2, 'No 
such file or directory'),), 'u': IOError(OSError(2, 'No such file or 
directory'),), 't': IOError(OSError(2, 'No such file or directory'),), 'y': 
IOError(OSError(2, 'No such file or directory'),)}
 Traceback for above exception (most recent call last):
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/utils/retry.py";,>
 line 175, in wrapper
    return fun(*args, **kwargs)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/testing/test_utils.py";,>
 line 91, in delete_files
    FileSystems.delete(file_paths)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/io/filesystems.py";,>
 line 224, in delete
    return filesystem.delete(paths)
  File 
"<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/sdks/python/apache_beam/io/localfilesystem.py";,>
 line 267, in delete
    raise BeamIOError("Delete operation failed", exceptions)

--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
Ran 2 tests in 369.498s

FAILED (errors=1)
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-08-08_13_48_31-14349531698595810259?project=apache-beam-testing
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2017-08-08_13_48_31-10685725663410924705?project=apache-beam-testing
Build step 'Execute shell' marked build as failure

Reply via email to