See 
<https://builds.apache.org/job/beam_PostCommit_Python_Verify/6654/display/redirect>

------------------------------------------
[...truncated 509.45 KB...]
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "monthly count/Combine.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s3"
        }, 
        "serialized_fn": "<string of 236 bytes>", 
        "user_name": "monthly count/Combine"
      }
    }, 
    {
      "kind": "ParallelDo", 
      "name": "s5", 
      "properties": {
        "display_data": [
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "<lambda>"
          }, 
          {
            "key": "fn", 
            "label": "Transform Function", 
            "namespace": "apache_beam.transforms.core.ParDo", 
            "shortValue": "CallableWrapperDoFn", 
            "type": "STRING", 
            "value": "apache_beam.transforms.core.CallableWrapperDoFn"
          }
        ], 
        "non_parallel_inputs": {}, 
        "output_info": [
          {
            "encoding": {
              "@type": "kind:windowed_value", 
              "component_encodings": [
                {
                  "@type": 
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                  "component_encodings": [
                    {
                      "@type": 
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                      "component_encodings": []
                    }, 
                    {
                      "@type": 
"FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/",
 
                      "component_encodings": []
                    }
                  ], 
                  "is_pair_like": true
                }, 
                {
                  "@type": "kind:global_window"
                }
              ], 
              "is_wrapper": true
            }, 
            "output_name": "out", 
            "user_name": "format.out"
          }
        ], 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s4"
        }, 
        "serialized_fn": "<string of 1028 bytes>", 
        "user_name": "format"
      }
    }, 
    {
      "kind": "ParallelWrite", 
      "name": "s6", 
      "properties": {
        "create_disposition": "CREATE_IF_NEEDED", 
        "dataset": "BigQueryTornadoesIT", 
        "display_data": [], 
        "encoding": {
          "@type": "kind:windowed_value", 
          "component_encodings": [
            {
              "@type": 
"RowAsDictJsonCoder$eNprYEpOLEhMzkiNT0pNzNXLzNdLTy7QS8pMLyxNLarkCsovdyx2yUwu8SrOz3POT0kt4ipk0GwsZKwtZErSAwCu1BVY",
 
              "component_encodings": []
            }, 
            {
              "@type": "kind:global_window"
            }
          ], 
          "is_wrapper": true
        }, 
        "format": "bigquery", 
        "parallel_input": {
          "@type": "OutputReference", 
          "output_name": "out", 
          "step_name": "s5"
        }, 
        "schema": "{\"fields\": [{\"type\": \"INTEGER\", \"name\": \"month\", 
\"mode\": \"NULLABLE\"}, {\"type\": \"INTEGER\", \"name\": \"tornado_count\", 
\"mode\": \"NULLABLE\"}]}", 
        "table": "monthly_tornadoes_1543169064064", 
        "user_name": "Write/WriteToBigQuery/NativeWrite", 
        "write_disposition": "WRITE_TRUNCATE"
      }
    }
  ], 
  "type": "JOB_TYPE_BATCH"
}
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
oauth2client.transport: INFO: Attempting refresh to obtain initial access_token
root: INFO: Create job: <Job
 createTime: u'2018-11-25T18:04:38.287442Z'
 currentStateTime: u'1970-01-01T00:00:00Z'
 id: u'2018-11-25_10_04_37-7849068734969825661'
 location: u'us-central1'
 name: u'beamapp-jenkins-1125180424-116725'
 projectId: u'apache-beam-testing'
 stageStates: []
 startTime: u'2018-11-25T18:04:38.287442Z'
 steps: []
 tempFiles: []
 type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)>
root: INFO: Created job with id: [2018-11-25_10_04_37-7849068734969825661]
root: INFO: To access the Dataflow monitoring console, please navigate to 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_04_37-7849068734969825661?project=apache-beam-testing
root: INFO: Job 2018-11-25_10_04_37-7849068734969825661 is in state 
JOB_STATE_RUNNING
root: INFO: 2018-11-25T18:04:37.385Z: JOB_MESSAGE_DETAILED: Autoscaling is 
enabled for job 2018-11-25_10_04_37-7849068734969825661. The number of workers 
will be between 1 and 1000.
root: INFO: 2018-11-25T18:04:37.435Z: JOB_MESSAGE_DETAILED: Autoscaling was 
automatically enabled for job 2018-11-25_10_04_37-7849068734969825661.
root: INFO: 2018-11-25T18:04:40.165Z: JOB_MESSAGE_DETAILED: Checking 
permissions granted to controller Service Account.
root: INFO: 2018-11-25T18:04:41.701Z: JOB_MESSAGE_BASIC: Worker configuration: 
n1-standard-1 in us-central1-b.
root: INFO: 2018-11-25T18:04:42.250Z: JOB_MESSAGE_DETAILED: Expanding 
CoGroupByKey operations into optimizable parts.
root: INFO: 2018-11-25T18:04:42.309Z: JOB_MESSAGE_DETAILED: Expanding 
GroupByKey operations into optimizable parts.
root: INFO: 2018-11-25T18:04:42.358Z: JOB_MESSAGE_DETAILED: Lifting 
ValueCombiningMappingFns into MergeBucketsMappingFns
root: INFO: 2018-11-25T18:04:42.459Z: JOB_MESSAGE_DEBUG: Annotating graph with 
Autotuner information.
root: INFO: 2018-11-25T18:04:42.584Z: JOB_MESSAGE_DETAILED: Fusing adjacent 
ParDo, Read, Write, and Flatten operations
root: INFO: 2018-11-25T18:04:42.625Z: JOB_MESSAGE_DETAILED: Fusing consumer 
months with tornadoes into read
root: INFO: 2018-11-25T18:04:42.672Z: JOB_MESSAGE_DETAILED: Fusing consumer 
monthly count/GroupByKey/Reify into monthly count/GroupByKey+monthly 
count/Combine/Partial
root: INFO: 2018-11-25T18:04:42.722Z: JOB_MESSAGE_DETAILED: Fusing consumer 
format into monthly count/Combine/Extract
root: INFO: 2018-11-25T18:04:42.774Z: JOB_MESSAGE_DETAILED: Fusing consumer 
monthly count/Combine/Extract into monthly count/Combine
root: INFO: 2018-11-25T18:04:42.824Z: JOB_MESSAGE_DETAILED: Fusing consumer 
Write/WriteToBigQuery/NativeWrite into format
root: INFO: 2018-11-25T18:04:42.881Z: JOB_MESSAGE_DETAILED: Fusing consumer 
monthly count/Combine into monthly count/GroupByKey/Read
root: INFO: 2018-11-25T18:04:42.928Z: JOB_MESSAGE_DETAILED: Fusing consumer 
monthly count/GroupByKey+monthly count/Combine/Partial into months with 
tornadoes
root: INFO: 2018-11-25T18:04:42.976Z: JOB_MESSAGE_DETAILED: Fusing consumer 
monthly count/GroupByKey/Write into monthly count/GroupByKey/Reify
root: INFO: 2018-11-25T18:04:43.024Z: JOB_MESSAGE_DEBUG: Workflow config is 
missing a default resource spec.
root: INFO: 2018-11-25T18:04:43.074Z: JOB_MESSAGE_DEBUG: Adding StepResource 
setup and teardown to workflow graph.
root: INFO: 2018-11-25T18:04:43.108Z: JOB_MESSAGE_DEBUG: Adding workflow start 
and stop steps.
root: INFO: 2018-11-25T18:04:43.148Z: JOB_MESSAGE_DEBUG: Assigning stage ids.
root: INFO: 2018-11-25T18:04:43.354Z: JOB_MESSAGE_DEBUG: Executing wait step 
start22
root: INFO: 2018-11-25T18:04:43.451Z: JOB_MESSAGE_BASIC: Executing operation 
monthly count/GroupByKey/Create
root: INFO: 2018-11-25T18:04:43.508Z: JOB_MESSAGE_DEBUG: Starting worker pool 
setup.
root: INFO: 2018-11-25T18:04:43.557Z: JOB_MESSAGE_BASIC: Starting 1 workers in 
us-central1-b...
root: INFO: 2018-11-25T18:04:43.656Z: JOB_MESSAGE_DEBUG: Value "monthly 
count/GroupByKey/Session" materialized.
root: INFO: 2018-11-25T18:04:43.751Z: JOB_MESSAGE_BASIC: Executing operation 
read+months with tornadoes+monthly count/GroupByKey+monthly 
count/Combine/Partial+monthly count/GroupByKey/Reify+monthly 
count/GroupByKey/Write
root: INFO: 2018-11-25T18:04:44.242Z: JOB_MESSAGE_BASIC: BigQuery export job 
"dataflow_job_13973052911007792222" started. You can check its status with the 
bq tool: "bq show -j --project_id=clouddataflow-readonly 
dataflow_job_13973052911007792222".
root: INFO: 2018-11-25T18:04:55.580Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised 
the number of workers to 0 based on the rate of progress in the currently 
running step(s).
root: INFO: 2018-11-25T18:05:14.682Z: JOB_MESSAGE_DETAILED: BigQuery export job 
progress: "dataflow_job_13973052911007792222" observed total of 1 exported 
files thus far.
root: INFO: 2018-11-25T18:05:14.734Z: JOB_MESSAGE_BASIC: BigQuery export job 
finished: "dataflow_job_13973052911007792222"
root: INFO: 2018-11-25T18:05:35.487Z: JOB_MESSAGE_DETAILED: Autoscaling: Raised 
the number of workers to 1 based on the rate of progress in the currently 
running step(s).
root: INFO: 2018-11-25T18:05:35.524Z: JOB_MESSAGE_DETAILED: Autoscaling: Would 
further reduce the number of workers but reached the minimum number allowed for 
the job.
root: INFO: 2018-11-25T18:07:27.532Z: JOB_MESSAGE_DETAILED: Workers have 
started successfully.
root: INFO: 2018-11-25T18:07:27.571Z: JOB_MESSAGE_DETAILED: Workers have 
started successfully.
root: INFO: 2018-11-25T18:07:28.682Z: JOB_MESSAGE_ERROR: A setup error was 
detected in beamapp-jenkins-112518042-11251004-4yem-harness-zk3v. Please refer 
to the worker-startup log for detailed information.
root: INFO: 2018-11-25T18:07:32.184Z: JOB_MESSAGE_ERROR: A setup error was 
detected in beamapp-jenkins-112518042-11251004-4yem-harness-zk3v. Please refer 
to the worker-startup log for detailed information.
root: INFO: 2018-11-25T18:07:48.003Z: JOB_MESSAGE_ERROR: A setup error was 
detected in beamapp-jenkins-112518042-11251004-4yem-harness-zk3v. Please refer 
to the worker-startup log for detailed information.
root: INFO: 2018-11-25T18:08:18.985Z: JOB_MESSAGE_ERROR: A setup error was 
detected in beamapp-jenkins-112518042-11251004-4yem-harness-zk3v. Please refer 
to the worker-startup log for detailed information.
root: INFO: 2018-11-25T18:08:19.149Z: JOB_MESSAGE_DEBUG: Executing failure step 
failure21
root: INFO: 2018-11-25T18:08:19.195Z: JOB_MESSAGE_ERROR: Workflow failed. 
Causes: S02:read+months with tornadoes+monthly count/GroupByKey+monthly 
count/Combine/Partial+monthly count/GroupByKey/Reify+monthly 
count/GroupByKey/Write failed., A work item was attempted 4 times without 
success. Each time the worker eventually lost contact with the service. The 
work item was attempted on: 
  beamapp-jenkins-112518042-11251004-4yem-harness-zk3v,
  beamapp-jenkins-112518042-11251004-4yem-harness-zk3v,
  beamapp-jenkins-112518042-11251004-4yem-harness-zk3v,
  beamapp-jenkins-112518042-11251004-4yem-harness-zk3v
root: INFO: 2018-11-25T18:08:19.328Z: JOB_MESSAGE_DETAILED: Cleaning up.
root: INFO: 2018-11-25T18:08:19.404Z: JOB_MESSAGE_DEBUG: Starting worker pool 
teardown.
root: INFO: 2018-11-25T18:08:19.450Z: JOB_MESSAGE_BASIC: Stopping worker pool...
root: INFO: 2018-11-25T18:10:11.019Z: JOB_MESSAGE_DETAILED: Autoscaling: 
Resized worker pool from 1 to 0.
root: INFO: 2018-11-25T18:10:11.066Z: JOB_MESSAGE_DETAILED: Autoscaling: Would 
further reduce the number of workers but reached the minimum number allowed for 
the job.
root: INFO: 2018-11-25T18:10:11.127Z: JOB_MESSAGE_BASIC: Worker pool stopped.
root: INFO: 2018-11-25T18:10:11.169Z: JOB_MESSAGE_DEBUG: Tearing down pending 
resources...
root: INFO: Job 2018-11-25_10_04_37-7849068734969825661 is in state 
JOB_STATE_FAILED
root: INFO: Clean up a BigQuery table with project: apache-beam-testing, 
dataset: BigQueryTornadoesIT, table: monthly_tornadoes_1543169064064.
google.auth.transport._http_client: DEBUG: Making request: GET 
http://169.254.169.254
google.auth.transport._http_client: DEBUG: Making request: GET 
http://metadata.google.internal/computeMetadata/v1/project/project-id
urllib3.util.retry: DEBUG: Converted retries value: 3 -> Retry(total=3, 
connect=None, read=None, redirect=None, status=None)
google.auth.transport.requests: DEBUG: Making request: GET 
http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/default/?recursive=true
urllib3.connectionpool: DEBUG: Starting new HTTP connection (1): 
metadata.google.internal:80
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET 
/computeMetadata/v1/instance/service-accounts/default/?recursive=true HTTP/1.1" 
200 144
google.auth.transport.requests: DEBUG: Making request: GET 
http://metadata.google.internal/computeMetadata/v1/instance/service-accounts/[email protected]/token
urllib3.connectionpool: DEBUG: http://metadata.google.internal:80 "GET 
/computeMetadata/v1/instance/service-accounts/[email protected]/token
 HTTP/1.1" 200 173
urllib3.connectionpool: DEBUG: Starting new HTTPS connection (1): 
www.googleapis.com:443
urllib3.connectionpool: DEBUG: https://www.googleapis.com:443 "DELETE 
/bigquery/v2/projects/apache-beam-testing/datasets/BigQueryTornadoesIT/tables/monthly_tornadoes_1543169064064
 HTTP/1.1" 404 None
--------------------- >> end captured logging << ---------------------

----------------------------------------------------------------------
XML: 
<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/src/sdks/python/nosetests.xml>
----------------------------------------------------------------------
Ran 18 tests in 2683.738s

FAILED (errors=2)
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_04_40-10619501220103224677?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_04_38-7927878155071309540?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_04_40-7161695053630940031?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_16_59-17704416867251687273?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_24_24-5742703706997750185?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_30_20-11186609359868357226?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_04_37-12176387962678726745?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_04_37-3706068116252383299?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_11_28-16762436658842793616?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_18_28-4550887450801010421?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_04_37-7849068734969825661?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_10_40-87527565036137448?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_18_30-1681529530696514830?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_24_25-15269715405729410454?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_30_35-9409211484788036621?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_37_19-9802234884212112180?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_43_29-12885495800942882741?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_04_38-3294149779166662752?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_13_07-5354828996564125796?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_04_38-37398875133442728?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_11_19-10545049789630682067?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_18_38-6784660283229927990?project=apache-beam-testing.
Found: 
https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-11-25_10_24_50-8518900843500055600?project=apache-beam-testing.

> Task :beam-sdks-python:postCommitIT FAILED
:beam-sdks-python:postCommitIT (Thread[Task worker for ':',5,main]) completed. 
Took 44 mins 44.576 secs.

FAILURE: Build failed with an exception.

* Where:
Build file 
'<https://builds.apache.org/job/beam_PostCommit_Python_Verify/ws/src/sdks/python/build.gradle'>
 line: 319

* What went wrong:
Execution failed for task ':beam-sdks-python:postCommitIT'.
> Process 'command 'sh'' finished with non-zero exit value 1

* Try:
Run with --stacktrace option to get the stack trace. Run with --debug option to 
get more log output. Run with --scan to get full insights.

* Get more help at https://help.gradle.org

BUILD FAILED in 48m 54s
6 actionable tasks: 6 executed

Publishing build scan...
https://gradle.com/s/y5rocr35ubw2i

Build step 'Invoke Gradle script' changed build result to FAILURE
Build step 'Invoke Gradle script' marked build as failure

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to