See <https://builds.apache.org/job/beam_PerformanceTests_Python/764/display/redirect?page=changes>
Changes: [iemejia] [BEAM-3404] Update KinesisIO to use AWS SDK 1.11.255 and KCL 1.8.8 [iemejia] [BEAM-3404] Fix and enable KinesisReaderIT ------------------------------------------ [...truncated 91.24 KB...] { "@type": "kind:windowed_value", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] }, { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] } ], "is_pair_like": true }, { "@type": "kind:global_window" } ], "is_wrapper": true } ] }, "output_name": "out", "user_name": "write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0).output" } ], "parallel_input": { "@type": "OutputReference", "output_name": "out", "step_name": "s8" }, "user_name": "write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(InitializeWrite.out.0)" } }, { "kind": "CollectionToSingleton", "name": "SideInput-s16", "properties": { "output_info": [ { "encoding": { "@type": "kind:windowed_value", "component_encodings": [ { "@type": "kind:windowed_value", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] }, { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] } ], "is_pair_like": true }, { "@type": "kind:global_window" } ], "is_wrapper": true } ] }, "output_name": "out", "user_name": "write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0).output" } ], "parallel_input": { "@type": "OutputReference", "output_name": "out", "step_name": "s14" }, "user_name": "write/Write/WriteImpl/FinalizeWrite/_UnpickledSideInput(Extract.out.0)" } }, { "kind": "ParallelDo", "name": "s17", "properties": { "display_data": [ { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.CallableWrapperDoFn", "type": "STRING", "value": "_finalize_write" }, { "key": "fn", "label": "Transform Function", "namespace": "apache_beam.transforms.core.ParDo", "shortValue": "CallableWrapperDoFn", "type": "STRING", "value": "apache_beam.transforms.core.CallableWrapperDoFn" } ], "non_parallel_inputs": { "SideInput-s15": { "@type": "OutputReference", "output_name": "out", "step_name": "SideInput-s15" }, "SideInput-s16": { "@type": "OutputReference", "output_name": "out", "step_name": "SideInput-s16" } }, "output_info": [ { "encoding": { "@type": "kind:windowed_value", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [ { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] }, { "@type": "FastPrimitivesCoder$eNprYEpOLEhMzkiNT0pNzNVLzk9JLSqGUlxuicUlAUWZuZklmWWpxc4gQa5CBs3GQsbaQqZQ/vi0xJycpMTk7Hiw+kJmPEYFZCZn56RCjWABGsFaW8iWVJykBwDlGS3/", "component_encodings": [] } ], "is_pair_like": true }, { "@type": "kind:global_window" } ], "is_wrapper": true }, "output_name": "out", "user_name": "write/Write/WriteImpl/FinalizeWrite.out" } ], "parallel_input": { "@type": "OutputReference", "output_name": "out", "step_name": "s7" }, "serialized_fn": "<string of 2420 bytes>", "user_name": "write/Write/WriteImpl/FinalizeWrite/FinalizeWrite" } } ], "type": "JOB_TYPE_BATCH" } root: INFO: Create job: <Job createTime: u'2018-01-08T07:21:54.931317Z' currentStateTime: u'1970-01-01T00:00:00Z' id: u'2018-01-07_23_21_54-18206030603986655304' location: u'us-central1' name: u'beamapp-jenkins-0108072152-616242' projectId: u'apache-beam-testing' stageStates: [] steps: [] tempFiles: [] type: TypeValueValuesEnum(JOB_TYPE_BATCH, 1)> root: INFO: Created job with id: [2018-01-07_23_21_54-18206030603986655304] root: INFO: To access the Dataflow monitoring console, please navigate to https://console.cloud.google.com/dataflow/jobsDetail/locations/us-central1/jobs/2018-01-07_23_21_54-18206030603986655304?project=apache-beam-testing root: INFO: Job 2018-01-07_23_21_54-18206030603986655304 is in state JOB_STATE_PENDING root: INFO: 2018-01-08T07:21:54.268Z: JOB_MESSAGE_DETAILED: (fca8d05fe883dfcc): Autoscaling is enabled for job 2018-01-07_23_21_54-18206030603986655304. The number of workers will be between 1 and 15. root: INFO: 2018-01-08T07:21:54.283Z: JOB_MESSAGE_DETAILED: (fca8d05fe883d1ad): Autoscaling was automatically enabled for job 2018-01-07_23_21_54-18206030603986655304. root: INFO: 2018-01-08T07:21:56.744Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed90b): Checking required Cloud APIs are enabled. root: INFO: 2018-01-08T07:21:57.522Z: JOB_MESSAGE_DETAILED: (3557ba8a2bedf15): Expanding CoGroupByKey operations into optimizable parts. root: INFO: 2018-01-08T07:21:57.554Z: JOB_MESSAGE_DEBUG: (3557ba8a2bedbce): Combiner lifting skipped for step write/Write/WriteImpl/GroupByKey: GroupByKey not followed by a combiner. root: INFO: 2018-01-08T07:21:57.590Z: JOB_MESSAGE_DEBUG: (3557ba8a2bedef4): Combiner lifting skipped for step group: GroupByKey not followed by a combiner. root: INFO: 2018-01-08T07:21:57.614Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed21a): Expanding GroupByKey operations into optimizable parts. root: INFO: 2018-01-08T07:21:57.642Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed540): Lifting ValueCombiningMappingFns into MergeBucketsMappingFns root: INFO: 2018-01-08T07:21:57.671Z: JOB_MESSAGE_DEBUG: (3557ba8a2bedeb2): Annotating graph with Autotuner information. root: INFO: 2018-01-08T07:21:57.719Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed4fe): Fusing adjacent ParDo, Read, Write, and Flatten operations root: INFO: 2018-01-08T07:21:57.746Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed824): Fusing consumer split into read/Read root: INFO: 2018-01-08T07:21:57.773Z: JOB_MESSAGE_DETAILED: (3557ba8a2bedb4a): Fusing consumer group/Write into group/Reify root: INFO: 2018-01-08T07:21:57.804Z: JOB_MESSAGE_DETAILED: (3557ba8a2bede70): Fusing consumer group/GroupByWindow into group/Read root: INFO: 2018-01-08T07:21:57.833Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed196): Fusing consumer write/Write/WriteImpl/GroupByKey/GroupByWindow into write/Write/WriteImpl/GroupByKey/Read root: INFO: 2018-01-08T07:21:57.860Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed4bc): Fusing consumer write/Write/WriteImpl/GroupByKey/Write into write/Write/WriteImpl/GroupByKey/Reify root: INFO: 2018-01-08T07:21:57.886Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed7e2): Fusing consumer write/Write/WriteImpl/WindowInto(WindowIntoFn) into write/Write/WriteImpl/Pair root: INFO: 2018-01-08T07:21:57.910Z: JOB_MESSAGE_DETAILED: (3557ba8a2bedb08): Fusing consumer write/Write/WriteImpl/GroupByKey/Reify into write/Write/WriteImpl/WindowInto(WindowIntoFn) root: INFO: 2018-01-08T07:21:57.934Z: JOB_MESSAGE_DETAILED: (3557ba8a2bede2e): Fusing consumer pair_with_one into split root: INFO: 2018-01-08T07:21:57.965Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed154): Fusing consumer group/Reify into pair_with_one root: INFO: 2018-01-08T07:21:57.988Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed47a): Fusing consumer write/Write/WriteImpl/WriteBundles/WriteBundles into format root: INFO: 2018-01-08T07:21:58.013Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed7a0): Fusing consumer write/Write/WriteImpl/Pair into write/Write/WriteImpl/WriteBundles/WriteBundles root: INFO: 2018-01-08T07:21:58.044Z: JOB_MESSAGE_DETAILED: (3557ba8a2bedac6): Fusing consumer format into count root: INFO: 2018-01-08T07:21:58.069Z: JOB_MESSAGE_DETAILED: (3557ba8a2beddec): Fusing consumer write/Write/WriteImpl/Extract into write/Write/WriteImpl/GroupByKey/GroupByWindow root: INFO: 2018-01-08T07:21:58.090Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed112): Fusing consumer count into group/GroupByWindow root: INFO: 2018-01-08T07:21:58.112Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed438): Fusing consumer write/Write/WriteImpl/InitializeWrite into write/Write/WriteImpl/DoOnce/Read root: INFO: 2018-01-08T07:21:58.138Z: JOB_MESSAGE_DEBUG: (3557ba8a2bed75e): Workflow config is missing a default resource spec. root: INFO: 2018-01-08T07:21:58.164Z: JOB_MESSAGE_DEBUG: (3557ba8a2beda84): Adding StepResource setup and teardown to workflow graph. root: INFO: 2018-01-08T07:21:58.186Z: JOB_MESSAGE_DEBUG: (3557ba8a2beddaa): Adding workflow start and stop steps. root: INFO: 2018-01-08T07:21:58.215Z: JOB_MESSAGE_DEBUG: (3557ba8a2bed0d0): Assigning stage ids. root: INFO: 2018-01-08T07:21:58.365Z: JOB_MESSAGE_DEBUG: (bf7b7f7391c0500): Executing wait step start25 root: INFO: 2018-01-08T07:21:58.428Z: JOB_MESSAGE_BASIC: (8c36565dbbe9a7e6): Executing operation write/Write/WriteImpl/DoOnce/Read+write/Write/WriteImpl/InitializeWrite root: INFO: 2018-01-08T07:21:58.466Z: JOB_MESSAGE_BASIC: (bf7b7f7391c001a): Executing operation group/Create root: INFO: 2018-01-08T07:21:58.479Z: JOB_MESSAGE_DEBUG: (cd7c291390209e30): Starting worker pool setup. root: INFO: 2018-01-08T07:21:58.502Z: JOB_MESSAGE_BASIC: (cd7c2913902095ee): Starting 1 workers in us-central1-f... root: INFO: 2018-01-08T07:21:58.592Z: JOB_MESSAGE_DEBUG: (9fa83853f30cd4f1): Value "group/Session" materialized. root: INFO: 2018-01-08T07:21:58.647Z: JOB_MESSAGE_BASIC: (bf7b7f7391c07f0): Executing operation read/Read+split+pair_with_one+group/Reify+group/Write root: INFO: Job 2018-01-07_23_21_54-18206030603986655304 is in state JOB_STATE_RUNNING root: INFO: 2018-01-08T07:22:05.184Z: JOB_MESSAGE_DETAILED: (7947adbcfc2fc3f8): Autoscaling: Raised the number of workers to 0 based on the rate of progress in the currently running step(s). root: INFO: 2018-01-08T07:22:40.074Z: JOB_MESSAGE_ERROR: (7947adbcfc2fc1cb): Startup of the worker pool in zone us-central1-f failed to bring up any of the desired 1 workers. QUOTA_EXCEEDED: Quota 'DISKS_TOTAL_GB' exceeded. Limit: 21000.0 in region us-central1. root: INFO: 2018-01-08T07:22:40.110Z: JOB_MESSAGE_ERROR: (7947adbcfc2fcd01): Workflow failed. root: INFO: 2018-01-08T07:22:40.255Z: JOB_MESSAGE_DETAILED: (3557ba8a2bed6fb): Cleaning up. root: INFO: 2018-01-08T07:22:40.315Z: JOB_MESSAGE_DEBUG: (3557ba8a2bedd47): Starting worker pool teardown. root: INFO: 2018-01-08T07:22:40.331Z: JOB_MESSAGE_BASIC: (3557ba8a2bed06d): Stopping worker pool... root: INFO: 2018-01-08T07:22:48.924Z: JOB_MESSAGE_DEBUG: (3557ba8a2bed02b): Tearing down pending resources... root: INFO: Job 2018-01-07_23_21_54-18206030603986655304 is in state JOB_STATE_FAILED --------------------- >> end captured logging << --------------------- ---------------------------------------------------------------------- Ran 2 tests in 139.839s FAILED (errors=2) 2018-01-08 07:23:00,158 a267542d MainThread beam_integration_benchmark(1/1) ERROR Error during benchmark beam_integration_benchmark Traceback (most recent call last): File "<https://builds.apache.org/job/beam_PerformanceTests_Python/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",> line 601, in RunBenchmark DoRunPhase(spec, collector, detailed_timer) File "<https://builds.apache.org/job/beam_PerformanceTests_Python/ws/PerfKitBenchmarker/perfkitbenchmarker/pkb.py",> line 504, in DoRunPhase samples = spec.BenchmarkRun(spec) File "<https://builds.apache.org/job/beam_PerformanceTests_Python/ws/PerfKitBenchmarker/perfkitbenchmarker/linux_benchmarks/beam_integration_benchmark.py",> line 159, in Run job_type=job_type) File "<https://builds.apache.org/job/beam_PerformanceTests_Python/ws/PerfKitBenchmarker/perfkitbenchmarker/providers/gcp/gcp_dpb_dataflow.py",> line 90, in SubmitJob assert retcode == 0, "Integration Test Failed." AssertionError: Integration Test Failed. 2018-01-08 07:23:00,159 a267542d MainThread beam_integration_benchmark(1/1) INFO Cleaning up benchmark beam_integration_benchmark 2018-01-08 07:23:00,161 a267542d MainThread beam_integration_benchmark(1/1) ERROR Benchmark 1/1 beam_integration_benchmark (UID: beam_integration_benchmark0) failed. Execution will continue. 2018-01-08 07:23:00,200 a267542d MainThread INFO Benchmark run statuses: --------------------------------------------------------------- Name UID Status --------------------------------------------------------------- beam_integration_benchmark beam_integration_benchmark0 FAILED --------------------------------------------------------------- Success rate: 0.00% (0/1) 2018-01-08 07:23:00,201 a267542d MainThread INFO Complete logs can be found at: /tmp/perfkitbenchmarker/runs/a267542d/pkb.log 2018-01-08 07:23:00,201 a267542d MainThread INFO Completion statuses can be found at: /tmp/perfkitbenchmarker/runs/a267542d/completion_statuses.json Build step 'Execute shell' marked build as failure Not sending mail to unregistered user git...@alasdairhodge.co.uk Not sending mail to unregistered user kirpic...@google.com Not sending mail to unregistered user mari...@mariagh.svl.corp.google.com