See <https://ci-beam.apache.org/job/beam_PostCommit_Python38/1046/display/redirect?page=changes>
Changes: [rohde.samuel] Update Dataflow V1Beta3 to newest version to add the [rohde.samuel] Add the 'enable_hot_key_logging' PipelineOption and plumb it to the [lpost] [BEAM-12079] Deterministic coding enforcement causes [noreply] fix: variables names in test-stream.md [Udi Meiri] Update dev Dataflow containers to latest version. [Kyle Weaver] [BEAM-12033] Create ZetaSqlException, which contains a GRPC Status code. [noreply] Merge pull request #14372 from [BEAM-12014] Add BIGNUMERIC support in ------------------------------------------ [...truncated 40.14 MB...] File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_test.py",> line 1247, in <lambda> table_map: table_map.get(dest, None), TypeError: unhashable type: 'TableReference' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.8/site-packages/dataflow_worker/batchworker.py", line 649, in do_work work_executor.execute() File "/usr/local/lib/python3.8/site-packages/dataflow_worker/executor.py", line 179, in execute op.start() File "dataflow_worker/shuffle_operations.py", line 63, in dataflow_worker.shuffle_operations.GroupedShuffleReadOperation.start File "dataflow_worker/shuffle_operations.py", line 64, in dataflow_worker.shuffle_operations.GroupedShuffleReadOperation.start File "dataflow_worker/shuffle_operations.py", line 79, in dataflow_worker.shuffle_operations.GroupedShuffleReadOperation.start File "dataflow_worker/shuffle_operations.py", line 80, in dataflow_worker.shuffle_operations.GroupedShuffleReadOperation.start File "dataflow_worker/shuffle_operations.py", line 84, in dataflow_worker.shuffle_operations.GroupedShuffleReadOperation.start File "apache_beam/runners/worker/operations.py", line 358, in apache_beam.runners.worker.operations.Operation.output File "apache_beam/runners/worker/operations.py", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive File "dataflow_worker/shuffle_operations.py", line 261, in dataflow_worker.shuffle_operations.BatchGroupAlsoByWindowsOperation.process File "dataflow_worker/shuffle_operations.py", line 268, in dataflow_worker.shuffle_operations.BatchGroupAlsoByWindowsOperation.process File "apache_beam/runners/worker/operations.py", line 358, in apache_beam.runners.worker.operations.Operation.output File "apache_beam/runners/worker/operations.py", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 717, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/worker/operations.py", line 718, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/common.py", line 1235, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 1300, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam/runners/common.py", line 1233, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam/runners/common.py", line 1395, in apache_beam.runners.common._OutputProcessor.process_outputs File "apache_beam/runners/worker/operations.py", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 717, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/worker/operations.py", line 718, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/common.py", line 1235, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 1300, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam/runners/common.py", line 1233, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam/runners/common.py", line 1395, in apache_beam.runners.common._OutputProcessor.process_outputs File "apache_beam/runners/worker/operations.py", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 717, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/worker/operations.py", line 718, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/common.py", line 1235, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 1300, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam/runners/common.py", line 1233, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam/runners/common.py", line 1395, in apache_beam.runners.common._OutputProcessor.process_outputs File "apache_beam/runners/worker/operations.py", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 717, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/worker/operations.py", line 718, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/common.py", line 1235, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 1315, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "/usr/local/lib/python3.8/site-packages/future/utils/__init__.py", line 446, in raise_with_traceback raise exc.with_traceback(traceback) File "apache_beam/runners/common.py", line 1233, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 762, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam/runners/common.py", line 887, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "/usr/local/lib/python3.8/site-packages/apache_beam/io/gcp/bigquery.py", line 1249, in process schema = self.schema(destination, *schema_side_inputs) File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_test.py",> line 1247, in <lambda> table_map: table_map.get(dest, None), TypeError: unhashable type: 'TableReference' [while running 'WriteWithMultipleDests/_StreamToBigQuery/StreamInsertRows/ParDo(BigQueryWriteFn)/ParDo(BigQueryWriteFn)'] apache_beam.runners.dataflow.dataflow_runner: INFO: 2021-04-02T00:51:38.664Z: JOB_MESSAGE_ERROR: Traceback (most recent call last): File "apache_beam/runners/common.py", line 1233, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 762, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam/runners/common.py", line 887, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "/usr/local/lib/python3.8/site-packages/apache_beam/io/gcp/bigquery.py", line 1249, in process schema = self.schema(destination, *schema_side_inputs) File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_test.py",> line 1247, in <lambda> table_map: table_map.get(dest, None), TypeError: unhashable type: 'TableReference' During handling of the above exception, another exception occurred: Traceback (most recent call last): File "/usr/local/lib/python3.8/site-packages/dataflow_worker/batchworker.py", line 649, in do_work work_executor.execute() File "/usr/local/lib/python3.8/site-packages/dataflow_worker/executor.py", line 179, in execute op.start() File "dataflow_worker/shuffle_operations.py", line 63, in dataflow_worker.shuffle_operations.GroupedShuffleReadOperation.start File "dataflow_worker/shuffle_operations.py", line 64, in dataflow_worker.shuffle_operations.GroupedShuffleReadOperation.start File "dataflow_worker/shuffle_operations.py", line 79, in dataflow_worker.shuffle_operations.GroupedShuffleReadOperation.start File "dataflow_worker/shuffle_operations.py", line 80, in dataflow_worker.shuffle_operations.GroupedShuffleReadOperation.start File "dataflow_worker/shuffle_operations.py", line 84, in dataflow_worker.shuffle_operations.GroupedShuffleReadOperation.start File "apache_beam/runners/worker/operations.py", line 358, in apache_beam.runners.worker.operations.Operation.output File "apache_beam/runners/worker/operations.py", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive File "dataflow_worker/shuffle_operations.py", line 261, in dataflow_worker.shuffle_operations.BatchGroupAlsoByWindowsOperation.process File "dataflow_worker/shuffle_operations.py", line 268, in dataflow_worker.shuffle_operations.BatchGroupAlsoByWindowsOperation.process File "apache_beam/runners/worker/operations.py", line 358, in apache_beam.runners.worker.operations.Operation.output File "apache_beam/runners/worker/operations.py", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 717, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/worker/operations.py", line 718, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/common.py", line 1235, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 1300, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam/runners/common.py", line 1233, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam/runners/common.py", line 1395, in apache_beam.runners.common._OutputProcessor.process_outputs File "apache_beam/runners/worker/operations.py", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 717, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/worker/operations.py", line 718, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/common.py", line 1235, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 1300, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam/runners/common.py", line 1233, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam/runners/common.py", line 1395, in apache_beam.runners.common._OutputProcessor.process_outputs File "apache_beam/runners/worker/operations.py", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 717, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/worker/operations.py", line 718, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/common.py", line 1235, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 1300, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "apache_beam/runners/common.py", line 1233, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 581, in apache_beam.runners.common.SimpleInvoker.invoke_process File "apache_beam/runners/common.py", line 1395, in apache_beam.runners.common._OutputProcessor.process_outputs File "apache_beam/runners/worker/operations.py", line 220, in apache_beam.runners.worker.operations.SingletonConsumerSet.receive File "apache_beam/runners/worker/operations.py", line 717, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/worker/operations.py", line 718, in apache_beam.runners.worker.operations.DoOperation.process File "apache_beam/runners/common.py", line 1235, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 1315, in apache_beam.runners.common.DoFnRunner._reraise_augmented File "/usr/local/lib/python3.8/site-packages/future/utils/__init__.py", line 446, in raise_with_traceback raise exc.with_traceback(traceback) File "apache_beam/runners/common.py", line 1233, in apache_beam.runners.common.DoFnRunner.process File "apache_beam/runners/common.py", line 762, in apache_beam.runners.common.PerWindowInvoker.invoke_process File "apache_beam/runners/common.py", line 887, in apache_beam.runners.common.PerWindowInvoker._invoke_process_per_window File "/usr/local/lib/python3.8/site-packages/apache_beam/io/gcp/bigquery.py", line 1249, in process schema = self.schema(destination, *schema_side_inputs) File "<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/apache_beam/io/gcp/bigquery_test.py",> line 1247, in <lambda> table_map: table_map.get(dest, None), TypeError: unhashable type: 'TableReference' [while running 'WriteWithMultipleDests/_StreamToBigQuery/StreamInsertRows/ParDo(BigQueryWriteFn)/ParDo(BigQueryWriteFn)'] apache_beam.runners.dataflow.dataflow_runner: INFO: 2021-04-02T00:51:39.391Z: JOB_MESSAGE_BASIC: Finished operation WriteWithMultipleDests/_StreamToBigQuery/CommitInsertIds/GroupByKey/Read+WriteWithMultipleDests/_StreamToBigQuery/CommitInsertIds/GroupByKey/GroupByWindow+WriteWithMultipleDests/_StreamToBigQuery/CommitInsertIds/FlatMap(restore_timestamps)+WriteWithMultipleDests/_StreamToBigQuery/DropShard+WriteWithMultipleDests/_StreamToBigQuery/FromHashableTableRef+WriteWithMultipleDests/_StreamToBigQuery/StreamInsertRows/ParDo(BigQueryWriteFn)/ParDo(BigQueryWriteFn)+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write apache_beam.runners.dataflow.dataflow_runner: INFO: 2021-04-02T00:51:39.889Z: JOB_MESSAGE_DEBUG: Executing failure step failure48 apache_beam.runners.dataflow.dataflow_runner: INFO: 2021-04-02T00:51:40.077Z: JOB_MESSAGE_ERROR: Workflow failed. Causes: S11:WriteWithMultipleDests/_StreamToBigQuery/CommitInsertIds/GroupByKey/Read+WriteWithMultipleDests/_StreamToBigQuery/CommitInsertIds/GroupByKey/GroupByWindow+WriteWithMultipleDests/_StreamToBigQuery/CommitInsertIds/FlatMap(restore_timestamps)+WriteWithMultipleDests/_StreamToBigQuery/DropShard+WriteWithMultipleDests/_StreamToBigQuery/FromHashableTableRef+WriteWithMultipleDests/_StreamToBigQuery/StreamInsertRows/ParDo(BigQueryWriteFn)/ParDo(BigQueryWriteFn)+assert_that/WindowInto(WindowIntoFn)+assert_that/ToVoidKey+assert_that/Group/pair_with_1+assert_that/Group/GroupByKey/Reify+assert_that/Group/GroupByKey/Write failed., The job failed because a work item has failed 4 times. Look in previous log entries for the cause of each one of the 4 failures. For more information, see https://cloud.google.com/dataflow/docs/guides/common-errors. The work item was attempted on these workers: beamapp-jenkins-040200435-04011744-f06y-harness-bvnc Root cause: Work item failed., beamapp-jenkins-040200435-04011744-f06y-harness-bvnc Root cause: Work item failed., beamapp-jenkins-040200435-04011744-f06y-harness-bvnc Root cause: Work item failed., beamapp-jenkins-040200435-04011744-f06y-harness-bvnc Root cause: Work item failed. apache_beam.runners.dataflow.dataflow_runner: INFO: 2021-04-02T00:51:40.401Z: JOB_MESSAGE_DETAILED: Cleaning up. apache_beam.runners.dataflow.dataflow_runner: INFO: 2021-04-02T00:51:40.474Z: JOB_MESSAGE_DEBUG: Starting worker pool teardown. apache_beam.runners.dataflow.dataflow_runner: INFO: 2021-04-02T00:51:40.514Z: JOB_MESSAGE_BASIC: Stopping worker pool... apache_beam.runners.dataflow.dataflow_runner: INFO: 2021-04-02T00:52:24.704Z: JOB_MESSAGE_DETAILED: Autoscaling: Resized worker pool from 1 to 0. apache_beam.runners.dataflow.dataflow_runner: INFO: 2021-04-02T00:52:24.748Z: JOB_MESSAGE_BASIC: Worker pool stopped. apache_beam.runners.dataflow.dataflow_runner: INFO: 2021-04-02T00:52:24.786Z: JOB_MESSAGE_DEBUG: Tearing down pending resources... apache_beam.runners.dataflow.dataflow_runner: INFO: Job 2021-04-01_17_44_02-13246612188569405916 is in state JOB_STATE_FAILED apache_beam.io.gcp.bigquery_test: INFO: Deleting dataset python_bq_streaming_inserts_16173242282899 in project apache-beam-testing --------------------- >> end captured logging << --------------------- Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_04_45-17499911106935513691?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_18_22-6111063415707104243?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_26_36-17327399244369260471?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_34_48-12739655610141691080?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_43_34-8904825658856291089?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_52_05-17663820542570565868?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_00_50-7910866993548419496?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_10_04-9541438453981390952?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_04_44-12666063734752426789?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_25_09-9358276951089896496?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_34_10-15095321732487843045?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_44_02-13246612188569405916?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_53_39-8169603264695623614?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_01_19-14121265306356075688?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_10_29-11813810985893855594?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_18_26-12975671982474253829?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_04_43-3481809735560369729?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_16_19-4146306868750046917?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_26_27-11655956157150394654?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_36_46-16532005754299127761?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_46_27-10239772575795602718?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_03_12-6103011993953709379?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_04_37-10196174445192590071?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_24_48-260043316342306826?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_35_27-4344886509974489191?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_45_43-17600804804208880445?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_55_21-761759604198406612?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_05_01-6211927963546889031?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_14_33-15586965309424652698?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_06_49-2849046091287359575?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_16_24-4490476733231249062?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_25_16-10937490425694747470?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_33_33-9130949177186503729?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_42_13-4434591154204181687?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_51_20-14217975704553693486?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_00_11-6287410410108792747?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_08_35-2107570108767478992?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_04_40-11923440318096290666?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_15_06-3036661315574918107?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_25_08-5400835150284889802?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_36_38-15516523646359458818?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_46_04-12916830124750699196?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_55_31-4493096053694879498?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_05_30-2231336251416920402?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_14_38-5677104774130447032?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_04_45-8796887955952029692?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_14_11-6823945865749004560?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_24_28-31274945521125562?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_36_26-6497033412415571447?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_45_12-13578391856526670800?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_54_08-11949829486094318114?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_02_58-6574231160160435072?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_12_11-18051810939494385200?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_04_39-2714590777239726958?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_14_26-17562351703165780310?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_23_41-9544847414413479870?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_31_44-9917211290700106218?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_39_58-16539915611439943901?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_48_16-542683157434030362?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_17_57_55-15994789226180786108?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_07_24-9871979369858046350?project=apache-beam-testing Worker logs: https://console.cloud.google.com/dataflow/jobs/us-central1/2021-04-01_18_17_22-8824302216191922078?project=apache-beam-testing ---------------------------------------------------------------------- XML: nosetests-postCommitIT-df-py38.xml ---------------------------------------------------------------------- XML: <https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/nosetests.xml> ---------------------------------------------------------------------- Ran 69 tests in 4968.423s FAILED (SKIP=6, errors=1) > Task :sdks:python:test-suites:dataflow:py38:postCommitIT FAILED FAILURE: Build failed with an exception. * Where: Script '<https://ci-beam.apache.org/job/beam_PostCommit_Python38/ws/src/sdks/python/test-suites/dataflow/common.gradle'> line: 118 * What went wrong: Execution failed for task ':sdks:python:test-suites:dataflow:py38:postCommitIT'. > Process 'command 'sh'' finished with non-zero exit value 1 * Try: Run with --stacktrace option to get the stack trace. Run with --info or --debug option to get more log output. Run with --scan to get full insights. * Get more help at https://help.gradle.org Deprecated Gradle features were used in this build, making it incompatible with Gradle 7.0. Use '--warning-mode all' to show the individual deprecation warnings. See https://docs.gradle.org/6.8/userguide/command_line_interface.html#sec:command_line_warnings BUILD FAILED in 1h 26m 30s 206 actionable tasks: 150 executed, 52 from cache, 4 up-to-date Gradle was unable to watch the file system for changes. The inotify watches limit is too low. Publishing build scan... https://gradle.com/s/f7yw2fcpocj74 Build step 'Invoke Gradle script' changed build result to FAILURE Build step 'Invoke Gradle script' marked build as failure --------------------------------------------------------------------- To unsubscribe, e-mail: [email protected] For additional commands, e-mail: [email protected]
