This is an automated email from the ASF dual-hosted git repository.

vincbeck pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/airflow.git


The following commit(s) were added to refs/heads/main by this push:
     new d65316f4cbd Increase the redshift cluster deletion to 25 minutes in 
system tests (#62413)
d65316f4cbd is described below

commit d65316f4cbdf2c6e2f32671243ed603432d68177
Author: Vincent <[email protected]>
AuthorDate: Tue Feb 24 13:28:51 2026 -0500

    Increase the redshift cluster deletion to 25 minutes in system tests 
(#62413)
---
 providers/amazon/tests/system/amazon/aws/example_redshift.py             | 1 +
 .../amazon/tests/system/amazon/aws/example_redshift_s3_transfers.py      | 1 +
 providers/amazon/tests/system/amazon/aws/example_s3_to_sql.py            | 1 +
 providers/amazon/tests/system/amazon/aws/example_sql_to_s3.py            | 1 +
 4 files changed, 4 insertions(+)

diff --git a/providers/amazon/tests/system/amazon/aws/example_redshift.py 
b/providers/amazon/tests/system/amazon/aws/example_redshift.py
index 8d9063db9b9..f54e3609305 100644
--- a/providers/amazon/tests/system/amazon/aws/example_redshift.py
+++ b/providers/amazon/tests/system/amazon/aws/example_redshift.py
@@ -220,6 +220,7 @@ with DAG(
     )
     # [END howto_operator_redshift_delete_cluster]
     delete_cluster.trigger_rule = TriggerRule.ALL_DONE
+    delete_cluster.max_attempts = 50
 
     # [START howto_operator_redshift_delete_cluster_snapshot]
     delete_cluster_snapshot = RedshiftDeleteClusterSnapshotOperator(
diff --git 
a/providers/amazon/tests/system/amazon/aws/example_redshift_s3_transfers.py 
b/providers/amazon/tests/system/amazon/aws/example_redshift_s3_transfers.py
index e4658a96567..46373164ebb 100644
--- a/providers/amazon/tests/system/amazon/aws/example_redshift_s3_transfers.py
+++ b/providers/amazon/tests/system/amazon/aws/example_redshift_s3_transfers.py
@@ -271,6 +271,7 @@ with DAG(
         task_id="delete_cluster",
         cluster_identifier=redshift_cluster_identifier,
         trigger_rule=TriggerRule.ALL_DONE,
+        max_attempts=50,
     )
 
     delete_bucket = S3DeleteBucketOperator(
diff --git a/providers/amazon/tests/system/amazon/aws/example_s3_to_sql.py 
b/providers/amazon/tests/system/amazon/aws/example_s3_to_sql.py
index e553f89b973..117de53376b 100644
--- a/providers/amazon/tests/system/amazon/aws/example_s3_to_sql.py
+++ b/providers/amazon/tests/system/amazon/aws/example_s3_to_sql.py
@@ -249,6 +249,7 @@ with DAG(
         task_id="delete_cluster",
         cluster_identifier=redshift_cluster_identifier,
         trigger_rule=TriggerRule.ALL_DONE,
+        max_attempts=50,
     )
 
     chain(
diff --git a/providers/amazon/tests/system/amazon/aws/example_sql_to_s3.py 
b/providers/amazon/tests/system/amazon/aws/example_sql_to_s3.py
index 0237d47f1cd..7d364da46e9 100644
--- a/providers/amazon/tests/system/amazon/aws/example_sql_to_s3.py
+++ b/providers/amazon/tests/system/amazon/aws/example_sql_to_s3.py
@@ -198,6 +198,7 @@ with DAG(
         task_id="delete_cluster",
         cluster_identifier=redshift_cluster_identifier,
         trigger_rule=TriggerRule.ALL_DONE,
+        max_attempts=50,
     )
 
     chain(

Reply via email to