Lee-W commented on code in PR #37000: URL: https://github.com/apache/airflow/pull/37000#discussion_r1465944418
########## airflow/providers/amazon/aws/operators/neptune.py: ########## @@ -0,0 +1,222 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from __future__ import annotations + +from typing import TYPE_CHECKING, Sequence + +from airflow.configuration import conf +from airflow.providers.amazon.aws.hooks.neptune import NeptuneHook +from airflow.providers.amazon.aws.operators.base_aws import AwsBaseOperator +from airflow.providers.amazon.aws.triggers.neptune import ( + NeptuneClusterAvailableTrigger, + NeptuneClusterStoppedTrigger, +) +from airflow.providers.amazon.aws.utils.mixins import aws_template_fields + +if TYPE_CHECKING: + from airflow.utils.context import Context + +AVAILABLE_STATES = ["available"] +STOPPED_STATES = ["stopped"] + + +class NeptuneStartDbClusterOperator(AwsBaseOperator[NeptuneHook]): + """Starts an Amazon Neptune DB cluster. + + Amazon Neptune Database is a serverless graph database designed for superior scalability + and availability. Neptune Database provides built-in security, continuous backups, and + integrations with other AWS services + + .. seealso:: + For more information on how to use this operator, take a look at the guide: + :ref:`howto/operator:NeptuneStartDbClusterOperator` + + :param db_cluster_id: The DB cluster identifier of the Neptune DB cluster to be started. + :param wait_for_completion: Whether to wait for the cluster to start. (default: True) + :param deferrable: If True, the operator will wait asynchronously for the cluster to start. + This implies waiting for completion. This mode requires aiobotocore module to be installed. + (default: False) + :param waiter_delay: Time in seconds to wait between status checks. + :param waiter_max_attempts: Maximum number of attempts to check for job completion. + :param aws_conn_id: The Airflow connection used for AWS credentials. + If this is ``None`` or empty then the default boto3 behaviour is used. If + running Airflow in a distributed manner and aws_conn_id is None or + empty, then default boto3 configuration would be used (and must be + maintained on each worker node). + :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. + + :param botocore_config: Configuration dictionary (key-values) for botocore client. See: + https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html + :return: dictionary with Neptune cluster id + """ + + aws_hook_class = NeptuneHook + template_fields: Sequence[str] = aws_template_fields("cluster_id") + + def __init__( + self, + db_cluster_id: str, + wait_for_completion: bool = True, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + deferrable: bool = conf.getboolean("operators", "default_deferrable", fallback=False), + **kwargs, + ): + super().__init__(**kwargs) + self.cluster_id = db_cluster_id + self.wait_for_completion = wait_for_completion + self.deferrable = deferrable + self.delay = waiter_delay + self.max_attempts = waiter_max_attempts + + """ @cached_property + def hook(self) -> NeptuneHook: + return NeptuneHook(aws_conn_id=self.aws_conn_id) """ Review Comment: ```suggestion @cached_property def hook(self) -> NeptuneHook: return NeptuneHook(aws_conn_id=self.aws_conn_id) ``` ########## airflow/providers/amazon/aws/operators/neptune.py: ########## @@ -0,0 +1,222 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from __future__ import annotations + +from typing import TYPE_CHECKING, Sequence + +from airflow.configuration import conf +from airflow.providers.amazon.aws.hooks.neptune import NeptuneHook +from airflow.providers.amazon.aws.operators.base_aws import AwsBaseOperator +from airflow.providers.amazon.aws.triggers.neptune import ( + NeptuneClusterAvailableTrigger, + NeptuneClusterStoppedTrigger, +) +from airflow.providers.amazon.aws.utils.mixins import aws_template_fields + +if TYPE_CHECKING: + from airflow.utils.context import Context + +AVAILABLE_STATES = ["available"] +STOPPED_STATES = ["stopped"] + + +class NeptuneStartDbClusterOperator(AwsBaseOperator[NeptuneHook]): + """Starts an Amazon Neptune DB cluster. + + Amazon Neptune Database is a serverless graph database designed for superior scalability + and availability. Neptune Database provides built-in security, continuous backups, and + integrations with other AWS services + + .. seealso:: + For more information on how to use this operator, take a look at the guide: + :ref:`howto/operator:NeptuneStartDbClusterOperator` + + :param db_cluster_id: The DB cluster identifier of the Neptune DB cluster to be started. + :param wait_for_completion: Whether to wait for the cluster to start. (default: True) + :param deferrable: If True, the operator will wait asynchronously for the cluster to start. + This implies waiting for completion. This mode requires aiobotocore module to be installed. + (default: False) + :param waiter_delay: Time in seconds to wait between status checks. + :param waiter_max_attempts: Maximum number of attempts to check for job completion. + :param aws_conn_id: The Airflow connection used for AWS credentials. + If this is ``None`` or empty then the default boto3 behaviour is used. If + running Airflow in a distributed manner and aws_conn_id is None or + empty, then default boto3 configuration would be used (and must be + maintained on each worker node). + :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. + + :param botocore_config: Configuration dictionary (key-values) for botocore client. See: + https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html + :return: dictionary with Neptune cluster id + """ + + aws_hook_class = NeptuneHook + template_fields: Sequence[str] = aws_template_fields("cluster_id") + + def __init__( + self, + db_cluster_id: str, + wait_for_completion: bool = True, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + deferrable: bool = conf.getboolean("operators", "default_deferrable", fallback=False), + **kwargs, + ): + super().__init__(**kwargs) + self.cluster_id = db_cluster_id + self.wait_for_completion = wait_for_completion + self.deferrable = deferrable + self.delay = waiter_delay + self.max_attempts = waiter_max_attempts + + """ @cached_property + def hook(self) -> NeptuneHook: + return NeptuneHook(aws_conn_id=self.aws_conn_id) """ + + def execute(self, context: Context): + self.log.info("Starting Neptune cluster: %s", self.cluster_id) + + # Check to make sure the cluster is not already available. + status = self.hook.get_cluster_status(self.cluster_id) + if status.lower() in AVAILABLE_STATES: + self.log.info("Neptune cluster %s is already available.", self.cluster_id) + return {"db_cluster_id": self.cluster_id} + + resp = self.hook.conn.start_db_cluster(DBClusterIdentifier=self.cluster_id) + status = resp.get("DBClusters", {}).get("Status", "Unknown") + + if self.deferrable: + self.log.info("Deferring for cluster start: %s", self.cluster_id) + + self.defer( + trigger=NeptuneClusterAvailableTrigger( + aws_conn_id=self.aws_conn_id, + db_cluster_id=self.cluster_id, + waiter_delay=self.delay, + waiter_max_attempts=self.max_attempts, + ), + method_name="execute_complete", + ) + + elif self.wait_for_completion: + self.log.info("Waiting for Neptune cluster %s to start.", self.cluster_id) + self.hook.wait_for_cluster_availability(self.cluster_id, self.delay, self.max_attempts) + + return {"db_cluster_id": self.cluster_id} + + def execute_complete(self, context: Context, event=None) -> dict[str, str]: + status = event.get("status", "") + cluster_id = event.get("cluster_id", "") + + self.log.info("Neptune cluster %s available with status: %s", cluster_id, status) + + return {"db_cluster_id": cluster_id} + + +class NeptuneStopDbClusterOperator(AwsBaseOperator[NeptuneHook]): + """ + Stops an Amazon Neptune DB cluster. + + Amazon Neptune Database is a serverless graph database designed for superior scalability + and availability. Neptune Database provides built-in security, continuous backups, and + integrations with other AWS services + + .. seealso:: + For more information on how to use this operator, take a look at the guide: + :ref:`howto/operator:NeptuneStartDbClusterOperator` + + :param db_cluster_id: The DB cluster identifier of the Neptune DB cluster to be stopped. + :param wait_for_completion: Whether to wait for cluster to stop. (default: True) + :param deferrable: If True, the operator will wait asynchronously for the cluster to stop. + This implies waiting for completion. This mode requires aiobotocore module to be installed. + (default: False) + :param waiter_delay: Time in seconds to wait between status checks. + :param waiter_max_attempts: Maximum number of attempts to check for job completion. + :param aws_conn_id: The Airflow connection used for AWS credentials. + If this is ``None`` or empty then the default boto3 behaviour is used. If + running Airflow in a distributed manner and aws_conn_id is None or + empty, then default boto3 configuration would be used (and must be + maintained on each worker node). + :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. + + :param botocore_config: Configuration dictionary (key-values) for botocore client. See: + https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html + :return: dictionary with Neptune cluster id + """ + + aws_hook_class = NeptuneHook + template_fields: Sequence[str] = aws_template_fields("cluster_id") + + def __init__( + self, + db_cluster_id: str, + wait_for_completion: bool = True, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + deferrable: bool = conf.getboolean("operators", "default_deferrable", fallback=False), + **kwargs, + ): + super().__init__(**kwargs) + self.cluster_id = db_cluster_id + self.wait_for_completion = wait_for_completion + self.deferrable = deferrable + self.delay = waiter_delay + self.max_attempts = waiter_max_attempts + + """ @cached_property + def hook(self) -> NeptuneHook: + return NeptuneHook(aws_conn_id=self.aws_conn_id) """ + + def execute(self, context: Context): Review Comment: ```suggestion def execute(self, context: Context) -> dict[str, str]: ``` ########## airflow/providers/amazon/aws/operators/neptune.py: ########## @@ -0,0 +1,222 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from __future__ import annotations + +from typing import TYPE_CHECKING, Sequence + +from airflow.configuration import conf +from airflow.providers.amazon.aws.hooks.neptune import NeptuneHook +from airflow.providers.amazon.aws.operators.base_aws import AwsBaseOperator +from airflow.providers.amazon.aws.triggers.neptune import ( + NeptuneClusterAvailableTrigger, + NeptuneClusterStoppedTrigger, +) +from airflow.providers.amazon.aws.utils.mixins import aws_template_fields + +if TYPE_CHECKING: + from airflow.utils.context import Context + +AVAILABLE_STATES = ["available"] +STOPPED_STATES = ["stopped"] + + +class NeptuneStartDbClusterOperator(AwsBaseOperator[NeptuneHook]): + """Starts an Amazon Neptune DB cluster. + + Amazon Neptune Database is a serverless graph database designed for superior scalability + and availability. Neptune Database provides built-in security, continuous backups, and + integrations with other AWS services + + .. seealso:: + For more information on how to use this operator, take a look at the guide: + :ref:`howto/operator:NeptuneStartDbClusterOperator` + + :param db_cluster_id: The DB cluster identifier of the Neptune DB cluster to be started. + :param wait_for_completion: Whether to wait for the cluster to start. (default: True) + :param deferrable: If True, the operator will wait asynchronously for the cluster to start. + This implies waiting for completion. This mode requires aiobotocore module to be installed. + (default: False) + :param waiter_delay: Time in seconds to wait between status checks. + :param waiter_max_attempts: Maximum number of attempts to check for job completion. + :param aws_conn_id: The Airflow connection used for AWS credentials. + If this is ``None`` or empty then the default boto3 behaviour is used. If + running Airflow in a distributed manner and aws_conn_id is None or + empty, then default boto3 configuration would be used (and must be + maintained on each worker node). + :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. + + :param botocore_config: Configuration dictionary (key-values) for botocore client. See: + https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html + :return: dictionary with Neptune cluster id + """ + + aws_hook_class = NeptuneHook + template_fields: Sequence[str] = aws_template_fields("cluster_id") + + def __init__( + self, + db_cluster_id: str, + wait_for_completion: bool = True, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + deferrable: bool = conf.getboolean("operators", "default_deferrable", fallback=False), + **kwargs, + ): + super().__init__(**kwargs) + self.cluster_id = db_cluster_id + self.wait_for_completion = wait_for_completion + self.deferrable = deferrable + self.delay = waiter_delay + self.max_attempts = waiter_max_attempts + + """ @cached_property + def hook(self) -> NeptuneHook: + return NeptuneHook(aws_conn_id=self.aws_conn_id) """ + + def execute(self, context: Context): + self.log.info("Starting Neptune cluster: %s", self.cluster_id) + + # Check to make sure the cluster is not already available. + status = self.hook.get_cluster_status(self.cluster_id) + if status.lower() in AVAILABLE_STATES: + self.log.info("Neptune cluster %s is already available.", self.cluster_id) + return {"db_cluster_id": self.cluster_id} + + resp = self.hook.conn.start_db_cluster(DBClusterIdentifier=self.cluster_id) + status = resp.get("DBClusters", {}).get("Status", "Unknown") + + if self.deferrable: + self.log.info("Deferring for cluster start: %s", self.cluster_id) + Review Comment: we might want to check once whether the cluster is ready here so that we might not actually need to defer it to the trigger something like https://github.com/apache/airflow/blob/390eacb01a710239e69f71d0c52134b7e87d17d4/airflow/providers/amazon/aws/operators/batch.py#L241-L251 ########## airflow/providers/amazon/aws/operators/neptune.py: ########## @@ -0,0 +1,222 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from __future__ import annotations + +from typing import TYPE_CHECKING, Sequence + +from airflow.configuration import conf +from airflow.providers.amazon.aws.hooks.neptune import NeptuneHook +from airflow.providers.amazon.aws.operators.base_aws import AwsBaseOperator +from airflow.providers.amazon.aws.triggers.neptune import ( + NeptuneClusterAvailableTrigger, + NeptuneClusterStoppedTrigger, +) +from airflow.providers.amazon.aws.utils.mixins import aws_template_fields + +if TYPE_CHECKING: + from airflow.utils.context import Context + +AVAILABLE_STATES = ["available"] +STOPPED_STATES = ["stopped"] + + +class NeptuneStartDbClusterOperator(AwsBaseOperator[NeptuneHook]): + """Starts an Amazon Neptune DB cluster. + + Amazon Neptune Database is a serverless graph database designed for superior scalability + and availability. Neptune Database provides built-in security, continuous backups, and + integrations with other AWS services + + .. seealso:: + For more information on how to use this operator, take a look at the guide: + :ref:`howto/operator:NeptuneStartDbClusterOperator` + + :param db_cluster_id: The DB cluster identifier of the Neptune DB cluster to be started. + :param wait_for_completion: Whether to wait for the cluster to start. (default: True) + :param deferrable: If True, the operator will wait asynchronously for the cluster to start. + This implies waiting for completion. This mode requires aiobotocore module to be installed. + (default: False) + :param waiter_delay: Time in seconds to wait between status checks. + :param waiter_max_attempts: Maximum number of attempts to check for job completion. + :param aws_conn_id: The Airflow connection used for AWS credentials. + If this is ``None`` or empty then the default boto3 behaviour is used. If + running Airflow in a distributed manner and aws_conn_id is None or + empty, then default boto3 configuration would be used (and must be + maintained on each worker node). + :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. + + :param botocore_config: Configuration dictionary (key-values) for botocore client. See: + https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html + :return: dictionary with Neptune cluster id + """ + + aws_hook_class = NeptuneHook + template_fields: Sequence[str] = aws_template_fields("cluster_id") + + def __init__( + self, + db_cluster_id: str, + wait_for_completion: bool = True, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + deferrable: bool = conf.getboolean("operators", "default_deferrable", fallback=False), + **kwargs, + ): + super().__init__(**kwargs) + self.cluster_id = db_cluster_id + self.wait_for_completion = wait_for_completion + self.deferrable = deferrable + self.delay = waiter_delay + self.max_attempts = waiter_max_attempts + + """ @cached_property + def hook(self) -> NeptuneHook: + return NeptuneHook(aws_conn_id=self.aws_conn_id) """ + + def execute(self, context: Context): + self.log.info("Starting Neptune cluster: %s", self.cluster_id) + + # Check to make sure the cluster is not already available. + status = self.hook.get_cluster_status(self.cluster_id) + if status.lower() in AVAILABLE_STATES: + self.log.info("Neptune cluster %s is already available.", self.cluster_id) + return {"db_cluster_id": self.cluster_id} + + resp = self.hook.conn.start_db_cluster(DBClusterIdentifier=self.cluster_id) + status = resp.get("DBClusters", {}).get("Status", "Unknown") + + if self.deferrable: + self.log.info("Deferring for cluster start: %s", self.cluster_id) + + self.defer( + trigger=NeptuneClusterAvailableTrigger( + aws_conn_id=self.aws_conn_id, + db_cluster_id=self.cluster_id, + waiter_delay=self.delay, + waiter_max_attempts=self.max_attempts, + ), + method_name="execute_complete", + ) + + elif self.wait_for_completion: + self.log.info("Waiting for Neptune cluster %s to start.", self.cluster_id) + self.hook.wait_for_cluster_availability(self.cluster_id, self.delay, self.max_attempts) + + return {"db_cluster_id": self.cluster_id} + + def execute_complete(self, context: Context, event=None) -> dict[str, str]: Review Comment: ```suggestion def execute_complete(self, context: Context, event: dict[str, Any] | None = None) -> dict[str, str]: ``` ########## airflow/providers/amazon/aws/operators/neptune.py: ########## @@ -0,0 +1,222 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from __future__ import annotations + +from typing import TYPE_CHECKING, Sequence + +from airflow.configuration import conf +from airflow.providers.amazon.aws.hooks.neptune import NeptuneHook +from airflow.providers.amazon.aws.operators.base_aws import AwsBaseOperator +from airflow.providers.amazon.aws.triggers.neptune import ( + NeptuneClusterAvailableTrigger, + NeptuneClusterStoppedTrigger, +) +from airflow.providers.amazon.aws.utils.mixins import aws_template_fields + +if TYPE_CHECKING: + from airflow.utils.context import Context + +AVAILABLE_STATES = ["available"] +STOPPED_STATES = ["stopped"] + + +class NeptuneStartDbClusterOperator(AwsBaseOperator[NeptuneHook]): + """Starts an Amazon Neptune DB cluster. + + Amazon Neptune Database is a serverless graph database designed for superior scalability + and availability. Neptune Database provides built-in security, continuous backups, and + integrations with other AWS services + + .. seealso:: + For more information on how to use this operator, take a look at the guide: + :ref:`howto/operator:NeptuneStartDbClusterOperator` + + :param db_cluster_id: The DB cluster identifier of the Neptune DB cluster to be started. + :param wait_for_completion: Whether to wait for the cluster to start. (default: True) + :param deferrable: If True, the operator will wait asynchronously for the cluster to start. + This implies waiting for completion. This mode requires aiobotocore module to be installed. + (default: False) + :param waiter_delay: Time in seconds to wait between status checks. + :param waiter_max_attempts: Maximum number of attempts to check for job completion. + :param aws_conn_id: The Airflow connection used for AWS credentials. + If this is ``None`` or empty then the default boto3 behaviour is used. If + running Airflow in a distributed manner and aws_conn_id is None or + empty, then default boto3 configuration would be used (and must be + maintained on each worker node). + :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. + + :param botocore_config: Configuration dictionary (key-values) for botocore client. See: + https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html + :return: dictionary with Neptune cluster id + """ + + aws_hook_class = NeptuneHook + template_fields: Sequence[str] = aws_template_fields("cluster_id") + + def __init__( + self, + db_cluster_id: str, + wait_for_completion: bool = True, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + deferrable: bool = conf.getboolean("operators", "default_deferrable", fallback=False), + **kwargs, + ): + super().__init__(**kwargs) + self.cluster_id = db_cluster_id + self.wait_for_completion = wait_for_completion + self.deferrable = deferrable + self.delay = waiter_delay + self.max_attempts = waiter_max_attempts + + """ @cached_property + def hook(self) -> NeptuneHook: + return NeptuneHook(aws_conn_id=self.aws_conn_id) """ + + def execute(self, context: Context): + self.log.info("Starting Neptune cluster: %s", self.cluster_id) + + # Check to make sure the cluster is not already available. + status = self.hook.get_cluster_status(self.cluster_id) + if status.lower() in AVAILABLE_STATES: + self.log.info("Neptune cluster %s is already available.", self.cluster_id) + return {"db_cluster_id": self.cluster_id} + + resp = self.hook.conn.start_db_cluster(DBClusterIdentifier=self.cluster_id) + status = resp.get("DBClusters", {}).get("Status", "Unknown") + + if self.deferrable: + self.log.info("Deferring for cluster start: %s", self.cluster_id) + + self.defer( + trigger=NeptuneClusterAvailableTrigger( + aws_conn_id=self.aws_conn_id, + db_cluster_id=self.cluster_id, + waiter_delay=self.delay, + waiter_max_attempts=self.max_attempts, + ), + method_name="execute_complete", + ) + + elif self.wait_for_completion: + self.log.info("Waiting for Neptune cluster %s to start.", self.cluster_id) + self.hook.wait_for_cluster_availability(self.cluster_id, self.delay, self.max_attempts) + + return {"db_cluster_id": self.cluster_id} + + def execute_complete(self, context: Context, event=None) -> dict[str, str]: + status = event.get("status", "") + cluster_id = event.get("cluster_id", "") + + self.log.info("Neptune cluster %s available with status: %s", cluster_id, status) + + return {"db_cluster_id": cluster_id} + + +class NeptuneStopDbClusterOperator(AwsBaseOperator[NeptuneHook]): + """ + Stops an Amazon Neptune DB cluster. + + Amazon Neptune Database is a serverless graph database designed for superior scalability + and availability. Neptune Database provides built-in security, continuous backups, and + integrations with other AWS services + + .. seealso:: + For more information on how to use this operator, take a look at the guide: + :ref:`howto/operator:NeptuneStartDbClusterOperator` + + :param db_cluster_id: The DB cluster identifier of the Neptune DB cluster to be stopped. + :param wait_for_completion: Whether to wait for cluster to stop. (default: True) + :param deferrable: If True, the operator will wait asynchronously for the cluster to stop. + This implies waiting for completion. This mode requires aiobotocore module to be installed. + (default: False) + :param waiter_delay: Time in seconds to wait between status checks. + :param waiter_max_attempts: Maximum number of attempts to check for job completion. + :param aws_conn_id: The Airflow connection used for AWS credentials. + If this is ``None`` or empty then the default boto3 behaviour is used. If + running Airflow in a distributed manner and aws_conn_id is None or + empty, then default boto3 configuration would be used (and must be + maintained on each worker node). + :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. + + :param botocore_config: Configuration dictionary (key-values) for botocore client. See: + https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html + :return: dictionary with Neptune cluster id + """ + + aws_hook_class = NeptuneHook + template_fields: Sequence[str] = aws_template_fields("cluster_id") + + def __init__( + self, + db_cluster_id: str, + wait_for_completion: bool = True, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + deferrable: bool = conf.getboolean("operators", "default_deferrable", fallback=False), + **kwargs, + ): + super().__init__(**kwargs) + self.cluster_id = db_cluster_id + self.wait_for_completion = wait_for_completion + self.deferrable = deferrable + self.delay = waiter_delay + self.max_attempts = waiter_max_attempts + + """ @cached_property + def hook(self) -> NeptuneHook: + return NeptuneHook(aws_conn_id=self.aws_conn_id) """ + + def execute(self, context: Context): + self.log.info("Stopping Neptune cluster: %s", self.cluster_id) + + # Check to make sure the cluster is not already available. + status = self.hook.get_cluster_status(self.cluster_id) + if status.lower() in STOPPED_STATES: + self.log.info("Neptune cluster %s is already stopped.", self.cluster_id) + return {"db_cluster_id": self.cluster_id} + + resp = self.hook.conn.stop_db_cluster(DBClusterIdentifier=self.cluster_id) + status = resp.get("DBClusters", {}).get("Status", "Unknown") + + if self.deferrable: + self.log.info("Deferring for cluster stop: %s", self.cluster_id) + Review Comment: same suggestion as above ########## airflow/providers/amazon/aws/operators/neptune.py: ########## @@ -0,0 +1,222 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from __future__ import annotations + +from typing import TYPE_CHECKING, Sequence + +from airflow.configuration import conf +from airflow.providers.amazon.aws.hooks.neptune import NeptuneHook +from airflow.providers.amazon.aws.operators.base_aws import AwsBaseOperator +from airflow.providers.amazon.aws.triggers.neptune import ( + NeptuneClusterAvailableTrigger, + NeptuneClusterStoppedTrigger, +) +from airflow.providers.amazon.aws.utils.mixins import aws_template_fields + +if TYPE_CHECKING: + from airflow.utils.context import Context + +AVAILABLE_STATES = ["available"] +STOPPED_STATES = ["stopped"] + + +class NeptuneStartDbClusterOperator(AwsBaseOperator[NeptuneHook]): + """Starts an Amazon Neptune DB cluster. + + Amazon Neptune Database is a serverless graph database designed for superior scalability + and availability. Neptune Database provides built-in security, continuous backups, and + integrations with other AWS services + + .. seealso:: + For more information on how to use this operator, take a look at the guide: + :ref:`howto/operator:NeptuneStartDbClusterOperator` + + :param db_cluster_id: The DB cluster identifier of the Neptune DB cluster to be started. + :param wait_for_completion: Whether to wait for the cluster to start. (default: True) + :param deferrable: If True, the operator will wait asynchronously for the cluster to start. + This implies waiting for completion. This mode requires aiobotocore module to be installed. + (default: False) + :param waiter_delay: Time in seconds to wait between status checks. + :param waiter_max_attempts: Maximum number of attempts to check for job completion. + :param aws_conn_id: The Airflow connection used for AWS credentials. + If this is ``None`` or empty then the default boto3 behaviour is used. If + running Airflow in a distributed manner and aws_conn_id is None or + empty, then default boto3 configuration would be used (and must be + maintained on each worker node). + :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. + + :param botocore_config: Configuration dictionary (key-values) for botocore client. See: + https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html + :return: dictionary with Neptune cluster id + """ + + aws_hook_class = NeptuneHook + template_fields: Sequence[str] = aws_template_fields("cluster_id") + + def __init__( + self, + db_cluster_id: str, + wait_for_completion: bool = True, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + deferrable: bool = conf.getboolean("operators", "default_deferrable", fallback=False), + **kwargs, + ): + super().__init__(**kwargs) + self.cluster_id = db_cluster_id + self.wait_for_completion = wait_for_completion + self.deferrable = deferrable + self.delay = waiter_delay + self.max_attempts = waiter_max_attempts + + """ @cached_property + def hook(self) -> NeptuneHook: + return NeptuneHook(aws_conn_id=self.aws_conn_id) """ + + def execute(self, context: Context): Review Comment: ```suggestion def execute(self, context: Context) -> dict[str, str]: ``` ########## airflow/providers/amazon/aws/operators/neptune.py: ########## @@ -0,0 +1,222 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from __future__ import annotations + +from typing import TYPE_CHECKING, Sequence + +from airflow.configuration import conf +from airflow.providers.amazon.aws.hooks.neptune import NeptuneHook +from airflow.providers.amazon.aws.operators.base_aws import AwsBaseOperator +from airflow.providers.amazon.aws.triggers.neptune import ( + NeptuneClusterAvailableTrigger, + NeptuneClusterStoppedTrigger, +) +from airflow.providers.amazon.aws.utils.mixins import aws_template_fields + +if TYPE_CHECKING: + from airflow.utils.context import Context + +AVAILABLE_STATES = ["available"] +STOPPED_STATES = ["stopped"] + + +class NeptuneStartDbClusterOperator(AwsBaseOperator[NeptuneHook]): + """Starts an Amazon Neptune DB cluster. + + Amazon Neptune Database is a serverless graph database designed for superior scalability + and availability. Neptune Database provides built-in security, continuous backups, and + integrations with other AWS services + + .. seealso:: + For more information on how to use this operator, take a look at the guide: + :ref:`howto/operator:NeptuneStartDbClusterOperator` + + :param db_cluster_id: The DB cluster identifier of the Neptune DB cluster to be started. + :param wait_for_completion: Whether to wait for the cluster to start. (default: True) + :param deferrable: If True, the operator will wait asynchronously for the cluster to start. + This implies waiting for completion. This mode requires aiobotocore module to be installed. + (default: False) + :param waiter_delay: Time in seconds to wait between status checks. + :param waiter_max_attempts: Maximum number of attempts to check for job completion. + :param aws_conn_id: The Airflow connection used for AWS credentials. + If this is ``None`` or empty then the default boto3 behaviour is used. If + running Airflow in a distributed manner and aws_conn_id is None or + empty, then default boto3 configuration would be used (and must be + maintained on each worker node). + :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. + + :param botocore_config: Configuration dictionary (key-values) for botocore client. See: + https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html + :return: dictionary with Neptune cluster id + """ + + aws_hook_class = NeptuneHook + template_fields: Sequence[str] = aws_template_fields("cluster_id") + + def __init__( + self, + db_cluster_id: str, + wait_for_completion: bool = True, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + deferrable: bool = conf.getboolean("operators", "default_deferrable", fallback=False), + **kwargs, + ): + super().__init__(**kwargs) + self.cluster_id = db_cluster_id + self.wait_for_completion = wait_for_completion + self.deferrable = deferrable + self.delay = waiter_delay + self.max_attempts = waiter_max_attempts + + """ @cached_property + def hook(self) -> NeptuneHook: + return NeptuneHook(aws_conn_id=self.aws_conn_id) """ + + def execute(self, context: Context): + self.log.info("Starting Neptune cluster: %s", self.cluster_id) + + # Check to make sure the cluster is not already available. + status = self.hook.get_cluster_status(self.cluster_id) + if status.lower() in AVAILABLE_STATES: + self.log.info("Neptune cluster %s is already available.", self.cluster_id) + return {"db_cluster_id": self.cluster_id} + + resp = self.hook.conn.start_db_cluster(DBClusterIdentifier=self.cluster_id) + status = resp.get("DBClusters", {}).get("Status", "Unknown") + + if self.deferrable: + self.log.info("Deferring for cluster start: %s", self.cluster_id) + + self.defer( + trigger=NeptuneClusterAvailableTrigger( + aws_conn_id=self.aws_conn_id, + db_cluster_id=self.cluster_id, + waiter_delay=self.delay, + waiter_max_attempts=self.max_attempts, + ), + method_name="execute_complete", + ) + + elif self.wait_for_completion: + self.log.info("Waiting for Neptune cluster %s to start.", self.cluster_id) + self.hook.wait_for_cluster_availability(self.cluster_id, self.delay, self.max_attempts) + + return {"db_cluster_id": self.cluster_id} + + def execute_complete(self, context: Context, event=None) -> dict[str, str]: + status = event.get("status", "") + cluster_id = event.get("cluster_id", "") + + self.log.info("Neptune cluster %s available with status: %s", cluster_id, status) + + return {"db_cluster_id": cluster_id} + + +class NeptuneStopDbClusterOperator(AwsBaseOperator[NeptuneHook]): + """ + Stops an Amazon Neptune DB cluster. + + Amazon Neptune Database is a serverless graph database designed for superior scalability + and availability. Neptune Database provides built-in security, continuous backups, and + integrations with other AWS services + + .. seealso:: + For more information on how to use this operator, take a look at the guide: + :ref:`howto/operator:NeptuneStartDbClusterOperator` + + :param db_cluster_id: The DB cluster identifier of the Neptune DB cluster to be stopped. + :param wait_for_completion: Whether to wait for cluster to stop. (default: True) + :param deferrable: If True, the operator will wait asynchronously for the cluster to stop. + This implies waiting for completion. This mode requires aiobotocore module to be installed. + (default: False) + :param waiter_delay: Time in seconds to wait between status checks. + :param waiter_max_attempts: Maximum number of attempts to check for job completion. + :param aws_conn_id: The Airflow connection used for AWS credentials. + If this is ``None`` or empty then the default boto3 behaviour is used. If + running Airflow in a distributed manner and aws_conn_id is None or + empty, then default boto3 configuration would be used (and must be + maintained on each worker node). + :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. + + :param botocore_config: Configuration dictionary (key-values) for botocore client. See: + https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html + :return: dictionary with Neptune cluster id + """ + + aws_hook_class = NeptuneHook + template_fields: Sequence[str] = aws_template_fields("cluster_id") + + def __init__( + self, + db_cluster_id: str, + wait_for_completion: bool = True, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + deferrable: bool = conf.getboolean("operators", "default_deferrable", fallback=False), + **kwargs, + ): + super().__init__(**kwargs) + self.cluster_id = db_cluster_id + self.wait_for_completion = wait_for_completion + self.deferrable = deferrable + self.delay = waiter_delay + self.max_attempts = waiter_max_attempts + + """ @cached_property + def hook(self) -> NeptuneHook: + return NeptuneHook(aws_conn_id=self.aws_conn_id) """ + + def execute(self, context: Context): + self.log.info("Stopping Neptune cluster: %s", self.cluster_id) + + # Check to make sure the cluster is not already available. + status = self.hook.get_cluster_status(self.cluster_id) + if status.lower() in STOPPED_STATES: + self.log.info("Neptune cluster %s is already stopped.", self.cluster_id) + return {"db_cluster_id": self.cluster_id} + + resp = self.hook.conn.stop_db_cluster(DBClusterIdentifier=self.cluster_id) + status = resp.get("DBClusters", {}).get("Status", "Unknown") + + if self.deferrable: + self.log.info("Deferring for cluster stop: %s", self.cluster_id) + + self.defer( + trigger=NeptuneClusterStoppedTrigger( + aws_conn_id=self.aws_conn_id, + db_cluster_id=self.cluster_id, + waiter_delay=self.delay, + waiter_max_attempts=self.max_attempts, + ), + method_name="execute_complete", + ) + + elif self.wait_for_completion: + self.log.info("Waiting for Neptune cluster %s to start.", self.cluster_id) + self.hook.wait_for_cluster_stopped(self.cluster_id, self.delay, self.max_attempts) + + return {"db_cluster_id": self.cluster_id} + + def execute_complete(self, context: Context, event=None) -> dict[str, str]: Review Comment: ```suggestion def execute_complete(self, context: Context, event: dict[str, Any] | None = None) -> dict[str, str]: ``` ########## airflow/providers/amazon/aws/operators/neptune.py: ########## @@ -0,0 +1,222 @@ +# +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from __future__ import annotations + +from typing import TYPE_CHECKING, Sequence + +from airflow.configuration import conf +from airflow.providers.amazon.aws.hooks.neptune import NeptuneHook +from airflow.providers.amazon.aws.operators.base_aws import AwsBaseOperator +from airflow.providers.amazon.aws.triggers.neptune import ( + NeptuneClusterAvailableTrigger, + NeptuneClusterStoppedTrigger, +) +from airflow.providers.amazon.aws.utils.mixins import aws_template_fields + +if TYPE_CHECKING: + from airflow.utils.context import Context + +AVAILABLE_STATES = ["available"] +STOPPED_STATES = ["stopped"] + + +class NeptuneStartDbClusterOperator(AwsBaseOperator[NeptuneHook]): + """Starts an Amazon Neptune DB cluster. + + Amazon Neptune Database is a serverless graph database designed for superior scalability + and availability. Neptune Database provides built-in security, continuous backups, and + integrations with other AWS services + + .. seealso:: + For more information on how to use this operator, take a look at the guide: + :ref:`howto/operator:NeptuneStartDbClusterOperator` + + :param db_cluster_id: The DB cluster identifier of the Neptune DB cluster to be started. + :param wait_for_completion: Whether to wait for the cluster to start. (default: True) + :param deferrable: If True, the operator will wait asynchronously for the cluster to start. + This implies waiting for completion. This mode requires aiobotocore module to be installed. + (default: False) + :param waiter_delay: Time in seconds to wait between status checks. + :param waiter_max_attempts: Maximum number of attempts to check for job completion. + :param aws_conn_id: The Airflow connection used for AWS credentials. + If this is ``None`` or empty then the default boto3 behaviour is used. If + running Airflow in a distributed manner and aws_conn_id is None or + empty, then default boto3 configuration would be used (and must be + maintained on each worker node). + :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. + + :param botocore_config: Configuration dictionary (key-values) for botocore client. See: + https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html + :return: dictionary with Neptune cluster id + """ + + aws_hook_class = NeptuneHook + template_fields: Sequence[str] = aws_template_fields("cluster_id") + + def __init__( + self, + db_cluster_id: str, + wait_for_completion: bool = True, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + deferrable: bool = conf.getboolean("operators", "default_deferrable", fallback=False), + **kwargs, + ): + super().__init__(**kwargs) + self.cluster_id = db_cluster_id + self.wait_for_completion = wait_for_completion + self.deferrable = deferrable + self.delay = waiter_delay + self.max_attempts = waiter_max_attempts + + """ @cached_property + def hook(self) -> NeptuneHook: + return NeptuneHook(aws_conn_id=self.aws_conn_id) """ + + def execute(self, context: Context): + self.log.info("Starting Neptune cluster: %s", self.cluster_id) + + # Check to make sure the cluster is not already available. + status = self.hook.get_cluster_status(self.cluster_id) + if status.lower() in AVAILABLE_STATES: + self.log.info("Neptune cluster %s is already available.", self.cluster_id) + return {"db_cluster_id": self.cluster_id} + + resp = self.hook.conn.start_db_cluster(DBClusterIdentifier=self.cluster_id) + status = resp.get("DBClusters", {}).get("Status", "Unknown") + + if self.deferrable: + self.log.info("Deferring for cluster start: %s", self.cluster_id) + + self.defer( + trigger=NeptuneClusterAvailableTrigger( + aws_conn_id=self.aws_conn_id, + db_cluster_id=self.cluster_id, + waiter_delay=self.delay, + waiter_max_attempts=self.max_attempts, + ), + method_name="execute_complete", + ) + + elif self.wait_for_completion: + self.log.info("Waiting for Neptune cluster %s to start.", self.cluster_id) + self.hook.wait_for_cluster_availability(self.cluster_id, self.delay, self.max_attempts) + + return {"db_cluster_id": self.cluster_id} + + def execute_complete(self, context: Context, event=None) -> dict[str, str]: + status = event.get("status", "") + cluster_id = event.get("cluster_id", "") + + self.log.info("Neptune cluster %s available with status: %s", cluster_id, status) + + return {"db_cluster_id": cluster_id} + + +class NeptuneStopDbClusterOperator(AwsBaseOperator[NeptuneHook]): + """ + Stops an Amazon Neptune DB cluster. + + Amazon Neptune Database is a serverless graph database designed for superior scalability + and availability. Neptune Database provides built-in security, continuous backups, and + integrations with other AWS services + + .. seealso:: + For more information on how to use this operator, take a look at the guide: + :ref:`howto/operator:NeptuneStartDbClusterOperator` + + :param db_cluster_id: The DB cluster identifier of the Neptune DB cluster to be stopped. + :param wait_for_completion: Whether to wait for cluster to stop. (default: True) + :param deferrable: If True, the operator will wait asynchronously for the cluster to stop. + This implies waiting for completion. This mode requires aiobotocore module to be installed. + (default: False) + :param waiter_delay: Time in seconds to wait between status checks. + :param waiter_max_attempts: Maximum number of attempts to check for job completion. + :param aws_conn_id: The Airflow connection used for AWS credentials. + If this is ``None`` or empty then the default boto3 behaviour is used. If + running Airflow in a distributed manner and aws_conn_id is None or + empty, then default boto3 configuration would be used (and must be + maintained on each worker node). + :param region_name: AWS region_name. If not specified then the default boto3 behaviour is used. + + :param botocore_config: Configuration dictionary (key-values) for botocore client. See: + https://botocore.amazonaws.com/v1/documentation/api/latest/reference/config.html + :return: dictionary with Neptune cluster id + """ + + aws_hook_class = NeptuneHook + template_fields: Sequence[str] = aws_template_fields("cluster_id") + + def __init__( + self, + db_cluster_id: str, + wait_for_completion: bool = True, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + deferrable: bool = conf.getboolean("operators", "default_deferrable", fallback=False), + **kwargs, + ): + super().__init__(**kwargs) + self.cluster_id = db_cluster_id + self.wait_for_completion = wait_for_completion + self.deferrable = deferrable + self.delay = waiter_delay + self.max_attempts = waiter_max_attempts + + """ @cached_property + def hook(self) -> NeptuneHook: + return NeptuneHook(aws_conn_id=self.aws_conn_id) """ Review Comment: ```suggestion @cached_property def hook(self) -> NeptuneHook: return NeptuneHook(aws_conn_id=self.aws_conn_id) ``` ########## airflow/providers/amazon/aws/triggers/neptune.py: ########## @@ -0,0 +1,115 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from __future__ import annotations + +from typing import TYPE_CHECKING + +from airflow.providers.amazon.aws.hooks.neptune import NeptuneHook +from airflow.providers.amazon.aws.triggers.base import AwsBaseWaiterTrigger + +if TYPE_CHECKING: + from airflow.providers.amazon.aws.hooks.base_aws import AwsGenericHook + + +class NeptuneClusterAvailableTrigger(AwsBaseWaiterTrigger): + """ + Triggers when a Neptune Cluster is available. + + :param db_cluster_id: Cluster ID to poll. + :param waiter_delay: The amount of time in seconds to wait between attempts. + :param waiter_max_attempts: The maximum number of attempts to be made. + :param aws_conn_id: The Airflow connection used for AWS credentials. + :param region_name: AWS region name (example: us-east-1) + """ + + def __init__( + self, + *, + db_cluster_id: str, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + aws_conn_id: str | None = None, + region_name: str | None = None, + **kwargs, + ) -> None: + super().__init__( + serialized_fields={"db_cluster_id": db_cluster_id}, + waiter_name="cluster_available", + waiter_args={"DBClusterIdentifier": db_cluster_id}, + failure_message="Neptune cluster failed", + status_message="Status of Neptune cluster is", + status_queries=["DBClusters[0].Status"], + return_key="db_cluster_id", + return_value=db_cluster_id, + waiter_delay=waiter_delay, + waiter_max_attempts=waiter_max_attempts, + aws_conn_id=aws_conn_id, + **kwargs, + ) + + def hook(self) -> AwsGenericHook: + return NeptuneHook( + aws_conn_id=self.aws_conn_id, + region_name=self.region_name, + verify=self.verify, + config=self.botocore_config, + ) + + +class NeptuneClusterStoppedTrigger(AwsBaseWaiterTrigger): + """ + Triggers when a Neptune Cluster is stopped. + + :param db_cluster_id: Cluster ID to poll. + :param waiter_delay: The amount of time in seconds to wait between attempts. + :param waiter_max_attempts: The maximum number of attempts to be made. + :param aws_conn_id: The Airflow connection used for AWS credentials. + :param region_name: AWS region name (example: us-east-1) + """ + + def __init__( + self, + *, + db_cluster_id: str, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + aws_conn_id: str | None = None, + region_name: str | None = None, + **kwargs, + ) -> None: + super().__init__( + serialized_fields={"db_cluster_id": db_cluster_id}, + waiter_name="cluster_stopped", + waiter_args={"DBClusterIdentifier": db_cluster_id}, + failure_message="Neptune cluster failed", Review Comment: Maybe we should change it to fail to stop Neptune? ########## airflow/providers/amazon/aws/triggers/neptune.py: ########## @@ -0,0 +1,115 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. +from __future__ import annotations + +from typing import TYPE_CHECKING + +from airflow.providers.amazon.aws.hooks.neptune import NeptuneHook +from airflow.providers.amazon.aws.triggers.base import AwsBaseWaiterTrigger + +if TYPE_CHECKING: + from airflow.providers.amazon.aws.hooks.base_aws import AwsGenericHook + + +class NeptuneClusterAvailableTrigger(AwsBaseWaiterTrigger): + """ + Triggers when a Neptune Cluster is available. + + :param db_cluster_id: Cluster ID to poll. + :param waiter_delay: The amount of time in seconds to wait between attempts. + :param waiter_max_attempts: The maximum number of attempts to be made. + :param aws_conn_id: The Airflow connection used for AWS credentials. + :param region_name: AWS region name (example: us-east-1) + """ + + def __init__( + self, + *, + db_cluster_id: str, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + aws_conn_id: str | None = None, + region_name: str | None = None, + **kwargs, + ) -> None: + super().__init__( + serialized_fields={"db_cluster_id": db_cluster_id}, + waiter_name="cluster_available", + waiter_args={"DBClusterIdentifier": db_cluster_id}, + failure_message="Neptune cluster failed", + status_message="Status of Neptune cluster is", + status_queries=["DBClusters[0].Status"], + return_key="db_cluster_id", + return_value=db_cluster_id, + waiter_delay=waiter_delay, + waiter_max_attempts=waiter_max_attempts, + aws_conn_id=aws_conn_id, + **kwargs, + ) + + def hook(self) -> AwsGenericHook: + return NeptuneHook( + aws_conn_id=self.aws_conn_id, + region_name=self.region_name, + verify=self.verify, + config=self.botocore_config, + ) + + +class NeptuneClusterStoppedTrigger(AwsBaseWaiterTrigger): + """ + Triggers when a Neptune Cluster is stopped. + + :param db_cluster_id: Cluster ID to poll. + :param waiter_delay: The amount of time in seconds to wait between attempts. + :param waiter_max_attempts: The maximum number of attempts to be made. + :param aws_conn_id: The Airflow connection used for AWS credentials. + :param region_name: AWS region name (example: us-east-1) + """ + + def __init__( + self, + *, + db_cluster_id: str, + waiter_delay: int = 30, + waiter_max_attempts: int = 60, + aws_conn_id: str | None = None, + region_name: str | None = None, + **kwargs, + ) -> None: + super().__init__( + serialized_fields={"db_cluster_id": db_cluster_id}, + waiter_name="cluster_stopped", + waiter_args={"DBClusterIdentifier": db_cluster_id}, + failure_message="Neptune cluster failed", + status_message="Status of Neptune cluster is", + status_queries=["DBClusters[0].Status"], + return_key="db_cluster_id", + return_value=db_cluster_id, + waiter_delay=waiter_delay, + waiter_max_attempts=waiter_max_attempts, + aws_conn_id=aws_conn_id, + **kwargs, + ) + + def hook(self) -> AwsGenericHook: + return NeptuneHook( + aws_conn_id=self.aws_conn_id, + region_name=self.region_name, + verify=self.verify, + config=self.botocore_config, + ) Review Comment: I am not sure whether creating yet another `NeptuneClusterGenricTrigger` makes sense. But by doing so, we could remove one duplicate `hook` method -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: [email protected] For queries about this service, please contact Infrastructure at: [email protected]
