This is an automated email from the ASF dual-hosted git repository.
potiuk pushed a commit to branch main
in repository https://gitbox.apache.org/repos/asf/airflow.git
The following commit(s) were added to refs/heads/main by this push:
new e8287734cb Fixup docstring for deprecated
DataprocSubmitHiveJobOperator (#32723)
e8287734cb is described below
commit e8287734cbc29d45eeab565936afe8c9eb2e5340
Author: max <[email protected]>
AuthorDate: Thu Jul 20 19:27:32 2023 +0200
Fixup docstring for deprecated DataprocSubmitHiveJobOperator (#32723)
---
airflow/providers/google/cloud/operators/dataproc.py | 4 ++++
.../providers/google/cloud/dataproc/example_dataproc_hive.py | 11 +++++++++--
2 files changed, 13 insertions(+), 2 deletions(-)
diff --git a/airflow/providers/google/cloud/operators/dataproc.py
b/airflow/providers/google/cloud/operators/dataproc.py
index 3fa6ae1e77..e84cbe32a8 100644
--- a/airflow/providers/google/cloud/operators/dataproc.py
+++ b/airflow/providers/google/cloud/operators/dataproc.py
@@ -1204,6 +1204,10 @@ class
DataprocSubmitPigJobOperator(DataprocJobBaseOperator):
class DataprocSubmitHiveJobOperator(DataprocJobBaseOperator):
"""Start a Hive query Job on a Cloud DataProc cluster.
+ .. seealso::
+ This operator is deprecated, please use
+
:class:`~airflow.providers.google.cloud.operators.dataproc.DataprocSubmitJobOperator`:
+
:param query: The query or reference to the query file (q extension).
:param query_uri: The HCFS URI of the script that contains the Hive
queries.
:param variables: Map of named parameters for the query.
diff --git
a/tests/system/providers/google/cloud/dataproc/example_dataproc_hive.py
b/tests/system/providers/google/cloud/dataproc/example_dataproc_hive.py
index ea98725b29..8d345f818d 100644
--- a/tests/system/providers/google/cloud/dataproc/example_dataproc_hive.py
+++ b/tests/system/providers/google/cloud/dataproc/example_dataproc_hive.py
@@ -74,7 +74,7 @@ with models.DAG(
schedule="@once",
start_date=datetime(2021, 1, 1),
catchup=False,
- tags=["example", "dataproc"],
+ tags=["example", "dataproc", "hive"],
) as dag:
# [START how_to_cloud_dataproc_create_cluster_operator]
create_cluster = DataprocCreateClusterOperator(
@@ -100,7 +100,14 @@ with models.DAG(
# [END how_to_cloud_dataproc_delete_cluster_operator]
delete_cluster.trigger_rule = TriggerRule.ALL_DONE
- create_cluster >> hive_task >> delete_cluster
+ (
+ # TEST SETUP
+ create_cluster
+ # TEST BODY
+ >> hive_task
+ # TEST TEARDOWN
+ >> delete_cluster
+ )
from tests.system.utils.watcher import watcher