denimalpaca commented on code in PR #25164:
URL: https://github.com/apache/airflow/pull/25164#discussion_r957336168
##########
airflow/providers/common/sql/operators/sql.py:
##########
@@ -273,38 +303,38 @@ def __init__(
self.table = table
self.checks = checks
+ self.partition_clause = partition_clause
# OpenLineage needs a valid SQL query with the input/output table(s)
to parse
self.sql = f"SELECT * FROM {self.table};"
def execute(self, context=None):
hook = self.get_db_hook()
-
- check_names = [*self.checks]
- check_mins_sql = ",".join(
- self.sql_min_template.replace("check_name", check_name) for
check_name in check_names
- )
- checks_sql = ",".join(
+ checks_sql = " UNION ALL ".join(
[
- self.sql_check_template.replace("check_statement",
value["check_statement"]).replace(
- "check_name", check_name
- )
+ self.sql_check_template.replace("check_statement",
value["check_statement"])
+ .replace("_check_name", check_name)
+ .replace("table", self.table)
for check_name, value in self.checks.items()
]
)
+ partition_clause_statement = f"WHERE {self.partition_clause}" if
self.partition_clause else ""
+ self.sql = f"SELECT check_name, check_result FROM ({checks_sql}) "
+ f"AS check_table {partition_clause_statement};"
- self.sql = f"SELECT {check_mins_sql} FROM (SELECT {checks_sql} FROM
{self.table});"
- records = hook.get_first(self.sql)
+ records = hook.get_pandas_df(self.sql)
Review Comment:
Ideally most users won't need to learn about why `.fetch_all` is being used
instead of `.get_first` 🙃 . I'm working with some users of the operator right
now and seeing what's complicated to make sure the docs are robust. I also have
a working example DAGs showing how to use the operator (with several more
planned)
[here](https://registry.astronomer.io/providers/common-sql/modules/sqltablecheckoperator/#example-dags).
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]