[
https://issues.apache.org/jira/browse/AIRFLOW-5006?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
]
t oo updated AIRFLOW-5006:
--------------------------
Description:
if a dagrun fails (and so some rows go to task_fail table in metastore) but
then after running 'clear' from cli to restart the dagrun (and its successful
run) the rows from metastore's task_fail table get deleted. is this expected?
or is there some way i can keep history of all failures in metastore?
even history of old successful runs are lost when doing clear for same exec date
was:if a dagrun fails (and so some rows go to task_fail table in metastore)
but then after running 'clear' from cli to restart the dagrun (and its
successful run) the rows from metastore's task_fail table get deleted. is this
expected? or is there some way i can keep history of all failures in metastore?
> Need task_failures audit trail for failed run in metastore db even after
> running 'clear' results in next DagRun successful
> --------------------------------------------------------------------------------------------------------------------------
>
> Key: AIRFLOW-5006
> URL: https://issues.apache.org/jira/browse/AIRFLOW-5006
> Project: Apache Airflow
> Issue Type: Improvement
> Components: database
> Affects Versions: 1.10.3, 1.10.5
> Environment: mysql RDS metastore, localexecutor
> Reporter: t oo
> Priority: Major
>
> if a dagrun fails (and so some rows go to task_fail table in metastore) but
> then after running 'clear' from cli to restart the dagrun (and its successful
> run) the rows from metastore's task_fail table get deleted. is this expected?
> or is there some way i can keep history of all failures in metastore?
>
> even history of old successful runs are lost when doing clear for same exec
> date
--
This message was sent by Atlassian Jira
(v8.3.4#803005)