podhornyi opened a new issue #15077: URL: https://github.com/apache/airflow/issues/15077
**Apache Airflow version**: 2.0.0 **Kubernetes version (if you are using kubernetes)** (use `kubectl version`): v1.15.10 **Environment**: ``` TZ: Etc/UTC ## ---------------- ## Airflow ## ---------------- AIRFLOW__CORE__DAGS_FOLDER: "/opt/airflow/git-sync" AIRFLOW__CORE__EXECUTOR: "KubernetesExecutor" AIRFLOW__CORE__FERNET_KEY: "{{ .Values.fernet.key }}" AIRFLOW__CORE__LOAD_EXAMPLES: "False" AIRFLOW__SCHEDULER__CHILD_PROCESS_LOG_DIRECTORY: "/opt/airflow/logs/scheduler" AIRFLOW__SCHEDULER__DAG_DIR_LIST_INTERVAL: "15" AIRFLOW__SCHEDULER__MIN_FILE_PROCESS_INTERVAL: "0" AIRFLOW__SCHEDULER__CATCHUP_BY_DEFAULT: "False" AIRFLOW__WEBSERVER__BASE_URL: "http://localhost:8080" AIRFLOW__WEBSERVER__WEB_SERVER_PORT: "8080" AIRFLOW__WEBSERVER__AUTHENTICATE: "True" AIRFLOW__WEBSERVER__AUTH_BACKEND: "airflow.contrib.auth.backends.password_auth" AIRFLOW__WEBSERVER__COOKIE_SAMESITE: "Lax" ## ---------------- ## Airflow - User Configs ## ---------------- AIRFLOW__API__AUTH_BACKEND: "airflow.api.auth.backend.deny_all" AIRFLOW__WEBSERVER__EXPOSE_CONFIG: "True" AIRFLOW__WEBSERVER__DEFAULT_DAG_RUN_DISPLAY_NUMBER: "50" AIRFLOW__CORE__DAG_CONCURRENCY: "1" AIRFLOW__CORE__MAX_ACTIVE_RUNS_PER_DAG: "1" AIRFLOW__SCHEDULER__MAX_DAGRUNS_PER_LOOP_TO_SCHEDULE: "1" AIRFLOW__CORE__DAG_RUN_CONF_OVERRIDES_PARAMS: "True" # Kubernetes section AIRFLOW__KUBERNETES__NAMESPACE: "default" AIRFLOW__KUBERNETES__DELETE_WORKER_PODS_ON_FAILURE: "False" AIRFLOW__KUBERNETES__DELETE_WORKER_PODS: "True" # LOGGING AIRFLOW__LOGGING__LOGGING_CONFIG_CLASS: "airflow_service.loggers.log_config.LOGGING_CONFIG" AIRFLOW__LOGGING__BASE_LOG_FOLDER: "/opt/airflow/logs" AIRFLOW__LOGGING__DAG_PROCESSOR_MANAGER_LOG_LOCATION: "/opt/airflow/logs/dag_processor_manager/dag_processor_manager.log" AIRFLOW__LOGGING__REMOTE_LOGGING: "True" AIRFLOW__LOGGING__COLORED_CONSOLE_LOG: "False" # ELK AIRFLOW__ELASTICSEARCH__JSON_FORMAT: "True" AIRFLOW__ELASTICSEARCH__WRITE_STDOUT: "True" AIRFLOW__ELASTICSEARCH__HOST: "http://elk.logging.svc.cluster.local:8080" AIRFLOW__ELASTICSEARCH__JSON_FIELDS: "asctime, name, levelname, filename, lineno, message" ``` - **Kernel** (e.g. `uname -a`): 4.9.0-11-amd64 #1 SMP Debian 4.9.189-3+deb9u2 (2019-11-11) x86_64 GNU/Linux **What happened**: After scheduler create k8s pod, and the node where that pod run goes down, scheduler delete the pod, set it to state up_fore_reschedule and task stuck in queue state forever. **What you expected to happen**: As for task defined retry=0, task should be failed or Scheduler should recreate in case of retry > 0 **How to reproduce it**: Delete k8s node when scheduler create a task, another words when scheduler send and API call to k8s and receive Ok status. **Anything else we need to know**: After scheduler restart, its saw task in queue state and execute it on via k8s call in regular way: ``` {"asctime": "2021-03-29 09:37:57,815", "processName": "MainProcess", "module": "kubernetes_executor", "filename": "kubernetes_executor.py", "lineno": 462, "levelname": "INFO", "message": "When executor started up, found 1 queued task instances"} {"asctime": "2021-03-29 09:37:57,846", "processName": "MainProcess", "module": "kubernetes_executor", "filename": "kubernetes_executor.py", "lineno": 480, "levelname": "INFO", "message": "TaskInstance: <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-29 09:34:01.780510+00:00 [queued]> found in queued state but was not launched, rescheduling"} ... {"asctime": "2021-03-29 09:37:58,153", "processName": "MainProcess", "module": "scheduler_job", "filename": "scheduler_job.py", "lineno": 938, "levelname": "INFO", "message": "1 tasks up for execution:\n\t<TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-29 09:34:01.780510+00:00 [scheduled]>"} {"asctime": "2021-03-29 09:37:58,157", "processName": "MainProcess", "module": "scheduler_job", "filename": "scheduler_job.py", "lineno": 967, "levelname": "INFO", "message": "Figuring out tasks to run in Pool(name=default_pool) with 128 open slots and 1 task instances ready to be queued"} {"asctime": "2021-03-29 09:37:58,157", "processName": "MainProcess", "module": "scheduler_job", "filename": "scheduler_job.py", "lineno": 995, "levelname": "INFO", "message": "DAG test_reschedule.up-for-reschedule has 0/1 running and queued tasks"} {"asctime": "2021-03-29 09:37:58,157", "processName": "MainProcess", "module": "scheduler_job", "filename": "scheduler_job.py", "lineno": 1060, "levelname": "INFO", "message": "Setting the following tasks to queued state:\n\t<TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-29 09:34:01.780510+00:00 [scheduled]>"} {"asctime": "2021-03-29 09:37:58,161", "processName": "MainProcess", "module": "scheduler_job", "filename": "scheduler_job.py", "lineno": 1102, "levelname": "INFO", "message": "Sending TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 29, 9, 34, 1, 780510, tzinfo=Timezone('UTC')), try_number=1) to executor with priority 3 and queue default"} ``` Scheduler logs Task which stuck in queue: `test_reschedule.up-for-reschedule.task_1` Pod name which end with up_for_reschedule: `estrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b` ``` [2021-03-26 16:34:07,585] [INFO] 1 tasks up for execution: <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [scheduled]> {scheduler_job.py:938} [2021-03-26 16:34:07,604] [INFO] Figuring out tasks to run in Pool(name=default_pool) with 128 open slots and 1 task instances ready to be queued {scheduler_job.py:967} [2021-03-26 16:34:07,604] [INFO] DAG test_reschedule.up-for-reschedule has 0/1 running and queued tasks {scheduler_job.py:995} [2021-03-26 16:34:07,604] [INFO] Setting the following tasks to queued state: <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [scheduled]> {scheduler_job.py:1060} [2021-03-26 16:34:07,610] [INFO] Sending TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=Timezone('UTC')), try_number=1) to executor with priority 3 and queue default {scheduler_job.py:1102} [2021-03-26 16:34:07,611] [INFO] Adding to queue: ['airflow', 'tasks', 'run', 'test_reschedule.up-for-reschedule', 'task_1', '2021-03-26T16:34:06.628707+00:00', '--local', '--pool', 'default_pool', '--subdir', '/opt/airflow/git-sync/data-import/reschedule/deployment/airflow_dags/dag.py'] {base_executor.py:79} [2021-03-26 16:34:07,613] [DEBUG] 0 running task instances {base_executor.py:147} [2021-03-26 16:34:07,613] [DEBUG] 1 in queue {base_executor.py:148} [2021-03-26 16:34:07,613] [DEBUG] 32 open slots {base_executor.py:149} [2021-03-26 16:34:07,613] [INFO] Add task TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=Timezone('UTC')), try_number=1) with command ['airflow', 'tasks', 'run', 'test_reschedule.up-for-reschedule', 'task_1', '2021-03-26T16:34:06.628707+00:00', '--local', '--pool', 'default_pool', '--subdir', '/opt/airflow/git-sync/data-import/reschedule/deployment/airflow_dags/dag.py'] with executor_config {...} {kubernetes_executor.py:510} [2021-03-26 16:34:07,629] [DEBUG] Calling the <class 'airflow.executors.kubernetes_executor.KubernetesExecutor'> sync method {base_executor.py:158} [2021-03-26 16:34:07,629] [DEBUG] self.running: {TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=Timezone('UTC')), try_number=1)} {kubernetes_executor.py:524} [2021-03-26 16:34:07,630] [DEBUG] Syncing KubernetesExecutor {kubernetes_executor.py:337} [2021-03-26 16:34:07,630] [DEBUG] KubeJobWatcher alive, continuing {kubernetes_executor.py:263} [2021-03-26 16:34:07,642] [INFO] Kubernetes job is (TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=Timezone('UTC')), try_number=1), ['airflow', 'tasks', 'run', 'test_reschedule.up-for-reschedule', 'task_1', '2021-03-26T16:34:06.628707+00:00', '--local', '--pool', 'default_pool', '--subdir', '/opt/airflow/git-sync/data-import/reschedule/deployment/airflow_dags/dag.py'], {...}, None) {kubernetes_executor.py:277} [2021-03-26 16:34:07,716] [DEBUG] Kubernetes running for command ['airflow', 'tasks', 'run', 'test_reschedule.up-for-reschedule', 'task_1', '2021-03-26T16:34:06.628707+00:00', '--local', '--pool', 'default_pool', '--subdir', '/opt/airflow/git-sync/data-import/reschedule/deployment/airflow_dags/dag.py'] {kubernetes_executor.py:306} [2021-03-26 16:34:07,716] [DEBUG] Kubernetes launching image XXXXXXXXXXX.dkr.ecr.eu-central-1.amazonaws.com/airflow:2.0.0 {kubernetes_executor.py:307} [2021-03-26 16:34:07,718] [DEBUG] Pod Creation Request: {...} {pod_launcher.py:79} [2021-03-26 16:34:07,750] [DEBUG] response body: {"kind":"Pod","apiVersion":"v1","metadata":{"name":"testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b", ...,"status":{"phase":"Pending","qosClass":"Burstable"}} {rest.py:230} [2021-03-26 16:34:07,751] [DEBUG] Disposing DB connection pool (PID 20086) {settings.py:290} [2021-03-26 16:34:07,760] [INFO] Event: testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b had an event of type ADDED {kubernetes_executor.py:147} [2021-03-26 16:34:07,761] [INFO] Event: testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b Pending {kubernetes_executor.py:202} [2021-03-26 16:34:07,808] [INFO] Event: testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b had an event of type MODIFIED {kubernetes_executor.py:147} [2021-03-26 16:34:07,808] [INFO] Event: testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b Pending {kubernetes_executor.py:202} [2021-03-26 16:34:07,810] [DEBUG] Disposing DB connection pool (PID 20072) {settings.py:290} [2021-03-26 16:34:07,760] [DEBUG] Pod Creation Response: {...} {pod_launcher.py:84} [2021-03-26 16:34:07,812] [DEBUG] Kubernetes Job created! {kubernetes_executor.py:311} [2021-03-26 16:34:07,813] [INFO] Executor reports execution of test_reschedule.up-for-reschedule.task_1 execution_date=2021-03-26 16:34:06.628707+00:00 exited with status queued for try_number 1 {scheduler_job.py:1193} [2021-03-26 16:34:07,822] [INFO] Setting external_id for <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> to 9 {scheduler_job.py:1220} [2021-03-26 16:34:07,906] [DEBUG] number of tis tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 3 task(s) {dagrun.py:490} [2021-03-26 16:34:07,906] [DEBUG] number of scheduleable tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 2 task(s) {dagrun.py:498} [2021-03-26 16:34:07,910] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:07,910] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:07,911] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:836} [2021-03-26 16:34:07,911] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:816} [2021-03-26 16:34:07,911] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:07,911] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:07,911] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:836} [2021-03-26 16:34:07,912] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:816} [2021-03-26 16:34:07,916] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:07,916] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Not In Retry Period' PASSED: True, The context specified that being in a retry period was permitted. {taskinstance.py:836} [2021-03-26 16:34:07,917] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Trigger Rule' PASSED: True, The task instance did not have any upstream tasks. {taskinstance.py:836} [2021-03-26 16:34:07,917] [DEBUG] Dependencies all met for <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> {taskinstance.py:826} [2021-03-26 16:34:07,917] [DEBUG] Skipping SLA check for <DAG: test_reschedule.up-for-reschedule> because no tasks in DAG have SLAs {scheduler_job.py:1720} [2021-03-26 16:34:07,930] [DEBUG] No tasks to consider for execution. {scheduler_job.py:933} [2021-03-26 16:34:07,931] [DEBUG] 1 running task instances {base_executor.py:147} [2021-03-26 16:34:07,932] [DEBUG] 0 in queue {base_executor.py:148} [2021-03-26 16:34:07,932] [DEBUG] 31 open slots {base_executor.py:149} [2021-03-26 16:34:07,932] [DEBUG] Calling the <class 'airflow.executors.kubernetes_executor.KubernetesExecutor'> sync method {base_executor.py:158} [2021-03-26 16:34:07,932] [DEBUG] self.running: {TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=Timezone('UTC')), try_number=1)} {kubernetes_executor.py:524} [2021-03-26 16:34:07,932] [DEBUG] Syncing KubernetesExecutor {kubernetes_executor.py:337} [2021-03-26 16:34:07,932] [DEBUG] KubeJobWatcher alive, continuing {kubernetes_executor.py:263} .... [2021-03-26 16:34:08,977] [DEBUG] number of tis tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 3 task(s) {dagrun.py:490} [2021-03-26 16:34:08,978] [DEBUG] number of scheduleable tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 2 task(s) {dagrun.py:498} [2021-03-26 16:34:08,977] [DEBUG] Started a process (PID: 20286) to generate tasks for /opt/airflow/git-sync/airflow/up-for-reschedule/airflow_service/loggers/log_config.py {dag_processing.py:1000} [2021-03-26 16:34:08,978] [DEBUG] 2/2 DAG parsing processes running {dag_processing.py:983} [2021-03-26 16:34:08,978] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:08,978] [DEBUG] 3 file paths queued for processing {dag_processing.py:985} [2021-03-26 16:34:08,978] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:08,979] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:836} [2021-03-26 16:34:08,979] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:816} [2021-03-26 16:34:08,979] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:08,979] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:08,980] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:836} [2021-03-26 16:34:08,980] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:816} [2021-03-26 16:34:08,984] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:08,985] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Not In Retry Period' PASSED: True, The context specified that being in a retry period was permitted. {taskinstance.py:836} [2021-03-26 16:34:08,985] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Trigger Rule' PASSED: True, The task instance did not have any upstream tasks. {taskinstance.py:836} [2021-03-26 16:34:08,985] [DEBUG] Dependencies all met for <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> {taskinstance.py:826} [2021-03-26 16:34:08,985] [DEBUG] Skipping SLA check for <DAG: test_reschedule.up-for-reschedule> because no tasks in DAG have SLAs {scheduler_job.py:1720} [2021-03-26 16:34:08,998] [DEBUG] No tasks to consider for execution. {scheduler_job.py:933} [2021-03-26 16:34:09,000] [DEBUG] 1 running task instances {base_executor.py:147} [2021-03-26 16:34:09,000] [DEBUG] 0 in queue {base_executor.py:148} [2021-03-26 16:34:09,000] [DEBUG] 31 open slots {base_executor.py:149} [2021-03-26 16:34:09,000] [DEBUG] Calling the <class 'airflow.executors.kubernetes_executor.KubernetesExecutor'> sync method {base_executor.py:158} [2021-03-26 16:34:09,000] [DEBUG] self.running: {TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=Timezone('UTC')), try_number=1)} {kubernetes_executor.py:524} [2021-03-26 16:34:09,000] [DEBUG] Syncing KubernetesExecutor {kubernetes_executor.py:337} ... [2021-03-26 16:34:10,048] [DEBUG] number of tis tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 3 task(s) {dagrun.py:490} [2021-03-26 16:34:10,048] [DEBUG] number of scheduleable tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 2 task(s) {dagrun.py:498} [2021-03-26 16:34:10,048] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:10,048] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:10,049] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:836} [2021-03-26 16:34:10,049] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:816} [2021-03-26 16:34:10,049] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:10,049] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:10,049] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:836} [2021-03-26 16:34:10,050] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:816} [2021-03-26 16:34:10,055] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:10,056] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Not In Retry Period' PASSED: True, The context specified that being in a retry period was permitted. {taskinstance.py:836} [2021-03-26 16:34:10,056] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Trigger Rule' PASSED: True, The task instance did not have any upstream tasks. {taskinstance.py:836} [2021-03-26 16:34:10,056] [DEBUG] Dependencies all met for <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> {taskinstance.py:826} [2021-03-26 16:34:10,056] [DEBUG] Skipping SLA check for <DAG: test_reschedule.up-for-reschedule> because no tasks in DAG have SLAs {scheduler_job.py:1720} [2021-03-26 16:34:10,066] [DEBUG] Processor for /opt/airflow/git-sync/airflow/up-for-reschedule/airflow_service/loggers/log_config.py finished {dag_processing.py:949} [2021-03-26 16:34:10,066] [DEBUG] Waiting for <ForkProcess name='DagFileProcessor2911-Process' pid=20464 parent=82 stopped exitcode=0> {scheduler_job.py:309} [2021-03-26 16:34:10,071] [DEBUG] No tasks to consider for execution. {scheduler_job.py:933} [2021-03-26 16:34:10,071] [DEBUG] Started a process (PID: 20478) to generate tasks for /opt/airflow/git-sync/data-import/reschedule/deployment/airflow_dags/dag.py {dag_processing.py:1000} [2021-03-26 16:34:10,072] [DEBUG] 2/2 DAG parsing processes running {dag_processing.py:983} [2021-03-26 16:34:10,072] [DEBUG] 1 file paths queued for processing {dag_processing.py:985} [2021-03-26 16:34:10,073] [DEBUG] 1 running task instances {base_executor.py:147} [2021-03-26 16:34:10,073] [DEBUG] 0 in queue {base_executor.py:148} [2021-03-26 16:34:10,073] [DEBUG] 31 open slots {base_executor.py:149} [2021-03-26 16:34:10,074] [DEBUG] Calling the <class 'airflow.executors.kubernetes_executor.KubernetesExecutor'> sync method {base_executor.py:158} [2021-03-26 16:34:10,074] [DEBUG] self.running: {TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=Timezone('UTC')), try_number=1)} {kubernetes_executor.py:524} [2021-03-26 16:34:10,074] [DEBUG] Syncing KubernetesExecutor {kubernetes_executor.py:337} [2021-03-26 16:34:10,074] [DEBUG] KubeJobWatcher alive, continuing {kubernetes_executor.py:263} ... [2021-03-26 16:34:10,315] [DEBUG] number of tis tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 3 task(s) {dagrun.py:490} [2021-03-26 16:34:10,315] [DEBUG] number of scheduleable tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 2 task(s) {dagrun.py:498} [2021-03-26 16:34:10,316] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:10,316] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:10,316] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:836} [2021-03-26 16:34:10,316] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:816} [2021-03-26 16:34:10,316] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:10,317] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:10,317] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:836} [2021-03-26 16:34:10,317] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:816} [2021-03-26 16:34:10,323] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:10,323] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Not In Retry Period' PASSED: True, The context specified that being in a retry period was permitted. {taskinstance.py:836} [2021-03-26 16:34:10,324] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Trigger Rule' PASSED: True, The task instance did not have any upstream tasks. {taskinstance.py:836} [2021-03-26 16:34:10,324] [DEBUG] Dependencies all met for <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> {taskinstance.py:826} [2021-03-26 16:34:10,324] [DEBUG] Skipping SLA check for <DAG: test_reschedule.up-for-reschedule> because no tasks in DAG have SLAs {scheduler_job.py:1720} [2021-03-26 16:34:10,342] [DEBUG] No tasks to consider for execution. {scheduler_job.py:933} [2021-03-26 16:34:10,344] [DEBUG] 1 running task instances {base_executor.py:147} [2021-03-26 16:34:10,344] [DEBUG] 0 in queue {base_executor.py:148} [2021-03-26 16:34:10,344] [DEBUG] 31 open slots {base_executor.py:149} [2021-03-26 16:34:10,344] [DEBUG] Calling the <class 'airflow.executors.kubernetes_executor.KubernetesExecutor'> sync method {base_executor.py:158} [2021-03-26 16:34:10,344] [DEBUG] self.running: {TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=Timezone('UTC')), try_number=1)} {kubernetes_executor.py:524} [2021-03-26 16:34:10,345] [DEBUG] Syncing KubernetesExecutor {kubernetes_executor.py:337} [2021-03-26 16:34:10,345] [DEBUG] KubeJobWatcher alive, continuing {kubernetes_executor.py:263} ... [2021-03-26 16:34:11,408] [DEBUG] number of tis tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 3 task(s) {dagrun.py:490} [2021-03-26 16:34:11,408] [DEBUG] number of scheduleable tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 2 task(s) {dagrun.py:498} [2021-03-26 16:34:11,409] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:11,409] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:11,409] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:836} [2021-03-26 16:34:11,409] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:816} [2021-03-26 16:34:11,409] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:11,409] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:11,410] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:836} [2021-03-26 16:34:11,410] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:816} [2021-03-26 16:34:11,412] [DEBUG] Started a process (PID: 20702) to generate tasks for /opt/airflow/git-sync/airflow/up-for-reschedule/airflow_service/dag.py {dag_processing.py:1000} [2021-03-26 16:34:11,413] [DEBUG] 2/2 DAG parsing processes running {dag_processing.py:983} [2021-03-26 16:34:11,414] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:11,415] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Not In Retry Period' PASSED: True, The context specified that being in a retry period was permitted. {taskinstance.py:836} [2021-03-26 16:34:11,415] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Trigger Rule' PASSED: True, The task instance did not have any upstream tasks. {taskinstance.py:836} [2021-03-26 16:34:11,415] [DEBUG] Dependencies all met for <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> {taskinstance.py:826} [2021-03-26 16:34:11,413] [DEBUG] 2 file paths queued for processing {dag_processing.py:985} [2021-03-26 16:34:11,415] [DEBUG] Skipping SLA check for <DAG: test_reschedule.up-for-reschedule> because no tasks in DAG have SLAs {scheduler_job.py:1720} [2021-03-26 16:34:11,416] [DEBUG] Disposing DB connection pool (PID 20695) {settings.py:290} [2021-03-26 16:34:11,428] [DEBUG] No tasks to consider for execution. {scheduler_job.py:933} [2021-03-26 16:34:11,429] [DEBUG] 1 running task instances {base_executor.py:147} [2021-03-26 16:34:11,430] [DEBUG] 0 in queue {base_executor.py:148} [2021-03-26 16:34:11,430] [DEBUG] 31 open slots {base_executor.py:149} [2021-03-26 16:34:11,430] [DEBUG] Calling the <class 'airflow.executors.kubernetes_executor.KubernetesExecutor'> sync method {base_executor.py:158} [2021-03-26 16:34:11,430] [DEBUG] self.running: {TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=Timezone('UTC')), try_number=1)} {kubernetes_executor.py:524} [2021-03-26 16:34:11,430] [DEBUG] Syncing KubernetesExecutor {kubernetes_executor.py:337} [2021-03-26 16:34:11,431] [DEBUG] KubeJobWatcher alive, continuing {kubernetes_executor.py:263} .... [2021-03-26 16:34:12,486] [DEBUG] number of tis tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 3 task(s) {dagrun.py:490} [2021-03-26 16:34:12,486] [DEBUG] number of scheduleable tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 2 task(s) {dagrun.py:498} [2021-03-26 16:34:12,486] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:12,486] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:12,487] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:836} [2021-03-26 16:34:12,487] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:816} [2021-03-26 16:34:12,487] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:12,488] [DEBUG] Processor for /opt/airflow/git-sync/airflow/up-for-reschedule/airflow_service/dag.py finished {dag_processing.py:949} [2021-03-26 16:34:12,487] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:12,489] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:836} [2021-03-26 16:34:12,489] [DEBUG] Waiting for <ForkProcess name='DagFileProcessor2969-Process' pid=20873 parent=82 started> {scheduler_job.py:309} [2021-03-26 16:34:12,495] [DEBUG] Started a process (PID: 20884) to generate tasks for /opt/airflow/git-sync/airflow/up-for-reschedule/airflow_service/utils.py {dag_processing.py:1000} [2021-03-26 16:34:12,490] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:816} [2021-03-26 16:34:12,497] [DEBUG] 2/2 DAG parsing processes running {dag_processing.py:983} [2021-03-26 16:34:12,497] [DEBUG] 0 file paths queued for processing {dag_processing.py:985} [2021-03-26 16:34:12,509] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:12,509] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Not In Retry Period' PASSED: True, The context specified that being in a retry period was permitted. {taskinstance.py:836} [2021-03-26 16:34:12,509] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Trigger Rule' PASSED: True, The task instance did not have any upstream tasks. {taskinstance.py:836} [2021-03-26 16:34:12,509] [DEBUG] Dependencies all met for <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> {taskinstance.py:826} [2021-03-26 16:34:12,510] [DEBUG] Skipping SLA check for <DAG: test_reschedule.up-for-reschedule> because no tasks in DAG have SLAs {scheduler_job.py:1720} [2021-03-26 16:34:12,529] [DEBUG] No tasks to consider for execution. {scheduler_job.py:933} [2021-03-26 16:34:12,531] [DEBUG] 1 running task instances {base_executor.py:147} [2021-03-26 16:34:12,531] [DEBUG] 0 in queue {base_executor.py:148} [2021-03-26 16:34:12,534] [DEBUG] 31 open slots {base_executor.py:149} [2021-03-26 16:34:12,534] [DEBUG] Disposing DB connection pool (PID 20884) {settings.py:290} [2021-03-26 16:34:12,534] [DEBUG] Calling the <class 'airflow.executors.kubernetes_executor.KubernetesExecutor'> sync method {base_executor.py:158} [2021-03-26 16:34:12,534] [DEBUG] self.running: {TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=Timezone('UTC')), try_number=1)} {kubernetes_executor.py:524} [2021-03-26 16:34:12,534] [DEBUG] Syncing KubernetesExecutor {kubernetes_executor.py:337} [2021-03-26 16:34:12,534] [DEBUG] KubeJobWatcher alive, continuing {kubernetes_executor.py:263} ... ... ... [2021-03-26 16:34:48,998] [INFO] Event: testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b had an event of type MODIFIED {kubernetes_executor.py:147} [2021-03-26 16:34:49,004] [INFO] Event: testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b Pending {kubernetes_executor.py:202} ... [2021-03-26 16:34:49,024] [INFO] Event: testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b had an event of type DELETED {kubernetes_executor.py:147} [2021-03-26 16:34:49,024] [INFO] Event: Failed to start pod testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b, will reschedule {kubernetes_executor.py:197} ... [2021-03-26 16:34:49,372] [DEBUG] DAG test_reschedule.up-for-reschedule not changed structure, skipping dagrun.verify_integrity {scheduler_job.py:1692} [2021-03-26 16:34:49,380] [DEBUG] number of tis tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 3 task(s) {dagrun.py:490} [2021-03-26 16:34:49,380] [DEBUG] number of scheduleable tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 2 task(s) {dagrun.py:498} [2021-03-26 16:34:49,380] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:49,381] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:49,381] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:836} [2021-03-26 16:34:49,381] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:816} [2021-03-26 16:34:49,381] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:49,382] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:49,382] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:836} [2021-03-26 16:34:49,382] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:816} [2021-03-26 16:34:49,388] [DEBUG] Disposing DB connection pool (PID 27073) {settings.py:290} [2021-03-26 16:34:49,390] [DEBUG] Processor for /opt/airflow/git-sync/airflow/up-for-reschedule/airflow_service/final_task.py finished {dag_processing.py:949} [2021-03-26 16:34:49,390] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:49,391] [DEBUG] Waiting for <ForkProcess name='DagFileProcessor3853-Process' pid=27062 parent=82 stopped exitcode=0> {scheduler_job.py:309} [2021-03-26 16:34:49,391] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Not In Retry Period' PASSED: True, The context specified that being in a retry period was permitted. {taskinstance.py:836} [2021-03-26 16:34:49,391] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Trigger Rule' PASSED: True, The task instance did not have any upstream tasks. {taskinstance.py:836} [2021-03-26 16:34:49,391] [DEBUG] Dependencies all met for <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> {taskinstance.py:826} [2021-03-26 16:34:49,395] [DEBUG] Started a process (PID: 27082) to generate tasks for /opt/airflow/git-sync/airflow/up-for-reschedule/airflow_service/dag.py {dag_processing.py:1000} [2021-03-26 16:34:49,403] [DEBUG] Skipping SLA check for <DAG: test_reschedule.up-for-reschedule> because no tasks in DAG have SLAs {scheduler_job.py:1720} [2021-03-26 16:34:49,404] [DEBUG] 2/2 DAG parsing processes running {dag_processing.py:983} [2021-03-26 16:34:49,404] [DEBUG] 2 file paths queued for processing {dag_processing.py:985} [2021-03-26 16:34:49,411] [DEBUG] Processor for /opt/airflow/git-sync/airflow/up-for-reschedule/airflow_service/loggers/log_config.py finished {dag_processing.py:949} [2021-03-26 16:34:49,412] [DEBUG] Waiting for <ForkProcess name='DagFileProcessor3855-Process' pid=27073 parent=82 stopped exitcode=0> {scheduler_job.py:309} [2021-03-26 16:34:49,416] [DEBUG] Started a process (PID: 27087) to generate tasks for /opt/airflow/git-sync/data-import/reschedule/deployment/airflow_dags/dag.py {dag_processing.py:1000} [2021-03-26 16:34:49,417] [DEBUG] 2/2 DAG parsing processes running {dag_processing.py:983} [2021-03-26 16:34:49,417] [DEBUG] 1 file paths queued for processing {dag_processing.py:985} [2021-03-26 16:34:49,419] [DEBUG] No tasks to consider for execution. {scheduler_job.py:933} [2021-03-26 16:34:49,420] [DEBUG] 1 running task instances {base_executor.py:147} [2021-03-26 16:34:49,420] [DEBUG] 0 in queue {base_executor.py:148} [2021-03-26 16:34:49,420] [DEBUG] 31 open slots {base_executor.py:149} [2021-03-26 16:34:49,420] [DEBUG] Calling the <class 'airflow.executors.kubernetes_executor.KubernetesExecutor'> sync method {base_executor.py:158} [2021-03-26 16:34:49,421] [DEBUG] self.running: {TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=Timezone('UTC')), try_number=1)} {kubernetes_executor.py:524} [2021-03-26 16:34:49,421] [DEBUG] Syncing KubernetesExecutor {kubernetes_executor.py:337} [2021-03-26 16:34:49,421] [DEBUG] KubeJobWatcher alive, continuing {kubernetes_executor.py:263} [2021-03-26 16:34:49,421] [DEBUG] Processing task ('testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b', 'et-airflow-qa', 'up_for_reschedule', {'dag_id': 'test_reschedule.up-for-reschedule', 'task_id': 'task_1', 'execution_date': '2021-03-26T16:34:06.628707+00:00', 'try_number': '1'}, '137947551') {kubernetes_executor.py:343} [2021-03-26 16:34:49,421] [INFO] Attempting to finish pod; pod_id: testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b; state: up_for_reschedule; annotations: {'dag_id': 'test_reschedule.up-for-reschedule', 'task_id': 'task_1', 'execution_date': '2021-03-26T16:34:06.628707+00:00', 'try_number': '1'} {kubernetes_executor.py:353} [2021-03-26 16:34:49,421] [DEBUG] Creating task key for annotations {'dag_id': 'test_reschedule.up-for-reschedule', 'task_id': 'task_1', 'execution_date': '2021-03-26T16:34:06.628707+00:00', 'try_number': '1'} {kubernetes_executor.py:362} [2021-03-26 16:34:49,422] [DEBUG] finishing job TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=tzlocal()), try_number=1) - up_for_reschedule (testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b) {kubernetes_executor.py:358} [2021-03-26 16:34:49,423] [INFO] Changing state of (TaskInstanceKey(dag_id='test_reschedule.up-for-reschedule', task_id='task_1', execution_date=datetime.datetime(2021, 3, 26, 16, 34, 6, 628707, tzinfo=tzlocal()), try_number=1), 'up_for_reschedule', 'testrescheduleupforrescheduletask1-90fe939148314774bfcc61079113f76b', 'et-airflow-qa', '137947551') to up_for_reschedule {kubernetes_executor.py:546} [2021-03-26 16:34:49,424] [INFO] Executor reports execution of test_reschedule.up-for-reschedule.task_1 execution_date=2021-03-26 16:34:06.628707+00:00 exited with status up_for_reschedule for try_number 1 {scheduler_job.py:1193} ... [2021-03-26 16:34:49,465] [DEBUG] DAG test_reschedule.up-for-reschedule not changed structure, skipping dagrun.verify_integrity {scheduler_job.py:1692} [2021-03-26 16:34:49,473] [DEBUG] number of tis tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 3 task(s) {dagrun.py:490} [2021-03-26 16:34:49,474] [DEBUG] number of scheduleable tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 2 task(s) {dagrun.py:498} [2021-03-26 16:34:49,474] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:49,474] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:49,474] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:836} [2021-03-26 16:34:49,474] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:816} [2021-03-26 16:34:49,475] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:49,475] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:49,475] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:836} [2021-03-26 16:34:49,475] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:816} [2021-03-26 16:34:49,475] [DEBUG] Processor for /opt/airflow/git-sync/airflow/up-for-reschedule/airflow_service/dag.py finished {dag_processing.py:949} [2021-03-26 16:34:49,476] [DEBUG] Waiting for <ForkProcess name='DagFileProcessor3856-Process' pid=27082 parent=82 stopped exitcode=0> {scheduler_job.py:309} [2021-03-26 16:34:49,480] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:49,480] [DEBUG] Started a process (PID: 27094) to generate tasks for /opt/airflow/git-sync/airflow/up-for-reschedule/airflow_service/utils.py {dag_processing.py:1000} [2021-03-26 16:34:49,480] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Not In Retry Period' PASSED: True, The context specified that being in a retry period was permitted. {taskinstance.py:836} [2021-03-26 16:34:49,480] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Trigger Rule' PASSED: True, The task instance did not have any upstream tasks. {taskinstance.py:836} [2021-03-26 16:34:49,480] [DEBUG] Dependencies all met for <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> {taskinstance.py:826} [2021-03-26 16:34:49,480] [DEBUG] 2/2 DAG parsing processes running {dag_processing.py:983} [2021-03-26 16:34:49,481] [DEBUG] 0 file paths queued for processing {dag_processing.py:985} [2021-03-26 16:34:49,481] [DEBUG] Skipping SLA check for <DAG: test_reschedule.up-for-reschedule> because no tasks in DAG have SLAs {scheduler_job.py:1720} [2021-03-26 16:34:49,512] [DEBUG] No tasks to consider for execution. {scheduler_job.py:933} [2021-03-26 16:34:49,513] [DEBUG] 0 running task instances {base_executor.py:147} [2021-03-26 16:34:49,514] [DEBUG] 0 in queue {base_executor.py:148} [2021-03-26 16:34:49,514] [DEBUG] 32 open slots {base_executor.py:149} [2021-03-26 16:34:49,514] [DEBUG] Calling the <class 'airflow.executors.kubernetes_executor.KubernetesExecutor'> sync method {base_executor.py:158} [2021-03-26 16:34:49,514] [DEBUG] Syncing KubernetesExecutor {kubernetes_executor.py:337} [2021-03-26 16:34:49,515] [DEBUG] KubeJobWatcher alive, continuing {kubernetes_executor.py:263} ... [2021-03-26 16:34:50,558] [DEBUG] DAG test_reschedule.up-for-reschedule not changed structure, skipping dagrun.verify_integrity {scheduler_job.py:1692} [2021-03-26 16:34:50,567] [DEBUG] number of tis tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 3 task(s) {dagrun.py:490} [2021-03-26 16:34:50,567] [DEBUG] number of scheduleable tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 2 task(s) {dagrun.py:498} [2021-03-26 16:34:50,568] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:50,568] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:50,569] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:836} [2021-03-26 16:34:50,569] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:816} [2021-03-26 16:34:50,569] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:50,570] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:50,570] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:836} [2021-03-26 16:34:50,570] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:816} [2021-03-26 16:34:50,575] [DEBUG] Disposing DB connection pool (PID 27269) {settings.py:290} [2021-03-26 16:34:50,575] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:50,575] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Not In Retry Period' PASSED: True, The context specified that being in a retry period was permitted. {taskinstance.py:836} [2021-03-26 16:34:50,575] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Trigger Rule' PASSED: True, The task instance did not have any upstream tasks. {taskinstance.py:836} [2021-03-26 16:34:50,575] [DEBUG] Dependencies all met for <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> {taskinstance.py:826} [2021-03-26 16:34:50,576] [DEBUG] Skipping SLA check for <DAG: test_reschedule.up-for-reschedule> because no tasks in DAG have SLAs {scheduler_job.py:1720} [2021-03-26 16:34:50,584] [DEBUG] Processor for /opt/airflow/git-sync/airflow/up-for-reschedule/airflow_service/loggers/log_config.py finished {dag_processing.py:949} [2021-03-26 16:34:50,584] [DEBUG] Waiting for <ForkProcess name='DagFileProcessor3883-Process' pid=27269 parent=82 started> {scheduler_job.py:309} [2021-03-26 16:34:50,590] [DEBUG] No tasks to consider for execution. {scheduler_job.py:933} [2021-03-26 16:34:50,604] [DEBUG] 0 running task instances {base_executor.py:147} [2021-03-26 16:34:50,604] [DEBUG] 0 in queue {base_executor.py:148} [2021-03-26 16:34:50,604] [DEBUG] 32 open slots {base_executor.py:149} [2021-03-26 16:34:50,604] [DEBUG] Calling the <class 'airflow.executors.kubernetes_executor.KubernetesExecutor'> sync method {base_executor.py:158} [2021-03-26 16:34:50,604] [DEBUG] Syncing KubernetesExecutor {kubernetes_executor.py:337} [2021-03-26 16:34:50,605] [DEBUG] KubeJobWatcher alive, continuing {kubernetes_executor.py:263} ... [2021-03-26 16:34:52,281] [DEBUG] DAG test_reschedule.up-for-reschedule not changed structure, skipping dagrun.verify_integrity {scheduler_job.py:1692} [2021-03-26 16:34:52,290] [DEBUG] number of tis tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 3 task(s) {dagrun.py:490} [2021-03-26 16:34:52,290] [DEBUG] number of scheduleable tasks for <DagRun test_reschedule.up-for-reschedule @ 2021-03-26 16:34:06.628707+00:00: manual__2021-03-26T16:34:06.628707+00:00, externally triggered: True>: 2 task(s) {dagrun.py:498} [2021-03-26 16:34:52,290] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:52,290] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:52,290] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:836} [2021-03-26 16:34:52,291] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_3 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_2'} {taskinstance.py:816} [2021-03-26 16:34:52,291] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:52,291] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Not In Retry Period' PASSED: True, The task instance was not marked for retrying. {taskinstance.py:836} [2021-03-26 16:34:52,291] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]> dependency 'Trigger Rule' PASSED: False, Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:836} [2021-03-26 16:34:52,304] [DEBUG] Disposing DB connection pool (PID 27542) {settings.py:290} [2021-03-26 16:34:52,306] [DEBUG] Dependencies not met for <TaskInstance: test_reschedule.up-for-reschedule.task_2 2021-03-26 16:34:06.628707+00:00 [None]>, dependency 'Trigger Rule' FAILED: Task's trigger rule 'all_success' requires all upstream tasks to have succeeded, but found 1 non-success(es). upstream_tasks_state={'total': 1, 'successes': 0, 'skipped': 0, 'failed': 0, 'upstream_failed': 0, 'done': 0}, upstream_task_ids={'task_1'} {taskinstance.py:816} [2021-03-26 16:34:52,311] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Previous Dagrun State' PASSED: True, The task did not have depends_on_past set. {taskinstance.py:836} [2021-03-26 16:34:52,311] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Not In Retry Period' PASSED: True, The context specified that being in a retry period was permitted. {taskinstance.py:836} [2021-03-26 16:34:52,311] [DEBUG] <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> dependency 'Trigger Rule' PASSED: True, The task instance did not have any upstream tasks. {taskinstance.py:836} [2021-03-26 16:34:52,311] [DEBUG] Dependencies all met for <TaskInstance: test_reschedule.up-for-reschedule.task_1 2021-03-26 16:34:06.628707+00:00 [queued]> {taskinstance.py:826} [2021-03-26 16:34:52,312] [DEBUG] Skipping SLA check for <DAG: test_reschedule.up-for-reschedule> because no tasks in DAG have SLAs {scheduler_job.py:1720} [2021-03-26 16:34:52,323] [DEBUG] Processor for /opt/airflow/git-sync/data-import/reschedule/deployment/airflow_dags/dag.py finished {dag_processing.py:949} [2021-03-26 16:34:52,324] [DEBUG] Waiting for <ForkProcess name='DagFileProcessor3922-Process' pid=27542 parent=82 stopped exitcode=0> {scheduler_job.py:309} [2021-03-26 16:34:52,324] [DEBUG] No tasks to consider for execution. {scheduler_job.py:933} [2021-03-26 16:34:52,325] [DEBUG] 0 running task instances {base_executor.py:147} [2021-03-26 16:34:52,326] [DEBUG] 0 in queue {base_executor.py:148} [2021-03-26 16:34:52,326] [DEBUG] 32 open slots {base_executor.py:149} [2021-03-26 16:34:52,326] [DEBUG] Calling the <class 'airflow.executors.kubernetes_executor.KubernetesExecutor'> sync method {base_executor.py:158} [2021-03-26 16:34:52,326] [DEBUG] Syncing KubernetesExecutor {kubernetes_executor.py:337} [2021-03-26 16:34:52,326] [DEBUG] KubeJobWatcher alive, continuing {kubernetes_executor.py:263} ``` -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. For queries about this service, please contact Infrastructure at: us...@infra.apache.org