vapiravfif commented on issue #16023:
URL: https://github.com/apache/airflow/issues/16023#issuecomment-877767772
@Jorricks Hi, it might be helpful to know that I couldn't reproduce the
issue on a DAG that was previously active, but it reproduced like a charm on a
brand new DAG with same setup. Both paused at the time of running the tasks.
I can't find anything related to either `Setting external_id for` or
`SchedulerJob instances as failed` in the logs for both instances (on DEBUG)
Here's a `grep -A 5 -B 5` from logs related to the task name (there wasn't
related logs on webserver...):
```
Scheduler:
[2021-07-11 08:28:08,961] {dag_processing.py:385} DEBUG - Received message
of type DagParsingStat
[2021-07-11 08:28:08,972] {scheduler_job.py:1854} INFO - Resetting orphaned
tasks for active dag runs
[2021-07-11 08:28:08,972] {scheduler_job.py:1862} DEBUG - Running
SchedulerJob.adopt_or_reset_orphaned_tasks with retries. Try 1 of 3
[2021-07-11 08:28:08,972] {scheduler_job.py:1864} DEBUG - Calling
SchedulerJob.adopt_or_reset_orphaned_tasks method
[2021-07-11 08:28:09,003] {scheduler_job.py:1921} INFO - Reset the following
1 orphaned TaskInstances:
<TaskInstance: timeout_testing.sleep_operator_task 2021-07-11
08:02:31.293769+00:00 [running]>
[2021-07-11 08:28:09,020] {scheduler_job.py:1399} DEBUG - Next timed event
is in 0.645959
[2021-07-11 08:28:09,020] {scheduler_job.py:1401} DEBUG - Ran scheduling
loop in 0.13 seconds
[2021-07-11 08:28:09,205] {settings.py:292} DEBUG - Disposing DB connection
pool (PID 1178)
[2021-07-11 08:28:09,223] {scheduler_job.py:310} DEBUG - Waiting for
<ForkProcess(DagFileProcessor921-Process, stopped)>
[2021-07-11 08:28:09,469] {settings.py:292} DEBUG - Disposing DB connection
pool (PID 1182)
Worker:
[2021-07-11 08:23:47,559: DEBUG/MainProcess] pidbox received method
enable_events() [reply_to:None ticket:None]
[2021-07-11 08:23:52,559: DEBUG/MainProcess] pidbox received method
enable_events() [reply_to:None ticket:None]
[2021-07-11 08:23:57,559: DEBUG/MainProcess] pidbox received method
enable_events() [reply_to:None ticket:None]
[2021-07-11 08:24:02,558: DEBUG/MainProcess] pidbox received method
enable_events() [reply_to:None ticket:None]
[2021-07-11 08:24:07,132: INFO/MainProcess] Received task:
airflow.executors.celery_executor.execute_command[ad28b4dc-6a78-4821-9a02-998aff8156b2]
[2021-07-11 08:24:07,133: DEBUG/MainProcess] TaskPool: Apply <function
_fast_trace_task at 0x7feff474d3b0>
(args:('airflow.executors.celery_executor.execute_command',
'ad28b4dc-6a78-4821-9a02-998aff8156b2', {'lang': 'py', 'task':
'airflow.executors.celery_executor.execute_command', 'id':
'ad28b4dc-6a78-4821-9a02-998aff8156b2', 'shadow': None, 'eta': None, 'expires':
None, 'group': None, 'group_index': None, 'retries': 0, 'timelimit': [None,
None], 'root_id': 'ad28b4dc-6a78-4821-9a02-998aff8156b2', 'parent_id': None,
'argsrepr': "[['airflow', 'tasks', 'run', 'timeout_testing',
'sleep_operator_task', '2021-07-11T08:02:31.293769+00:00', '--local', '--pool',
'default_pool', '--subdir', 'DAGS_FOLDER/timeout_testing.py']]", 'kwargsrepr':
'{}', 'origin': 'gen268@airflow-dev-web-6d79645c68-tzbnv', 'reply_to':
'a3aa91fd-479b-3d76-893e-7a8e8d23c454', 'correlation_id':
'ad28b4dc-6a78-4821-9a02-998aff8156b2', 'hostname':
'celery@airflow-dev-worker-0', 'delivery_info': {'exchange': '', 'routin
g_key': 'default', 'priority': 0, 'redelivered': None}, 'args': [['airflow',
'tasks', 'run', 'timeout_testing', 'sleep_operator_task',... kwargs:{})
[2021-07-11 08:24:07,134: DEBUG/MainProcess] Task accepted:
airflow.executors.celery_executor.execute_command[ad28b4dc-6a78-4821-9a02-998aff8156b2]
pid:45
--
[2021-07-11 08:24:07,134: DEBUG/MainProcess] Task accepted:
airflow.executors.celery_executor.execute_command[ad28b4dc-6a78-4821-9a02-998aff8156b2]
pid:45
[2021-07-11 08:24:07,186: INFO/ForkPoolWorker-15] Executing command in
Celery: ['airflow', 'tasks', 'run', 'timeout_testing', 'sleep_operator_task',
'2021-07-11T08:02:31.293769+00:00', '--local', '--pool', 'default_pool',
'--subdir', 'DAGS_FOLDER/timeout_testing.py']
[2021-07-11 08:24:07,329: DEBUG/ForkPoolWorker-15] Calling callbacks:
[<function default_action_log at 0x7ff0008deb90>]
[2021-07-11 08:24:07,350: DEBUG/ForkPoolWorker-15] Setting up DB connection
pool (PID 47)
[2021-07-11 08:24:07,351: DEBUG/ForkPoolWorker-15]
settings.prepare_engine_args(): Using NullPool
[2021-07-11 08:24:07,353: INFO/ForkPoolWorker-15] Filling up the DagBag from
...dags_dev/timeout_testing.py
[2021-07-11 08:24:07,354: DEBUG/ForkPoolWorker-15] Importing
...dags_dev/timeout_testing.py
--
--
[2021-07-11 08:24:07,357: DEBUG/ForkPoolWorker-15] Loaded DAG <DAG:
timeout_testing>
[2021-07-11 08:24:07,399: DEBUG/ForkPoolWorker-15] Loading plugins
[2021-07-11 08:24:07,399: DEBUG/ForkPoolWorker-15] Loading plugins from
directory: /opt/airflow/plugins
[2021-07-11 08:24:07,399: DEBUG/ForkPoolWorker-15] Loading plugins from
entrypoints
[2021-07-11 08:24:07,473: DEBUG/ForkPoolWorker-15] Integrate DAG plugins
[2021-07-11 08:24:07,501: WARNING/ForkPoolWorker-15] Running <TaskInstance:
timeout_testing.sleep_operator_task 2021-07-11T08:02:31.293769+00:00 [None]> on
host airflow-dev-worker-0.airflow-dev-worker.airflow-dev.svc.cluster.local
[2021-07-11 08:24:07,559: DEBUG/MainProcess] pidbox received method
enable_events() [reply_to:None ticket:None]
[2021-07-11 08:24:12,562: DEBUG/MainProcess] pidbox received method
enable_events() [reply_to:None ticket:None]
[2021-07-11 08:24:17,559: DEBUG/MainProcess] pidbox received method
enable_events() [reply_to:None ticket:None]
[2021-07-11 08:24:22,563: DEBUG/MainProcess] pidbox received method
enable_events() [reply_to:None ticket:None]
[2021-07-11 08:24:27,559: DEBUG/MainProcess] pidbox received method
enable_events() [reply_to:None ticket:None]
--
```
I cannot share the full logs as I'm testing in real environment and there
are many S3 requests that need hashing, but I saved the logs and happy to
provide any part of them as needed. Thank you!
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]