I'm experiencing the same issue. I've built a simple DAG with no external
dependencies other than bash that illustrates the problem consistently on
my machine, find it here:
https://gist.github.com/slotrans/b3e475c2b9789c4efc41876567902425
If you run it as e.g. airflow backfill tasks_never_run -s 2017-06-06 -e
2017-06-06 you should see some tasks permanently remain in a state of "no
status". Restarting the scheduler will not help. Ctrl-C-ing the backfill
command and running it again *may* resolve it. The scheduler will
continually log messages like the following:
[2017-06-05 18:42:49,372] {jobs.py:1408} INFO - Heartbeating the process
manager
[2017-06-05 18:42:49,375] {dag_processing.py:559} INFO - Processor for
/Users/noah.yetter/airflow/dags/tasks_never_run.py finished
[2017-06-05 18:42:49,428] {jobs.py:1007} INFO - Tasks up for execution:
<TaskInstance: tasks_never_run.t_leaf_five__axe 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_five__arrows 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_two__determined 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_two__beyond 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_two__axe 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_two__arrows 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_three__determined 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_three__beyond 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_three__arrows 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_one__beyond 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_one__arrows 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_four__determined 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_four__arrows 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_five__beyond 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_one__determined 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_one__axe 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_five__determined 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_three__axe 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_four__axe 2017-06-05 00:00:00
[scheduled]>
<TaskInstance: tasks_never_run.t_leaf_four__beyond 2017-06-05 00:00:00
[scheduled]>
[2017-06-05 18:42:49,430] {jobs.py:1030} INFO - Figuring out tasks to run
in Pool(name=None) with 128 open slots and 20 task instances in queue
[2017-06-05 18:42:49,463] {jobs.py:1444} INFO - Heartbeating the executor
On 2017-06-01 15:18 (-0600), "Josef [email protected]> wrote:
> Hi!>
>
> We have a problem with our airflow. Sometimes, several tasks get queued
but they never get run and remain in Queud state forever. Other tasks from
the same schedule interval run. And next schedule interval runs normally
too. But these several tasks remain queued.>
>
> We are using Airflow 1.8.1. Currently with CeleryExecutor and redis, but
we had the same problem with LocalExecutor as well (actually switching to
Celery helped quite a bit, the problem now happens way less often, but
still it happens). We have 18 DAGs total, 13 active. Some have just 1-2
tasks, but some are more complex, like 8 tasks or so and with upstreams.
There are also ExternalTaskSensor tasks used. >
>
> I tried playing around with DAG configurations (limiting concurrency,
max_active_runs, ...), tried switching off some DAGs completely (not all
but most) etc., so far nothing helped. Right now, I am not really sure,
what else to try to identify a solve the issue.>
>
> I am getting a bit desperate, so I would really appreciate any help with
this. Thank you all in advance!>
>
> Joe>
>