Re: Canceling a DAG run

2017-02-28 Thread Maxime Beauchemin
Changing the state of a dag run in the UI will only stop the scheduler from evaluating that dag run, it does not sigterm or anything like that. "Clearing" running task instances (UI/CLI) will result in failing the task instances. Also as of 1.8.x if the state of the task in changed externally from

Task instance enqueued but never runs, holds up next run of the DAG

2017-02-28 Thread Vijay Ramesh
I have a large DAG (32 tasks) with concurrency=2 and max_active_runs=1. Most of the tasks also use a redshift_pool, and this is running the LocalExecutor on 1.8.0RC4. When the DAG kicks off things seem to generally function, but a few of the tasks get moved to queued status (appropriately) but

Re: Article: Airflow and the Future of Data Engineering: A Q

2017-02-28 Thread Taylor Edmiston
Thanks for the share, Max! I really enjoyed working on and co-publishing the post. I know everyone is quite busy with the 1.8 release, so I especially appreciate taking the time to spread the word and think about the big picture. I'm happy to field any questions on the post or Astronomer. You

Article: Airflow and the Future of Data Engineering: A Q

2017-02-28 Thread Maxime Beauchemin
Jointly publishing this short Q with Taylor D. Edmiston of Atronomer.io https://medium.com/the-astronomer-journey/airflow-and-the-future-of-data-engineering-a-q-a-266f68d956a9#.cpast6vjd

Canceling a DAG run

2017-02-28 Thread Arpan Dhandhania
Hi folks, We have been using Airflow to run ETL jobs but sometimes require to cancel a DAG run. Since I could not find a way to cancel a DAG run (from CLI or UI), we are using a crude work-around that is to cancel the tasks that are running and thus failing the DAG run. What I expect (and I

Re: Failing jobs not properly terminated

2017-02-28 Thread EKC (Erik Cederstrand)
Here's a follow-up with more information. I think my suspicion is supported by the Airflow database. It has the task marked as failed in the task_instance table, and the PID matches the dead job: airflow=> select job_id, task_id, dag_id, state, pid from task_instance where job_id=127022;

Failing jobs not properly terminated

2017-02-28 Thread EKC (Erik Cederstrand)
Hi folks, We're experiencing an issue where a failing DAG is not properly terminated. The log file contains a Python stack trace of the failure and then ends with: {base_task_runner.py:112} INFO - Running: ['bash', '-c', 'airflow run my_dag my_task 2017-02-22T00:00:00 --job_id 127022

Re: [RESULT] [VOTE] Release Airflow 1.8.0 based on Airflow 1.8.0rc4

2017-02-28 Thread Bolke de Bruin
Gotcha: It works, but then slightly different. If you added the subdag, do not zoom in, but click on the subdag in the main dag. Use mark success there. It will then allow you to mark all tasks successful that are part of the subdag. Do we still consider this a blocker? Imho, no as a