thiagophx opened a new issue #11936:
URL: https://github.com/apache/airflow/issues/11936


   **Apache Airflow version**: 1.10.12
   
   
   **Kubernetes version (if you are using kubernetes)** (use `kubectl 
version`): 1.10
   
   **Environment**: 
   
   - **Cloud provider or hardware configuration**: AWS
   - **OS** (e.g. from /etc/os-release): Debian GNU/Linux 9 (stretch)
   - **Kernel** (e.g. `uname -a`): 4.19.123-coreos
   
   **What happened**:
   
   After rolling out to Airflow 1.10.12, we've started to notice a huge 
increase on loading up DAG files. Some of our DAGs started timing out, and we 
see the following error on the logs:
   ```
   Processor for /airflow/airflow-dags/dags/my_dag.py with PID 27954 started at 
2020-10-29T02:21:04.630540+00:00 has timed out, killing it
   Processor for /airflow/airflow-dags/dags/my_dag.py exited with return code -9
   ```
   
   **What you expected to happen**:
   
   We didn't have this issue on the 1.10.9. Our DAGs load time went from 1 
second to over 40 seconds, and therefore going over the 
`dag_file_processor_timeout` setting.
   
   **How to reproduce it**:
   Any DAG with more than 200 tasks seem to be affected. On our case we have 
one DAG full of Sensors to other DAGs, but the issue doesn't seem to be related 
to a particular type of task, but rather the total size of the DAG.
   
   **Anything else we need to know**:
   
   The problem occurs every time the scheduled time is reached, so that the 
tasks are created, and the process times out.


----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to