[ https://issues.apache.org/jira/browse/AIRFLOW-1667?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16358757#comment-16358757 ]
Ash Berlin-Taylor commented on AIRFLOW-1667: -------------------------------------------- The process that writes to the log files is a sub-process of the celery worker itself – that just invokes {{airflow run --local}} - and that means the flush should happen as soon the task instance finishes running. I do not see this behaivour on Py3/1.9.0 - our tasks appear in S3 when the task instance is finished. Are you saying you have to stop the {{airflow worker}} process for the logs to appear in S3? > Remote log handlers don't upload logs on task finish > ---------------------------------------------------- > > Key: AIRFLOW-1667 > URL: https://issues.apache.org/jira/browse/AIRFLOW-1667 > Project: Apache Airflow > Issue Type: Bug > Components: logging > Affects Versions: 1.9.0, 1.10.0 > Reporter: Arthur Vigil > Priority: Major > > AIRFLOW-1385 revised logging for configurability, but the provided remote log > handlers (S3TaskHandler and GCSTaskHandler) only upload on close (flush is > left at the default implementation provided by `logging.FileHandler`). A > handler will be closed on process exit by `logging.shutdown()`, but depending > on the Executor used worker processes may not regularly shutdown, and can > very likely persist between tasks. This means during normal execution log > files are never uploaded. > Need to find a way to flush remote log handlers in a timely manner, but > without hitting the target resources unnecessarily. -- This message was sent by Atlassian JIRA (v7.6.3#76005)