[ 
https://issues.apache.org/jira/browse/AIRFLOW-1667?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16358757#comment-16358757
 ] 

Ash Berlin-Taylor commented on AIRFLOW-1667:
--------------------------------------------

The process that writes to the log files is a sub-process of the celery worker 
itself – that just invokes {{airflow run --local}} - and that means the flush 
should happen as soon the task instance finishes running.

I do not see this behaivour on Py3/1.9.0 - our tasks appear in S3 when the task 
instance is finished. Are you saying you have to stop the {{airflow worker}} 
process for the logs to appear in S3?

> Remote log handlers don't upload logs on task finish
> ----------------------------------------------------
>
>                 Key: AIRFLOW-1667
>                 URL: https://issues.apache.org/jira/browse/AIRFLOW-1667
>             Project: Apache Airflow
>          Issue Type: Bug
>          Components: logging
>    Affects Versions: 1.9.0, 1.10.0
>            Reporter: Arthur Vigil
>            Priority: Major
>
> AIRFLOW-1385 revised logging for configurability, but the provided remote log 
> handlers (S3TaskHandler and GCSTaskHandler) only upload on close (flush is 
> left at the default implementation provided by `logging.FileHandler`). A 
> handler will be closed on process exit by `logging.shutdown()`, but depending 
> on the Executor used worker processes may not regularly shutdown, and can 
> very likely persist between tasks. This means during normal execution log 
> files are never uploaded.
> Need to find a way to flush remote log handlers in a timely manner, but 
> without hitting the target resources unnecessarily.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to