Aidar Mamytov created AIRFLOW-5935:
--------------------------------------

             Summary: Logs are not sent S3 by S3TaskHandler
                 Key: AIRFLOW-5935
                 URL: https://issues.apache.org/jira/browse/AIRFLOW-5935
             Project: Apache Airflow
          Issue Type: Task
          Components: logging
    Affects Versions: 1.10.6
            Reporter: Aidar Mamytov
            Assignee: Aidar Mamytov


When exactly is S3TaskHandler supposed to have its *s3_write* or *close* method 
called? The logs are written locally but are not appearing in S3.I've 
pdb-debugged my custom log_config.py file and Airflow reads configs 
successfully and loads *S3TaskHandler* configs successfully. I also 
pdb-debugged and checked another thing with print statements - whenever I try 
to open "_View Log_" for any task in the admin dashboard, it definitely calls 
*S3TaskHandler.s3_read* and *S3TaskHandler.s3_log_exists* and successfully 
connects to S3. I also checked if Airflow is able to connect to S3 in Python 
console: imported *S3Hook* and *S3TaskHandler* and tried to connect to S3, read 
objects and write new ones to my bucket - all good.

The problem is that although Airflow is able to connect to S3 bucket and 
interact with it with read/write operations, it just does not upload logs to 
it. What might I do wrong or what do I not understand about airflow remote 
logging?



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

Reply via email to