Aidar Mamytov created AIRFLOW-5935:
--------------------------------------
Summary: Logs are not sent S3 by S3TaskHandler
Key: AIRFLOW-5935
URL: https://issues.apache.org/jira/browse/AIRFLOW-5935
Project: Apache Airflow
Issue Type: Task
Components: logging
Affects Versions: 1.10.6
Reporter: Aidar Mamytov
Assignee: Aidar Mamytov
When exactly is S3TaskHandler supposed to have its *s3_write* or *close* method
called? The logs are written locally but are not appearing in S3.I've
pdb-debugged my custom log_config.py file and Airflow reads configs
successfully and loads *S3TaskHandler* configs successfully. I also
pdb-debugged and checked another thing with print statements - whenever I try
to open "_View Log_" for any task in the admin dashboard, it definitely calls
*S3TaskHandler.s3_read* and *S3TaskHandler.s3_log_exists* and successfully
connects to S3. I also checked if Airflow is able to connect to S3 in Python
console: imported *S3Hook* and *S3TaskHandler* and tried to connect to S3, read
objects and write new ones to my bucket - all good.
The problem is that although Airflow is able to connect to S3 bucket and
interact with it with read/write operations, it just does not upload logs to
it. What might I do wrong or what do I not understand about airflow remote
logging?
--
This message was sent by Atlassian Jira
(v8.3.4#803005)