Hi all, I've read the docs, I've looked at the source for s3Hook, and I've brute forced different combinations in my config file but I just can't get this working so I'm hoping someone can give me some insite.
I need my logs to go into s3 storage, but no matter how I configure it my s3 bucket remains empty, and I never see any errors in my airflow scheduler or workers regarding issues connection, or that it's even attempted to connect. Pip install airflow[s3] Set up a connection in web UI called s3_conn, with type S3, and extra set to {"aws_access_key_id": "mykey", "aws_secret_access_key": "mykey"} In airflow config I set the following remote_base_log_folder = s3://bucket/ (I've tried with and without the trailing slash) remote_log_conn_id = s3_conn encrypt_s3_logs = False Is there some other step that I'm missing? Thanks, Jason Kromm This email and any attachments may contain confidential and proprietary information of Blackboard that is for the sole use of the intended recipient. If you are not the intended recipient, disclosure, copying, re-distribution or other use of any of this information is strictly prohibited. Please immediately notify the sender and delete this transmission if you received this email in error.