gopidesupavan opened a new issue, #50583:
URL: https://github.com/apache/airflow/issues/50583
### Body
When remote logging enabled, am seeing connection errors in scheduler.
```
[2025-05-14T07:38:05.993+0000] {base_aws.py:609} WARNING - Unable to find
AWS Connection ID 's3_conn', switching to empty.
[2025-05-14T07:38:05.995+0000] {base_aws.py:185} INFO - No connection ID
provided. Fallback on boto3 credential strategy (region_name=None). See:
https://boto3.amazonaws.com/v1/documentation
/api/latest/guide/configuration.html
[2025-05-14T07:38:06.273+0000] {scheduler_job_runner.py:450} INFO - 1 tasks
up for execution:
<TaskInstance: example_xcom_args.print_value
manual__2025-05-14T07:38:02.428050+00:00 [scheduled]>
[2025-05-14T07:38:06.274+0000] {scheduler_job_runner.py:522} INFO - DAG
example_xcom_args has 0/16 running and queued tasks
[2025-05-14T07:38:06.275+0000] {scheduler_job_runner.py:661} INFO - Setting
the following tasks to queued state:
<TaskInstance: example_xcom_args.print_value
manual__2025-05-14T07:38:02.428050+00:00 [scheduled]>
[2025-05-14T07:38:06.280+0000] {scheduler_job_runner.py:767} INFO - Trying
to enqueue tasks: [<TaskInstance: example_xcom_args.print_value
manual__2025-05-14T07:38:02.428050+00:00 [scheduled
]>] for executor: LocalExecutor(parallelism=32)
[2025-05-14T07:38:08.936+0000] {s3_task_handler.py:124} ERROR - Could not
verify previous log to append
Traceback (most recent call last):
File
"/opt/airflow/providers/amazon/src/airflow/providers/amazon/aws/log/s3_task_handler.py",
line 120, in write
if append and self.s3_log_exists(remote_log_location):
File
"/opt/airflow/providers/amazon/src/airflow/providers/amazon/aws/log/s3_task_handler.py",
line 81, in s3_log_exists
return self.hook.check_for_key(remote_log_location)
File
"/opt/airflow/providers/amazon/src/airflow/providers/amazon/aws/hooks/s3.py",
line 153, in wrapper
return func(*bound_args.args, **bound_args.kwargs)
File
"/opt/airflow/providers/amazon/src/airflow/providers/amazon/aws/hooks/s3.py",
line 126, in wrapper
return func(*bound_args.args, **bound_args.kwargs)
File
"/opt/airflow/providers/amazon/src/airflow/providers/amazon/aws/hooks/s3.py",
line 953, in check_for_key
obj = self.head_object(key, bucket_name)
File
"/opt/airflow/providers/amazon/src/airflow/providers/amazon/aws/hooks/s3.py",
line 153, in wrapper
return func(*bound_args.args, **bound_args.kwargs)
File
"/opt/airflow/providers/amazon/src/airflow/providers/amazon/aws/hooks/s3.py",
line 126, in wrapper
return func(*bound_args.args, **bound_args.kwargs)
File
"/opt/airflow/providers/amazon/src/airflow/providers/amazon/aws/hooks/s3.py",
line 934, in head_object
return self.get_conn().head_object(Bucket=bucket_name, Key=key)
File "/usr/local/lib/python3.9/site-packages/botocore/client.py", line
569, in _api_call
return self._make_api_call(operation_name, kwargs)
File "/usr/local/lib/python3.9/site-packages/botocore/client.py", line
1005, in _make_api_call
http, parsed_response = self._make_request(
```
To produce:
Set the remote logging enabled in airflow.cfg and connection id.
### Committer
- [x] I acknowledge that I am a maintainer/committer of the Apache Airflow
project.
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]