kajanajustin opened a new issue, #32167:
URL: https://github.com/apache/airflow/issues/32167
### Apache Airflow version
Other Airflow 2 version (please specify below)
### What happened
Hi Team,
Currently, I am running docker-compose using the airflow:2.5.3-python3.8
image. Currently, when I go running my dag everything looks good except I am
getting these ugly asterisks everywhere the word 'airflow' is supposed to be. I
double-checked my config and paths and still have the same outcome. Please let
me know if there is something I may be doing incorrectly.
here are the logs below
```
*** Reading local file:
/opt/airflow/logs/dag_id=heartbeat/run_id=manual__2023-06-27T00:25:10.808687+00:00/task_id=heartbeat/attempt=1.log
[2023-06-26, 20:25:22 EDT] {taskinstance.py:1090} INFO - Dependencies all
met for dep_context=non-requeueable deps ti=<TaskInstance: heartbeat.heartbeat
manual__2023-06-27T00:25:10.808687+00:00 [queued]>
[2023-06-26, 20:25:22 EDT] {taskinstance.py:1090} INFO - Dependencies all
met for dep_context=requeueable deps ti=<TaskInstance: heartbeat.heartbeat
manual__2023-06-27T00:25:10.808687+00:00 [queued]>
[2023-06-26, 20:25:22 EDT] {taskinstance.py:1288} INFO -
--------------------------------------------------------------------------------
[2023-06-26, 20:25:22 EDT] {taskinstance.py:1289} INFO - Starting attempt 1
of 1
[2023-06-26, 20:25:22 EDT] {taskinstance.py:1290} INFO -
--------------------------------------------------------------------------------
[2023-06-26, 20:25:22 EDT] {taskinstance.py:1309} INFO - Executing
<Task(PipelineOperator): heartbeat> on 2023-06-27 00:25:10.808687+00:00
[2023-06-26, 20:25:22 EDT] {standard_task_runner.py:55} INFO - Started
process 21159 to run task
[2023-06-26, 20:25:22 EDT] {standard_task_runner.py:82} INFO - Running:
['***', 'tasks', 'run', 'heartbeat', 'heartbeat',
'manual__2023-06-27T00:25:10.808687+00:00', '--job-id', '26', '--raw',
'--subdir', 'DAGS_FOLDER/heartbeat.py', '--cfg-path', '/tmp/tmpp2vbdho7']
[2023-06-26, 20:25:22 EDT] {standard_task_runner.py:83} INFO - Job 26:
Subtask heartbeat
[2023-06-26, 20:25:23 EDT] {task_command.py:389} INFO - Running
<TaskInstance: heartbeat.heartbeat manual__2023-06-27T00:25:10.808687+00:00
[running]> on host c8908875256e
[2023-06-26, 20:25:23 EDT] {taskinstance.py:1516} INFO - Exporting the
following env vars:
AIRFLOW_CTX_DAG_OWNER=***
AIRFLOW_CTX_DAG_ID=heartbeat
AIRFLOW_CTX_TASK_ID=heartbeat
AIRFLOW_CTX_EXECUTION_DATE=2023-06-27T00:25:10.808687+00:00
AIRFLOW_CTX_TRY_NUMBER=1
AIRFLOW_CTX_DAG_RUN_ID=manual__2023-06-27T00:25:10.808687+00:00
[2023-06-26, 20:25:23 EDT] {operators.py:514} INFO - Using common lib
version [1.1.91].
[2023-06-26, 20:25:23 EDT] {operators.py:171} INFO - Building end_interval
using data_interval_end: 2023-06-27T00:25:00+00:00, data_interval_type :
SINCE_LAST_SUCCESS
[2023-06-26, 20:25:23 EDT] {operators.py:179} INFO - end_interval =
2023-06-27T00:25:00+00:00
[2023-06-26, 20:25:23 EDT] {logging_mixin.py:137} INFO - Run id for the
Lineage API is being set to: 88562ae5-3511-37a6-a361-59ebb4b3e111
[2023-06-26, 20:25:23 EDT] {operators.py:139} INFO - Building start_interval
using data_interval_start: 2023-06-27T00:20:00+00:00,
prev_data_interval_end_success : None,data_interval_type : SINCE_LAST_SUCCESS
[2023-06-26, 20:25:23 EDT] {operators.py:166} INFO - start_interval =
2023-06-27T00:20:00+00:00
[2023-06-26, 20:25:23 EDT] {operators.py:171} INFO - Building end_interval
using data_interval_end: 2023-06-27T00:25:00+00:00, data_interval_type :
SINCE_LAST_SUCCESS
[2023-06-26, 20:25:23 EDT] {operators.py:179} INFO - end_interval =
2023-06-27T00:25:00+00:00
[2023-06-26, 20:25:23 EDT] {operators.py:611} INFO - Bash command
configuration: /opt/***/run-pipeline-shared.sh general heartbeat/heartbeat.py
```
All of my other airflow servers don't have this issue, and I have checked to
ensure that the `sensitive_var_conn_names = False` Not sure what else to do at
this point in time....
### What you think should happen instead
Instead, the *** should be shown as `airflow`
### How to reproduce
Utilizing docker-compose and the airflow:2.5.3-python3.8 image, essentially
use any basic dag and have airflow pick up on it then go to logs and see the
asterisks
### Operating System
Debian GNU/Linux 11 (bullseye)
### Versions of Apache Airflow Providers
_No response_
### Deployment
Docker-Compose
### Deployment details
_No response_
### Anything else
_No response_
### Are you willing to submit PR?
- [X] Yes I am willing to submit a PR!
### Code of Conduct
- [X] I agree to follow this project's [Code of
Conduct](https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md)
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]