michaelgaotile opened a new issue #22291:
URL: https://github.com/apache/airflow/issues/22291


   ### Apache Airflow version
   
   2.2.2
   
   ### What happened
   
   This is on Amazon MWAA Airflow
   Tasks are scheduled for @daily 0:00 UTC, 
   
   Sometimes the task will fail right away, logs show no reason. The second 
attempt will kick off at past 14:00 UTC, a 14 hour delay. 
   
   Log 1:
   *** Reading remote log from Cloudwatch log_group: xxxxx 
log_stream:xxxxxx/2022-03-14T00_00_00+00_00/1.log.
   [2022-03-15 00:05:03,281] {{logging_mixin.py:104}} INFO - [2022-03-15 
00:05:03,281] {{base_aws.py:368}} INFO - Airflow Connection: xxxxx
   [2022-03-15 00:05:03,308] {{logging_mixin.py:104}} INFO - [2022-03-15 
00:05:03,308] {{base_aws.py:179}} INFO - No credentials retrieved from 
Connection
   [2022-03-15 00:05:03,308] {{logging_mixin.py:104}} INFO - [2022-03-15 
00:05:03,308] {{base_aws.py:87}} INFO - Creating session with 
aws_access_key_id=None region_name=xx
   [2022-03-15 00:05:03,481] {{logging_mixin.py:104}} INFO - [2022-03-15 
00:05:03,481] {{base_aws.py:157}} INFO - role_arn is None
   [2022-03-15 00:05:03,848] {{taskinstance.py:877}} INFO - Dependencies all 
met for <TaskInstance: xxxxxxxxx[queued]>
   [2022-03-15 00:05:04,011] {{taskinstance.py:877}} INFO - Dependencies all 
met for <TaskInstance: x2022-03-14T00:00:00+00:00 xxxxxxxxxxx[queued]>
   [2022-03-15 00:05:04,011] {{taskinstance.py:1068}} INFO - 
   
--------------------------------------------------------------------------------
   [2022-03-15 00:05:04,012] {{taskinstance.py:1069}} INFO - Starting attempt 1 
of 4
   [2022-03-15 00:05:04,012] {{taskinstance.py:1070}} INFO - 
   
--------------------------------------------------------------------------------
   [2022-03-15 00:05:04,100] {{taskinstance.py:1089}} INFO - Executing 
<Task(DatabricksRunNowOperator): xxxxxxxxx> on 2022-03-14T00:00:00+00:00
   [2022-03-15 00:05:04,152] {{standard_task_runner.py:52}} INFO - Started 
process 82 to run task
   
   
   Log 2:
   
   *** Reading remote log from Cloudwatch log_group:xxxxxxxxlog_stream: 
xxxxxxxxt/2022-03-14T00_00_00+00_00/2.log.
   [2022-03-15 14:12:00,700] {{logging_mixin.py:104}} INFO - [2022-03-15 
14:12:00,699] {{base_aws.py:368}} INFO - Airflow Connection: 
aws_conn_id=aws_default
   [2022-03-15 14:12:00,709] {{logging_mixin.py:104}} INFO - [2022-03-15 
14:12:00,709] {{base_aws.py:179}} INFO - No credentials retrieved from 
Connection
   [2022-03-15 14:12:00,709] {{logging_mixin.py:104}} INFO - [2022-03-15 
14:12:00,709] {{base_aws.py:87}} INFO - Creating session with 
aws_access_key_id=None region_name=xxxx
   [2022-03-15 14:12:00,725] {{logging_mixin.py:104}} INFO - [2022-03-15 
14:12:00,725] {{base_aws.py:157}} INFO - role_arn is None
   [2022-03-15 14:12:00,851] {{taskinstance.py:877}} INFO - Dependencies all 
met for <TaskInstance: xxxxxxxxxxxxx 2022-03-14T00:00:00+00:00 [queued]>
   [2022-03-15 14:12:00,871] {{taskinstance.py:877}} INFO - Dependencies all 
met for <TaskInstance: xxxxxxxxxxx2022-03-14T00:00:00+00:00 [queued]>
   [2022-03-15 14:12:00,871] {{taskinstance.py:1068}} INFO - 
   
--------------------------------------------------------------------------------
   [2022-03-15 14:12:00,871] {{taskinstance.py:1069}} INFO - Starting attempt 2 
of 4
   [2022-03-15 14:12:00,872] {{taskinstance.py:1070}} INFO - 
   
--------------------------------------------------------------------------------
   [2022-03-15 14:12:00,883] {{taskinstance.py:1089}} INFO - Executing 
<Task(DatabricksRunNowOperator): xxx> on 2022-03-14T00:00:00+00:00
   [2022-03-15 14:12:00,886] {{standard_task_runner.py:52}} INFO - Started 
process 16758 to run task
   [2022-03-15 14:12:00,980] {{logging_mixin.py:104}} INFO - Running 
<TaskInstance: xxxxxxxxxxxxxxxx 2022-03-14T00:00:00+00:00 [running]> on host 
xxxxxxx
   [2022-03-15 14:12:01,041] {{taskinstance.py:1283}} INFO - Exporting the 
following env vars:
   AIRFLOW_CTX_DAG_OWNER=xxxxxxx
   AIRFLOW_CTX_DAG_ID=xxxxxxxxx
   AIRFLOW_CTX_TASK_ID=xxxxxxxxxxx
   AIRFLOW_CTX_EXECUTION_DATE=2022-03-14T00:00:00+00:00
   AIRFLOW_CTX_DAG_RUN_ID=scheduled__2022-03-14T00:00:00+00:00
   [2022-03-15 14:12:01,058] {{logging_mixin.py:104}} INFO - [2022-03-15 
14:12:01,058] {{base.py:78}} INFO - Using connection to: id: databricks. 
Host:xxxxxxxxx, Port: None, Schema: , Login: , Password: 
   [2022-03-15 14:12:01,059] {{logging_mixin.py:104}} INFO - [2022-03-15 
14:12:01,059] {{databricks.py:170}} INFO - Using token auth. 
   [2022-03-15 14:12:01,241] {{databricks.py:73}} INFO - Run submitted with 
run_id: 5177061
   [2022-03-15 14:12:01,242] {{logging_mixin.py:104}} INFO - [2022-03-15 
14:12:01,242] {{databricks.py:170}} INFO - Using token auth. 
   [2022-03-15 14:12:01,439] {{databricks.py:78}} INFO - View run status, Spark 
UI, and logs at https:/xxxxxxxxxxxx
   
   ### What you think should happen instead
   
   I expect the task to start at the correct time, and in the event of failure, 
to restart at the specified restart interval, not 14 hours later. 
   
   ### How to reproduce
   
   Seems to happen randomly across different jobs.
   
   ### Operating System
   
   MWAA
   
   ### Versions of Apache Airflow Providers
   
   _No response_
   
   ### Deployment
   
   MWAA
   
   ### Deployment details
   
   mw1.medium
   max worker count = 20
   min worker count = 1
   scheduler count = 4
   
   ### Anything else
   
   _No response_
   
   ### Are you willing to submit PR?
   
   - [ ] Yes I am willing to submit a PR!
   
   ### Code of Conduct
   
   - [X] I agree to follow this project's [Code of 
Conduct](https://github.com/apache/airflow/blob/main/CODE_OF_CONDUCT.md)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


Reply via email to