Gyula Fora created FLINK-27802:
----------------------------------

             Summary: Job submission errors are swallowed for Flink 1.15 + HA
                 Key: FLINK-27802
                 URL: https://issues.apache.org/jira/browse/FLINK-27802
             Project: Flink
          Issue Type: Improvement
            Reporter: Gyula Fora


We are currently setting both a result store and the 
"execution.submit-failed-job-on-application-error" config for HA jobs.

This leads to swallowed job submission errors that only show up in the result 
store, but the flink job is not actually displayed in the failed state:


2022-05-26 12:34:43,497 WARN 
org.apache.flink.runtime.dispatcher.StandaloneDispatcher [] - Ignoring JobGraph 
submission 'State machine job' (00000000000000000000000000000000) because the 
job already reached a globally-terminal state (i.e. FAILED, CANCELED, FINISHED) 
in a previous execution.
2022-05-26 12:34:43,552 INFO 
org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap 
[] - Application completed SUCCESSFULLY


The easiest way to reproduce this is to create a new deployment and set 
initialSavepointPath to a random missing path.

I consider this a bug in Flink but we should simply disable the 
execution.submit-failed-job-on-application-error config.



--
This message was sent by Atlassian Jira
(v8.20.7#820007)

Reply via email to