Gyula Fora created FLINK-27802: ---------------------------------- Summary: Job submission errors are swallowed for Flink 1.15 + HA Key: FLINK-27802 URL: https://issues.apache.org/jira/browse/FLINK-27802 Project: Flink Issue Type: Improvement Reporter: Gyula Fora
We are currently setting both a result store and the "execution.submit-failed-job-on-application-error" config for HA jobs. This leads to swallowed job submission errors that only show up in the result store, but the flink job is not actually displayed in the failed state: 2022-05-26 12:34:43,497 WARN org.apache.flink.runtime.dispatcher.StandaloneDispatcher [] - Ignoring JobGraph submission 'State machine job' (00000000000000000000000000000000) because the job already reached a globally-terminal state (i.e. FAILED, CANCELED, FINISHED) in a previous execution. 2022-05-26 12:34:43,552 INFO org.apache.flink.client.deployment.application.ApplicationDispatcherBootstrap [] - Application completed SUCCESSFULLY The easiest way to reproduce this is to create a new deployment and set initialSavepointPath to a random missing path. I consider this a bug in Flink but we should simply disable the execution.submit-failed-job-on-application-error config. -- This message was sent by Atlassian Jira (v8.20.7#820007)