Max Feng created FLINK-37483:
--------------------------------
Summary: Native kubernetes clusters losing checkpoint state on
FAILED
Key: FLINK-37483
URL: https://issues.apache.org/jira/browse/FLINK-37483
Project: Flink
Issue Type: Bug
Components: Deployment / Kubernetes
Affects Versions: 1.20.1
Reporter: Max Feng
We're running Flink 1.20, native kubernetes application-mode clusters, and
we're running into an issue where clusters are restarting without checkpoints
from HA configmaps.
To the best of our understanding, here's what's happening:
1) We're running application-mode clusters in native kubernetes with
externalized checkpoints, retained on cancellation. We're attempting to restore
a job from a checkpoint; the checkpoint reference is held in the Kubernetes HA
configmap.
2) The job goes to state FAILED.
3) The HA configmap containing the checkpoint reference is cleaned up
4) The Kubernetes pod exits. Because it is a Kubernetes deployment, the pod is
immediately restarted.
5) Upon restart, the new Jobmanager finds no checkpoints to restore from.
We think this is a bad combination of the following behaviors:
* FAILED triggers cleanup, which cleans up HA configmaps in native kubernetes
mode
* FAILED does not actually stop a job in native kubernetes mode, instead it is
immediately retried
--
This message was sent by Atlassian Jira
(v8.20.10#820010)