[ 
https://issues.apache.org/jira/browse/FLINK-37483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Max Feng updated FLINK-37483:
-----------------------------
    Description: 
We're running Flink 1.20, native kubernetes application-mode clusters, and 
we're running into an issue where clusters are restarting without checkpoints 
from HA configmaps.

To the best of our understanding, here's what's happening: 

1) We're running application-mode clusters in native kubernetes with 
externalized checkpoints, retained on cancellation. We're attempting to restore 
a job from a checkpoint; the checkpoint reference is held in the Kubernetes HA 
configmap. 
2) The jobmanager encounters an issue during startup, and the job goes to state 
FAILED.
3) The HA configmap containing the checkpoint reference is cleaned up.
4) The Kubernetes pod exits. Because it is a Kubernetes deployment, the pod is 
immediately restarted. 
5) Upon restart, the new Jobmanager finds no checkpoints to restore from.

We think this is a bad combination of the following behaviors:
* FAILED triggers cleanup, which cleans up HA configmaps in native kubernetes 
mode
* FAILED does not actually stop a job in native kubernetes mode, instead it is 
immediately retried

  was:
We're running Flink 1.20, native kubernetes application-mode clusters, and 
we're running into an issue where clusters are restarting without checkpoints 
from HA configmaps.

To the best of our understanding, here's what's happening: 

1) We're running application-mode clusters in native kubernetes with 
externalized checkpoints, retained on cancellation. We're attempting to restore 
a job from a checkpoint; the checkpoint reference is held in the Kubernetes HA 
configmap. 
2) The jobmanager encounters an issue during startup, and the job goes to state 
FAILED.
3) The HA configmap containing the checkpoint reference is cleaned up
4) The Kubernetes pod exits. Because it is a Kubernetes deployment, the pod is 
immediately restarted. 
5) Upon restart, the new Jobmanager finds no checkpoints to restore from.

We think this is a bad combination of the following behaviors:
* FAILED triggers cleanup, which cleans up HA configmaps in native kubernetes 
mode
* FAILED does not actually stop a job in native kubernetes mode, instead it is 
immediately retried


> Native kubernetes clusters losing checkpoint state on FAILED
> ------------------------------------------------------------
>
>                 Key: FLINK-37483
>                 URL: https://issues.apache.org/jira/browse/FLINK-37483
>             Project: Flink
>          Issue Type: Bug
>          Components: Deployment / Kubernetes
>    Affects Versions: 1.20.1
>            Reporter: Max Feng
>            Priority: Major
>
> We're running Flink 1.20, native kubernetes application-mode clusters, and 
> we're running into an issue where clusters are restarting without checkpoints 
> from HA configmaps.
> To the best of our understanding, here's what's happening: 
> 1) We're running application-mode clusters in native kubernetes with 
> externalized checkpoints, retained on cancellation. We're attempting to 
> restore a job from a checkpoint; the checkpoint reference is held in the 
> Kubernetes HA configmap. 
> 2) The jobmanager encounters an issue during startup, and the job goes to 
> state FAILED.
> 3) The HA configmap containing the checkpoint reference is cleaned up.
> 4) The Kubernetes pod exits. Because it is a Kubernetes deployment, the pod 
> is immediately restarted. 
> 5) Upon restart, the new Jobmanager finds no checkpoints to restore from.
> We think this is a bad combination of the following behaviors:
> * FAILED triggers cleanup, which cleans up HA configmaps in native kubernetes 
> mode
> * FAILED does not actually stop a job in native kubernetes mode, instead it 
> is immediately retried



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to