[ 
https://issues.apache.org/jira/browse/FLINK-34518?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17821080#comment-17821080
 ] 

Matthias Pohl edited comment on FLINK-34518 at 2/27/24 2:13 PM:
----------------------------------------------------------------

master: 
[3ba6f15043aac3e2d4744ad2431b4805ce5dec5a|https://github.com/apache/flink/commit/3ba6f15043aac3e2d4744ad2431b4805ce5dec5a]
1.19: 
[d743ee33067dad18817d158f91db821cb494a31e|https://github.com/apache/flink/commit/d743ee33067dad18817d158f91db821cb494a31e]
1.18: 
[384f6b24de9941d452ec84f9a7f617a65cb207a8|https://github.com/apache/flink/commit/384f6b24de9941d452ec84f9a7f617a65cb207a8]
1.17: 1.17: 
[539a6fefa4cd049139e3be04f1672437e12744d5|https://github.com/apache/flink/commit/539a6fefa4cd049139e3be04f1672437e12744d5]


was (Author: mapohl):
master: 
[3ba6f15043aac3e2d4744ad2431b4805ce5dec5a|https://github.com/apache/flink/commit/3ba6f15043aac3e2d4744ad2431b4805ce5dec5a]
1.19: 
[d743ee33067dad18817d158f91db821cb494a31e|https://github.com/apache/flink/commit/d743ee33067dad18817d158f91db821cb494a31e]
1.18: 
[384f6b24de9941d452ec84f9a7f617a65cb207a8|https://github.com/apache/flink/commit/384f6b24de9941d452ec84f9a7f617a65cb207a8]
1.17: tba

> Adaptive Scheduler restores from empty state if JM fails during restarting 
> state
> --------------------------------------------------------------------------------
>
>                 Key: FLINK-34518
>                 URL: https://issues.apache.org/jira/browse/FLINK-34518
>             Project: Flink
>          Issue Type: Bug
>          Components: Runtime / Checkpointing, Runtime / Coordination
>    Affects Versions: 1.17.2, 1.19.0, 1.18.1, 1.20.0
>            Reporter: Gyula Fora
>            Assignee: Matthias Pohl
>            Priority: Critical
>              Labels: pull-request-available
>
> If a JobManager failover occurs while the Job is in a Restarting state, the 
> HA metadata is deleted (as if it was a globally terminal state) and the job 
> restarts from an empty state after the JM comes back up:
> Jobmanager killed after killing Taskmanager (restarting phase):
> {noformat}
> 2024-02-26 10:10:12,147 DEBUG 
> org.apache.flink.kubernetes.KubernetesResourceManagerDriver  [] - Ignore 
> TaskManager pod that is already added: autoscaling-example-taskmanager-3-2
> 2024-02-26 10:10:13,799 DEBUG 
> org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] - 
> Trigger heartbeat request.
> 2024-02-26 10:10:13,799 DEBUG 
> org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] - 
> Trigger heartbeat request.
> 2024-02-26 10:10:13,799 DEBUG org.apache.flink.runtime.jobmaster.JobMaster    
>              [] - Received heartbeat request from 
> 9b7e17b75812ab60ecf028e02368d0c2.
> 2024-02-26 10:10:13,799 DEBUG 
> org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] - 
> Received heartbeat from 251c25cf794e3c9396fc02306613507b.
> 2024-02-26 10:10:14,091 DEBUG 
> org.apache.pekko.remote.transport.netty.NettyTransport       [] - Remote 
> connection to [/10.244.0.120:55647] was disconnected because of [id: 
> 0x4a61a791, /10.244.0.120:55647 :> /10.244.0.118:6123] DISCONNECTED
> 2024-02-26 10:10:14,091 DEBUG 
> org.apache.pekko.remote.transport.ProtocolStateActor         [] - Association 
> between local [tcp://[email protected]:6123] and remote 
> [tcp://[email protected]:55647] was disassociated because the 
> ProtocolStateActor failed: Unknown
> 2024-02-26 10:10:14,092 INFO  
> org.apache.flink.runtime.entrypoint.ClusterEntrypoint        [] - RECEIVED 
> SIGNAL 15: SIGTERM. Shutting down as requested.
> 2024-02-26 10:10:14,094 INFO  
> org.apache.flink.runtime.entrypoint.ClusterEntrypoint        [] - Shutting 
> KubernetesApplicationClusterEntrypoint down with application status UNKNOWN. 
> Diagnostics Cluster entrypoint has been closed externally..
> 2024-02-26 10:10:14,095 INFO  org.apache.flink.runtime.blob.BlobServer        
>              [] - Stopped BLOB server at 0.0.0.0:6124
> 2024-02-26 10:10:14,095 INFO  
> org.apache.flink.runtime.jobmaster.MiniDispatcherRestEndpoint [] - Shutting 
> down rest endpoint.
> 2024-02-26 10:10:14,315 DEBUG 
> org.apache.flink.kubernetes.shaded.io.fabric8.kubernetes.client.Watcher [] - 
> Watcher closed
> 2024-02-26 10:10:14,511 DEBUG org.apache.pekko.actor.CoordinatedShutdown      
>              [] - Performing task [terminate-system] in CoordinatedShutdown 
> phase [actor-system-terminate]
> 2024-02-26 10:10:14,595 INFO  
> org.apache.pekko.remote.RemoteActorRefProvider$RemotingTerminator [] - 
> Shutting down remote daemon.
> 2024-02-26 10:10:14,596 INFO  
> org.apache.pekko.remote.RemoteActorRefProvider$RemotingTerminator [] - Remote 
> daemon shut down; proceeding with flushing remote transports.{noformat}
> Then the new JM comes back it doesn't find any checkpoints as the HA metadata 
> was deleted (we couldn't see this in the logs of the shutting down JM):
> {noformat}
> 2024-02-26 10:10:30,294 INFO  
> org.apache.flink.runtime.checkpoint.DefaultCompletedCheckpointStoreUtils [] - 
> Recovering checkpoints from 
> KubernetesStateHandleStore{configMapName='autoscaling-example-5ddd0b1ba346d3bfd5ef53a63772e43c-config-map'}.2024-02-26
>  10:10:30,394 INFO  
> org.apache.flink.runtime.checkpoint.DefaultCompletedCheckpointStoreUtils [] - 
> Found 0 checkpoints in 
> KubernetesStateHandleStore{configMapName='autoscaling-example-5ddd0b1ba346d3bfd5ef53a63772e43c-config-map'}.{noformat}
> Even the main method is re-run and the jobgraph is regenerated (which is 
> expected given the HA metadata was removed incorrectly)



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to