[
https://issues.apache.org/jira/browse/FLINK-34518?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17821080#comment-17821080
]
Matthias Pohl edited comment on FLINK-34518 at 2/27/24 2:13 PM:
----------------------------------------------------------------
master:
[3ba6f15043aac3e2d4744ad2431b4805ce5dec5a|https://github.com/apache/flink/commit/3ba6f15043aac3e2d4744ad2431b4805ce5dec5a]
1.19:
[d743ee33067dad18817d158f91db821cb494a31e|https://github.com/apache/flink/commit/d743ee33067dad18817d158f91db821cb494a31e]
1.18:
[384f6b24de9941d452ec84f9a7f617a65cb207a8|https://github.com/apache/flink/commit/384f6b24de9941d452ec84f9a7f617a65cb207a8]
1.17: 1.17:
[539a6fefa4cd049139e3be04f1672437e12744d5|https://github.com/apache/flink/commit/539a6fefa4cd049139e3be04f1672437e12744d5]
was (Author: mapohl):
master:
[3ba6f15043aac3e2d4744ad2431b4805ce5dec5a|https://github.com/apache/flink/commit/3ba6f15043aac3e2d4744ad2431b4805ce5dec5a]
1.19:
[d743ee33067dad18817d158f91db821cb494a31e|https://github.com/apache/flink/commit/d743ee33067dad18817d158f91db821cb494a31e]
1.18:
[384f6b24de9941d452ec84f9a7f617a65cb207a8|https://github.com/apache/flink/commit/384f6b24de9941d452ec84f9a7f617a65cb207a8]
1.17: tba
> Adaptive Scheduler restores from empty state if JM fails during restarting
> state
> --------------------------------------------------------------------------------
>
> Key: FLINK-34518
> URL: https://issues.apache.org/jira/browse/FLINK-34518
> Project: Flink
> Issue Type: Bug
> Components: Runtime / Checkpointing, Runtime / Coordination
> Affects Versions: 1.17.2, 1.19.0, 1.18.1, 1.20.0
> Reporter: Gyula Fora
> Assignee: Matthias Pohl
> Priority: Critical
> Labels: pull-request-available
>
> If a JobManager failover occurs while the Job is in a Restarting state, the
> HA metadata is deleted (as if it was a globally terminal state) and the job
> restarts from an empty state after the JM comes back up:
> Jobmanager killed after killing Taskmanager (restarting phase):
> {noformat}
> 2024-02-26 10:10:12,147 DEBUG
> org.apache.flink.kubernetes.KubernetesResourceManagerDriver [] - Ignore
> TaskManager pod that is already added: autoscaling-example-taskmanager-3-2
> 2024-02-26 10:10:13,799 DEBUG
> org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] -
> Trigger heartbeat request.
> 2024-02-26 10:10:13,799 DEBUG
> org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] -
> Trigger heartbeat request.
> 2024-02-26 10:10:13,799 DEBUG org.apache.flink.runtime.jobmaster.JobMaster
> [] - Received heartbeat request from
> 9b7e17b75812ab60ecf028e02368d0c2.
> 2024-02-26 10:10:13,799 DEBUG
> org.apache.flink.runtime.resourcemanager.active.ActiveResourceManager [] -
> Received heartbeat from 251c25cf794e3c9396fc02306613507b.
> 2024-02-26 10:10:14,091 DEBUG
> org.apache.pekko.remote.transport.netty.NettyTransport [] - Remote
> connection to [/10.244.0.120:55647] was disconnected because of [id:
> 0x4a61a791, /10.244.0.120:55647 :> /10.244.0.118:6123] DISCONNECTED
> 2024-02-26 10:10:14,091 DEBUG
> org.apache.pekko.remote.transport.ProtocolStateActor [] - Association
> between local [tcp://[email protected]:6123] and remote
> [tcp://[email protected]:55647] was disassociated because the
> ProtocolStateActor failed: Unknown
> 2024-02-26 10:10:14,092 INFO
> org.apache.flink.runtime.entrypoint.ClusterEntrypoint [] - RECEIVED
> SIGNAL 15: SIGTERM. Shutting down as requested.
> 2024-02-26 10:10:14,094 INFO
> org.apache.flink.runtime.entrypoint.ClusterEntrypoint [] - Shutting
> KubernetesApplicationClusterEntrypoint down with application status UNKNOWN.
> Diagnostics Cluster entrypoint has been closed externally..
> 2024-02-26 10:10:14,095 INFO org.apache.flink.runtime.blob.BlobServer
> [] - Stopped BLOB server at 0.0.0.0:6124
> 2024-02-26 10:10:14,095 INFO
> org.apache.flink.runtime.jobmaster.MiniDispatcherRestEndpoint [] - Shutting
> down rest endpoint.
> 2024-02-26 10:10:14,315 DEBUG
> org.apache.flink.kubernetes.shaded.io.fabric8.kubernetes.client.Watcher [] -
> Watcher closed
> 2024-02-26 10:10:14,511 DEBUG org.apache.pekko.actor.CoordinatedShutdown
> [] - Performing task [terminate-system] in CoordinatedShutdown
> phase [actor-system-terminate]
> 2024-02-26 10:10:14,595 INFO
> org.apache.pekko.remote.RemoteActorRefProvider$RemotingTerminator [] -
> Shutting down remote daemon.
> 2024-02-26 10:10:14,596 INFO
> org.apache.pekko.remote.RemoteActorRefProvider$RemotingTerminator [] - Remote
> daemon shut down; proceeding with flushing remote transports.{noformat}
> Then the new JM comes back it doesn't find any checkpoints as the HA metadata
> was deleted (we couldn't see this in the logs of the shutting down JM):
> {noformat}
> 2024-02-26 10:10:30,294 INFO
> org.apache.flink.runtime.checkpoint.DefaultCompletedCheckpointStoreUtils [] -
> Recovering checkpoints from
> KubernetesStateHandleStore{configMapName='autoscaling-example-5ddd0b1ba346d3bfd5ef53a63772e43c-config-map'}.2024-02-26
> 10:10:30,394 INFO
> org.apache.flink.runtime.checkpoint.DefaultCompletedCheckpointStoreUtils [] -
> Found 0 checkpoints in
> KubernetesStateHandleStore{configMapName='autoscaling-example-5ddd0b1ba346d3bfd5ef53a63772e43c-config-map'}.{noformat}
> Even the main method is re-run and the jobgraph is regenerated (which is
> expected given the HA metadata was removed incorrectly)
--
This message was sent by Atlassian Jira
(v8.20.10#820010)