[ https://issues.apache.org/jira/browse/FLINK-16215?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17046195#comment-17046195 ]
YufeiLiu commented on FLINK-16215: ---------------------------------- [~xintongsong] This sounds good. We can do some actual recovery work like release unused containers rather than just put all previous containers into WorkerMap. It need some changes in {{YarnResourceManager}}, It's kind of like {{MesosResourceManager}} recovery process. I think put the recovery work in {{prepareLeadershipAsync}} would be nice, which don't confirm leadership until the recovery is done. Do we have plans to improve on this? Besides, I had some question about "RM not assuming all TMs have the same resource", than which decide the specification of TM? > Start redundant TaskExecutor when JM failed > ------------------------------------------- > > Key: FLINK-16215 > URL: https://issues.apache.org/jira/browse/FLINK-16215 > Project: Flink > Issue Type: Bug > Components: Runtime / Coordination > Affects Versions: 1.10.0 > Reporter: YufeiLiu > Priority: Major > > TaskExecutor will reconnect to the new ResourceManager leader when JM failed, > and JobMaster will restart and reschedule job. If job slot request arrive > earlier than TM registration, RM will start new workers rather than reuse the > existing TMs. > It‘s hard to reproduce becasue TM registration usually come first, and > timeout check will stop redundant TMs. > But I think it would be better if we make the {{recoverWokerNode}} to > interface, and put recovered slots in {{pendingSlots}} wait for TM > reconnection. -- This message was sent by Atlassian Jira (v8.3.4#803005)