[ https://issues.apache.org/jira/browse/YARN-7695?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16309278#comment-16309278 ]
stefanlee commented on YARN-7695: --------------------------------- there is another problem in this scenario, when i turn on *ContinuousScheduling* , and submit a lot of applications, then my cluster have no available resource, active RM1's log print {code:java} 2018-01-03 16:05:49,860 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.AppSchedulable: Making reservation: node=datanode2 app_id=application_1514952157240_0019 2018-01-03 16:05:49,860 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt: Application application_1514952157240_0019 reserved container container_1514952157240_0019_02_000003 on node host: datanode2:37528 #containers=2 available=<memory:2048, vCores:1> used=<memory:4096, vCores:2>, currently has 1 at priority 10; currentReservation 3072 2018-01-03 16:05:49,860 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FSSchedulerNode: Updated reserved container container_1514952157240_0019_02_000003 on node host: datanode2:37528 #containers=2 available=<memory:2048, vCores:1> used=<memory:4096, vCores:2> for application org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FSSchedulerApp@a9790a8 2018-01-03 16:05:49,868 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairScheduler: Trying to fulfill reservation for application appattempt_1514952157240_0019_000002 on node: host: datanode2:37528 #containers=2 available=<memory:2048, vCores:1> used=<memory:4096, vCores:2> 2018-01-03 16:05:49,868 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.AppSchedulable: Making reservation: node=datanode2 app_id=application_1514952157240_0019 2018-01-03 16:05:49,868 INFO org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerApplicationAttempt: Application application_1514952157240_0019 reserved container container_1514952157240_0019_02_000003 on node host: datanode2:37528 #containers=2 available=<memory:2048, vCores:1> used=<memory:4096, vCores:2>, currently has 1 at priority 10; currentReservation 3072 {code} , then i repeat 3 step in description, active RM1 transit to standby, RM2 runs normaly, but standby RM1's log still print above info, it seems *ContinuousScheduling* runs in dead loop. so i think it is best to stop these thread when RM1 transit to standby in this scenario. > when active RM transit to standby , this RM will new another > FairSchedulerUpdate Thread > --------------------------------------------------------------------------------------- > > Key: YARN-7695 > URL: https://issues.apache.org/jira/browse/YARN-7695 > Project: Hadoop YARN > Issue Type: Bug > Components: fairscheduler, resourcemanager > Affects Versions: 2.4.0 > Reporter: stefanlee > > 1. i test haoop-2.4.0 in my cluster. > 2. RM1 is active and RM2 is standby > 3. i delete /yarn-leader-election/Yarn/ActiveStandbyElectorLock from ZK > 4. RM1 then transit from active to standby success. > 5. at last ,i print RM1 jstack info and found two "AllocationFileReloader" > and two "FairSchedulerUpdateThread" in RM1. -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org