[ https://issues.apache.org/jira/browse/YARN-3894?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14623972#comment-14623972 ]
Hadoop QA commented on YARN-3894: --------------------------------- \\ \\ | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | pre-patch | 15m 59s | Pre-patch trunk compilation is healthy. | | {color:green}+1{color} | @author | 0m 0s | The patch does not contain any @author tags. | | {color:green}+1{color} | tests included | 0m 0s | The patch appears to include 1 new or modified test files. | | {color:green}+1{color} | javac | 7m 39s | There were no new javac warning messages. | | {color:green}+1{color} | javadoc | 9m 35s | There were no new javadoc warning messages. | | {color:green}+1{color} | release audit | 0m 22s | The applied patch does not increase the total number of release audit warnings. | | {color:green}+1{color} | checkstyle | 0m 46s | There were no new checkstyle issues. | | {color:green}+1{color} | whitespace | 0m 0s | The patch has no lines that end in whitespace. | | {color:green}+1{color} | install | 1m 21s | mvn install still works. | | {color:green}+1{color} | eclipse:eclipse | 0m 32s | The patch built with eclipse:eclipse. | | {color:green}+1{color} | findbugs | 1m 26s | The patch does not introduce any new Findbugs (version 3.0.0) warnings. | | {color:red}-1{color} | yarn tests | 51m 19s | Tests failed in hadoop-yarn-server-resourcemanager. | | | | 89m 2s | | \\ \\ || Reason || Tests || | Failed unit tests | hadoop.yarn.server.resourcemanager.applicationsmanager.TestAMRMRPCNodeUpdates | | | hadoop.yarn.server.resourcemanager.rmapp.TestRMAppTransitions | | | hadoop.yarn.server.resourcemanager.TestApplicationCleanup | | | hadoop.yarn.server.resourcemanager.TestResourceTrackerService | \\ \\ || Subsystem || Report/Notes || | Patch URL | http://issues.apache.org/jira/secure/attachment/12744946/0002-YARN-3894.patch | | Optional Tests | javadoc javac unit findbugs checkstyle | | git revision | trunk / d7319de | | hadoop-yarn-server-resourcemanager test log | https://builds.apache.org/job/PreCommit-YARN-Build/8514/artifact/patchprocess/testrun_hadoop-yarn-server-resourcemanager.txt | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/8514/testReport/ | | Java | 1.7.0_55 | | uname | Linux asf906.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/8514/console | This message was automatically generated. > RM startup should fail for wrong CS xml NodeLabel capacity configuration > ------------------------------------------------------------------------- > > Key: YARN-3894 > URL: https://issues.apache.org/jira/browse/YARN-3894 > Project: Hadoop YARN > Issue Type: Bug > Components: capacityscheduler > Reporter: Bibin A Chundatt > Assignee: Bibin A Chundatt > Priority: Critical > Attachments: 0001-YARN-3894.patch, 0002-YARN-3894.patch, > capacity-scheduler.xml > > > Currently in capacity Scheduler when capacity configuration is wrong > RM will shutdown, but not incase of NodeLabels capacity mismatch > In {{CapacityScheduler#initializeQueues}} > {code} > private void initializeQueues(CapacitySchedulerConfiguration conf) > throws IOException { > root = > parseQueue(this, conf, null, CapacitySchedulerConfiguration.ROOT, > queues, queues, noop); > labelManager.reinitializeQueueLabels(getQueueToLabels()); > root = > parseQueue(this, conf, null, CapacitySchedulerConfiguration.ROOT, > queues, queues, noop); > LOG.info("Initialized root queue " + root); > initializeQueueMappings(); > setQueueAcls(authorizer, queues); > } > {code} > {{labelManager}} is initialized from queues and calculation for Label level > capacity mismatch happens in {{parseQueue}} . So during initialization > {{parseQueue}} the labels will be empty . > *Steps to reproduce* > # Configure RM with capacity scheduler > # Add one or two node label from rmadmin > # Configure capacity xml with nodelabel but issue with capacity configuration > for already added label > # Restart both RM > # Check on service init of capacity scheduler node label list is populated > *Expected* > RM should not start > *Current exception on reintialize check* > {code} > 2015-07-07 19:18:25,655 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler: > Initialized queue: default: capacity=0.5, absoluteCapacity=0.5, > usedResources=<memory:0, vCores:0>, usedCapacity=0.0, > absoluteUsedCapacity=0.0, numApps=0, numContainers=0 > 2015-07-07 19:18:25,656 WARN > org.apache.hadoop.yarn.server.resourcemanager.AdminService: Exception refresh > queues. > java.io.IOException: Failed to re-init queues > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler.reinitialize(CapacityScheduler.java:383) > at > org.apache.hadoop.yarn.server.resourcemanager.AdminService.refreshQueues(AdminService.java:376) > at > org.apache.hadoop.yarn.server.resourcemanager.AdminService.refreshAll(AdminService.java:605) > at > org.apache.hadoop.yarn.server.resourcemanager.AdminService.transitionToActive(AdminService.java:314) > at > org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService.becomeActive(EmbeddedElectorService.java:126) > at > org.apache.hadoop.ha.ActiveStandbyElector.becomeActive(ActiveStandbyElector.java:824) > at > org.apache.hadoop.ha.ActiveStandbyElector.processResult(ActiveStandbyElector.java:420) > at > org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:599) > at > org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:498) > Caused by: java.lang.IllegalArgumentException: Illegal capacity of 0.5 for > children of queue root for label=node2 > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue.setChildQueues(ParentQueue.java:159) > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler.parseQueue(CapacityScheduler.java:639) > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler.reinitializeQueues(CapacityScheduler.java:503) > at > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler.reinitialize(CapacityScheduler.java:379) > ... 8 more > 2015-07-07 19:18:25,656 WARN > org.apache.hadoop.yarn.server.resourcemanager.RMAuditLogger: USER=dsperf > OPERATION=refreshQueues TARGET=AdminService RESULT=FAILURE > DESCRIPTION=Exception refresh queues. PERMISSIONS= > 2015-07-07 19:18:25,656 WARN > org.apache.hadoop.yarn.server.resourcemanager.RMAuditLogger: USER=dsperf > OPERATION=transitionToActive TARGET=RMHAProtocolService > RESULT=FAILURE DESCRIPTION=Exception transitioning to active PERMISSIONS= > 2015-07-07 19:18:25,656 WARN org.apache.hadoop.ha.ActiveStandbyElector: > Exception handling the winning of election > org.apache.hadoop.ha.ServiceFailedException: RM could not transition to Active > at > org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService.becomeActive(EmbeddedElectorService.java:128) > at > org.apache.hadoop.ha.ActiveStandbyElector.becomeActive(ActiveStandbyElector.java:824) > at > org.apache.hadoop.ha.ActiveStandbyElector.processResult(ActiveStandbyElector.java:420) > at > org.apache.zookeeper.ClientCnxn$EventThread.processEvent(ClientCnxn.java:599) > at > org.apache.zookeeper.ClientCnxn$EventThread.run(ClientCnxn.java:498) > Caused by: org.apache.hadoop.ha.ServiceFailedException: Error when > transitioning to Active mode > at > org.apache.hadoop.yarn.server.resourcemanager.AdminService.transitionToActive(AdminService.java:321) > at > org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService.becomeActive(EmbeddedElectorService.java:126) > ... 4 more > Caused by: org.apache.hadoop.ha.ServiceFailedException: java.io.IOException: > Failed to re-init queues > at > org.apache.hadoop.yarn.server.resourcemanager.AdminService.refreshAll(AdminService.java:617) > at > org.apache.hadoop.yarn.server.resourcemanager.AdminService.transitionToActive(AdminService.java:314) > ... 5 more > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332)