[jira] [Commented] (YARN-2194) Cgroups cease to work in RHEL7
[ https://issues.apache.org/jira/browse/YARN-2194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15680343#comment-15680343 ] carlhe commented on YARN-2194: -- My hadoop version is 2.6.4 which is also affected by such bug. Obviously, the patch from (v5) is not available for me, because 2.6.4 even not has the directory "hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/linux/..." How can I solve this issue? Thank you~ > Cgroups cease to work in RHEL7 > -- > > Key: YARN-2194 > URL: https://issues.apache.org/jira/browse/YARN-2194 > Project: Hadoop YARN > Issue Type: Bug > Components: nodemanager >Affects Versions: 2.7.0 >Reporter: Wei Yan >Assignee: Wei Yan >Priority: Critical > Fix For: 2.8.0, 3.0.0-alpha1 > > Attachments: YARN-2194-1.patch, YARN-2194-2.patch, YARN-2194-3.patch, > YARN-2194-4.patch, YARN-2194-5.patch, YARN-2194-6.patch, YARN-2194-7.patch > > > In RHEL7, the CPU controller is named "cpu,cpuacct". The comma in the > controller name leads to container launch failure. > RHEL7 deprecates libcgroup and recommends the user of systemd. However, > systemd has certain shortcomings as identified in this JIRA (see comments). > This JIRA only fixes the failure, and doesn't try to use systemd. -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5915) ATS 1.5 FileSystemTimelineWriter causes flush() to be called after every event write
[ https://issues.apache.org/jira/browse/YARN-5915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15680063#comment-15680063 ] Junping Du commented on YARN-5915: -- Also, CC [~gtCarrera9] as well. > ATS 1.5 FileSystemTimelineWriter causes flush() to be called after every > event write > > > Key: YARN-5915 > URL: https://issues.apache.org/jira/browse/YARN-5915 > Project: Hadoop YARN > Issue Type: Bug > Components: timelineserver >Affects Versions: 3.0.0-alpha1 >Reporter: Atul Sikaria >Assignee: Atul Sikaria > Attachments: YARN-5915.01.patch > > -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5915) ATS 1.5 FileSystemTimelineWriter causes flush() to be called after every event write
[ https://issues.apache.org/jira/browse/YARN-5915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15680062#comment-15680062 ] Junping Du commented on YARN-5915: -- Thanks [~ASikaria] for reporting the issue and delivering a fix. I have added you to yarn contributor list, and assign this jira to you. The fix sounds reasonable to me. CC [~xgong] and [~jlowe] for more comments who worked on YARN-4814 before. > ATS 1.5 FileSystemTimelineWriter causes flush() to be called after every > event write > > > Key: YARN-5915 > URL: https://issues.apache.org/jira/browse/YARN-5915 > Project: Hadoop YARN > Issue Type: Bug > Components: timelineserver >Affects Versions: 3.0.0-alpha1 >Reporter: Atul Sikaria >Assignee: Atul Sikaria > Attachments: YARN-5915.01.patch > > -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5915) ATS 1.5 FileSystemTimelineWriter causes flush() to be called after every event write
[ https://issues.apache.org/jira/browse/YARN-5915?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Junping Du updated YARN-5915: - Assignee: Atul Sikaria > ATS 1.5 FileSystemTimelineWriter causes flush() to be called after every > event write > > > Key: YARN-5915 > URL: https://issues.apache.org/jira/browse/YARN-5915 > Project: Hadoop YARN > Issue Type: Bug > Components: timelineserver >Affects Versions: 3.0.0-alpha1 >Reporter: Atul Sikaria >Assignee: Atul Sikaria > Attachments: YARN-5915.01.patch > > -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5877) Allow all nm-whitelist-env to get overridden during launch
[ https://issues.apache.org/jira/browse/YARN-5877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679848#comment-15679848 ] Varun Saxena commented on YARN-5877: Sorry for the confusion. Correct. We do not inherit parent environment while launching the container. We explicitly clear the environment in ProcessBuilder and use the environment in container launch context instead. But can a similar thing happen in docker scenario ? The main thing which I wanted to highlight was that we should probably give preference to environment set in container launch context if its also configured in the whitelist no matter what the environment is set to. > Allow all nm-whitelist-env to get overridden during launch > -- > > Key: YARN-5877 > URL: https://issues.apache.org/jira/browse/YARN-5877 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Attachments: Dockerfile, YARN-5877.0001.patch, YARN-5877.0002.patch, > bootstrap.sh, yarn-site.xml > > > As per the {{yarn.nodemanager.env-whitelist}} for the configured values > should containers may override rather than use NodeManager's default. > {code} > > Environment variables that containers may override rather > than use NodeManager's default. > yarn.nodemanager.env-whitelist > > JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,CLASSPATH_PREPEND_DISTCACHE,HADOOP_YARN_HOME > > {code} > But only the following containers can override > {code} > whitelist.add(ApplicationConstants.Environment.HADOOP_YARN_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_COMMON_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_HDFS_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_CONF_DIR.name()); > whitelist.add(ApplicationConstants.Environment.JAVA_HOME.name()); > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (YARN-5548) Use MockRMMemoryStateStore to reduce test failures
[ https://issues.apache.org/jira/browse/YARN-5548?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679043#comment-15679043 ] Varun Saxena edited comment on YARN-5548 at 11/19/16 5:36 PM: -- [~sunilg], I am fine with your suggestion. It can be useful elsewhere i.e. where MockRM isn't used as well. Although its public static class inside MockRM, referring to MockRM in a test where it isn't used would look a little awkward. was (Author: varun_saxena): [~sunilg], I am fine with your suggestion. It can be useful elsewhere i.e. where MockRM isn't used as well. Although its public class inside MockRM, referring to MockRM in a test where it isn't used would look a little awkward. > Use MockRMMemoryStateStore to reduce test failures > -- > > Key: YARN-5548 > URL: https://issues.apache.org/jira/browse/YARN-5548 > Project: Hadoop YARN > Issue Type: Test >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Labels: oct16-easy, test > Attachments: YARN-5548.0001.patch, YARN-5548.0002.patch, > YARN-5548.0003.patch, YARN-5548.0004.patch, YARN-5548.0005.patch > > > https://builds.apache.org/job/PreCommit-YARN-Build/12850/testReport/org.apache.hadoop.yarn.server.resourcemanager/TestRMRestart/testFinishedAppRemovalAfterRMRestart/ > {noformat} > Error Message > Stacktrace > java.lang.AssertionError: expected null, but was: application_submission_context { application_id { id: 1 cluster_timestamp: > 1471885197388 } application_name: "" queue: "default" priority { priority: 0 > } am_container_spec { } cancel_tokens_when_complete: true maxAppAttempts: 2 > resource { memory: 1024 virtual_cores: 1 } applicationType: "YARN" > keep_containers_across_application_attempts: false > attempt_failures_validity_interval: 0 am_container_resource_request { > priority { priority: 0 } resource_name: "*" capability { memory: 1024 > virtual_cores: 1 } num_containers: 0 relax_locality: true > node_label_expression: "" execution_type_request { execution_type: GUARANTEED > enforce_execution_type: false } } } user: "jenkins" start_time: 1471885197417 > application_state: RMAPP_FINISHED finish_time: 1471885197478> > at org.junit.Assert.fail(Assert.java:88) > at org.junit.Assert.failNotNull(Assert.java:664) > at org.junit.Assert.assertNull(Assert.java:646) > at org.junit.Assert.assertNull(Assert.java:656) > at > org.apache.hadoop.yarn.server.resourcemanager.TestRMRestart.testFinishedAppRemovalAfterRMRestart(TestRMRestart.java:1656) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5916) Rack shown as null for node decommisioned during registration.
[ https://issues.apache.org/jira/browse/YARN-5916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679584#comment-15679584 ] Hadoop QA commented on YARN-5916: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 1m 0s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 46s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 33s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 20s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 39s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 17s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 1s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 22s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 32s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 30s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 30s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 17s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 36s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 14s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 5s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 19s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 41m 41s{color} | {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 15s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 57m 52s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.yarn.server.resourcemanager.TestRMRestart | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:a9ad5d6 | | JIRA Issue | YARN-5916 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12839691/YARN-5916.0001.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 3559bc6849f8 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 7584fbf | | Default Java | 1.8.0_111 | | findbugs | v3.0.0 | | unit | https://builds.apache.org/job/PreCommit-YARN-Build/13984/artifact/patchprocess/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/13984/testReport/ | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/13984/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Rack shown as null for node decommisioned during registration. > --
[jira] [Updated] (YARN-5916) Rack shown as null for node decommisioned during registration.
[ https://issues.apache.org/jira/browse/YARN-5916?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bibin A Chundatt updated YARN-5916: --- Attachment: YARN-5916.0001.patch > Rack shown as null for node decommisioned during registration. > -- > > Key: YARN-5916 > URL: https://issues.apache.org/jira/browse/YARN-5916 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Minor > Attachments: Screenshot from 2016-11-19 15-22-23.png, > YARN-5916.0001.patch > > > Steps to reproduce > # configure exclude node as nm1 hostname > # start resourcemanager > # start nm1 hostname > # with current configuration during node registration the node gets rejected. > # Now check RM UI -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5916) Rack shown as null for node decommisioned during registration.
[ https://issues.apache.org/jira/browse/YARN-5916?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bibin A Chundatt updated YARN-5916: --- Attachment: (was: YARN-5916.0001.patch) > Rack shown as null for node decommisioned during registration. > -- > > Key: YARN-5916 > URL: https://issues.apache.org/jira/browse/YARN-5916 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Minor > Attachments: Screenshot from 2016-11-19 15-22-23.png, > YARN-5916.0001.patch > > > Steps to reproduce > # configure exclude node as nm1 hostname > # start resourcemanager > # start nm1 hostname > # with current configuration during node registration the node gets rejected. > # Now check RM UI -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-3884) RMContainerImpl transition from RESERVED to KILL apphistory status not updated
[ https://issues.apache.org/jira/browse/YARN-3884?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679424#comment-15679424 ] Hadoop QA commented on YARN-3884: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 25s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 2 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 9m 4s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 42s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 31s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 50s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 22s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 13s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 26s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 40s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 37s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 37s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 27s{color} | {color:orange} hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager: The patch generated 10 new + 452 unchanged - 1 fixed = 462 total (was 453) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 47s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 21s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 30s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 27s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 48m 25s{color} | {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 18s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 68m 30s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.yarn.server.resourcemanager.scheduler.fair.TestContinuousScheduling | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:a9ad5d6 | | JIRA Issue | YARN-3884 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12839688/YARN-3884.0003.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 2e3dfd81c17b 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 7584fbf | | Default Java | 1.8.0_111 | | findbugs | v3.0.0 | | checkstyle | https://builds.apache.org/job/PreCommit-YARN-Build/13983/artifact/patchprocess/diff-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt | | unit | https://builds.apache.org/job/PreCommit-YARN-Build/13983/artifact/patchprocess/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/13983/testReport/ | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager | | Console output | https://builds.apac
[jira] [Comment Edited] (YARN-3884) RMContainerImpl transition from RESERVED to KILL apphistory status not updated
[ https://issues.apache.org/jira/browse/YARN-3884?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679330#comment-15679330 ] Bibin A Chundatt edited comment on YARN-3884 at 11/19/16 2:12 PM: -- Attaching patch after handling comments was (Author: bibinchundatt): Attaching patch after handling patch > RMContainerImpl transition from RESERVED to KILL apphistory status not updated > -- > > Key: YARN-3884 > URL: https://issues.apache.org/jira/browse/YARN-3884 > Project: Hadoop YARN > Issue Type: Bug > Components: resourcemanager > Environment: Suse11 Sp3 >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Labels: oct16-easy > Attachments: 0001-YARN-3884.patch, Apphistory Container Status.jpg, > Elapsed Time.jpg, Test Result-Container status.jpg, YARN-3884.0002.patch, > YARN-3884.0003.patch > > > Setup > === > 1 NM 3072 16 cores each > Steps to reproduce > === > 1.Submit apps to Queue 1 with 512 mb 1 core > 2.Submit apps to Queue 2 with 512 mb and 5 core > lots of containers get reserved and unreserved in this case > {code} > 2015-07-02 20:45:31,169 INFO > org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: > container_e24_1435849994778_0002_01_13 Container Transitioned from NEW to > RESERVED > 2015-07-02 20:45:31,170 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue: > Reserved container application=application_1435849994778_0002 > resource= queue=QueueA: capacity=0.4, > absoluteCapacity=0.4, usedResources=, > usedCapacity=1.6410257, absoluteUsedCapacity=0.65625, numApps=1, > numContainers=5 usedCapacity=1.6410257 absoluteUsedCapacity=0.65625 > used= cluster= > 2015-07-02 20:45:31,170 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: > Re-sorting assigned queue: root.QueueA stats: QueueA: capacity=0.4, > absoluteCapacity=0.4, usedResources=, > usedCapacity=2.0317461, absoluteUsedCapacity=0.8125, numApps=1, > numContainers=6 > 2015-07-02 20:45:31,170 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: > assignedContainer queue=root usedCapacity=0.96875 > absoluteUsedCapacity=0.96875 used= > cluster= > 2015-07-02 20:45:31,191 INFO > org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: > container_e24_1435849994778_0001_01_14 Container Transitioned from NEW to > ALLOCATED > 2015-07-02 20:45:31,191 INFO > org.apache.hadoop.yarn.server.resourcemanager.RMAuditLogger: USER=dsperf > OPERATION=AM Allocated ContainerTARGET=SchedulerApp > RESULT=SUCCESS APPID=application_1435849994778_0001 > CONTAINERID=container_e24_1435849994778_0001_01_14 > 2015-07-02 20:45:31,191 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerNode: > Assigned container container_e24_1435849994778_0001_01_14 of capacity > on host host-10-19-92-117:64318, which has 6 > containers, used and available > after allocation > 2015-07-02 20:45:31,191 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue: > assignedContainer application attempt=appattempt_1435849994778_0001_01 > container=Container: [ContainerId: > container_e24_1435849994778_0001_01_14, NodeId: host-10-19-92-117:64318, > NodeHttpAddress: host-10-19-92-117:65321, Resource: , > Priority: 20, Token: null, ] queue=default: capacity=0.2, > absoluteCapacity=0.2, usedResources=, > usedCapacity=2.0846906, absoluteUsedCapacity=0.4166, numApps=1, > numContainers=5 clusterResource= > 2015-07-02 20:45:31,191 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: > Re-sorting assigned queue: root.default stats: default: capacity=0.2, > absoluteCapacity=0.2, usedResources=, > usedCapacity=2.5016286, absoluteUsedCapacity=0.5, numApps=1, numContainers=6 > 2015-07-02 20:45:31,191 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: > assignedContainer queue=root usedCapacity=1.0 absoluteUsedCapacity=1.0 > used= cluster= > 2015-07-02 20:45:32,143 INFO > org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: > container_e24_1435849994778_0001_01_14 Container Transitioned from > ALLOCATED to ACQUIRED > 2015-07-02 20:45:32,174 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler: > Trying to fulfill reservation for application application_1435849994778_0002 > on node: host-10-19-92-143:64318 > 2015-07-02 20:45:32,174 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue: > Reserved container application=application_1435849994778_0
[jira] [Updated] (YARN-3884) RMContainerImpl transition from RESERVED to KILL apphistory status not updated
[ https://issues.apache.org/jira/browse/YARN-3884?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bibin A Chundatt updated YARN-3884: --- Attachment: YARN-3884.0003.patch Attaching patch after handling patch > RMContainerImpl transition from RESERVED to KILL apphistory status not updated > -- > > Key: YARN-3884 > URL: https://issues.apache.org/jira/browse/YARN-3884 > Project: Hadoop YARN > Issue Type: Bug > Components: resourcemanager > Environment: Suse11 Sp3 >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Labels: oct16-easy > Attachments: 0001-YARN-3884.patch, Apphistory Container Status.jpg, > Elapsed Time.jpg, Test Result-Container status.jpg, YARN-3884.0002.patch, > YARN-3884.0003.patch > > > Setup > === > 1 NM 3072 16 cores each > Steps to reproduce > === > 1.Submit apps to Queue 1 with 512 mb 1 core > 2.Submit apps to Queue 2 with 512 mb and 5 core > lots of containers get reserved and unreserved in this case > {code} > 2015-07-02 20:45:31,169 INFO > org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: > container_e24_1435849994778_0002_01_13 Container Transitioned from NEW to > RESERVED > 2015-07-02 20:45:31,170 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue: > Reserved container application=application_1435849994778_0002 > resource= queue=QueueA: capacity=0.4, > absoluteCapacity=0.4, usedResources=, > usedCapacity=1.6410257, absoluteUsedCapacity=0.65625, numApps=1, > numContainers=5 usedCapacity=1.6410257 absoluteUsedCapacity=0.65625 > used= cluster= > 2015-07-02 20:45:31,170 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: > Re-sorting assigned queue: root.QueueA stats: QueueA: capacity=0.4, > absoluteCapacity=0.4, usedResources=, > usedCapacity=2.0317461, absoluteUsedCapacity=0.8125, numApps=1, > numContainers=6 > 2015-07-02 20:45:31,170 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: > assignedContainer queue=root usedCapacity=0.96875 > absoluteUsedCapacity=0.96875 used= > cluster= > 2015-07-02 20:45:31,191 INFO > org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: > container_e24_1435849994778_0001_01_14 Container Transitioned from NEW to > ALLOCATED > 2015-07-02 20:45:31,191 INFO > org.apache.hadoop.yarn.server.resourcemanager.RMAuditLogger: USER=dsperf > OPERATION=AM Allocated ContainerTARGET=SchedulerApp > RESULT=SUCCESS APPID=application_1435849994778_0001 > CONTAINERID=container_e24_1435849994778_0001_01_14 > 2015-07-02 20:45:31,191 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.SchedulerNode: > Assigned container container_e24_1435849994778_0001_01_14 of capacity > on host host-10-19-92-117:64318, which has 6 > containers, used and available > after allocation > 2015-07-02 20:45:31,191 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue: > assignedContainer application attempt=appattempt_1435849994778_0001_01 > container=Container: [ContainerId: > container_e24_1435849994778_0001_01_14, NodeId: host-10-19-92-117:64318, > NodeHttpAddress: host-10-19-92-117:65321, Resource: , > Priority: 20, Token: null, ] queue=default: capacity=0.2, > absoluteCapacity=0.2, usedResources=, > usedCapacity=2.0846906, absoluteUsedCapacity=0.4166, numApps=1, > numContainers=5 clusterResource= > 2015-07-02 20:45:31,191 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: > Re-sorting assigned queue: root.default stats: default: capacity=0.2, > absoluteCapacity=0.2, usedResources=, > usedCapacity=2.5016286, absoluteUsedCapacity=0.5, numApps=1, numContainers=6 > 2015-07-02 20:45:31,191 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.ParentQueue: > assignedContainer queue=root usedCapacity=1.0 absoluteUsedCapacity=1.0 > used= cluster= > 2015-07-02 20:45:32,143 INFO > org.apache.hadoop.yarn.server.resourcemanager.rmcontainer.RMContainerImpl: > container_e24_1435849994778_0001_01_14 Container Transitioned from > ALLOCATED to ACQUIRED > 2015-07-02 20:45:32,174 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler: > Trying to fulfill reservation for application application_1435849994778_0002 > on node: host-10-19-92-143:64318 > 2015-07-02 20:45:32,174 INFO > org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.LeafQueue: > Reserved container application=application_1435849994778_0002 > resource= queue=QueueA: capacity=0.4, > absoluteCapacity=0.4, usedResources=, > usedCapacity=2.0317461, absoluteUsedCapacity=0.8125,
[jira] [Commented] (YARN-5877) Allow all nm-whitelist-env to get overridden during launch
[ https://issues.apache.org/jira/browse/YARN-5877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679311#comment-15679311 ] Bibin A Chundatt commented on YARN-5877: I am open to adding new configuration as mentioned earlier. > Allow all nm-whitelist-env to get overridden during launch > -- > > Key: YARN-5877 > URL: https://issues.apache.org/jira/browse/YARN-5877 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Attachments: Dockerfile, YARN-5877.0001.patch, YARN-5877.0002.patch, > bootstrap.sh, yarn-site.xml > > > As per the {{yarn.nodemanager.env-whitelist}} for the configured values > should containers may override rather than use NodeManager's default. > {code} > > Environment variables that containers may override rather > than use NodeManager's default. > yarn.nodemanager.env-whitelist > > JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,CLASSPATH_PREPEND_DISTCACHE,HADOOP_YARN_HOME > > {code} > But only the following containers can override > {code} > whitelist.add(ApplicationConstants.Environment.HADOOP_YARN_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_COMMON_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_HDFS_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_CONF_DIR.name()); > whitelist.add(ApplicationConstants.Environment.JAVA_HOME.name()); > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5148) [YARN-3368] Add page to new YARN UI to view server side configurations/logs/JVM-metrics
[ https://issues.apache.org/jira/browse/YARN-5148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679304#comment-15679304 ] Sunil G commented on YARN-5148: --- Thanks [~lewuathe] Yes, you are correct. In ember, we also need to add some more changes in css etc. [~Sreenath], could you please help to provide some thoughts here. > [YARN-3368] Add page to new YARN UI to view server side > configurations/logs/JVM-metrics > --- > > Key: YARN-5148 > URL: https://issues.apache.org/jira/browse/YARN-5148 > Project: Hadoop YARN > Issue Type: Sub-task > Components: webapp, yarn-ui-v2 >Reporter: Wangda Tan >Assignee: Kai Sasaki > Labels: oct16-medium > Attachments: Screen Shot 2016-09-11 at 23.28.31.png, Screen Shot > 2016-09-13 at 22.27.00.png, UsingStringifyPrint.png, > YARN-5148-YARN-3368.01.patch, YARN-5148-YARN-3368.02.patch, > YARN-5148-YARN-3368.03.patch, YARN-5148-YARN-3368.04.patch, > YARN-5148-YARN-3368.05.patch, YARN-5148-YARN-3368.06.patch, > YARN-5148.07.patch, yarn-conf.png, yarn-tools.png > > -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5888) [YARN-3368] Improve unit tests for YARN UI
[ https://issues.apache.org/jira/browse/YARN-5888?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679300#comment-15679300 ] Sunil G commented on YARN-5888: --- cc/ [~Sreenath] > [YARN-3368] Improve unit tests for YARN UI > -- > > Key: YARN-5888 > URL: https://issues.apache.org/jira/browse/YARN-5888 > Project: Hadoop YARN > Issue Type: Sub-task > Components: yarn-ui-v2 >Reporter: Akhil PB >Assignee: Akhil PB >Priority: Minor > Attachments: YARN-5888.001.patch > > > - Add missing test cases in new YARN UI > - Fix test cases errors in new YARN UI -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5877) Allow all nm-whitelist-env to get overridden during launch
[ https://issues.apache.org/jira/browse/YARN-5877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679295#comment-15679295 ] Bibin A Chundatt commented on YARN-5877: {quote} Before the change in the patch, we could have set DUMMY_ENV in NM whitelist and NM would have picked up value for this environment variable from container launch context (if available) otherwise taken it from NM environment. After the change in this patch, NM would keep the value coming from container launch context as default and use that only if the environment variable on NM is not set. {quote} Could please recheck this as per current Shell launch the system env will not be inherited from parent to child process. So will thr be a case other than docker this could happen?? > Allow all nm-whitelist-env to get overridden during launch > -- > > Key: YARN-5877 > URL: https://issues.apache.org/jira/browse/YARN-5877 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Attachments: Dockerfile, YARN-5877.0001.patch, YARN-5877.0002.patch, > bootstrap.sh, yarn-site.xml > > > As per the {{yarn.nodemanager.env-whitelist}} for the configured values > should containers may override rather than use NodeManager's default. > {code} > > Environment variables that containers may override rather > than use NodeManager's default. > yarn.nodemanager.env-whitelist > > JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,CLASSPATH_PREPEND_DISTCACHE,HADOOP_YARN_HOME > > {code} > But only the following containers can override > {code} > whitelist.add(ApplicationConstants.Environment.HADOOP_YARN_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_COMMON_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_HDFS_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_CONF_DIR.name()); > whitelist.add(ApplicationConstants.Environment.JAVA_HOME.name()); > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (YARN-5877) Allow all nm-whitelist-env to get overridden during launch
[ https://issues.apache.org/jira/browse/YARN-5877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679132#comment-15679132 ] Varun Saxena edited comment on YARN-5877 at 11/19/16 11:55 AM: --- Is this the right thing to do ? NM whitelist configuration is used to allow containers to override NM's default environment. Let us assume there is an environment variable DUMMY_ENV. Assume in some deployment, for some jobs, its set in the ContainerLaunchContext and for all other jobs we want a default which can be set in NM environment. Before the change in the patch, we could have set DUMMY_ENV in NM whitelist and NM would have picked up value for this environment variable from container launch context (if available) otherwise taken it from NM environment. After the change in this patch, NM would keep the value coming from container launch context as default and use that only if the environment variable on NM is not set. This is a departure from current behavior. Probably you can argue none of us are using it this way but its still a change in behavior. We have done the same for HADOOP_YARN_HOME, HADOOP_COMMON_HOME, etc. in YARN-3853 but these are well known environment variables so probably nobody uses them the way I mentioned above. Atleast we do not. So it may be safe to use them this way. If we were to have a different configuration, we can probably get rid of these concerns. Thoughts ? was (Author: varun_saxena): Is this the right thing to do ? NM whitelist configuration is used to allow containers to override NM's default environment. Let us assume there is an environment variable DUMMY_ENV. Assume in some deployment, for some jobs, its set in the ContainerLaunchContext and for all other jobs we want a default which can be set in NM environment. Before the change in the patch, we could have set DUMMY_ENV in NM whitelist and NM would have picked up value for this environment variable from container launch context (if available) otherwise taken it from NM environment. After the change in this patch, NM would keep the value coming from container launch context as default and use that only if the environment variable on NM is not set. This is a departure from current behavior. Probably you can argue none of us are using it this way but its still a change in behavior. We have done the same for HADOOP_YARN_HOME, HADOOP_COMMON_HOME, etc. in YARN-3853 but these are well known environment variables so probably nobody uses them the way I mentioned above. So its safe to use them this way. If we were to have a different configuration, we can probably get rid of these concerns. Thoughts ? > Allow all nm-whitelist-env to get overridden during launch > -- > > Key: YARN-5877 > URL: https://issues.apache.org/jira/browse/YARN-5877 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Attachments: Dockerfile, YARN-5877.0001.patch, YARN-5877.0002.patch, > bootstrap.sh, yarn-site.xml > > > As per the {{yarn.nodemanager.env-whitelist}} for the configured values > should containers may override rather than use NodeManager's default. > {code} > > Environment variables that containers may override rather > than use NodeManager's default. > yarn.nodemanager.env-whitelist > > JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,CLASSPATH_PREPEND_DISTCACHE,HADOOP_YARN_HOME > > {code} > But only the following containers can override > {code} > whitelist.add(ApplicationConstants.Environment.HADOOP_YARN_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_COMMON_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_HDFS_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_CONF_DIR.name()); > whitelist.add(ApplicationConstants.Environment.JAVA_HOME.name()); > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5877) Allow all nm-whitelist-env to get overridden during launch
[ https://issues.apache.org/jira/browse/YARN-5877?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679132#comment-15679132 ] Varun Saxena commented on YARN-5877: Is this the right thing to do ? NM whitelist configuration is used to allow containers to override NM's default environment. Let us assume there is an environment variable DUMMY_ENV. Assume in some deployment, for some jobs, its set in the ContainerLaunchContext and for all other jobs we want a default which can be set in NM environment. Before the change in the patch, we could have set DUMMY_ENV in NM whitelist and NM would have picked up value for this environment variable from container launch context (if available) otherwise taken it from NM environment. After the change in this patch, NM would keep the value coming from container launch context as default and use that only if the environment variable on NM is not set. This is a departure from current behavior. Probably you can argue none of us are using it this way but its still a change in behavior. We have done the same for HADOOP_YARN_HOME, HADOOP_COMMON_HOME, etc. in YARN-3853 but these are well known environment variables so probably nobody uses them the way I mentioned above. So its safe to use them this way. If we were to have a different configuration, we can probably get rid of these concerns. Thoughts ? > Allow all nm-whitelist-env to get overridden during launch > -- > > Key: YARN-5877 > URL: https://issues.apache.org/jira/browse/YARN-5877 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Attachments: Dockerfile, YARN-5877.0001.patch, YARN-5877.0002.patch, > bootstrap.sh, yarn-site.xml > > > As per the {{yarn.nodemanager.env-whitelist}} for the configured values > should containers may override rather than use NodeManager's default. > {code} > > Environment variables that containers may override rather > than use NodeManager's default. > yarn.nodemanager.env-whitelist > > JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,CLASSPATH_PREPEND_DISTCACHE,HADOOP_YARN_HOME > > {code} > But only the following containers can override > {code} > whitelist.add(ApplicationConstants.Environment.HADOOP_YARN_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_COMMON_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_HDFS_HOME.name()); > whitelist.add(ApplicationConstants.Environment.HADOOP_CONF_DIR.name()); > whitelist.add(ApplicationConstants.Environment.JAVA_HOME.name()); > {code} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5548) Use MockRMMemoryStateStore to reduce test failures
[ https://issues.apache.org/jira/browse/YARN-5548?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Varun Saxena updated YARN-5548: --- Summary: Use MockRMMemoryStateStore to reduce test failures (was: Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart) > Use MockRMMemoryStateStore to reduce test failures > -- > > Key: YARN-5548 > URL: https://issues.apache.org/jira/browse/YARN-5548 > Project: Hadoop YARN > Issue Type: Test >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Labels: oct16-easy, test > Attachments: YARN-5548.0001.patch, YARN-5548.0002.patch, > YARN-5548.0003.patch, YARN-5548.0004.patch, YARN-5548.0005.patch > > > https://builds.apache.org/job/PreCommit-YARN-Build/12850/testReport/org.apache.hadoop.yarn.server.resourcemanager/TestRMRestart/testFinishedAppRemovalAfterRMRestart/ > {noformat} > Error Message > Stacktrace > java.lang.AssertionError: expected null, but was: application_submission_context { application_id { id: 1 cluster_timestamp: > 1471885197388 } application_name: "" queue: "default" priority { priority: 0 > } am_container_spec { } cancel_tokens_when_complete: true maxAppAttempts: 2 > resource { memory: 1024 virtual_cores: 1 } applicationType: "YARN" > keep_containers_across_application_attempts: false > attempt_failures_validity_interval: 0 am_container_resource_request { > priority { priority: 0 } resource_name: "*" capability { memory: 1024 > virtual_cores: 1 } num_containers: 0 relax_locality: true > node_label_expression: "" execution_type_request { execution_type: GUARANTEED > enforce_execution_type: false } } } user: "jenkins" start_time: 1471885197417 > application_state: RMAPP_FINISHED finish_time: 1471885197478> > at org.junit.Assert.fail(Assert.java:88) > at org.junit.Assert.failNotNull(Assert.java:664) > at org.junit.Assert.assertNull(Assert.java:646) > at org.junit.Assert.assertNull(Assert.java:656) > at > org.apache.hadoop.yarn.server.resourcemanager.TestRMRestart.testFinishedAppRemovalAfterRMRestart(TestRMRestart.java:1656) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5548) Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart
[ https://issues.apache.org/jira/browse/YARN-5548?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679043#comment-15679043 ] Varun Saxena commented on YARN-5548: [~sunilg], I am fine with your suggestion. It can be useful elsewhere i.e. where MockRM isn't used as well. Although its public class inside MockRM, referring to MockRM in a test where it isn't used would look a little awkward. > Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart > -- > > Key: YARN-5548 > URL: https://issues.apache.org/jira/browse/YARN-5548 > Project: Hadoop YARN > Issue Type: Test >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Labels: oct16-easy, test > Attachments: YARN-5548.0001.patch, YARN-5548.0002.patch, > YARN-5548.0003.patch, YARN-5548.0004.patch, YARN-5548.0005.patch > > > https://builds.apache.org/job/PreCommit-YARN-Build/12850/testReport/org.apache.hadoop.yarn.server.resourcemanager/TestRMRestart/testFinishedAppRemovalAfterRMRestart/ > {noformat} > Error Message > Stacktrace > java.lang.AssertionError: expected null, but was: application_submission_context { application_id { id: 1 cluster_timestamp: > 1471885197388 } application_name: "" queue: "default" priority { priority: 0 > } am_container_spec { } cancel_tokens_when_complete: true maxAppAttempts: 2 > resource { memory: 1024 virtual_cores: 1 } applicationType: "YARN" > keep_containers_across_application_attempts: false > attempt_failures_validity_interval: 0 am_container_resource_request { > priority { priority: 0 } resource_name: "*" capability { memory: 1024 > virtual_cores: 1 } num_containers: 0 relax_locality: true > node_label_expression: "" execution_type_request { execution_type: GUARANTEED > enforce_execution_type: false } } } user: "jenkins" start_time: 1471885197417 > application_state: RMAPP_FINISHED finish_time: 1471885197478> > at org.junit.Assert.fail(Assert.java:88) > at org.junit.Assert.failNotNull(Assert.java:664) > at org.junit.Assert.assertNull(Assert.java:646) > at org.junit.Assert.assertNull(Assert.java:656) > at > org.apache.hadoop.yarn.server.resourcemanager.TestRMRestart.testFinishedAppRemovalAfterRMRestart(TestRMRestart.java:1656) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5548) Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart
[ https://issues.apache.org/jira/browse/YARN-5548?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15679028#comment-15679028 ] Hadoop QA commented on YARN-5548: - | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 16s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 15 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 7m 10s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 34s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 31s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 41s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 18s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 4s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 22s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 35s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 32s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 32s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 29s{color} | {color:orange} hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager: The patch generated 2 new + 799 unchanged - 2 fixed = 801 total (was 801) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 39s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 14s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 13s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 20s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 46m 47s{color} | {color:green} hadoop-yarn-server-resourcemanager in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 21s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 63m 26s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:a9ad5d6 | | JIRA Issue | YARN-5548 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12839680/YARN-5548.0005.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux 35e35cb5f427 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 7584fbf | | Default Java | 1.8.0_111 | | findbugs | v3.0.0 | | checkstyle | https://builds.apache.org/job/PreCommit-YARN-Build/13980/artifact/patchprocess/diff-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/13980/testReport/ | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/13980/console | | Powered by | Apache Yetus 0.4.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart > --
[jira] [Commented] (YARN-5916) Rack shown as null for node decommisioned during registration.
[ https://issues.apache.org/jira/browse/YARN-5916?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15678990#comment-15678990 ] Sunil G commented on YARN-5916: --- Thanks [~bibinchundatt] Patch look straight forward to me. Pending jenkins. > Rack shown as null for node decommisioned during registration. > -- > > Key: YARN-5916 > URL: https://issues.apache.org/jira/browse/YARN-5916 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Minor > Attachments: Screenshot from 2016-11-19 15-22-23.png, > YARN-5916.0001.patch > > > Steps to reproduce > # configure exclude node as nm1 hostname > # start resourcemanager > # start nm1 hostname > # with current configuration during node registration the node gets rejected. > # Now check RM UI -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5916) Rack shown as null for node decommisioned during registration.
[ https://issues.apache.org/jira/browse/YARN-5916?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bibin A Chundatt updated YARN-5916: --- Attachment: Screenshot from 2016-11-19 15-22-23.png > Rack shown as null for node decommisioned during registration. > -- > > Key: YARN-5916 > URL: https://issues.apache.org/jira/browse/YARN-5916 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Minor > Attachments: Screenshot from 2016-11-19 15-22-23.png, > YARN-5916.0001.patch > > > Steps to reproduce > # configure exclude node as nm1 hostname > # start resourcemanager > # start nm1 hostname > # with current configuration during node registration the node gets rejected. > # Now check RM UI -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5916) Rack shown as null for node decommisioned during registration.
[ https://issues.apache.org/jira/browse/YARN-5916?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bibin A Chundatt updated YARN-5916: --- Attachment: YARN-5916.0001.patch Simple fix attaching patch > Rack shown as null for node decommisioned during registration. > -- > > Key: YARN-5916 > URL: https://issues.apache.org/jira/browse/YARN-5916 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt >Priority: Minor > Attachments: Screenshot from 2016-11-19 15-22-23.png, > YARN-5916.0001.patch > > > Steps to reproduce > # configure exclude node as nm1 hostname > # start resourcemanager > # start nm1 hostname > # with current configuration during node registration the node gets rejected. > # Now check RM UI -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Created] (YARN-5916) Rack shown as null for node decommisioned during registration.
Bibin A Chundatt created YARN-5916: -- Summary: Rack shown as null for node decommisioned during registration. Key: YARN-5916 URL: https://issues.apache.org/jira/browse/YARN-5916 Project: Hadoop YARN Issue Type: Bug Reporter: Bibin A Chundatt Assignee: Bibin A Chundatt Priority: Minor Steps to reproduce # configure exclude node as nm1 hostname # start resourcemanager # start nm1 hostname # with current configuration during node registration the node gets rejected. # Now check RM UI -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5548) Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart
[ https://issues.apache.org/jira/browse/YARN-5548?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15678948#comment-15678948 ] Sunil G commented on YARN-5548: --- [~bibinchundatt] [~varun_saxena] Could we take out {{MockRMMemoryStateStore}} from MockRM and place it as a common class in same package. So all other tests could use the same w/o referring MockRM. > Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart > -- > > Key: YARN-5548 > URL: https://issues.apache.org/jira/browse/YARN-5548 > Project: Hadoop YARN > Issue Type: Test >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Labels: oct16-easy, test > Attachments: YARN-5548.0001.patch, YARN-5548.0002.patch, > YARN-5548.0003.patch, YARN-5548.0004.patch, YARN-5548.0005.patch > > > https://builds.apache.org/job/PreCommit-YARN-Build/12850/testReport/org.apache.hadoop.yarn.server.resourcemanager/TestRMRestart/testFinishedAppRemovalAfterRMRestart/ > {noformat} > Error Message > Stacktrace > java.lang.AssertionError: expected null, but was: application_submission_context { application_id { id: 1 cluster_timestamp: > 1471885197388 } application_name: "" queue: "default" priority { priority: 0 > } am_container_spec { } cancel_tokens_when_complete: true maxAppAttempts: 2 > resource { memory: 1024 virtual_cores: 1 } applicationType: "YARN" > keep_containers_across_application_attempts: false > attempt_failures_validity_interval: 0 am_container_resource_request { > priority { priority: 0 } resource_name: "*" capability { memory: 1024 > virtual_cores: 1 } num_containers: 0 relax_locality: true > node_label_expression: "" execution_type_request { execution_type: GUARANTEED > enforce_execution_type: false } } } user: "jenkins" start_time: 1471885197417 > application_state: RMAPP_FINISHED finish_time: 1471885197478> > at org.junit.Assert.fail(Assert.java:88) > at org.junit.Assert.failNotNull(Assert.java:664) > at org.junit.Assert.assertNull(Assert.java:646) > at org.junit.Assert.assertNull(Assert.java:656) > at > org.apache.hadoop.yarn.server.resourcemanager.TestRMRestart.testFinishedAppRemovalAfterRMRestart(TestRMRestart.java:1656) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-5548) Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart
[ https://issues.apache.org/jira/browse/YARN-5548?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bibin A Chundatt updated YARN-5548: --- Attachment: YARN-5548.0005.patch Attaching patch after fix > Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart > -- > > Key: YARN-5548 > URL: https://issues.apache.org/jira/browse/YARN-5548 > Project: Hadoop YARN > Issue Type: Test >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Labels: oct16-easy, test > Attachments: YARN-5548.0001.patch, YARN-5548.0002.patch, > YARN-5548.0003.patch, YARN-5548.0004.patch, YARN-5548.0005.patch > > > https://builds.apache.org/job/PreCommit-YARN-Build/12850/testReport/org.apache.hadoop.yarn.server.resourcemanager/TestRMRestart/testFinishedAppRemovalAfterRMRestart/ > {noformat} > Error Message > Stacktrace > java.lang.AssertionError: expected null, but was: application_submission_context { application_id { id: 1 cluster_timestamp: > 1471885197388 } application_name: "" queue: "default" priority { priority: 0 > } am_container_spec { } cancel_tokens_when_complete: true maxAppAttempts: 2 > resource { memory: 1024 virtual_cores: 1 } applicationType: "YARN" > keep_containers_across_application_attempts: false > attempt_failures_validity_interval: 0 am_container_resource_request { > priority { priority: 0 } resource_name: "*" capability { memory: 1024 > virtual_cores: 1 } num_containers: 0 relax_locality: true > node_label_expression: "" execution_type_request { execution_type: GUARANTEED > enforce_execution_type: false } } } user: "jenkins" start_time: 1471885197417 > application_state: RMAPP_FINISHED finish_time: 1471885197478> > at org.junit.Assert.fail(Assert.java:88) > at org.junit.Assert.failNotNull(Assert.java:664) > at org.junit.Assert.assertNull(Assert.java:646) > at org.junit.Assert.assertNull(Assert.java:656) > at > org.apache.hadoop.yarn.server.resourcemanager.TestRMRestart.testFinishedAppRemovalAfterRMRestart(TestRMRestart.java:1656) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5548) Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart
[ https://issues.apache.org/jira/browse/YARN-5548?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15678898#comment-15678898 ] Varun Saxena commented on YARN-5548: Test failures are related. They are because of the outer MockRM object created while calling MockRM#getMockRMMemoryStateStore. Anyways this will be fixed once you make the change as per suggestion above. > Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart > -- > > Key: YARN-5548 > URL: https://issues.apache.org/jira/browse/YARN-5548 > Project: Hadoop YARN > Issue Type: Test >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Labels: oct16-easy, test > Attachments: YARN-5548.0001.patch, YARN-5548.0002.patch, > YARN-5548.0003.patch, YARN-5548.0004.patch > > > https://builds.apache.org/job/PreCommit-YARN-Build/12850/testReport/org.apache.hadoop.yarn.server.resourcemanager/TestRMRestart/testFinishedAppRemovalAfterRMRestart/ > {noformat} > Error Message > Stacktrace > java.lang.AssertionError: expected null, but was: application_submission_context { application_id { id: 1 cluster_timestamp: > 1471885197388 } application_name: "" queue: "default" priority { priority: 0 > } am_container_spec { } cancel_tokens_when_complete: true maxAppAttempts: 2 > resource { memory: 1024 virtual_cores: 1 } applicationType: "YARN" > keep_containers_across_application_attempts: false > attempt_failures_validity_interval: 0 am_container_resource_request { > priority { priority: 0 } resource_name: "*" capability { memory: 1024 > virtual_cores: 1 } num_containers: 0 relax_locality: true > node_label_expression: "" execution_type_request { execution_type: GUARANTEED > enforce_execution_type: false } } } user: "jenkins" start_time: 1471885197417 > application_state: RMAPP_FINISHED finish_time: 1471885197478> > at org.junit.Assert.fail(Assert.java:88) > at org.junit.Assert.failNotNull(Assert.java:664) > at org.junit.Assert.assertNull(Assert.java:646) > at org.junit.Assert.assertNull(Assert.java:656) > at > org.apache.hadoop.yarn.server.resourcemanager.TestRMRestart.testFinishedAppRemovalAfterRMRestart(TestRMRestart.java:1656) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5548) Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart
[ https://issues.apache.org/jira/browse/YARN-5548?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15678884#comment-15678884 ] Varun Saxena commented on YARN-5548: Thanks [~bibinchundatt] for the patch. I would turn MockRMMemoryStateStore to a public static class so that we need not create outer MockRM object. Then we can either call new MockRM.MockRMMemoryStateStore() from the test or you can provide a utility method in MockRM too. We can also get the store from RM context because by default store chosen is MemoryRMStateStore and it will be overriden by MockRMMemoryStateStore in MockRM while creating it. > Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart > -- > > Key: YARN-5548 > URL: https://issues.apache.org/jira/browse/YARN-5548 > Project: Hadoop YARN > Issue Type: Test >Reporter: Bibin A Chundatt >Assignee: Bibin A Chundatt > Labels: oct16-easy, test > Attachments: YARN-5548.0001.patch, YARN-5548.0002.patch, > YARN-5548.0003.patch, YARN-5548.0004.patch > > > https://builds.apache.org/job/PreCommit-YARN-Build/12850/testReport/org.apache.hadoop.yarn.server.resourcemanager/TestRMRestart/testFinishedAppRemovalAfterRMRestart/ > {noformat} > Error Message > Stacktrace > java.lang.AssertionError: expected null, but was: application_submission_context { application_id { id: 1 cluster_timestamp: > 1471885197388 } application_name: "" queue: "default" priority { priority: 0 > } am_container_spec { } cancel_tokens_when_complete: true maxAppAttempts: 2 > resource { memory: 1024 virtual_cores: 1 } applicationType: "YARN" > keep_containers_across_application_attempts: false > attempt_failures_validity_interval: 0 am_container_resource_request { > priority { priority: 0 } resource_name: "*" capability { memory: 1024 > virtual_cores: 1 } num_containers: 0 relax_locality: true > node_label_expression: "" execution_type_request { execution_type: GUARANTEED > enforce_execution_type: false } } } user: "jenkins" start_time: 1471885197417 > application_state: RMAPP_FINISHED finish_time: 1471885197478> > at org.junit.Assert.fail(Assert.java:88) > at org.junit.Assert.failNotNull(Assert.java:664) > at org.junit.Assert.assertNull(Assert.java:646) > at org.junit.Assert.assertNull(Assert.java:656) > at > org.apache.hadoop.yarn.server.resourcemanager.TestRMRestart.testFinishedAppRemovalAfterRMRestart(TestRMRestart.java:1656) > {noformat} -- This message was sent by Atlassian JIRA (v6.3.4#6332) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-5548) Random test failure TestRMRestart#testFinishedAppRemovalAfterRMRestart
[ https://issues.apache.org/jira/browse/YARN-5548?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15678879#comment-15678879 ] Hadoop QA commented on YARN-5548: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 17s{color} | {color:blue} Docker mode activated. {color} | | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 15 new or modified test files. {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 6m 47s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 33s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 30s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 39s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 18s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 57s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 21s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 31s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 31s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 31s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 28s{color} | {color:orange} hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager: The patch generated 1 new + 800 unchanged - 1 fixed = 801 total (was 801) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 36s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 15s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 3s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 19s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 42m 0s{color} | {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 16s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 57m 44s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.yarn.server.resourcemanager.security.TestRMDelegationTokens | | | hadoop.yarn.server.resourcemanager.TestRMRestart | | | hadoop.yarn.server.resourcemanager.scheduler.fair.TestFairScheduler | | | hadoop.yarn.server.resourcemanager.security.TestDelegationTokenRenewer | | | hadoop.yarn.server.resourcemanager.TestWorkPreservingRMRestart | \\ \\ || Subsystem || Report/Notes || | Docker | Image:yetus/hadoop:a9ad5d6 | | JIRA Issue | YARN-5548 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12839679/YARN-5548.0004.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit findbugs checkstyle | | uname | Linux cdf892b15f7e 3.13.0-95-generic #142-Ubuntu SMP Fri Aug 12 17:00:09 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 7584fbf | | Default Java | 1.8.0_111 | | findbugs | v3.0.0 | | checkstyle | https://builds.apache.org/job/PreCommit-YARN-Build/13979/artifact/patchprocess/diff-checkstyle-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt | | unit | https://builds.apache.org/job/PreCommit-YARN-Build/13979/artifact/patchprocess/patch-unit-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-resourcemanager.txt | | Test Results | https://builds.apache.org/job/P