[jira] [Commented] (YARN-9996) Code cleanup in QueueAdminConfigurationMutationACLPolicy
[ https://issues.apache.org/jira/browse/YARN-9996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082813#comment-17082813 ] Hadoop QA commented on YARN-9996: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 43s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 21m 32s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 48s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 34s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 53s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 16m 2s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 44s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 31s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 42s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 42s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 31s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 14m 38s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 50s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 28s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 95m 40s{color} | {color:green} hadoop-yarn-server-resourcemanager in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 28s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}158m 28s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=19.03.8 Server=19.03.8 Image:yetus/hadoop:e6455cc864d | | JIRA Issue | YARN-9996 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12999623/YARN-9996.001.patch | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux 2febc83a53f7 4.15.0-74-generic #84-Ubuntu SMP Thu Dec 19 08:06:28 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 3edbe87 | | maven | version: Apache Maven 3.3.9 | | Default Java | 1.8.0_242 | | findbugs | v3.1.0-RC1 | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/25887/testReport/ | | Max. process+thread count | 821 (vs. ulimit of 5500) | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/25887/console | | Powered by | Apache Yetus 0.8.0
[jira] [Commented] (YARN-9996) Code cleanup in QueueAdminConfigurationMutationACLPolicy
[ https://issues.apache.org/jira/browse/YARN-9996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082768#comment-17082768 ] Siddharth Ahuja commented on YARN-9996: --- My apologies. [~wilfreds] just pointed me to the fact that I hadn't actually submitted the patch which would have changed the status of the JIRA to Patch Available and this would have triggered the build. I assumed that attaching a patch would trigger the build which is incorrect. > Code cleanup in QueueAdminConfigurationMutationACLPolicy > > > Key: YARN-9996 > URL: https://issues.apache.org/jira/browse/YARN-9996 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Szilard Nemeth >Assignee: Siddharth Ahuja >Priority: Major > Attachments: YARN-9996.001.patch > > > Method 'isMutationAllowed' contains many uses of substring and lastIndexOf. > These could be extracted and simplified. > Also, some logging could be added as well. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-8680) YARN NM: Implement Iterable Abstraction for LocalResourceTracker state
[ https://issues.apache.org/jira/browse/YARN-8680?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082724#comment-17082724 ] Hadoop QA commented on YARN-8680: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 16m 46s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 2 new or modified test files. {color} | || || || || {color:brown} branch-2.10 Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 9m 59s{color} | {color:green} branch-2.10 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 54s{color} | {color:green} branch-2.10 passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 48s{color} | {color:green} branch-2.10 passed with JDK v1.8.0_242 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 25s{color} | {color:green} branch-2.10 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 37s{color} | {color:green} branch-2.10 passed {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 57s{color} | {color:red} hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager in branch-2.10 has 4 extant Findbugs warnings. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 26s{color} | {color:green} branch-2.10 passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 23s{color} | {color:green} branch-2.10 passed with JDK v1.8.0_242 {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 32s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 51s{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 51s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 46s{color} | {color:green} the patch passed with JDK v1.8.0_242 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 46s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 20s{color} | {color:orange} hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager: The patch generated 1 new + 242 unchanged - 5 fixed = 243 total (was 247) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 31s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 4s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 23s{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 19s{color} | {color:green} the patch passed with JDK v1.8.0_242 {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 15m 7s{color} | {color:green} hadoop-yarn-server-nodemanager in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 24s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 53m 14s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=19.03.8 Server=19.03.8 Image:yetus/hadoop:bfeec47759e | | JIRA Issue | YARN-8680 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12999831/YARN-8680-branch-2.10.001.patch | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux 9ab0a74b4ae1 4.15.0-74-generic #84-Ubuntu SMP Thu Dec 19 08:06:28 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality |
[jira] [Reopened] (YARN-8680) YARN NM: Implement Iterable Abstraction for LocalResourceTracker state
[ https://issues.apache.org/jira/browse/YARN-8680?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Jim Brennan reopened YARN-8680: --- This was a follow-up to YARN-8242, which was pulled back to branch-2.10 via YARN-10227. It would be good to pull this back as well. > YARN NM: Implement Iterable Abstraction for LocalResourceTracker state > -- > > Key: YARN-8680 > URL: https://issues.apache.org/jira/browse/YARN-8680 > Project: Hadoop YARN > Issue Type: Improvement > Components: yarn >Reporter: Pradeep Ambati >Assignee: Pradeep Ambati >Priority: Critical > Fix For: 3.2.0, 3.1.2 > > Attachments: YARN-8680.00.patch, YARN-8680.01.patch, > YARN-8680.02.patch, YARN-8680.03.patch, YARN-8680.04.patch > > > Similar to YARN-8242, implement iterable abstraction for > LocalResourceTrackerState to load completed and in progress resources when > needed rather than loading them all at a time for a respective state. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10219) YARN service placement constraints is broken
[ https://issues.apache.org/jira/browse/YARN-10219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082652#comment-17082652 ] Hadoop QA commented on YARN-10219: -- | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 1m 31s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 2 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 19s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 23m 7s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 7s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 25s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 6s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 17m 38s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 0s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 49s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 17s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 41s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 7m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 7m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 0s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 14m 53s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 6s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 48s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 19m 48s{color} | {color:green} hadoop-yarn-services-core in the patch passed. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 0m 21s{color} | {color:green} hadoop-yarn-site in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 42s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}103m 14s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=19.03.8 Server=19.03.8 Image:yetus/hadoop:e6455cc864d | | JIRA Issue | YARN-10219 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12999814/YARN-10219.005.patch | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux 36df2906a9c2 4.15.0-74-generic #84-Ubuntu SMP Thu Dec 19 08:06:28 UTC 2019 x86_64
[jira] [Updated] (YARN-10219) YARN service placement constraints is broken
[ https://issues.apache.org/jira/browse/YARN-10219?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eric Yang updated YARN-10219: - Attachment: YARN-10219.005.patch > YARN service placement constraints is broken > > > Key: YARN-10219 > URL: https://issues.apache.org/jira/browse/YARN-10219 > Project: Hadoop YARN > Issue Type: Bug >Affects Versions: 3.1.0, 3.2.0, 3.1.1, 3.1.2, 3.3.0, 3.2.1, 3.1.3 >Reporter: Eric Yang >Assignee: Eric Yang >Priority: Blocker > Attachments: YARN-10219.001.patch, YARN-10219.002.patch, > YARN-10219.003.patch, YARN-10219.004.patch, YARN-10219.005.patch > > > YARN service placement constraint does not work with node label nor node > attributes. Example of placement constraints: > {code} > "placement_policy": { > "constraints": [ > { > "type": "AFFINITY", > "scope": "NODE", > "node_attributes": { > "label":["genfile"] > }, > "target_tags": [ > "ping" > ] > } > ] > }, > {code} > Node attribute added: > {code} ./bin/yarn nodeattributes -add "host-3.example.com:label=genfile" > {code} > Scheduling activities shows: > {code} Node does not match partition or placement constraints, > unsatisfied PC expression="in,node,ping", target-type=ALLOCATION_TAG > > 1 > host-3.example.com:45454{code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10219) YARN service placement constraints is broken
[ https://issues.apache.org/jira/browse/YARN-10219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082519#comment-17082519 ] Hadoop QA commented on YARN-10219: -- | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 39s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 2 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 24s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 21m 33s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 8m 17s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 1m 26s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 8s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 17m 43s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 1s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 50s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 16s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 40s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 7m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 7m 45s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 1m 28s{color} | {color:orange} hadoop-yarn-project/hadoop-yarn: The patch generated 2 new + 47 unchanged - 0 fixed = 49 total (was 47) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 3s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 15m 2s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 6s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 46s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 19m 31s{color} | {color:green} hadoop-yarn-services-core in the patch passed. {color} | | {color:green}+1{color} | {color:green} unit {color} | {color:green} 0m 21s{color} | {color:green} hadoop-yarn-site in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 42s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}101m 23s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=19.03.8 Server=19.03.8 Image:yetus/hadoop:e6455cc864d | | JIRA Issue | YARN-10219 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12999803/YARN-10219.004.patch | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | |
[jira] [Assigned] (YARN-8450) Blocking resources such as GPU/FPGA etc tend to release actual device slowly even after RM identifies it as COMPLETED
[ https://issues.apache.org/jira/browse/YARN-8450?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bilwa S T reassigned YARN-8450: --- Assignee: (was: Bilwa S T) > Blocking resources such as GPU/FPGA etc tend to release actual device slowly > even after RM identifies it as COMPLETED > - > > Key: YARN-8450 > URL: https://issues.apache.org/jira/browse/YARN-8450 > Project: Hadoop YARN > Issue Type: Bug > Components: nodemanager >Affects Versions: 3.0.2 >Reporter: Sunil G >Priority: Major > > For resources such as GPU/FPGA or similar resources, sometimes we have seen > that device is not released from a container even after container is in > completed states. > In such cases, we need a common way of handling from NM level. YARN-8423 is > only handling this for GPU. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Assigned] (YARN-6431) make DELETE/STOP/CONVERT queues work in reservation system
[ https://issues.apache.org/jira/browse/YARN-6431?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bilwa S T reassigned YARN-6431: --- Assignee: (was: Bilwa S T) > make DELETE/STOP/CONVERT queues work in reservation system > -- > > Key: YARN-6431 > URL: https://issues.apache.org/jira/browse/YARN-6431 > Project: Hadoop YARN > Issue Type: Sub-task > Components: yarn >Reporter: Xuan Gong >Priority: Major > > Previous, we have made some enhancements on DELETE/STOP/CONVERT queues. We > need to make sure that those enhancements work for reservation system as well. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Assigned] (YARN-9459) ConfiguredYarnAuthorizer should not now any details about queues
[ https://issues.apache.org/jira/browse/YARN-9459?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bilwa S T reassigned YARN-9459: --- Assignee: Bilwa S T > ConfiguredYarnAuthorizer should not now any details about queues > > > Key: YARN-9459 > URL: https://issues.apache.org/jira/browse/YARN-9459 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Szilard Nemeth >Assignee: Bilwa S T >Priority: Major > > ConfiguredYarnAuthorizer should not now anything about queues, as this is not > its responsibility, see > [here|https://github.com/apache/hadoop/blob/39b4a37e02e929a698fcf9e32f1f71bb6b977635/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/security/ConfiguredYarnAuthorizer.java#L70] > Code like this could be part of the QueueACLsManager. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Assigned] (YARN-9641) ApplicationMasterService#AllocateResponseLock need to hold only ResponseID
[ https://issues.apache.org/jira/browse/YARN-9641?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bilwa S T reassigned YARN-9641: --- Assignee: Bilwa S T > ApplicationMasterService#AllocateResponseLock need to hold only ResponseID > --- > > Key: YARN-9641 > URL: https://issues.apache.org/jira/browse/YARN-9641 > Project: Hadoop YARN > Issue Type: Bug >Reporter: Bibin Chundatt >Assignee: Bilwa S T >Priority: Major > > ApplicationMasterService ApplicationResponseLock could occupy too much memory. > # 5K+ node > # AllocateResponse could contain 5K nodes NodeReport, Container Allocated > data too.. > Only the responseId is used in ApplicationMaster Service.. No need to hold > all the data. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-10219) YARN service placement constraints is broken
[ https://issues.apache.org/jira/browse/YARN-10219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082437#comment-17082437 ] Eric Yang commented on YARN-10219: -- [~prabhujoseph] Patch 4 updated the indentation and reduce number of vcore that can be used for the test to pass on more powerful nodes. > YARN service placement constraints is broken > > > Key: YARN-10219 > URL: https://issues.apache.org/jira/browse/YARN-10219 > Project: Hadoop YARN > Issue Type: Bug >Affects Versions: 3.1.0, 3.2.0, 3.1.1, 3.1.2, 3.3.0, 3.2.1, 3.1.3 >Reporter: Eric Yang >Assignee: Eric Yang >Priority: Blocker > Attachments: YARN-10219.001.patch, YARN-10219.002.patch, > YARN-10219.003.patch, YARN-10219.004.patch > > > YARN service placement constraint does not work with node label nor node > attributes. Example of placement constraints: > {code} > "placement_policy": { > "constraints": [ > { > "type": "AFFINITY", > "scope": "NODE", > "node_attributes": { > "label":["genfile"] > }, > "target_tags": [ > "ping" > ] > } > ] > }, > {code} > Node attribute added: > {code} ./bin/yarn nodeattributes -add "host-3.example.com:label=genfile" > {code} > Scheduling activities shows: > {code} Node does not match partition or placement constraints, > unsatisfied PC expression="in,node,ping", target-type=ALLOCATION_TAG > > 1 > host-3.example.com:45454{code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10219) YARN service placement constraints is broken
[ https://issues.apache.org/jira/browse/YARN-10219?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eric Yang updated YARN-10219: - Attachment: YARN-10219.004.patch > YARN service placement constraints is broken > > > Key: YARN-10219 > URL: https://issues.apache.org/jira/browse/YARN-10219 > Project: Hadoop YARN > Issue Type: Bug >Affects Versions: 3.1.0, 3.2.0, 3.1.1, 3.1.2, 3.3.0, 3.2.1, 3.1.3 >Reporter: Eric Yang >Assignee: Eric Yang >Priority: Blocker > Attachments: YARN-10219.001.patch, YARN-10219.002.patch, > YARN-10219.003.patch, YARN-10219.004.patch > > > YARN service placement constraint does not work with node label nor node > attributes. Example of placement constraints: > {code} > "placement_policy": { > "constraints": [ > { > "type": "AFFINITY", > "scope": "NODE", > "node_attributes": { > "label":["genfile"] > }, > "target_tags": [ > "ping" > ] > } > ] > }, > {code} > Node attribute added: > {code} ./bin/yarn nodeattributes -add "host-3.example.com:label=genfile" > {code} > Scheduling activities shows: > {code} Node does not match partition or placement constraints, > unsatisfied PC expression="in,node,ping", target-type=ALLOCATION_TAG > > 1 > host-3.example.com:45454{code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-10219) YARN service placement constraints is broken
[ https://issues.apache.org/jira/browse/YARN-10219?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eric Yang updated YARN-10219: - Target Version/s: 3.3.0 (was: 3.4.0) Priority: Blocker (was: Major) > YARN service placement constraints is broken > > > Key: YARN-10219 > URL: https://issues.apache.org/jira/browse/YARN-10219 > Project: Hadoop YARN > Issue Type: Bug >Affects Versions: 3.1.0, 3.2.0, 3.1.1, 3.1.2, 3.3.0, 3.2.1, 3.1.3 >Reporter: Eric Yang >Assignee: Eric Yang >Priority: Blocker > Attachments: YARN-10219.001.patch, YARN-10219.002.patch, > YARN-10219.003.patch > > > YARN service placement constraint does not work with node label nor node > attributes. Example of placement constraints: > {code} > "placement_policy": { > "constraints": [ > { > "type": "AFFINITY", > "scope": "NODE", > "node_attributes": { > "label":["genfile"] > }, > "target_tags": [ > "ping" > ] > } > ] > }, > {code} > Node attribute added: > {code} ./bin/yarn nodeattributes -add "host-3.example.com:label=genfile" > {code} > Scheduling activities shows: > {code} Node does not match partition or placement constraints, > unsatisfied PC expression="in,node,ping", target-type=ALLOCATION_TAG > > 1 > host-3.example.com:45454{code} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-9954) Configurable max application tags and max tag length
[ https://issues.apache.org/jira/browse/YARN-9954?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082306#comment-17082306 ] Bilwa S T commented on YARN-9954: - Hi [~jhung] / [~brahma] could u please help to review the patch ?? > Configurable max application tags and max tag length > > > Key: YARN-9954 > URL: https://issues.apache.org/jira/browse/YARN-9954 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Jonathan Hung >Assignee: Bilwa S T >Priority: Major > Attachments: YARN-9954.001.patch > > > Currently max tags and max tag length is hardcoded, it should be configurable > {noformat} > @Evolving > public static final int APPLICATION_MAX_TAGS = 10; > @Evolving > public static final int APPLICATION_MAX_TAG_LENGTH = 100; {noformat} -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-9355) RMContainerRequestor#makeRemoteRequest has confusing log message
[ https://issues.apache.org/jira/browse/YARN-9355?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082142#comment-17082142 ] Hadoop QA commented on YARN-9355: - | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 44s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:red}-1{color} | {color:red} test4tests {color} | {color:red} 0m 0s{color} | {color:red} The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 22m 50s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 29s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 24s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 32s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 15m 31s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 45s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 24s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 32s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 25s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 25s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 17s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 14m 38s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 50s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 20s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 8m 18s{color} | {color:green} hadoop-mapreduce-client-app in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 30s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 68m 16s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=19.03.8 Server=19.03.8 Image:yetus/hadoop:e6455cc864d | | JIRA Issue | YARN-9355 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12999750/YARN-9355.001.patch | | Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux 79de89c66fe0 4.15.0-74-generic #84-Ubuntu SMP Thu Dec 19 08:06:28 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 8d49229 | | maven | version: Apache Maven 3.3.9 | | Default Java | 1.8.0_242 | | findbugs | v3.1.0-RC1 | | Test Results | https://builds.apache.org/job/PreCommit-YARN-Build/25883/testReport/ | | Max. process+thread count | 571 (vs. ulimit of 5500) | | modules | C: hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app U: hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app | | Console output | https://builds.apache.org/job/PreCommit-YARN-Build/25883/console | | Powered by | Apache Yetus 0.8.0 http://yetus.apache.org |
[jira] [Comment Edited] (YARN-9355) RMContainerRequestor#makeRemoteRequest has confusing log message
[ https://issues.apache.org/jira/browse/YARN-9355?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082109#comment-17082109 ] Umesh Mittal edited comment on YARN-9355 at 4/13/20, 7:07 AM: -- Hi [Szilard,|https://issues.apache.org/jira/secure/ViewProfile.jspa?name=snemeth] I have finished writing up a patch for this improvement request. There in no JUNIT updated as this is a very cosmetic change. Thanks was (Author: umittal): Hi [Szilard|https://issues.apache.org/jira/secure/ViewProfile.jspa?name=snemeth] I have finished writing up a patch for this improvement request. There in no JUNIT updated as this is a very cosmetic change. Thanks > RMContainerRequestor#makeRemoteRequest has confusing log message > > > Key: YARN-9355 > URL: https://issues.apache.org/jira/browse/YARN-9355 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Szilard Nemeth >Assignee: Umesh Mittal >Priority: Trivial > Labels: newbie, newbie++ > Attachments: YARN-9355.001.patch > > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor#makeRemoteRequest > has this log: > {code:java} > if (ask.size() > 0 || release.size() > 0) { > LOG.info("getResources() for " + applicationId + ":" + " ask=" > + ask.size() + " release= " + release.size() + " newContainers=" > + allocateResponse.getAllocatedContainers().size() > + " finishedContainers=" + numCompletedContainers > + " resourcelimit=" + availableResources + " knownNMs=" > + clusterNmCount); > } > {code} > The reason why "getResources()" is printed because > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator#getResources > invokes makeRemoteRequest. This is not too informative and error-prone as > name of getResources could change over time and the log will be outdated. > Moreover, it's not a good idea to print a method name from a method below the > current one in the stack. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (YARN-9355) RMContainerRequestor#makeRemoteRequest has confusing log message
[ https://issues.apache.org/jira/browse/YARN-9355?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082109#comment-17082109 ] Umesh Mittal edited comment on YARN-9355 at 4/13/20, 7:06 AM: -- Hi [Szilard|https://issues.apache.org/jira/secure/ViewProfile.jspa?name=snemeth] I have finished writing up a patch for this improvement request. There in no JUNIT updated as this is a very cosmetic change. Thanks was (Author: umittal): Hi [Szilard|https://issues.apache.org/jira/secure/ViewProfile.jspa?name=snemeth] I have finished writing up a patch for this improvement request. Thanks > RMContainerRequestor#makeRemoteRequest has confusing log message > > > Key: YARN-9355 > URL: https://issues.apache.org/jira/browse/YARN-9355 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Szilard Nemeth >Assignee: Umesh Mittal >Priority: Trivial > Labels: newbie, newbie++ > Attachments: YARN-9355.001.patch > > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor#makeRemoteRequest > has this log: > {code:java} > if (ask.size() > 0 || release.size() > 0) { > LOG.info("getResources() for " + applicationId + ":" + " ask=" > + ask.size() + " release= " + release.size() + " newContainers=" > + allocateResponse.getAllocatedContainers().size() > + " finishedContainers=" + numCompletedContainers > + " resourcelimit=" + availableResources + " knownNMs=" > + clusterNmCount); > } > {code} > The reason why "getResources()" is printed because > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator#getResources > invokes makeRemoteRequest. This is not too informative and error-prone as > name of getResources could change over time and the log will be outdated. > Moreover, it's not a good idea to print a method name from a method below the > current one in the stack. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (YARN-9355) RMContainerRequestor#makeRemoteRequest has confusing log message
[ https://issues.apache.org/jira/browse/YARN-9355?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082109#comment-17082109 ] Umesh Mittal edited comment on YARN-9355 at 4/13/20, 7:03 AM: -- Hi [Szilard|https://issues.apache.org/jira/secure/ViewProfile.jspa?name=snemeth] I have finished writing up a patch for this improvement request. was (Author: umittal): Hi [Szilard.|https://issues.apache.org/jira/secure/ViewProfile.jspa?name=snemeth] I have finished writing up a patch for this improvement request. > RMContainerRequestor#makeRemoteRequest has confusing log message > > > Key: YARN-9355 > URL: https://issues.apache.org/jira/browse/YARN-9355 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Szilard Nemeth >Assignee: Umesh Mittal >Priority: Trivial > Labels: newbie, newbie++ > Attachments: YARN-9355.001.patch > > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor#makeRemoteRequest > has this log: > {code:java} > if (ask.size() > 0 || release.size() > 0) { > LOG.info("getResources() for " + applicationId + ":" + " ask=" > + ask.size() + " release= " + release.size() + " newContainers=" > + allocateResponse.getAllocatedContainers().size() > + " finishedContainers=" + numCompletedContainers > + " resourcelimit=" + availableResources + " knownNMs=" > + clusterNmCount); > } > {code} > The reason why "getResources()" is printed because > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator#getResources > invokes makeRemoteRequest. This is not too informative and error-prone as > name of getResources could change over time and the log will be outdated. > Moreover, it's not a good idea to print a method name from a method below the > current one in the stack. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Comment Edited] (YARN-9355) RMContainerRequestor#makeRemoteRequest has confusing log message
[ https://issues.apache.org/jira/browse/YARN-9355?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082109#comment-17082109 ] Umesh Mittal edited comment on YARN-9355 at 4/13/20, 7:03 AM: -- Hi [Szilard|https://issues.apache.org/jira/secure/ViewProfile.jspa?name=snemeth] I have finished writing up a patch for this improvement request. Thanks was (Author: umittal): Hi [Szilard|https://issues.apache.org/jira/secure/ViewProfile.jspa?name=snemeth] I have finished writing up a patch for this improvement request. > RMContainerRequestor#makeRemoteRequest has confusing log message > > > Key: YARN-9355 > URL: https://issues.apache.org/jira/browse/YARN-9355 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Szilard Nemeth >Assignee: Umesh Mittal >Priority: Trivial > Labels: newbie, newbie++ > Attachments: YARN-9355.001.patch > > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor#makeRemoteRequest > has this log: > {code:java} > if (ask.size() > 0 || release.size() > 0) { > LOG.info("getResources() for " + applicationId + ":" + " ask=" > + ask.size() + " release= " + release.size() + " newContainers=" > + allocateResponse.getAllocatedContainers().size() > + " finishedContainers=" + numCompletedContainers > + " resourcelimit=" + availableResources + " knownNMs=" > + clusterNmCount); > } > {code} > The reason why "getResources()" is printed because > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator#getResources > invokes makeRemoteRequest. This is not too informative and error-prone as > name of getResources could change over time and the log will be outdated. > Moreover, it's not a good idea to print a method name from a method below the > current one in the stack. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Commented] (YARN-9355) RMContainerRequestor#makeRemoteRequest has confusing log message
[ https://issues.apache.org/jira/browse/YARN-9355?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17082109#comment-17082109 ] Umesh Mittal commented on YARN-9355: Hi [Szilard.|https://issues.apache.org/jira/secure/ViewProfile.jspa?name=snemeth] I have finished writing up a patch for this improvement request. > RMContainerRequestor#makeRemoteRequest has confusing log message > > > Key: YARN-9355 > URL: https://issues.apache.org/jira/browse/YARN-9355 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Szilard Nemeth >Assignee: Umesh Mittal >Priority: Trivial > Labels: newbie, newbie++ > Attachments: YARN-9355.001.patch > > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor#makeRemoteRequest > has this log: > {code:java} > if (ask.size() > 0 || release.size() > 0) { > LOG.info("getResources() for " + applicationId + ":" + " ask=" > + ask.size() + " release= " + release.size() + " newContainers=" > + allocateResponse.getAllocatedContainers().size() > + " finishedContainers=" + numCompletedContainers > + " resourcelimit=" + availableResources + " knownNMs=" > + clusterNmCount); > } > {code} > The reason why "getResources()" is printed because > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator#getResources > invokes makeRemoteRequest. This is not too informative and error-prone as > name of getResources could change over time and the log will be outdated. > Moreover, it's not a good idea to print a method name from a method below the > current one in the stack. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org
[jira] [Updated] (YARN-9355) RMContainerRequestor#makeRemoteRequest has confusing log message
[ https://issues.apache.org/jira/browse/YARN-9355?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Umesh Mittal updated YARN-9355: --- Attachment: YARN-9355.001.patch > RMContainerRequestor#makeRemoteRequest has confusing log message > > > Key: YARN-9355 > URL: https://issues.apache.org/jira/browse/YARN-9355 > Project: Hadoop YARN > Issue Type: Improvement >Reporter: Szilard Nemeth >Assignee: Umesh Mittal >Priority: Trivial > Labels: newbie, newbie++ > Attachments: YARN-9355.001.patch > > > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerRequestor#makeRemoteRequest > has this log: > {code:java} > if (ask.size() > 0 || release.size() > 0) { > LOG.info("getResources() for " + applicationId + ":" + " ask=" > + ask.size() + " release= " + release.size() + " newContainers=" > + allocateResponse.getAllocatedContainers().size() > + " finishedContainers=" + numCompletedContainers > + " resourcelimit=" + availableResources + " knownNMs=" > + clusterNmCount); > } > {code} > The reason why "getResources()" is printed because > org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator#getResources > invokes makeRemoteRequest. This is not too informative and error-prone as > name of getResources could change over time and the log will be outdated. > Moreover, it's not a good idea to print a method name from a method below the > current one in the stack. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org