[jira] [Commented] (YARN-9694) UI always show default-rack for all the nodes while running SLS.

2019-07-26 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9694?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16894259#comment-16894259
 ] 

Hadoop QA commented on YARN-9694:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
23s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 20m 
 8s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
28s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
28s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
52s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
11m 54s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
37s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
25s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
15s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
26s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
12m 24s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  0m 
50s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
18s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 10m 26s{color} 
| {color:red} hadoop-sls in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
29s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 61m  9s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | hadoop.yarn.sls.TestSLSRunner |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=19.03.1 Server=19.03.1 Image:yetus/hadoop:bdbca0e |
| JIRA Issue | YARN-9694 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12976019/YARN-9694.001.patch |
| Optional Tests |  dupname  asflicense  compile  javac  javadoc  mvninstall  
mvnsite  unit  shadedclient  findbugs  checkstyle  |
| uname | Linux c2662973b50e 4.4.0-138-generic #164-Ubuntu SMP Tue Oct 2 
17:16:02 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/patchprocess/precommit/personality/provided.sh |
| git revision | trunk / 2fe450c |
| maven | version: Apache Maven 3.3.9 |
| Default Java | 1.8.0_212 |
| findbugs | v3.1.0-RC1 |
| unit | 
https://builds.apache.org/job/PreCommit-YARN-Build/24430/artifact/out/patch-unit-hadoop-tools_hadoop-sls.txt
 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/24430/testReport/ |
| Max. process+thread count | 449 (vs. ulimit of 1) |
| modules | C: hadoop-tools/hadoop-sls U: hadoop-tools/hadoop-sls |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/24430/console |
| Powered by | Apache Yetus 0.8.0   http://yetus.apache.org |


This message was automatically generated.



> UI always show 

[jira] [Comment Edited] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16894019#comment-16894019
 ] 

Babble Shack edited comment on YARN-9690 at 7/26/19 6:08 PM:
-

I also tried creating a new configuration directory and disabling 
_yarn.resourcemanager.scheduler_ and using it default value at the resource 
manager.
{code:java}



{code}
The new configuration:
{code:java}

 
       yarn.resourcemanager.scheduler.address
       localhost:8049
   {code}
And launching the job with
{code:java}
yarn --config /debug/hadoop/etc/hadoop jar $MAPRED_EXAMPLES pi 
-Dmapreduce.job.num-opportunistic-maps-percent="40" 10 10
{code}
The nodemanager then tries to connect with scheduler on 0.0.0.0:8030, however 
the amrmproxy has already started, and so the amrmproxy should respond by 
redirecting the request to the Local RM
{code:java}
//Nodemanager logs
2019-07-26 17:32:25,480 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 container_1564162333195_0001_01_01's ip = 10.244.29.84, and hostname = 
yarn-worker-2.yarn-service.yarn.svc.cluster.local
2019-07-26 17:32:26,962 INFO SecurityLogger.org.apache.hadoop.ipc.Server: Auth 
successful for appattempt_1564162333195_0001_01 (auth:SIMPLE)
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.AMRMProxyService: 
Registering application master. 
Host:yarn-worker-2.yarn-service.yarn.svc.cluster.local Port:42172 Tracking 
Url:http://yarn-worker-2.yarn-service.yarn.svc.cluster.local:37110
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler: 
Forwarding registration request to theDistributed Scheduler Service on YARN RM
2019-07-26 17:32:26,966 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.DefaultRequestInterceptor: 
Forwarding registerApplicationMasterForDistributedSchedulingrequest to the real 
YARN RM
2019-07-26 17:32:27,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 0 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
2019-07-26 17:32:28,485 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 Skipping monitoring container container_1564162333195_0001_01_01 since CPU 
usage is not yet available.
2019-07-26 17:32:28,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 1 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

{code}
However there is no service listening on 0.0.0.0:8030

The application log, shows AM is connecting to AMRMproxy
{code:java}
// AM Logs
2019-07-26 17:50:17,238 INFO [main] org.apache.hadoop.yarn.client.RMProxy: 
Connecting to ResourceManager at localhost/127.0.0.1:8049
End of LogType:syslog.This log file belongs to a running container 
(container_1564163395606_0001_01_01) and so may not be complete.
***
{code}


was (Author: babbleshack):
I also tried disabling _yarn.resourcemanager.scheduler_ and using it default 
value at the resource manager.
{code:java}



{code}
I then create a new config directory with the following
{code:java}

 
       yarn.resourcemanager.scheduler.address
       localhost:8049
   {code}
And launching the job with
{code:java}
yarn --config /debug/hadoop/etc/hadoop jar $MAPRED_EXAMPLES pi 
-Dmapreduce.job.num-opportunistic-maps-percent="40" 10 10
{code}
The nodemanager then tries to connect with scheduler on 0.0.0.0:8030, however 
the amrmproxy has already started, and so the amrmproxy should respond by 
redirecting the request to the Local RM
{code:java}
//Nodemanager logs
2019-07-26 17:32:25,480 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 container_1564162333195_0001_01_01's ip = 10.244.29.84, and hostname = 
yarn-worker-2.yarn-service.yarn.svc.cluster.local
2019-07-26 17:32:26,962 INFO SecurityLogger.org.apache.hadoop.ipc.Server: Auth 
successful for appattempt_1564162333195_0001_01 (auth:SIMPLE)
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.AMRMProxyService: 
Registering application master. 
Host:yarn-worker-2.yarn-service.yarn.svc.cluster.local Port:42172 Tracking 
Url:http://yarn-worker-2.yarn-service.yarn.svc.cluster.local:37110
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler: 
Forwarding registration request to theDistributed Scheduler Service on YARN RM
2019-07-26 17:32:26,966 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.DefaultRequestInterceptor: 
Forwarding registerApplicationMasterForDistributedSchedulingrequest to the real 
YARN RM
2019-07-26 17:32:27,971 INFO 

[jira] [Comment Edited] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16894019#comment-16894019
 ] 

Babble Shack edited comment on YARN-9690 at 7/26/19 5:56 PM:
-

I also tried disabling _yarn.resourcemanager.scheduler_ and using it default 
value at the resource manager.
{code:java}



{code}
I then create a new config directory with the following
{code:java}

 
       yarn.resourcemanager.scheduler.address
       localhost:8049
   {code}
And launching the job with
{code:java}
yarn --config /debug/hadoop/etc/hadoop jar $MAPRED_EXAMPLES pi 
-Dmapreduce.job.num-opportunistic-maps-percent="40" 10 10
{code}
The nodemanager then tries to connect with scheduler on 0.0.0.0:8030, however 
the amrmproxy has already started, and so the amrmproxy should respond by 
redirecting the request to the Local RM
{code:java}
//Nodemanager logs
2019-07-26 17:32:25,480 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 container_1564162333195_0001_01_01's ip = 10.244.29.84, and hostname = 
yarn-worker-2.yarn-service.yarn.svc.cluster.local
2019-07-26 17:32:26,962 INFO SecurityLogger.org.apache.hadoop.ipc.Server: Auth 
successful for appattempt_1564162333195_0001_01 (auth:SIMPLE)
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.AMRMProxyService: 
Registering application master. 
Host:yarn-worker-2.yarn-service.yarn.svc.cluster.local Port:42172 Tracking 
Url:http://yarn-worker-2.yarn-service.yarn.svc.cluster.local:37110
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler: 
Forwarding registration request to theDistributed Scheduler Service on YARN RM
2019-07-26 17:32:26,966 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.DefaultRequestInterceptor: 
Forwarding registerApplicationMasterForDistributedSchedulingrequest to the real 
YARN RM
2019-07-26 17:32:27,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 0 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
2019-07-26 17:32:28,485 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 Skipping monitoring container container_1564162333195_0001_01_01 since CPU 
usage is not yet available.
2019-07-26 17:32:28,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 1 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

{code}
However there is no service listening on 0.0.0.0:8030

The application log, shows AM is connecting to AMRMproxy


{code:java}
// AM Logs
2019-07-26 17:50:17,238 INFO [main] org.apache.hadoop.yarn.client.RMProxy: 
Connecting to ResourceManager at localhost/127.0.0.1:8049
End of LogType:syslog.This log file belongs to a running container 
(container_1564163395606_0001_01_01) and so may not be complete.
***
{code}


was (Author: babbleshack):
I also tried disabling _yarn.resourcemanager.scheduler_ and using it default 
value at the resource manager.
{code:java}



{code}
I then create a new config directory with the following
{code:java}

 
       yarn.resourcemanager.scheduler.address
       localhost:8049
   {code}
And launching the job with
{code:java}
yarn --config /debug/hadoop/etc/hadoop jar $MAPRED_EXAMPLES pi 
-Dmapreduce.job.num-opportunistic-maps-percent="40" 10 10
{code}
The nodemanager then tries to connect with scheduler on 0.0.0.0:8030, however 
the amrmproxy has already started, and so the amrmproxy should respond by 
redirecting the request to the Local RM
{code:java}
//Nodemanager logs
2019-07-26 17:32:25,480 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 container_1564162333195_0001_01_01's ip = 10.244.29.84, and hostname = 
yarn-worker-2.yarn-service.yarn.svc.cluster.local
2019-07-26 17:32:26,962 INFO SecurityLogger.org.apache.hadoop.ipc.Server: Auth 
successful for appattempt_1564162333195_0001_01 (auth:SIMPLE)
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.AMRMProxyService: 
Registering application master. 
Host:yarn-worker-2.yarn-service.yarn.svc.cluster.local Port:42172 Tracking 
Url:http://yarn-worker-2.yarn-service.yarn.svc.cluster.local:37110
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler: 
Forwarding registration request to theDistributed Scheduler Service on YARN RM
2019-07-26 17:32:26,966 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.DefaultRequestInterceptor: 
Forwarding registerApplicationMasterForDistributedSchedulingrequest to the real 
YARN RM
2019-07-26 17:32:27,971 INFO 

[jira] [Comment Edited] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16894019#comment-16894019
 ] 

Babble Shack edited comment on YARN-9690 at 7/26/19 5:51 PM:
-

I also tried disabling _yarn.resourcemanager.scheduler_ and using it default 
value at the resource manager.
{code:java}



{code}
I then create a new config directory with the following
{code:java}

 
       yarn.resourcemanager.scheduler.address
       localhost:8049
   {code}
And launching the job with
{code:java}
yarn --config /debug/hadoop/etc/hadoop jar $MAPRED_EXAMPLES pi 
-Dmapreduce.job.num-opportunistic-maps-percent="40" 10 10
{code}
The nodemanager then tries to connect with scheduler on 0.0.0.0:8030, however 
the amrmproxy has already started, and so the amrmproxy should respond by 
redirecting the request to the Local RM
{code:java}
//Nodemanager logs
2019-07-26 17:32:25,480 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 container_1564162333195_0001_01_01's ip = 10.244.29.84, and hostname = 
yarn-worker-2.yarn-service.yarn.svc.cluster.local
2019-07-26 17:32:26,962 INFO SecurityLogger.org.apache.hadoop.ipc.Server: Auth 
successful for appattempt_1564162333195_0001_01 (auth:SIMPLE)
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.AMRMProxyService: 
Registering application master. 
Host:yarn-worker-2.yarn-service.yarn.svc.cluster.local Port:42172 Tracking 
Url:http://yarn-worker-2.yarn-service.yarn.svc.cluster.local:37110
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler: 
Forwarding registration request to theDistributed Scheduler Service on YARN RM
2019-07-26 17:32:26,966 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.DefaultRequestInterceptor: 
Forwarding registerApplicationMasterForDistributedSchedulingrequest to the real 
YARN RM
2019-07-26 17:32:27,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 0 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
2019-07-26 17:32:28,485 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 Skipping monitoring container container_1564162333195_0001_01_01 since CPU 
usage is not yet available.
2019-07-26 17:32:28,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 1 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

{code}
However there is no service listening on 0.0.0.0:8030


was (Author: babbleshack):
I also tried disabling _yarn.resourcemanager.scheduler_ and using it default 
value at the resource manager.
{code:java}



{code}
I then create a new config directory with the following
{code:java}

 
       yarn.resourcemanager.scheduler.address
       localhost:8049
   {code}
And launching the job with
{code:java}
yarn --config /debug/hadoop/etc/hadoop jar $MAPRED_EXAMPLES pi 
-Dmapreduce.job.num-opportunistic-maps-percent="40" 10 10
{code}
The nodemanager then tries to connect with scheduler on 0.0.0.0:8030, however 
the amrmproxy has already started, and so the amrmproxy should respond by 
redirecting the request to the Local RM
{code:java}
2019-07-26 17:32:25,480 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 container_1564162333195_0001_01_01's ip = 10.244.29.84, and hostname = 
yarn-worker-2.yarn-service.yarn.svc.cluster.local
2019-07-26 17:32:26,962 INFO SecurityLogger.org.apache.hadoop.ipc.Server: Auth 
successful for appattempt_1564162333195_0001_01 (auth:SIMPLE)
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.AMRMProxyService: 
Registering application master. 
Host:yarn-worker-2.yarn-service.yarn.svc.cluster.local Port:42172 Tracking 
Url:http://yarn-worker-2.yarn-service.yarn.svc.cluster.local:37110
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler: 
Forwarding registration request to theDistributed Scheduler Service on YARN RM
2019-07-26 17:32:26,966 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.DefaultRequestInterceptor: 
Forwarding registerApplicationMasterForDistributedSchedulingrequest to the real 
YARN RM
2019-07-26 17:32:27,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 0 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
2019-07-26 17:32:28,485 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 Skipping monitoring container container_1564162333195_0001_01_01 since CPU 
usage is not yet available.
2019-07-26 17:32:28,971 INFO 

[jira] [Commented] (YARN-8480) Add boolean option for resources

2019-07-26 Thread Daniel Templeton (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-8480?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16894020#comment-16894020
 ] 

Daniel Templeton commented on YARN-8480:


I still hold that boolean resources are the way to go, but given that I'm no 
longer active in YARN, and that Yunicorn aspires to be the scheduler to rule 
them all, it's not worth the fight.  Go ahead and close it.

> Add boolean option for resources
> 
>
> Key: YARN-8480
> URL: https://issues.apache.org/jira/browse/YARN-8480
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Daniel Templeton
>Assignee: Szilard Nemeth
>Priority: Major
> Attachments: YARN-8480.001.patch, YARN-8480.002.patch
>
>
> Make it possible to define a resource with a boolean value.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16894019#comment-16894019
 ] 

Babble Shack edited comment on YARN-9690 at 7/26/19 5:43 PM:
-

I also tried disabling _yarn.resourcemanager.scheduler_ and using it default 
value at the resource manager.
{code:java}



{code}
I then create a new config directory with the following
{code:java}

 
       yarn.resourcemanager.scheduler.address
       localhost:8049
   {code}
And launching the job with
{code:java}
yarn --config /debug/hadoop/etc/hadoop jar $MAPRED_EXAMPLES pi 
-Dmapreduce.job.num-opportunistic-maps-percent="40" 10 10
{code}
The nodemanager then tries to connect with scheduler on 0.0.0.0:8030, however 
the amrmproxy has already started, and so the amrmproxy should respond by 
redirecting the request to the Local RM
{code:java}
2019-07-26 17:32:25,480 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 container_1564162333195_0001_01_01's ip = 10.244.29.84, and hostname = 
yarn-worker-2.yarn-service.yarn.svc.cluster.local
2019-07-26 17:32:26,962 INFO SecurityLogger.org.apache.hadoop.ipc.Server: Auth 
successful for appattempt_1564162333195_0001_01 (auth:SIMPLE)
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.AMRMProxyService: 
Registering application master. 
Host:yarn-worker-2.yarn-service.yarn.svc.cluster.local Port:42172 Tracking 
Url:http://yarn-worker-2.yarn-service.yarn.svc.cluster.local:37110
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler: 
Forwarding registration request to theDistributed Scheduler Service on YARN RM
2019-07-26 17:32:26,966 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.DefaultRequestInterceptor: 
Forwarding registerApplicationMasterForDistributedSchedulingrequest to the real 
YARN RM
2019-07-26 17:32:27,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 0 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
2019-07-26 17:32:28,485 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 Skipping monitoring container container_1564162333195_0001_01_01 since CPU 
usage is not yet available.
2019-07-26 17:32:28,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 1 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

{code}
However there is no service listening on 0.0.0.0:8030


was (Author: babbleshack):
I also tried disabling _yarn.resourcemanager.scheduler_ and using it default 
value at the resource manager.
{code:java}


 
      yarn.resourcemanager.scheduler.address
      localhost:8049
  
-->
{code}
I then create a new config directory with the following
{code:java}

 
       yarn.resourcemanager.scheduler.address
       localhost:8049
   {code}

 And launching the job with
{code:java}
yarn --config /debug/hadoop/etc/hadoop jar $MAPRED_EXAMPLES pi 
-Dmapreduce.job.num-opportunistic-maps-percent="40" 10 10
{code}
The nodemanager then tries to connect with scheduler on 0.0.0.0:8030, however 
the amrmproxy has already started, and so the amrmproxy should respond by 
redirecting the request to the Local RM
{code:java}
2019-07-26 17:32:25,480 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 container_1564162333195_0001_01_01's ip = 10.244.29.84, and hostname = 
yarn-worker-2.yarn-service.yarn.svc.cluster.local
2019-07-26 17:32:26,962 INFO SecurityLogger.org.apache.hadoop.ipc.Server: Auth 
successful for appattempt_1564162333195_0001_01 (auth:SIMPLE)
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.AMRMProxyService: 
Registering application master. 
Host:yarn-worker-2.yarn-service.yarn.svc.cluster.local Port:42172 Tracking 
Url:http://yarn-worker-2.yarn-service.yarn.svc.cluster.local:37110
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler: 
Forwarding registration request to theDistributed Scheduler Service on YARN RM
2019-07-26 17:32:26,966 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.DefaultRequestInterceptor: 
Forwarding registerApplicationMasterForDistributedSchedulingrequest to the real 
YARN RM
2019-07-26 17:32:27,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 0 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
2019-07-26 17:32:28,485 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 Skipping monitoring container container_1564162333195_0001_01_01 since CPU 
usage is 

[jira] [Comment Edited] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893981#comment-16893981
 ] 

Babble Shack edited comment on YARN-9690 at 7/26/19 5:43 PM:
-

Setting `yarn.resourcemanager.scheduler.address` does not throw the Invalid 
AMRMToken, however the application master is unable to register with the 
resource manager.

I have configured my resource manager, which is also the machine I submit MR 
jobs from with the following
{code:java}


      yarn.resourcemanager.scheduler.address
      localhost:8049
  
{code}
Whilst my nodemanager is configured as:
{code:java}
 

 
{code}
Doing so enable the AM to connect to RM at 127.0.0.1,
{code:java}
// Application Log
/
2019-07-26 16:47:46,928 INFO [main] org.apache.hadoop.security.SecurityUtil: 
Updating Configuration
2019-07-26 16:47:46,993 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Executing with tokens: [Kind: 
YARN_AM_RM_TOKEN, Service: , Ident: (appAttemptId { application_id { id: 1 
cluster_timestamp: 1564159650361 } attemptId: 1 } keyId: -803833355)]
2019-07-26 16:47:47,021 INFO [main] org.apache.hadoop.conf.Configuration: 
resource-types.xml not found
2019-07-26 16:47:47,021 INFO [main] 
org.apache.hadoop.yarn.util.resource.ResourceUtils: Unable to find 
'resource-types.xml'.
2019-07-26 16:47:47,028 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Using mapred newApiCommitter.
2019-07-26 16:47:47,029 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter set in config 
null
2019-07-26 16:47:47,048 INFO [main] 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: File Output 
Committer Algorithm version is 2
2019-07-26 16:47:47,048 INFO [main] 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: FileOutputCommitter 
skip cleanup _temporary folders under output directory:false, ignore cleanup 
failures: false
2019-07-26 16:47:47,389 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter is 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
2019-07-26 16:47:47,490 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.jobhistory.EventType for class 
org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler
2019-07-26 16:47:47,491 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.JobEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobEventDispatcher
2019-07-26 16:47:47,491 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.TaskEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.TaskAttemptEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskAttemptEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventType for class 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.speculate.Speculator$EventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$SpeculatorEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.rm.ContainerAllocator$EventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter
2019-07-26 16:47:47,493 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncher$EventType for 
class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerLauncherRouter
2019-07-26 16:47:47,507 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,514 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,521 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,526 INFO [main] 
org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Emitting job 
history data to the timeline server is not enabled
2019-07-26 16:47:47,550 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.JobFinishEvent$Type for class 

[jira] [Comment Edited] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16894019#comment-16894019
 ] 

Babble Shack edited comment on YARN-9690 at 7/26/19 5:42 PM:
-

I also tried disabling _yarn.resourcemanager.scheduler_ and using it default 
value at the resource manager.
{code:java}


 
      yarn.resourcemanager.scheduler.address
      localhost:8049
  
-->
{code}
I then create a new config directory with the following
{code:java}

 
       yarn.resourcemanager.scheduler.address
       localhost:8049
   {code}

 And launching the job with
{code:java}
yarn --config /debug/hadoop/etc/hadoop jar $MAPRED_EXAMPLES pi 
-Dmapreduce.job.num-opportunistic-maps-percent="40" 10 10
{code}
The nodemanager then tries to connect with scheduler on 0.0.0.0:8030, however 
the amrmproxy has already started, and so the amrmproxy should respond by 
redirecting the request to the Local RM
{code:java}
2019-07-26 17:32:25,480 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 container_1564162333195_0001_01_01's ip = 10.244.29.84, and hostname = 
yarn-worker-2.yarn-service.yarn.svc.cluster.local
2019-07-26 17:32:26,962 INFO SecurityLogger.org.apache.hadoop.ipc.Server: Auth 
successful for appattempt_1564162333195_0001_01 (auth:SIMPLE)
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.AMRMProxyService: 
Registering application master. 
Host:yarn-worker-2.yarn-service.yarn.svc.cluster.local Port:42172 Tracking 
Url:http://yarn-worker-2.yarn-service.yarn.svc.cluster.local:37110
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler: 
Forwarding registration request to theDistributed Scheduler Service on YARN RM
2019-07-26 17:32:26,966 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.DefaultRequestInterceptor: 
Forwarding registerApplicationMasterForDistributedSchedulingrequest to the real 
YARN RM
2019-07-26 17:32:27,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 0 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
2019-07-26 17:32:28,485 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 Skipping monitoring container container_1564162333195_0001_01_01 since CPU 
usage is not yet available.
2019-07-26 17:32:28,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 1 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

{code}
However there is no service listening on 0.0.0.0:8030


was (Author: babbleshack):
I also tried disabling _yarn.resourcemanager.scheduler_ and using it default 
value at the resource manager.
{code:java}


 
      yarn.resourcemanager.scheduler.address
      localhost:8049
  
-->
{code}
I then create a new config directory with the following

 
      yarn.resourcemanager.scheduler.address
      localhost:8049
  
And launching the job with
{code:java}
yarn --config /debug/hadoop/etc/hadoop jar $MAPRED_EXAMPLES pi 
-Dmapreduce.job.num-opportunistic-maps-percent="40" 10 10
{code}

The nodemanager then tries to connect with scheduler on 0.0.0.0:8030, however 
the amrmproxy has already started, and so the amrmproxy should respond by 
redirecting the request to the Local RM
{code:java}
2019-07-26 17:32:25,480 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 container_1564162333195_0001_01_01's ip = 10.244.29.84, and hostname = 
yarn-worker-2.yarn-service.yarn.svc.cluster.local
2019-07-26 17:32:26,962 INFO SecurityLogger.org.apache.hadoop.ipc.Server: Auth 
successful for appattempt_1564162333195_0001_01 (auth:SIMPLE)
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.AMRMProxyService: 
Registering application master. 
Host:yarn-worker-2.yarn-service.yarn.svc.cluster.local Port:42172 Tracking 
Url:http://yarn-worker-2.yarn-service.yarn.svc.cluster.local:37110
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler: 
Forwarding registration request to theDistributed Scheduler Service on YARN RM
2019-07-26 17:32:26,966 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.DefaultRequestInterceptor: 
Forwarding registerApplicationMasterForDistributedSchedulingrequest to the real 
YARN RM
2019-07-26 17:32:27,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 0 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
2019-07-26 17:32:28,485 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 Skipping monitoring container 

[jira] [Comment Edited] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893981#comment-16893981
 ] 

Babble Shack edited comment on YARN-9690 at 7/26/19 5:42 PM:
-

Setting `yarn.resourcemanager.scheduler.address` does not throw the Invalid 
AMRMToken, however the application master is unable to register with the 
resource manager.

I have configured my resource manager, which is also the machine I submit MR 
jobs from with the following
{code:java}

 

 
{code}
Doing so enable the AM to connect to RM at 127.0.0.1,
{code:java}
// Application Log
/
2019-07-26 16:47:46,928 INFO [main] org.apache.hadoop.security.SecurityUtil: 
Updating Configuration
2019-07-26 16:47:46,993 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Executing with tokens: [Kind: 
YARN_AM_RM_TOKEN, Service: , Ident: (appAttemptId { application_id { id: 1 
cluster_timestamp: 1564159650361 } attemptId: 1 } keyId: -803833355)]
2019-07-26 16:47:47,021 INFO [main] org.apache.hadoop.conf.Configuration: 
resource-types.xml not found
2019-07-26 16:47:47,021 INFO [main] 
org.apache.hadoop.yarn.util.resource.ResourceUtils: Unable to find 
'resource-types.xml'.
2019-07-26 16:47:47,028 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Using mapred newApiCommitter.
2019-07-26 16:47:47,029 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter set in config 
null
2019-07-26 16:47:47,048 INFO [main] 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: File Output 
Committer Algorithm version is 2
2019-07-26 16:47:47,048 INFO [main] 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: FileOutputCommitter 
skip cleanup _temporary folders under output directory:false, ignore cleanup 
failures: false
2019-07-26 16:47:47,389 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter is 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
2019-07-26 16:47:47,490 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.jobhistory.EventType for class 
org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler
2019-07-26 16:47:47,491 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.JobEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobEventDispatcher
2019-07-26 16:47:47,491 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.TaskEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.TaskAttemptEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskAttemptEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventType for class 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.speculate.Speculator$EventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$SpeculatorEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.rm.ContainerAllocator$EventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter
2019-07-26 16:47:47,493 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncher$EventType for 
class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerLauncherRouter
2019-07-26 16:47:47,507 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,514 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,521 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,526 INFO [main] 
org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Emitting job 
history data to the timeline server is not enabled
2019-07-26 16:47:47,550 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.JobFinishEvent$Type for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobFinishEventHandler
2019-07-26 16:47:47,677 INFO [main] 

[jira] [Commented] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16894019#comment-16894019
 ] 

Babble Shack commented on YARN-9690:


I also tried disabling _yarn.resourcemanager.scheduler_ and using it default 
value at the resource manager.
{code:java}


 
      yarn.resourcemanager.scheduler.address
      localhost:8049
  
-->
{code}
I then create a new config directory with the following

 
      yarn.resourcemanager.scheduler.address
      localhost:8049
  
And launching the job with
{code:java}
yarn --config /debug/hadoop/etc/hadoop jar $MAPRED_EXAMPLES pi 
-Dmapreduce.job.num-opportunistic-maps-percent="40" 10 10
{code}

The nodemanager then tries to connect with scheduler on 0.0.0.0:8030, however 
the amrmproxy has already started, and so the amrmproxy should respond by 
redirecting the request to the Local RM
{code:java}
2019-07-26 17:32:25,480 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 container_1564162333195_0001_01_01's ip = 10.244.29.84, and hostname = 
yarn-worker-2.yarn-service.yarn.svc.cluster.local
2019-07-26 17:32:26,962 INFO SecurityLogger.org.apache.hadoop.ipc.Server: Auth 
successful for appattempt_1564162333195_0001_01 (auth:SIMPLE)
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.AMRMProxyService: 
Registering application master. 
Host:yarn-worker-2.yarn-service.yarn.svc.cluster.local Port:42172 Tracking 
Url:http://yarn-worker-2.yarn-service.yarn.svc.cluster.local:37110
2019-07-26 17:32:26,965 INFO 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler: 
Forwarding registration request to theDistributed Scheduler Service on YARN RM
2019-07-26 17:32:26,966 INFO 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.DefaultRequestInterceptor: 
Forwarding registerApplicationMasterForDistributedSchedulingrequest to the real 
YARN RM
2019-07-26 17:32:27,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 0 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)
2019-07-26 17:32:28,485 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.monitor.ContainersMonitorImpl:
 Skipping monitoring container container_1564162333195_0001_01_01 since CPU 
usage is not yet available.
2019-07-26 17:32:28,971 INFO org.apache.hadoop.ipc.Client: Retrying connect to 
server: 0.0.0.0/0.0.0.0:8030. Already tried 1 time(s); retry policy is 
RetryUpToMaximumCountWithFixedSleep(maxRetries=10, sleepTime=1000 MILLISECONDS)

{code}
However there is no service listening on 0.0.0.0:8030

> Invalid AMRM token when distributed scheduling is enabled.
> --
>
> Key: YARN-9690
> URL: https://issues.apache.org/jira/browse/YARN-9690
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: distributed-scheduling, yarn
>Affects Versions: 2.9.2, 3.1.2
> Environment: OS: Ubuntu 18.04
> JVM: 1.8.0_212-8u212-b03-0ubuntu1.18.04.1-b03
>Reporter: Babble Shack
>Priority: Major
> Attachments: applicationlog, distributed_log, ds_application.log, 
> image-2019-07-26-18-00-14-980.png, nodemanager-yarn-site.xml, 
> nodemanager.log, rm-yarn-site.xml, yarn-site.xml
>
>
> Applications fail to start due to invalild AMRM from application attempt.
> I have tested this with 0/100% opportunistic maps and the same issue occurs 
> regardless.
> {code:java}
> 
> -->
> 
>   
>     mapreduceyarn.nodemanager.aux-services
>     mapreduce_shuffle
>   
>   
>       yarn.resourcemanager.address
>       yarn-master-0.yarn-service.yarn:8032
>   
>   
>       yarn.resourcemanager.scheduler.address
>       0.0.0.0:8049
>   
>   
>     
> yarn.resourcemanager.opportunistic-container-allocation.enabled
>     true
>   
>   
>     yarn.nodemanager.opportunistic-containers-max-queue-length
>     10
>   
>   
>     yarn.nodemanager.distributed-scheduling.enabled
>     true
>   
>  
>   
>     yarn.webapp.ui2.enable
>     true
>   
>   
>       yarn.resourcemanager.resource-tracker.address
>       yarn-master-0.yarn-service.yarn:8031
>   
>   
>     yarn.log-aggregation-enable
>     true
>   
>   
>       yarn.nodemanager.aux-services
>       mapreduce_shuffle
>   
>   
>   
>   
>   
>     yarn.nodemanager.resource.memory-mb
>     7168
>   
>   
>     yarn.scheduler.minimum-allocation-mb
>     3584
>   
>   
>     yarn.scheduler.maximum-allocation-mb
>     7168
>   
>   
>     yarn.app.mapreduce.am.resource.mb
>     7168
>   
>   
>   
>     yarn.app.mapreduce.am.command-opts
>     -Xmx5734m
>   
>   
>   
>     yarn.timeline-service.enabled
>     true
>   
>   
>     yarn.resourcemanager.system-metrics-publisher.enabled
>     true
>   
>   
>     

[jira] [Comment Edited] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893981#comment-16893981
 ] 

Babble Shack edited comment on YARN-9690 at 7/26/19 5:13 PM:
-

Setting `yarn.resourcemanager.scheduler.address` does not throw the Invalid 
AMRMToken, however the application master is unable to register with the 
resource manager.

I have configured my resource manager, which is also the machine I submit MR 
jobs from with the following
{code:java}

 
      yarn.resourcemanager.scheduler.address
      localhost:8049
  
{code}
Whilst my nodemanager is configured as:
{code:java}
 

 
{code}
Doing so enable the AM to connect to RM at 127.0.0.1,
{code:java}
// Application Log
/
2019-07-26 16:47:46,928 INFO [main] org.apache.hadoop.security.SecurityUtil: 
Updating Configuration
2019-07-26 16:47:46,993 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Executing with tokens: [Kind: 
YARN_AM_RM_TOKEN, Service: , Ident: (appAttemptId { application_id { id: 1 
cluster_timestamp: 1564159650361 } attemptId: 1 } keyId: -803833355)]
2019-07-26 16:47:47,021 INFO [main] org.apache.hadoop.conf.Configuration: 
resource-types.xml not found
2019-07-26 16:47:47,021 INFO [main] 
org.apache.hadoop.yarn.util.resource.ResourceUtils: Unable to find 
'resource-types.xml'.
2019-07-26 16:47:47,028 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Using mapred newApiCommitter.
2019-07-26 16:47:47,029 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter set in config 
null
2019-07-26 16:47:47,048 INFO [main] 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: File Output 
Committer Algorithm version is 2
2019-07-26 16:47:47,048 INFO [main] 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: FileOutputCommitter 
skip cleanup _temporary folders under output directory:false, ignore cleanup 
failures: false
2019-07-26 16:47:47,389 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter is 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
2019-07-26 16:47:47,490 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.jobhistory.EventType for class 
org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler
2019-07-26 16:47:47,491 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.JobEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobEventDispatcher
2019-07-26 16:47:47,491 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.TaskEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.TaskAttemptEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskAttemptEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventType for class 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.speculate.Speculator$EventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$SpeculatorEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.rm.ContainerAllocator$EventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter
2019-07-26 16:47:47,493 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncher$EventType for 
class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerLauncherRouter
2019-07-26 16:47:47,507 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,514 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,521 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,526 INFO [main] 
org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Emitting job 
history data to the timeline server is not enabled
2019-07-26 16:47:47,550 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.JobFinishEvent$Type for class 

[jira] [Commented] (YARN-9479) Change String.equals to Objects.equals(String,String) to avoid possible NullPointerException

2019-07-26 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9479?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893994#comment-16893994
 ] 

Hadoop QA commented on YARN-9479:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  1m 
12s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} dupname {color} | {color:green}  0m  
0s{color} | {color:green} No case conflicting files found. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 20m 
30s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
59s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
36s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
55s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
13m 46s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
36s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue}  1m 
55s{color} | {color:blue} Used deprecated FindBugs config; considering 
switching to SpotBugs. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
52s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
54s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
51s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
51s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
0m 34s{color} | {color:orange} 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager:
 The patch generated 2 new + 15 unchanged - 0 fixed = 17 total (was 15) {color} 
|
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
54s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
12m 48s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
29s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m  
1s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} unit {color} | {color:red} 88m 18s{color} 
| {color:red} hadoop-yarn-server-resourcemanager in the patch failed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
29s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}148m 36s{color} | 
{color:black} {color} |
\\
\\
|| Reason || Tests ||
| Failed junit tests | 
hadoop.yarn.server.resourcemanager.security.TestDelegationTokenRenewer |
|   | 
hadoop.yarn.server.resourcemanager.reservation.TestCapacitySchedulerPlanFollower
 |
|   | 
hadoop.yarn.server.resourcemanager.scheduler.fair.TestFairSchedulerPreemption |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=18.09.7 Server=18.09.7 base: 
https://builds.apache.org/job/hadoop-multibranch/job/PR-738/4/artifact/out/Dockerfile
 |
| GITHUB PR | https://github.com/apache/hadoop/pull/738 |
| JIRA Issue | YARN-9479 |
| Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite 
unit shadedclient findbugs checkstyle |
| uname | Linux b1db80e57343 

[jira] [Comment Edited] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893981#comment-16893981
 ] 

Babble Shack edited comment on YARN-9690 at 7/26/19 5:12 PM:
-

Setting `yarn.resourcemanager.scheduler.address` does not throw the Invalid 
AMRMToken, however the application master is unable to register with the 
resource manager.

I have configured my resource manager, which is also the machine I submit MR 
jobs from with the following
{code:java}

 
      yarn.resourcemanager.scheduler.address
      localhost:8049
  
{code}
Whilst my nodemanager is configured as:
{code:java}
 
 
{code}
Doing so enable the AM to connect to RM at 127.0.0.1,
{code:java}
// Application Log
/
2019-07-26 16:47:46,928 INFO [main] org.apache.hadoop.security.SecurityUtil: 
Updating Configuration
2019-07-26 16:47:46,993 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Executing with tokens: [Kind: 
YARN_AM_RM_TOKEN, Service: , Ident: (appAttemptId { application_id { id: 1 
cluster_timestamp: 1564159650361 } attemptId: 1 } keyId: -803833355)]
2019-07-26 16:47:47,021 INFO [main] org.apache.hadoop.conf.Configuration: 
resource-types.xml not found
2019-07-26 16:47:47,021 INFO [main] 
org.apache.hadoop.yarn.util.resource.ResourceUtils: Unable to find 
'resource-types.xml'.
2019-07-26 16:47:47,028 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Using mapred newApiCommitter.
2019-07-26 16:47:47,029 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter set in config 
null
2019-07-26 16:47:47,048 INFO [main] 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: File Output 
Committer Algorithm version is 2
2019-07-26 16:47:47,048 INFO [main] 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: FileOutputCommitter 
skip cleanup _temporary folders under output directory:false, ignore cleanup 
failures: false
2019-07-26 16:47:47,389 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter is 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
2019-07-26 16:47:47,490 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.jobhistory.EventType for class 
org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler
2019-07-26 16:47:47,491 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.JobEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobEventDispatcher
2019-07-26 16:47:47,491 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.TaskEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.TaskAttemptEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskAttemptEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventType for class 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.speculate.Speculator$EventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$SpeculatorEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.rm.ContainerAllocator$EventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter
2019-07-26 16:47:47,493 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncher$EventType for 
class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerLauncherRouter
2019-07-26 16:47:47,507 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,514 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,521 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,526 INFO [main] 
org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Emitting job 
history data to the timeline server is not enabled
2019-07-26 16:47:47,550 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.JobFinishEvent$Type for class 

[jira] [Comment Edited] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893981#comment-16893981
 ] 

Babble Shack edited comment on YARN-9690 at 7/26/19 5:10 PM:
-

Setting `yarn.resourcemanager.scheduler.address` does not throw the Invalid 
AMRMToken, however the application master is unable to register with the 
resource manager.

I have configured my resource manager, which is also the machine I submit MR 
jobs from with the following
{code:java}

 
      yarn.resourcemanager.scheduler.address
      localhost:8049
  
{code}
Whilst my nodemanager is configured as:
{code:java}
 
 
{code}

Doing so enable the AM to connect to RM at 127.0.0.1,
{code:java}
// Application Log
/
2019-07-26 16:47:46,928 INFO [main] org.apache.hadoop.security.SecurityUtil: 
Updating Configuration
2019-07-26 16:47:46,993 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Executing with tokens: [Kind: 
YARN_AM_RM_TOKEN, Service: , Ident: (appAttemptId { application_id { id: 1 
cluster_timestamp: 1564159650361 } attemptId: 1 } keyId: -803833355)]
2019-07-26 16:47:47,021 INFO [main] org.apache.hadoop.conf.Configuration: 
resource-types.xml not found
2019-07-26 16:47:47,021 INFO [main] 
org.apache.hadoop.yarn.util.resource.ResourceUtils: Unable to find 
'resource-types.xml'.
2019-07-26 16:47:47,028 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: Using mapred newApiCommitter.
2019-07-26 16:47:47,029 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter set in config 
null
2019-07-26 16:47:47,048 INFO [main] 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: File Output 
Committer Algorithm version is 2
2019-07-26 16:47:47,048 INFO [main] 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter: FileOutputCommitter 
skip cleanup _temporary folders under output directory:false, ignore cleanup 
failures: false
2019-07-26 16:47:47,389 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster: OutputCommitter is 
org.apache.hadoop.mapreduce.lib.output.FileOutputCommitter
2019-07-26 16:47:47,490 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.jobhistory.EventType for class 
org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler
2019-07-26 16:47:47,491 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.JobEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$JobEventDispatcher
2019-07-26 16:47:47,491 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.TaskEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.TaskAttemptEventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$TaskAttemptEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventType for class 
org.apache.hadoop.mapreduce.v2.app.commit.CommitterEventHandler
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.speculate.Speculator$EventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$SpeculatorEventDispatcher
2019-07-26 16:47:47,492 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.rm.ContainerAllocator$EventType for class 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter
2019-07-26 16:47:47,493 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.launcher.ContainerLauncher$EventType for 
class org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerLauncherRouter
2019-07-26 16:47:47,507 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,514 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,521 INFO [main] 
org.apache.hadoop.mapreduce.v2.jobhistory.JobHistoryUtils: Default file system 
[hdfs://hdfs-master-0.hdfs-service.hdfs:9000]
2019-07-26 16:47:47,526 INFO [main] 
org.apache.hadoop.mapreduce.jobhistory.JobHistoryEventHandler: Emitting job 
history data to the timeline server is not enabled
2019-07-26 16:47:47,550 INFO [main] 
org.apache.hadoop.yarn.event.AsyncDispatcher: Registering class 
org.apache.hadoop.mapreduce.v2.app.job.event.JobFinishEvent$Type for class 

[jira] [Commented] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893981#comment-16893981
 ] 

Babble Shack commented on YARN-9690:


Setting `yarn.resourcemanager.scheduler.address` does not throw the Invalid 
AMRMToken, however the application master is unable to register with the 
resource manager.

I have configured my resource manager, which is also the machine I submit MR 
jobs from with the following
{code:java}

 
      yarn.resourcemanager.scheduler.address
      localhost:8049
  
{code}
{code:java}
 
 
{code}
 

> Invalid AMRM token when distributed scheduling is enabled.
> --
>
> Key: YARN-9690
> URL: https://issues.apache.org/jira/browse/YARN-9690
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: distributed-scheduling, yarn
>Affects Versions: 2.9.2, 3.1.2
> Environment: OS: Ubuntu 18.04
> JVM: 1.8.0_212-8u212-b03-0ubuntu1.18.04.1-b03
>Reporter: Babble Shack
>Priority: Major
> Attachments: applicationlog, distributed_log, yarn-site.xml
>
>
> Applications fail to start due to invalild AMRM from application attempt.
> I have tested this with 0/100% opportunistic maps and the same issue occurs 
> regardless.
> {code:java}
> 
> -->
> 
>   
>     mapreduceyarn.nodemanager.aux-services
>     mapreduce_shuffle
>   
>   
>       yarn.resourcemanager.address
>       yarn-master-0.yarn-service.yarn:8032
>   
>   
>       yarn.resourcemanager.scheduler.address
>       0.0.0.0:8049
>   
>   
>     
> yarn.resourcemanager.opportunistic-container-allocation.enabled
>     true
>   
>   
>     yarn.nodemanager.opportunistic-containers-max-queue-length
>     10
>   
>   
>     yarn.nodemanager.distributed-scheduling.enabled
>     true
>   
>  
>   
>     yarn.webapp.ui2.enable
>     true
>   
>   
>       yarn.resourcemanager.resource-tracker.address
>       yarn-master-0.yarn-service.yarn:8031
>   
>   
>     yarn.log-aggregation-enable
>     true
>   
>   
>       yarn.nodemanager.aux-services
>       mapreduce_shuffle
>   
>   
>   
>   
>   
>     yarn.nodemanager.resource.memory-mb
>     7168
>   
>   
>     yarn.scheduler.minimum-allocation-mb
>     3584
>   
>   
>     yarn.scheduler.maximum-allocation-mb
>     7168
>   
>   
>     yarn.app.mapreduce.am.resource.mb
>     7168
>   
>   
>   
>     yarn.app.mapreduce.am.command-opts
>     -Xmx5734m
>   
>   
>   
>     yarn.timeline-service.enabled
>     true
>   
>   
>     yarn.resourcemanager.system-metrics-publisher.enabled
>     true
>   
>   
>     yarn.timeline-service.generic-application-history.enabled
>     true
>   
>   
>     yarn.timeline-service.bind-host
>     0.0.0.0
>   
> 
> {code}
> Relevant logs:
> {code:java}
> 2019-07-22 14:56:37,104 INFO [main] 
> org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: 100% of the 
> mappers will be scheduled using OPPORTUNISTIC containers
> 2019-07-22 14:56:37,117 INFO [main] org.apache.hadoop.yarn.client.RMProxy: 
> Connecting to ResourceManager at 
> yarn-master-0.yarn-service.yarn/10.244.1.134:8030
> 2019-07-22 14:56:37,150 WARN [main] org.apache.hadoop.ipc.Client: Exception 
> encountered while connecting to the server : 
> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.token.SecretManager$InvalidToken):
>  Invalid AMRMToken from appattempt_1563805140414_0002_02
> 2019-07-22 14:56:37,152 ERROR [main] 
> org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator: Exception while 
> registering
> org.apache.hadoop.security.token.SecretManager$InvalidToken: Invalid 
> AMRMToken from appattempt_1563805140414_0002_02
>     at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
>     at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
>     at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
>     at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
>     at 
> org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53)
>     at 
> org.apache.hadoop.yarn.ipc.RPCUtil.instantiateIOException(RPCUtil.java:80)
>     at 
> org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:119)
>     at 
> org.apache.hadoop.yarn.api.impl.pb.client.ApplicationMasterProtocolPBClientImpl.registerApplicationMaster(ApplicationMasterProtocolPBClientImpl.java:109)
>     at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>     at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>     at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>     at java.lang.reflect.Method.invoke(Method.java:498)
>     at 
> org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422)
> 

[jira] [Commented] (YARN-9509) Capped cpu usage with cgroup strict-resource-usage based on a mulitplier

2019-07-26 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9509?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893970#comment-16893970
 ] 

Hadoop QA commented on YARN-9509:
-

| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  2m 
22s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} dupname {color} | {color:green}  0m  
0s{color} | {color:green} No case conflicting files found. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  1m  
5s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 21m 
29s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 10m 
12s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  1m 
21s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m 
32s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
16m 49s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  2m  
3s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue}  1m 
40s{color} | {color:blue} Used deprecated FindBugs config; considering 
switching to SpotBugs. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  5m 
48s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
16s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
57s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  7m 
55s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  7m 
55s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
1m 18s{color} | {color:orange} hadoop-yarn-project/hadoop-yarn: The patch 
generated 5 new + 219 unchanged - 0 fixed = 224 total (was 219) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  2m 
31s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
1s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
13m 28s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  2m  
1s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  5m 
39s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  0m 
58s{color} | {color:green} hadoop-yarn-api in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  3m 
57s{color} | {color:green} hadoop-yarn-common in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 21m 
31s{color} | {color:green} hadoop-yarn-server-nodemanager in the patch passed. 
{color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
45s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}126m  4s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=18.09.7 Server=18.09.7 base: 

[jira] [Commented] (YARN-9009) Fix flaky test TestEntityGroupFSTimelineStore.testCleanLogs

2019-07-26 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9009?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893933#comment-16893933
 ] 

Hadoop QA commented on YARN-9009:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m  
0s{color} | {color:blue} Docker mode activated. {color} |
| {color:red}-1{color} | {color:red} patch {color} | {color:red}  0m 14s{color} 
| {color:red} https://github.com/apache/hadoop/pull/438 does not apply to 
trunk. Rebase required? Wrong Branch? See 
https://wiki.apache.org/hadoop/HowToContribute for help. {color} |
\\
\\
|| Subsystem || Report/Notes ||
| GITHUB PR | https://github.com/apache/hadoop/pull/438 |
| JIRA Issue | YARN-9009 |
| Console output | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-438/2/console |
| versions | git=2.7.4 |
| Powered by | Apache Yetus 0.10.0 http://yetus.apache.org |


This message was automatically generated.



> Fix flaky test TestEntityGroupFSTimelineStore.testCleanLogs
> ---
>
> Key: YARN-9009
> URL: https://issues.apache.org/jira/browse/YARN-9009
> Project: Hadoop YARN
>  Issue Type: Bug
> Environment: Ubuntu 18.04
> java version "1.8.0_181"
> Java(TM) SE Runtime Environment (build 1.8.0_181-b13)
> Java HotSpot(TM) 64-Bit Server VM (build 25.181-b13, mixed mode)
>  
> Apache Maven 3.5.4 (1edded0938998edf8bf061f1ceb3cfdeccf443fe; 
> 2018-06-17T13:33:14-05:00)
>Reporter: OrDTesters
>Assignee: OrDTesters
>Priority: Minor
> Fix For: 3.0.4, 3.1.2, 3.3.0, 3.2.1
>
> Attachments: YARN-9009-trunk-001.patch
>
>
> In TestEntityGroupFSTimelineStore, testCleanLogs fails when run after 
> testMoveToDone.
> testCleanLogs fails because testMoveToDone moves a file into the same 
> directory that testCleanLogs cleans, causing testCleanLogs to clean 3 files, 
> instead of 2 as testCleanLogs expects.
> To fix the failure of testCleanLogs, we can delete the file after the file is 
> moved by testMoveToDone.
> Pull request link: [https://github.com/apache/hadoop/pull/438]



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-9468) Fix inaccurate documentations in Placement Constraints

2019-07-26 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893920#comment-16893920
 ] 

Hadoop QA commented on YARN-9468:
-

| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  2m 
15s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} dupname {color} | {color:green}  0m  
0s{color} | {color:green} No case conflicting files found. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 22m 
23s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
22s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
35m 18s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
14s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
20s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
13m 19s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
26s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 53m 46s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=19.03.1 Server=19.03.1 base: 
https://builds.apache.org/job/hadoop-multibranch/job/PR-717/2/artifact/out/Dockerfile
 |
| GITHUB PR | https://github.com/apache/hadoop/pull/717 |
| JIRA Issue | YARN-9468 |
| Optional Tests | dupname asflicense mvnsite |
| uname | Linux 2bb965e54e85 4.15.0-48-generic #51-Ubuntu SMP Wed Apr 3 
08:28:49 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | personality/hadoop.sh |
| git revision | trunk / c0a0c35 |
| Max. process+thread count | 341 (vs. ulimit of 5500) |
| modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site U: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-site |
| Console output | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-717/2/console |
| versions | git=2.7.4 maven=3.3.9 |
| Powered by | Apache Yetus 0.10.0 http://yetus.apache.org |


This message was automatically generated.



> Fix inaccurate documentations in Placement Constraints
> --
>
> Key: YARN-9468
> URL: https://issues.apache.org/jira/browse/YARN-9468
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: yarn
>Affects Versions: 3.2.0
>Reporter: hunshenshi
>Assignee: hunshenshi
>Priority: Major
>
> Document Placement Constraints
> *First* 
> {code:java}
> zk=3,NOTIN,NODE,zk:hbase=5,IN,RACK,zk:spark=7,CARDINALITY,NODE,hbase,1,3{code}
>  * place 5 containers with tag “hbase” with affinity to a rack on which 
> containers with tag “zk” are running (i.e., an “hbase” container 
> should{color:#ff} not{color} be placed at a rack where an “zk” container 
> is running, given that “zk” is the TargetTag of the second constraint);
> The _*not*_ word in brackets should be delete.
>  
> *Second*
> {code:java}
> PlacementSpec => "" | KeyVal;PlacementSpec
> {code}
> The semicolon should be replaced by colon
>  



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Created] (YARN-9710) [UI2] Yarn Daemon Logs displays the URL instead of log name

2019-07-26 Thread Prabhu Joseph (JIRA)
Prabhu Joseph created YARN-9710:
---

 Summary: [UI2] Yarn Daemon Logs displays the URL instead of log 
name
 Key: YARN-9710
 URL: https://issues.apache.org/jira/browse/YARN-9710
 Project: Hadoop YARN
  Issue Type: Bug
  Components: yarn-ui-v2
Affects Versions: 3.2.0
Reporter: Prabhu Joseph
 Attachments: Screen Shot 2019-07-26 at 8.53.50 PM.png

[UI2] Yarn Daemon Logs displays the URL instead of log name

 !Screen Shot 2019-07-26 at 8.53.50 PM.png|height=150|width=300! 



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-9710) [UI2] Yarn Daemon Logs displays the URL instead of log name

2019-07-26 Thread Prabhu Joseph (JIRA)


 [ 
https://issues.apache.org/jira/browse/YARN-9710?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prabhu Joseph updated YARN-9710:

Description: 
Yarn Daemon Logs displays the URL instead of log name.

!Screen Shot 2019-07-26 at 8.53.50 PM.png|height=300!

  was:
[UI2] Yarn Daemon Logs displays the URL instead of log name

 !Screen Shot 2019-07-26 at 8.53.50 PM.png|height=300|width=500! 


> [UI2] Yarn Daemon Logs displays the URL instead of log name
> ---
>
> Key: YARN-9710
> URL: https://issues.apache.org/jira/browse/YARN-9710
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: yarn-ui-v2
>Affects Versions: 3.2.0
>Reporter: Prabhu Joseph
>Priority: Minor
> Attachments: Screen Shot 2019-07-26 at 8.53.50 PM.png
>
>
> Yarn Daemon Logs displays the URL instead of log name.
> !Screen Shot 2019-07-26 at 8.53.50 PM.png|height=300!



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Assigned] (YARN-9710) [UI2] Yarn Daemon Logs displays the URL instead of log name

2019-07-26 Thread Prabhu Joseph (JIRA)


 [ 
https://issues.apache.org/jira/browse/YARN-9710?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prabhu Joseph reassigned YARN-9710:
---

Assignee: Prabhu Joseph

> [UI2] Yarn Daemon Logs displays the URL instead of log name
> ---
>
> Key: YARN-9710
> URL: https://issues.apache.org/jira/browse/YARN-9710
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: yarn-ui-v2
>Affects Versions: 3.2.0
>Reporter: Prabhu Joseph
>Assignee: Prabhu Joseph
>Priority: Minor
> Attachments: Screen Shot 2019-07-26 at 8.53.50 PM.png
>
>
> Yarn Daemon Logs displays the URL instead of log name.
> !Screen Shot 2019-07-26 at 8.53.50 PM.png|height=300!



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-9710) [UI2] Yarn Daemon Logs displays the URL instead of log name

2019-07-26 Thread Prabhu Joseph (JIRA)


 [ 
https://issues.apache.org/jira/browse/YARN-9710?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Prabhu Joseph updated YARN-9710:

Description: 
[UI2] Yarn Daemon Logs displays the URL instead of log name

 !Screen Shot 2019-07-26 at 8.53.50 PM.png|height=300|width=500! 

  was:
[UI2] Yarn Daemon Logs displays the URL instead of log name

 !Screen Shot 2019-07-26 at 8.53.50 PM.png|height=150|width=300! 


> [UI2] Yarn Daemon Logs displays the URL instead of log name
> ---
>
> Key: YARN-9710
> URL: https://issues.apache.org/jira/browse/YARN-9710
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: yarn-ui-v2
>Affects Versions: 3.2.0
>Reporter: Prabhu Joseph
>Priority: Minor
> Attachments: Screen Shot 2019-07-26 at 8.53.50 PM.png
>
>
> [UI2] Yarn Daemon Logs displays the URL instead of log name
>  !Screen Shot 2019-07-26 at 8.53.50 PM.png|height=300|width=500! 



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-9579) the property of sharedcache in mapred-default.xml

2019-07-26 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9579?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893873#comment-16893873
 ] 

Hadoop QA commented on YARN-9579:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
41s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} dupname {color} | {color:green}  0m  
0s{color} | {color:green} No case conflicting files found. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 17m 
18s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
32s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
37s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
28m 38s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
17s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
30s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
30s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
35s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
3s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
11m 52s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
15s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  5m 
18s{color} | {color:green} hadoop-mapreduce-client-core in the patch passed. 
{color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
28s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 50m 58s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=19.03.1 Server=19.03.1 base: 
https://builds.apache.org/job/hadoop-multibranch/job/PR-848/3/artifact/out/Dockerfile
 |
| GITHUB PR | https://github.com/apache/hadoop/pull/848 |
| JIRA Issue | YARN-9579 |
| Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite 
unit shadedclient xml |
| uname | Linux 4d46d27e8af9 4.4.0-139-generic #165-Ubuntu SMP Wed Oct 24 
10:58:50 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | personality/hadoop.sh |
| git revision | trunk / c0a0c35 |
| Default Java | 1.8.0_212 |
|  Test Results | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-848/3/testReport/ |
| Max. process+thread count | 1620 (vs. ulimit of 5500) |
| modules | C: 
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core 
U: 
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core |
| Console output | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-848/3/console |
| versions | git=2.7.4 maven=3.3.9 |
| Powered by | Apache Yetus 0.10.0 http://yetus.apache.org |


This message was automatically generated.



> the property of sharedcache in mapred-default.xml
> -
>
> Key: YARN-9579
> URL: https://issues.apache.org/jira/browse/YARN-9579
> Project: Hadoop 

[jira] [Commented] (YARN-9601) Potential NPE in ZookeeperFederationStateStore#getPoliciesConfigurations

2019-07-26 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9601?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893870#comment-16893870
 ] 

Hadoop QA commented on YARN-9601:
-

| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  1m  
3s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} dupname {color} | {color:green}  0m  
0s{color} | {color:green} No case conflicting files found. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 20m 
39s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
38s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
21s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
40s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
11m 59s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
23s{color} | {color:green} trunk passed {color} |
| {color:blue}0{color} | {color:blue} spotbugs {color} | {color:blue}  1m 
16s{color} | {color:blue} Used deprecated FindBugs config; considering 
switching to SpotBugs. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
13s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
32s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
30s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
30s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
14s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
34s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
12m 50s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
24s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  2m 
53s{color} | {color:green} hadoop-yarn-server-common in the patch passed. 
{color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
28s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 57m 53s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=19.03.1 Server=19.03.1 base: 
https://builds.apache.org/job/hadoop-multibranch/job/PR-908/3/artifact/out/Dockerfile
 |
| GITHUB PR | https://github.com/apache/hadoop/pull/908 |
| JIRA Issue | YARN-9601 |
| Optional Tests | dupname asflicense compile javac javadoc mvninstall mvnsite 
unit shadedclient findbugs checkstyle |
| uname | Linux 19d3ebdf9e02 4.4.0-138-generic #164-Ubuntu SMP Tue Oct 2 
17:16:02 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | personality/hadoop.sh |
| git revision | trunk / c0a0c35 |
| Default Java | 1.8.0_212 |
|  Test Results | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-908/3/testReport/ |
| Max. process+thread count | 412 (vs. ulimit of 5500) |
| modules | C: 

[jira] [Comment Edited] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16890883#comment-16890883
 ] 

Babble Shack edited comment on YARN-9690 at 7/26/19 2:30 PM:
-

[~bibinchundatt]

I have tried with
{code:java}
  
      yarn.resourcemanager.scheduler.address
      0.0.0.0:8049
  {code}
+I also tried with your suggestion (localhost instead of 0.0.0.0).+

but that did not work either. I have attached an additional log

The stack trace is different with the localhost modification, however the cause 
is still "Invalid AMRMToken from appattempt_1563878096961_0001_01"
{code:java}
2019-07-23 10:35:21,517 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: 40% of the mappers 
will be scheduled using OPPORTUNISTIC containers
2019-07-23 10:35:21,530 INFO [main] org.apache.hadoop.yarn.client.RMProxy: 
Connecting to ResourceManager at localhost/127.0.0.1:8049
2019-07-23 10:35:21,568 ERROR [main] 
org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator: Exception while 
registering
org.apache.hadoop.security.token.SecretManager$InvalidToken: Invalid AMRMToken 
from appattempt_1563878096961_0001_02
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
    at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53)
    at 
org.apache.hadoop.yarn.ipc.RPCUtil.instantiateIOException(RPCUtil.java:80)
    at 
org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:119)
    at 
org.apache.hadoop.yarn.server.api.impl.pb.client.DistributedSchedulingAMProtocolPBClientImpl.registerApplicationMasterForDistributedScheduling(DistributedSchedulingAMProtocolPBClientImpl.java:90)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:498)
    at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422)
    at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165)
    at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157)
    at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
    at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359)
    at 
com.sun.proxy.$Proxy79.registerApplicationMasterForDistributedScheduling(Unknown
 Source)
    at 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.DefaultRequestInterceptor.registerApplicationMasterForDistributedScheduling(DefaultRequestInterceptor.java:154)
    at 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler.registerApplicationMasterForDistributedScheduling(DistributedScheduler.java:214)
    at 
org.apache.hadoop.yarn.server.nodemanager.scheduler.DistributedScheduler.registerApplicationMaster(DistributedScheduler.java:133)
    at 
org.apache.hadoop.yarn.server.nodemanager.amrmproxy.AMRMProxyService.registerApplicationMaster(AMRMProxyService.java:295)
    at 
org.apache.hadoop.yarn.api.impl.pb.service.ApplicationMasterProtocolPBServiceImpl.registerApplicationMaster(ApplicationMasterProtocolPBServiceImpl.java:90)
    at 
org.apache.hadoop.yarn.proto.ApplicationMasterProtocol$ApplicationMasterProtocolService$2.callBlockingMethod(ApplicationMasterProtocol.java:95)
    at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:523)
    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:991)
    at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:872)
    at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:818)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:422)
    at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2678)
Caused by: 
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.token.SecretManager$InvalidToken):
 Invalid AMRMToken from appattempt_1563878096961_0001_02
{code}
[^distributed_log]


was (Author: babbleshack):
[~bibinchundatt]

I have tried with
{code:java}
  
      yarn.resourcemanager.scheduler.address
      0.0.0.0:8049
  {code}
I also tried with your suggestion (localhost instead of 0.0.0.0).

but that did not work either. I have attached an additional log

The stack trace is different with the 

[jira] [Updated] (YARN-9690) Invalid AMRM token when distributed scheduling is enabled.

2019-07-26 Thread Babble Shack (JIRA)


 [ 
https://issues.apache.org/jira/browse/YARN-9690?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Babble Shack updated YARN-9690:
---
Description: 
Applications fail to start due to invalild AMRM from application attempt.

I have tested this with 0/100% opportunistic maps and the same issue occurs 
regardless.
{code:java}

-->

  
    mapreduceyarn.nodemanager.aux-services
    mapreduce_shuffle
  
  
      yarn.resourcemanager.address
      yarn-master-0.yarn-service.yarn:8032
  
  
      yarn.resourcemanager.scheduler.address
      0.0.0.0:8049
  

  
    yarn.resourcemanager.opportunistic-container-allocation.enabled
    true
  
  
    yarn.nodemanager.opportunistic-containers-max-queue-length
    10
  
  
    yarn.nodemanager.distributed-scheduling.enabled
    true
  
 
  
    yarn.webapp.ui2.enable
    true
  
  
      yarn.resourcemanager.resource-tracker.address
      yarn-master-0.yarn-service.yarn:8031
  
  
    yarn.log-aggregation-enable
    true
  
  
      yarn.nodemanager.aux-services
      mapreduce_shuffle
  

  
  

  
  
    yarn.nodemanager.resource.memory-mb
    7168
  
  
    yarn.scheduler.minimum-allocation-mb
    3584
  
  
    yarn.scheduler.maximum-allocation-mb
    7168
  
  
    yarn.app.mapreduce.am.resource.mb
    7168
  
  
  
    yarn.app.mapreduce.am.command-opts
    -Xmx5734m
  

  
  
    yarn.timeline-service.enabled
    true
  
  
    yarn.resourcemanager.system-metrics-publisher.enabled
    true
  
  
    yarn.timeline-service.generic-application-history.enabled
    true
  
  
    yarn.timeline-service.bind-host
    0.0.0.0
  

{code}
Relevant logs:
{code:java}
2019-07-22 14:56:37,104 INFO [main] 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator: 100% of the mappers 
will be scheduled using OPPORTUNISTIC containers
2019-07-22 14:56:37,117 INFO [main] org.apache.hadoop.yarn.client.RMProxy: 
Connecting to ResourceManager at 
yarn-master-0.yarn-service.yarn/10.244.1.134:8030
2019-07-22 14:56:37,150 WARN [main] org.apache.hadoop.ipc.Client: Exception 
encountered while connecting to the server : 
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.token.SecretManager$InvalidToken):
 Invalid AMRMToken from appattempt_1563805140414_0002_02
2019-07-22 14:56:37,152 ERROR [main] 
org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator: Exception while 
registering
org.apache.hadoop.security.token.SecretManager$InvalidToken: Invalid AMRMToken 
from appattempt_1563805140414_0002_02
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at 
sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    at 
sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
    at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53)
    at 
org.apache.hadoop.yarn.ipc.RPCUtil.instantiateIOException(RPCUtil.java:80)
    at 
org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:119)
    at 
org.apache.hadoop.yarn.api.impl.pb.client.ApplicationMasterProtocolPBClientImpl.registerApplicationMaster(ApplicationMasterProtocolPBClientImpl.java:109)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
    at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:498)
    at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invokeMethod(RetryInvocationHandler.java:422)
    at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeMethod(RetryInvocationHandler.java:165)
    at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invoke(RetryInvocationHandler.java:157)
    at 
org.apache.hadoop.io.retry.RetryInvocationHandler$Call.invokeOnce(RetryInvocationHandler.java:95)
    at 
org.apache.hadoop.io.retry.RetryInvocationHandler.invoke(RetryInvocationHandler.java:359)
    at com.sun.proxy.$Proxy82.registerApplicationMaster(Unknown Source)
    at 
org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator.register(RMCommunicator.java:160)
    at 
org.apache.hadoop.mapreduce.v2.app.rm.RMCommunicator.serviceStart(RMCommunicator.java:121)
    at 
org.apache.hadoop.mapreduce.v2.app.rm.RMContainerAllocator.serviceStart(RMContainerAllocator.java:274)
    at org.apache.hadoop.service.AbstractService.start(AbstractService.java:194)
    at 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster$ContainerAllocatorRouter.serviceStart(MRAppMaster.java:959)
    at org.apache.hadoop.service.AbstractService.start(AbstractService.java:194)
    at 
org.apache.hadoop.service.CompositeService.serviceStart(CompositeService.java:121)
    at 
org.apache.hadoop.mapreduce.v2.app.MRAppMaster.serviceStart(MRAppMaster.java:1272)
    at 

[jira] [Commented] (YARN-7621) Support submitting apps with queue path for CapacityScheduler

2019-07-26 Thread Weiwei Yang (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-7621?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893856#comment-16893856
 ] 

Weiwei Yang commented on YARN-7621:
---

[~cane], could you pls help to review [~Tao Yang]'s patch? Just want to cross 
check.

Thanks

> Support submitting apps with queue path for CapacityScheduler
> -
>
> Key: YARN-7621
> URL: https://issues.apache.org/jira/browse/YARN-7621
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler
>Reporter: Tao Yang
>Assignee: Tao Yang
>Priority: Major
>  Labels: fs2cs
> Attachments: YARN-7621.001.patch, YARN-7621.002.patch
>
>
> Currently there is a difference of queue definition in 
> ApplicationSubmissionContext between CapacityScheduler and FairScheduler. 
> FairScheduler needs queue path but CapacityScheduler needs queue name. There 
> is no doubt of the correction of queue definition for CapacityScheduler 
> because it does not allow duplicate leaf queue names, but it's hard to switch 
> between FairScheduler and CapacityScheduler. I propose to support submitting 
> apps with queue path for CapacityScheduler to make the interface clearer and 
> scheduler switch smoothly.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-8045) Reduce log output from container status calls

2019-07-26 Thread Jim Brennan (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-8045?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893841#comment-16893841
 ] 

Jim Brennan commented on YARN-8045:
---

Would really like to see this pulled back to 2.8 - it looks like it will be 
clean.

> Reduce log output from container status calls
> -
>
> Key: YARN-8045
> URL: https://issues.apache.org/jira/browse/YARN-8045
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Shane Kumpf
>Assignee: Craig Condit
>Priority: Major
> Fix For: 3.2.0
>
> Attachments: YARN-8045.001.patch
>
>
> Each time a container's status is returned a log entry is produced in the NM 
> from {{ContainerManagerImpl}}. The container status includes the diagnostics 
> field for the container. If the diagnostics field contains an exception, it 
> can appear as if the exception is logged repeatedly every second. The 
> diagnostics message can also span many lines, which puts pressure on the logs 
> and makes it harder to read.
> For example:
> {code}
> 2018-03-17 22:01:11,632 INFO 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.ContainerManagerImpl:
>  Getting container-status for container_e01_1521323860653_0001_01_05
> 2018-03-17 22:01:11,632 INFO 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.ContainerManagerImpl:
>  Returning ContainerStatus: [ContainerId: 
> container_e01_1521323860653_0001_01_05, ExecutionType: GUARANTEED, State: 
> RUNNING, Capability: , Diagnostics: [2018-03-17 
> 22:01:00.675]Exception from container-launch.
> Container id: container_e01_1521323860653_0001_01_05
> Exit code: -1
> Exception message: 
> Shell ouput: 
> [2018-03-17 22:01:00.750]Diagnostic message from attempt :
> [2018-03-17 22:01:00.750]Container exited with a non-zero exit code -1.
> , ExitStatus: -1, IP: null, Host: null, ContainerSubState: SCHEDULED]
> {code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-9698) [Umbrella] Tools to help migration from Fair Scheduler to Capacity Scheduler

2019-07-26 Thread Tan, Wangda (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9698?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893792#comment-16893792
 ] 

Tan, Wangda commented on YARN-9698:
---

[~cane] , is the feature you mentioned supported by FairScheduler? Or it is 
just a new feature. 

> [Umbrella] Tools to help migration from Fair Scheduler to Capacity Scheduler
> 
>
> Key: YARN-9698
> URL: https://issues.apache.org/jira/browse/YARN-9698
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacity scheduler
>Reporter: Weiwei Yang
>Priority: Major
>  Labels: fs2cs
>
> We see some users want to migrate from Fair Scheduler to Capacity Scheduler, 
> this Jira is created as an umbrella to track all related efforts for the 
> migration, the scope contains
>  * Bug fixes
>  * Add missing features
>  * Migration tools that help to generate CS configs based on FS, validate 
> configs etc
>  * Documents
> this is part of CS component, the purpose is to make the migration process 
> smooth.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (YARN-9698) [Umbrella] Tools to help migration from Fair Scheduler to Capacity Scheduler

2019-07-26 Thread zhoukang (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9698?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893744#comment-16893744
 ] 

zhoukang edited comment on YARN-9698 at 7/26/19 11:37 AM:
--

How about the idea that support ACL proxy user in CapacityScheduler.
Explanation: submit an app use command as below
{code:java}
./bin/yarn  jar  ~/workspace/yarn/Test3.1/wordcount-1.0-SNAPSHOT.jar 
com.xxx.xxx.WordCount -Dmapred.job.queue.name=user@default /tmp/input.dat 
/tmp/output
{code}
[~cheersyang]


was (Author: cane):
How about the idea that support ACL proxy user in CapacityScheduler.
Explanation: submit an app use command below
{code:java}
./bin/yarn  jar  ~/workspace/yarn/Test3.1/wordcount-1.0-SNAPSHOT.jar 
com.xxx.xxx.WordCount -Dmapred.job.queue.name=user@default /tmp/input.dat 
/tmp/output
{code}
[~cheersyang]

> [Umbrella] Tools to help migration from Fair Scheduler to Capacity Scheduler
> 
>
> Key: YARN-9698
> URL: https://issues.apache.org/jira/browse/YARN-9698
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacity scheduler
>Reporter: Weiwei Yang
>Priority: Major
>  Labels: fs2cs
>
> We see some users want to migrate from Fair Scheduler to Capacity Scheduler, 
> this Jira is created as an umbrella to track all related efforts for the 
> migration, the scope contains
>  * Bug fixes
>  * Add missing features
>  * Migration tools that help to generate CS configs based on FS, validate 
> configs etc
>  * Documents
> this is part of CS component, the purpose is to make the migration process 
> smooth.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-9698) [Umbrella] Tools to help migration from Fair Scheduler to Capacity Scheduler

2019-07-26 Thread zhoukang (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9698?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893744#comment-16893744
 ] 

zhoukang commented on YARN-9698:


How about the idea that support ACL proxy user in CapacityScheduler.
Explanation: submit an app use command below
{code:java}
./bin/yarn  jar  ~/workspace/yarn/Test3.1/wordcount-1.0-SNAPSHOT.jar 
com.xxx.xxx.WordCount -Dmapred.job.queue.name=user@default /tmp/input.dat 
/tmp/output
{code}
[~cheersyang]

> [Umbrella] Tools to help migration from Fair Scheduler to Capacity Scheduler
> 
>
> Key: YARN-9698
> URL: https://issues.apache.org/jira/browse/YARN-9698
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: capacity scheduler
>Reporter: Weiwei Yang
>Priority: Major
>  Labels: fs2cs
>
> We see some users want to migrate from Fair Scheduler to Capacity Scheduler, 
> this Jira is created as an umbrella to track all related efforts for the 
> migration, the scope contains
>  * Bug fixes
>  * Add missing features
>  * Migration tools that help to generate CS configs based on FS, validate 
> configs etc
>  * Documents
> this is part of CS component, the purpose is to make the migration process 
> smooth.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-9709) When we expanding queue list the scheduler page will not show any applications

2019-07-26 Thread zhoukang (JIRA)


 [ 
https://issues.apache.org/jira/browse/YARN-9709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

zhoukang updated YARN-9709:
---
Attachment: (was: list2.png)

> When we expanding queue list the scheduler page will not show any applications
> --
>
> Key: YARN-9709
> URL: https://issues.apache.org/jira/browse/YARN-9709
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler
>Affects Versions: 3.1.2
>Reporter: zhoukang
>Priority: Major
> Attachments: list1.png, list3.png
>
>
> When we expanding queue list the scheduler page will not show any 
> applications.But it works well in FairScheduler.
>  !list1.png!  !list2.png! 



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-9709) When we expanding queue list the scheduler page will not show any applications

2019-07-26 Thread zhoukang (JIRA)


 [ 
https://issues.apache.org/jira/browse/YARN-9709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

zhoukang updated YARN-9709:
---
Description: 
When we expanding queue list the scheduler page will not show any 
applications.But it works well in FairScheduler.
 !list1.png! 
 !list3.png! 

  was:
When we expanding queue list the scheduler page will not show any 
applications.But it works well in FairScheduler.
 !list1.png!  !list2.png! 


> When we expanding queue list the scheduler page will not show any applications
> --
>
> Key: YARN-9709
> URL: https://issues.apache.org/jira/browse/YARN-9709
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler
>Affects Versions: 3.1.2
>Reporter: zhoukang
>Priority: Major
> Attachments: list1.png, list3.png
>
>
> When we expanding queue list the scheduler page will not show any 
> applications.But it works well in FairScheduler.
>  !list1.png! 
>  !list3.png! 



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Updated] (YARN-9709) When we expanding queue list the scheduler page will not show any applications

2019-07-26 Thread zhoukang (JIRA)


 [ 
https://issues.apache.org/jira/browse/YARN-9709?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

zhoukang updated YARN-9709:
---
Attachment: list3.png

> When we expanding queue list the scheduler page will not show any applications
> --
>
> Key: YARN-9709
> URL: https://issues.apache.org/jira/browse/YARN-9709
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler
>Affects Versions: 3.1.2
>Reporter: zhoukang
>Priority: Major
> Attachments: list1.png, list3.png
>
>
> When we expanding queue list the scheduler page will not show any 
> applications.But it works well in FairScheduler.
>  !list1.png!  !list2.png! 



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Created] (YARN-9709) When we expanding queue list the scheduler page will not show any applications

2019-07-26 Thread zhoukang (JIRA)
zhoukang created YARN-9709:
--

 Summary: When we expanding queue list the scheduler page will not 
show any applications
 Key: YARN-9709
 URL: https://issues.apache.org/jira/browse/YARN-9709
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: capacityscheduler
Affects Versions: 3.1.2
Reporter: zhoukang
 Attachments: list1.png, list2.png

When we expanding queue list the scheduler page will not show any 
applications.But it works well in FairScheduler.
 !list1.png!  !list2.png! 



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-7621) Support submitting apps with queue path for CapacityScheduler

2019-07-26 Thread zhoukang (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-7621?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893728#comment-16893728
 ] 

zhoukang commented on YARN-7621:


Good job.
We also encountered the same problem and have made some compatibility

> Support submitting apps with queue path for CapacityScheduler
> -
>
> Key: YARN-7621
> URL: https://issues.apache.org/jira/browse/YARN-7621
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: capacityscheduler
>Reporter: Tao Yang
>Assignee: Tao Yang
>Priority: Major
>  Labels: fs2cs
> Attachments: YARN-7621.001.patch, YARN-7621.002.patch
>
>
> Currently there is a difference of queue definition in 
> ApplicationSubmissionContext between CapacityScheduler and FairScheduler. 
> FairScheduler needs queue path but CapacityScheduler needs queue name. There 
> is no doubt of the correction of queue definition for CapacityScheduler 
> because it does not allow duplicate leaf queue names, but it's hard to switch 
> between FairScheduler and CapacityScheduler. I propose to support submitting 
> apps with queue path for CapacityScheduler to make the interface clearer and 
> scheduler switch smoothly.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Created] (YARN-9708) Add Yarnclient#getDelegationToken API implementation and SecureLogin in router

2019-07-26 Thread Xie YiFan (JIRA)
Xie YiFan created YARN-9708:
---

 Summary: Add Yarnclient#getDelegationToken API implementation and 
SecureLogin in router
 Key: YARN-9708
 URL: https://issues.apache.org/jira/browse/YARN-9708
 Project: Hadoop YARN
  Issue Type: New Feature
  Components: router
Affects Versions: 3.1.1
Reporter: Xie YiFan
 Attachments: Add_getDelegationToken_and_SecureLogin_in_router.patch

1.we use router as proxy to manage multiple cluster which be independent of 
each other in order to apply unified client. Thus, we implement our customized 
AMRMProxyPolicy that doesn't broadcast ResourceRequest to other cluster.

2.Our production environment need kerberos. But router doesn't support 
SecureLogin for now.
https://issues.apache.org/jira/browse/YARN-6539 desn't work. So we improvement 
it.

3.Some framework like oozie would get Token via yarnclient#getDelegationToken 
which router doesn't support. Our solution is that adding homeCluster to 
ApplicationSubmissionContextProto & GetDelegationTokenRequestProto. Job would 
be submitted with specified clusterid so that router knows which cluster to 
submit this job. Router would get Token from one RM according to specified 
clusterid when client call getDelegation meanwhile apply some mechanism to save 
this token in memory.

 



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Assigned] (YARN-9707) [UI2] App Attempt state data is missing

2019-07-26 Thread Szilard Nemeth (JIRA)


 [ 
https://issues.apache.org/jira/browse/YARN-9707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szilard Nemeth reassigned YARN-9707:


Assignee: Zoltan Siegl

> [UI2] App Attempt state data is missing
> ---
>
> Key: YARN-9707
> URL: https://issues.apache.org/jira/browse/YARN-9707
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Yesha Vora
>Assignee: Zoltan Siegl
>Priority: Critical
>
> Steps:
> 1) Launch a Dshell application or Yarn service application.
> 2) Go to app attempt page Grid view. State column shows N/A.
> Yarn UI1 shows app attempt state for Running and Finished application. This 
> ability is missing from UI2.
> UI2 is using below rest call. This rest call does not show the app attempt 
> state details.
> {code:title=ws/v1/cluster/apps/application_1563946396350_0002/appattempts?_=1564004553389}
> 115640045242901564004541852container_1563946396350_0002_01_01xx:yyxx:yyhttp://ixx:yy/node/containerlogs/container_1563946396350_0002_01_01/hrt_qaappattempt_1563946396350_0002_01{code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Assigned] (YARN-9704) [UI2] Fix Pending, Allocated, Reserved Containers information for Fair Scheduler

2019-07-26 Thread Szilard Nemeth (JIRA)


 [ 
https://issues.apache.org/jira/browse/YARN-9704?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szilard Nemeth reassigned YARN-9704:


Assignee: Zoltan Siegl

> [UI2] Fix Pending, Allocated, Reserved Containers information for Fair 
> Scheduler
> 
>
> Key: YARN-9704
> URL: https://issues.apache.org/jira/browse/YARN-9704
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Yesha Vora
>Assignee: Zoltan Siegl
>Priority: Critical
> Attachments: Screenshot 2019-07-25 at 10.51.08 PM.png
>
>
> YARN UI2 shows "Pending, Allocated, Reserved Containers" information for fair 
> scheduler. In here, pending container information is not printed. UI2 shows 
> ",0,0" instead "0,0,0".
> In UI1, This same information is displayed as Num of active Application & 
> Pending applications.
> Num Active Applications:  0
> Num Pending Applications: 0
> It's not clear from UI2 what do we intend to show in "Pending, Allocated, 
> Reserved Containers"? Is it really containers or apps?



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Assigned] (YARN-9706) [UI2] App Attempt state missing from Graph view

2019-07-26 Thread Szilard Nemeth (JIRA)


 [ 
https://issues.apache.org/jira/browse/YARN-9706?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szilard Nemeth reassigned YARN-9706:


Assignee: Zoltan Siegl

> [UI2] App Attempt state missing from Graph view
> ---
>
> Key: YARN-9706
> URL: https://issues.apache.org/jira/browse/YARN-9706
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Yesha Vora
>Assignee: Zoltan Siegl
>Priority: Major
>
> 1) Launch a Dshell application or Yarn service application.
> 2) Go to app attempt page Grid view. State column shows N/A.
> 3) Go to app attempt Graph view. State data is not present on this page.
> Apparently, app attempt data is only shown in Grid view. Grid and Graph view 
> should show the same details.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Assigned] (YARN-9705) [UI2] AM Node Web UI should not display full link

2019-07-26 Thread Szilard Nemeth (JIRA)


 [ 
https://issues.apache.org/jira/browse/YARN-9705?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szilard Nemeth reassigned YARN-9705:


Assignee: Zoltan Siegl

> [UI2] AM Node Web UI should not display full link
> -
>
> Key: YARN-9705
> URL: https://issues.apache.org/jira/browse/YARN-9705
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Yesha Vora
>Assignee: Zoltan Siegl
>Priority: Major
>
> App Attempt page shows AM Node Web UI. It displays the full link. 
> It should not print out full link as display text. Rather, It should use 
> display AM Node name which links to the node.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Commented] (YARN-9681) AM resource limit is incorrect for queue

2019-07-26 Thread Hadoop QA (JIRA)


[ 
https://issues.apache.org/jira/browse/YARN-9681?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16893408#comment-16893408
 ] 

Hadoop QA commented on YARN-9681:
-

| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
22s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 18m 
 0s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
48s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
35s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
47s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
11m 30s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
13s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
31s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
45s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
42s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
42s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
27s{color} | {color:green} 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager:
 The patch generated 0 new + 43 unchanged - 2 fixed = 43 total (was 45) {color} 
|
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
47s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
11m 35s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m 
21s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
29s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 81m 
39s{color} | {color:green} hadoop-yarn-server-resourcemanager in the patch 
passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
28s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}132m  3s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=19.03.1 Server=19.03.1 Image:yetus/hadoop:bdbca0e |
| JIRA Issue | YARN-9681 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12975926/YARN-9681.0003.patch |
| Optional Tests |  dupname  asflicense  compile  javac  javadoc  mvninstall  
mvnsite  unit  shadedclient  findbugs  checkstyle  |
| uname | Linux 3b7c3ab53f8f 4.4.0-139-generic #165-Ubuntu SMP Wed Oct 24 
10:58:50 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/patchprocess/precommit/personality/provided.sh |
| git revision | trunk / ce99cc3 |
| maven | version: Apache Maven 3.3.9 |
| Default Java | 1.8.0_212 |
| findbugs | v3.1.0-RC1 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-YARN-Build/24429/testReport/ |
| Max. process+thread count | 926 (vs. ulimit of 1) |
| modules | C: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager
 U: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager
 |
| Console output | 
https://builds.apache.org/job/PreCommit-YARN-Build/24429/console |
| Powered by | Apache 

[jira] [Updated] (YARN-8468) Enable the use of queue based maximum container allocation limit and implement it in FairScheduler

2019-07-26 Thread Szilard Nemeth (JIRA)


 [ 
https://issues.apache.org/jira/browse/YARN-8468?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Szilard Nemeth updated YARN-8468:
-
Description: 
When using any scheduler, you can use "yarn.scheduler.maximum-allocation-mb" to 
limit the overall size of a container. This applies globally to all containers 
and cannot be limited by queue or and is not scheduler dependent.

The goal of this ticket is to allow this value to be set on a per queue basis.

The use case: User has two pools, one for ad hoc jobs and one for enterprise 
apps. User wants to limit ad hoc jobs to small containers but allow enterprise 
apps to request as many resources as needed. Setting 
yarn.scheduler.maximum-allocation-mb sets a default value for maximum container 
size for all queues and setting maximum resources per queue with 
“maxContainerResources” queue config value.
NOTE: "maxContainerResources" has been changed to "maxContainerAllocation" as 
per the review comments.

Suggested solution:

All the infrastructure is already in the code. We need to do the following:
 * add the setting to the queue properties for all queue types (parent and 
leaf), this will cover dynamically created queues.
 * if we set it on the root we override the scheduler setting and we should not 
allow that.
 * make sure that queue resource cap can not be larger than scheduler max 
resource cap in the config.
 * implement getMaximumResourceCapability(String queueName) in the FairScheduler
 * implement getMaximumResourceCapability(String queueName) in both 
FSParentQueue and FSLeafQueue as follows
 * expose the setting in the queue information in the RM web UI.
 * expose the setting in the metrics etc for the queue.
 * Enforce the use of queue based maximum allocation limit if it is available, 
if not use the general scheduler level setting
 ** Use it during validation and normalization of requests in 
scheduler.allocate, app submit and resource request

  was:
When using any scheduler, you can use "yarn.scheduler.maximum-allocation-mb" to 
limit the overall size of a container. This applies globally to all containers 
and cannot be limited by queue or and is not scheduler dependent.

The goal of this ticket is to allow this value to be set on a per queue basis.

The use case: User has two pools, one for ad hoc jobs and one for enterprise 
apps. User wants to limit ad hoc jobs to small containers but allow enterprise 
apps to request as many resources as needed. Setting 
yarn.scheduler.maximum-allocation-mb sets a default value for maximum container 
size for all queues and setting maximum resources per queue with 
“maxContainerResources” queue config value.

Suggested solution:

All the infrastructure is already in the code. We need to do the following:
 * add the setting to the queue properties for all queue types (parent and 
leaf), this will cover dynamically created queues.
 * if we set it on the root we override the scheduler setting and we should not 
allow that.
 * make sure that queue resource cap can not be larger than scheduler max 
resource cap in the config.
 * implement getMaximumResourceCapability(String queueName) in the FairScheduler
 * implement getMaximumResourceCapability(String queueName) in both 
FSParentQueue and FSLeafQueue as follows
 * expose the setting in the queue information in the RM web UI.
 * expose the setting in the metrics etc for the queue.
 * Enforce the use of queue based maximum allocation limit if it is available, 
if not use the general scheduler level setting
 ** Use it during validation and normalization of requests in 
scheduler.allocate, app submit and resource request


> Enable the use of queue based maximum container allocation limit and 
> implement it in FairScheduler
> --
>
> Key: YARN-8468
> URL: https://issues.apache.org/jira/browse/YARN-8468
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: fairscheduler, scheduler
>Affects Versions: 3.1.0
>Reporter: Antal Bálint Steinbach
>Assignee: Antal Bálint Steinbach
>Priority: Critical
> Fix For: 3.2.0
>
> Attachments: YARN-8468-branch-3.1.018.patch, 
> YARN-8468-branch-3.1.019.patch, YARN-8468-branch-3.1.020.patch, 
> YARN-8468-branch-3.1.021.patch, YARN-8468-branch-3.1.022.patch, 
> YARN-8468.000.patch, YARN-8468.001.patch, YARN-8468.002.patch, 
> YARN-8468.003.patch, YARN-8468.004.patch, YARN-8468.005.patch, 
> YARN-8468.006.patch, YARN-8468.007.patch, YARN-8468.008.patch, 
> YARN-8468.009.patch, YARN-8468.010.patch, YARN-8468.011.patch, 
> YARN-8468.012.patch, YARN-8468.013.patch, YARN-8468.014.patch, 
> YARN-8468.015.patch, YARN-8468.016.patch, YARN-8468.017.patch, 
> YARN-8468.018.patch
>
>
> When using any scheduler, you can use 

[jira] [Created] (YARN-9706) [UI2] App Attempt state missing from Graph view

2019-07-26 Thread Yesha Vora (JIRA)
Yesha Vora created YARN-9706:


 Summary: [UI2] App Attempt state missing from Graph view
 Key: YARN-9706
 URL: https://issues.apache.org/jira/browse/YARN-9706
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Yesha Vora



1) Launch a Dshell application or Yarn service application.
2) Go to app attempt page Grid view. State column shows N/A.
3) Go to app attempt Graph view. State data is not present on this page.

Apparently, app attempt data is only shown in Grid view. Grid and Graph view 
should show the same details.



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org



[jira] [Created] (YARN-9707) [UI2] App Attempt state data is missing

2019-07-26 Thread Yesha Vora (JIRA)
Yesha Vora created YARN-9707:


 Summary: [UI2] App Attempt state data is missing
 Key: YARN-9707
 URL: https://issues.apache.org/jira/browse/YARN-9707
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Yesha Vora


Steps:

1) Launch a Dshell application or Yarn service application.
2) Go to app attempt page Grid view. State column shows N/A.

Yarn UI1 shows app attempt state for Running and Finished application. This 
ability is missing from UI2.

UI2 is using below rest call. This rest call does not show the app attempt 
state details.

{code:title=ws/v1/cluster/apps/application_1563946396350_0002/appattempts?_=1564004553389}
115640045242901564004541852container_1563946396350_0002_01_01xx:yyxx:yyhttp://ixx:yy/node/containerlogs/container_1563946396350_0002_01_01/hrt_qaappattempt_1563946396350_0002_01{code}



--
This message was sent by Atlassian JIRA
(v7.6.14#76016)

-
To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org