[jira] [Assigned] (YARN-1531) Update yarn command document

2013-12-26 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA reassigned YARN-1531:
---

Assignee: Akira AJISAKA

 Update yarn command document
 

 Key: YARN-1531
 URL: https://issues.apache.org/jira/browse/YARN-1531
 Project: Hadoop YARN
  Issue Type: Bug
  Components: documentation
Reporter: Akira AJISAKA
Assignee: Akira AJISAKA

 There are some options which are not written to Yarn Command document.
 For example, yarn rmadmin command options are as follows:
 {code}
  Usage: yarn rmadmin
-refreshQueues 
-refreshNodes 
-refreshSuperUserGroupsConfiguration 
-refreshUserToGroupsMappings 
-refreshAdminAcls 
-refreshServiceAcl 
-getGroups [username]
-help [cmd]
-transitionToActive serviceId
-transitionToStandby serviceId
-failover [--forcefence] [--forceactive] serviceId serviceId
-getServiceState serviceId
-checkHealth serviceId
 {code}
 But some of the new options such as -getGroups, -transitionToActive, and 
 -transitionToStandby are not documented.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (YARN-1409) NonAggregatingLogHandler can throw RejectedExecutionException

2013-12-26 Thread Tsuyoshi OZAWA (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1409?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsuyoshi OZAWA updated YARN-1409:
-

Attachment: YARN-1409.3.patch

Thank you for the review, [~jlowe]. I apologize for the delay.
1. Added a test case to reproduce this issue
2. Fixed to avoid creating LogDeleterRunnable instances twice.

 NonAggregatingLogHandler can throw RejectedExecutionException
 -

 Key: YARN-1409
 URL: https://issues.apache.org/jira/browse/YARN-1409
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Tsuyoshi OZAWA
Assignee: Tsuyoshi OZAWA
 Attachments: YARN-1409.1.patch, YARN-1409.2.patch, YARN-1409.3.patch


 This problem is caused by handling APPLICATION_FINISHED events after calling 
 sched.shotdown() in NonAggregatingLongHandler#serviceStop(). 
 org.apache.hadoop.mapred.TestJobCleanup can fail because of 
 RejectedExecutionException by NonAggregatingLogHandler.
 {code}
 2013-11-13 10:53:06,970 FATAL [AsyncDispatcher event handler] 
 event.AsyncDispatcher (AsyncDispatcher.java:dispatch(166)) - Error in 
 dispatcher thread
 java.util.concurrent.RejectedExecutionException: Task 
 java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask@d51df63 
 rejected from 
 java.util.concurrent.ScheduledThreadPoolExecutor@7a20e369[Shutting down, pool 
 size = 4, active threads = 0, queued tasks = 7, completed tasks = 0]
 at 
 java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2048)
 at 
 java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:821)
 at 
 java.util.concurrent.ScheduledThreadPoolExecutor.delayedExecute(ScheduledThreadPoolExecutor.java:325)
 at 
 java.util.concurrent.ScheduledThreadPoolExecutor.schedule(ScheduledThreadPoolExecutor.java:530)
 at 
 org.apache.hadoop.yarn.server.nodemanager.containermanager.loghandler.NonAggregatingLogHandler.handle(NonAggregatingLogHandler.java:121)
 at 
 org.apache.hadoop.yarn.server.nodemanager.containermanager.loghandler.NonAggregatingLogHandler.handle(NonAggregatingLogHandler.java:49)
 at 
 org.apache.hadoop.yarn.event.AsyncDispatcher.dispatch(AsyncDispatcher.java:159)
 at 
 org.apache.hadoop.yarn.event.AsyncDispatcher$1.run(AsyncDispatcher.java:95)
 at java.lang.Thread.run(Thread.java:724)
 {code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1409) NonAggregatingLogHandler can throw RejectedExecutionException

2013-12-26 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1409?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13856816#comment-13856816
 ] 

Hadoop QA commented on YARN-1409:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12620481/YARN-1409.3.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/2729//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/2729//console

This message is automatically generated.

 NonAggregatingLogHandler can throw RejectedExecutionException
 -

 Key: YARN-1409
 URL: https://issues.apache.org/jira/browse/YARN-1409
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Tsuyoshi OZAWA
Assignee: Tsuyoshi OZAWA
 Attachments: YARN-1409.1.patch, YARN-1409.2.patch, YARN-1409.3.patch


 This problem is caused by handling APPLICATION_FINISHED events after calling 
 sched.shotdown() in NonAggregatingLongHandler#serviceStop(). 
 org.apache.hadoop.mapred.TestJobCleanup can fail because of 
 RejectedExecutionException by NonAggregatingLogHandler.
 {code}
 2013-11-13 10:53:06,970 FATAL [AsyncDispatcher event handler] 
 event.AsyncDispatcher (AsyncDispatcher.java:dispatch(166)) - Error in 
 dispatcher thread
 java.util.concurrent.RejectedExecutionException: Task 
 java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask@d51df63 
 rejected from 
 java.util.concurrent.ScheduledThreadPoolExecutor@7a20e369[Shutting down, pool 
 size = 4, active threads = 0, queued tasks = 7, completed tasks = 0]
 at 
 java.util.concurrent.ThreadPoolExecutor$AbortPolicy.rejectedExecution(ThreadPoolExecutor.java:2048)
 at 
 java.util.concurrent.ThreadPoolExecutor.reject(ThreadPoolExecutor.java:821)
 at 
 java.util.concurrent.ScheduledThreadPoolExecutor.delayedExecute(ScheduledThreadPoolExecutor.java:325)
 at 
 java.util.concurrent.ScheduledThreadPoolExecutor.schedule(ScheduledThreadPoolExecutor.java:530)
 at 
 org.apache.hadoop.yarn.server.nodemanager.containermanager.loghandler.NonAggregatingLogHandler.handle(NonAggregatingLogHandler.java:121)
 at 
 org.apache.hadoop.yarn.server.nodemanager.containermanager.loghandler.NonAggregatingLogHandler.handle(NonAggregatingLogHandler.java:49)
 at 
 org.apache.hadoop.yarn.event.AsyncDispatcher.dispatch(AsyncDispatcher.java:159)
 at 
 org.apache.hadoop.yarn.event.AsyncDispatcher$1.run(AsyncDispatcher.java:95)
 at java.lang.Thread.run(Thread.java:724)
 {code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1172) Convert *SecretManagers in the RM to services

2013-12-26 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1172?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13856824#comment-13856824
 ] 

Hudson commented on YARN-1172:
--

FAILURE: Integrated in Hadoop-Yarn-trunk #433 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/433/])
YARN-1172. Convert SecretManagers in RM to services (Tsuyoshi OZAWA via kasha) 
(kasha: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1553431)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMSecretManagerService.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockRM.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/QueueACLsTestBase.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestClientRMService.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRMRestart.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/security/TestClientToAMTokens.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/security/TestRMDelegationTokens.java


 Convert *SecretManagers in the RM to services
 -

 Key: YARN-1172
 URL: https://issues.apache.org/jira/browse/YARN-1172
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: resourcemanager
Affects Versions: 2.1.0-beta
Reporter: Karthik Kambatla
Assignee: Tsuyoshi OZAWA
 Fix For: 2.4.0

 Attachments: YARN-1172.1.patch, YARN-1172.10.patch, 
 YARN-1172.11.patch, YARN-1172.12.patch, YARN-1172.13.patch, 
 YARN-1172.14.patch, YARN-1172.15.patch, YARN-1172.2.patch, YARN-1172.3.patch, 
 YARN-1172.4.patch, YARN-1172.5.patch, YARN-1172.6.patch, YARN-1172.7.patch, 
 YARN-1172.8.patch, YARN-1172.9.patch, yarn-1172-16.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1172) Convert *SecretManagers in the RM to services

2013-12-26 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1172?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13856869#comment-13856869
 ] 

Hudson commented on YARN-1172:
--

FAILURE: Integrated in Hadoop-Hdfs-trunk #1624 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1624/])
YARN-1172. Convert SecretManagers in RM to services (Tsuyoshi OZAWA via kasha) 
(kasha: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1553431)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMSecretManagerService.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockRM.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/QueueACLsTestBase.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestClientRMService.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRMRestart.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/security/TestClientToAMTokens.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/security/TestRMDelegationTokens.java


 Convert *SecretManagers in the RM to services
 -

 Key: YARN-1172
 URL: https://issues.apache.org/jira/browse/YARN-1172
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: resourcemanager
Affects Versions: 2.1.0-beta
Reporter: Karthik Kambatla
Assignee: Tsuyoshi OZAWA
 Fix For: 2.4.0

 Attachments: YARN-1172.1.patch, YARN-1172.10.patch, 
 YARN-1172.11.patch, YARN-1172.12.patch, YARN-1172.13.patch, 
 YARN-1172.14.patch, YARN-1172.15.patch, YARN-1172.2.patch, YARN-1172.3.patch, 
 YARN-1172.4.patch, YARN-1172.5.patch, YARN-1172.6.patch, YARN-1172.7.patch, 
 YARN-1172.8.patch, YARN-1172.9.patch, yarn-1172-16.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1172) Convert *SecretManagers in the RM to services

2013-12-26 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1172?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13856899#comment-13856899
 ] 

Hudson commented on YARN-1172:
--

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1650 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1650/])
YARN-1172. Convert SecretManagers in RM to services (Tsuyoshi OZAWA via kasha) 
(kasha: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1553431)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMSecretManagerService.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/MockRM.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/QueueACLsTestBase.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestClientRMService.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/TestRMRestart.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/security/TestClientToAMTokens.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/security/TestRMDelegationTokens.java


 Convert *SecretManagers in the RM to services
 -

 Key: YARN-1172
 URL: https://issues.apache.org/jira/browse/YARN-1172
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: resourcemanager
Affects Versions: 2.1.0-beta
Reporter: Karthik Kambatla
Assignee: Tsuyoshi OZAWA
 Fix For: 2.4.0

 Attachments: YARN-1172.1.patch, YARN-1172.10.patch, 
 YARN-1172.11.patch, YARN-1172.12.patch, YARN-1172.13.patch, 
 YARN-1172.14.patch, YARN-1172.15.patch, YARN-1172.2.patch, YARN-1172.3.patch, 
 YARN-1172.4.patch, YARN-1172.5.patch, YARN-1172.6.patch, YARN-1172.7.patch, 
 YARN-1172.8.patch, YARN-1172.9.patch, yarn-1172-16.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Resolved] (YARN-1532) Unable to run any MR program on Windows 7 for hadoop2.2 GA Exception from container-launch:

2013-12-26 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth resolved YARN-1532.
-

Resolution: Duplicate

This looks like a duplicate of YARN-1138, which reports that the default 
configuration of {{yarn.application.classpath}} uses Unix-specific syntax for 
environment variables.  Until a patch is available for that, you can work 
around the problem by overriding your configuration in yarn-site.xml to use 
Windows-style environment variables, like this:

{code}
  property
nameyarn.application.classpath/name

value%HADOOP_CONF_DIR%,%HADOOP_COMMON_HOME%/share/hadoop/common/*,%HADOOP_COMMON_HOME%/share/hadoop/common/lib/*,%HADOOP_HDFS_HOME%/share/hadoop/hdfs/*,%HADOOP_HDFS_HOME%/share/hadoop/hdfs/lib/*,%HADOOP_MAPRED_HOME%/share/hadoop/mapreduce/*,%HADOOP_MAPRED_HOME%/share/hadoop/mapreduce/lib/*,%HADOOP_YARN_HOME%/share/hadoop/yarn/*,%HADOOP_YARN_HOME%/share/hadoop/yarn/lib/*/value
  /property
{code}


 Unable to run any MR program on  Windows 7 for hadoop2.2 GA Exception from 
 container-launch:
 --

 Key: YARN-1532
 URL: https://issues.apache.org/jira/browse/YARN-1532
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.2.0
 Environment: Windows 7(64 bit)
Reporter: Hadoop Developer
Priority: Critical

 Unable to run any MR program on Hadoop 2.2 because of some classpath issue 
 which is failing container launch.
 user logs on namenode shows:
 java.lang.NoClassDefFoundError: org/apache/hadoop/service/CompositeService
 and 
 Could not find the main class: 
 org.apache.hadoop.mapreduce.v2.app.MRAppMaster.  Program will exit.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (YARN-1538) Support web-services for application-submission and AM protocols

2013-12-26 Thread Vinod Kumar Vavilapalli (JIRA)
Vinod Kumar Vavilapalli created YARN-1538:
-

 Summary: Support web-services for application-submission and AM 
protocols
 Key: YARN-1538
 URL: https://issues.apache.org/jira/browse/YARN-1538
 Project: Hadoop YARN
  Issue Type: New Feature
Reporter: Vinod Kumar Vavilapalli


We already have read-only web-services for YARN - MAPREDUCE-2863. It'll be 
great to support APIs for application submission and the scheduling protocol - 
as alternatives to RPCs.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1538) Support web-services for application-submission and AM protocols

2013-12-26 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1538?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857030#comment-13857030
 ] 

Alejandro Abdelnur commented on YARN-1538:
--

Is the idea to fully duplicate all AM facing RPC interfaces over HTTP? Having 
scheduling over HTTP but not the NM interface over HTTP does not buy much as 
you'll need to use RPC for starting the containers.

Also, how about security and token support over HTTP?

 Support web-services for application-submission and AM protocols
 

 Key: YARN-1538
 URL: https://issues.apache.org/jira/browse/YARN-1538
 Project: Hadoop YARN
  Issue Type: New Feature
Reporter: Vinod Kumar Vavilapalli

 We already have read-only web-services for YARN - MAPREDUCE-2863. It'll be 
 great to support APIs for application submission and the scheduling protocol 
 - as alternatives to RPCs.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (YARN-1244) Missing yarn queue-cli

2013-12-26 Thread Vinod Kumar Vavilapalli (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1244?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinod Kumar Vavilapalli updated YARN-1244:
--

Target Version/s: 2.3.0
  Labels: newbie  (was: )
  Issue Type: New Feature  (was: Bug)

 Missing yarn queue-cli
 --

 Key: YARN-1244
 URL: https://issues.apache.org/jira/browse/YARN-1244
 Project: Hadoop YARN
  Issue Type: New Feature
Reporter: Vinod Kumar Vavilapalli
Assignee: Vinod Kumar Vavilapalli
  Labels: newbie

 We don't have a yarn queue CLI. For now mapred still has one that is 
 working, but we need to move over that functionality to yarn CLI itself.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Assigned] (YARN-1244) Missing yarn queue-cli

2013-12-26 Thread Vinod Kumar Vavilapalli (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1244?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinod Kumar Vavilapalli reassigned YARN-1244:
-

Assignee: (was: Vinod Kumar Vavilapalli)

 Missing yarn queue-cli
 --

 Key: YARN-1244
 URL: https://issues.apache.org/jira/browse/YARN-1244
 Project: Hadoop YARN
  Issue Type: New Feature
Reporter: Vinod Kumar Vavilapalli
  Labels: newbie

 We don't have a yarn queue CLI. For now mapred still has one that is 
 working, but we need to move over that functionality to yarn CLI itself.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1244) Missing yarn queue-cli

2013-12-26 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1244?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857034#comment-13857034
 ] 

Vinod Kumar Vavilapalli commented on YARN-1244:
---

Today we have and so need to support at least some of 
 - mapred queue -list
 - mapred queue -showacls
 - mapred queue -info
 - mapred queue -showJobs


 Missing yarn queue-cli
 --

 Key: YARN-1244
 URL: https://issues.apache.org/jira/browse/YARN-1244
 Project: Hadoop YARN
  Issue Type: New Feature
Reporter: Vinod Kumar Vavilapalli
  Labels: newbie

 We don't have a yarn queue CLI. For now mapred still has one that is 
 working, but we need to move over that functionality to yarn CLI itself.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (YARN-1539) Queue admin ACLs should NOT be similar to submit-acls w.r.t hierarchy.

2013-12-26 Thread Vinod Kumar Vavilapalli (JIRA)
Vinod Kumar Vavilapalli created YARN-1539:
-

 Summary: Queue admin ACLs should NOT be similar to submit-acls 
w.r.t hierarchy.
 Key: YARN-1539
 URL: https://issues.apache.org/jira/browse/YARN-1539
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Vinod Kumar Vavilapalli
Priority: Critical


Today, Queue admin ACLs are similar to submit-acls w.r.t hierarchy in that if 
one has to be able to administer a queue, he/she should be an admin of all the 
queues in the ancestry - an unnecessary burden.

This was added in YARN-899 and I believe is wrong semantics as well as 
implementation.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1539) Queue admin ACLs should NOT be similar to submit-acls w.r.t hierarchy.

2013-12-26 Thread Sandy Ryza (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1539?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857050#comment-13857050
 ] 

Sandy Ryza commented on YARN-1539:
--

My understanding was that the way that both submit and admin work is that: to 
have access to a queue, you need to be in its access control list OR in the 
access control list of any of its ancestors.  Where is the unnecessary burden?  
Also, would this not be an incompatible change?

{code}
  @Override
  public boolean hasAccess(QueueACL acl, UserGroupInformation user) {
synchronized (this) {
  if (acls.get(acl).isUserAllowed(user)) {
return true;
  }
}

if (parent != null) {
  return parent.hasAccess(acl, user);
}

return false;
  }
{code}

 Queue admin ACLs should NOT be similar to submit-acls w.r.t hierarchy.
 --

 Key: YARN-1539
 URL: https://issues.apache.org/jira/browse/YARN-1539
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Vinod Kumar Vavilapalli
Priority: Critical

 Today, Queue admin ACLs are similar to submit-acls w.r.t hierarchy in that if 
 one has to be able to administer a queue, he/she should be an admin of all 
 the queues in the ancestry - an unnecessary burden.
 This was added in YARN-899 and I believe is wrong semantics as well as 
 implementation.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1539) Queue admin ACLs should NOT be similar to submit-acls w.r.t hierarchy.

2013-12-26 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1539?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857064#comment-13857064
 ] 

Vinod Kumar Vavilapalli commented on YARN-1539:
---

bq. you need to be in its access control list OR in the access control list of 
any of its ancestors.
It is AND not OR in Capacity-Scheduler. And that makes sense for submit-acls. 
You need to have permissions to submit to a parent-queue, child-queue, leaf 
queue etc. But not for admin ACLs.

Seems like that is different for FairScheduler. All the more we should fix this 
to be consistent. FairScheduler should also change w.r.t submit-acls.

Yes, this would be an incompatible change, but don't know of a solution 
otherwise. I found this a while back 
([here|https://issues.apache.org/jira/browse/YARN-899?focusedCommentId=13781287page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13781287])
 but failed to file the ticket and fix it in time.

 Queue admin ACLs should NOT be similar to submit-acls w.r.t hierarchy.
 --

 Key: YARN-1539
 URL: https://issues.apache.org/jira/browse/YARN-1539
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Vinod Kumar Vavilapalli
Priority: Critical

 Today, Queue admin ACLs are similar to submit-acls w.r.t hierarchy in that if 
 one has to be able to administer a queue, he/she should be an admin of all 
 the queues in the ancestry - an unnecessary burden.
 This was added in YARN-899 and I believe is wrong semantics as well as 
 implementation.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (YARN-1539) Queue admin ACLs should NOT be similar to submit-acls w.r.t hierarchy.

2013-12-26 Thread Vinod Kumar Vavilapalli (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1539?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinod Kumar Vavilapalli updated YARN-1539:
--

Issue Type: Sub-task  (was: Bug)
Parent: YARN-47

 Queue admin ACLs should NOT be similar to submit-acls w.r.t hierarchy.
 --

 Key: YARN-1539
 URL: https://issues.apache.org/jira/browse/YARN-1539
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Vinod Kumar Vavilapalli
Priority: Critical

 Today, Queue admin ACLs are similar to submit-acls w.r.t hierarchy in that if 
 one has to be able to administer a queue, he/she should be an admin of all 
 the queues in the ancestry - an unnecessary burden.
 This was added in YARN-899 and I believe is wrong semantics as well as 
 implementation.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (YARN-1538) Support web-services for application-submission and AM protocols

2013-12-26 Thread Vinod Kumar Vavilapalli (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1538?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinod Kumar Vavilapalli updated YARN-1538:
--

Description: We already have read-only web-services for YARN - 
MAPREDUCE-2863. It'll be great to support APIs for application submission and 
the AM protocols - scheduling and container-management - as alternatives to 
RPCs.  (was: We already have read-only web-services for YARN - MAPREDUCE-2863. 
It'll be great to support APIs for application submission and the scheduling 
protocol - as alternatives to RPCs.)

Yes I meant all of them by saying AM protocols. Fixing description.

 Support web-services for application-submission and AM protocols
 

 Key: YARN-1538
 URL: https://issues.apache.org/jira/browse/YARN-1538
 Project: Hadoop YARN
  Issue Type: New Feature
Reporter: Vinod Kumar Vavilapalli

 We already have read-only web-services for YARN - MAPREDUCE-2863. It'll be 
 great to support APIs for application submission and the AM protocols - 
 scheduling and container-management - as alternatives to RPCs.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1523) Replace RMNotYetReadyException with the standard hadoop io StandbyException

2013-12-26 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857071#comment-13857071
 ] 

Karthik Kambatla commented on YARN-1523:


bq. Should we make the throwStandbyException() a method of HAUtil so that we 
could reuse it in other places if need be?
Like the idea. Let us do it when we have another place that throws 
StandbyException. 

bq. When we want the standby RM to re-direct the client to the active RM then 
how do you think we would do that?
The two options appear to be (1) notifying the client (through a derivate of 
StandbyException) that we are not Active and that the Active is so-and-so (2) 
relaying the client request to the Active RM and responding to the Client. The 
first approach might be simpler and have fewer round-trips overall, 
particularly if there are only few kinds of clients.

Will commit this later today. 


 Replace RMNotYetReadyException with the standard hadoop io StandbyException
 ---

 Key: YARN-1523
 URL: https://issues.apache.org/jira/browse/YARN-1523
 Project: Hadoop YARN
  Issue Type: Sub-task
Affects Versions: 2.4.0
Reporter: Bikas Saha
Assignee: Karthik Kambatla
 Attachments: yarn-1523-1.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (YARN-1539) Queue admin ACLs should NOT be similar to submit-acls w.r.t hierarchy.

2013-12-26 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1539?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated YARN-1539:
---

Hadoop Flags: Incompatible change

 Queue admin ACLs should NOT be similar to submit-acls w.r.t hierarchy.
 --

 Key: YARN-1539
 URL: https://issues.apache.org/jira/browse/YARN-1539
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Vinod Kumar Vavilapalli
Priority: Critical

 Today, Queue admin ACLs are similar to submit-acls w.r.t hierarchy in that if 
 one has to be able to administer a queue, he/she should be an admin of all 
 the queues in the ancestry - an unnecessary burden.
 This was added in YARN-899 and I believe is wrong semantics as well as 
 implementation.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1539) Queue admin ACLs should NOT be similar to submit-acls w.r.t hierarchy.

2013-12-26 Thread Sandy Ryza (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1539?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857074#comment-13857074
 ] 

Sandy Ryza commented on YARN-1539:
--

The code I posted above is from the Capacity Scheduler.  It looks like OR to 
me.  The Fair Scheduler has slightly different code, but we took the semantics 
from this.  Am I missing something?

 Queue admin ACLs should NOT be similar to submit-acls w.r.t hierarchy.
 --

 Key: YARN-1539
 URL: https://issues.apache.org/jira/browse/YARN-1539
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Vinod Kumar Vavilapalli
Priority: Critical

 Today, Queue admin ACLs are similar to submit-acls w.r.t hierarchy in that if 
 one has to be able to administer a queue, he/she should be an admin of all 
 the queues in the ancestry - an unnecessary burden.
 This was added in YARN-899 and I believe is wrong semantics as well as 
 implementation.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (YARN-1540) Add an easy way to turn on HA

2013-12-26 Thread Karthik Kambatla (JIRA)
Karthik Kambatla created YARN-1540:
--

 Summary: Add an easy way to turn on HA
 Key: YARN-1540
 URL: https://issues.apache.org/jira/browse/YARN-1540
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: resourcemanager
Affects Versions: 2.4.0
Reporter: Karthik Kambatla


Users will have to modify the configuration significantly to turn on HA. It 
would be nice to have a simpler way of doing this.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1540) Add an easy way to turn on HA

2013-12-26 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1540?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857089#comment-13857089
 ] 

Karthik Kambatla commented on YARN-1540:


The bare minimum configs required to be specified should be: 
# Enable HA with default options
# ZK address for the embedded elector and store
# Hostnames for the two RMs 

A utility method, to be invoked before anything else, could set all other 
configs based on these. 

 Add an easy way to turn on HA
 -

 Key: YARN-1540
 URL: https://issues.apache.org/jira/browse/YARN-1540
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: resourcemanager
Affects Versions: 2.4.0
Reporter: Karthik Kambatla

 Users will have to modify the configuration significantly to turn on HA. It 
 would be nice to have a simpler way of doing this.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1527) yarn rmadmin command prints wrong usage info:

2013-12-26 Thread Jian He (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857096#comment-13857096
 ] 

Jian He commented on YARN-1527:
---

patch looks good, +1. 

 yarn rmadmin command prints wrong usage info:
 -

 Key: YARN-1527
 URL: https://issues.apache.org/jira/browse/YARN-1527
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.2.0
Reporter: Jian He
Assignee: Akira AJISAKA
  Labels: newbie
 Attachments: YARN-1527.patch


 The usage should be: yarn rmadmin, instead of java RMAdmin, and the 
 -refreshQueues should be in the second line.
 {code} Usage: java RMAdmin   -refreshQueues 
-refreshNodes 
-refreshSuperUserGroupsConfiguration 
-refreshUserToGroupsMappings 
-refreshAdminAcls 
-refreshServiceAcl 
-getGroups [username]
-help [cmd]
-transitionToActive serviceId
-transitionToStandby serviceId
-failover [--forcefence] [--forceactive] serviceId serviceId
-getServiceState serviceId
-checkHealth serviceId
 {code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1527) yarn rmadmin command prints wrong usage info:

2013-12-26 Thread Jian He (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857099#comment-13857099
 ] 

Jian He commented on YARN-1527:
---

Compiled and run the command myself, now the usage looks as the following:
{code}
Usage: yarn rmadmin
   -refreshQueues 
   -refreshNodes 
   -refreshSuperUserGroupsConfiguration 
   -refreshUserToGroupsMappings 
   -refreshAdminAcls 
   -refreshServiceAcl 
   -getGroups [username]
   -help [cmd]
   -transitionToActive serviceId
   -transitionToStandby serviceId
   -failover [--forcefence] [--forceactive] serviceId serviceId
   -getServiceState serviceId
   -checkHealth serviceId
{code}

 yarn rmadmin command prints wrong usage info:
 -

 Key: YARN-1527
 URL: https://issues.apache.org/jira/browse/YARN-1527
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.2.0
Reporter: Jian He
Assignee: Akira AJISAKA
  Labels: newbie
 Attachments: YARN-1527.patch


 The usage should be: yarn rmadmin, instead of java RMAdmin, and the 
 -refreshQueues should be in the second line.
 {code} Usage: java RMAdmin   -refreshQueues 
-refreshNodes 
-refreshSuperUserGroupsConfiguration 
-refreshUserToGroupsMappings 
-refreshAdminAcls 
-refreshServiceAcl 
-getGroups [username]
-help [cmd]
-transitionToActive serviceId
-transitionToStandby serviceId
-failover [--forcefence] [--forceactive] serviceId serviceId
-getServiceState serviceId
-checkHealth serviceId
 {code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1527) yarn rmadmin command prints wrong usage info:

2013-12-26 Thread Jian He (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857126#comment-13857126
 ] 

Jian He commented on YARN-1527:
---

Committed to trunk and branch-2,  thanks [~ajisakaa] !

 yarn rmadmin command prints wrong usage info:
 -

 Key: YARN-1527
 URL: https://issues.apache.org/jira/browse/YARN-1527
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.2.0
Reporter: Jian He
Assignee: Akira AJISAKA
  Labels: newbie
 Fix For: 2.4.0

 Attachments: YARN-1527.patch


 The usage should be: yarn rmadmin, instead of java RMAdmin, and the 
 -refreshQueues should be in the second line.
 {code} Usage: java RMAdmin   -refreshQueues 
-refreshNodes 
-refreshSuperUserGroupsConfiguration 
-refreshUserToGroupsMappings 
-refreshAdminAcls 
-refreshServiceAcl 
-getGroups [username]
-help [cmd]
-transitionToActive serviceId
-transitionToStandby serviceId
-failover [--forcefence] [--forceactive] serviceId serviceId
-getServiceState serviceId
-checkHealth serviceId
 {code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1527) yarn rmadmin command prints wrong usage info:

2013-12-26 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857127#comment-13857127
 ] 

Hudson commented on YARN-1527:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #4929 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4929/])
YARN-1527. Fixed yarn rmadmin command to print the correct usage info. 
Contributed by Akira AJISAKA. (jianhe: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1553596)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/main/java/org/apache/hadoop/yarn/client/cli/RMAdminCLI.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client/src/test/java/org/apache/hadoop/yarn/client/TestRMAdminCLI.java


 yarn rmadmin command prints wrong usage info:
 -

 Key: YARN-1527
 URL: https://issues.apache.org/jira/browse/YARN-1527
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.2.0
Reporter: Jian He
Assignee: Akira AJISAKA
  Labels: newbie
 Fix For: 2.4.0

 Attachments: YARN-1527.patch


 The usage should be: yarn rmadmin, instead of java RMAdmin, and the 
 -refreshQueues should be in the second line.
 {code} Usage: java RMAdmin   -refreshQueues 
-refreshNodes 
-refreshSuperUserGroupsConfiguration 
-refreshUserToGroupsMappings 
-refreshAdminAcls 
-refreshServiceAcl 
-getGroups [username]
-help [cmd]
-transitionToActive serviceId
-transitionToStandby serviceId
-failover [--forcefence] [--forceactive] serviceId serviceId
-getServiceState serviceId
-checkHealth serviceId
 {code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (YARN-1541) Invalidate AM Host/Port when app attempt is done so that in the mean-while client doesn’t get wrong information.

2013-12-26 Thread Jian He (JIRA)
Jian He created YARN-1541:
-

 Summary: Invalidate AM Host/Port when app attempt is done so that 
in the mean-while client doesn’t get wrong information.
 Key: YARN-1541
 URL: https://issues.apache.org/jira/browse/YARN-1541
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Jian He
Assignee: Jian He






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (YARN-1541) Invalidate AM Host/Port when app attempt is done so that in the mean-while client doesn’t get wrong information.

2013-12-26 Thread Jian He (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1541?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jian He updated YARN-1541:
--

Attachment: YARN-1541.1.patch

New patch invalidates the AM's host and roc port when the AM attempt is done

 Invalidate AM Host/Port when app attempt is done so that in the mean-while 
 client doesn’t get wrong information.
 

 Key: YARN-1541
 URL: https://issues.apache.org/jira/browse/YARN-1541
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Jian He
Assignee: Jian He
 Attachments: YARN-1541.1.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1541) Invalidate AM Host/Port when app attempt is done so that in the mean-while client doesn’t get wrong information.

2013-12-26 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1541?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857171#comment-13857171
 ] 

Hadoop QA commented on YARN-1541:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12620560/YARN-1541.1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/2730//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/2730//console

This message is automatically generated.

 Invalidate AM Host/Port when app attempt is done so that in the mean-while 
 client doesn’t get wrong information.
 

 Key: YARN-1541
 URL: https://issues.apache.org/jira/browse/YARN-1541
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Jian He
Assignee: Jian He
 Attachments: YARN-1541.1.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (YARN-1461) RM API and RM changes to handle tags for running jobs

2013-12-26 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1461?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated YARN-1461:
---

Attachment: yarn-1461-1.patch

Patch that adds tags.

 RM API and RM changes to handle tags for running jobs
 -

 Key: YARN-1461
 URL: https://issues.apache.org/jira/browse/YARN-1461
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: resourcemanager
Affects Versions: 2.2.0
Reporter: Karthik Kambatla
Assignee: Karthik Kambatla
 Attachments: yarn-1461-1.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1413) [YARN-321] AHS WebUI should server aggregated logs as well

2013-12-26 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1413?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857194#comment-13857194
 ] 

Vinod Kumar Vavilapalli commented on YARN-1413:
---

There's a minor merge conflict. Can you please update? Tx.

Also, there seem to be issues with container-logs URL on AHS redirect which in 
turn will redirect to the node once the log-aggregation finishes - causing an 
infinite redirect cycle. Are you not seeing this? Is this tested with 
log-aggregation on?

 [YARN-321] AHS WebUI should server aggregated logs as well
 --

 Key: YARN-1413
 URL: https://issues.apache.org/jira/browse/YARN-1413
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Zhijie Shen
Assignee: Mayank Bansal
 Attachments: YARN-1413-1.patch, YARN-1413-2.patch, YARN-1413-3.patch, 
 YARN-1413-4.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (YARN-1542) Add unit test for public resource on viewfs

2013-12-26 Thread Gera Shegalov (JIRA)
Gera Shegalov created YARN-1542:
---

 Summary: Add unit test for public resource on viewfs
 Key: YARN-1542
 URL: https://issues.apache.org/jira/browse/YARN-1542
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: nodemanager
Reporter: Gera Shegalov






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1461) RM API and RM changes to handle tags for running jobs

2013-12-26 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1461?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857219#comment-13857219
 ] 

Hadoop QA commented on YARN-1461:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12620567/yarn-1461-1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 5 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:red}-1 findbugs{color}.  The patch appears to introduce 2 new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager:

  
org.apache.hadoop.yarn.server.resourcemanager.TestClientRMService

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/2731//testReport/
Findbugs warnings: 
https://builds.apache.org/job/PreCommit-YARN-Build/2731//artifact/trunk/patchprocess/newPatchFindbugsWarningshadoop-yarn-common.html
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/2731//console

This message is automatically generated.

 RM API and RM changes to handle tags for running jobs
 -

 Key: YARN-1461
 URL: https://issues.apache.org/jira/browse/YARN-1461
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: resourcemanager
Affects Versions: 2.2.0
Reporter: Karthik Kambatla
Assignee: Karthik Kambatla
 Attachments: yarn-1461-1.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (YARN-1461) RM API and RM changes to handle tags for running jobs

2013-12-26 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1461?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated YARN-1461:
---

Attachment: yarn-1461-2.patch

Patch to fix the findbugs warnings. TestClientRMService passes locally. Will 
look deeper if it fails again. 

 RM API and RM changes to handle tags for running jobs
 -

 Key: YARN-1461
 URL: https://issues.apache.org/jira/browse/YARN-1461
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: resourcemanager
Affects Versions: 2.2.0
Reporter: Karthik Kambatla
Assignee: Karthik Kambatla
 Attachments: yarn-1461-1.patch, yarn-1461-2.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (YARN-1542) Add unit test for public resource on viewfs

2013-12-26 Thread Gera Shegalov (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1542?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gera Shegalov updated YARN-1542:


Attachment: YARN-1542.v01.patch

 Add unit test for public resource on viewfs
 ---

 Key: YARN-1542
 URL: https://issues.apache.org/jira/browse/YARN-1542
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: nodemanager
Reporter: Gera Shegalov
 Attachments: YARN-1542.v01.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1461) RM API and RM changes to handle tags for running jobs

2013-12-26 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1461?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857241#comment-13857241
 ] 

Hadoop QA commented on YARN-1461:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12620574/yarn-1461-2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 5 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/2732//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/2732//console

This message is automatically generated.

 RM API and RM changes to handle tags for running jobs
 -

 Key: YARN-1461
 URL: https://issues.apache.org/jira/browse/YARN-1461
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: resourcemanager
Affects Versions: 2.2.0
Reporter: Karthik Kambatla
Assignee: Karthik Kambatla
 Attachments: yarn-1461-1.patch, yarn-1461-2.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1542) Add unit test for public resource on viewfs

2013-12-26 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857243#comment-13857243
 ] 

Hadoop QA commented on YARN-1542:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12620577/YARN-1542.v01.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common:

  org.apache.hadoop.yarn.util.TestFSDownload

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/2733//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/2733//console

This message is automatically generated.

 Add unit test for public resource on viewfs
 ---

 Key: YARN-1542
 URL: https://issues.apache.org/jira/browse/YARN-1542
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: nodemanager
Reporter: Gera Shegalov
Assignee: Gera Shegalov
 Attachments: YARN-1542.v01.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1542) Add unit test for public resource on viewfs

2013-12-26 Thread Gera Shegalov (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1542?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857255#comment-13857255
 ] 

Gera Shegalov commented on YARN-1542:
-

The failed unit test is being added by this JIRA to demonstrate the problem. it 
would be fixed by HADOOP-10191

 Add unit test for public resource on viewfs
 ---

 Key: YARN-1542
 URL: https://issues.apache.org/jira/browse/YARN-1542
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: nodemanager
Reporter: Gera Shegalov
Assignee: Gera Shegalov
 Attachments: YARN-1542.v01.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (YARN-1523) Use StandbyException instead of RMNotYetReadyException

2013-12-26 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1523?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated YARN-1523:
---

Summary: Use StandbyException instead of RMNotYetReadyException  (was: 
Replace RMNotYetReadyException with the standard hadoop io StandbyException)

 Use StandbyException instead of RMNotYetReadyException
 --

 Key: YARN-1523
 URL: https://issues.apache.org/jira/browse/YARN-1523
 Project: Hadoop YARN
  Issue Type: Sub-task
Affects Versions: 2.4.0
Reporter: Bikas Saha
Assignee: Karthik Kambatla
 Attachments: yarn-1523-1.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1541) Invalidate AM Host/Port when app attempt is done so that in the mean-while client doesn’t get wrong information.

2013-12-26 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1541?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857283#comment-13857283
 ] 

Vinod Kumar Vavilapalli commented on YARN-1541:
---

Can you add a test in TestRM? Validating transitions directly is not so useful.

 Invalidate AM Host/Port when app attempt is done so that in the mean-while 
 client doesn’t get wrong information.
 

 Key: YARN-1541
 URL: https://issues.apache.org/jira/browse/YARN-1541
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Jian He
Assignee: Jian He
 Attachments: YARN-1541.1.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1523) Use StandbyException instead of RMNotYetReadyException

2013-12-26 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857287#comment-13857287
 ] 

Hudson commented on YARN-1523:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #4930 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/4930/])
YARN-1523. Use StandbyException instead of RMNotYetReadyException (kasha) 
(kasha: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1553616)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/exceptions/RMNotYetActiveException.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/server/api/ResourceManagerAdministrationProtocol.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/AdminService.java


 Use StandbyException instead of RMNotYetReadyException
 --

 Key: YARN-1523
 URL: https://issues.apache.org/jira/browse/YARN-1523
 Project: Hadoop YARN
  Issue Type: Sub-task
Affects Versions: 2.4.0
Reporter: Bikas Saha
Assignee: Karthik Kambatla
 Fix For: 2.4.0

 Attachments: yarn-1523-1.patch






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1521) Mark appropriate protocol methods with the idempotent annotation

2013-12-26 Thread Xuan Gong (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1521?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857315#comment-13857315
 ] 

Xuan Gong commented on YARN-1521:
-

an idempotent operation is one that has no additional effect if it is called 
more than once with the same input parameters. That means no matter how many 
times you call the operation with the same parameter, the result will be the 
same. In our HA case, i think that it means no matter which RM is active, by 
calling the operations, we will get the same result. 

Based on this, here is my proposal:
{code}
In ApplicationClientProtocol : 
{code}
a. getNewApplication(), submitApplication(), forceKillApplication(), 
getApplicationReport(), 
getApplications(),getQueueUserAclInfo(),getDelegationToken(),renewDelegationToken(),
 cancelDelegationToken() need to add idempotent annotation.

When RM becomes Active, it will read back all the Applications, and all related 
DelegationTokens back from the zk, in that case, no matter which RM is active, 
it will get all the information for all the applications, either running or 
finished. That is the reason why we can always get the same results if we call 
getApplicationReport(), 
getApplications(),getDelegationToken(),renewDelegationToken(), 
cancelDelegationToken(). So, the idempotent annotation need to be added for 
those operations.

For forceKillApplication(), getNewApplication(), submitApplication(), those 
operation need to add idempotent annotation too. Because, when the user calls 
them, no matter which RM is active, the application will be submitted or killed.

For getQueueUserAclInfo(), YARN-1459 will guarantee that other RMs will get the 
latest configuration if one of RMs are changed its configuration. In this case, 
when getQueueUserAclInfo is called, we will always get the same results no 
matter which RM is active.


b. getClusterMetrics(), getClusterNodes(),getQueueInfo() do not need to add.

I do not want to add idempotent annotation. Because those are based on the 
Nodes information or queue information. The current active RM has the resource 
information about all the NMs, and queue info for all the applications. But 
when failover happens, the next active RMs will not have all the NMs 
information immediately, and it does not have any queue info (such as which 
application in which queue). 


{code}
In ResourceTracker:
{code}
a.registerNodeManager needs to add idempotent annotation.
Same reason why I want to add idempotent annotation for submitApplication()

b. nodeHeartbeat does not need to add

The new active RM does not know all the NMs immediately. 


{code}
In ApplicationMasterProtocol
{code}
Does not need to add for all the methods in ApplicationMasterProtocol.
There are three operations: registerApplicationMaster, finishApplicationMaster, 
allocate(). When the failover happens, a new RM will start a new Attempt for 
the running AM (just like the RM restart). 


{code}
In ResourceManagerAdministrationProtocol
{code}
a.refreshQueues(), refreshNodes(), 
refreshSuperUserGroupsConfiguration(),RefreshUserToGroupsMappingsRequest(), 
refreshAdminAcls(), refreshServiceAcls() need to add idempotent annotation.

YARN-1459 will guarantee that other RMs will get the latest configuration if 
one of RMs are changed its configuration.

b.updateNodeResource() do not need to add
The new active RM does not know all the NMs immediately. 

 Mark appropriate protocol methods with the idempotent annotation
 

 Key: YARN-1521
 URL: https://issues.apache.org/jira/browse/YARN-1521
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Xuan Gong
Assignee: Xuan Gong

 After YARN-1028, we add the automatically failover into RMProxy. This JIRA is 
 to identify whether we need to add idempotent annotation and which methods 
 can be marked as idempotent.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1521) Mark appropriate protocol methods with the idempotent annotation

2013-12-26 Thread Xuan Gong (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1521?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857318#comment-13857318
 ] 

Xuan Gong commented on YARN-1521:
-

[~bikassaha], [~vinodkv], [~kkambatl]

Comments, please

 Mark appropriate protocol methods with the idempotent annotation
 

 Key: YARN-1521
 URL: https://issues.apache.org/jira/browse/YARN-1521
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Xuan Gong
Assignee: Xuan Gong

 After YARN-1028, we add the automatically failover into RMProxy. This JIRA is 
 to identify whether we need to add idempotent annotation and which methods 
 can be marked as idempotent.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (YARN-1543) Stateless multiple resource manager

2013-12-26 Thread Tsuyoshi OZAWA (JIRA)
Tsuyoshi OZAWA created YARN-1543:


 Summary: Stateless multiple resource manager
 Key: YARN-1543
 URL: https://issues.apache.org/jira/browse/YARN-1543
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Tsuyoshi OZAWA


Currently, YARN is implemented based on statefull single resource manager 
architecture. This approach can raise operation costs compare to stateless 
multi resource manager architecture, because we need to care which resource 
manager is active to ensure availability. 

We can avoid this problem by saving all states into updatable and transactional 
RMStateStore such as ZooKeeper or RDBMS. If we take this approach, multiple 
resource managers can be launched at the same time, we can solve this problem. 
Operation cost can be reduced, because we don't need to maintain 
specific(current master) resource manager.

This idea is based on a paper: “Omega: flexible, scalable schedulers for large 
compute clusters”[1].

[1] http://eurosys2013.tudos.org/wp-content/uploads/2013/paper/Schwarzkopf.pdf




--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1521) Mark appropriate protocol methods with the idempotent annotation

2013-12-26 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1521?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857328#comment-13857328
 ] 

Karthik Kambatla commented on YARN-1521:


bq. an idempotent operation is one that has no additional effect if it is 
called more than once with the same input parameters. That means no matter how 
many times you call the operation with the same parameter, the result will be 
the same. In our HA case, i think that it means no matter which RM is active, 
by calling the operations, we will get the same result.
I am not sure if this is completely accurate. 

Rephrasing the definition from wiki - an idempotent operation can be applied 
multiple times without changing the result beyond the initial application. The 
relevance to that in the context of HA and client failover, I believe, has to 
do with whether an operation can be retried. For instance, submitting an 
application is not idempotent. 

Suppose a client is in the process of submitting an application to the Active 
(say, RM1) and RM1 goes down. RM2 takes over. Now, we don't know the status of 
the app submission and would be wrong to retry submitting to RM2. In this case, 
the client shouldn't failover to RM2. 

Unfortunately, there might be retries in MR and YARN without checking for 
idempotence or status of the previous attempt. I haven't run into any issues 
yet, but I think it would be nice to handle all of them cleanly. 

 Mark appropriate protocol methods with the idempotent annotation
 

 Key: YARN-1521
 URL: https://issues.apache.org/jira/browse/YARN-1521
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Xuan Gong
Assignee: Xuan Gong

 After YARN-1028, we add the automatically failover into RMProxy. This JIRA is 
 to identify whether we need to add idempotent annotation and which methods 
 can be marked as idempotent.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (YARN-1543) Stateless multiple resource manager

2013-12-26 Thread Tsuyoshi OZAWA (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1543?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsuyoshi OZAWA updated YARN-1543:
-

Description: 
Currently, YARN is implemented based on statefull single resource manager 
architecture. This approach can raise operation costs compare to stateless 
multi resource manager architecture, because we need to care which resource 
manager is active to ensure availability. 

We can avoid this problem by saving all states into updatable and transactional 
RMStateStore such as ZooKeeper or RDBMS. If we take this approach, multiple 
resource managers can be launched at the same time. As a result, operation cost 
can be reduced, because we don't need to maintain specific(current master) 
resource manager.

This idea is based on a paper: “Omega: flexible, scalable schedulers for large 
compute clusters”[1].

[1] http://eurosys2013.tudos.org/wp-content/uploads/2013/paper/Schwarzkopf.pdf


  was:
Currently, YARN is implemented based on statefull single resource manager 
architecture. This approach can raise operation costs compare to stateless 
multi resource manager architecture, because we need to care which resource 
manager is active to ensure availability. 

We can avoid this problem by saving all states into updatable and transactional 
RMStateStore such as ZooKeeper or RDBMS. If we take this approach, multiple 
resource managers can be launched at the same time, we can solve this problem. 
Operation cost can be reduced, because we don't need to maintain 
specific(current master) resource manager.

This idea is based on a paper: “Omega: flexible, scalable schedulers for large 
compute clusters”[1].

[1] http://eurosys2013.tudos.org/wp-content/uploads/2013/paper/Schwarzkopf.pdf



 Stateless multiple resource manager
 ---

 Key: YARN-1543
 URL: https://issues.apache.org/jira/browse/YARN-1543
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Tsuyoshi OZAWA

 Currently, YARN is implemented based on statefull single resource manager 
 architecture. This approach can raise operation costs compare to stateless 
 multi resource manager architecture, because we need to care which resource 
 manager is active to ensure availability. 
 We can avoid this problem by saving all states into updatable and 
 transactional RMStateStore such as ZooKeeper or RDBMS. If we take this 
 approach, multiple resource managers can be launched at the same time. As a 
 result, operation cost can be reduced, because we don't need to maintain 
 specific(current master) resource manager.
 This idea is based on a paper: “Omega: flexible, scalable schedulers for large 
 compute clusters”[1].
 [1] http://eurosys2013.tudos.org/wp-content/uploads/2013/paper/Schwarzkopf.pdf



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (YARN-1463) TestContainerManagerSecurity#testContainerManager fails

2013-12-26 Thread Vinod Kumar Vavilapalli (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1463?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinod Kumar Vavilapalli updated YARN-1463:
--

Attachment: YARN-1463-20131226.txt

Taking this over as it continues to break builds.

Here's a patch, that fixes the fundamental issue, and avoids starting 
http-server where possible or creates spnego keytab/princs where needed.

 TestContainerManagerSecurity#testContainerManager fails
 ---

 Key: YARN-1463
 URL: https://issues.apache.org/jira/browse/YARN-1463
 Project: Hadoop YARN
  Issue Type: Test
Affects Versions: 2.4.0
Reporter: Ted Yu
Assignee: Binglin Chang
 Attachments: YARN-1463-20131226.txt, YARN-1463.v1.patch, 
 YARN-1463.v2.patch


 Here is stack trace:
 {code}
 testContainerManager[1](org.apache.hadoop.yarn.server.TestContainerManagerSecurity)
   Time elapsed: 1.756 sec   ERROR!
 org.apache.hadoop.yarn.exceptions.YarnRuntimeException: java.io.IOException: 
 ResourceManager failed to start. Final state is STOPPED
   at 
 org.apache.hadoop.yarn.server.MiniYARNCluster$ResourceManagerWrapper.serviceStart(MiniYARNCluster.java:253)
   at org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
   at 
 org.apache.hadoop.service.CompositeService.serviceStart(CompositeService.java:121)
   at org.apache.hadoop.service.AbstractService.start(AbstractService.java:193)
   at 
 org.apache.hadoop.yarn.server.TestContainerManagerSecurity.testContainerManager(TestContainerManagerSecurity.java:110)
 {code}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1482) WebApplicationProxy should be always-on w.r.t HA even if it is embedded in the RM

2013-12-26 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1482?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857351#comment-13857351
 ] 

Vinod Kumar Vavilapalli commented on YARN-1482:
---

bq. try to try
I meant try to test.

 WebApplicationProxy should be always-on w.r.t HA even if it is embedded in 
 the RM
 -

 Key: YARN-1482
 URL: https://issues.apache.org/jira/browse/YARN-1482
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Vinod Kumar Vavilapalli
Assignee: Xuan Gong
 Attachments: YARN-1482.1.patch, YARN-1482.2.patch


 This way, even if an RM goes to standby mode, we can affect a redirect to the 
 active. And more importantly, users will not suddenly see all their links 
 stop working.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (YARN-1544) Add Tools menu to web interface of ApplicationHistoryServer

2013-12-26 Thread Shinichi Yamashita (JIRA)
Shinichi Yamashita created YARN-1544:


 Summary: Add Tools menu to web interface of 
ApplicationHistoryServer
 Key: YARN-1544
 URL: https://issues.apache.org/jira/browse/YARN-1544
 Project: Hadoop YARN
  Issue Type: Improvement
Affects Versions: YARN-321
Reporter: Shinichi Yamashita
Priority: Minor


It should add the link of the Tools menu like ResourceManager and NodeManager 
in web interface of ApplicationHistoryServer.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1544) Add Tools menu to web interface of ApplicationHistoryServer

2013-12-26 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1544?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857354#comment-13857354
 ] 

Hadoop QA commented on YARN-1544:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12620593/sample-webui.png
  against trunk revision .

{color:red}-1 patch{color}.  The patch command could not apply the patch.

Console output: https://builds.apache.org/job/PreCommit-YARN-Build/2735//console

This message is automatically generated.

 Add Tools menu to web interface of ApplicationHistoryServer
 ---

 Key: YARN-1544
 URL: https://issues.apache.org/jira/browse/YARN-1544
 Project: Hadoop YARN
  Issue Type: Improvement
Affects Versions: YARN-321
Reporter: Shinichi Yamashita
Priority: Minor
 Attachments: YARN-1544.patch, sample-webui.png


 It should add the link of the Tools menu like ResourceManager and 
 NodeManager in web interface of ApplicationHistoryServer.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1521) Mark appropriate protocol methods with the idempotent annotation

2013-12-26 Thread Bikas Saha (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1521?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857358#comment-13857358
 ] 

Bikas Saha commented on YARN-1521:
--

After the first application of an idempotent operation, the subsequent 
application of the operation should not change the state of the system in any 
substantial manner. So its not clear to me how submitApplication() is 
idempotent. If the first operation succeeds then the second must fail. 
getNewApplication() will create a new application id on the RM for every 
invocation. It may be harmless to retry but its probably not idempotent. 
renewDelegationToken() - again harmless to retry but probably not idempotent. 
killApplication() is idempotent. getApplicationReport() is idempotent. 
getDelegationToken() - probably not since it results in creation of a new 
delegation token.
Then there is the point of which operations are transferable via retries across 
RM failover. 

 Mark appropriate protocol methods with the idempotent annotation
 

 Key: YARN-1521
 URL: https://issues.apache.org/jira/browse/YARN-1521
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Xuan Gong
Assignee: Xuan Gong

 After YARN-1028, we add the automatically failover into RMProxy. This JIRA is 
 to identify whether we need to add idempotent annotation and which methods 
 can be marked as idempotent.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Commented] (YARN-1029) Allow embedding leader election into the RM

2013-12-26 Thread Bikas Saha (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1029?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13857368#comment-13857368
 ] 

Bikas Saha commented on YARN-1029:
--

bq. The method should have been synchronized in YARN-1481, Vinod Kumar 
Vavilapalli and I thought we could handle it here. Could do an addendum patch 
there instead if that is preferred.
Please take care of it wherever appropriate.

bq. Not sure how advantageous that will be - we ll end up calling the common 
method instead of ExitUtil.terminate only for the common method to call it? 
Also, getCause() doesn't exist in AbstractEvent requiring us to add a new kind 
of event (CausedEvent?) that both these events extend. Seems too complicated 
for the gain.
The advantage is that future changes will need to edit one place in the code 
and forget about other places. Not a major point though. Making life easier for 
future devs.

bq. Agree - 100%, but would like to do it lazily when another such case pops up.
Again, if we organize the newly added code such that its a common event for any 
module to inform the RM about a fatal error then we are good for the future. 
Embedded elector can use that event instead of a custom named event.

I am sorry I could not understand your comment explaining how the test passes 
with these timeouts. Been sick for a while. Probably my brain is running slow :P




 Allow embedding leader election into the RM
 ---

 Key: YARN-1029
 URL: https://issues.apache.org/jira/browse/YARN-1029
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Bikas Saha
Assignee: Karthik Kambatla
 Attachments: embedded-zkfc-approach.patch, yarn-1029-0.patch, 
 yarn-1029-0.patch, yarn-1029-1.patch, yarn-1029-2.patch, yarn-1029-3.patch, 
 yarn-1029-4.patch, yarn-1029-approach.patch


 It should be possible to embed common ActiveStandyElector into the RM such 
 that ZooKeeper based leader election and notification is in-built. In 
 conjunction with a ZK state store, this configuration will be a simple 
 deployment option.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)