[jira] [Assigned] (YARN-50) Implement renewal / cancellation of Delegation Tokens

2012-12-20 Thread Siddharth Seth (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-50?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Seth reassigned YARN-50:
--

Assignee: Siddharth Seth

 Implement renewal / cancellation of Delegation Tokens
 -

 Key: YARN-50
 URL: https://issues.apache.org/jira/browse/YARN-50
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Siddharth Seth
Assignee: Siddharth Seth

 Currently, delegation tokens issues by the RM and History server cannot be 
 renewed or cancelled. This needs to be implemented.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-50) Implement renewal / cancellation of Delegation Tokens

2012-12-20 Thread Siddharth Seth (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-50?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Seth updated YARN-50:
---

Attachment: YARN-50_wip.txt

Work in progress. Some of the classes can be moved to different packages. Needs 
to be tested.

One flaw with renewing tokens in their current form in the RM, is that the RM 
ends up depending on the token Renewers to be in the classpath.
HDFS token renewal already needs the HDFS renewer (DFSClient.Renewer) to be in 
the RM classpath.
With this patch, MR JobHistoryServer token renewal requires MR jars in the RM 
classpath. (The patch does not add them - but the default startup scripts 
already do). It also ends up requiring HistoryServer configuration parameters 
to be part of the RM configuration.
Avoiding these dependencies doesn't seem like it'll be a small change.

 Implement renewal / cancellation of Delegation Tokens
 -

 Key: YARN-50
 URL: https://issues.apache.org/jira/browse/YARN-50
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Siddharth Seth
Assignee: Siddharth Seth
 Attachments: YARN-50_wip.txt


 Currently, delegation tokens issues by the RM and History server cannot be 
 renewed or cancelled. This needs to be implemented.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-103) Add a yarn AM - RM client module

2012-12-20 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-103?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13536908#comment-13536908
 ] 

Alejandro Abdelnur commented on YARN-103:
-

If developers (of AMs) suppose to use the interface only, then they should not 
see the implementing class. They should get an interface instance via a 
factory. The implementing class should not be in the same package or it should 
be package private. Using annotations in public API on methods to state they 
are for testing only is calling for trouble as developers will use those 
methods.

As I assume this pattern will hold true for more than one interface/impl set, 
we could have a generic factory for this:

{code}
  public T T create(ClassT klass, Configuration conf);
{code}

Internally we can use our good old {{ReflectionUtils.newInstance()}} and drive 
the implementation from a configuration property.






 Add a yarn AM - RM client module
 

 Key: YARN-103
 URL: https://issues.apache.org/jira/browse/YARN-103
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Bikas Saha
Assignee: Bikas Saha
 Attachments: YARN-103.1.patch, YARN-103.2.patch, YARN-103.3.patch, 
 YARN-103.4.patch, YARN-103.4.wrapper.patch, YARN-103.5.patch, 
 YARN-103.6.patch, YARN-103.7.patch


 Add a basic client wrapper library to the AM RM protocol in order to prevent 
 proliferation of code being duplicated everywhere. Provide helper functions 
 to perform reverse mapping of container requests to RM allocation resource 
 request table format.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-103) Add a yarn AM - RM client module

2012-12-20 Thread Tom White (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-103?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13536929#comment-13536929
 ] 

Tom White commented on YARN-103:


 The interface shouldn't really be implemented by anyone outside of YARN

This is the heart of the problem. We don't have a way to say (via the audience 
annotations) that an interface is for read-only use only - and not for users to 
implement. An interface may be @Public @Stable from the point of view of a user 
who wants to call it, but that doesn't mean that folks should implement it 
themselves, since for interfaces like the one we are discussing we might want 
to add a new method, say (note that such a change is compatible with @Stable). 
Adding a new method is fine for the first type of user, but not for the second, 
since their implementation breaks.

In this case, I think it's likely we'll add more methods. For example, it would 
be useful to add a waitForState method to YarnClient (which is also an 
interface), which waits for a given application to reach a particular 
YarnApplicationState. If YarnClient were a class then this would be a 
compatible change, but if it's an interface then it is not.

I think we should do one of the following:
1. Change YarnClient and AMRMClient to be concrete implementations.
2. Leave the interface/implementation distinction and make the interfaces 
@Public @Unstable.

I prefer 1. since these classes are helper classes - they are not a 
tightly-defined interface.

 Add a yarn AM - RM client module
 

 Key: YARN-103
 URL: https://issues.apache.org/jira/browse/YARN-103
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Bikas Saha
Assignee: Bikas Saha
 Attachments: YARN-103.1.patch, YARN-103.2.patch, YARN-103.3.patch, 
 YARN-103.4.patch, YARN-103.4.wrapper.patch, YARN-103.5.patch, 
 YARN-103.6.patch, YARN-103.7.patch


 Add a basic client wrapper library to the AM RM protocol in order to prevent 
 proliferation of code being duplicated everywhere. Provide helper functions 
 to perform reverse mapping of container requests to RM allocation resource 
 request table format.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-103) Add a yarn AM - RM client module

2012-12-20 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-103?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13536940#comment-13536940
 ] 

Alejandro Abdelnur commented on YARN-103:
-

If we go that route I'd also prefer #1 and any method required for testing 
should be package private.

 Add a yarn AM - RM client module
 

 Key: YARN-103
 URL: https://issues.apache.org/jira/browse/YARN-103
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Bikas Saha
Assignee: Bikas Saha
 Attachments: YARN-103.1.patch, YARN-103.2.patch, YARN-103.3.patch, 
 YARN-103.4.patch, YARN-103.4.wrapper.patch, YARN-103.5.patch, 
 YARN-103.6.patch, YARN-103.7.patch


 Add a basic client wrapper library to the AM RM protocol in order to prevent 
 proliferation of code being duplicated everywhere. Provide helper functions 
 to perform reverse mapping of container requests to RM allocation resource 
 request table format.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-223) Change processTree interface to work better with native code

2012-12-20 Thread Dennis Y (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13536945#comment-13536945
 ] 

Dennis Y commented on YARN-223:
---

This patch breaks compilation of 'Apache Hadoop Gridmix' on branch-2.
There is an import of 
org.apache.hadoop.yarn.util.ResourceCalculatorPlugin.ProcResourceValues which 
were removed by this patch.



2012-12-20 13:04:28.354 [INFO] 

2012-12-20 13:04:28.356 [INFO] Building Apache Hadoop Gridmix 2.0.3-SNAPSHOT
2012-12-20 13:04:28.357 [INFO] 

2012-12-20 13:04:28.423 [INFO] 
2012-12-20 13:04:28.425 [INFO] --- maven-clean-plugin:2.4.1:clean 
(default-clean) @ hadoop-gridmix ---
2012-12-20 13:04:28.426 [INFO] 
2012-12-20 13:04:28.428 [INFO] --- maven-antrun-plugin:1.6:run 
(create-testdirs) @ hadoop-gridmix ---
2012-12-20 13:04:28.443 [INFO] Executing tasks
2012-12-20 13:04:28.444 
2012-12-20 13:04:28.445 main:
2012-12-20 13:04:28.446 [mkdir] Created dir: 
/home/jenkins/jenkins/workspace/Hadoop-common-Integration-branch-2-gd/yhadoop-common/hadoop-tools/hadoop-gridmix/target/test-dir
2012-12-20 13:04:28.448 [INFO] Executed tasks
2012-12-20 13:04:28.450 [INFO] 
2012-12-20 13:04:28.452 [INFO] --- maven-resources-plugin:2.2:resources 
(default-resources) @ hadoop-gridmix ---
2012-12-20 13:04:28.454 [INFO] Using default encoding to copy filtered 
resources.
2012-12-20 13:04:28.455 [INFO] 
2012-12-20 13:04:28.457 [INFO] --- maven-compiler-plugin:2.5.1:compile 
(default-compile) @ hadoop-gridmix ---
2012-12-20 13:04:28.459 [INFO] Compiling 44 source files to 
/home/jenkins/jenkins/workspace/Hadoop-common-Integration-branch-2-gd/yhadoop-common/hadoop-tools/hadoop-gridmix/target/classes
2012-12-20 13:04:28.638 [INFO] 
2012-12-20 13:04:28.640 [INFO] --- maven-resources-plugin:2.2:testResources 
(default-testResources) @ hadoop-gridmix ---
2012-12-20 13:04:28.642 [INFO] Using default encoding to copy filtered 
resources.
2012-12-20 13:04:28.643 [INFO] 
2012-12-20 13:04:28.645 [INFO] --- maven-antrun-plugin:1.6:run (create-log-dir) 
@ hadoop-gridmix ---
2012-12-20 13:04:28.660 [INFO] Executing tasks
2012-12-20 13:04:28.661 
2012-12-20 13:04:28.662 main:
2012-12-20 13:04:28.663[delete] Deleting directory 
/home/jenkins/jenkins/workspace/Hadoop-common-Integration-branch-2-gd/yhadoop-common/hadoop-tools/hadoop-gridmix/target/test-dir
2012-12-20 13:04:28.665 [mkdir] Created dir: 
/home/jenkins/jenkins/workspace/Hadoop-common-Integration-branch-2-gd/yhadoop-common/hadoop-tools/hadoop-gridmix/target/test-dir
2012-12-20 13:04:28.667 [mkdir] Created dir: 
/home/jenkins/jenkins/workspace/Hadoop-common-Integration-branch-2-gd/yhadoop-common/hadoop-tools/hadoop-gridmix/target/log
2012-12-20 13:04:28.669 [INFO] Executed tasks
2012-12-20 13:04:28.670 [INFO] 
2012-12-20 13:04:28.672 [INFO] --- maven-compiler-plugin:2.5.1:testCompile 
(default-testCompile) @ hadoop-gridmix ---
2012-12-20 13:04:28.674 [INFO] Compiling 16 source files to 
/home/jenkins/jenkins/workspace/Hadoop-common-Integration-branch-2-gd/yhadoop-common/hadoop-tools/hadoop-gridmix/target/test-classes
2012-12-20 13:04:28.774 [INFO] 
-
2012-12-20 13:04:28.776 [ERROR] COMPILATION ERROR : 
2012-12-20 13:04:28.778 [INFO] 
-
2012-12-20 13:04:28.780 [ERROR] 
/home/jenkins/jenkins/workspace/Hadoop-common-Integration-branch-2-gd/yhadoop-common/hadoop-tools/hadoop-gridmix/src/test/java/org/apache/hadoop/mapred/gridmix/TestResourceUsageEmulators.java:[35,59]
 cannot find symbol
2012-12-20 13:04:28.781 symbol  : class ProcResourceValues
2012-12-20 13:04:28.782 location: class 
org.apache.hadoop.yarn.util.ResourceCalculatorPlugin
2012-12-20 13:04:28.784 [INFO] 1 error
2012-12-20 13:04:28.786 [INFO] 
-

 Change processTree interface to work better with native code
 

 Key: YARN-223
 URL: https://issues.apache.org/jira/browse/YARN-223
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Radim Kolar
Assignee: Radim Kolar
Priority: Critical
 Fix For: 3.0.0, 2.0.3-alpha

 Attachments: pstree-update4.txt, pstree-update6.txt, 
 pstree-update6.txt


 Problem is that on every update of processTree new object is required. This 
 is undesired when working with processTree implementation in native code.
 replace ProcessTree.getProcessTree() with updateProcessTree(). No new object 
 allocation is needed and it simplify application code a bit.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact 

[jira] [Commented] (YARN-223) Change processTree interface to work better with native code

2012-12-20 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13536956#comment-13536956
 ] 

Hudson commented on YARN-223:
-

Integrated in Hadoop-Yarn-trunk #71 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/71/])
YARN-223. Update process tree instead of getting new process trees. (Radim 
Kolar via llu) (Revision 1424244)

 Result = SUCCESS
llu : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1424244
Files : 
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Task.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRConfig.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/util/LinuxResourceCalculatorPlugin.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/util/ProcfsBasedProcessTree.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/util/TestProcfsBasedProcessTree.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/LinuxResourceCalculatorPlugin.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/ProcfsBasedProcessTree.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/ResourceCalculatorPlugin.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/ResourceCalculatorProcessTree.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/test/java/org/apache/hadoop/yarn/util/TestProcfsBasedProcessTree.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/test/java/org/apache/hadoop/yarn/util/TestResourceCalculatorProcessTree.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/monitor/ContainersMonitorImpl.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/monitor/TestContainersMonitor.java


 Change processTree interface to work better with native code
 

 Key: YARN-223
 URL: https://issues.apache.org/jira/browse/YARN-223
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Radim Kolar
Assignee: Radim Kolar
Priority: Critical
 Fix For: 3.0.0, 2.0.3-alpha

 Attachments: pstree-update4.txt, pstree-update6.txt, 
 pstree-update6.txt


 Problem is that on every update of processTree new object is required. This 
 is undesired when working with processTree implementation in native code.
 replace ProcessTree.getProcessTree() with updateProcessTree(). No new object 
 allocation is needed and it simplify application code a bit.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-223) Change processTree interface to work better with native code

2012-12-20 Thread Dennis Y (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13536969#comment-13536969
 ] 

Dennis Y commented on YARN-223:
---

I've created https://issues.apache.org/jira/browse/MAPREDUCE-4895 and submitted 
patch

 Change processTree interface to work better with native code
 

 Key: YARN-223
 URL: https://issues.apache.org/jira/browse/YARN-223
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Radim Kolar
Assignee: Radim Kolar
Priority: Critical
 Fix For: 3.0.0, 2.0.3-alpha

 Attachments: pstree-update4.txt, pstree-update6.txt, 
 pstree-update6.txt


 Problem is that on every update of processTree new object is required. This 
 is undesired when working with processTree implementation in native code.
 replace ProcessTree.getProcessTree() with updateProcessTree(). No new object 
 allocation is needed and it simplify application code a bit.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-103) Add a yarn AM - RM client module

2012-12-20 Thread Tom White (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-103?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13536990#comment-13536990
 ] 

Tom White commented on YARN-103:


Some feedback on the current patch:
* Why not use the existing ResourceRequest rather than creating a new type 
ContainerRequest? Having two equivalent types seems confusing.
* Is the ANY constant meant to be used by users? It looks like you specify a 
ContainerRequest with null hosts and racks in this case. If so, then it would 
be useful to add a constructor that doesn't take hosts or racks for that case, 
although given my previous point it would be easier to use ResourceRequest.
* {{assertTrue(containersRequestedRack == 2);}} appears on two successive 
lines. I think the assertion should be about containersRequestedAny.
* // do a few iterations to ensure RM is not going send new containers - wait 
in the loop to allow NMs to heartbeat?

 Add a yarn AM - RM client module
 

 Key: YARN-103
 URL: https://issues.apache.org/jira/browse/YARN-103
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Bikas Saha
Assignee: Bikas Saha
 Attachments: YARN-103.1.patch, YARN-103.2.patch, YARN-103.3.patch, 
 YARN-103.4.patch, YARN-103.4.wrapper.patch, YARN-103.5.patch, 
 YARN-103.6.patch, YARN-103.7.patch


 Add a basic client wrapper library to the AM RM protocol in order to prevent 
 proliferation of code being duplicated everywhere. Provide helper functions 
 to perform reverse mapping of container requests to RM allocation resource 
 request table format.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-223) Change processTree interface to work better with native code

2012-12-20 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537010#comment-13537010
 ] 

Hudson commented on YARN-223:
-

Integrated in Hadoop-Hdfs-trunk #1260 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1260/])
YARN-223. Update process tree instead of getting new process trees. (Radim 
Kolar via llu) (Revision 1424244)

 Result = FAILURE
llu : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1424244
Files : 
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapred/Task.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/MRConfig.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/util/LinuxResourceCalculatorPlugin.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/util/ProcfsBasedProcessTree.java
* 
/hadoop/common/trunk/hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient/src/test/java/org/apache/hadoop/mapreduce/util/TestProcfsBasedProcessTree.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/LinuxResourceCalculatorPlugin.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/ProcfsBasedProcessTree.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/ResourceCalculatorPlugin.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/java/org/apache/hadoop/yarn/util/ResourceCalculatorProcessTree.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/test/java/org/apache/hadoop/yarn/util/TestProcfsBasedProcessTree.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/test/java/org/apache/hadoop/yarn/util/TestResourceCalculatorProcessTree.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/monitor/ContainersMonitorImpl.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/monitor/TestContainersMonitor.java


 Change processTree interface to work better with native code
 

 Key: YARN-223
 URL: https://issues.apache.org/jira/browse/YARN-223
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Radim Kolar
Assignee: Radim Kolar
Priority: Critical
 Fix For: 3.0.0, 2.0.3-alpha

 Attachments: pstree-update4.txt, pstree-update6.txt, 
 pstree-update6.txt


 Problem is that on every update of processTree new object is required. This 
 is undesired when working with processTree implementation in native code.
 replace ProcessTree.getProcessTree() with updateProcessTree(). No new object 
 allocation is needed and it simplify application code a bit.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-50) Implement renewal / cancellation of Delegation Tokens

2012-12-20 Thread Daryn Sharp (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-50?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537075#comment-13537075
 ] 

Daryn Sharp commented on YARN-50:
-

A few minor questions/comments:
# Should the ctors for {{RenewDelegationTokenRequestPBImpl(proto}} and 
{{RenewDelegationTokenResponsePBImpl(proto)}} discard the given proto?  I'm not 
fluent in the protobuf implementation.
# Very minor, a comment in {{YarnClientImpl}} has misspelled ClietnRMProtocol
# Why do HS parameters need to be in RM config?  It should be sufficient to 
have MR in the class path to pickup the token renewers.

Issues:
# In {{RMDelegationTokenRenewer}}:
## The yarn client should connect to the service address within the token, 
rather than assume all RM tokens are for the local RM.  It's possible that a 
job might have a RM token for a different cluster.
## Are null checks on yarnClient within the try/finally needed?  If 
{{getYarnClient}} can return null, NPEs may be generated on the renew/cancel 
calls which would imply the conditional should replace the try block, or the 
conditional should be removed.
# {{ClientRMService#renewDelegationToken/cancelDelegationToken}} throw an 
exception when the connection is token based, which is correct, but the error 
message says tokens cannot be issued instead of cancelled/renewed.
# Based on the SASL work I'm doing, {{isAllowedDelegationTokenOp}} should be 
more like this: {{return 
UserGroupInformation.getCurrentUser().getRealAuthenticationMethod() != 
AuthenticationMethod.TOKEN}}.  (Aside: I've actually been meaning to push this 
check down into the ADTSM but I'll do that on another jira)

 Implement renewal / cancellation of Delegation Tokens
 -

 Key: YARN-50
 URL: https://issues.apache.org/jira/browse/YARN-50
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Siddharth Seth
Assignee: Siddharth Seth
 Attachments: YARN-50_wip.txt


 Currently, delegation tokens issues by the RM and History server cannot be 
 renewed or cancelled. This needs to be implemented.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-279) Generalize RM token management

2012-12-20 Thread Daryn Sharp (JIRA)
Daryn Sharp created YARN-279:


 Summary: Generalize RM token management
 Key: YARN-279
 URL: https://issues.apache.org/jira/browse/YARN-279
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: resourcemanager
Affects Versions: 0.23.5, 2.0.0-alpha, 3.0.0
Reporter: Daryn Sharp


Token renewal/cancelation in the RM presents challenges to support arbitrary 
tokens.  The RM's CLASSPATH is currently required to have token renewer classes 
and all associated classes for the project's client.  The logistics of having 
installs on the RM of all hadoop projects that submit jobs - just to support 
client connections to renew/cancel tokens - are untenable.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-279) Generalize RM token management

2012-12-20 Thread Daryn Sharp (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-279?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537148#comment-13537148
 ] 

Daryn Sharp commented on YARN-279:
--

The current token management (renew/cancel) has contaminated the RM with 
specific knowledge of other systems.  Unfortunately the token management cannot 
be pushed down to the AM because a token may expire before the RM launches the 
job.  Also, the AM would suffer from the same problem of needing other 
project's jars just to manage the tokens.

For the sake of discussion, the implementation of a token's management should 
be cleanly abstracted and removed from the RM.  One approach would be to 
require resources with tokens to provide a web service that implements renew 
and cancel.  The RM's renew timers then only call a REST api to manage the 
tokens.  A token would only need to contain a base uri for the REST call.

Thoughts?

 Generalize RM token management
 --

 Key: YARN-279
 URL: https://issues.apache.org/jira/browse/YARN-279
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: resourcemanager
Affects Versions: 3.0.0, 2.0.0-alpha, 0.23.5
Reporter: Daryn Sharp

 Token renewal/cancelation in the RM presents challenges to support arbitrary 
 tokens.  The RM's CLASSPATH is currently required to have token renewer 
 classes and all associated classes for the project's client.  The logistics 
 of having installs on the RM of all hadoop projects that submit jobs - just 
 to support client connections to renew/cancel tokens - are untenable.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-223) Change processTree interface to work better with native code

2012-12-20 Thread Thomas Graves (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-223?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thomas Graves updated YARN-223:
---

Fix Version/s: 0.23.6

I pulled this into branch-0.23

 Change processTree interface to work better with native code
 

 Key: YARN-223
 URL: https://issues.apache.org/jira/browse/YARN-223
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Radim Kolar
Assignee: Radim Kolar
Priority: Critical
 Fix For: 3.0.0, 2.0.3-alpha, 0.23.6

 Attachments: pstree-update4.txt, pstree-update6.txt, 
 pstree-update6.txt


 Problem is that on every update of processTree new object is required. This 
 is undesired when working with processTree implementation in native code.
 replace ProcessTree.getProcessTree() with updateProcessTree(). No new object 
 allocation is needed and it simplify application code a bit.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-223) Change processTree interface to work better with native code

2012-12-20 Thread Luke Lu (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537165#comment-13537165
 ] 

Luke Lu commented on YARN-223:
--

Sorry folks! I got too used to seeing mapreduce code in YARN :)

 Change processTree interface to work better with native code
 

 Key: YARN-223
 URL: https://issues.apache.org/jira/browse/YARN-223
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Radim Kolar
Assignee: Radim Kolar
Priority: Critical
 Fix For: 3.0.0, 2.0.3-alpha, 0.23.6

 Attachments: pstree-update4.txt, pstree-update6.txt, 
 pstree-update6.txt


 Problem is that on every update of processTree new object is required. This 
 is undesired when working with processTree implementation in native code.
 replace ProcessTree.getProcessTree() with updateProcessTree(). No new object 
 allocation is needed and it simplify application code a bit.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-223) Change processTree interface to work better with native code

2012-12-20 Thread Luke Lu (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-223?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537166#comment-13537166
 ] 

Luke Lu commented on YARN-223:
--

I think the PreCommit test patch system should either handle cross project 
patches or outright reject such patches.

 Change processTree interface to work better with native code
 

 Key: YARN-223
 URL: https://issues.apache.org/jira/browse/YARN-223
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Radim Kolar
Assignee: Radim Kolar
Priority: Critical
 Fix For: 3.0.0, 2.0.3-alpha, 0.23.6

 Attachments: pstree-update4.txt, pstree-update6.txt, 
 pstree-update6.txt


 Problem is that on every update of processTree new object is required. This 
 is undesired when working with processTree implementation in native code.
 replace ProcessTree.getProcessTree() with updateProcessTree(). No new object 
 allocation is needed and it simplify application code a bit.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-276) Capacity Scheduler can hang when submit many jobs concurrently

2012-12-20 Thread Robert Joseph Evans (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-276?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537168#comment-13537168
 ] 

Robert Joseph Evans commented on YARN-276:
--

I am not an expert on the scheduler code, so I have not done an in depth review 
of the patch.  My biggest concern with this is that there is no visibility in 
the UI/web services about why an app may not have been scheduled.  It would be 
great if you could update CapacitySchedulerLeafQueueInfo.java and the web page 
that uses it CapacitySchedulerPage.java.

 Capacity Scheduler can hang when submit many jobs concurrently
 --

 Key: YARN-276
 URL: https://issues.apache.org/jira/browse/YARN-276
 Project: Hadoop YARN
  Issue Type: Bug
  Components: capacityscheduler
Affects Versions: 3.0.0, 2.0.1-alpha
Reporter: nemon lou
 Attachments: YARN-276.patch, YARN-276.patch, YARN-276.patch, 
 YARN-276.patch, YARN-276.patch

   Original Estimate: 24h
  Remaining Estimate: 24h

 In hadoop2.0.1,When i submit many jobs concurrently at the same time,Capacity 
 scheduler can hang with most resources taken up by AM and don't have enough 
 resources for tasks.And then all applications hang there.
 The cause is that yarn.scheduler.capacity.maximum-am-resource-percent not 
 check directly.Instead ,this property only used for maxActiveApplications. 
 And maxActiveApplications is computed by minimumAllocation (not by Am 
 actually used).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-280) RM does not reject app submission with invalid tokens

2012-12-20 Thread Daryn Sharp (JIRA)
Daryn Sharp created YARN-280:


 Summary: RM does not reject app submission with invalid tokens
 Key: YARN-280
 URL: https://issues.apache.org/jira/browse/YARN-280
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 0.23.5, 2.0.0-alpha, 3.0.0
Reporter: Daryn Sharp
Assignee: Daryn Sharp


The RM will launch an app with invalid tokens.  The tasks will languish with 
failed connection retries, followed by task reattempts, followed by app 
reattempts.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-280) RM does not reject app submission with invalid tokens

2012-12-20 Thread Daryn Sharp (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-280?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daryn Sharp updated YARN-280:
-

Attachment: YARN-280.patch

Instead of scheduling tokens for immediate renewal in the background, the first 
renewal happens in the current thread so an invalid token will fail the app 
submission.  Once all tokens are known to be valid, they are scheduled for 
renewal.

 RM does not reject app submission with invalid tokens
 -

 Key: YARN-280
 URL: https://issues.apache.org/jira/browse/YARN-280
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 3.0.0, 2.0.0-alpha, 0.23.5
Reporter: Daryn Sharp
Assignee: Daryn Sharp
 Attachments: YARN-280.patch


 The RM will launch an app with invalid tokens.  The tasks will languish with 
 failed connection retries, followed by task reattempts, followed by app 
 reattempts.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Moved] (YARN-281) Add a test for YARN Schedulers' MAXIMUM_ALLOCATION limits

2012-12-20 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-281?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J moved MAPREDUCE-4171 to YARN-281:
-

  Component/s: (was: mrv2)
   (was: test)
   scheduler
Affects Version/s: (was: 2.0.0-alpha)
   2.0.0-alpha
  Key: YARN-281  (was: MAPREDUCE-4171)
  Project: Hadoop YARN  (was: Hadoop Map/Reduce)

 Add a test for YARN Schedulers' MAXIMUM_ALLOCATION limits
 -

 Key: YARN-281
 URL: https://issues.apache.org/jira/browse/YARN-281
 Project: Hadoop YARN
  Issue Type: Test
  Components: scheduler
Affects Versions: 2.0.0-alpha
Reporter: Harsh J
Assignee: Harsh J
  Labels: test

 We currently have tests that test MINIMUM_ALLOCATION limits for FifoScheduler 
 and the likes, but no test for MAXIMUM_ALLOCATION yet. We should add a test 
 to prevent regressions of any kind on such limits.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-278) Fair scheduler maxRunningApps config causes no apps to make progress

2012-12-20 Thread Sandy Ryza (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-278?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sandy Ryza updated YARN-278:


Attachment: YARN-278.patch

 Fair scheduler maxRunningApps config causes no apps to make progress
 

 Key: YARN-278
 URL: https://issues.apache.org/jira/browse/YARN-278
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager, scheduler
Affects Versions: 2.0.2-alpha
Reporter: Sandy Ryza
Assignee: Sandy Ryza
 Attachments: YARN-278.patch, YARN-278.patch


 This occurs because the scheduler erroneously chooses apps to offer resources 
 to that are not runnable, then later decides they are not runnable, and 
 doesn't try to give the resources to anyone else.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-282) Fair scheduler web UI double counts Apps Submitted

2012-12-20 Thread Sandy Ryza (JIRA)
Sandy Ryza created YARN-282:
---

 Summary: Fair scheduler web UI double counts Apps Submitted
 Key: YARN-282
 URL: https://issues.apache.org/jira/browse/YARN-282
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.0.2-alpha
Reporter: Sandy Ryza
Assignee: Sandy Ryza


Each app submitted is reported twice under Apps Submitted

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-282) Fair scheduler web UI double counts Apps Submitted

2012-12-20 Thread Sandy Ryza (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-282?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sandy Ryza updated YARN-282:


Attachment: YARN-282.patch

 Fair scheduler web UI double counts Apps Submitted
 --

 Key: YARN-282
 URL: https://issues.apache.org/jira/browse/YARN-282
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.0.2-alpha
Reporter: Sandy Ryza
Assignee: Sandy Ryza
 Attachments: YARN-282.patch


 Each app submitted is reported twice under Apps Submitted

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-278) Fair scheduler maxRunningApps config causes no apps to make progress

2012-12-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-278?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537291#comment-13537291
 ] 

Hadoop QA commented on YARN-278:


{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12561959/YARN-278.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/241//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/241//console

This message is automatically generated.

 Fair scheduler maxRunningApps config causes no apps to make progress
 

 Key: YARN-278
 URL: https://issues.apache.org/jira/browse/YARN-278
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager, scheduler
Affects Versions: 2.0.2-alpha
Reporter: Sandy Ryza
Assignee: Sandy Ryza
 Attachments: YARN-278.patch, YARN-278.patch


 This occurs because the scheduler erroneously chooses apps to offer resources 
 to that are not runnable, then later decides they are not runnable, and 
 doesn't try to give the resources to anyone else.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-50) Implement renewal / cancellation of Delegation Tokens

2012-12-20 Thread Siddharth Seth (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-50?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Siddharth Seth updated YARN-50:
---

Attachment: YARN-50.txt

Thanks for the quick reviews Daryn and Bobby. Updated patch to address the 
issues. Since the service address is being used - MR config isn't really 
required (in the MR part of the patch).

bq. Based on the SASL work I'm doing, isAllowedDelegationTokenOp should be more 
like this: return 
UserGroupInformation.getCurrentUser().getRealAuthenticationMethod() != 
AuthenticationMethod.TOKEN. (Aside: I've actually been meaning to push this 
check down into the ADTSM but I'll do that on another jira)
For security related checks, I think an explicit whitelist is a better option. 
Have included KERBEROS, KERBEROS_SSL and CERTIFICATE. Feel free to change this.

bq. The yarn client should connect to the service address within the token, 
rather than assume all RM tokens are for the local RM. It's possible that a job 
might have a RM token for a different cluster.
I'm not sure this is a valid use case. Get a token from one RM, submit a job to 
RM2 with RM1s token, which then may submit back to RM1... both RMs end up 
renewing the token. IAC, using the service address to remove the dependency on 
the configs.

 Implement renewal / cancellation of Delegation Tokens
 -

 Key: YARN-50
 URL: https://issues.apache.org/jira/browse/YARN-50
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Siddharth Seth
Assignee: Siddharth Seth
Priority: Blocker
 Attachments: YARN-50.txt, YARN-50_wip.txt


 Currently, delegation tokens issues by the RM and History server cannot be 
 renewed or cancelled. This needs to be implemented.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-282) Fair scheduler web UI double counts Apps Submitted

2012-12-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-282?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537308#comment-13537308
 ] 

Hadoop QA commented on YARN-282:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12561966/YARN-282.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/242//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/242//console

This message is automatically generated.

 Fair scheduler web UI double counts Apps Submitted
 --

 Key: YARN-282
 URL: https://issues.apache.org/jira/browse/YARN-282
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 2.0.2-alpha
Reporter: Sandy Ryza
Assignee: Sandy Ryza
 Attachments: YARN-282.patch


 Each app submitted is reported twice under Apps Submitted

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-50) Implement renewal / cancellation of Delegation Tokens

2012-12-20 Thread Daryn Sharp (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-50?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537329#comment-13537329
 ] 

Daryn Sharp commented on YARN-50:
-

I think it looks like my concerns are addressed.  Only remark is that 
{{ResourceManager}} appears to have some spurious imports added to it.  It 
looks like changes were made and then reverted except the imports?

 Implement renewal / cancellation of Delegation Tokens
 -

 Key: YARN-50
 URL: https://issues.apache.org/jira/browse/YARN-50
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Siddharth Seth
Assignee: Siddharth Seth
Priority: Blocker
 Attachments: YARN-50.txt, YARN-50_wip.txt


 Currently, delegation tokens issues by the RM and History server cannot be 
 renewed or cancelled. This needs to be implemented.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-50) Implement renewal / cancellation of Delegation Tokens

2012-12-20 Thread Daryn Sharp (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-50?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537335#comment-13537335
 ] 

Daryn Sharp commented on YARN-50:
-

Actually, it looks like {{CancelDelegationTokenResponsePBImpl(port)}} should be 
setting viaProto = true?

 Implement renewal / cancellation of Delegation Tokens
 -

 Key: YARN-50
 URL: https://issues.apache.org/jira/browse/YARN-50
 Project: Hadoop YARN
  Issue Type: Sub-task
Reporter: Siddharth Seth
Assignee: Siddharth Seth
Priority: Blocker
 Attachments: YARN-50.txt, YARN-50_wip.txt


 Currently, delegation tokens issues by the RM and History server cannot be 
 renewed or cancelled. This needs to be implemented.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-283) Fair scheduler fails to get queue info without root prefix

2012-12-20 Thread Sandy Ryza (JIRA)
Sandy Ryza created YARN-283:
---

 Summary: Fair scheduler fails to get queue info without root prefix
 Key: YARN-283
 URL: https://issues.apache.org/jira/browse/YARN-283
 Project: Hadoop YARN
  Issue Type: Bug
  Components: scheduler
Affects Versions: 2.0.2-alpha
Reporter: Sandy Ryza
Assignee: Sandy Ryza


If queue1 exists, and a client calls mapred queue -info queue1, an exception 
is thrown.  If they use root.queue1, it works correctly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-283) Fair scheduler fails to get queue info without root prefix

2012-12-20 Thread Sandy Ryza (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-283?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sandy Ryza updated YARN-283:


Attachment: YARN-283.patch

 Fair scheduler fails to get queue info without root prefix
 --

 Key: YARN-283
 URL: https://issues.apache.org/jira/browse/YARN-283
 Project: Hadoop YARN
  Issue Type: Bug
  Components: scheduler
Affects Versions: 2.0.2-alpha
Reporter: Sandy Ryza
Assignee: Sandy Ryza
 Attachments: YARN-283.patch


 If queue1 exists, and a client calls mapred queue -info queue1, an 
 exception is thrown.  If they use root.queue1, it works correctly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-283) Fair scheduler fails to get queue info without root prefix

2012-12-20 Thread Sandy Ryza (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-283?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537446#comment-13537446
 ] 

Sandy Ryza commented on YARN-283:
-

Tested manually that it works.

 Fair scheduler fails to get queue info without root prefix
 --

 Key: YARN-283
 URL: https://issues.apache.org/jira/browse/YARN-283
 Project: Hadoop YARN
  Issue Type: Bug
  Components: scheduler
Affects Versions: 2.0.2-alpha
Reporter: Sandy Ryza
Assignee: Sandy Ryza
 Attachments: YARN-283.patch


 If queue1 exists, and a client calls mapred queue -info queue1, an 
 exception is thrown.  If they use root.queue1, it works correctly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-283) Fair scheduler fails to get queue info without root prefix

2012-12-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-283?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537457#comment-13537457
 ] 

Hadoop QA commented on YARN-283:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12561998/YARN-283.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  The javadoc tool did not generate any 
warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/243//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/243//console

This message is automatically generated.

 Fair scheduler fails to get queue info without root prefix
 --

 Key: YARN-283
 URL: https://issues.apache.org/jira/browse/YARN-283
 Project: Hadoop YARN
  Issue Type: Bug
  Components: scheduler
Affects Versions: 2.0.2-alpha
Reporter: Sandy Ryza
Assignee: Sandy Ryza
 Attachments: YARN-283.patch


 If queue1 exists, and a client calls mapred queue -info queue1, an 
 exception is thrown.  If they use root.queue1, it works correctly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-142) Change YARN APIs to throw IOException

2012-12-20 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-142?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated YARN-142:
---

Attachment: YARN-142.1.patch

Use IOException to replace the YarnRemoteException 

 Change YARN APIs to throw IOException
 -

 Key: YARN-142
 URL: https://issues.apache.org/jira/browse/YARN-142
 Project: Hadoop YARN
  Issue Type: Bug
Affects Versions: 0.23.3, 2.0.0-alpha
Reporter: Siddharth Seth
Assignee: Xuan Gong
Priority: Critical
 Attachments: YARN-142.1.patch


 Ref: MAPREDUCE-4067
 All YARN APIs currently throw YarnRemoteException.
 1) This cannot be extended in it's current form.
 2) The RPC layer can throw IOExceptions. These end up showing up as 
 UndeclaredThrowableExceptions.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-283) Fair scheduler fails to get queue info without root prefix

2012-12-20 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-283?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537537#comment-13537537
 ] 

Karthik Kambatla commented on YARN-283:
---

+1

 Fair scheduler fails to get queue info without root prefix
 --

 Key: YARN-283
 URL: https://issues.apache.org/jira/browse/YARN-283
 Project: Hadoop YARN
  Issue Type: Bug
  Components: scheduler
Affects Versions: 2.0.2-alpha
Reporter: Sandy Ryza
Assignee: Sandy Ryza
 Attachments: YARN-283.patch


 If queue1 exists, and a client calls mapred queue -info queue1, an 
 exception is thrown.  If they use root.queue1, it works correctly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-276) Capacity Scheduler can hang when submit many jobs concurrently

2012-12-20 Thread nemon lou (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-276?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13537643#comment-13537643
 ] 

nemon lou commented on YARN-276:


Good idea ,Robert.Thank you for your comment.
I think it's good to display  AM used resources and AM percent limit(or max 
resources that AMs can use) for each leaf queue on capacity scheduler page.



 Capacity Scheduler can hang when submit many jobs concurrently
 --

 Key: YARN-276
 URL: https://issues.apache.org/jira/browse/YARN-276
 Project: Hadoop YARN
  Issue Type: Bug
  Components: capacityscheduler
Affects Versions: 3.0.0, 2.0.1-alpha
Reporter: nemon lou
 Attachments: YARN-276.patch, YARN-276.patch, YARN-276.patch, 
 YARN-276.patch, YARN-276.patch

   Original Estimate: 24h
  Remaining Estimate: 24h

 In hadoop2.0.1,When i submit many jobs concurrently at the same time,Capacity 
 scheduler can hang with most resources taken up by AM and don't have enough 
 resources for tasks.And then all applications hang there.
 The cause is that yarn.scheduler.capacity.maximum-am-resource-percent not 
 check directly.Instead ,this property only used for maxActiveApplications. 
 And maxActiveApplications is computed by minimumAllocation (not by Am 
 actually used).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Moved] (YARN-284) YARN capacity scheduler doesn't spread MR tasks evenly on an underutilized cluster

2012-12-20 Thread Harsh J (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-284?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Harsh J moved MAPREDUCE-3268 to YARN-284:
-

  Component/s: (was: scheduler)
   scheduler
Affects Version/s: (was: 0.23.0)
   2.0.0-alpha
  Key: YARN-284  (was: MAPREDUCE-3268)
  Project: Hadoop YARN  (was: Hadoop Map/Reduce)

 YARN capacity scheduler doesn't spread MR tasks evenly on an underutilized 
 cluster
 --

 Key: YARN-284
 URL: https://issues.apache.org/jira/browse/YARN-284
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: scheduler
Affects Versions: 2.0.0-alpha
Reporter: Todd Lipcon

 The fair scheduler in MR1 has the behavior that, if a job is submitted to an 
 under-utilized cluster and the cluster has more open slots than tasks in the 
 job, the tasks are spread evenly throughout the cluster. This improves job 
 latency since more spindles and NICs are utilized to complete the job. In MR2 
 I see this issue causing significantly longer job runtimes when there is 
 excess capacity in the cluster -- especially on reducers which sometimmes end 
 up clumping together on a smaller set of nodes which then become disk/network 
 constrained.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-285) RM should be able to provide a tracking link for apps that have already been purged

2012-12-20 Thread Derek Dagit (JIRA)
Derek Dagit created YARN-285:


 Summary: RM should be able to provide a tracking link for apps 
that have already been purged
 Key: YARN-285
 URL: https://issues.apache.org/jira/browse/YARN-285
 Project: Hadoop YARN
  Issue Type: Improvement
Affects Versions: 0.23.5, 3.0.0
Reporter: Derek Dagit
Assignee: Derek Dagit


As applications complete, the RM tracks their IDs in a completed list.  This 
list is routinely truncated to limit the total number of application remembered 
by the RM.

When a user clicks the History for a job, either the browser is redirected to 
the application's tracking link obtained from the stored application instance.  
But when the application has been purged from the RM, an error is displayed.

In very busy clusters the rate at which applications complete can cause 
applications to be purged from the RM's internal list within hours, which 
breaks the proxy URLs users have saved for their jobs.

We would like the RM to provide valid tracking links persist so that users are 
not frustrated by broken links.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-285) RM should be able to provide a tracking link for apps that have already been purged

2012-12-20 Thread Derek Dagit (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-285?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Derek Dagit updated YARN-285:
-

Attachment: YARN-285-branch-0.23.patch

 RM should be able to provide a tracking link for apps that have already been 
 purged
 ---

 Key: YARN-285
 URL: https://issues.apache.org/jira/browse/YARN-285
 Project: Hadoop YARN
  Issue Type: Improvement
Affects Versions: 3.0.0, 0.23.5
Reporter: Derek Dagit
Assignee: Derek Dagit
 Attachments: YARN-285-branch-0.23.patch


 As applications complete, the RM tracks their IDs in a completed list.  This 
 list is routinely truncated to limit the total number of application 
 remembered by the RM.
 When a user clicks the History for a job, either the browser is redirected to 
 the application's tracking link obtained from the stored application 
 instance.  But when the application has been purged from the RM, an error is 
 displayed.
 In very busy clusters the rate at which applications complete can cause 
 applications to be purged from the RM's internal list within hours, which 
 breaks the proxy URLs users have saved for their jobs.
 We would like the RM to provide valid tracking links persist so that users 
 are not frustrated by broken links.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-285) RM should be able to provide a tracking link for apps that have already been purged

2012-12-20 Thread Derek Dagit (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-285?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Derek Dagit updated YARN-285:
-

Attachment: YARN-285.patch

This is a patch that introduces handling for an optional tracking URI plugin, 
specified via a new config, that should accept a given ApplicationId and return 
a valid URI.

The plugin is used in cases where a proxy link is requested, but the RM has 
already purged the given application.  The plugin will use the ApplicationId to 
rewrite the URL, and the browser will be redirected just as when the RM uses 
the tracking URL of a application the it knows about.

The intention of the plugin instead of a simple URL rewrite is to keep 
architectural separation between Yarn and MapReduce.

The mechanism is similar to that involving the ResourceCalculatorPlugin.

 RM should be able to provide a tracking link for apps that have already been 
 purged
 ---

 Key: YARN-285
 URL: https://issues.apache.org/jira/browse/YARN-285
 Project: Hadoop YARN
  Issue Type: Improvement
Affects Versions: 3.0.0, 0.23.5
Reporter: Derek Dagit
Assignee: Derek Dagit
 Attachments: YARN-285-branch-0.23.patch, YARN-285.patch


 As applications complete, the RM tracks their IDs in a completed list.  This 
 list is routinely truncated to limit the total number of application 
 remembered by the RM.
 When a user clicks the History for a job, either the browser is redirected to 
 the application's tracking link obtained from the stored application 
 instance.  But when the application has been purged from the RM, an error is 
 displayed.
 In very busy clusters the rate at which applications complete can cause 
 applications to be purged from the RM's internal list within hours, which 
 breaks the proxy URLs users have saved for their jobs.
 We would like the RM to provide valid tracking links persist so that users 
 are not frustrated by broken links.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira