[jira] [Commented] (YARN-3200) Factor OSType out from Shell: changes in YARN

2015-02-15 Thread Yongjun Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3200?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14321812#comment-14321812
 ] 

Yongjun Zhang commented on YARN-3200:
-

This is a subtask that I will be working on, but it seems I can't assign the 
jira to myself. Anyone who has the privilege, appreciate your help to assign it 
to me. Thanks.



 Factor OSType out from Shell: changes in YARN
 -

 Key: YARN-3200
 URL: https://issues.apache.org/jira/browse/YARN-3200
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Yongjun Zhang
 Fix For: 2.7.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (YARN-3200) Factor OSType out from Shell: changes in YARN

2015-02-15 Thread Brahma Reddy Battula (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3200?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brahma Reddy Battula reassigned YARN-3200:
--

Assignee: Brahma Reddy Battula

 Factor OSType out from Shell: changes in YARN
 -

 Key: YARN-3200
 URL: https://issues.apache.org/jira/browse/YARN-3200
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Yongjun Zhang
Assignee: Brahma Reddy Battula
 Fix For: 2.7.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-2799) cleanup TestLogAggregationService based on the change in YARN-90

2015-02-15 Thread zhihai xu (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-2799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14321754#comment-14321754
 ] 

zhihai xu commented on YARN-2799:
-

Hi [~djp],
Yes, you are right. that is a good catch. I uploaded a new patch 
YARN-2799.002.patch which addressed your comment.
Please review it.
thanks
zhihai

 cleanup TestLogAggregationService based on the change in YARN-90
 

 Key: YARN-2799
 URL: https://issues.apache.org/jira/browse/YARN-2799
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: test
Reporter: zhihai xu
Assignee: zhihai xu
Priority: Minor
 Attachments: YARN-2799.000.patch, YARN-2799.001.patch, 
 YARN-2799.002.patch


 cleanup TestLogAggregationService based on the change in YARN-90.
 The following code is added to setup in YARN-90, 
 {code}
 dispatcher = createDispatcher();
 appEventHandler = mock(EventHandler.class);
 dispatcher.register(ApplicationEventType.class, appEventHandler);
 {code}
 In this case, we should remove all these code from each test function to 
 avoid duplicate code.
 Same for dispatcher.stop() which is in tearDown,
 we can remove dispatcher.stop() from from each test function also because it 
 will always be called from tearDown for each test.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-2799) cleanup TestLogAggregationService based on the change in YARN-90

2015-02-15 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-2799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14321762#comment-14321762
 ] 

Hadoop QA commented on YARN-2799:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12698939/YARN-2799.002.patch
  against trunk revision ef950ea.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/6639//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/6639//console

This message is automatically generated.

 cleanup TestLogAggregationService based on the change in YARN-90
 

 Key: YARN-2799
 URL: https://issues.apache.org/jira/browse/YARN-2799
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: test
Reporter: zhihai xu
Assignee: zhihai xu
Priority: Minor
 Attachments: YARN-2799.000.patch, YARN-2799.001.patch, 
 YARN-2799.002.patch


 cleanup TestLogAggregationService based on the change in YARN-90.
 The following code is added to setup in YARN-90, 
 {code}
 dispatcher = createDispatcher();
 appEventHandler = mock(EventHandler.class);
 dispatcher.register(ApplicationEventType.class, appEventHandler);
 {code}
 In this case, we should remove all these code from each test function to 
 avoid duplicate code.
 Same for dispatcher.stop() which is in tearDown,
 we can remove dispatcher.stop() from from each test function also because it 
 will always be called from tearDown for each test.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3199) Fair Scheduler documentation improvements

2015-02-15 Thread Rohit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14321798#comment-14321798
 ] 

Rohit Agarwal commented on YARN-3199:
-

[~aw], [~gururaj] Please also pick this up when converting the docs from apt to 
markdown as part of YARN-3168.

 Fair Scheduler documentation improvements
 -

 Key: YARN-3199
 URL: https://issues.apache.org/jira/browse/YARN-3199
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: fairscheduler
Affects Versions: 2.6.0
Reporter: Rohit Agarwal
Priority: Minor
  Labels: documentation
 Attachments: YARN-3199.patch


 {{yarn.scheduler.increment-allocation-mb}} and 
 {{yarn.scheduler.increment-allocation-vcores}} are not documented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3199) Fair Scheduler documentation improvements

2015-02-15 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3199?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14321805#comment-14321805
 ] 

Hadoop QA commented on YARN-3199:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12698946/YARN-3199.patch
  against trunk revision 3338f6d.

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+0 tests included{color}.  The patch appears to be a 
documentation patch that doesn't require tests.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in .

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/6640//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/6640//console

This message is automatically generated.

 Fair Scheduler documentation improvements
 -

 Key: YARN-3199
 URL: https://issues.apache.org/jira/browse/YARN-3199
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: fairscheduler
Affects Versions: 2.6.0
Reporter: Rohit Agarwal
Priority: Minor
  Labels: documentation
 Attachments: YARN-3199.patch


 {{yarn.scheduler.increment-allocation-mb}} and 
 {{yarn.scheduler.increment-allocation-vcores}} are not documented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (YARN-3199) Fair Scheduler documentation improvements

2015-02-15 Thread Rohit Agarwal (JIRA)
Rohit Agarwal created YARN-3199:
---

 Summary: Fair Scheduler documentation improvements
 Key: YARN-3199
 URL: https://issues.apache.org/jira/browse/YARN-3199
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: fairscheduler
Affects Versions: 2.6.0
Reporter: Rohit Agarwal
Priority: Minor


{{yarn.scheduler.increment-allocation-mb}} and 
{{yarn.scheduler.increment-allocation-vcores}} are not documented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3199) Fair Scheduler documentation improvements

2015-02-15 Thread Rohit Agarwal (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3199?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rohit Agarwal updated YARN-3199:

Attachment: YARN-3199.patch

 Fair Scheduler documentation improvements
 -

 Key: YARN-3199
 URL: https://issues.apache.org/jira/browse/YARN-3199
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: fairscheduler
Affects Versions: 2.6.0
Reporter: Rohit Agarwal
Priority: Minor
  Labels: documentation
 Attachments: YARN-3199.patch


 {{yarn.scheduler.increment-allocation-mb}} and 
 {{yarn.scheduler.increment-allocation-vcores}} are not documented.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-2749) Some testcases from TestLogAggregationService fails in trunk

2015-02-15 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-2749?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14322007#comment-14322007
 ] 

Hudson commented on YARN-2749:
--

FAILURE: Integrated in Hadoop-trunk-Commit #7117 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/7117/])
YARN-2749. Fix some testcases from TestLogAggregationService fails in trunk. 
(Contributed by Xuan Gong) (junping_du: rev 
ab0b958a522d502426b91b6e4ab6dd29caccc372)
* hadoop-yarn-project/CHANGES.txt
* 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/logaggregation/AppLogAggregatorImpl.java


 Some testcases from TestLogAggregationService fails in trunk
 

 Key: YARN-2749
 URL: https://issues.apache.org/jira/browse/YARN-2749
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Xuan Gong
Assignee: Xuan Gong
 Fix For: 2.7.0

 Attachments: YARN-2749.1.patch, YARN-2749.2.patch, YARN-2749.2.patch


 Some testcases from TestLogAggregationService fails in trunk. 
 Those can be reproduced in centos
 Stack Trace:
 java.lang.AssertionError: null
   at org.junit.Assert.fail(Assert.java:86)
   at org.junit.Assert.assertTrue(Assert.java:41)
   at org.junit.Assert.assertTrue(Assert.java:52)
   at 
 org.apache.hadoop.yarn.server.nodemanager.containermanager.logaggregation.TestLogAggregationService.testLogAggregationService(TestLogAggregationService.java:1362)
   at 
 org.apache.hadoop.yarn.server.nodemanager.containermanager.logaggregation.TestLogAggregationService.testLogAggregationServiceWithRetention(TestLogAggregationService.java:1290)
 Stack Trace:
 java.lang.AssertionError: null
   at org.junit.Assert.fail(Assert.java:86)
   at org.junit.Assert.assertTrue(Assert.java:41)
   at org.junit.Assert.assertTrue(Assert.java:52)
   at 
 org.apache.hadoop.yarn.server.nodemanager.containermanager.logaggregation.TestLogAggregationService.testLogAggregationService(TestLogAggregationService.java:1362)
   at 
 org.apache.hadoop.yarn.server.nodemanager.containermanager.logaggregation.TestLogAggregationService.testLogAggregationServiceWithRetention(TestLogAggregationService.java:1290)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3200) Factor OSType out from Shell: changes in YARN

2015-02-15 Thread Tsuyoshi OZAWA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3200?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14322064#comment-14322064
 ] 

Tsuyoshi OZAWA commented on YARN-3200:
--

s/head/ahead/

 Factor OSType out from Shell: changes in YARN
 -

 Key: YARN-3200
 URL: https://issues.apache.org/jira/browse/YARN-3200
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Yongjun Zhang
Assignee: Yongjun Zhang
 Fix For: 2.7.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3200) Factor OSType out from Shell: changes in YARN

2015-02-15 Thread Tsuyoshi OZAWA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3200?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14322063#comment-14322063
 ] 

Tsuyoshi OZAWA commented on YARN-3200:
--

[~yzhangal] done. Please go head.

 Factor OSType out from Shell: changes in YARN
 -

 Key: YARN-3200
 URL: https://issues.apache.org/jira/browse/YARN-3200
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Yongjun Zhang
Assignee: Yongjun Zhang
 Fix For: 2.7.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3200) Factor OSType out from Shell: changes in YARN

2015-02-15 Thread Yongjun Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3200?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14322070#comment-14322070
 ] 

Yongjun Zhang commented on YARN-3200:
-

Thanks a lot [~brahmareddy] and [~ozawa]! I will submit patch after 
HADOOP-11597 is resolved because of the dependency.



 Factor OSType out from Shell: changes in YARN
 -

 Key: YARN-3200
 URL: https://issues.apache.org/jira/browse/YARN-3200
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Yongjun Zhang
Assignee: Yongjun Zhang
 Fix For: 2.7.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3200) Factor OSType out from Shell: changes in YARN

2015-02-15 Thread Tsuyoshi OZAWA (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3200?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsuyoshi OZAWA updated YARN-3200:
-
Assignee: Brahma Reddy Battula  (was: Tsuyoshi OZAWA)

 Factor OSType out from Shell: changes in YARN
 -

 Key: YARN-3200
 URL: https://issues.apache.org/jira/browse/YARN-3200
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Yongjun Zhang
Assignee: Brahma Reddy Battula
 Fix For: 2.7.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3200) Factor OSType out from Shell: changes in YARN

2015-02-15 Thread Tsuyoshi OZAWA (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3200?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsuyoshi OZAWA updated YARN-3200:
-
Assignee: (was: Brahma Reddy Battula)

 Factor OSType out from Shell: changes in YARN
 -

 Key: YARN-3200
 URL: https://issues.apache.org/jira/browse/YARN-3200
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Yongjun Zhang
 Fix For: 2.7.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3200) Factor OSType out from Shell: changes in YARN

2015-02-15 Thread Tsuyoshi OZAWA (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3200?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Tsuyoshi OZAWA updated YARN-3200:
-
Assignee: Yongjun Zhang

 Factor OSType out from Shell: changes in YARN
 -

 Key: YARN-3200
 URL: https://issues.apache.org/jira/browse/YARN-3200
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Yongjun Zhang
Assignee: Yongjun Zhang
 Fix For: 2.7.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (YARN-3200) Factor OSType out from Shell: changes in YARN

2015-02-15 Thread Yongjun Zhang (JIRA)
Yongjun Zhang created YARN-3200:
---

 Summary: Factor OSType out from Shell: changes in YARN
 Key: YARN-3200
 URL: https://issues.apache.org/jira/browse/YARN-3200
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Yongjun Zhang
 Fix For: 2.7.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-2799) cleanup TestLogAggregationService based on the change in YARN-90

2015-02-15 Thread zhihai xu (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-2799?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

zhihai xu updated YARN-2799:

Attachment: YARN-2799.002.patch

 cleanup TestLogAggregationService based on the change in YARN-90
 

 Key: YARN-2799
 URL: https://issues.apache.org/jira/browse/YARN-2799
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: test
Reporter: zhihai xu
Assignee: zhihai xu
Priority: Minor
 Attachments: YARN-2799.000.patch, YARN-2799.001.patch, 
 YARN-2799.002.patch


 cleanup TestLogAggregationService based on the change in YARN-90.
 The following code is added to setup in YARN-90, 
 {code}
 dispatcher = createDispatcher();
 appEventHandler = mock(EventHandler.class);
 dispatcher.register(ApplicationEventType.class, appEventHandler);
 {code}
 In this case, we should remove all these code from each test function to 
 avoid duplicate code.
 Same for dispatcher.stop() which is in tearDown,
 we can remove dispatcher.stop() from from each test function also because it 
 will always be called from tearDown for each test.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3194) After NM restart,completed containers are not released which are sent during NM registration

2015-02-15 Thread Rohith (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14322277#comment-14322277
 ] 

Rohith commented on YARN-3194:
--

bq. it's removing the old node and adding the newly connected node. RM is also 
not restarted. 
{{RMNodeImpl#ReconnectNodeTransition#.transition}} does not remove old node if 
any applications are running. In the below code, if noRunningApps is false then 
Node is not removed. Instead just handling running applications.
{code}
public void transition(RMNodeImpl rmNode, RMNodeEvent event) {
  RMNodeReconnectEvent reconnectEvent = (RMNodeReconnectEvent) event;
  RMNode newNode = reconnectEvent.getReconnectedNode();
  rmNode.nodeManagerVersion = newNode.getNodeManagerVersion();
  ListApplicationId runningApps = reconnectEvent.getRunningApplications();
  boolean noRunningApps = 
  (runningApps == null) || (runningApps.size() == 0);
  
  // No application running on the node, so send node-removal event with 
  // cleaning up old container info.
  if (noRunningApps) {
// Remove the node from scheduler
// Add node to the scheduler
  } else {
rmNode.httpPort = newNode.getHttpPort();
rmNode.httpAddress = newNode.getHttpAddress();
rmNode.totalCapability = newNode.getTotalCapability();
  
// Reset heartbeat ID since node just restarted.
rmNode.getLastNodeHeartBeatResponse().setResponseId(0);
  }

  // Handles running app on this node
// resource update to schedule code
  }
  
}
{code}

 After NM restart,completed containers are not released which are sent during 
 NM registration
 

 Key: YARN-3194
 URL: https://issues.apache.org/jira/browse/YARN-3194
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.6.0
 Environment: NM restart is enabled
Reporter: Rohith
Assignee: Rohith

 On NM restart ,NM sends all the outstanding NMContainerStatus to RM. But RM 
 process only ContainerState.RUNNING. If container is completed when NM was 
 down then those containers resources wont be release which result in 
 applications to hang.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3194) After NM restart,completed containers are not released which are sent during NM registration

2015-02-15 Thread Jian He (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14322154#comment-14322154
 ] 

Jian He commented on YARN-3194:
---

bq. I have one doubt on this method 
ResourceTrackerService#handleNMContainerStatus
This is legacy code for non-work-preserving restart. we could remove that. Just 
disregard this method.
bq. NM RESTART is Enabled – Problem is here
For node_reconnect event, it's removing the old node and adding the newly 
connected node. RM is also not restarted. I don't think we need to handle the 
RMNodeReconnectEvent


 After NM restart,completed containers are not released which are sent during 
 NM registration
 

 Key: YARN-3194
 URL: https://issues.apache.org/jira/browse/YARN-3194
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.6.0
 Environment: NM restart is enabled
Reporter: Rohith
Assignee: Rohith

 On NM restart ,NM sends all the outstanding NMContainerStatus to RM. But RM 
 process only ContainerState.RUNNING. If container is completed when NM was 
 down then those containers resources wont be release which result in 
 applications to hang.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3194) After NM restart,completed containers are not released which are sent during NM registration

2015-02-15 Thread Rohith (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14322144#comment-14322144
 ] 

Rohith commented on YARN-3194:
--

And Not related specific to this jira, I have one doubt on this method 
{{ResourceTrackerService#handleNMContainerStatus}} is sending 
container_finished event only to master container. Why other containers are not 
considered? I think it is made intentionally for optimization so if 
container_finished event would release other containers resources. Is this is 
the reason?

 After NM restart,completed containers are not released which are sent during 
 NM registration
 

 Key: YARN-3194
 URL: https://issues.apache.org/jira/browse/YARN-3194
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.6.0
 Environment: NM restart is enabled
Reporter: Rohith
Assignee: Rohith

 On NM restart ,NM sends all the outstanding NMContainerStatus to RM. But RM 
 process only ContainerState.RUNNING. If container is completed when NM was 
 down then those containers resources wont be release which result in 
 applications to hang.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3194) After NM restart,completed containers are not released which are sent during NM registration

2015-02-15 Thread Rohith (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3194?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14322143#comment-14322143
 ] 

Rohith commented on YARN-3194:
--

Thanks [~jianhe] for pointing me out container recovery flow!! Issue priority 
can decided later,not a problem.

I had deeper look about NM registration flow. There are 2 scenario's can occur
# Node added event : Again here 2 scenario's can occur
## New node is registering with different ip:port -- NOT A PROBLEM
## Old node is re-registering because of RESYNC command from RM when RM restart 
-- NOT A PROBLEM
# Node reconnected event : 
## Existing node is re-registering i.e RM treat it as reconnecting node when RM 
is not restarted
### NM RESTART NOT Enabled -- NOT A PROBLEM
### NM RESTART is Enabled -- {color:red}Problem is here{color}
When Node is reconnected and applications are running in that node, 
NMContainerStatus are ignored. I think RMNodeReconnectEvent should consider 
NMContainerStatus and process it.

 After NM restart,completed containers are not released which are sent during 
 NM registration
 

 Key: YARN-3194
 URL: https://issues.apache.org/jira/browse/YARN-3194
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.6.0
 Environment: NM restart is enabled
Reporter: Rohith
Assignee: Rohith

 On NM restart ,NM sends all the outstanding NMContainerStatus to RM. But RM 
 process only ContainerState.RUNNING. If container is completed when NM was 
 down then those containers resources wont be release which result in 
 applications to hang.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3110) Faulty link and state in ApplicationHistory when aplication is in unassigned state

2015-02-15 Thread Naganarasimha G R (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3110?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14321775#comment-14321775
 ] 

Naganarasimha G R commented on YARN-3110:
-

Hi [~ozawa], Can you please review this patch too ?

 Faulty link and state in ApplicationHistory when aplication is in unassigned 
 state
 --

 Key: YARN-3110
 URL: https://issues.apache.org/jira/browse/YARN-3110
 Project: Hadoop YARN
  Issue Type: Bug
  Components: applications, timelineserver
Affects Versions: 2.6.0
Reporter: Bibin A Chundatt
Assignee: Naganarasimha G R
Priority: Minor
 Attachments: YARN-3110.20150209-1.patch


 Application state and History link wrong when Application is in unassigned 
 state
  
 1.Configure capacity schedular with queue size as 1  also max Absolute Max 
 Capacity:  10.0%
 (Current application state is Accepted and Unassigned from resource manager 
 side)
 2.Submit application to queue and check the state and link in Application 
 history
 State= null and History link shown as N/A in applicationhistory page
 Kill the same application . In timeline server logs the below is show when 
 selecting application link.
 {quote}
 2015-01-29 15:39:50,956 ERROR org.apache.hadoop.yarn.webapp.View: Failed to 
 read the AM container of the application attempt 
 appattempt_1422467063659_0007_01.
 java.lang.NullPointerException
   at 
 org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryManagerOnTimelineStore.getContainer(ApplicationHistoryManagerOnTimelineStore.java:162)
   at 
 org.apache.hadoop.yarn.server.applicationhistoryservice.ApplicationHistoryManagerOnTimelineStore.getAMContainer(ApplicationHistoryManagerOnTimelineStore.java:184)
   at 
 org.apache.hadoop.yarn.server.webapp.AppBlock$3.run(AppBlock.java:160)
   at 
 org.apache.hadoop.yarn.server.webapp.AppBlock$3.run(AppBlock.java:157)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:415)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628)
   at 
 org.apache.hadoop.yarn.server.webapp.AppBlock.render(AppBlock.java:156)
   at 
 org.apache.hadoop.yarn.webapp.view.HtmlBlock.render(HtmlBlock.java:67)
   at 
 org.apache.hadoop.yarn.webapp.view.HtmlBlock.renderPartial(HtmlBlock.java:77)
   at org.apache.hadoop.yarn.webapp.View.render(View.java:235)
   at 
 org.apache.hadoop.yarn.webapp.view.HtmlPage$Page.subView(HtmlPage.java:49)
   at 
 org.apache.hadoop.yarn.webapp.hamlet.HamletImpl$EImp._v(HamletImpl.java:117)
   at org.apache.hadoop.yarn.webapp.hamlet.Hamlet$TD._(Hamlet.java:845)
   at 
 org.apache.hadoop.yarn.webapp.view.TwoColumnLayout.render(TwoColumnLayout.java:56)
   at org.apache.hadoop.yarn.webapp.view.HtmlPage.render(HtmlPage.java:82)
   at org.apache.hadoop.yarn.webapp.Controller.render(Controller.java:212)
   at 
 org.apache.hadoop.yarn.server.applicationhistoryservice.webapp.AHSController.app(AHSController.java:38)
   at sun.reflect.GeneratedMethodAccessor63.invoke(Unknown Source)
   at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
   at java.lang.reflect.Method.invoke(Method.java:606)
   at org.apache.hadoop.yarn.webapp.Dispatcher.service(Dispatcher.java:153)
   at javax.servlet.http.HttpServlet.service(HttpServlet.java:820)
   at 
 com.google.inject.servlet.ServletDefinition.doService(ServletDefinition.java:263)
   at 
 com.google.inject.servlet.ServletDefinition.service(ServletDefinition.java:178)
   at 
 com.google.inject.servlet.ManagedServletPipeline.service(ManagedServletPipeline.java:91)
   at 
 com.google.inject.servlet.FilterChainInvocation.doFilter(FilterChainInvocation.java:62)
   at 
 com.sun.jersey.spi.container.servlet.ServletContainer.doFilter(ServletContainer.java:900)
   at 
 com.sun.jersey.spi.container.servlet.ServletContainer.doFilter(ServletContainer.java:834)
   at 
 com.sun.jersey.spi.container.servlet.ServletContainer.doFilter(ServletContainer.java:795)
   at 
 com.google.inject.servlet.FilterDefinition.doFilter(FilterDefinition.java:163)
   at 
 com.google.inject.servlet.FilterChainInvocation.doFilter(FilterChainInvocation.java:58)
   at 
 com.google.inject.servlet.ManagedFilterPipeline.dispatch(ManagedFilterPipeline.java:118)
   at com.google.inject.servlet.GuiceFilter.doFilter(GuiceFilter.java:113)
   at 
 org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212)
   at 
 org.apache.hadoop.http.lib.StaticUserWebFilter$StaticUserFilter.doFilter(StaticUserWebFilter.java:109)
   at 
 

[jira] [Assigned] (YARN-3200) Factor OSType out from Shell: changes in YARN

2015-02-15 Thread Brahma Reddy Battula (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3200?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brahma Reddy Battula reassigned YARN-3200:
--

Assignee: Tsuyoshi OZAWA  (was: Brahma Reddy Battula)

Hi [~ozawa]

can you please assign to [~yzhangal]..

 Factor OSType out from Shell: changes in YARN
 -

 Key: YARN-3200
 URL: https://issues.apache.org/jira/browse/YARN-3200
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Yongjun Zhang
Assignee: Tsuyoshi OZAWA
 Fix For: 2.7.0






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (YARN-3201) add args for DistributedShell to specify a image for tasks that will run on docker

2015-02-15 Thread zhangwei (JIRA)
zhangwei created YARN-3201:
--

 Summary: add args for DistributedShell to specify a image for 
tasks that will run on docker
 Key: YARN-3201
 URL: https://issues.apache.org/jira/browse/YARN-3201
 Project: Hadoop YARN
  Issue Type: Wish
  Components: applications/distributed-shell
Reporter: zhangwei


It's very useful to execute a script on docker to do some test, but the 
distributedshell has no args to set the image.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (YARN-3195) [YARN]Missing uniformity In Yarn Queue CLI command

2015-02-15 Thread Devaraj K (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3195?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Devaraj K reassigned YARN-3195:
---

Assignee: Jagadesh Kiran N

 [YARN]Missing uniformity  In Yarn Queue CLI command
 ---

 Key: YARN-3195
 URL: https://issues.apache.org/jira/browse/YARN-3195
 Project: Hadoop YARN
  Issue Type: Bug
  Components: client
Affects Versions: 2.6.0
 Environment: SUSE Linux SP3
Reporter: Jagadesh Kiran N
Assignee: Jagadesh Kiran N
Priority: Minor
 Fix For: 2.7.0

 Attachments: Helptobe removed in Queue.png, YARN-3195.patch, 
 YARN-3195.patch


 Help is generic command should not be placed here because of this uniformity 
 is missing compared to other commands.Remove -help command inside ./yarn 
 queue as uniformity with respect to other commands 
 {code}
 SO486LDPag65:/home/OpenSource/HA/install/hadoop/resourcemanager/bin # ./yarn 
 queue -help
 15/02/13 19:30:20 WARN util.NativeCodeLoader: Unable to load native-hadoop 
 library for your platform... using builtin-java classes where applicable
 usage: queue
 * -help  Displays help for all commands.*
  -status Queue Name   List queue information about given queue.
 SO486LDPag65:/home/OpenSource/HA/install/hadoop/resourcemanager/bin # ./yarn 
 queue
 15/02/13 19:33:14 WARN util.NativeCodeLoader: Unable to load native-hadoop 
 library for your platform... using builtin-java classes where applicable
 Invalid Command Usage :
 usage: queue
 * -help  Displays help for all commands.*
  -status Queue Name   List queue information about given queue.
 {code}
 * -help  Displays help for all commands.*



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (YARN-3025) Provide API for retrieving blacklisted nodes

2015-02-15 Thread Ted Yu (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-3025?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ted Yu updated YARN-3025:
-
Attachment: yarn-3025-v2.txt

Patch v2 does what was proposed above.

Next step is to add getter for black listed nodes in ApplicationMasterProtocol

 Provide API for retrieving blacklisted nodes
 

 Key: YARN-3025
 URL: https://issues.apache.org/jira/browse/YARN-3025
 Project: Hadoop YARN
  Issue Type: Improvement
Reporter: Ted Yu
Assignee: Ted Yu
 Attachments: yarn-3025-v1.txt, yarn-3025-v2.txt


 We have the following method which updates blacklist:
 {code}
   public synchronized void updateBlacklist(ListString blacklistAdditions,
   ListString blacklistRemovals) {
 {code}
 Upon AM failover, there should be an API which returns the blacklisted nodes 
 so that the new AM can make consistent decisions.
 The new API can be:
 {code}
   public synchronized ListString getBlacklistedNodes()
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3076) YarnClient implementation to retrieve label to node mapping

2015-02-15 Thread Varun Saxena (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3076?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14322395#comment-14322395
 ] 

Varun Saxena commented on YARN-3076:


[~leftnoteasy], ok will update the cases and upload a new patch today.

 YarnClient implementation to retrieve label to node mapping
 ---

 Key: YARN-3076
 URL: https://issues.apache.org/jira/browse/YARN-3076
 Project: Hadoop YARN
  Issue Type: Sub-task
  Components: client
Affects Versions: 2.7.0
Reporter: Varun Saxena
Assignee: Varun Saxena
 Attachments: YARN-3076.001.patch, YARN-3076.002.patch






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-3168) Convert site documentation from apt to markdown

2015-02-15 Thread Gururaj Shetty (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-3168?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14322414#comment-14322414
 ] 

Gururaj Shetty commented on YARN-3168:
--

Thanks Allen Wittenauer for the patch. Do you convert the .apt to markdown 
files manually or do you use any tool to do the same?

 Convert site documentation from apt to markdown
 ---

 Key: YARN-3168
 URL: https://issues.apache.org/jira/browse/YARN-3168
 Project: Hadoop YARN
  Issue Type: Improvement
  Components: documentation
Affects Versions: 3.0.0
Reporter: Allen Wittenauer
Assignee: Gururaj Shetty
 Attachments: YARN-3168-00.patch


 YARN analog to HADOOP-11495



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (YARN-44) NM webapp returns 500 error when container log directory is missing

2015-02-15 Thread Devaraj K (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-44?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14322453#comment-14322453
 ] 

Devaraj K commented on YARN-44:
---

It doesn't seem to be a problem anymore with the changes happened. [~jlowe], 
can you confirm to close this issue?

 NM webapp returns 500 error when container log directory is missing
 ---

 Key: YARN-44
 URL: https://issues.apache.org/jira/browse/YARN-44
 Project: Hadoop YARN
  Issue Type: Bug
  Components: nodemanager
Affects Versions: 0.23.3
Reporter: Jason Lowe

 When log aggregation fails to create the application log directory it kills 
 the application.  This apparently can leave the container with no log 
 directory, as the user encounters a 500 error when trying to access the AM 
 logs via the webapp.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)