[jira] [Updated] (YARN-1919) Log yarn.resourcemanager.cluster-id is required for HA instead of throwing NPE

2014-04-09 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated YARN-1919:
---

Summary: Log yarn.resourcemanager.cluster-id is required for HA instead of 
throwing NPE  (was: NPE in Resource Manager if we don't set 
'yarn.resourcemanager.cluster-id' for HA)

> Log yarn.resourcemanager.cluster-id is required for HA instead of throwing NPE
> --
>
> Key: YARN-1919
> URL: https://issues.apache.org/jira/browse/YARN-1919
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Affects Versions: 2.3.0
>Reporter: Devaraj K
>Assignee: Karthik Kambatla
>Priority: Minor
>
> {code:xml}
> 2014-04-09 16:14:16,392 WARN org.apache.hadoop.service.AbstractService: When 
> stopping the service 
> org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService : 
> java.lang.NullPointerException
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService.serviceStop(EmbeddedElectorService.java:108)
>   at 
> org.apache.hadoop.service.AbstractService.stop(AbstractService.java:221)
>   at 
> org.apache.hadoop.service.ServiceOperations.stop(ServiceOperations.java:52)
>   at 
> org.apache.hadoop.service.ServiceOperations.stopQuietly(ServiceOperations.java:80)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:171)
>   at 
> org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.AdminService.serviceInit(AdminService.java:122)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>   at 
> org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.serviceInit(ResourceManager.java:232)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.main(ResourceManager.java:1038)
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Reopened] (YARN-1919) NPE in Resource Manager if we don't set 'yarn.resourcemanager.cluster-id' for HA

2014-04-09 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla reopened YARN-1919:


  Assignee: Karthik Kambatla

> NPE in Resource Manager if we don't set 'yarn.resourcemanager.cluster-id' for 
> HA
> 
>
> Key: YARN-1919
> URL: https://issues.apache.org/jira/browse/YARN-1919
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Affects Versions: 2.3.0
>Reporter: Devaraj K
>Assignee: Karthik Kambatla
>Priority: Minor
>
> {code:xml}
> 2014-04-09 16:14:16,392 WARN org.apache.hadoop.service.AbstractService: When 
> stopping the service 
> org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService : 
> java.lang.NullPointerException
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService.serviceStop(EmbeddedElectorService.java:108)
>   at 
> org.apache.hadoop.service.AbstractService.stop(AbstractService.java:221)
>   at 
> org.apache.hadoop.service.ServiceOperations.stop(ServiceOperations.java:52)
>   at 
> org.apache.hadoop.service.ServiceOperations.stopQuietly(ServiceOperations.java:80)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:171)
>   at 
> org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.AdminService.serviceInit(AdminService.java:122)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>   at 
> org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.serviceInit(ResourceManager.java:232)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.main(ResourceManager.java:1038)
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1919) NPE in Resource Manager if we don't set 'yarn.resourcemanager.cluster-id' for HA

2014-04-09 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964982#comment-13964982
 ] 

Karthik Kambatla commented on YARN-1919:


Indeed, that makes more sense.

> NPE in Resource Manager if we don't set 'yarn.resourcemanager.cluster-id' for 
> HA
> 
>
> Key: YARN-1919
> URL: https://issues.apache.org/jira/browse/YARN-1919
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Affects Versions: 2.3.0
>Reporter: Devaraj K
>Priority: Minor
>
> {code:xml}
> 2014-04-09 16:14:16,392 WARN org.apache.hadoop.service.AbstractService: When 
> stopping the service 
> org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService : 
> java.lang.NullPointerException
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService.serviceStop(EmbeddedElectorService.java:108)
>   at 
> org.apache.hadoop.service.AbstractService.stop(AbstractService.java:221)
>   at 
> org.apache.hadoop.service.ServiceOperations.stop(ServiceOperations.java:52)
>   at 
> org.apache.hadoop.service.ServiceOperations.stopQuietly(ServiceOperations.java:80)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:171)
>   at 
> org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.AdminService.serviceInit(AdminService.java:122)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>   at 
> org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.serviceInit(ResourceManager.java:232)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.main(ResourceManager.java:1038)
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1919) NPE in Resource Manager if we don't set 'yarn.resourcemanager.cluster-id' for HA

2014-04-09 Thread Jian He (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1919?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964972#comment-13964972
 ] 

Jian He commented on YARN-1919:
---

Does it make sense to print a help message instead of throwing NPE ?

> NPE in Resource Manager if we don't set 'yarn.resourcemanager.cluster-id' for 
> HA
> 
>
> Key: YARN-1919
> URL: https://issues.apache.org/jira/browse/YARN-1919
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Affects Versions: 2.3.0
>Reporter: Devaraj K
>Priority: Minor
>
> {code:xml}
> 2014-04-09 16:14:16,392 WARN org.apache.hadoop.service.AbstractService: When 
> stopping the service 
> org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService : 
> java.lang.NullPointerException
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService.serviceStop(EmbeddedElectorService.java:108)
>   at 
> org.apache.hadoop.service.AbstractService.stop(AbstractService.java:221)
>   at 
> org.apache.hadoop.service.ServiceOperations.stop(ServiceOperations.java:52)
>   at 
> org.apache.hadoop.service.ServiceOperations.stopQuietly(ServiceOperations.java:80)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:171)
>   at 
> org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.AdminService.serviceInit(AdminService.java:122)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>   at 
> org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.serviceInit(ResourceManager.java:232)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.main(ResourceManager.java:1038)
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1920) TestFileSystemApplicationHistoryStore.testMissingApplicationAttemptHistoryData fails in windows

2014-04-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1920?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964921#comment-13964921
 ] 

Hadoop QA commented on YARN-1920:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12639515/YARN-1920.txt
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-applicationhistoryservice.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3541//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3541//console

This message is automatically generated.

> TestFileSystemApplicationHistoryStore.testMissingApplicationAttemptHistoryData
>  fails in windows
> ---
>
> Key: YARN-1920
> URL: https://issues.apache.org/jira/browse/YARN-1920
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Vinod Kumar Vavilapalli
>Assignee: Vinod Kumar Vavilapalli
> Attachments: YARN-1920.txt
>
>
> Though this was only failing in Windows, after debugging, I realized that the 
> test fails because we are leaking a file-handle in the history service.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1914) Test TestFSDownload.testDownloadPublicWithStatCache fails on Windows

2014-04-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1914?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964914#comment-13964914
 ] 

Hadoop QA commented on YARN-1914:
-

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12639511/apache-yarn-1914.1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3540//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3540//console

This message is automatically generated.

> Test TestFSDownload.testDownloadPublicWithStatCache fails on Windows
> 
>
> Key: YARN-1914
> URL: https://issues.apache.org/jira/browse/YARN-1914
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Varun Vasudev
>Assignee: Varun Vasudev
> Attachments: apache-yarn-1914.0.patch, apache-yarn-1914.1.patch
>
>
> The TestFSDownload.testDownloadPublicWithStatCache test in hadoop-yarn-common 
> consistently fails on Windows environments.
> The root cause is that the test checks for execute permission for all users 
> on every ancestor of the target directory. In windows, by default, group 
> "Everyone" has no permissions on any directory in the install drive. It's 
> unreasonable to expect this test to pass and we should skip it on Windows.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-996) REST API support for node resource configuration

2014-04-09 Thread Junping Du (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junping Du updated YARN-996:


Assignee: Kenji Kikushima  (was: Junping Du)

> REST API support for node resource configuration
> 
>
> Key: YARN-996
> URL: https://issues.apache.org/jira/browse/YARN-996
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager, scheduler
>Reporter: Junping Du
>Assignee: Kenji Kikushima
> Attachments: YARN-996-sample.patch
>
>
> Besides admin protocol and CLI, REST API should also be supported for node 
> resource configuration



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (YARN-996) REST API support for node resource configuration

2014-04-09 Thread Junping Du (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Junping Du reassigned YARN-996:
---

Assignee: Junping Du

> REST API support for node resource configuration
> 
>
> Key: YARN-996
> URL: https://issues.apache.org/jira/browse/YARN-996
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager, scheduler
>Reporter: Junping Du
>Assignee: Junping Du
> Attachments: YARN-996-sample.patch
>
>
> Besides admin protocol and CLI, REST API should also be supported for node 
> resource configuration



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-996) REST API support for node resource configuration

2014-04-09 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-996?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964906#comment-13964906
 ] 

Junping Du commented on YARN-996:
-

Hi [~kj-ki], Thanks for delivering a patch for this. I already assign this JIRA 
to you and will review your patch later. Thx!

> REST API support for node resource configuration
> 
>
> Key: YARN-996
> URL: https://issues.apache.org/jira/browse/YARN-996
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager, scheduler
>Reporter: Junping Du
>Assignee: Kenji Kikushima
> Attachments: YARN-996-sample.patch
>
>
> Besides admin protocol and CLI, REST API should also be supported for node 
> resource configuration



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-1920) TestFileSystemApplicationHistoryStore.testMissingApplicationAttemptHistoryData fails in windows

2014-04-09 Thread Vinod Kumar Vavilapalli (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1920?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinod Kumar Vavilapalli updated YARN-1920:
--

Attachment: YARN-1920.txt

Here's a patch that fixes the leak and passes the test on Windows.

> TestFileSystemApplicationHistoryStore.testMissingApplicationAttemptHistoryData
>  fails in windows
> ---
>
> Key: YARN-1920
> URL: https://issues.apache.org/jira/browse/YARN-1920
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Vinod Kumar Vavilapalli
>Assignee: Vinod Kumar Vavilapalli
> Attachments: YARN-1920.txt
>
>
> Though this was only failing in Windows, after debugging, I realized that the 
> test fails because we are leaking a file-handle in the history service.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (YARN-1920) TestFileSystemApplicationHistoryStore.testMissingApplicationAttemptHistoryData fails in windows

2014-04-09 Thread Vinod Kumar Vavilapalli (JIRA)
Vinod Kumar Vavilapalli created YARN-1920:
-

 Summary: 
TestFileSystemApplicationHistoryStore.testMissingApplicationAttemptHistoryData 
fails in windows
 Key: YARN-1920
 URL: https://issues.apache.org/jira/browse/YARN-1920
 Project: Hadoop YARN
  Issue Type: Bug
Reporter: Vinod Kumar Vavilapalli
Assignee: Vinod Kumar Vavilapalli


Though this was only failing in Windows, after debugging, I realized that the 
test fails because we are leaking a file-handle in the history service.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-1914) Test TestFSDownload.testDownloadPublicWithStatCache fails on Windows

2014-04-09 Thread Varun Vasudev (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1914?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Varun Vasudev updated YARN-1914:


Attachment: apache-yarn-1914.1.patch

Attached new patch fixing root cause in FSDownload.java. The patch should fail 
the HadoopQA because it fixes a currently failing test on Windows. Thank you 
[~bikassaha] and [~ivanmi].

> Test TestFSDownload.testDownloadPublicWithStatCache fails on Windows
> 
>
> Key: YARN-1914
> URL: https://issues.apache.org/jira/browse/YARN-1914
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Varun Vasudev
>Assignee: Varun Vasudev
> Attachments: apache-yarn-1914.0.patch, apache-yarn-1914.1.patch
>
>
> The TestFSDownload.testDownloadPublicWithStatCache test in hadoop-yarn-common 
> consistently fails on Windows environments.
> The root cause is that the test checks for execute permission for all users 
> on every ancestor of the target directory. In windows, by default, group 
> "Everyone" has no permissions on any directory in the install drive. It's 
> unreasonable to expect this test to pass and we should skip it on Windows.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1910) TestAMRMTokens fails on windows

2014-04-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1910?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964887#comment-13964887
 ] 

Hudson commented on YARN-1910:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5485 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5485/])
YARN-1910. Fixed a race condition in TestAMRMTokens that causes the test to 
fail more often on Windows. Contributed by Xuan Gong. (vinodkv: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1586192)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/security/TestAMRMTokens.java


> TestAMRMTokens fails on windows
> ---
>
> Key: YARN-1910
> URL: https://issues.apache.org/jira/browse/YARN-1910
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Xuan Gong
>Assignee: Xuan Gong
> Fix For: 2.4.0
>
> Attachments: YARN-1910.1.patch, YARN-1910.2.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1910) TestAMRMTokens fails on windows

2014-04-09 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1910?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964882#comment-13964882
 ] 

Vinod Kumar Vavilapalli commented on YARN-1910:
---

Alright, looks good to me too. +1. The test fails without the patch and passes 
with on Windows. It passes on Linux before and after the patch. Checking this 
in.

> TestAMRMTokens fails on windows
> ---
>
> Key: YARN-1910
> URL: https://issues.apache.org/jira/browse/YARN-1910
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Xuan Gong
>Assignee: Xuan Gong
> Fix For: 2.4.0
>
> Attachments: YARN-1910.1.patch, YARN-1910.2.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-996) REST API support for node resource configuration

2014-04-09 Thread Kenji Kikushima (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-996?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kenji Kikushima updated YARN-996:
-

Attachment: YARN-996-sample.patch

Hi, I'm trying to implement REST API. Here is a sample patch.
- We can get Resource infomation from getting "/ws/v1/cluster/nodes/{nodeid}"
- We can update Resources. For example, putting 
"/ws/v1/cluster/nodes/{nodeid}/vcores/10" means updates to 10 vcores
- At present, Resource Manager REST API only
- This patch needs YARN-1911.patch in YARN-1911 to avoid NPE
Please comment if you have interest. Thanks.

> REST API support for node resource configuration
> 
>
> Key: YARN-996
> URL: https://issues.apache.org/jira/browse/YARN-996
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager, scheduler
>Reporter: Junping Du
> Attachments: YARN-996-sample.patch
>
>
> Besides admin protocol and CLI, REST API should also be supported for node 
> resource configuration



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Resolved] (YARN-1919) NPE in Resource Manager if we don't set 'yarn.resourcemanager.cluster-id' for HA

2014-04-09 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1919?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla resolved YARN-1919.


Resolution: Not a Problem

I understand this is an inconvenience, but this is by design. One of those 
bug-or-feature cases. See 
https://issues.apache.org/jira/browse/YARN-1029?focusedCommentId=13861990&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13861990

> NPE in Resource Manager if we don't set 'yarn.resourcemanager.cluster-id' for 
> HA
> 
>
> Key: YARN-1919
> URL: https://issues.apache.org/jira/browse/YARN-1919
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Affects Versions: 2.3.0
>Reporter: Devaraj K
>Priority: Minor
>
> {code:xml}
> 2014-04-09 16:14:16,392 WARN org.apache.hadoop.service.AbstractService: When 
> stopping the service 
> org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService : 
> java.lang.NullPointerException
> java.lang.NullPointerException
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService.serviceStop(EmbeddedElectorService.java:108)
>   at 
> org.apache.hadoop.service.AbstractService.stop(AbstractService.java:221)
>   at 
> org.apache.hadoop.service.ServiceOperations.stop(ServiceOperations.java:52)
>   at 
> org.apache.hadoop.service.ServiceOperations.stopQuietly(ServiceOperations.java:80)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:171)
>   at 
> org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.AdminService.serviceInit(AdminService.java:122)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>   at 
> org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.serviceInit(ResourceManager.java:232)
>   at 
> org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
>   at 
> org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.main(ResourceManager.java:1038)
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1914) Test TestFSDownload.testDownloadPublicWithStatCache fails on Windows

2014-04-09 Thread Ivan Mitic (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1914?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964326#comment-13964326
 ] 

Ivan Mitic commented on YARN-1914:
--

Thanks Bikas for looping us in. Varun, HADOOP-8731 might be able to help. 

> Test TestFSDownload.testDownloadPublicWithStatCache fails on Windows
> 
>
> Key: YARN-1914
> URL: https://issues.apache.org/jira/browse/YARN-1914
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Varun Vasudev
>Assignee: Varun Vasudev
> Attachments: apache-yarn-1914.0.patch
>
>
> The TestFSDownload.testDownloadPublicWithStatCache test in hadoop-yarn-common 
> consistently fails on Windows environments.
> The root cause is that the test checks for execute permission for all users 
> on every ancestor of the target directory. In windows, by default, group 
> "Everyone" has no permissions on any directory in the install drive. It's 
> unreasonable to expect this test to pass and we should skip it on Windows.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1338) Recover localized resource cache state upon nodemanager restart

2014-04-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1338?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964313#comment-13964313
 ] 

Hadoop QA commented on YARN-1338:
-

{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12639419/YARN-1338v2.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 15 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 1.3.9) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-YARN-Build/3539//testReport/
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/3539//console

This message is automatically generated.

> Recover localized resource cache state upon nodemanager restart
> ---
>
> Key: YARN-1338
> URL: https://issues.apache.org/jira/browse/YARN-1338
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager
>Affects Versions: 2.3.0
>Reporter: Jason Lowe
>Assignee: Jason Lowe
> Attachments: YARN-1338.patch, YARN-1338v2.patch
>
>
> Today when node manager restarts we clean up all the distributed cache files 
> from disk. This is definitely not ideal from 2 aspects.
> * For work preserving restart we definitely want them as running containers 
> are using them
> * For even non work preserving restart this will be useful in the sense that 
> we don't have to download them again if needed by future tasks.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1907) TestRMApplicationHistoryWriter#testRMWritingMassiveHistory runs slow and intermittently fails

2014-04-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964294#comment-13964294
 ] 

Hudson commented on YARN-1907:
--

SUCCESS: Integrated in Hadoop-trunk-Commit #5475 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5475/])
YARN-1907. TestRMApplicationHistoryWriter#testRMWritingMassiveHistory 
intermittently fails. Contributed by Mit Desai. (kihwal: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1585992)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/ahs/TestRMApplicationHistoryWriter.java


> TestRMApplicationHistoryWriter#testRMWritingMassiveHistory runs slow and 
> intermittently fails
> -
>
> Key: YARN-1907
> URL: https://issues.apache.org/jira/browse/YARN-1907
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.5.0
>Reporter: Mit Desai
>Assignee: Mit Desai
> Fix For: 3.0.0, 2.5.0
>
> Attachments: HDFS-6195.patch
>
>
> The test has 1 containers that it tries to cleanup.
> The cleanup has a timeout of 2ms in which the test sometimes cannot do 
> the cleanup completely and gives out an Assertion Failure.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1907) TestRMApplicationHistoryWriter#testRMWritingMassiveHistory runs slow and intermittently fails

2014-04-09 Thread Kihwal Lee (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1907?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964284#comment-13964284
 ] 

Kihwal Lee commented on YARN-1907:
--

+1 for the change.

bq. However, is it a better code practice to loop until all the containers are 
cleaned up (removing the 200 round bounds), and set a suitable timeout for this 
test case?
The unit tests do not always run in a controlled environment, so timing can 
vary widely. Tight timeout or timing-based performance checks are not 
recommended in unit tests.

> TestRMApplicationHistoryWriter#testRMWritingMassiveHistory runs slow and 
> intermittently fails
> -
>
> Key: YARN-1907
> URL: https://issues.apache.org/jira/browse/YARN-1907
> Project: Hadoop YARN
>  Issue Type: Bug
>Affects Versions: 3.0.0, 2.5.0
>Reporter: Mit Desai
>Assignee: Mit Desai
> Attachments: HDFS-6195.patch
>
>
> The test has 1 containers that it tries to cleanup.
> The cleanup has a timeout of 2ms in which the test sometimes cannot do 
> the cleanup completely and gives out an Assertion Failure.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-1338) Recover localized resource cache state upon nodemanager restart

2014-04-09 Thread Jason Lowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1338?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Lowe updated YARN-1338:
-

Attachment: YARN-1338v2.patch

Updating patch after YARN-1757 and other recent changes on trunk.

> Recover localized resource cache state upon nodemanager restart
> ---
>
> Key: YARN-1338
> URL: https://issues.apache.org/jira/browse/YARN-1338
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager
>Affects Versions: 2.3.0
>Reporter: Jason Lowe
>Assignee: Jason Lowe
> Attachments: YARN-1338.patch, YARN-1338v2.patch
>
>
> Today when node manager restarts we clean up all the distributed cache files 
> from disk. This is definitely not ideal from 2 aspects.
> * For work preserving restart we definitely want them as running containers 
> are using them
> * For even non work preserving restart this will be useful in the sense that 
> we don't have to download them again if needed by future tasks.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1912) ResourceLocalizer started without any jvm memory control

2014-04-09 Thread Nathan Roberts (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1912?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964198#comment-13964198
 ] 

Nathan Roberts commented on YARN-1912:
--

Ah. Thanks Stanley for the pointer and clarification.  

> ResourceLocalizer started without any jvm memory control
> 
>
> Key: YARN-1912
> URL: https://issues.apache.org/jira/browse/YARN-1912
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: nodemanager
>Affects Versions: 2.2.0
>Reporter: stanley shi
>
> In the LinuxContainerExecutor.java#startLocalizer, it does not specify any 
> "-Xmx" configurations in the command, this caused the ResourceLocalizer to be 
> started with default memory setting.
> In an server-level hardware, it will use 25% of the system memory as the max 
> heap size, this will cause memory issue in some cases.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1908) Distributed shell with custom script has permission error.

2014-04-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1908?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964168#comment-13964168
 ] 

Hudson commented on YARN-1908:
--

SUCCESS: Integrated in Hadoop-Hdfs-trunk #1726 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1726/])
YARN-1908. Fixed DistributedShell to not fail in secure clusters. Contributed 
by Vinod Kumar Vavilapalli and Jian He. (vinodkv: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1585849)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/src/main/java/org/apache/hadoop/yarn/applications/distributedshell/ApplicationMaster.java


> Distributed shell with custom script has permission error.
> --
>
> Key: YARN-1908
> URL: https://issues.apache.org/jira/browse/YARN-1908
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: applications/distributed-shell
>Affects Versions: 2.4.0
>Reporter: Tassapol Athiapinya
>Assignee: Vinod Kumar Vavilapalli
> Fix For: 2.4.1
>
> Attachments: YARN-1908.1.patch, YARN-1908.2.patch, YARN-1908.3.patch, 
> YARN-1908.4.patch
>
>
> Create test1.sh having "pwd".
> Run this command as user1:
> hadoop jar /usr/lib/hadoop-yarn/hadoop-yarn-applications-distributedshell.jar 
> -jar /usr/lib/hadoop-yarn/hadoop-yarn-applications-distributedshell.jar 
> -shell_script test1.sh
> NM is run by yarn user. An exception is thrown because yarn user has no 
> permissions on custom script in hdfs path. The custom script is created with 
> distributed shell app.
> {code}
> Caused by: 
> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException):
>  Permission denied: user=yarn, access=WRITE, 
> inode="/user/user1/DistributedShell/70":user1:user1:drwxr-xr-x
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1757) NM Recovery. Auxiliary service support.

2014-04-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964164#comment-13964164
 ] 

Hudson commented on YARN-1757:
--

SUCCESS: Integrated in Hadoop-Hdfs-trunk #1726 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1726/])
YARN-1757. NM Recovery. Auxiliary service support. (Jason Lowe via kasha) 
(kasha: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1585783)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/server/api/AuxiliaryService.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/resources/yarn-default.xml
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/NodeManager.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/AuxServices.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/TestAuxServices.java


> NM Recovery. Auxiliary service support.
> ---
>
> Key: YARN-1757
> URL: https://issues.apache.org/jira/browse/YARN-1757
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager
>Affects Versions: 2.3.0
>Reporter: Jason Lowe
>Assignee: Jason Lowe
> Fix For: 2.5.0
>
> Attachments: YARN-1757-v2.patch, YARN-1757.patch, YARN-1757.patch
>
>
> There needs to be a mechanism for communicating to auxiliary services whether 
> nodemanager recovery is enabled and where they should store their state.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1784) TestContainerAllocation assumes CapacityScheduler

2014-04-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964169#comment-13964169
 ] 

Hudson commented on YARN-1784:
--

SUCCESS: Integrated in Hadoop-Hdfs-trunk #1726 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1726/])
YARN-1784. TestContainerAllocation assumes CapacityScheduler. (Robert Kanter 
via kasha) (kasha: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1585894)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestContainerAllocation.java


> TestContainerAllocation assumes CapacityScheduler
> -
>
> Key: YARN-1784
> URL: https://issues.apache.org/jira/browse/YARN-1784
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Affects Versions: 2.3.0
>Reporter: Karthik Kambatla
>Assignee: Robert Kanter
>Priority: Minor
> Fix For: 2.5.0
>
> Attachments: YARN-1784.patch, YARN-1784.patch
>
>
> TestContainerAllocation assumes CapacityScheduler



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1784) TestContainerAllocation assumes CapacityScheduler

2014-04-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964139#comment-13964139
 ] 

Hudson commented on YARN-1784:
--

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1752 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1752/])
YARN-1784. TestContainerAllocation assumes CapacityScheduler. (Robert Kanter 
via kasha) (kasha: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1585894)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestContainerAllocation.java


> TestContainerAllocation assumes CapacityScheduler
> -
>
> Key: YARN-1784
> URL: https://issues.apache.org/jira/browse/YARN-1784
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Affects Versions: 2.3.0
>Reporter: Karthik Kambatla
>Assignee: Robert Kanter
>Priority: Minor
> Fix For: 2.5.0
>
> Attachments: YARN-1784.patch, YARN-1784.patch
>
>
> TestContainerAllocation assumes CapacityScheduler



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1757) NM Recovery. Auxiliary service support.

2014-04-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964135#comment-13964135
 ] 

Hudson commented on YARN-1757:
--

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1752 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1752/])
YARN-1757. NM Recovery. Auxiliary service support. (Jason Lowe via kasha) 
(kasha: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1585783)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/server/api/AuxiliaryService.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/resources/yarn-default.xml
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/NodeManager.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/AuxServices.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/TestAuxServices.java


> NM Recovery. Auxiliary service support.
> ---
>
> Key: YARN-1757
> URL: https://issues.apache.org/jira/browse/YARN-1757
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager
>Affects Versions: 2.3.0
>Reporter: Jason Lowe
>Assignee: Jason Lowe
> Fix For: 2.5.0
>
> Attachments: YARN-1757-v2.patch, YARN-1757.patch, YARN-1757.patch
>
>
> There needs to be a mechanism for communicating to auxiliary services whether 
> nodemanager recovery is enabled and where they should store their state.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1908) Distributed shell with custom script has permission error.

2014-04-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1908?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964138#comment-13964138
 ] 

Hudson commented on YARN-1908:
--

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1752 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1752/])
YARN-1908. Fixed DistributedShell to not fail in secure clusters. Contributed 
by Vinod Kumar Vavilapalli and Jian He. (vinodkv: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1585849)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/src/main/java/org/apache/hadoop/yarn/applications/distributedshell/ApplicationMaster.java


> Distributed shell with custom script has permission error.
> --
>
> Key: YARN-1908
> URL: https://issues.apache.org/jira/browse/YARN-1908
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: applications/distributed-shell
>Affects Versions: 2.4.0
>Reporter: Tassapol Athiapinya
>Assignee: Vinod Kumar Vavilapalli
> Fix For: 2.4.1
>
> Attachments: YARN-1908.1.patch, YARN-1908.2.patch, YARN-1908.3.patch, 
> YARN-1908.4.patch
>
>
> Create test1.sh having "pwd".
> Run this command as user1:
> hadoop jar /usr/lib/hadoop-yarn/hadoop-yarn-applications-distributedshell.jar 
> -jar /usr/lib/hadoop-yarn/hadoop-yarn-applications-distributedshell.jar 
> -shell_script test1.sh
> NM is run by yarn user. An exception is thrown because yarn user has no 
> permissions on custom script in hdfs path. The custom script is created with 
> distributed shell app.
> {code}
> Caused by: 
> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException):
>  Permission denied: user=yarn, access=WRITE, 
> inode="/user/user1/DistributedShell/70":user1:user1:drwxr-xr-x
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1784) TestContainerAllocation assumes CapacityScheduler

2014-04-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964034#comment-13964034
 ] 

Hudson commented on YARN-1784:
--

SUCCESS: Integrated in Hadoop-Yarn-trunk #534 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/534/])
YARN-1784. TestContainerAllocation assumes CapacityScheduler. (Robert Kanter 
via kasha) (kasha: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1585894)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/capacity/TestContainerAllocation.java


> TestContainerAllocation assumes CapacityScheduler
> -
>
> Key: YARN-1784
> URL: https://issues.apache.org/jira/browse/YARN-1784
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Affects Versions: 2.3.0
>Reporter: Karthik Kambatla
>Assignee: Robert Kanter
>Priority: Minor
> Fix For: 2.5.0
>
> Attachments: YARN-1784.patch, YARN-1784.patch
>
>
> TestContainerAllocation assumes CapacityScheduler



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1757) NM Recovery. Auxiliary service support.

2014-04-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1757?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964031#comment-13964031
 ] 

Hudson commented on YARN-1757:
--

SUCCESS: Integrated in Hadoop-Yarn-trunk #534 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/534/])
YARN-1757. NM Recovery. Auxiliary service support. (Jason Lowe via kasha) 
(kasha: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1585783)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/conf/YarnConfiguration.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/server/api/AuxiliaryService.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common/src/main/resources/yarn-default.xml
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/NodeManager.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/AuxServices.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/TestAuxServices.java


> NM Recovery. Auxiliary service support.
> ---
>
> Key: YARN-1757
> URL: https://issues.apache.org/jira/browse/YARN-1757
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: nodemanager
>Affects Versions: 2.3.0
>Reporter: Jason Lowe
>Assignee: Jason Lowe
> Fix For: 2.5.0
>
> Attachments: YARN-1757-v2.patch, YARN-1757.patch, YARN-1757.patch
>
>
> There needs to be a mechanism for communicating to auxiliary services whether 
> nodemanager recovery is enabled and where they should store their state.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1908) Distributed shell with custom script has permission error.

2014-04-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1908?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13964033#comment-13964033
 ] 

Hudson commented on YARN-1908:
--

SUCCESS: Integrated in Hadoop-Yarn-trunk #534 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/534/])
YARN-1908. Fixed DistributedShell to not fail in secure clusters. Contributed 
by Vinod Kumar Vavilapalli and Jian He. (vinodkv: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1585849)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell/src/main/java/org/apache/hadoop/yarn/applications/distributedshell/ApplicationMaster.java


> Distributed shell with custom script has permission error.
> --
>
> Key: YARN-1908
> URL: https://issues.apache.org/jira/browse/YARN-1908
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: applications/distributed-shell
>Affects Versions: 2.4.0
>Reporter: Tassapol Athiapinya
>Assignee: Vinod Kumar Vavilapalli
> Fix For: 2.4.1
>
> Attachments: YARN-1908.1.patch, YARN-1908.2.patch, YARN-1908.3.patch, 
> YARN-1908.4.patch
>
>
> Create test1.sh having "pwd".
> Run this command as user1:
> hadoop jar /usr/lib/hadoop-yarn/hadoop-yarn-applications-distributedshell.jar 
> -jar /usr/lib/hadoop-yarn/hadoop-yarn-applications-distributedshell.jar 
> -shell_script test1.sh
> NM is run by yarn user. An exception is thrown because yarn user has no 
> permissions on custom script in hdfs path. The custom script is created with 
> distributed shell app.
> {code}
> Caused by: 
> org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.security.AccessControlException):
>  Permission denied: user=yarn, access=WRITE, 
> inode="/user/user1/DistributedShell/70":user1:user1:drwxr-xr-x
>   at 
> org.apache.hadoop.hdfs.server.namenode.FSPermissionChecker.checkFsPermission(FSPermissionChecker.java:265)
> {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (YARN-1919) NPE in Resource Manager if we don't set 'yarn.resourcemanager.cluster-id' for HA

2014-04-09 Thread Devaraj K (JIRA)
Devaraj K created YARN-1919:
---

 Summary: NPE in Resource Manager if we don't set 
'yarn.resourcemanager.cluster-id' for HA
 Key: YARN-1919
 URL: https://issues.apache.org/jira/browse/YARN-1919
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.3.0
Reporter: Devaraj K
Priority: Minor


{code:xml}
2014-04-09 16:14:16,392 WARN org.apache.hadoop.service.AbstractService: When 
stopping the service 
org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService : 
java.lang.NullPointerException
java.lang.NullPointerException
at 
org.apache.hadoop.yarn.server.resourcemanager.EmbeddedElectorService.serviceStop(EmbeddedElectorService.java:108)
at 
org.apache.hadoop.service.AbstractService.stop(AbstractService.java:221)
at 
org.apache.hadoop.service.ServiceOperations.stop(ServiceOperations.java:52)
at 
org.apache.hadoop.service.ServiceOperations.stopQuietly(ServiceOperations.java:80)
at 
org.apache.hadoop.service.AbstractService.init(AbstractService.java:171)
at 
org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
at 
org.apache.hadoop.yarn.server.resourcemanager.AdminService.serviceInit(AdminService.java:122)
at 
org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
at 
org.apache.hadoop.service.CompositeService.serviceInit(CompositeService.java:107)
at 
org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.serviceInit(ResourceManager.java:232)
at 
org.apache.hadoop.service.AbstractService.init(AbstractService.java:163)
at 
org.apache.hadoop.yarn.server.resourcemanager.ResourceManager.main(ResourceManager.java:1038)
{code}




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (YARN-1918) Typo in description and error message for 'yarn.resourcemanager.cluster-id'

2014-04-09 Thread Devaraj K (JIRA)
Devaraj K created YARN-1918:
---

 Summary: Typo in description and error message for 
'yarn.resourcemanager.cluster-id'
 Key: YARN-1918
 URL: https://issues.apache.org/jira/browse/YARN-1918
 Project: Hadoop YARN
  Issue Type: Improvement
Affects Versions: 2.3.0
Reporter: Devaraj K
Priority: Trivial


1.  In yarn-default.xml
{code:xml}

Name of the cluster. In a HA setting,
  this is used to ensure the RM participates in leader
  election fo this cluster and ensures it does not affect
  other clusters
yarn.resourcemanager.cluster-id

  
{code}
Here the line 'election fo this cluster and ensures it does not affect' should 
be replaced with  'election for this cluster and ensures it does not affect'.

2. 
{code:xml}
org.apache.hadoop.HadoopIllegalArgumentException: Configuration doesn't 
specifyyarn.resourcemanager.cluster-id
at 
org.apache.hadoop.yarn.conf.YarnConfiguration.getClusterId(YarnConfiguration.java:1336)
{code}
In the above exception message, it is missing a space between message and 
configuration name.




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (YARN-1917) Add "waitForCompletion" interface to YarnClient

2014-04-09 Thread Wangda Tan (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1917?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13963959#comment-13963959
 ] 

Wangda Tan commented on YARN-1917:
--

My basic thinking is we can leverage existing implementation in 
UnmanagedAMLauncher.monitorApplication, it should be enough for many use cases. 
The method could be, YarnClient.waitForCompletion(ApplicationId appId). 
Appreciate your comments :)

> Add "waitForCompletion" interface to YarnClient
> ---
>
> Key: YARN-1917
> URL: https://issues.apache.org/jira/browse/YARN-1917
> Project: Hadoop YARN
>  Issue Type: New Feature
>  Components: client
>Affects Versions: 2.4.0
>Reporter: Wangda Tan
>
> Currently, YARN dosen't have this method. Users needs to write 
> implementations like UnmanagedAMLauncher.monitorApplication or 
> mapreduce.Job.monitorAndPrintJob on their own. This feature should be helpful 
> to end users.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (YARN-1917) Add "waitForCompletion" interface to YarnClient

2014-04-09 Thread Wangda Tan (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1917?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wangda Tan updated YARN-1917:
-

Priority: Minor  (was: Major)

> Add "waitForCompletion" interface to YarnClient
> ---
>
> Key: YARN-1917
> URL: https://issues.apache.org/jira/browse/YARN-1917
> Project: Hadoop YARN
>  Issue Type: New Feature
>  Components: client
>Affects Versions: 2.4.0
>Reporter: Wangda Tan
>Priority: Minor
>
> Currently, YARN dosen't have this method. Users needs to write 
> implementations like UnmanagedAMLauncher.monitorApplication or 
> mapreduce.Job.monitorAndPrintJob on their own. This feature should be helpful 
> to end users.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (YARN-1917) Add "waitForCompletion" interface to YarnClient

2014-04-09 Thread Wangda Tan (JIRA)
Wangda Tan created YARN-1917:


 Summary: Add "waitForCompletion" interface to YarnClient
 Key: YARN-1917
 URL: https://issues.apache.org/jira/browse/YARN-1917
 Project: Hadoop YARN
  Issue Type: New Feature
  Components: client
Affects Versions: 2.4.0
Reporter: Wangda Tan


Currently, YARN dosen't have this method. Users needs to write implementations 
like UnmanagedAMLauncher.monitorApplication or mapreduce.Job.monitorAndPrintJob 
on their own. This feature should be helpful to end users.




--
This message was sent by Atlassian JIRA
(v6.2#6252)