[jira] [Commented] (YARN-1074) Clean up YARN CLI app list to show only running apps.

2013-08-22 Thread Xuan Gong (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1074?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13748241#comment-13748241
 ] 

Xuan Gong commented on YARN-1074:
-

New patch addresses all the comments

> Clean up YARN CLI app list to show only running apps.
> -
>
> Key: YARN-1074
> URL: https://issues.apache.org/jira/browse/YARN-1074
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: client
>Reporter: Tassapol Athiapinya
>Assignee: Xuan Gong
> Attachments: YARN-1074.1.patch, YARN-1074.2.patch, YARN-1074.3.patch, 
> YARN-1074.4.patch, YARN-1074.5.patch, YARN-1074.6.patch
>
>
> Once a user brings up YARN daemon, runs jobs, jobs will stay in output 
> returned by $ yarn application -list even after jobs complete already. We 
> want YARN command line to clean up this list. Specifically, we want to remove 
> applications with FINISHED state(not Final-State) or KILLED state from the 
> result.
> {code}
> [user1@host1 ~]$ yarn application -list
> Total Applications:150
> Application-IdApplication-Name
> Application-Type  User   Queue   State   
> Final-State   ProgressTracking-URL
> application_1374638600275_0109   Sleep job   
> MAPREDUCEuser1  default  KILLED
> KILLED   100%host1:54059
> application_1374638600275_0121   Sleep job   
> MAPREDUCEuser1  defaultFINISHED 
> SUCCEEDED   100% host1:19888/jobhistory/job/job_1374638600275_0121
> application_1374638600275_0020   Sleep job   
> MAPREDUCEuser1  defaultFINISHED 
> SUCCEEDED   100% host1:19888/jobhistory/job/job_1374638600275_0020
> application_1374638600275_0038   Sleep job   
> MAPREDUCEuser1  default  
> 
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1074) Clean up YARN CLI app list to show only running apps.

2013-08-22 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1074?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated YARN-1074:


Attachment: YARN-1074.6.patch

> Clean up YARN CLI app list to show only running apps.
> -
>
> Key: YARN-1074
> URL: https://issues.apache.org/jira/browse/YARN-1074
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: client
>Reporter: Tassapol Athiapinya
>Assignee: Xuan Gong
> Attachments: YARN-1074.1.patch, YARN-1074.2.patch, YARN-1074.3.patch, 
> YARN-1074.4.patch, YARN-1074.5.patch, YARN-1074.6.patch
>
>
> Once a user brings up YARN daemon, runs jobs, jobs will stay in output 
> returned by $ yarn application -list even after jobs complete already. We 
> want YARN command line to clean up this list. Specifically, we want to remove 
> applications with FINISHED state(not Final-State) or KILLED state from the 
> result.
> {code}
> [user1@host1 ~]$ yarn application -list
> Total Applications:150
> Application-IdApplication-Name
> Application-Type  User   Queue   State   
> Final-State   ProgressTracking-URL
> application_1374638600275_0109   Sleep job   
> MAPREDUCEuser1  default  KILLED
> KILLED   100%host1:54059
> application_1374638600275_0121   Sleep job   
> MAPREDUCEuser1  defaultFINISHED 
> SUCCEEDED   100% host1:19888/jobhistory/job/job_1374638600275_0121
> application_1374638600275_0020   Sleep job   
> MAPREDUCEuser1  defaultFINISHED 
> SUCCEEDED   100% host1:19888/jobhistory/job/job_1374638600275_0020
> application_1374638600275_0038   Sleep job   
> MAPREDUCEuser1  default  
> 
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-707) Add user info in the YARN ClientToken

2013-08-22 Thread Vinod Kumar Vavilapalli (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinod Kumar Vavilapalli updated YARN-707:
-

Attachment: YARN-707-20130822.txt

Here's a quick patch to add user info to the ClientToAMToken.

Also added a test to verify denial to tampering of user info.

Have to do sanity single node tests.

> Add user info in the YARN ClientToken
> -
>
> Key: YARN-707
> URL: https://issues.apache.org/jira/browse/YARN-707
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Bikas Saha
>Assignee: Vinod Kumar Vavilapalli
> Fix For: 2.1.0-beta
>
> Attachments: YARN-707-20130822.txt
>
>
> If user info is present in the client token then it can be used to do limited 
> authz in the AM.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-556) RM Restart phase 2 - Work preserving restart

2013-08-22 Thread Bikas Saha (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-556?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bikas Saha updated YARN-556:


Attachment: Work Preserving RM Restart.pdf

Attaching a proposal with details. I may have missed writing something even 
though I thought of it or may have missed something altogether. Will 
incorporate feedback as it comes. Will soon start creating sub-tasks that make 
sense in a chronological ordering of work. Making incremental progress while 
keeping the RM stable is the desired course of action (like YARN-128).

> RM Restart phase 2 - Work preserving restart
> 
>
> Key: YARN-556
> URL: https://issues.apache.org/jira/browse/YARN-556
> Project: Hadoop YARN
>  Issue Type: New Feature
>  Components: resourcemanager
>Reporter: Bikas Saha
>Assignee: Bikas Saha
> Attachments: Work Preserving RM Restart.pdf
>
>
> YARN-128 covered storing the state needed for the RM to recover critical 
> information. This umbrella jira will track changes needed to recover the 
> running state of the cluster so that work can be preserved across RM restarts.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1074) Clean up YARN CLI app list to show only running apps.

2013-08-22 Thread Xuan Gong (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1074?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13748068#comment-13748068
 ] 

Xuan Gong commented on YARN-1074:
-

Use EnumSet in the new patch.

> Clean up YARN CLI app list to show only running apps.
> -
>
> Key: YARN-1074
> URL: https://issues.apache.org/jira/browse/YARN-1074
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: client
>Reporter: Tassapol Athiapinya
>Assignee: Xuan Gong
> Attachments: YARN-1074.1.patch, YARN-1074.2.patch, YARN-1074.3.patch, 
> YARN-1074.4.patch, YARN-1074.5.patch
>
>
> Once a user brings up YARN daemon, runs jobs, jobs will stay in output 
> returned by $ yarn application -list even after jobs complete already. We 
> want YARN command line to clean up this list. Specifically, we want to remove 
> applications with FINISHED state(not Final-State) or KILLED state from the 
> result.
> {code}
> [user1@host1 ~]$ yarn application -list
> Total Applications:150
> Application-IdApplication-Name
> Application-Type  User   Queue   State   
> Final-State   ProgressTracking-URL
> application_1374638600275_0109   Sleep job   
> MAPREDUCEuser1  default  KILLED
> KILLED   100%host1:54059
> application_1374638600275_0121   Sleep job   
> MAPREDUCEuser1  defaultFINISHED 
> SUCCEEDED   100% host1:19888/jobhistory/job/job_1374638600275_0121
> application_1374638600275_0020   Sleep job   
> MAPREDUCEuser1  defaultFINISHED 
> SUCCEEDED   100% host1:19888/jobhistory/job/job_1374638600275_0020
> application_1374638600275_0038   Sleep job   
> MAPREDUCEuser1  default  
> 
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1074) Clean up YARN CLI app list to show only running apps.

2013-08-22 Thread Xuan Gong (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1074?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13748067#comment-13748067
 ] 

Xuan Gong commented on YARN-1074:
-

bq.TestApplicationACLs changes seem unnecessary. Shouldn't the default 
behaviour be enough for this?

Yes, we do not need to make any changes. Revert the changes from 
TestApplicationACLs and TestClientRMServices.


> Clean up YARN CLI app list to show only running apps.
> -
>
> Key: YARN-1074
> URL: https://issues.apache.org/jira/browse/YARN-1074
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: client
>Reporter: Tassapol Athiapinya
>Assignee: Xuan Gong
> Attachments: YARN-1074.1.patch, YARN-1074.2.patch, YARN-1074.3.patch, 
> YARN-1074.4.patch, YARN-1074.5.patch
>
>
> Once a user brings up YARN daemon, runs jobs, jobs will stay in output 
> returned by $ yarn application -list even after jobs complete already. We 
> want YARN command line to clean up this list. Specifically, we want to remove 
> applications with FINISHED state(not Final-State) or KILLED state from the 
> result.
> {code}
> [user1@host1 ~]$ yarn application -list
> Total Applications:150
> Application-IdApplication-Name
> Application-Type  User   Queue   State   
> Final-State   ProgressTracking-URL
> application_1374638600275_0109   Sleep job   
> MAPREDUCEuser1  default  KILLED
> KILLED   100%host1:54059
> application_1374638600275_0121   Sleep job   
> MAPREDUCEuser1  defaultFINISHED 
> SUCCEEDED   100% host1:19888/jobhistory/job/job_1374638600275_0121
> application_1374638600275_0020   Sleep job   
> MAPREDUCEuser1  defaultFINISHED 
> SUCCEEDED   100% host1:19888/jobhistory/job/job_1374638600275_0020
> application_1374638600275_0038   Sleep job   
> MAPREDUCEuser1  default  
> 
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1074) Clean up YARN CLI app list to show only running apps.

2013-08-22 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1074?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated YARN-1074:


Attachment: YARN-1074.5.patch

> Clean up YARN CLI app list to show only running apps.
> -
>
> Key: YARN-1074
> URL: https://issues.apache.org/jira/browse/YARN-1074
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: client
>Reporter: Tassapol Athiapinya
>Assignee: Xuan Gong
> Attachments: YARN-1074.1.patch, YARN-1074.2.patch, YARN-1074.3.patch, 
> YARN-1074.4.patch, YARN-1074.5.patch
>
>
> Once a user brings up YARN daemon, runs jobs, jobs will stay in output 
> returned by $ yarn application -list even after jobs complete already. We 
> want YARN command line to clean up this list. Specifically, we want to remove 
> applications with FINISHED state(not Final-State) or KILLED state from the 
> result.
> {code}
> [user1@host1 ~]$ yarn application -list
> Total Applications:150
> Application-IdApplication-Name
> Application-Type  User   Queue   State   
> Final-State   ProgressTracking-URL
> application_1374638600275_0109   Sleep job   
> MAPREDUCEuser1  default  KILLED
> KILLED   100%host1:54059
> application_1374638600275_0121   Sleep job   
> MAPREDUCEuser1  defaultFINISHED 
> SUCCEEDED   100% host1:19888/jobhistory/job/job_1374638600275_0121
> application_1374638600275_0020   Sleep job   
> MAPREDUCEuser1  defaultFINISHED 
> SUCCEEDED   100% host1:19888/jobhistory/job/job_1374638600275_0020
> application_1374638600275_0038   Sleep job   
> MAPREDUCEuser1  default  
> 
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1024) Define a virtual core unambigiously

2013-08-22 Thread Sandy Ryza (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1024?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sandy Ryza updated YARN-1024:
-

Attachment: CPUasaYARNresource.pdf

> Define a virtual core unambigiously
> ---
>
> Key: YARN-1024
> URL: https://issues.apache.org/jira/browse/YARN-1024
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Arun C Murthy
>Assignee: Arun C Murthy
> Attachments: CPUasaYARNresource.pdf
>
>
> We need to clearly define the meaning of a virtual core unambiguously so that 
> it's easy to migrate applications between clusters.
> For e.g. here is Amazon EC2 definition of ECU: 
> http://aws.amazon.com/ec2/faqs/#What_is_an_EC2_Compute_Unit_and_why_did_you_introduce_it
> Essentially we need to clearly define a YARN Virtual Core (YVC).
> Equivalently, we can use ECU itself: *One EC2 Compute Unit provides the 
> equivalent CPU capacity of a 1.0-1.2 GHz 2007 Opteron or 2007 Xeon processor.*

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1024) Define a virtual core unambigiously

2013-08-22 Thread Sandy Ryza (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1024?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13748062#comment-13748062
 ] 

Sandy Ryza commented on YARN-1024:
--

I wrote up a more detailed proposal and attached a PDF of it.

> Define a virtual core unambigiously
> ---
>
> Key: YARN-1024
> URL: https://issues.apache.org/jira/browse/YARN-1024
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Arun C Murthy
>Assignee: Arun C Murthy
> Attachments: CPUasaYARNresource.pdf
>
>
> We need to clearly define the meaning of a virtual core unambiguously so that 
> it's easy to migrate applications between clusters.
> For e.g. here is Amazon EC2 definition of ECU: 
> http://aws.amazon.com/ec2/faqs/#What_is_an_EC2_Compute_Unit_and_why_did_you_introduce_it
> Essentially we need to clearly define a YARN Virtual Core (YVC).
> Equivalently, we can use ECU itself: *One EC2 Compute Unit provides the 
> equivalent CPU capacity of a 1.0-1.2 GHz 2007 Opteron or 2007 Xeon processor.*

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-540) Race condition causing RM to potentially relaunch already unregistered AMs on RM restart

2013-08-22 Thread Vinod Kumar Vavilapalli (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-540?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinod Kumar Vavilapalli updated YARN-540:
-

Summary: Race condition causing RM to potentially relaunch already 
unregistered AMs on RM restart  (was: RM state store not cleaned if job 
succeeds but RM shutdown and restart-dispatcher stopped before it can process 
REMOVE_APP event)

> Race condition causing RM to potentially relaunch already unregistered AMs on 
> RM restart
> 
>
> Key: YARN-540
> URL: https://issues.apache.org/jira/browse/YARN-540
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: resourcemanager
>Reporter: Jian He
>Assignee: Jian He
>Priority: Blocker
> Attachments: YARN-540.patch, YARN-540.patch
>
>
> When job succeeds and successfully call finishApplicationMaster, RM shutdown 
> and restart-dispatcher is stopped before it can process REMOVE_APP event. The 
> next time RM comes back, it will reload the existing state files even though 
> the job is succeeded

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-540) RM state store not cleaned if job succeeds but RM shutdown and restart-dispatcher stopped before it can process REMOVE_APP event

2013-08-22 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-540?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13748024#comment-13748024
 ] 

Vinod Kumar Vavilapalli commented on YARN-540:
--

I think we should fix it the right way. And after things like changes in 
MAPREDUCE-5476, AM's spending time in FINISHING state is going to be more 
common.

So, I am +1 to Bikas's proposal (2). But in the interesting of not making 
incompatible changes, let's do the following:
 - Let's change FinishApplicationMasterResponse to also container a 
response-completed field. If it is true, it means that RM has finished the 
finalization of AM, otherwise, AM is supposed to retry till it becomes true
 - Let RM do the state-store changes asynchronously.

It'll still be a behavior change, but clients which don't follow the multi-step 
unregister will risk only getting restarted.

> RM state store not cleaned if job succeeds but RM shutdown and 
> restart-dispatcher stopped before it can process REMOVE_APP event
> 
>
> Key: YARN-540
> URL: https://issues.apache.org/jira/browse/YARN-540
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: resourcemanager
>Reporter: Jian He
>Assignee: Jian He
>Priority: Blocker
> Attachments: YARN-540.patch, YARN-540.patch
>
>
> When job succeeds and successfully call finishApplicationMaster, RM shutdown 
> and restart-dispatcher is stopped before it can process REMOVE_APP event. The 
> next time RM comes back, it will reload the existing state files even though 
> the job is succeeded

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1092) RM proxy URLs invalid when running without a YARN proxy service

2013-08-22 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1092?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747896#comment-13747896
 ] 

Steve Loughran commented on YARN-1092:
--

# My AM generates a URL to the hbase master, such as {{http://ubuntu:8080}}, a 
URL registered with an AM that is not running a proxy service, and whose 
yarn-site.xml does not declare one
# the link on the GUI is {{http://ubuntu:9081//ubuntu:8080}}
# Which leads to a 404 page

> RM proxy URLs invalid when running without a YARN proxy service
> ---
>
> Key: YARN-1092
> URL: https://issues.apache.org/jira/browse/YARN-1092
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Affects Versions: 2.1.0-beta
> Environment: single host cluster without any yarn proxy service
>Reporter: Steve Loughran
>Priority: Minor
>
> The URLs generated in the RM GUI aren't valid if you don't have a proxy 
> server running

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-1092) RM proxy URLs invalid when running without a YARN proxy service

2013-08-22 Thread Steve Loughran (JIRA)
Steve Loughran created YARN-1092:


 Summary: RM proxy URLs invalid when running without a YARN proxy 
service
 Key: YARN-1092
 URL: https://issues.apache.org/jira/browse/YARN-1092
 Project: Hadoop YARN
  Issue Type: Bug
  Components: resourcemanager
Affects Versions: 2.1.0-beta
 Environment: single host cluster without any yarn proxy service
Reporter: Steve Loughran
Priority: Minor


The URLs generated in the RM GUI aren't valid if you don't have a proxy server 
running

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1089) Add YARN compute units alongside virtual cores

2013-08-22 Thread Sandy Ryza (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1089?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747881#comment-13747881
 ] 

Sandy Ryza commented on YARN-1089:
--

Yeah, I'll write up a document and post it on YARN-1024.  I'm hoping to keep 
the broader discussion there so we can use this (and perhaps additional JIRAs) 
for the actual implementation.

> Add YARN compute units alongside virtual cores
> --
>
> Key: YARN-1089
> URL: https://issues.apache.org/jira/browse/YARN-1089
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: api
>Affects Versions: 2.1.0-beta
>Reporter: Sandy Ryza
>Assignee: Sandy Ryza
>
> Based on discussion in YARN-1024, we will add YARN compute units as a 
> resource for requesting and scheduling CPU processing power.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1089) Add YARN compute units alongside virtual cores

2013-08-22 Thread Hitesh Shah (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1089?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747855#comment-13747855
 ] 

Hitesh Shah commented on YARN-1089:
---

+1 to [~vinodkv]'s request. [~sandyr] Could you ensure that the proposal also 
clearly explains how an application developer is meant to use the compute units 
and/or virtual cores when defining an allocation request and how the allocation 
(based on these 2 params) will be enforced on a container.

> Add YARN compute units alongside virtual cores
> --
>
> Key: YARN-1089
> URL: https://issues.apache.org/jira/browse/YARN-1089
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: api
>Affects Versions: 2.1.0-beta
>Reporter: Sandy Ryza
>Assignee: Sandy Ryza
>
> Based on discussion in YARN-1024, we will add YARN compute units as a 
> resource for requesting and scheduling CPU processing power.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1089) Add YARN compute units alongside virtual cores

2013-08-22 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1089?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747851#comment-13747851
 ] 

Vinod Kumar Vavilapalli commented on YARN-1089:
---

I haven't been following YARN-1024, can you please summarize the proposal here 
for discussion? Thanks.

> Add YARN compute units alongside virtual cores
> --
>
> Key: YARN-1089
> URL: https://issues.apache.org/jira/browse/YARN-1089
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: api
>Affects Versions: 2.1.0-beta
>Reporter: Sandy Ryza
>Assignee: Sandy Ryza
>
> Based on discussion in YARN-1024, we will add YARN compute units as a 
> resource for requesting and scheduling CPU processing power.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1089) Add YARN compute units alongside virtual cores

2013-08-22 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1089?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747750#comment-13747750
 ] 

Arun C Murthy commented on YARN-1089:
-

+1 for this enhancement.

> Add YARN compute units alongside virtual cores
> --
>
> Key: YARN-1089
> URL: https://issues.apache.org/jira/browse/YARN-1089
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: api
>Affects Versions: 2.1.0-beta
>Reporter: Sandy Ryza
>Assignee: Sandy Ryza
>
> Based on discussion in YARN-1024, we will add YARN compute units as a 
> resource for requesting and scheduling CPU processing power.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-899) Get queue administration ACLs working

2013-08-22 Thread Xuan Gong (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-899?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xuan Gong updated YARN-899:
---

Attachment: YARN-899.6.patch

create the patch based on the latest trunk

> Get queue administration ACLs working
> -
>
> Key: YARN-899
> URL: https://issues.apache.org/jira/browse/YARN-899
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: scheduler
>Affects Versions: 2.1.0-beta
>Reporter: Sandy Ryza
>Assignee: Xuan Gong
> Attachments: YARN-899.1.patch, YARN-899.2.patch, YARN-899.3.patch, 
> YARN-899.4.patch, YARN-899.5.patch, YARN-899.5.patch, YARN-899.6.patch
>
>
> The Capacity Scheduler documents the 
> yarn.scheduler.capacity.root..acl_administer_queue config option 
> for controlling who can administer a queue, but it is not hooked up to 
> anything.  The Fair Scheduler could make use of a similar option as well.  
> This is a feature-parity regression from MR1.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1089) Add YARN compute units alongside virtual cores

2013-08-22 Thread Sandy Ryza (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sandy Ryza updated YARN-1089:
-

Description: Based on discussion in YARN-1024, we will add YARN compute 
units as a resource for requesting and scheduling CPU processing power.

> Add YARN compute units alongside virtual cores
> --
>
> Key: YARN-1089
> URL: https://issues.apache.org/jira/browse/YARN-1089
> Project: Hadoop YARN
>  Issue Type: Improvement
>  Components: api
>Affects Versions: 2.1.0-beta
>Reporter: Sandy Ryza
>Assignee: Sandy Ryza
>
> Based on discussion in YARN-1024, we will add YARN compute units as a 
> resource for requesting and scheduling CPU processing power.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1024) Define a virtual core unambigiously

2013-08-22 Thread Sandy Ryza (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1024?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747701#comment-13747701
 ] 

Sandy Ryza commented on YARN-1024:
--

Filed YARN-1089 for adding YCUs.

> Define a virtual core unambigiously
> ---
>
> Key: YARN-1024
> URL: https://issues.apache.org/jira/browse/YARN-1024
> Project: Hadoop YARN
>  Issue Type: Improvement
>Reporter: Arun C Murthy
>Assignee: Arun C Murthy
>
> We need to clearly define the meaning of a virtual core unambiguously so that 
> it's easy to migrate applications between clusters.
> For e.g. here is Amazon EC2 definition of ECU: 
> http://aws.amazon.com/ec2/faqs/#What_is_an_EC2_Compute_Unit_and_why_did_you_introduce_it
> Essentially we need to clearly define a YARN Virtual Core (YVC).
> Equivalently, we can use ECU itself: *One EC2 Compute Unit provides the 
> equivalent CPU capacity of a 1.0-1.2 GHz 2007 Opteron or 2007 Xeon processor.*

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (YARN-1091) All containers localization fails in NM when any one of the configured nm local-dir disk becomes full

2013-08-22 Thread Jason Lowe (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1091?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jason Lowe resolved YARN-1091.
--

Resolution: Duplicate

Duplicate of YARN-257.

> All containers localization fails in NM when any one of the configured nm 
> local-dir disk becomes full
> -
>
> Key: YARN-1091
> URL: https://issues.apache.org/jira/browse/YARN-1091
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: nodemanager
>Affects Versions: 2.0.5-alpha
>Reporter: Devaraj K
>Assignee: Devaraj K
>Priority: Critical
>
> {code:xml}
> 2013-08-22 13:54:22,100 WARN 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: Unable to 
> create app directory 
> /opt/nish/usercache/nish/appcache/application_1377151891396_0017
> java.io.IOException: mkdir of 
> /opt/nish/usercache/nish/appcache/application_1377151891396_0017 failed
>   at org.apache.hadoop.fs.FileSystem.primitiveMkdir(FileSystem.java:1125)
>   at 
> org.apache.hadoop.fs.DelegateToFileSystem.mkdir(DelegateToFileSystem.java:150)
>   at org.apache.hadoop.fs.FilterFs.mkdir(FilterFs.java:187)
>   at org.apache.hadoop.fs.FileContext$4.next(FileContext.java:730)
>   at org.apache.hadoop.fs.FileContext$4.next(FileContext.java:726)
>   at 
> org.apache.hadoop.fs.FileContext$FSLinkResolver.resolve(FileContext.java:2379)
>   at org.apache.hadoop.fs.FileContext.mkdir(FileContext.java:726)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.createDir(DefaultContainerExecutor.java:330)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.createAppDirs(DefaultContainerExecutor.java:426)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.startLocalizer(DefaultContainerExecutor.java:90)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService$LocalizerRunner.run(ResourceLocalizationService.java:859)
> 2013-08-22 13:54:22,102 INFO 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: Copying 
> from 
> /home/nish/new/JAN_4/nmlocal/nmPrivate/container_1377151891396_0017_01_000263.tokens
>  to 
> /home/nish/new/JAN_4/nmlocal/usercache/nish/appcache/application_1377151891396_0017/container_1377151891396_0017_01_000263.tokens
> 2013-08-22 13:54:22,102 INFO 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: CWD set 
> to 
> /home/nish/new/JAN_4/nmlocal/usercache/nish/appcache/application_1377151891396_0017
>  = 
> file:/home/nish/new/JAN_4/nmlocal/usercache/nish/appcache/application_1377151891396_0017
> 2013-08-22 13:54:22,103 INFO 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService:
>  Localizer failed
> java.io.FileNotFoundException: File 
> file:/opt/nish/usercache/nish/appcache/application_1377151891396_0017 does 
> not exist
>   at 
> org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:492)
>   at org.apache.hadoop.fs.FileSystem.primitiveMkdir(FileSystem.java:1112)
>   at 
> org.apache.hadoop.fs.DelegateToFileSystem.mkdir(DelegateToFileSystem.java:150)
>   at org.apache.hadoop.fs.FilterFs.mkdir(FilterFs.java:187)
>   at org.apache.hadoop.fs.FileContext$4.next(FileContext.java:730)
>   at org.apache.hadoop.fs.FileContext$4.next(FileContext.java:726)
>   at 
> org.apache.hadoop.fs.FileContext$FSLinkResolver.resolve(FileContext.java:2379)
>   at org.apache.hadoop.fs.FileContext.mkdir(FileContext.java:726)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ContainerLocalizer.initDirs(ContainerLocalizer.java:391)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ContainerLocalizer.runLocalization(ContainerLocalizer.java:130)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.startLocalizer(DefaultContainerExecutor.java:103)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService$LocalizerRunner.run(ResourceLocalizationService.java:859)
> 2013-08-22 13:54:22,104 INFO 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.container.Container:
>  Container container_1377151891396_0017_01_000263 transitioned from 
> LOCALIZING to LOCALIZATION_FAILED
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1082) Secure RM with recovery enabled and rm state store on hdfs fails with gss exception

2013-08-22 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1082?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747539#comment-13747539
 ] 

Hudson commented on YARN-1082:
--

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1526 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1526/])
YARN-1082. Addendum patch. (acmurthy: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1516352)
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/FileSystemRMStateStore.java
YARN-1082. Create base directories on HDFS after RM login to ensure RM recovery 
doesn't fail in secure mode. Contributed by Vinod K V. (acmurthy: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1516337)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMContextImpl.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/FileSystemRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/MemoryRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/NullRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/RMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/TestRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/TestFairScheduler.java


> Secure RM with recovery enabled and rm state store on hdfs fails with gss 
> exception
> ---
>
> Key: YARN-1082
> URL: https://issues.apache.org/jira/browse/YARN-1082
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Affects Versions: 2.1.0-beta
>Reporter: Arpit Gupta
>Assignee: Vinod Kumar Vavilapalli
>Priority: Blocker
> Fix For: 2.1.1-beta
>
> Attachments: YARN-1082-20130820.2.txt, YARN-1082-20130821.txt
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-881) Priority#compareTo method seems to be wrong.

2013-08-22 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747543#comment-13747543
 ] 

Hudson commented on YARN-881:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1526 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1526/])
YARN-881. Priority#compareTo method seems to be wrong. (Jian He via bikas) 
(bikas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1516331)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/api/records/Priority.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FairScheduler.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/TestSchedulerUtils.java


> Priority#compareTo method seems to be wrong.
> 
>
> Key: YARN-881
> URL: https://issues.apache.org/jira/browse/YARN-881
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Jian He
>Assignee: Jian He
> Fix For: 2.1.1-beta
>
> Attachments: YARN-881.1.patch, YARN-881.patch
>
>
> if lower int value means higher priority, shouldn't we "return 
> other.getPriority() - this.getPriority() " 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1082) Secure RM with recovery enabled and rm state store on hdfs fails with gss exception

2013-08-22 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1082?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747512#comment-13747512
 ] 

Hudson commented on YARN-1082:
--

SUCCESS: Integrated in Hadoop-Hdfs-trunk #1499 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1499/])
YARN-1082. Addendum patch. (acmurthy: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1516352)
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/FileSystemRMStateStore.java
YARN-1082. Create base directories on HDFS after RM login to ensure RM recovery 
doesn't fail in secure mode. Contributed by Vinod K V. (acmurthy: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1516337)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMContextImpl.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/FileSystemRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/MemoryRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/NullRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/RMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/TestRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/TestFairScheduler.java


> Secure RM with recovery enabled and rm state store on hdfs fails with gss 
> exception
> ---
>
> Key: YARN-1082
> URL: https://issues.apache.org/jira/browse/YARN-1082
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Affects Versions: 2.1.0-beta
>Reporter: Arpit Gupta
>Assignee: Vinod Kumar Vavilapalli
>Priority: Blocker
> Fix For: 2.1.1-beta
>
> Attachments: YARN-1082-20130820.2.txt, YARN-1082-20130821.txt
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-881) Priority#compareTo method seems to be wrong.

2013-08-22 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747516#comment-13747516
 ] 

Hudson commented on YARN-881:
-

SUCCESS: Integrated in Hadoop-Hdfs-trunk #1499 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1499/])
YARN-881. Priority#compareTo method seems to be wrong. (Jian He via bikas) 
(bikas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1516331)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/api/records/Priority.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FairScheduler.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/TestSchedulerUtils.java


> Priority#compareTo method seems to be wrong.
> 
>
> Key: YARN-881
> URL: https://issues.apache.org/jira/browse/YARN-881
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Jian He
>Assignee: Jian He
> Fix For: 2.1.1-beta
>
> Attachments: YARN-881.1.patch, YARN-881.patch
>
>
> if lower int value means higher priority, shouldn't we "return 
> other.getPriority() - this.getPriority() " 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-881) Priority#compareTo method seems to be wrong.

2013-08-22 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-881?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747426#comment-13747426
 ] 

Hudson commented on YARN-881:
-

SUCCESS: Integrated in Hadoop-Yarn-trunk #309 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/309/])
YARN-881. Priority#compareTo method seems to be wrong. (Jian He via bikas) 
(bikas: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1516331)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api/src/main/java/org/apache/hadoop/yarn/api/records/Priority.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/FairScheduler.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/TestSchedulerUtils.java


> Priority#compareTo method seems to be wrong.
> 
>
> Key: YARN-881
> URL: https://issues.apache.org/jira/browse/YARN-881
> Project: Hadoop YARN
>  Issue Type: Bug
>Reporter: Jian He
>Assignee: Jian He
> Fix For: 2.1.1-beta
>
> Attachments: YARN-881.1.patch, YARN-881.patch
>
>
> if lower int value means higher priority, shouldn't we "return 
> other.getPriority() - this.getPriority() " 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-1082) Secure RM with recovery enabled and rm state store on hdfs fails with gss exception

2013-08-22 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-1082?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747422#comment-13747422
 ] 

Hudson commented on YARN-1082:
--

SUCCESS: Integrated in Hadoop-Yarn-trunk #309 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/309/])
YARN-1082. Addendum patch. (acmurthy: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1516352)
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/FileSystemRMStateStore.java
YARN-1082. Create base directories on HDFS after RM login to ensure RM recovery 
doesn't fail in secure mode. Contributed by Vinod K V. (acmurthy: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVN&view=rev&rev=1516337)
* /hadoop/common/trunk/hadoop-yarn-project/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/RMContextImpl.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/ResourceManager.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/FileSystemRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/MemoryRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/NullRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/main/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/RMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/recovery/TestRMStateStore.java
* 
/hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/java/org/apache/hadoop/yarn/server/resourcemanager/scheduler/fair/TestFairScheduler.java


> Secure RM with recovery enabled and rm state store on hdfs fails with gss 
> exception
> ---
>
> Key: YARN-1082
> URL: https://issues.apache.org/jira/browse/YARN-1082
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: resourcemanager
>Affects Versions: 2.1.0-beta
>Reporter: Arpit Gupta
>Assignee: Vinod Kumar Vavilapalli
>Priority: Blocker
> Fix For: 2.1.1-beta
>
> Attachments: YARN-1082-20130820.2.txt, YARN-1082-20130821.txt
>
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (YARN-1091) All containers localization fails in NM when any one of the configured nm local-dir disk becomes full

2013-08-22 Thread Devaraj K (JIRA)

 [ 
https://issues.apache.org/jira/browse/YARN-1091?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Devaraj K updated YARN-1091:


Summary: All containers localization fails in NM when any one of the 
configured nm local-dir disk becomes full  (was: All containers localization 
fails when any one of the configured nm local-dir disk becomes full)

> All containers localization fails in NM when any one of the configured nm 
> local-dir disk becomes full
> -
>
> Key: YARN-1091
> URL: https://issues.apache.org/jira/browse/YARN-1091
> Project: Hadoop YARN
>  Issue Type: Bug
>  Components: nodemanager
>Affects Versions: 2.0.5-alpha
>Reporter: Devaraj K
>Assignee: Devaraj K
>Priority: Critical
>
> {code:xml}
> 2013-08-22 13:54:22,100 WARN 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: Unable to 
> create app directory 
> /opt/nish/usercache/nish/appcache/application_1377151891396_0017
> java.io.IOException: mkdir of 
> /opt/nish/usercache/nish/appcache/application_1377151891396_0017 failed
>   at org.apache.hadoop.fs.FileSystem.primitiveMkdir(FileSystem.java:1125)
>   at 
> org.apache.hadoop.fs.DelegateToFileSystem.mkdir(DelegateToFileSystem.java:150)
>   at org.apache.hadoop.fs.FilterFs.mkdir(FilterFs.java:187)
>   at org.apache.hadoop.fs.FileContext$4.next(FileContext.java:730)
>   at org.apache.hadoop.fs.FileContext$4.next(FileContext.java:726)
>   at 
> org.apache.hadoop.fs.FileContext$FSLinkResolver.resolve(FileContext.java:2379)
>   at org.apache.hadoop.fs.FileContext.mkdir(FileContext.java:726)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.createDir(DefaultContainerExecutor.java:330)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.createAppDirs(DefaultContainerExecutor.java:426)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.startLocalizer(DefaultContainerExecutor.java:90)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService$LocalizerRunner.run(ResourceLocalizationService.java:859)
> 2013-08-22 13:54:22,102 INFO 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: Copying 
> from 
> /home/nish/new/JAN_4/nmlocal/nmPrivate/container_1377151891396_0017_01_000263.tokens
>  to 
> /home/nish/new/JAN_4/nmlocal/usercache/nish/appcache/application_1377151891396_0017/container_1377151891396_0017_01_000263.tokens
> 2013-08-22 13:54:22,102 INFO 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: CWD set 
> to 
> /home/nish/new/JAN_4/nmlocal/usercache/nish/appcache/application_1377151891396_0017
>  = 
> file:/home/nish/new/JAN_4/nmlocal/usercache/nish/appcache/application_1377151891396_0017
> 2013-08-22 13:54:22,103 INFO 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService:
>  Localizer failed
> java.io.FileNotFoundException: File 
> file:/opt/nish/usercache/nish/appcache/application_1377151891396_0017 does 
> not exist
>   at 
> org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:492)
>   at org.apache.hadoop.fs.FileSystem.primitiveMkdir(FileSystem.java:1112)
>   at 
> org.apache.hadoop.fs.DelegateToFileSystem.mkdir(DelegateToFileSystem.java:150)
>   at org.apache.hadoop.fs.FilterFs.mkdir(FilterFs.java:187)
>   at org.apache.hadoop.fs.FileContext$4.next(FileContext.java:730)
>   at org.apache.hadoop.fs.FileContext$4.next(FileContext.java:726)
>   at 
> org.apache.hadoop.fs.FileContext$FSLinkResolver.resolve(FileContext.java:2379)
>   at org.apache.hadoop.fs.FileContext.mkdir(FileContext.java:726)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ContainerLocalizer.initDirs(ContainerLocalizer.java:391)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ContainerLocalizer.runLocalization(ContainerLocalizer.java:130)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.startLocalizer(DefaultContainerExecutor.java:103)
>   at 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService$LocalizerRunner.run(ResourceLocalizationService.java:859)
> 2013-08-22 13:54:22,104 INFO 
> org.apache.hadoop.yarn.server.nodemanager.containermanager.container.Container:
>  Container container_1377151891396_0017_01_000263 transitioned from 
> LOCALIZING to LOCALIZATION_FAILED
> {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (YARN-1091) All containers localization fails when any one of the configured nm local-dir disk becomes full

2013-08-22 Thread Devaraj K (JIRA)
Devaraj K created YARN-1091:
---

 Summary: All containers localization fails when any one of the 
configured nm local-dir disk becomes full
 Key: YARN-1091
 URL: https://issues.apache.org/jira/browse/YARN-1091
 Project: Hadoop YARN
  Issue Type: Bug
  Components: nodemanager
Affects Versions: 2.0.5-alpha
Reporter: Devaraj K
Assignee: Devaraj K
Priority: Critical


{code:xml}
2013-08-22 13:54:22,100 WARN 
org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: Unable to 
create app directory 
/opt/nish/usercache/nish/appcache/application_1377151891396_0017
java.io.IOException: mkdir of 
/opt/nish/usercache/nish/appcache/application_1377151891396_0017 failed
at org.apache.hadoop.fs.FileSystem.primitiveMkdir(FileSystem.java:1125)
at 
org.apache.hadoop.fs.DelegateToFileSystem.mkdir(DelegateToFileSystem.java:150)
at org.apache.hadoop.fs.FilterFs.mkdir(FilterFs.java:187)
at org.apache.hadoop.fs.FileContext$4.next(FileContext.java:730)
at org.apache.hadoop.fs.FileContext$4.next(FileContext.java:726)
at 
org.apache.hadoop.fs.FileContext$FSLinkResolver.resolve(FileContext.java:2379)
at org.apache.hadoop.fs.FileContext.mkdir(FileContext.java:726)
at 
org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.createDir(DefaultContainerExecutor.java:330)
at 
org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.createAppDirs(DefaultContainerExecutor.java:426)
at 
org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.startLocalizer(DefaultContainerExecutor.java:90)
at 
org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService$LocalizerRunner.run(ResourceLocalizationService.java:859)
2013-08-22 13:54:22,102 INFO 
org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: Copying 
from 
/home/nish/new/JAN_4/nmlocal/nmPrivate/container_1377151891396_0017_01_000263.tokens
 to 
/home/nish/new/JAN_4/nmlocal/usercache/nish/appcache/application_1377151891396_0017/container_1377151891396_0017_01_000263.tokens
2013-08-22 13:54:22,102 INFO 
org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor: CWD set to 
/home/nish/new/JAN_4/nmlocal/usercache/nish/appcache/application_1377151891396_0017
 = 
file:/home/nish/new/JAN_4/nmlocal/usercache/nish/appcache/application_1377151891396_0017
2013-08-22 13:54:22,103 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService:
 Localizer failed
java.io.FileNotFoundException: File 
file:/opt/nish/usercache/nish/appcache/application_1377151891396_0017 does not 
exist
at 
org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:492)
at org.apache.hadoop.fs.FileSystem.primitiveMkdir(FileSystem.java:1112)
at 
org.apache.hadoop.fs.DelegateToFileSystem.mkdir(DelegateToFileSystem.java:150)
at org.apache.hadoop.fs.FilterFs.mkdir(FilterFs.java:187)
at org.apache.hadoop.fs.FileContext$4.next(FileContext.java:730)
at org.apache.hadoop.fs.FileContext$4.next(FileContext.java:726)
at 
org.apache.hadoop.fs.FileContext$FSLinkResolver.resolve(FileContext.java:2379)
at org.apache.hadoop.fs.FileContext.mkdir(FileContext.java:726)
at 
org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ContainerLocalizer.initDirs(ContainerLocalizer.java:391)
at 
org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ContainerLocalizer.runLocalization(ContainerLocalizer.java:130)
at 
org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor.startLocalizer(DefaultContainerExecutor.java:103)
at 
org.apache.hadoop.yarn.server.nodemanager.containermanager.localizer.ResourceLocalizationService$LocalizerRunner.run(ResourceLocalizationService.java:859)
2013-08-22 13:54:22,104 INFO 
org.apache.hadoop.yarn.server.nodemanager.containermanager.container.Container: 
Container container_1377151891396_0017_01_000263 transitioned from LOCALIZING 
to LOCALIZATION_FAILED
{code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (YARN-49) Improve distributed shell application to work on a secure cluster

2013-08-22 Thread Mohammad Kamrul Islam (JIRA)

[ 
https://issues.apache.org/jira/browse/YARN-49?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13747319#comment-13747319
 ] 

Mohammad Kamrul Islam commented on YARN-49:
---

[~ojoshi] do you have WIP patch that i can use for new Giraph AM? It doesn't 
need to work though.



> Improve distributed shell application to work on a secure cluster
> -
>
> Key: YARN-49
> URL: https://issues.apache.org/jira/browse/YARN-49
> Project: Hadoop YARN
>  Issue Type: Sub-task
>  Components: applications/distributed-shell
>Reporter: Hitesh Shah
>Assignee: Omkar Vinit Joshi
>


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira