[jira] [Updated] (MAPREDUCE-6468) Consistent log severity level guards and statements in MapReduce project

2015-09-09 Thread Jagadesh Kiran N (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-6468?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jagadesh Kiran N updated MAPREDUCE-6468:

Attachment: MAPREDUCE-6468-03.patch

Attaching the patch.Addressed white spaces and check style.Findbugs need to 
check as no issue displayed. Test failures are not part of patch updation.

> Consistent log severity level guards and statements in MapReduce project
> 
>
> Key: MAPREDUCE-6468
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6468
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>Reporter: Jackie Chang
>Assignee: Jagadesh Kiran N
>Priority: Minor
>  Labels: BB2015-05-TBR
> Attachments: HADOOP-9995-00.patch, HADOOP-9995.patch, 
> MAPREDUCE-6468-01.patch, MAPREDUCE-6468-02.patch, MAPREDUCE-6468-03.patch
>
>
> Developers use logs to do in-house debugging. These log statements are later 
> demoted to less severe levels and usually are guarded by their matching 
> severity levels. However, we do see inconsistencies in trunk. A log statement 
> like 
> {code}
>if (LOG.isDebugEnabled()) {
> LOG.info("Assigned container (" + allocated + ") "
> {code}
> doesn't make much sense because the log message is actually only printed out 
> in DEBUG-level. We do see previous issues tried to correct this 
> inconsistency. I am proposing a comprehensive correction over trunk.
> Doug Cutting pointed it out in HADOOP-312: 
> https://issues.apache.org/jira/browse/HADOOP-312?focusedCommentId=12429498=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-12429498
> HDFS-1611 also corrected this inconsistency.
> This could have been avoided by switching from log4j to slf4j's {} format 
> like CASSANDRA-625 (2010/3) and ZOOKEEPER-850 (2012/1), which gives cleaner 
> code and slightly higher performance.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6468) Consistent log severity level guards and statements in MapReduce project

2015-09-09 Thread Tsuyoshi Ozawa (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14736364#comment-14736364
 ] 

Tsuyoshi Ozawa commented on MAPREDUCE-6468:
---

[~jagadesh.kiran] Could you double check that the patch you uploaded is 
correct? It seems old one.

{code}
+++ 
b/hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/ha/BootstrapStandby.java
@@ -374,9 +374,7 @@ private boolean checkLogsAvailableForRead(FSImage image, 
long imageTxId,
   "or call saveNamespace on the active node.\n" +
   "Error: " + e.getLocalizedMessage();
   if (LOG.isDebugEnabled()) {
-LOG.fatal(msg, e);
-  } else {
-LOG.fatal(msg);
+LOG.debug("", e);
   }
   return false;
 }
{code}

Also, in BackupImage, the log level looks wrong. Please fix it.
{code}
 if (LOG.isDebugEnabled()) {
-  LOG.debug("State transition " + bnState + " -> " + newState);
+  LOG.trace("State transition " + bnState + " -> " + newState);
 }
{code}

> Consistent log severity level guards and statements in MapReduce project
> 
>
> Key: MAPREDUCE-6468
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6468
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>Reporter: Jackie Chang
>Assignee: Jagadesh Kiran N
>Priority: Minor
>  Labels: BB2015-05-TBR
> Attachments: HADOOP-9995-00.patch, HADOOP-9995.patch, 
> MAPREDUCE-6468-01.patch, MAPREDUCE-6468-02.patch, MAPREDUCE-6468-03.patch
>
>
> Developers use logs to do in-house debugging. These log statements are later 
> demoted to less severe levels and usually are guarded by their matching 
> severity levels. However, we do see inconsistencies in trunk. A log statement 
> like 
> {code}
>if (LOG.isDebugEnabled()) {
> LOG.info("Assigned container (" + allocated + ") "
> {code}
> doesn't make much sense because the log message is actually only printed out 
> in DEBUG-level. We do see previous issues tried to correct this 
> inconsistency. I am proposing a comprehensive correction over trunk.
> Doug Cutting pointed it out in HADOOP-312: 
> https://issues.apache.org/jira/browse/HADOOP-312?focusedCommentId=12429498=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-12429498
> HDFS-1611 also corrected this inconsistency.
> This could have been avoided by switching from log4j to slf4j's {} format 
> like CASSANDRA-625 (2010/3) and ZOOKEEPER-850 (2012/1), which gives cleaner 
> code and slightly higher performance.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6468) Consistent log severity level guards and statements in MapReduce project

2015-09-09 Thread Tsuyoshi Ozawa (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14736315#comment-14736315
 ] 

Tsuyoshi Ozawa commented on MAPREDUCE-6468:
---

[~jagadesh.kiran] thank you for updating. In BootstrapStandby, the message 
should be used instead of passing Exception object directly since it helps 
users to understand what's happened.
{code}
@@ -374,9 +374,7 @@ private boolean checkLogsAvailableForRead(FSImage image, 
long imageTxId,
   "or call saveNamespace on the active node.\n" +
   "Error: " + e.getLocalizedMessage();
   if (LOG.isDebugEnabled()) {
-LOG.fatal(msg, e);
-  } else {
-LOG.fatal(msg);
+LOG.debug("", e);
   }
{code}

Also, the log looks fatal one: we shouldn't remove it. Could you update it as 
follows?:
{code}
   if (LOG.isDebugEnabled()) {
LOG.log(msg, e); // this line should be fixed as LOG.debug
  } else {
LOG.fatal(msg);  // this line and else statement should remain here.
   }
{code}



> Consistent log severity level guards and statements in MapReduce project
> 
>
> Key: MAPREDUCE-6468
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6468
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>Reporter: Jackie Chang
>Assignee: Jagadesh Kiran N
>Priority: Minor
>  Labels: BB2015-05-TBR
> Attachments: HADOOP-9995-00.patch, HADOOP-9995.patch, 
> MAPREDUCE-6468-01.patch, MAPREDUCE-6468-02.patch
>
>
> Developers use logs to do in-house debugging. These log statements are later 
> demoted to less severe levels and usually are guarded by their matching 
> severity levels. However, we do see inconsistencies in trunk. A log statement 
> like 
> {code}
>if (LOG.isDebugEnabled()) {
> LOG.info("Assigned container (" + allocated + ") "
> {code}
> doesn't make much sense because the log message is actually only printed out 
> in DEBUG-level. We do see previous issues tried to correct this 
> inconsistency. I am proposing a comprehensive correction over trunk.
> Doug Cutting pointed it out in HADOOP-312: 
> https://issues.apache.org/jira/browse/HADOOP-312?focusedCommentId=12429498=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-12429498
> HDFS-1611 also corrected this inconsistency.
> This could have been avoided by switching from log4j to slf4j's {} format 
> like CASSANDRA-625 (2010/3) and ZOOKEEPER-850 (2012/1), which gives cleaner 
> code and slightly higher performance.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (MAPREDUCE-6468) Consistent log severity level guards and statements in MapReduce project

2015-09-09 Thread Jagadesh Kiran N (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-6468?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jagadesh Kiran N updated MAPREDUCE-6468:

Attachment: MAPREDUCE-6468-04.patch

[~ozawa] thanks for your review, addressed all your comments in this patch 
,please review

> Consistent log severity level guards and statements in MapReduce project
> 
>
> Key: MAPREDUCE-6468
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6468
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>Reporter: Jackie Chang
>Assignee: Jagadesh Kiran N
>Priority: Minor
>  Labels: BB2015-05-TBR
> Attachments: HADOOP-9995-00.patch, HADOOP-9995.patch, 
> MAPREDUCE-6468-01.patch, MAPREDUCE-6468-02.patch, MAPREDUCE-6468-03.patch, 
> MAPREDUCE-6468-04.patch
>
>
> Developers use logs to do in-house debugging. These log statements are later 
> demoted to less severe levels and usually are guarded by their matching 
> severity levels. However, we do see inconsistencies in trunk. A log statement 
> like 
> {code}
>if (LOG.isDebugEnabled()) {
> LOG.info("Assigned container (" + allocated + ") "
> {code}
> doesn't make much sense because the log message is actually only printed out 
> in DEBUG-level. We do see previous issues tried to correct this 
> inconsistency. I am proposing a comprehensive correction over trunk.
> Doug Cutting pointed it out in HADOOP-312: 
> https://issues.apache.org/jira/browse/HADOOP-312?focusedCommentId=12429498=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-12429498
> HDFS-1611 also corrected this inconsistency.
> This could have been avoided by switching from log4j to slf4j's {} format 
> like CASSANDRA-625 (2010/3) and ZOOKEEPER-850 (2012/1), which gives cleaner 
> code and slightly higher performance.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6468) Consistent log severity level guards and statements in MapReduce project

2015-09-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14736761#comment-14736761
 ] 

Hadoop QA commented on MAPREDUCE-6468:
--

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  25m 40s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   9m 31s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  12m 15s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 28s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   6m 37s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 40s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 37s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |  10m 39s | The patch appears to introduce 1 
new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 41s | Pre-build of native portion |
| {color:green}+1{color} | mapreduce tests |   9m 50s | Tests passed in 
hadoop-mapreduce-client-app. |
| {color:green}+1{color} | tools/hadoop tests |   6m 59s | Tests passed in 
hadoop-distcp. |
| {color:red}-1{color} | tools/hadoop tests |  10m 14s | Tests failed in 
hadoop-gridmix. |
| {color:red}-1{color} | yarn tests |   5m 32s | Tests failed in 
hadoop-yarn-client. |
| {color:green}+1{color} | yarn tests |   1m 59s | Tests passed in 
hadoop-yarn-common. |
| {color:green}+1{color} | yarn tests |  54m 34s | Tests passed in 
hadoop-yarn-server-resourcemanager. |
| {color:red}-1{color} | hdfs tests | 162m 40s | Tests failed in hadoop-hdfs. |
| | | 323m  0s | |
\\
\\
|| Reason || Tests ||
| FindBugs | module:hadoop-hdfs |
| Failed unit tests | hadoop.yarn.client.TestApplicationClientProtocolOnHA |
|   | hadoop.yarn.client.api.impl.TestYarnClient |
|   | hadoop.hdfs.server.namenode.ha.TestBootstrapStandby |
| Timed out tests | org.apache.hadoop.mapred.gridmix.TestGridmixSubmission |
|   | org.apache.hadoop.mapred.gridmix.TestSleepJob |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12754844/MAPREDUCE-6468-03.patch
 |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / 94cf7ab |
| Findbugs warnings | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5975/artifact/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
 |
| hadoop-mapreduce-client-app test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5975/artifact/patchprocess/testrun_hadoop-mapreduce-client-app.txt
 |
| hadoop-distcp test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5975/artifact/patchprocess/testrun_hadoop-distcp.txt
 |
| hadoop-gridmix test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5975/artifact/patchprocess/testrun_hadoop-gridmix.txt
 |
| hadoop-yarn-client test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5975/artifact/patchprocess/testrun_hadoop-yarn-client.txt
 |
| hadoop-yarn-common test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5975/artifact/patchprocess/testrun_hadoop-yarn-common.txt
 |
| hadoop-yarn-server-resourcemanager test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5975/artifact/patchprocess/testrun_hadoop-yarn-server-resourcemanager.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5975/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5975/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf901.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5975/console |


This message was automatically generated.

> Consistent log severity level guards and statements in MapReduce project
> 
>
> Key: MAPREDUCE-6468
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6468
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>Reporter: Jackie Chang
>Assignee: Jagadesh Kiran N
>

[jira] [Commented] (MAPREDUCE-6468) Consistent log severity level guards and statements in MapReduce project

2015-09-09 Thread Jagadesh Kiran N (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14736793#comment-14736793
 ] 

Jagadesh Kiran N commented on MAPREDUCE-6468:
-

No Findbugs in the Report, And the test case failures are not related to the 
Patch. 

> Consistent log severity level guards and statements in MapReduce project
> 
>
> Key: MAPREDUCE-6468
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6468
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>Reporter: Jackie Chang
>Assignee: Jagadesh Kiran N
>Priority: Minor
>  Labels: BB2015-05-TBR
> Attachments: HADOOP-9995-00.patch, HADOOP-9995.patch, 
> MAPREDUCE-6468-01.patch, MAPREDUCE-6468-02.patch, MAPREDUCE-6468-03.patch, 
> MAPREDUCE-6468-04.patch
>
>
> Developers use logs to do in-house debugging. These log statements are later 
> demoted to less severe levels and usually are guarded by their matching 
> severity levels. However, we do see inconsistencies in trunk. A log statement 
> like 
> {code}
>if (LOG.isDebugEnabled()) {
> LOG.info("Assigned container (" + allocated + ") "
> {code}
> doesn't make much sense because the log message is actually only printed out 
> in DEBUG-level. We do see previous issues tried to correct this 
> inconsistency. I am proposing a comprehensive correction over trunk.
> Doug Cutting pointed it out in HADOOP-312: 
> https://issues.apache.org/jira/browse/HADOOP-312?focusedCommentId=12429498=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-12429498
> HDFS-1611 also corrected this inconsistency.
> This could have been avoided by switching from log4j to slf4j's {} format 
> like CASSANDRA-625 (2010/3) and ZOOKEEPER-850 (2012/1), which gives cleaner 
> code and slightly higher performance.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Jason Lowe (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14736983#comment-14736983
 ] 

Jason Lowe commented on MAPREDUCE-6415:
---

+1 latest patch looks good to me.  Will commit this later today if there are no 
objections.

> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-5870) Support for passing Job priority through Application Submission Context in Mapreduce Side

2015-09-09 Thread Sunil G (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5870?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14736993#comment-14736993
 ] 

Sunil G commented on MAPREDUCE-5870:


Thank you [~jlowe] for sharing the comments.

As discussed offline, I had doubts on couple of points here.

bq.We are already returning the job priority via getJobStatus
As I see, we are invoking {{YARNRunner#getJobStatus}} to get the job status 
while we try to call {{Job#getPriority}}.
And in {{ClientServiceDelegate#getJobStatus}}, *JobStatus* is created from 
*JobReport* (from AM).
{code}
  public JobStatus getJobStatus(JobID oldJobID) throws IOException {
org.apache.hadoop.mapreduce.v2.api.records.JobId jobId =
  TypeConverter.toYarn(oldJobID);
GetJobReportRequest request =
recordFactory.newRecordInstance(GetJobReportRequest.class);
request.setJobId(jobId);
JobReport report = ((GetJobReportResponse) invoke("getJobReport",
GetJobReportRequest.class, request)).getJobReport();
...
...
jobStatus = TypeConverter.fromYarn(report, url);
...
{code}

As this JobReport doesnt have priority information, current code in 
{{TypeConverter#fromYarn}} is to hardcode the status.

{code}
  public static JobStatus fromYarn(JobReport jobreport, String trackingUrl) {
JobPriority jobPriority = JobPriority.NORMAL;
JobStatus jobStatus = new org.apache.hadoop.mapred.JobStatus(
fromYarn(jobreport.getJobId()), jobreport.getSetupProgress(), jobreport

{code}

AppReport has the information about the priority of the application. So a 
second call has to be made to fetch the same from {{YarnRunner#getJobStatus}}. 
And it can be set to JobStatus separately.
Kindly help to share your thoughts here if I am missing something important.

bq.Why was the call to testChangingJobPriority removed from TestMRJobClient?
{{TestMRJobClient#testChangingJobPriority}} was invoked after job was 
completed. And YARN throws excpetion for change priority application if its not 
RUNNING or ACCEPTED. As updated offline, I agree that YARN need not have to 
throw excpetion if the app is completed, rather it can be skipped. I will file 
a ticket in YARN to fix the same. Thoughts?

For LocalJobRunner exception case, I will file another ticket to fix the 
exception and to support correct priority.
Also I will add all possible test cases for Job and JobConf to see all possible 
integer and enums are set/get correctly.
Thank you for the support.

> Support for passing Job priority through Application Submission Context in 
> Mapreduce Side
> -
>
> Key: MAPREDUCE-5870
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5870
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: client
>Reporter: Sunil G
>Assignee: Sunil G
> Attachments: 0001-MAPREDUCE-5870.patch, 0002-MAPREDUCE-5870.patch, 
> 0003-MAPREDUCE-5870.patch, 0004-MAPREDUCE-5870.patch, 
> 0005-MAPREDUCE-5870.patch, 0006-MAPREDUCE-5870.patch, Yarn-2002.1.patch
>
>
> Job Prioirty can be set from client side as below [Configuration and api].
>   a.  JobConf.getJobPriority() and 
> Job.setPriority(JobPriority priority) 
>   b.  We can also use configuration 
> "mapreduce.job.priority".
>   Now this Job priority can be passed in Application Submission 
> context from Client side.
>   Here we can reuse the MRJobConfig.PRIORITY configuration. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-5870) Support for passing Job priority through Application Submission Context in Mapreduce Side

2015-09-09 Thread Jason Lowe (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5870?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14737035#comment-14737035
 ] 

Jason Lowe commented on MAPREDUCE-5870:
---

bq. AppReport has the information about the priority of the application. So a 
second call has to be made to fetch the same from YarnRunner#getJobStatus. And 
it can be set to JobStatus separately.

The problem is that the ApplicationMaster doesn't actually know what its job 
priority actually is, as that's a concept tracked by the RM not the AM.  So to 
report the priority we'd have to make a separate call to the RM to determine 
the priority and meld that into the rest of the information returned for the 
job status.  I'm not thrilled with the thought of adding an additional call to 
the RM for every time a client wants to get the job status, especially for a 
field that most clients won't need.  Clients often bang on the job status to 
see when a job is done, and that means they'll start banging on the RM as well 
if we make the delegate start hitting both the AM and the RM to generate the 
JobStatus.

There's only two ways I can currently think of to get around that.  Either we 
need to obsolete the priority field from the job status and provide a separate 
call to get the priority when the client really wants to know it, or we have to 
find a way for the AM to know its job priority so it can return it in its 
JobStatus responses.  For the latter we could have the RM send it down in 
heartbeat responses, but there would be a delay between when a client updates 
the priority and the AM reports the updated value.  We could have the AM 
proactively ask the RM what its priority is, but then we're back to banging on 
the RM for every job status call (only this time from AMs instead of the 
client, but same number of RM calls overall).

Any other thoughts on what we should do here?  I'm not thrilled about the 
thought of ditching the priority in the JobStatus field.  However it's been 
worthless ever since we went to YARN (i.e.: broken since 0.23 and 2.x), so it's 
tempting to declare it officially obsolete (as it in practice already is) and 
provide a different API to retrieve it.

bq. And YARN throws excpetion for change priority application if its not 
RUNNING or ACCEPTED. 
Yes, this to me is a bug.  We don't throw exceptions when clients try to kill 
apps that have terminated, and similarly we shouldn't throw exceptions if 
clients try to change the priority of terminated apps.


> Support for passing Job priority through Application Submission Context in 
> Mapreduce Side
> -
>
> Key: MAPREDUCE-5870
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5870
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: client
>Reporter: Sunil G
>Assignee: Sunil G
> Attachments: 0001-MAPREDUCE-5870.patch, 0002-MAPREDUCE-5870.patch, 
> 0003-MAPREDUCE-5870.patch, 0004-MAPREDUCE-5870.patch, 
> 0005-MAPREDUCE-5870.patch, 0006-MAPREDUCE-5870.patch, Yarn-2002.1.patch
>
>
> Job Prioirty can be set from client side as below [Configuration and api].
>   a.  JobConf.getJobPriority() and 
> Job.setPriority(JobPriority priority) 
>   b.  We can also use configuration 
> "mapreduce.job.priority".
>   Now this Job priority can be passed in Application Submission 
> context from Client side.
>   Here we can reuse the MRJobConfig.PRIORITY configuration. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-5870) Support for passing Job priority through Application Submission Context in Mapreduce Side

2015-09-09 Thread Eric Payne (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5870?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14737076#comment-14737076
 ] 

Eric Payne commented on MAPREDUCE-5870:
---

{quote}
There's only two ways I can currently think of to get around that. Either we 
need to obsolete the priority field from the job status and provide a separate 
call to get the priority when the client really wants to know it, or we have to 
find a way for the AM to know its job priority so it can return it in its 
JobStatus responses. For the latter we could have the RM send it down in 
heartbeat responses, but there would be a delay between when a client updates 
the priority and the AM reports the updated value.
{quote}
[~jlowe] and [~sunilg],
I would vote for the second option. That is, update the priority in the RM and 
then tell the AM about it when it hearbeats in. I think it would be fine to 
have the short delay between when the priority was updated and the AM knows 
about it. I can't foresee a use case where the client would atomically need to 
know about the change. The one thing that might cause confusion is if the 
client set the priority and immediately read the priority which might still be 
the old value. This may cause confusion, but as long as this behavior is well 
documented it should be fine.

> Support for passing Job priority through Application Submission Context in 
> Mapreduce Side
> -
>
> Key: MAPREDUCE-5870
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5870
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: client
>Reporter: Sunil G
>Assignee: Sunil G
> Attachments: 0001-MAPREDUCE-5870.patch, 0002-MAPREDUCE-5870.patch, 
> 0003-MAPREDUCE-5870.patch, 0004-MAPREDUCE-5870.patch, 
> 0005-MAPREDUCE-5870.patch, 0006-MAPREDUCE-5870.patch, Yarn-2002.1.patch
>
>
> Job Prioirty can be set from client side as below [Configuration and api].
>   a.  JobConf.getJobPriority() and 
> Job.setPriority(JobPriority priority) 
>   b.  We can also use configuration 
> "mapreduce.job.priority".
>   Now this Job priority can be passed in Application Submission 
> context from Client side.
>   Here we can reuse the MRJobConfig.PRIORITY configuration. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6468) Consistent log severity level guards and statements in MapReduce project

2015-09-09 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14737006#comment-14737006
 ] 

Hadoop QA commented on MAPREDUCE-6468:
--

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  22m 59s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  1s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   7m 55s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m 12s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 21s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   4m 55s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 29s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 32s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   9m  6s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 15s | Pre-build of native portion |
| {color:red}-1{color} | mapreduce tests |   8m 52s | Tests failed in 
hadoop-mapreduce-client-app. |
| {color:green}+1{color} | tools/hadoop tests |   6m 23s | Tests passed in 
hadoop-distcp. |
| {color:green}+1{color} | tools/hadoop tests |  14m 48s | Tests passed in 
hadoop-gridmix. |
| {color:green}+1{color} | yarn tests |   6m 56s | Tests passed in 
hadoop-yarn-client. |
| {color:red}-1{color} | yarn tests |   0m 20s | Tests failed in 
hadoop-yarn-common. |
| {color:red}-1{color} | yarn tests |   0m 20s | Tests failed in 
hadoop-yarn-server-resourcemanager. |
| {color:red}-1{color} | hdfs tests | 162m  1s | Tests failed in hadoop-hdfs. |
| | | 260m 30s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.mapreduce.v2.app.job.impl.TestShuffleProvider |
|   | hadoop.mapreduce.v2.app.TestJobEndNotifier |
|   | hadoop.mapreduce.v2.app.TestMRApp |
|   | hadoop.hdfs.server.blockmanagement.TestBlockManager |
| Timed out tests | org.apache.hadoop.mapreduce.v2.app.job.impl.TestTaskAttempt 
|
| Failed build | hadoop-yarn-common |
|   | hadoop-yarn-server-resourcemanager |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12754880/MAPREDUCE-6468-04.patch
 |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / 0113e45 |
| hadoop-mapreduce-client-app test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5976/artifact/patchprocess/testrun_hadoop-mapreduce-client-app.txt
 |
| hadoop-distcp test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5976/artifact/patchprocess/testrun_hadoop-distcp.txt
 |
| hadoop-gridmix test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5976/artifact/patchprocess/testrun_hadoop-gridmix.txt
 |
| hadoop-yarn-client test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5976/artifact/patchprocess/testrun_hadoop-yarn-client.txt
 |
| hadoop-yarn-common test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5976/artifact/patchprocess/testrun_hadoop-yarn-common.txt
 |
| hadoop-yarn-server-resourcemanager test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5976/artifact/patchprocess/testrun_hadoop-yarn-server-resourcemanager.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5976/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5976/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf905.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-MAPREDUCE-Build/5976/console |


This message was automatically generated.

> Consistent log severity level guards and statements in MapReduce project
> 
>
> Key: MAPREDUCE-6468
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6468
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>Reporter: Jackie Chang
>Assignee: Jagadesh Kiran N
>Priority: Minor
>  Labels: BB2015-05-TBR
> Attachments: HADOOP-9995-00.patch, 

[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14737892#comment-14737892
 ] 

Karthik Kambatla commented on MAPREDUCE-6415:
-

Checking this in. (Jason, sorry for the jumping the gun here).

> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14738006#comment-14738006
 ] 

Hudson commented on MAPREDUCE-6415:
---

FAILURE: Integrated in Hadoop-Yarn-trunk #1102 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/1102/])
MAPREDUCE-6415. Create a tool to combine aggregated logs into HAR files. 
(Robert Kanter via kasha) (kasha: rev 119cc75e7ebd723790f6326498383304aba384a2)
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogs.java
* MAPREDUCE-6415.003.patch
* hadoop-mapreduce-project/CHANGES.txt
* hadoop-assemblies/src/main/resources/assemblies/hadoop-tools.xml
* hadoop-mapreduce-project/bin/mapred
* hadoop-tools/pom.xml
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogsRunner.java
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogsRunner.java
* hadoop-tools/hadoop-archive-logs/pom.xml
* hadoop-project/pom.xml
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogs.java
* hadoop-tools/hadoop-tools-dist/pom.xml


> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Fix For: 2.8.0
>
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Robert Kanter (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14737909#comment-14737909
 ] 

Robert Kanter commented on MAPREDUCE-6415:
--

Thanks everyone!  I'm glad we finally have a workable solution to this issue in 
now

> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14737978#comment-14737978
 ] 

Hudson commented on MAPREDUCE-6415:
---

FAILURE: Integrated in Hadoop-trunk-Commit #8426 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/8426/])
removing accidental file in MAPREDUCE-6415 (rkanter: rev 
f15371062f1bbcbb79bf44fd67ec647020d56c69)
* MAPREDUCE-6415.003.patch


> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Fix For: 2.8.0
>
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14737907#comment-14737907
 ] 

Hudson commented on MAPREDUCE-6415:
---

FAILURE: Integrated in Hadoop-trunk-Commit #8424 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/8424/])
MAPREDUCE-6415. Create a tool to combine aggregated logs into HAR files. 
(Robert Kanter via kasha) (kasha: rev 119cc75e7ebd723790f6326498383304aba384a2)
* hadoop-mapreduce-project/CHANGES.txt
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogsRunner.java
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogs.java
* hadoop-mapreduce-project/bin/mapred
* MAPREDUCE-6415.003.patch
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogs.java
* hadoop-tools/pom.xml
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogsRunner.java
* hadoop-assemblies/src/main/resources/assemblies/hadoop-tools.xml
* hadoop-tools/hadoop-tools-dist/pom.xml
* hadoop-project/pom.xml
* hadoop-tools/hadoop-archive-logs/pom.xml


> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14737906#comment-14737906
 ] 

Karthik Kambatla commented on MAPREDUCE-6415:
-

Committed to trunk and branch-2. Thanks [~rkanter] for this handy tool, and 
[~jlowe] for your reviews. 

> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Robert Kanter (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Robert Kanter updated MAPREDUCE-6415:
-
   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 2.8.0
   Status: Resolved  (was: Patch Available)

> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Fix For: 2.8.0
>
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14737968#comment-14737968
 ] 

Hudson commented on MAPREDUCE-6415:
---

FAILURE: Integrated in Hadoop-Mapreduce-trunk-Java8 #364 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/364/])
MAPREDUCE-6415. Create a tool to combine aggregated logs into HAR files. 
(Robert Kanter via kasha) (kasha: rev 119cc75e7ebd723790f6326498383304aba384a2)
* hadoop-tools/hadoop-tools-dist/pom.xml
* hadoop-tools/hadoop-archive-logs/pom.xml
* hadoop-project/pom.xml
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogs.java
* hadoop-mapreduce-project/CHANGES.txt
* MAPREDUCE-6415.003.patch
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogs.java
* hadoop-assemblies/src/main/resources/assemblies/hadoop-tools.xml
* hadoop-tools/pom.xml
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogsRunner.java
* hadoop-mapreduce-project/bin/mapred
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogsRunner.java


> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Fix For: 2.8.0
>
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14738136#comment-14738136
 ] 

Hudson commented on MAPREDUCE-6415:
---

SUCCESS: Integrated in Hadoop-Mapreduce-trunk #2313 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2313/])
MAPREDUCE-6415. Create a tool to combine aggregated logs into HAR files. 
(Robert Kanter via kasha) (kasha: rev 119cc75e7ebd723790f6326498383304aba384a2)
* hadoop-tools/hadoop-tools-dist/pom.xml
* hadoop-tools/hadoop-archive-logs/pom.xml
* hadoop-project/pom.xml
* MAPREDUCE-6415.003.patch
* hadoop-tools/pom.xml
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogs.java
* hadoop-mapreduce-project/bin/mapred
* hadoop-mapreduce-project/CHANGES.txt
* hadoop-assemblies/src/main/resources/assemblies/hadoop-tools.xml
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogsRunner.java
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogs.java
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogsRunner.java
removing accidental file in MAPREDUCE-6415 (rkanter: rev 
f15371062f1bbcbb79bf44fd67ec647020d56c69)
* MAPREDUCE-6415.003.patch


> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Fix For: 2.8.0
>
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14738148#comment-14738148
 ] 

Hudson commented on MAPREDUCE-6415:
---

SUCCESS: Integrated in Hadoop-Mapreduce-trunk-Java8 #365 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/365/])
removing accidental file in MAPREDUCE-6415 (rkanter: rev 
f15371062f1bbcbb79bf44fd67ec647020d56c69)
* MAPREDUCE-6415.003.patch


> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Fix For: 2.8.0
>
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14738153#comment-14738153
 ] 

Hudson commented on MAPREDUCE-6415:
---

FAILURE: Integrated in Hadoop-Hdfs-trunk #2290 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/2290/])
MAPREDUCE-6415. Create a tool to combine aggregated logs into HAR files. 
(Robert Kanter via kasha) (kasha: rev 119cc75e7ebd723790f6326498383304aba384a2)
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogsRunner.java
* hadoop-tools/hadoop-tools-dist/pom.xml
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogsRunner.java
* hadoop-tools/hadoop-archive-logs/pom.xml
* hadoop-mapreduce-project/bin/mapred
* MAPREDUCE-6415.003.patch
* hadoop-assemblies/src/main/resources/assemblies/hadoop-tools.xml
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogs.java
* hadoop-tools/pom.xml
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogs.java
* hadoop-project/pom.xml
* hadoop-mapreduce-project/CHANGES.txt
removing accidental file in MAPREDUCE-6415 (rkanter: rev 
f15371062f1bbcbb79bf44fd67ec647020d56c69)
* MAPREDUCE-6415.003.patch


> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Fix For: 2.8.0
>
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14738164#comment-14738164
 ] 

Hudson commented on MAPREDUCE-6415:
---

FAILURE: Integrated in Hadoop-Hdfs-trunk-Java8 #351 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/351/])
MAPREDUCE-6415. Create a tool to combine aggregated logs into HAR files. 
(Robert Kanter via kasha) (kasha: rev 119cc75e7ebd723790f6326498383304aba384a2)
* hadoop-tools/pom.xml
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogsRunner.java
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogsRunner.java
* hadoop-mapreduce-project/bin/mapred
* hadoop-mapreduce-project/CHANGES.txt
* hadoop-project/pom.xml
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogs.java
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogs.java
* hadoop-assemblies/src/main/resources/assemblies/hadoop-tools.xml
* hadoop-tools/hadoop-tools-dist/pom.xml
* MAPREDUCE-6415.003.patch
* hadoop-tools/hadoop-archive-logs/pom.xml
removing accidental file in MAPREDUCE-6415 (rkanter: rev 
f15371062f1bbcbb79bf44fd67ec647020d56c69)
* MAPREDUCE-6415.003.patch


> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Fix For: 2.8.0
>
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14738092#comment-14738092
 ] 

Hudson commented on MAPREDUCE-6415:
---

FAILURE: Integrated in Hadoop-Yarn-trunk-Java8 #371 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk-Java8/371/])
MAPREDUCE-6415. Create a tool to combine aggregated logs into HAR files. 
(Robert Kanter via kasha) (kasha: rev 119cc75e7ebd723790f6326498383304aba384a2)
* hadoop-tools/pom.xml
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogs.java
* hadoop-mapreduce-project/CHANGES.txt
* hadoop-project/pom.xml
* hadoop-assemblies/src/main/resources/assemblies/hadoop-tools.xml
* MAPREDUCE-6415.003.patch
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogs.java
* hadoop-tools/hadoop-tools-dist/pom.xml
* 
hadoop-tools/hadoop-archive-logs/src/test/java/org/apache/hadoop/tools/TestHadoopArchiveLogsRunner.java
* hadoop-mapreduce-project/bin/mapred
* 
hadoop-tools/hadoop-archive-logs/src/main/java/org/apache/hadoop/tools/HadoopArchiveLogsRunner.java
* hadoop-tools/hadoop-archive-logs/pom.xml
removing accidental file in MAPREDUCE-6415 (rkanter: rev 
f15371062f1bbcbb79bf44fd67ec647020d56c69)
* MAPREDUCE-6415.003.patch


> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Fix For: 2.8.0
>
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14738108#comment-14738108
 ] 

Hudson commented on MAPREDUCE-6415:
---

SUCCESS: Integrated in Hadoop-Yarn-trunk #1103 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/1103/])
removing accidental file in MAPREDUCE-6415 (rkanter: rev 
f15371062f1bbcbb79bf44fd67ec647020d56c69)
* MAPREDUCE-6415.003.patch


> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Fix For: 2.8.0
>
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-5870) Support for passing Job priority through Application Submission Context in Mapreduce Side

2015-09-09 Thread Sunil G (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5870?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14737270#comment-14737270
 ] 

Sunil G commented on MAPREDUCE-5870:


Hi [~jlowe] and [~eepayne]
Thank you very much for the comments.

Yes. Option 2 will be helpful to cover the existing functionality. And Priority 
is logically a part of JobStatus, so its good to keep that maximum possible 
way. However, if AM can get to know the Application Priority, is there any 
other advantage AM will have other than updating JobStatus. Still its good to 
do this way as we can avoid 2 RPC calls during every status query.

> Support for passing Job priority through Application Submission Context in 
> Mapreduce Side
> -
>
> Key: MAPREDUCE-5870
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5870
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: client
>Reporter: Sunil G
>Assignee: Sunil G
> Attachments: 0001-MAPREDUCE-5870.patch, 0002-MAPREDUCE-5870.patch, 
> 0003-MAPREDUCE-5870.patch, 0004-MAPREDUCE-5870.patch, 
> 0005-MAPREDUCE-5870.patch, 0006-MAPREDUCE-5870.patch, Yarn-2002.1.patch
>
>
> Job Prioirty can be set from client side as below [Configuration and api].
>   a.  JobConf.getJobPriority() and 
> Job.setPriority(JobPriority priority) 
>   b.  We can also use configuration 
> "mapreduce.job.priority".
>   Now this Job priority can be passed in Application Submission 
> context from Client side.
>   Here we can reuse the MRJobConfig.PRIORITY configuration. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6415) Create a tool to combine aggregated logs into HAR files

2015-09-09 Thread Karthik Kambatla (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6415?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14737264#comment-14737264
 ] 

Karthik Kambatla commented on MAPREDUCE-6415:
-

LGTM too. Checked with Robert on the findbugs, looks like it was broken before 
the patch as well. 

+1

> Create a tool to combine aggregated logs into HAR files
> ---
>
> Key: MAPREDUCE-6415
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6415
> Project: Hadoop Map/Reduce
>  Issue Type: New Feature
>Affects Versions: 2.8.0
>Reporter: Robert Kanter
>Assignee: Robert Kanter
> Attachments: HAR-ableAggregatedLogs_v1.pdf, MAPREDUCE-6415.001.patch, 
> MAPREDUCE-6415.002.patch, MAPREDUCE-6415.002.patch, MAPREDUCE-6415.003.patch, 
> MAPREDUCE-6415_branch-2.001.patch, MAPREDUCE-6415_branch-2.002.patch, 
> MAPREDUCE-6415_branch-2.003.patch, MAPREDUCE-6415_branch-2_prelim_001.patch, 
> MAPREDUCE-6415_branch-2_prelim_002.patch, MAPREDUCE-6415_prelim_001.patch, 
> MAPREDUCE-6415_prelim_002.patch
>
>
> While we wait for YARN-2942 to become viable, it would still be great to 
> improve the aggregated logs problem.  We can write a tool that combines 
> aggregated log files into a single HAR file per application, which should 
> solve the too many files and too many blocks problems.  See the design 
> document for details.
> See YARN-2942 for more context.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-5003) AM recovery should recreate records for attempts that were incomplete

2015-09-09 Thread Jason Lowe (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5003?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=14737484#comment-14737484
 ] 

Jason Lowe commented on MAPREDUCE-5003:
---

Thanks for updating the patch, Chang!  I tried it out with a sleep job where I 
manually failed and killed various task attempts and then killed the AM attempt 
so it would recover.  It recovers information for task attempts that had 
completed, but for task attempts that were active at the time where the AM 
failed (i.e.: those with a diagnostic of "Killed during application recovery") 
the log link is broken and the host info is missing.  That means we can't 
figure out where the task attempts were running and can't get to their logs.

Other comments on the patch:

completedTasksFromPreviousRun is probably not the best name given the code is 
now placing tasks in that collection that have not completed.

Nit: rather than returning from the middle of the TaskImpl.recover function for 
a running task, I think it would be a bit cleaner to have the code recover a 
missing taskStatus as taskState=RUNNING and add a RUNNING case to the existing 
switch statement.


> AM recovery should recreate records for attempts that were incomplete
> -
>
> Key: MAPREDUCE-5003
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-5003
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: mr-am
>Reporter: Jason Lowe
>Assignee: Chang Li
> Attachments: MAPREDUCE-5003.1.patch, MAPREDUCE-5003.2.patch, 
> MAPREDUCE-5003.3.patch, MAPREDUCE-5003.4.patch, MAPREDUCE-5003.5.patch, 
> MAPREDUCE-5003.5.patch, MAPREDUCE-5003.6.patch
>
>
> As discussed in MAPREDUCE-4992, it would be nice if the AM recovered task 
> attempt entries for *all* task attempts launched by the prior app attempt 
> even if those task attempts did not complete.  The attempts would have to be 
> marked as killed or something similar to indicate it is no longer running.  
> Having records for the task attempts enables the user to see what nodes were 
> associated with the attempts and potentially access their logs.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Work started] (MAPREDUCE-6473) Job submission can take a long time during Cluster initialization

2015-09-09 Thread Kuhu Shukla (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-6473?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on MAPREDUCE-6473 started by Kuhu Shukla.
--
> Job submission can take a long time during Cluster initialization
> -
>
> Key: MAPREDUCE-6473
> URL: https://issues.apache.org/jira/browse/MAPREDUCE-6473
> Project: Hadoop Map/Reduce
>  Issue Type: Improvement
>  Components: performance
>Reporter: Kuhu Shukla
>Assignee: Kuhu Shukla
>
> During initialization in Cluster.java, the framework provider classes are 
> loaded inside a sync block which can considerably increase job submission 
> time when the number of submissions are high. The motive is to reduce time 
> spent in this sync block safely to improve performance.
> {noformat}
> synchronized (frameworkLoader) {
>   for (ClientProtocolProvider provider : frameworkLoader) {
> LOG.debug("Trying ClientProtocolProvider : "
> + provider.getClass().getName());
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (MAPREDUCE-6473) Job submission can take a long time during Cluster initialization

2015-09-09 Thread Kuhu Shukla (JIRA)
Kuhu Shukla created MAPREDUCE-6473:
--

 Summary: Job submission can take a long time during Cluster 
initialization
 Key: MAPREDUCE-6473
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6473
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
Reporter: Kuhu Shukla
Assignee: Kuhu Shukla


During initialization in Cluster.java, the framework provider classes are 
loaded inside a sync block which can considerably increase job submission time 
when the number of submissions are high. The motive is to reduce time spent in 
this sync block safely to improve performance.
{noformat}
synchronized (frameworkLoader) {
  for (ClientProtocolProvider provider : frameworkLoader) {
LOG.debug("Trying ClientProtocolProvider : "
+ provider.getClass().getName());
{noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)