[jira] [Updated] (MAPREDUCE-5728) Check NPE for serializer/deserializer in MapTask

2014-11-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5728?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5728:
-
Fix Version/s: (was: 2.6.0)
   2.7.0

 Check NPE for serializer/deserializer in MapTask
 

 Key: MAPREDUCE-5728
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5728
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: client
Affects Versions: 2.2.0
Reporter: Jerry He
Assignee: Jerry He
Priority: Minor
 Fix For: 2.4.0, 2.7.0

 Attachments: MAPREDUCE-5728-trunk.patch


 Currently we will get NPE if the serializer/deserializer is not configured 
 correctly.
 {code}
 14/01/14 11:52:35 INFO mapred.JobClient: Task Id : 
 attempt_201401072154_0027_m_02_2, Status : FAILED
 java.lang.NullPointerException
 at 
 org.apache.hadoop.mapred.MapTask$MapOutputBuffer.init(MapTask.java:944)
 at 
 org.apache.hadoop.mapred.MapTask$NewOutputCollector.init(MapTask.java:672)
 at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:740)
 at org.apache.hadoop.mapred.MapTask.run(MapTask.java:368)
 at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
 at 
 java.security.AccessController.doPrivileged(AccessController.java:362)
 at javax.security.auth.Subject.doAs(Subject.java:573)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1502)
 at org.apache.hadoop.mapred.Child.main(Child.java:249)
 {code}
 serializationFactory.getSerializer and serializationFactory.getDeserializer 
 returns NULL in this case.
 Let's check NPE for serializer/deserializer in MapTask so that we don't get 
 meaningless NPE.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6156) Fetcher - connect() doesn't handle connection refused correctly

2014-11-13 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6156?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14210145#comment-14210145
 ] 

Arun C Murthy commented on MAPREDUCE-6156:
--

[~jlowe] Thanks. I've merged this into branch-2.6.0 also for hadoop-2.6.0-rc1.

 Fetcher - connect() doesn't handle connection refused correctly 
 

 Key: MAPREDUCE-6156
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6156
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Reporter: Sidharta Seethana
Assignee: Junping Du
Priority: Blocker
 Fix For: 2.6.0

 Attachments: MAPREDUCE-6156-v2.patch, MAPREDUCE-6156-v3.patch, 
 MAPREDUCE-6156.patch


 The connect() function in the fetcher assumes that whenever an IOException is 
 thrown, the amount of time passed equals connectionTimeout ( see code 
 snippet below ). This is incorrect. For example, in case the NM is down, an 
 ConnectException is thrown immediately - and the catch block assumes a minute 
 has passed when it is not the case.
 {code}
   if (connectionTimeout  0) {
   throw new IOException(Invalid timeout 
 + [timeout =  + connectionTimeout +  ms]);
 } else if (connectionTimeout  0) {
   unit = Math.min(UNIT_CONNECT_TIMEOUT, connectionTimeout);
 }
 // set the connect timeout to the unit-connect-timeout
 connection.setConnectTimeout(unit);
 while (true) {
   try {
 connection.connect();
 break;
   } catch (IOException ioe) {
 // update the total remaining connect-timeout
 connectionTimeout -= unit;
 // throw an exception if we have waited for timeout amount of time
 // note that the updated value if timeout is used here
 if (connectionTimeout == 0) {
   throw ioe;
 }
 // reset the connect timeout for the last try
 if (connectionTimeout  unit) {
   unit = connectionTimeout;
   // reset the connect time out for the final connect
   connection.setConnectTimeout(unit);
 }
   }
 }
 {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-6083) Map/Reduce dangerously adds Guava @Beta class to CryptoUtils

2014-11-05 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-6083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14198360#comment-14198360
 ] 

Arun C Murthy commented on MAPREDUCE-6083:
--

[~ctubbsii] - Apologies for the late response. Unfortunately, we can't change 
guava versions in 2.6 since it would be incompatible. 

 Map/Reduce dangerously adds Guava @Beta class to CryptoUtils
 

 Key: MAPREDUCE-6083
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6083
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Affects Versions: 2.6.0
Reporter: Christopher Tubbs
Priority: Blocker
  Labels: beta, deprecated, guava
 Attachments: 
 0001-MAPREDUCE-6083-Avoid-client-use-of-deprecated-LimitI.patch


 See HDFS-7040 for more background/details.
 In recent 2.6.0-SNAPSHOTs, the use of LimitInputStream was added to 
 CryptoUtils. This is part of the API components of Hadoop, which severely 
 impacts users who were utilizing newer versions of Guava, where the @Beta and 
 @Deprecated class, LimitInputStream, has been removed (removed in version 15 
 and later), beyond the impact already experienced in 2.4.0 as identified in 
 HDFS-7040.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (MAPREDUCE-6083) Map/Reduce dangerously adds Guava @Beta class to CryptoUtils

2014-11-05 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-6083?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-6083:
-
Status: Open  (was: Patch Available)

Yes, I'm cool with that. Tx!

 Map/Reduce dangerously adds Guava @Beta class to CryptoUtils
 

 Key: MAPREDUCE-6083
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-6083
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Affects Versions: 2.6.0
Reporter: Christopher Tubbs
Priority: Blocker
  Labels: beta, deprecated, guava
 Attachments: 
 0001-MAPREDUCE-6083-Avoid-client-use-of-deprecated-LimitI.patch


 See HDFS-7040 for more background/details.
 In recent 2.6.0-SNAPSHOTs, the use of LimitInputStream was added to 
 CryptoUtils. This is part of the API components of Hadoop, which severely 
 impacts users who were utilizing newer versions of Guava, where the @Beta and 
 @Deprecated class, LimitInputStream, has been removed (removed in version 15 
 and later), beyond the impact already experienced in 2.4.0 as identified in 
 HDFS-7040.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (MAPREDUCE-2841) Task level native optimization

2014-06-24 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-2841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14041754#comment-14041754
 ] 

Arun C Murthy commented on MAPREDUCE-2841:
--

I also noticed that the github has a bunch of code related to Pig, Hive etc. - 
I think we'd all agree that they need to be in respective projects eventually.

 Task level native optimization
 --

 Key: MAPREDUCE-2841
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-2841
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: task
 Environment: x86-64 Linux/Unix
Reporter: Binglin Chang
Assignee: Sean Zhong
 Attachments: DESIGN.html, MAPREDUCE-2841.v1.patch, 
 MAPREDUCE-2841.v2.patch, dualpivot-0.patch, dualpivotv20-0.patch, 
 fb-shuffle.patch


 I'm recently working on native optimization for MapTask based on JNI. 
 The basic idea is that, add a NativeMapOutputCollector to handle k/v pairs 
 emitted by mapper, therefore sort, spill, IFile serialization can all be done 
 in native code, preliminary test(on Xeon E5410, jdk6u24) showed promising 
 results:
 1. Sort is about 3x-10x as fast as java(only binary string compare is 
 supported)
 2. IFile serialization speed is about 3x of java, about 500MB/s, if hardware 
 CRC32C is used, things can get much faster(1G/
 3. Merge code is not completed yet, so the test use enough io.sort.mb to 
 prevent mid-spill
 This leads to a total speed up of 2x~3x for the whole MapTask, if 
 IdentityMapper(mapper does nothing) is used
 There are limitations of course, currently only Text and BytesWritable is 
 supported, and I have not think through many things right now, such as how to 
 support map side combine. I had some discussion with somebody familiar with 
 hive, it seems that these limitations won't be much problem for Hive to 
 benefit from those optimizations, at least. Advices or discussions about 
 improving compatibility are most welcome:) 
 Currently NativeMapOutputCollector has a static method called canEnable(), 
 which checks if key/value type, comparator type, combiner are all compatible, 
 then MapTask can choose to enable NativeMapOutputCollector.
 This is only a preliminary test, more work need to be done. I expect better 
 final results, and I believe similar optimization can be adopt to reduce task 
 and shuffle too. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (MAPREDUCE-2841) Task level native optimization

2014-06-24 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-2841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14041752#comment-14041752
 ] 

Arun C Murthy commented on MAPREDUCE-2841:
--

{quote}
If the MR developer community generally agrees this belongs in the core, I'd 
like to start a feature branch for it in order to import the current code, sort 
out the build/integration issues, and take care of the remaining items that 
Sean mentioned above.
{quote}

[~tlipcon] Thanks for starting this discussion. I have a few thoughts I'd like 
to run by you.

I think the eventual goal of this (looking at 
https://github.com/intel-hadoop/nativetask/blob/master/README.md) is a 
full-native runtime for MapReduce including sort, shuffle, merge etc.

Hence, it does look like we will achieve a compatible, but alternate 
implementation of MapReduce runtime. Hence, this is similar to other alternate 
runtimes for MapReduce such as Apache Tez.

Furthermore, this is implemented in C++ - which is, frankly, a concern for the 
poor job C++ has done with ABI. I'm glad to see that it doesn't rely on boost - 
the worst affender. This is the same reason the native Hadoop client 
(HADOOP-10388) is being done purely in C. Also, the MR development community is 
pre-dominantly Java, which is something to keep in mind. This is a big concern 
for me.

In all, it seems to me we could consider having this not in Apache Hadoop, but 
as an incubator project to develop a native, MR compatible runtime. 

This will allow it to develop a like-minded community (C++ skills etc.) and not 
be bogged down by *all* of Hadoop's requirements such as security (how/when 
will this allow for secure shuffle or encrypted shuffle etc.), compatibility 
with several OSes (flavours of Linux, MacOSX, Windows) etc. It will also allow 
them to ship independently and get user feedback more quickly.

Similarly, I am wary of importing a nearly 75K LOC codebase into a stable 
project and it's impact on our releases on breakage - particularly given the 
difference in skills of the community i.e. Java v/s C++ etc.

What do you think Todd  Sean? I'm more than happy to help with incubator 
process if required.

 Task level native optimization
 --

 Key: MAPREDUCE-2841
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-2841
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: task
 Environment: x86-64 Linux/Unix
Reporter: Binglin Chang
Assignee: Sean Zhong
 Attachments: DESIGN.html, MAPREDUCE-2841.v1.patch, 
 MAPREDUCE-2841.v2.patch, dualpivot-0.patch, dualpivotv20-0.patch, 
 fb-shuffle.patch


 I'm recently working on native optimization for MapTask based on JNI. 
 The basic idea is that, add a NativeMapOutputCollector to handle k/v pairs 
 emitted by mapper, therefore sort, spill, IFile serialization can all be done 
 in native code, preliminary test(on Xeon E5410, jdk6u24) showed promising 
 results:
 1. Sort is about 3x-10x as fast as java(only binary string compare is 
 supported)
 2. IFile serialization speed is about 3x of java, about 500MB/s, if hardware 
 CRC32C is used, things can get much faster(1G/
 3. Merge code is not completed yet, so the test use enough io.sort.mb to 
 prevent mid-spill
 This leads to a total speed up of 2x~3x for the whole MapTask, if 
 IdentityMapper(mapper does nothing) is used
 There are limitations of course, currently only Text and BytesWritable is 
 supported, and I have not think through many things right now, such as how to 
 support map side combine. I had some discussion with somebody familiar with 
 hive, it seems that these limitations won't be much problem for Hive to 
 benefit from those optimizations, at least. Advices or discussions about 
 improving compatibility are most welcome:) 
 Currently NativeMapOutputCollector has a static method called canEnable(), 
 which checks if key/value type, comparator type, combiner are all compatible, 
 then MapTask can choose to enable NativeMapOutputCollector.
 This is only a preliminary test, more work need to be done. I expect better 
 final results, and I believe similar optimization can be adopt to reduce task 
 and shuffle too. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (MAPREDUCE-2841) Task level native optimization

2014-06-24 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-2841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14041757#comment-14041757
 ] 

Arun C Murthy commented on MAPREDUCE-2841:
--

On related thought to Pig/Hive etc. - I see Hadoop MapReduce fading away fast 
particularly since projects using MR such as Pig, Hive, Cascading etc. 
re-vector on other projects like Apache Tez or Apache Spark.

For e.g.
# Hive-on-Tez (https://issues.apache.org/jira/browse/HIVE-4660) - The hive 
community has already moved it's major investments away from MR to Tez.
# Pig-on-Tez (https://issues.apache.org/jira/browse/PIG-3446) - The pig 
community is very close to shipping this in pig-0.14 and again is investing 
heavily on Tez.

Given that, Sean/Todd, would it be useful to discuss contributing this to Tez 
instead? 

This way the work here would continue to stay relevant in the context of the 
majority users of MapReduce who use Pig, Hive, Cascading etc. 

Of course, I'm sure another option is Apache Spark, but given that Tez is much 
more closer (code-base wise) to MR, it would be much easier to contribute to 
Tez. Happy to help if that makes sense too. Thanks.

 Task level native optimization
 --

 Key: MAPREDUCE-2841
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-2841
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: task
 Environment: x86-64 Linux/Unix
Reporter: Binglin Chang
Assignee: Sean Zhong
 Attachments: DESIGN.html, MAPREDUCE-2841.v1.patch, 
 MAPREDUCE-2841.v2.patch, dualpivot-0.patch, dualpivotv20-0.patch, 
 fb-shuffle.patch


 I'm recently working on native optimization for MapTask based on JNI. 
 The basic idea is that, add a NativeMapOutputCollector to handle k/v pairs 
 emitted by mapper, therefore sort, spill, IFile serialization can all be done 
 in native code, preliminary test(on Xeon E5410, jdk6u24) showed promising 
 results:
 1. Sort is about 3x-10x as fast as java(only binary string compare is 
 supported)
 2. IFile serialization speed is about 3x of java, about 500MB/s, if hardware 
 CRC32C is used, things can get much faster(1G/
 3. Merge code is not completed yet, so the test use enough io.sort.mb to 
 prevent mid-spill
 This leads to a total speed up of 2x~3x for the whole MapTask, if 
 IdentityMapper(mapper does nothing) is used
 There are limitations of course, currently only Text and BytesWritable is 
 supported, and I have not think through many things right now, such as how to 
 support map side combine. I had some discussion with somebody familiar with 
 hive, it seems that these limitations won't be much problem for Hive to 
 benefit from those optimizations, at least. Advices or discussions about 
 improving compatibility are most welcome:) 
 Currently NativeMapOutputCollector has a static method called canEnable(), 
 which checks if key/value type, comparator type, combiner are all compatible, 
 then MapTask can choose to enable NativeMapOutputCollector.
 This is only a preliminary test, more work need to be done. I expect better 
 final results, and I believe similar optimization can be adopt to reduce task 
 and shuffle too. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (MAPREDUCE-2841) Task level native optimization

2014-06-24 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-2841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14041816#comment-14041816
 ] 

Arun C Murthy commented on MAPREDUCE-2841:
--

Todd,

bq. I agree that building a completely parallel C++ MR runtime is a much larger 
project that should not be part of Hadoop. 

I'm confused. There already exists large amounts of code on the github for a 
the full task runtime. Is that abandoned? Are you saying there no intention to 
contribute that to Hadoop, ever? Why would that be? Would that be a separate 
project?

With or without ABI, C++ still is a major problem w.r.t different compiler 
versions, different platforms we support etc. That is precisely why 
HADOOP-10388 chose to use pure-C only. A similar switch makes me *much* more 
comfortable, aside from the disparity in skills in the Hadoop community. 

Furthermore, there are considerably more security issues which open up in C++ 
land such as buffer overflow etc.



bq. I think the 75k you're counting may include the auto-generated shell 
scripts.

From the github:

{noformat}
$ find . -name *.java | xargs wc -l
   11988 total
$ find . -name *.h | xargs wc -l
   27269 total
$ find . -name *.cc | xargs wc -l
   26276 total
{noformat}

Whether it's test or non-test, we are still importing a *lot* of code - code 
for which the Hadoop community does need to maintain?



bq. So, it's not a tiny import by any means, but for 2x improvement on terasort 
wallclock, my opinion is that the maintenance burden is worth it.

Todd, as we both know, there are many, many ways to get 2x improvement on 
terasort...
... nor is it worth a lot in real-world outside of benchmarks. 

I'm sure we both would take 2x on Pig/Hive anyday... *smile*



bq. As for importing to Tez, I don't think the community has generally agreed 
to EOL MapReduce

Regardless of whether or not we pull this into MR, it would be useful to pull 
it into Tez too - if Sean wants to do it. Let's not discourage them.

I'm sure we both agree, and want to see real world workloads improve and that 
Hive/Pig/Cascading etc. represent that.

IAC, hopefully we can stop this meme that I'm trying to *preclude* you from 
doing anything regardless of my religious beliefs. IAC, we both realize MR is 
reasonably stable and won't get a lot of investment, and so do our employers:
http://vision.cloudera.com/mapreduce-spark/
http://hortonworks.com/hadoop/tez/

So, I'd appreciate if we don't misinterpret each others' technical opinions and 
concerns during this discussion. Thanks.

FTR: I'll restate my concerns about C++, roadmap for C++ runtime, 
maintainability, support for all of Hadoop (security, platforms etc.). 

Furthermore, this jira was opened nearly 3 years ago and only has sporadic 
bursts of activity - not a good sign for long-term maintainability.

I've stated my concerns, let's try get through them by focussing on those 
aspects.



Finally, what is the concern you see with starting this as an incubator project 
and allowing folks to develop a community around it? We can certainly help on 
our end by making it easy for them to plug in via interfaces etc. 

Thanks.

 Task level native optimization
 --

 Key: MAPREDUCE-2841
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-2841
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: task
 Environment: x86-64 Linux/Unix
Reporter: Binglin Chang
Assignee: Sean Zhong
 Attachments: DESIGN.html, MAPREDUCE-2841.v1.patch, 
 MAPREDUCE-2841.v2.patch, dualpivot-0.patch, dualpivotv20-0.patch, 
 fb-shuffle.patch


 I'm recently working on native optimization for MapTask based on JNI. 
 The basic idea is that, add a NativeMapOutputCollector to handle k/v pairs 
 emitted by mapper, therefore sort, spill, IFile serialization can all be done 
 in native code, preliminary test(on Xeon E5410, jdk6u24) showed promising 
 results:
 1. Sort is about 3x-10x as fast as java(only binary string compare is 
 supported)
 2. IFile serialization speed is about 3x of java, about 500MB/s, if hardware 
 CRC32C is used, things can get much faster(1G/
 3. Merge code is not completed yet, so the test use enough io.sort.mb to 
 prevent mid-spill
 This leads to a total speed up of 2x~3x for the whole MapTask, if 
 IdentityMapper(mapper does nothing) is used
 There are limitations of course, currently only Text and BytesWritable is 
 supported, and I have not think through many things right now, such as how to 
 support map side combine. I had some discussion with somebody familiar with 
 hive, it seems that these limitations won't be much problem for Hive to 
 benefit from those optimizations, at least. Advices or discussions about 
 improving compatibility are most welcome:) 
 Currently NativeMapOutputCollector has a 

[jira] [Comment Edited] (MAPREDUCE-2841) Task level native optimization

2014-06-24 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-2841?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14041816#comment-14041816
 ] 

Arun C Murthy edited comment on MAPREDUCE-2841 at 6/24/14 8:30 AM:
---

Todd,

bq. I agree that building a completely parallel C++ MR runtime is a much larger 
project that should not be part of Hadoop. 

I'm confused. There already exists large amounts of code on the github for a 
the full task runtime. Is that abandoned? Are you saying there no intention to 
contribute that to Hadoop, ever? Why would that be? Would that be a separate 
project?

With or without ABI, C++ still is a major problem w.r.t different compiler 
versions, different platforms we support etc. That is precisely why 
HADOOP-10388 chose to use pure-C only. A similar switch makes me *much* more 
comfortable, aside from the disparity in skills in the Hadoop community. 

Furthermore, there are considerably more security issues which open up in C++ 
land such as buffer overflow etc.



bq. I think the 75k you're counting may include the auto-generated shell 
scripts.

From the github:

{noformat}
$ find . -name *.java | xargs wc -l
   11988 total
$ find . -name *.h | xargs wc -l
   27269 total
$ find . -name *.cc | xargs wc -l
   26276 total
{noformat}

Whether it's test or non-test, we are still importing a *lot* of code - code 
for which the Hadoop community does need to maintain?



bq. So, it's not a tiny import by any means, but for 2x improvement on terasort 
wallclock, my opinion is that the maintenance burden is worth it.

Todd, as we both know, there are many, many ways to get 2x improvement on 
terasort...
... nor is it worth a lot in real-world outside of benchmarks. 

I'm sure we both would take 2x on Pig/Hive anyday... *smile*



bq. As for importing to Tez, I don't think the community has generally agreed 
to EOL MapReduce

Regardless of whether or not we pull this into MR, it would be useful to pull 
it into Tez too - if Sean wants to do it. Let's not discourage them.

I'm sure we both agree, and want to see real world workloads improve and that 
Hive/Pig/Cascading etc. represent that.

IAC, hopefully we can stop this meme that I'm trying to *preclude* you from 
doing anything regardless of my beliefs. IAC, we both realize MR is reasonably 
stable and won't get a lot of investment, and so do our employers:
http://vision.cloudera.com/mapreduce-spark/
http://hortonworks.com/hadoop/tez/

Essentially, you asked for feedback from the MapReduce community; and this is 
my honest feedback - as someone who has actively helped maintain this codebase 
for more than 8 years now. So, I'd appreciate if we don't misinterpret each 
others' technical opinions and concerns during this discussion. Thanks in 
advance.

FTR: I'll restate my concerns about C++, roadmap for C++ runtime, 
maintainability, support for all of Hadoop (new security bugs, future security 
features, platforms etc.). 

Furthermore, this jira was opened nearly 3 years ago and only has sporadic 
bursts of activity - not a good sign for long-term maintainability.

I've stated my concerns, let's try get through them by focussing on those 
aspects.



Finally, what is the concern you see with starting this as an incubator project 
and allowing folks to develop a community around it? We can certainly help on 
our end by making it easy for them to plug in via interfaces etc. 

Thanks.


was (Author: acmurthy):
Todd,

bq. I agree that building a completely parallel C++ MR runtime is a much larger 
project that should not be part of Hadoop. 

I'm confused. There already exists large amounts of code on the github for a 
the full task runtime. Is that abandoned? Are you saying there no intention to 
contribute that to Hadoop, ever? Why would that be? Would that be a separate 
project?

With or without ABI, C++ still is a major problem w.r.t different compiler 
versions, different platforms we support etc. That is precisely why 
HADOOP-10388 chose to use pure-C only. A similar switch makes me *much* more 
comfortable, aside from the disparity in skills in the Hadoop community. 

Furthermore, there are considerably more security issues which open up in C++ 
land such as buffer overflow etc.



bq. I think the 75k you're counting may include the auto-generated shell 
scripts.

From the github:

{noformat}
$ find . -name *.java | xargs wc -l
   11988 total
$ find . -name *.h | xargs wc -l
   27269 total
$ find . -name *.cc | xargs wc -l
   26276 total
{noformat}

Whether it's test or non-test, we are still importing a *lot* of code - code 
for which the Hadoop community does need to maintain?



bq. So, it's not a tiny import by any means, but for 2x improvement on terasort 
wallclock, my opinion is that the maintenance burden is worth it.

Todd, as we both know, there are many, many ways to get 2x improvement on 

[jira] [Commented] (MAPREDUCE-5831) Old MR client is not compatible with new MR application

2014-06-20 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14039136#comment-14039136
 ] 

Arun C Murthy commented on MAPREDUCE-5831:
--

I think the real way, going fwd, to fix this is to ensure the MR client and the 
application are on the same version - MAPREDUCE-4421 should solve that; and be 
the default deployment model.

 Old MR client is not compatible with new MR application
 ---

 Key: MAPREDUCE-5831
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5831
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: client, mr-am
Affects Versions: 2.2.0, 2.3.0
Reporter: Zhijie Shen
Assignee: Tan, Wangda
Priority: Critical

 Recently, we saw the following scenario:
 1. The user setup a cluster of hadoop 2.3., which contains YARN 2.3 and MR  
 2.3.
 2. The user client on a machine that MR 2.2 is installed and in the classpath.
 Then, when the user submitted a simple wordcount job, he saw the following 
 message:
 {code}
 16:00:41,027  INFO main mapreduce.Job:1345 -  map 100% reduce 100%
 16:00:41,036  INFO main mapreduce.Job:1356 - Job job_1396468045458_0006 
 completed successfully
 16:02:20,535  WARN main mapreduce.JobRunner:212 - Cannot start job 
 [wordcountJob]
 java.lang.IllegalArgumentException: No enum constant 
 org.apache.hadoop.mapreduce.JobCounter.MB_MILLIS_REDUCES
   at java.lang.Enum.valueOf(Enum.java:236)
   at 
 org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.valueOf(FrameworkCounterGroup.java:148)
   at 
 org.apache.hadoop.mapreduce.counters.FrameworkCounterGroup.findCounter(FrameworkCounterGroup.java:182)
   at 
 org.apache.hadoop.mapreduce.counters.AbstractCounters.findCounter(AbstractCounters.java:154)
   at 
 org.apache.hadoop.mapreduce.TypeConverter.fromYarn(TypeConverter.java:240)
   at 
 org.apache.hadoop.mapred.ClientServiceDelegate.getJobCounters(ClientServiceDelegate.java:370)
   at 
 org.apache.hadoop.mapred.YARNRunner.getJobCounters(YARNRunner.java:511)
   at org.apache.hadoop.mapreduce.Job$7.run(Job.java:756)
   at org.apache.hadoop.mapreduce.Job$7.run(Job.java:753)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:415)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1491)
   at org.apache.hadoop.mapreduce.Job.getCounters(Job.java:753)
   at org.apache.hadoop.mapreduce.Job.monitorAndPrintJob(Job.java:1361)
   at org.apache.hadoop.mapreduce.Job.waitForCompletion(Job.java:1289)
 . . .
 {code}
 The problem is that the wordcount job was running on one or more than one 
 nodes of the YARN cluster, where MR 2.3 libs were installed, and 
 JobCounter.MB_MILLIS_REDUCES is available in the counters. On the other side, 
 due to the classpath setting, the client was likely to run with MR 2.2 libs. 
 After the client retrieved the counters from MR AM, it tried to construct the 
 Counter object with the received counter name. Unfortunately, the enum didn't 
 exist in the client's classpath. Therefore, No enum constant exception is 
 thrown here.
 JobCounter.MB_MILLIS_REDUCES is brought to MR2 via MAPREDUCE-5464 since 
 Hadoop 2.3.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (MAPREDUCE-5728) Check NPE for serializer/deserializer in MapTask

2014-04-10 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5728?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5728:
-

Fix Version/s: (was: 2.2.1)
   2.5.0

 Check NPE for serializer/deserializer in MapTask
 

 Key: MAPREDUCE-5728
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5728
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: client
Affects Versions: 2.2.0
Reporter: Jerry He
Assignee: Jerry He
Priority: Minor
 Fix For: 2.4.0, 2.5.0

 Attachments: MAPREDUCE-5728-trunk.patch


 Currently we will get NPE if the serializer/deserializer is not configured 
 correctly.
 {code}
 14/01/14 11:52:35 INFO mapred.JobClient: Task Id : 
 attempt_201401072154_0027_m_02_2, Status : FAILED
 java.lang.NullPointerException
 at 
 org.apache.hadoop.mapred.MapTask$MapOutputBuffer.init(MapTask.java:944)
 at 
 org.apache.hadoop.mapred.MapTask$NewOutputCollector.init(MapTask.java:672)
 at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:740)
 at org.apache.hadoop.mapred.MapTask.run(MapTask.java:368)
 at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
 at 
 java.security.AccessController.doPrivileged(AccessController.java:362)
 at javax.security.auth.Subject.doAs(Subject.java:573)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1502)
 at org.apache.hadoop.mapred.Child.main(Child.java:249)
 {code}
 serializationFactory.getSerializer and serializationFactory.getDeserializer 
 returns NULL in this case.
 Let's check NPE for serializer/deserializer in MapTask so that we don't get 
 meaningless NPE.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (MAPREDUCE-5815) Fix NPE in TestMRAppMaster

2014-03-29 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5815?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5815:
-

Priority: Blocker  (was: Major)

 Fix NPE in TestMRAppMaster
 --

 Key: MAPREDUCE-5815
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5815
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: client, mrv2
Reporter: Gera Shegalov
Assignee: Gera Shegalov
Priority: Blocker
 Attachments: MAPREDUCE-5815.v01.patch


 Working MAPREDUCE-5813 I stumbled on NPE's in TestMRAppMaster. They seem to 
 be introduced by MAPREDUCE-5805.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (MAPREDUCE-5815) Fix NPE in TestMRAppMaster

2014-03-29 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5815?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13954377#comment-13954377
 ] 

Arun C Murthy commented on MAPREDUCE-5815:
--

[~jira.shegalov] - I'm not sure we should quietly ignore the queue-name... 
maybe the right fix is to ensure TestMRAppMaster actually sets a proper queue 
name? 

[~vinodkv] - Thoughts?

 Fix NPE in TestMRAppMaster
 --

 Key: MAPREDUCE-5815
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5815
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: client, mrv2
Reporter: Gera Shegalov
Assignee: Gera Shegalov
Priority: Blocker
 Attachments: MAPREDUCE-5815.v01.patch


 Working MAPREDUCE-5813 I stumbled on NPE's in TestMRAppMaster. They seem to 
 be introduced by MAPREDUCE-5805.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (MAPREDUCE-5815) Fix NPE in TestMRAppMaster

2014-03-29 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5815?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5815:
-

Status: Open  (was: Patch Available)

 Fix NPE in TestMRAppMaster
 --

 Key: MAPREDUCE-5815
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5815
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: client, mrv2
Reporter: Gera Shegalov
Assignee: Gera Shegalov
Priority: Blocker
 Attachments: MAPREDUCE-5815.v01.patch


 Working MAPREDUCE-5813 I stumbled on NPE's in TestMRAppMaster. They seem to 
 be introduced by MAPREDUCE-5805.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (MAPREDUCE-5813) YarnChild does not load job.xml with mapreduce.job.classloader=true

2014-03-29 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5813?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5813:
-

   Resolution: Fixed
Fix Version/s: 2.4.0
   Status: Resolved  (was: Patch Available)

I just committed this. Thanks [~jira.shegalov]!

 YarnChild does not load job.xml with mapreduce.job.classloader=true 
 

 Key: MAPREDUCE-5813
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5813
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv2, task
Affects Versions: 2.3.0
Reporter: Gera Shegalov
Assignee: Gera Shegalov
Priority: Blocker
 Fix For: 2.4.0

 Attachments: MAPREDUCE-5813.v01.patch, MAPREDUCE-5813.v02.patch


 {{YarnChild.main}} uses {{JobConf.addResource(String)}} to load {{job.xml}} 
 that relies on class loading. When {{mapreduce.job.classloader=true}} the 
 job-speicific part of the class path is separated from {{CLASSPATH}} into 
 {{APP_CLASSPATH}}. Therefore {{job.xml}} is inaccessible for the default 
 class loader. Later {{writeLocalJobFile}} overwrites the correct localized 
 {{job.xml}} on disk as well.
 This problem is easily avoided by using  {{JobConf.addResource(Path)}} to 
 read the localized {{job.xml}} without relying on class loading.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (MAPREDUCE-5813) YarnChild does not load job.xml with mapreduce.job.classloader=true

2014-03-28 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5813?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5813:
-

Status: Open  (was: Patch Available)

Agree w/ [~sjlee0]. 

[~jira.shegalov] - mind updating the patch? Tx.

 YarnChild does not load job.xml with mapreduce.job.classloader=true 
 

 Key: MAPREDUCE-5813
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5813
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv2, task
Affects Versions: 2.3.0
Reporter: Gera Shegalov
Assignee: Gera Shegalov
Priority: Blocker
 Attachments: MAPREDUCE-5813.v01.patch


 {{YarnChild.main}} uses {{JobConf.addResource(String)}} to load {{job.xml}} 
 that relies on class loading. When {{mapreduce.job.classloader=true}} the 
 job-speicific part of the class path is separated from {{CLASSPATH}} into 
 {{APP_CLASSPATH}}. Therefore {{job.xml}} is inaccessible for the default 
 class loader. Later {{writeLocalJobFile}} overwrites the correct localized 
 {{job.xml}} on disk as well.
 This problem is easily avoided by using  {{JobConf.addResource(Path)}} to 
 read the localized {{job.xml}} without relying on class loading.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (MAPREDUCE-5553) Add task state filters on Application/MRJob page for MR Application master

2014-03-12 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5553?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5553:
-

Assignee: Paul Han

 Add task state filters on Application/MRJob page for MR Application master 
 ---

 Key: MAPREDUCE-5553
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5553
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: applicationmaster
Affects Versions: 2.0.5-alpha
Reporter: Paul Han
Assignee: Paul Han
Priority: Minor
 Attachments: FilteredTaskPage.png, MAPREDUCE-5553.patch, 
 MRJobWithTaskStateFilter.png, MRJobWithoutTaskStateFilters.png


 On Job page of MR application master, the task attempts have a nice breakdown 
 of different state: running, failed But for map/reduce task level, 
 there's only one links to all tasks without breakdown for different states:
 pending, running, complete, etc.
 With Hadoop 1, the similar functionality exists: user can go thru 
 jobtracker- job- map/reduce tasks - attempts.
 This ticket will improve the usability to allow user to go job- total, 
 pending, running, complete Map/Reduce task.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (MAPREDUCE-5553) Add task state filters on Application/MRJob page for MR Application master

2014-03-12 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5553?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13932480#comment-13932480
 ] 

Arun C Murthy commented on MAPREDUCE-5553:
--

I just committed this. Apologies for taking so long to get back to you - thanks 
[~paulhan]!

 Add task state filters on Application/MRJob page for MR Application master 
 ---

 Key: MAPREDUCE-5553
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5553
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: applicationmaster
Affects Versions: 2.0.5-alpha
Reporter: Paul Han
Assignee: Paul Han
Priority: Minor
 Fix For: 2.4.0

 Attachments: FilteredTaskPage.png, MAPREDUCE-5553.patch, 
 MRJobWithTaskStateFilter.png, MRJobWithoutTaskStateFilters.png


 On Job page of MR application master, the task attempts have a nice breakdown 
 of different state: running, failed But for map/reduce task level, 
 there's only one links to all tasks without breakdown for different states:
 pending, running, complete, etc.
 With Hadoop 1, the similar functionality exists: user can go thru 
 jobtracker- job- map/reduce tasks - attempts.
 This ticket will improve the usability to allow user to go job- total, 
 pending, running, complete Map/Reduce task.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (MAPREDUCE-5553) Add task state filters on Application/MRJob page for MR Application master

2014-03-12 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5553?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5553:
-

   Resolution: Fixed
Fix Version/s: 2.4.0
   Status: Resolved  (was: Patch Available)

 Add task state filters on Application/MRJob page for MR Application master 
 ---

 Key: MAPREDUCE-5553
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5553
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: applicationmaster
Affects Versions: 2.0.5-alpha
Reporter: Paul Han
Assignee: Paul Han
Priority: Minor
 Fix For: 2.4.0

 Attachments: FilteredTaskPage.png, MAPREDUCE-5553.patch, 
 MRJobWithTaskStateFilter.png, MRJobWithoutTaskStateFilters.png


 On Job page of MR application master, the task attempts have a nice breakdown 
 of different state: running, failed But for map/reduce task level, 
 there's only one links to all tasks without breakdown for different states:
 pending, running, complete, etc.
 With Hadoop 1, the similar functionality exists: user can go thru 
 jobtracker- job- map/reduce tasks - attempts.
 This ticket will improve the usability to allow user to go job- total, 
 pending, running, complete Map/Reduce task.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (MAPREDUCE-4868) Allow multiple iteration for map

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4868?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-4868:
-

Fix Version/s: (was: 2.3.0)
   2.4.0

 Allow multiple iteration for map
 

 Key: MAPREDUCE-4868
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-4868
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: mrv2
Affects Versions: 3.0.0, 2.0.3-alpha
Reporter: Jerry Chen
 Fix For: 3.0.0, 2.4.0

   Original Estimate: 168h
  Remaining Estimate: 168h

 Currently, the Mapper class allows advanced users to override public void 
 run(Context context) method for more control over the execution of the 
 mapper, while Context interface limit the operations over the data which is 
 the foundation of more control.
 One of use cases is that when I am considering a hive optimziation problem, I 
 want to go two passes over the input data instead of using a another job or 
 task ( which may slower the whole process). Each pass do the same thing but 
 with a different parameters.
 This is a new paradigm of Map Reduce usage and can be archived easily by 
 extend Context interface a little with the more control over the data such as 
 reset the input.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (MAPREDUCE-5728) Check NPE for serializer/deserializer in MapTask

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5728?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5728:
-

Fix Version/s: (was: 2.3.0)
   2.4.0

 Check NPE for serializer/deserializer in MapTask
 

 Key: MAPREDUCE-5728
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5728
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: client
Affects Versions: 2.2.0
Reporter: Jerry He
Assignee: Jerry He
Priority: Minor
 Fix For: 2.2.1, 2.4.0

 Attachments: MAPREDUCE-5728-trunk.patch


 Currently we will get NPE if the serializer/deserializer is not configured 
 correctly.
 {code}
 14/01/14 11:52:35 INFO mapred.JobClient: Task Id : 
 attempt_201401072154_0027_m_02_2, Status : FAILED
 java.lang.NullPointerException
 at 
 org.apache.hadoop.mapred.MapTask$MapOutputBuffer.init(MapTask.java:944)
 at 
 org.apache.hadoop.mapred.MapTask$NewOutputCollector.init(MapTask.java:672)
 at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:740)
 at org.apache.hadoop.mapred.MapTask.run(MapTask.java:368)
 at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
 at 
 java.security.AccessController.doPrivileged(AccessController.java:362)
 at javax.security.auth.Subject.doAs(Subject.java:573)
 at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1502)
 at org.apache.hadoop.mapred.Child.main(Child.java:249)
 {code}
 serializationFactory.getSerializer and serializationFactory.getDeserializer 
 returns NULL in this case.
 Let's check NPE for serializer/deserializer in MapTask so that we don't get 
 meaningless NPE.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (MAPREDUCE-4462) Enhance readability of TestFairScheduler.java

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4462?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-4462:
-

Fix Version/s: (was: 2.3.0)
   2.4.0

 Enhance readability of TestFairScheduler.java
 -

 Key: MAPREDUCE-4462
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-4462
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: scheduler, test
Reporter: Ryan Hennig
Priority: Minor
  Labels: comments, test
 Fix For: 2.4.0

 Attachments: MAPREDUCE-4462.patch

   Original Estimate: 2h
  Remaining Estimate: 2h

 While reading over the unit tests for the Fair Scheduler introduced by 
 MAPREDUCE-3451, I added comments to make the logic of the test easier to grok 
 quickly.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (MAPREDUCE-5559) Reconsidering the policy of ignoring the blacklist after reaching the threshold

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5559?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5559:
-

Fix Version/s: (was: 2.3.0)
   2.4.0

 Reconsidering the policy of ignoring the blacklist after reaching the 
 threshold
 ---

 Key: MAPREDUCE-5559
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5559
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
Affects Versions: 2.1.1-beta
Reporter: Zhijie Shen
Assignee: Zhijie Shen
 Fix For: 2.4.0


 Nowadays, when MR AM find the number of blacklisted nodes reaches one 
 threshold, the blacklist will be totally ignored. The newly assigned 
 containers on the blacklisted nodes will be allocated. This may be not the 
 best practice. We need to reconsider of it.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (MAPREDUCE-5028) Maps fail when io.sort.mb is set to high value

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5028?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5028:
-

Fix Version/s: (was: 2.3.0)
   2.4.0

 Maps fail when io.sort.mb is set to high value
 --

 Key: MAPREDUCE-5028
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5028
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Affects Versions: 1.1.1, 2.0.3-alpha, 0.23.5
Reporter: Karthik Kambatla
Assignee: Karthik Kambatla
Priority: Critical
 Fix For: 1.2.0, 2.4.0

 Attachments: MR-5028_testapp.patch, mr-5028-branch1.patch, 
 mr-5028-branch1.patch, mr-5028-branch1.patch, mr-5028-trunk.patch, 
 mr-5028-trunk.patch, mr-5028-trunk.patch, repro-mr-5028.patch


 Verified the problem exists on branch-1 with the following configuration:
 Pseudo-dist mode: 2 maps/ 1 reduce, mapred.child.java.opts=-Xmx2048m, 
 io.sort.mb=1280, dfs.block.size=2147483648
 Run teragen to generate 4 GB data
 Maps fail when you run wordcount on this configuration with the following 
 error: 
 {noformat}
 java.io.IOException: Spill failed
   at 
 org.apache.hadoop.mapred.MapTask$MapOutputBuffer.collect(MapTask.java:1031)
   at 
 org.apache.hadoop.mapred.MapTask$NewOutputCollector.write(MapTask.java:692)
   at 
 org.apache.hadoop.mapreduce.TaskInputOutputContext.write(TaskInputOutputContext.java:80)
   at 
 org.apache.hadoop.examples.WordCount$TokenizerMapper.map(WordCount.java:45)
   at 
 org.apache.hadoop.examples.WordCount$TokenizerMapper.map(WordCount.java:34)
   at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
   at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:766)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:396)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1149)
   at org.apache.hadoop.mapred.Child.main(Child.java:249)
 Caused by: java.io.EOFException
   at java.io.DataInputStream.readInt(DataInputStream.java:375)
   at org.apache.hadoop.io.IntWritable.readFields(IntWritable.java:38)
   at 
 org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:67)
   at 
 org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:40)
   at 
 org.apache.hadoop.mapreduce.ReduceContext.nextKeyValue(ReduceContext.java:116)
   at 
 org.apache.hadoop.mapreduce.ReduceContext.nextKey(ReduceContext.java:92)
   at org.apache.hadoop.mapreduce.Reducer.run(Reducer.java:175)
   at 
 org.apache.hadoop.mapred.Task$NewCombinerRunner.combine(Task.java:1505)
   at 
 org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpill(MapTask.java:1438)
   at 
 org.apache.hadoop.mapred.MapTask$MapOutputBuffer.access$1800(MapTask.java:855)
   at 
 org.apache.hadoop.mapred.MapTask$MapOutputBuffer$SpillThread.run(MapTask.java:1346)
 {noformat}



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (MAPREDUCE-5036) Default shuffle handler port should not be 8080

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5036?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5036:
-

Fix Version/s: (was: 2.3.0)
   2.4.0

 Default shuffle handler port should not be 8080
 ---

 Key: MAPREDUCE-5036
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5036
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
Affects Versions: 2.0.3-alpha
Reporter: Sandy Ryza
Assignee: Sandy Ryza
 Fix For: 2.4.0

 Attachments: MAPREDUCE-5036-13562.patch, MAPREDUCE-5036-2.patch, 
 MAPREDUCE-5036.patch


 The shuffle handler port (mapreduce.shuffle.port) defaults to 8080.  This is 
 a pretty common port for web services, and is likely to cause unnecessary 
 port conflicts.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (MAPREDUCE-4253) Tests for mapreduce-client-core are lying under mapreduce-client-jobclient

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4253?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-4253:
-

Fix Version/s: (was: 2.3.0)
   2.4.0

 Tests for mapreduce-client-core are lying under mapreduce-client-jobclient
 --

 Key: MAPREDUCE-4253
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-4253
 Project: Hadoop Map/Reduce
  Issue Type: Task
  Components: client
Affects Versions: 2.0.0-alpha
Reporter: Harsh J
Assignee: Tsuyoshi OZAWA
 Fix For: 2.4.0

 Attachments: MR-4253.1.patch, MR-4253.2.patch, 
 crossing_project_checker.rb, result.txt


 Many of the tests for client libs from mapreduce-client-core are lying under 
 mapreduce-client-jobclient.
 We should investigate if this is the right thing to do and if not, move the 
 tests back into client-core.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (MAPREDUCE-4468) Encapsulate FairScheduler preemption logic into helper class

2014-02-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4468?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-4468:
-

Fix Version/s: (was: 2.3.0)
   2.4.0

 Encapsulate FairScheduler preemption logic into helper class
 

 Key: MAPREDUCE-4468
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-4468
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: scheduler
Reporter: Ryan Hennig
Priority: Minor
  Labels: refactoring, scheduler
 Fix For: 2.4.0

 Attachments: MAPREDUCE-4468.patch

   Original Estimate: 4h
  Remaining Estimate: 4h

 I've extracted the preemption logic from the Fair Scheduler into a helper 
 class so that FairScheduler is closer to following the Single Responsibility 
 Principle.  This may eventually evolve into a generalized preemption module 
 which could be leveraged by other schedulers.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Updated] (MAPREDUCE-5663) Add an interface to Input/Ouput Formats to obtain delegation tokens

2014-01-13 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5663?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5663:
-

Status: Open  (was: Patch Available)

+1, mostly lgtm.

One nit: I'd like to change the api to 'obtainCredentials' rather than 
'addCredentials'. That threw me off a little. Last thing - a more descriptive 
javadoc would help too. Thanks.

 Add an interface to Input/Ouput Formats to obtain delegation tokens
 ---

 Key: MAPREDUCE-5663
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5663
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
Reporter: Siddharth Seth
Assignee: Michael Weng
 Attachments: MAPREDUCE-5663.4.txt, MAPREDUCE-5663.5.txt, 
 MAPREDUCE-5663.patch.txt, MAPREDUCE-5663.patch.txt2, MAPREDUCE-5663.patch.txt3


 Currently, delegation tokens are obtained as part of the getSplits / 
 checkOutputSpecs calls to the InputFormat / OutputFormat respectively.
 This works as long as the splits are generated on a node with kerberos 
 credentials. For split generation elsewhere (AM for example), an explicit 
 interface is required.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Issue Comment Deleted] (MAPREDUCE-5541) Improved algorithm for whether need speculative task

2013-10-21 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5541?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5541:
-

Comment: was deleted

(was: Committed to hadoop-longwing/condor-branch-1.)

 Improved algorithm for whether need speculative task
 

 Key: MAPREDUCE-5541
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5541
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: mrv1
Affects Versions: 1.2.1
Reporter: zhaoyunjiong
Assignee: zhaoyunjiong
 Fix For: 1.2.2

 Attachments: MAPREDUCE-5541-branch-1.2.patch, 
 MAPREDUCE-5541-branch-1.2.patch


 Most of time, tasks won't start running at same time.
 In this case hasSpeculativeTask in TaskInProgress not working very well.
 Some times, some tasks just start running, and scheduler already decide it 
 need speculative task to run.
 And this waste a lot of resource.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Commented] (MAPREDUCE-5541) Improved algorithm for whether need speculative task

2013-10-21 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5541?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13801360#comment-13801360
 ] 

Arun C Murthy commented on MAPREDUCE-5541:
--

[~zhaoyunjiong] The patch looks good. Can you pls share your experience running 
with this patch? What kind of tests have you done? Tx.

 Improved algorithm for whether need speculative task
 

 Key: MAPREDUCE-5541
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5541
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: mrv1
Affects Versions: 1.2.1
Reporter: zhaoyunjiong
Assignee: zhaoyunjiong
 Fix For: 1.2.2

 Attachments: MAPREDUCE-5541-branch-1.2.patch, 
 MAPREDUCE-5541-branch-1.2.patch


 Most of time, tasks won't start running at same time.
 In this case hasSpeculativeTask in TaskInProgress not working very well.
 Some times, some tasks just start running, and scheduler already decide it 
 need speculative task to run.
 And this waste a lot of resource.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Commented] (MAPREDUCE-5541) Improved algorithm for whether need speculative task

2013-10-21 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5541?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13801488#comment-13801488
 ] 

Arun C Murthy commented on MAPREDUCE-5541:
--

Thanks. Meanwhile, can you please do a full run of 'ant test' and report 
results? I'm almost sure this will break some unit tests. Thanks.

 Improved algorithm for whether need speculative task
 

 Key: MAPREDUCE-5541
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5541
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: mrv1
Affects Versions: 1.2.1
Reporter: zhaoyunjiong
Assignee: zhaoyunjiong
 Fix For: 1.2.2

 Attachments: MAPREDUCE-5541-branch-1.2.patch, 
 MAPREDUCE-5541-branch-1.2.patch


 Most of time, tasks won't start running at same time.
 In this case hasSpeculativeTask in TaskInProgress not working very well.
 Some times, some tasks just start running, and scheduler already decide it 
 need speculative task to run.
 And this waste a lot of resource.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Commented] (MAPREDUCE-5583) Ability to limit running map and reduce tasks

2013-10-15 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5583?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13795429#comment-13795429
 ] 

Arun C Murthy commented on MAPREDUCE-5583:
--

Looks like we are rehashing the discussion on HADOOP-5170.

 Ability to limit running map and reduce tasks
 -

 Key: MAPREDUCE-5583
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5583
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: mr-am, mrv2
Affects Versions: 0.23.9, 2.1.1-beta
Reporter: Jason Lowe

 It would be nice if users could specify a limit to the number of map or 
 reduce tasks that are running simultaneously.  Occasionally users are 
 performing operations in tasks that can lead to DDoS scenarios if too many 
 tasks run simultaneously (e.g.: accessing a database, web service, etc.).  
 Having the ability to throttle the number of tasks simultaneously running 
 would provide users a way to mitigate issues with too many tasks on a large 
 cluster attempting to access a serivce at any one time.
 This is similar to the functionality requested by MAPREDUCE-224 and 
 implemented by HADOOP-3412 but was dropped in mrv2.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Commented] (MAPREDUCE-5583) Ability to limit running map and reduce tasks

2013-10-15 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5583?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13795427#comment-13795427
 ] 

Arun C Murthy commented on MAPREDUCE-5583:
--

I'm very concerned about allowing this sort of control to the end user without 
checks/balances. If every one does the same I want max of n tasks, all 
resources in the cluster can deadlock. This is one of the main reasons we 
haven't supported this. I'm still against it.

 Ability to limit running map and reduce tasks
 -

 Key: MAPREDUCE-5583
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5583
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: mr-am, mrv2
Affects Versions: 0.23.9, 2.1.1-beta
Reporter: Jason Lowe

 It would be nice if users could specify a limit to the number of map or 
 reduce tasks that are running simultaneously.  Occasionally users are 
 performing operations in tasks that can lead to DDoS scenarios if too many 
 tasks run simultaneously (e.g.: accessing a database, web service, etc.).  
 Having the ability to throttle the number of tasks simultaneously running 
 would provide users a way to mitigate issues with too many tasks on a large 
 cluster attempting to access a serivce at any one time.
 This is similar to the functionality requested by MAPREDUCE-224 and 
 implemented by HADOOP-3412 but was dropped in mrv2.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Commented] (MAPREDUCE-5583) Ability to limit running map and reduce tasks

2013-10-15 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5583?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13795443#comment-13795443
 ] 

Arun C Murthy commented on MAPREDUCE-5583:
--

Cluster with 100,000 containers, 1,000 jobs, each with 10 tasks, and 
specifies that they can only run 5 tasks. So, you are now only using 5% of the 
cluster and no one makes progress leading to very poor utilization and 
peanut-buttering effect.

Admittedly it's a contrived example and yes, I agree a user can hack his own AM 
to do this - but let's not make this trivial for normal users. This leads to 
all sorts of bad side-effects by supporting it out of the box.

Some form of admin control (e.g. queue with a max-cap) for a small number of 
use-cases where you *actually* need this feature is much safer.

 Ability to limit running map and reduce tasks
 -

 Key: MAPREDUCE-5583
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5583
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: mr-am, mrv2
Affects Versions: 0.23.9, 2.1.1-beta
Reporter: Jason Lowe

 It would be nice if users could specify a limit to the number of map or 
 reduce tasks that are running simultaneously.  Occasionally users are 
 performing operations in tasks that can lead to DDoS scenarios if too many 
 tasks run simultaneously (e.g.: accessing a database, web service, etc.).  
 Having the ability to throttle the number of tasks simultaneously running 
 would provide users a way to mitigate issues with too many tasks on a large 
 cluster attempting to access a serivce at any one time.
 This is similar to the functionality requested by MAPREDUCE-224 and 
 implemented by HADOOP-3412 but was dropped in mrv2.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Commented] (MAPREDUCE-5583) Ability to limit running map and reduce tasks

2013-10-14 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5583?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13794782#comment-13794782
 ] 

Arun C Murthy commented on MAPREDUCE-5583:
--

[~jlowe] We can already accomplish this with queue/user limits? 

 Ability to limit running map and reduce tasks
 -

 Key: MAPREDUCE-5583
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5583
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: mr-am, mrv2
Affects Versions: 0.23.9, 2.1.1-beta
Reporter: Jason Lowe

 It would be nice if users could specify a limit to the number of map or 
 reduce tasks that are running simultaneously.  Occasionally users are 
 performing operations in tasks that can lead to DDoS scenarios if too many 
 tasks run simultaneously (e.g.: accessing a database, web service, etc.).  
 Having the ability to throttle the number of tasks simultaneously running 
 would provide users a way to mitigate issues with too many tasks on a large 
 cluster attempting to access a serivce at any one time.
 This is similar to the functionality requested by MAPREDUCE-224 and 
 implemented by HADOOP-3412 but was dropped in mrv2.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5565) job clean up fails on secure cluster as the file system is not created in the context of the ugi running the job

2013-10-09 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5565?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5565:
-

Resolution: Duplicate
Status: Resolved  (was: Patch Available)

Duplicate of MAPREDUCE-5508.

 job clean up fails on secure cluster as the file system is not created in the 
 context of the ugi running the job
 

 Key: MAPREDUCE-5565
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5565
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: jobtracker
Affects Versions: 1.2.1
Reporter: Arpit Gupta
Assignee: Arun C Murthy
Priority: Critical
 Fix For: 1.3.0

 Attachments: MAPREDUCE-5565.patch, MAPREDUCE-5565.patch


 On secure clusters we see the following exceptions in the jt log
 {code}
 2013-10-04 04:52:31,753 ERROR 
 org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException 
 as:tt/host@REALM cause:javax.security.sasl.SaslException: GSS
 initiate failed [Caused by GSSException: No valid credentials provided 
 (Mechanism level: Failed to find any Kerberos tgt)]
 {code}
 And after the job finishes the staging dir is not cleaned up. While debugging 
 with [~acmurthy] we determined that file system object needs to be created in 
 the the context of the user who ran the job.
 Job however successfully completes



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5432) JobHistoryParser does not fetch clockSplits, cpuUsages, vMemKbytes, physMemKbytes from history file

2013-10-05 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5432?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5432:
-

Status: Open  (was: Patch Available)

[~ozawa] Patch looks good. Can you pls modify existing tests to check for the 
new values? Tx.

 JobHistoryParser does not fetch clockSplits, cpuUsages, vMemKbytes, 
 physMemKbytes from history file
 ---

 Key: MAPREDUCE-5432
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5432
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: jobhistoryserver, mrv2
Affects Versions: 2.0.4-alpha
Reporter: Vrushali C
Assignee: Tsuyoshi OZAWA
 Attachments: MAPREDUCE-5432.1.patch, MAPREDUCE-5432.2.patch


 JobHistoryParser's handleMapAttemptFinishedEvent() function does not look at 
 MapAttemptFinishedEvent's   
   int[] clockSplits;
   int[] cpuUsages;
   int[] vMemKbytes;
   int[] physMemKbytes;
 JobHistoryParser's inner class TaskAttemptInfo also needs to be enhanced to 
 have these as members so that handleMapAttemptFinishedEvent() can get them 
 and store them.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5565) job clean up fails on secure cluster as the file system is not created in the context of the ugi running the job

2013-10-04 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5565?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5565:
-

Fix Version/s: 1.3.0

 job clean up fails on secure cluster as the file system is not created in the 
 context of the ugi running the job
 

 Key: MAPREDUCE-5565
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5565
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: jobtracker
Affects Versions: 1.2.1
Reporter: Arpit Gupta
Assignee: Arun C Murthy
Priority: Critical
 Fix For: 1.3.0

 Attachments: MAPREDUCE-5565.patch


 On secure clusters we see the following exceptions in the jt log
 {code}
 2013-10-04 04:52:31,753 ERROR 
 org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException 
 as:tt/host@REALM cause:javax.security.sasl.SaslException: GSS
 initiate failed [Caused by GSSException: No valid credentials provided 
 (Mechanism level: Failed to find any Kerberos tgt)]
 {code}
 And after the job finishes the staging dir is not cleaned up. While debugging 
 with [~acmurthy] we determined that file system object needs to be created in 
 the the context of the user who ran the job.
 Job however successfully completes



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5565) job clean up fails on secure cluster as the file system is not created in the context of the ugi running the job

2013-10-04 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5565?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5565:
-

Status: Patch Available  (was: Open)

 job clean up fails on secure cluster as the file system is not created in the 
 context of the ugi running the job
 

 Key: MAPREDUCE-5565
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5565
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: jobtracker
Affects Versions: 1.2.1
Reporter: Arpit Gupta
Assignee: Arun C Murthy
Priority: Critical
 Attachments: MAPREDUCE-5565.patch


 On secure clusters we see the following exceptions in the jt log
 {code}
 2013-10-04 04:52:31,753 ERROR 
 org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException 
 as:tt/host@REALM cause:javax.security.sasl.SaslException: GSS
 initiate failed [Caused by GSSException: No valid credentials provided 
 (Mechanism level: Failed to find any Kerberos tgt)]
 {code}
 And after the job finishes the staging dir is not cleaned up. While debugging 
 with [~acmurthy] we determined that file system object needs to be created in 
 the the context of the user who ran the job.
 Job however successfully completes



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5565) job clean up fails on secure cluster as the file system is not created in the context of the ugi running the job

2013-10-04 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5565?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5565:
-

Attachment: MAPREDUCE-5565.patch

 Thanks for helping to track this down [~arpitgupta]!

Here is a straight-fwd patch which should fix the regression due to 
MAPREDUCE-5351.

 job clean up fails on secure cluster as the file system is not created in the 
 context of the ugi running the job
 

 Key: MAPREDUCE-5565
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5565
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: jobtracker
Affects Versions: 1.2.1
Reporter: Arpit Gupta
Assignee: Arun C Murthy
Priority: Critical
 Attachments: MAPREDUCE-5565.patch


 On secure clusters we see the following exceptions in the jt log
 {code}
 2013-10-04 04:52:31,753 ERROR 
 org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException 
 as:tt/host@REALM cause:javax.security.sasl.SaslException: GSS
 initiate failed [Caused by GSSException: No valid credentials provided 
 (Mechanism level: Failed to find any Kerberos tgt)]
 {code}
 And after the job finishes the staging dir is not cleaned up. While debugging 
 with [~acmurthy] we determined that file system object needs to be created in 
 the the context of the user who ran the job.
 Job however successfully completes



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5565) job clean up fails on secure cluster as the file system is not created in the context of the ugi running the job

2013-10-04 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5565?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5565:
-

Attachment: MAPREDUCE-5565.patch

Slightly more conservative patch to reset tempDirFs in catch blocks.

 job clean up fails on secure cluster as the file system is not created in the 
 context of the ugi running the job
 

 Key: MAPREDUCE-5565
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5565
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: jobtracker
Affects Versions: 1.2.1
Reporter: Arpit Gupta
Assignee: Arun C Murthy
Priority: Critical
 Fix For: 1.3.0

 Attachments: MAPREDUCE-5565.patch, MAPREDUCE-5565.patch


 On secure clusters we see the following exceptions in the jt log
 {code}
 2013-10-04 04:52:31,753 ERROR 
 org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException 
 as:tt/host@REALM cause:javax.security.sasl.SaslException: GSS
 initiate failed [Caused by GSSException: No valid credentials provided 
 (Mechanism level: Failed to find any Kerberos tgt)]
 {code}
 And after the job finishes the staging dir is not cleaned up. While debugging 
 with [~acmurthy] we determined that file system object needs to be created in 
 the the context of the user who ran the job.
 Job however successfully completes



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5530) Binary and source incompatibility in mapred.lib.CombineFileInputFormat between branch-1 and branch-2

2013-10-02 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5530:
-

   Resolution: Fixed
Fix Version/s: 2.1.2-beta
   Status: Resolved  (was: Patch Available)

I just committed this. Thanks [~rkanter]!

Also, thanks to [~sandyr] and [~zjshen] for reviews and feedback.

 Binary and source incompatibility in mapred.lib.CombineFileInputFormat 
 between branch-1 and branch-2
 

 Key: MAPREDUCE-5530
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5530
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv1, mrv2
Affects Versions: 2.1.1-beta
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker
 Fix For: 2.1.2-beta

 Attachments: MAPREDUCE-5530.patch, MAPREDUCE-5530.patch, 
 MAPREDUCE-5530.patch, MAPREDUCE-5530.patch


 {{mapred.lib.CombineFileInputFormat}} in branch-1 has this method:
 {code:java}
 protected boolean isSplitable(FileSystem fs, Path file)
 {code}
 In branch-2, {{mapred.lib.CombineFileInputFormat}} is now a subclass of 
 {{mapreduce.lib.input.CombineFileInputFormat}}, from which it inherits the 
 similar method:
 {code:java}
 protected boolean isSplitable(JobContext context, Path file)
 {code}
 This means that any code that subclasses 
 {{mapred.lib.CombineFileInputFormat}} and does not provide its own 
 implementation of {{protected boolean isSplitable(FileSystem fs, Path file)}} 
 will not be binary or source compatible if it tries to call {{isSplitable}} 
 with a {{FileSystem}} argument anywhere (that is, if compiled against 
 branch-1, it will throw a {{NoSuchMethodError}} if run against branch-2; 
 also, it won't even compile against branch-2).  



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5459) Update the doc of running MRv1 examples jar on YARN

2013-10-02 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5459?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5459:
-

   Resolution: Fixed
Fix Version/s: 2.1.2-beta
   Status: Resolved  (was: Patch Available)

I just committed this. Thanks [~zjshen]!

 Update the doc of running MRv1 examples jar on YARN
 ---

 Key: MAPREDUCE-5459
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5459
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Reporter: Zhijie Shen
Assignee: Zhijie Shen
 Fix For: 2.1.2-beta

 Attachments: MAPREDUCE-5459.1.patch


 In addition to adding two env vars: HADOOP_USER_CLASSPATH_FIRST and 
 HADOOP_CLASSPATH, we still need to add
 {code}
 property
 namemapreduce.job.user.classpath.first/name
 valuetrue/value
 /property
 {code}
 in mapred-site.xml to make sure that the MRv1 examples jar runs correctly on 
 YARN. Some examples will use Java reflection to find the classes in the 
 examples jar dynamically when they are running. With this configuration, the 
 MRv1 examples jar will appear before the MRv2 examples jar in CLASSPATH of 
 the processes in YARN containers. Therefore, the classes found via reflection 
 will be picked from MRv1 examples jar instead of MRv2 examples jar as well.
 MapReduce_Compatibility_Hadoop1_Hadoop2.apt.vm needs to be updated to 
 document this.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Commented] (MAPREDUCE-5530) Binary and source incompatibility in mapred.lib.CombineFileInputFormat between branch-1 and branch-2

2013-10-01 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13783099#comment-13783099
 ] 

Arun C Murthy commented on MAPREDUCE-5530:
--

[~rkanter]: did my comment make sense, is there a chance you can quickly 
incorporate them so we can resolve this last MR issue? Thanks!

 Binary and source incompatibility in mapred.lib.CombineFileInputFormat 
 between branch-1 and branch-2
 

 Key: MAPREDUCE-5530
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5530
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv1, mrv2
Affects Versions: 2.1.1-beta
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker
 Attachments: MAPREDUCE-5530.patch, MAPREDUCE-5530.patch


 {{mapred.lib.CombineFileInputFormat}} in branch-1 has this method:
 {code:java}
 protected boolean isSplitable(FileSystem fs, Path file)
 {code}
 In branch-2, {{mapred.lib.CombineFileInputFormat}} is now a subclass of 
 {{mapreduce.lib.input.CombineFileInputFormat}}, from which it inherits the 
 similar method:
 {code:java}
 protected boolean isSplitable(JobContext context, Path file)
 {code}
 This means that any code that subclasses 
 {{mapred.lib.CombineFileInputFormat}} and does not provide its own 
 implementation of {{protected boolean isSplitable(FileSystem fs, Path file)}} 
 will not be binary or source compatible if it tries to call {{isSplitable}} 
 with a {{FileSystem}} argument anywhere (that is, if compiled against 
 branch-1, it will throw a {{NoSuchMethodError}} if run against branch-2; 
 also, it won't even compile against branch-2).  



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5536) mapreduce.jobhistory.webapp.https.address property is not respected

2013-10-01 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5536?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5536:
-

Target Version/s: 2.1.2-beta

 mapreduce.jobhistory.webapp.https.address property is not respected
 ---

 Key: MAPREDUCE-5536
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5536
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Affects Versions: 2.1.1-beta
Reporter: Yesha Vora
Assignee: Omkar Vinit Joshi
Priority: Blocker
 Attachments: MAPREDUCE-5536.20131027.1.patch, 
 MAPREDUCE-5536.20131030.1.patch, MAPREDUCE-5536.20131030.1.patch, 
 MAPREDUCE-5536.20131030.3.patch, YARN-1240.20131025.1.patch


 The jobhistory server starts on port defined by 
 mapreduce.jobhistory.webapp.address property instead 
 mapreduce.jobhistory.webapp.https.address when hadoop.ssl.enabled=true.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Commented] (MAPREDUCE-5530) Binary and source incompatibility in mapred.lib.CombineFileInputFormat between branch-1 and branch-2

2013-10-01 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13783447#comment-13783447
 ] 

Arun C Murthy commented on MAPREDUCE-5530:
--

[~rkanter] People *really* should NOT be overriding 
mapred.lib.CombineFileInputFormat.isSplitable(JobContext, Path) because 
JobContext does not exist in the mapred api... hence we really should mark it 
as @Private.

 Binary and source incompatibility in mapred.lib.CombineFileInputFormat 
 between branch-1 and branch-2
 

 Key: MAPREDUCE-5530
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5530
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv1, mrv2
Affects Versions: 2.1.1-beta
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker
 Attachments: MAPREDUCE-5530.patch, MAPREDUCE-5530.patch, 
 MAPREDUCE-5530.patch


 {{mapred.lib.CombineFileInputFormat}} in branch-1 has this method:
 {code:java}
 protected boolean isSplitable(FileSystem fs, Path file)
 {code}
 In branch-2, {{mapred.lib.CombineFileInputFormat}} is now a subclass of 
 {{mapreduce.lib.input.CombineFileInputFormat}}, from which it inherits the 
 similar method:
 {code:java}
 protected boolean isSplitable(JobContext context, Path file)
 {code}
 This means that any code that subclasses 
 {{mapred.lib.CombineFileInputFormat}} and does not provide its own 
 implementation of {{protected boolean isSplitable(FileSystem fs, Path file)}} 
 will not be binary or source compatible if it tries to call {{isSplitable}} 
 with a {{FileSystem}} argument anywhere (that is, if compiled against 
 branch-1, it will throw a {{NoSuchMethodError}} if run against branch-2; 
 also, it won't even compile against branch-2).  



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5530) Binary and source incompatibility in mapred.lib.CombineFileInputFormat between branch-1 and branch-2

2013-10-01 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5530:
-

Status: Patch Available  (was: Open)

Thanks [~rkanter]. Submitting patch for hudson on your behalf so that we can 
commit this asap.

 Binary and source incompatibility in mapred.lib.CombineFileInputFormat 
 between branch-1 and branch-2
 

 Key: MAPREDUCE-5530
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5530
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv1, mrv2
Affects Versions: 2.1.1-beta
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker
 Attachments: MAPREDUCE-5530.patch, MAPREDUCE-5530.patch, 
 MAPREDUCE-5530.patch, MAPREDUCE-5530.patch


 {{mapred.lib.CombineFileInputFormat}} in branch-1 has this method:
 {code:java}
 protected boolean isSplitable(FileSystem fs, Path file)
 {code}
 In branch-2, {{mapred.lib.CombineFileInputFormat}} is now a subclass of 
 {{mapreduce.lib.input.CombineFileInputFormat}}, from which it inherits the 
 similar method:
 {code:java}
 protected boolean isSplitable(JobContext context, Path file)
 {code}
 This means that any code that subclasses 
 {{mapred.lib.CombineFileInputFormat}} and does not provide its own 
 implementation of {{protected boolean isSplitable(FileSystem fs, Path file)}} 
 will not be binary or source compatible if it tries to call {{isSplitable}} 
 with a {{FileSystem}} argument anywhere (that is, if compiled against 
 branch-1, it will throw a {{NoSuchMethodError}} if run against branch-2; 
 also, it won't even compile against branch-2).  



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5551) Binary Incompatibility of O.A.H.U.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes

2013-09-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5551?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5551:
-

   Resolution: Fixed
Fix Version/s: 2.1.2-beta
   Status: Resolved  (was: Patch Available)

I just committed this. Thanks [~zjshen]!

 Binary Incompatibility of 
 O.A.H.U.mapred.SequenceFileAsBinaryOutputFormat.WritableValueBytes
 

 Key: MAPREDUCE-5551
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5551
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
Reporter: Zhijie Shen
Assignee: Zhijie Shen
Priority: Blocker
 Fix For: 2.1.2-beta

 Attachments: MAPREDUCE-5551.1.patch


 The non-default constructor is moved to the super class, but it cannot be 
 inherited.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5530) Binary and source incompatibility in mapred.lib.CombineFileInputFormat between branch-1 and branch-2

2013-09-30 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5530?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5530:
-

Status: Open  (was: Patch Available)

 Binary and source incompatibility in mapred.lib.CombineFileInputFormat 
 between branch-1 and branch-2
 

 Key: MAPREDUCE-5530
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5530
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv1, mrv2
Affects Versions: 2.1.1-beta
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker
 Attachments: MAPREDUCE-5530.patch, MAPREDUCE-5530.patch


 {{mapred.lib.CombineFileInputFormat}} in branch-1 has this method:
 {code:java}
 protected boolean isSplitable(FileSystem fs, Path file)
 {code}
 In branch-2, {{mapred.lib.CombineFileInputFormat}} is now a subclass of 
 {{mapreduce.lib.input.CombineFileInputFormat}}, from which it inherits the 
 similar method:
 {code:java}
 protected boolean isSplitable(JobContext context, Path file)
 {code}
 This means that any code that subclasses 
 {{mapred.lib.CombineFileInputFormat}} and does not provide its own 
 implementation of {{protected boolean isSplitable(FileSystem fs, Path file)}} 
 will not be binary or source compatible if it tries to call {{isSplitable}} 
 with a {{FileSystem}} argument anywhere (that is, if compiled against 
 branch-1, it will throw a {{NoSuchMethodError}} if run against branch-2; 
 also, it won't even compile against branch-2).  



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Commented] (MAPREDUCE-5530) Binary and source incompatibility in mapred.lib.CombineFileInputFormat between branch-1 and branch-2

2013-09-30 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5530?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13782567#comment-13782567
 ] 

Arun C Murthy commented on MAPREDUCE-5530:
--

[~rkanter] I agree the patch isn't perfect, but that's the only option we have 
now... thought I wonder if we should just document this and not fix it... 
thoughts?

IAC, couple of comments:
# Please mark 
org.apache.hadoop.mapred.lib.CombineFileInputFormat.isSplitable(JobContext, 
Path) as @Private 
# Also add some documentation to the method with a pointer to this jira 
explaining why end-users should never override 
org.apache.hadoop.mapred.lib.CombineFileInputFormat.isSplitable(JobContext, 
Path).

Thanks.

 Binary and source incompatibility in mapred.lib.CombineFileInputFormat 
 between branch-1 and branch-2
 

 Key: MAPREDUCE-5530
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5530
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv1, mrv2
Affects Versions: 2.1.1-beta
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker
 Attachments: MAPREDUCE-5530.patch, MAPREDUCE-5530.patch


 {{mapred.lib.CombineFileInputFormat}} in branch-1 has this method:
 {code:java}
 protected boolean isSplitable(FileSystem fs, Path file)
 {code}
 In branch-2, {{mapred.lib.CombineFileInputFormat}} is now a subclass of 
 {{mapreduce.lib.input.CombineFileInputFormat}}, from which it inherits the 
 similar method:
 {code:java}
 protected boolean isSplitable(JobContext context, Path file)
 {code}
 This means that any code that subclasses 
 {{mapred.lib.CombineFileInputFormat}} and does not provide its own 
 implementation of {{protected boolean isSplitable(FileSystem fs, Path file)}} 
 will not be binary or source compatible if it tries to call {{isSplitable}} 
 with a {{FileSystem}} argument anywhere (that is, if compiled against 
 branch-1, it will throw a {{NoSuchMethodError}} if run against branch-2; 
 also, it won't even compile against branch-2).  



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Updated] (MAPREDUCE-5531) Binary and source incompatibility in mapreduce.TaskID and mapreduce.TaskAttemptID between branch-1 and branch-2

2013-09-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5531?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5531:
-

   Resolution: Fixed
Fix Version/s: 2.1.2-beta
   Status: Resolved  (was: Patch Available)

+1 from me too.

I just committed this. Thanks [~rkanter]! 

Also, [~zjshen] thanks for the review.

 Binary and source incompatibility in mapreduce.TaskID and 
 mapreduce.TaskAttemptID between branch-1 and branch-2
 ---

 Key: MAPREDUCE-5531
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5531
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv1, mrv2
Affects Versions: 2.1.1-beta
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker
 Fix For: 2.1.2-beta

 Attachments: MAPREDUCE-5531.patch, MAPREDUCE-5531.patch


 {{mapreduce.TaskID}} in branch-1 has these two constructors:
 {code:java}
 public TaskID(JobID jobId, boolean isMap, int id)
 public TaskID(String jtIdentifier, int jobId, boolean isMap, int id)
 {code}
 In branch-2, {{mapreduce.TaskID}} no longer has either of the above two 
 constructors.  
 Also, {{mapreduce.TaskAttemptID}} in branch-1 has this constructor:
 {code:java}
 public TaskAttemptID(String jtIdentifier, int jobId, boolean isMap, int 
 taskId, int id)
 {code}
 In branch-2, {{mapreduce.TaskAttemptID}} no longer his this constructor.
 It looks like these constructors were probably removed because the {{boolean 
 isMap}} was replaced by an enum, {{TaskType}}.
 This means that any code that tries to use any of those constructors will not 
 be binary or source compatible (in fact, the missing {{TaskAttemptID}} 
 constructor calls one of the missing {{TaskID}} constructors).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5529) Binary incompatibilities in mapred.lib.TotalOrderPartitioner between branch-1 and branch-2

2013-09-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5529?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5529:
-

Attachment: MAPREDUCE-5529_branch_2.patch

Sigh, I needed to modify the branch-2 patch since MAPREDUCE-4594 is only on 
trunk

[~qwertymaniac]: in future, please don't commit simple enhancements such as 
MAPREDU-4594 only to trunk. Thanks.

 Binary incompatibilities in mapred.lib.TotalOrderPartitioner between branch-1 
 and branch-2
 --

 Key: MAPREDUCE-5529
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5529
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv1, mrv2
Affects Versions: 2.1.1-beta
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker
 Attachments: MAPREDUCE-5529_branch_2.patch, MAPREDUCE-5529.patch, 
 MAPREDUCE-5529.patch


 {{mapred.lib.TotalOrderPartitioner}} in branch-1 has these two methods:
 {code:java}
 public static String getPartitionFile(JobConf job)
 public static void setPartitionFile(JobConf job, Path p)
 {code}
 In branch-2, {{mapred.lib.TotalOrderPartitioner}} is now a subclass of 
 {{mapred.lib.TotalOrderPartitioner}}, from which it inherits the similar 
 methods:
 {code:java}
 public static String getPartitionFile(Configuration conf)
 public static void setPartitionFile(Configuration conf, Path p)
 {code}
 This means that any code that does either of the following:
 {code:java}
 TotalOrderPartitioner.setPartitionFile(new JobConf(), new Path(/));
 String str = TotalOrderPartitioner.getPartitionFile(new JobConf());
 {code}
 will not be binary compatible (that is, if compiled against branch-1, it will 
 throw a {{NoSuchMethodError}} if run against branch-2).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5529) Binary incompatibilities in mapred.lib.TotalOrderPartitioner between branch-1 and branch-2

2013-09-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5529?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5529:
-

   Resolution: Fixed
Fix Version/s: 2.1.2-beta
   Status: Resolved  (was: Patch Available)

+1. I just committed this, thanks [~rkanter]!

Thanks for the reviews [~zjshen]!

 Binary incompatibilities in mapred.lib.TotalOrderPartitioner between branch-1 
 and branch-2
 --

 Key: MAPREDUCE-5529
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5529
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv1, mrv2
Affects Versions: 2.1.1-beta
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker
 Fix For: 2.1.2-beta

 Attachments: MAPREDUCE-5529_branch_2.patch, MAPREDUCE-5529.patch, 
 MAPREDUCE-5529.patch


 {{mapred.lib.TotalOrderPartitioner}} in branch-1 has these two methods:
 {code:java}
 public static String getPartitionFile(JobConf job)
 public static void setPartitionFile(JobConf job, Path p)
 {code}
 In branch-2, {{mapred.lib.TotalOrderPartitioner}} is now a subclass of 
 {{mapred.lib.TotalOrderPartitioner}}, from which it inherits the similar 
 methods:
 {code:java}
 public static String getPartitionFile(Configuration conf)
 public static void setPartitionFile(Configuration conf, Path p)
 {code}
 This means that any code that does either of the following:
 {code:java}
 TotalOrderPartitioner.setPartitionFile(new JobConf(), new Path(/));
 String str = TotalOrderPartitioner.getPartitionFile(new JobConf());
 {code}
 will not be binary compatible (that is, if compiled against branch-1, it will 
 throw a {{NoSuchMethodError}} if run against branch-2).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5514) TestRMContainerAllocator fails on trunk

2013-09-26 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5514?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5514:
-

   Resolution: Fixed
Fix Version/s: 2.3.0
   Status: Resolved  (was: Patch Available)

 TestRMContainerAllocator fails on trunk
 ---

 Key: MAPREDUCE-5514
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5514
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Reporter: Zhijie Shen
Assignee: Zhijie Shen
Priority: Blocker
 Fix For: 2.3.0

 Attachments: MAPREDUCE-5514.1.patch, MAPREDUCE-5514.2.patch, 
 MAPREDUCE-5514.3.patch, 
 org.apache.hadoop.mapreduce.v2.app.TestRMContainerAllocator-output.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5514) TestRMContainerAllocator fails on trunk

2013-09-26 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13779419#comment-13779419
 ] 

Arun C Murthy commented on MAPREDUCE-5514:
--

I just committed this. Thanks [~zjshen]!

 TestRMContainerAllocator fails on trunk
 ---

 Key: MAPREDUCE-5514
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5514
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Reporter: Zhijie Shen
Assignee: Zhijie Shen
Priority: Blocker
 Fix For: 2.3.0

 Attachments: MAPREDUCE-5514.1.patch, MAPREDUCE-5514.2.patch, 
 MAPREDUCE-5514.3.patch, 
 org.apache.hadoop.mapreduce.v2.app.TestRMContainerAllocator-output.txt




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5529) Binary incompatibilities in mapred.lib.TotalOrderPartitioner between branch-1 and branch-2

2013-09-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5529?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5529:
-

Target Version/s: 2.1.2-beta

 Binary incompatibilities in mapred.lib.TotalOrderPartitioner between branch-1 
 and branch-2
 --

 Key: MAPREDUCE-5529
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5529
 Project: Hadoop Map/Reduce
  Issue Type: Sub-task
  Components: mrv1, mrv2
Affects Versions: 2.1.1-beta
Reporter: Robert Kanter
Assignee: Robert Kanter
Priority: Blocker
 Attachments: MAPREDUCE-5529.patch


 {{mapred.lib.TotalOrderPartitioner}} in branch-1 has these two methods:
 {code:java}
 public static String getPartitionFile(JobConf job)
 public static void setPartitionFile(JobConf job, Path p)
 {code}
 In branch-2, {{mapred.lib.TotalOrderPartitioner}} is now a subclass of 
 {{mapred.lib.TotalOrderPartitioner}}, from which it inherits the similar 
 methods:
 {code:java}
 public static String getPartitionFile(Configuration conf)
 public static void setPartitionFile(Configuration conf, Path p)
 {code}
 This means that any code that does either of the following:
 {code:java}
 TotalOrderPartitioner.setPartitionFile(new JobConf(), new Path(/));
 String str = TotalOrderPartitioner.getPartitionFile(new JobConf());
 {code}
 will not be binary compatible (that is, if compiled against branch-1, it will 
 throw a {{NoSuchMethodError}} if run against branch-2).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5503) TestMRJobClient.testJobClient is failing

2013-09-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5503?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5503:
-

Target Version/s: 3.0.0, 2.1.2-beta  (was: 3.0.0, 2.1.1-beta)

 TestMRJobClient.testJobClient is failing
 

 Key: MAPREDUCE-5503
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5503
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv2
Affects Versions: 3.0.0, 2.1.0-beta
Reporter: Jason Lowe
Assignee: Jian He
Priority: Blocker
 Attachments: MAPREDUCE-5503.patch, MAPREDUCE-5503.patch


 TestMRJobClient.testJobClient is failing on trunk and causing precommit 
 builds to complain:
 {noformat}
 testJobClient(org.apache.hadoop.mapreduce.TestMRJobClient)  Time elapsed: 
 26.361 sec   FAILURE!
 junit.framework.AssertionFailedError: expected:1 but was:0
   at junit.framework.Assert.fail(Assert.java:50)
   at junit.framework.Assert.failNotEquals(Assert.java:287)
   at junit.framework.Assert.assertEquals(Assert.java:67)
   at junit.framework.Assert.assertEquals(Assert.java:199)
   at junit.framework.Assert.assertEquals(Assert.java:205)
   at 
 org.apache.hadoop.mapreduce.TestMRJobClient.testJobList(TestMRJobClient.java:474)
   at 
 org.apache.hadoop.mapreduce.TestMRJobClient.testJobClient(TestMRJobClient.java:112)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5448) MapFileOutputFormat#getReaders bug with invisible files/folders

2013-09-18 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5448?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5448:
-

Status: Patch Available  (was: Open)

 MapFileOutputFormat#getReaders bug with invisible files/folders
 ---

 Key: MAPREDUCE-5448
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5448
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Reporter: Maysam Yabandeh
Assignee: Maysam Yabandeh
Priority: Minor
 Attachments: MAPREDUCE-5448.patch

   Original Estimate: 1h
  Remaining Estimate: 1h

 MapReduce jobs also produce some invisible files such as _SUCCESS, even when 
 the output format is MapFileOutputFormat. MapFileOutputFormat#getReaders 
 however reads the entire content of the job output, assming that they are 
 MapFiles.
 {code}
 Path[] names = FileUtil.stat2Paths(fs.listStatus(dir));
 {code}
 It should use a filter to skip the files that start with . or _.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5493) In-memory map outputs can be leaked after shuffle completes

2013-09-16 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5493?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5493:
-

   Resolution: Fixed
Fix Version/s: 2.1.1-beta
   Status: Resolved  (was: Patch Available)

+1, looks good - thanks for tracking this down [~jlowe]!

I just committed this.

 In-memory map outputs can be leaked after shuffle completes
 ---

 Key: MAPREDUCE-5493
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5493
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv2
Affects Versions: 2.1.0-beta, 0.23.9
Reporter: Jason Lowe
Assignee: Jason Lowe
Priority: Blocker
 Fix For: 2.1.1-beta

 Attachments: MAPREDUCE-5493.patch


 MergeManagerImpl#close adds the contents of inMemoryMergedMapOutputs and 
 inMemoryMapOutputs to a list of map outputs that is subsequently processed, 
 but it does not clear those sets.  This prevents some of the map outputs from 
 being garbage collected and significantly reduces the memory available for 
 the subsequent reduce phase.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5480) TestJSHSecurity.testDelegationToken is breaking after YARN-1085

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5480:
-

Resolution: Duplicate
Status: Resolved  (was: Patch Available)

Fixed via YARN-1085

 TestJSHSecurity.testDelegationToken is breaking after YARN-1085
 ---

 Key: MAPREDUCE-5480
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5480
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Affects Versions: 2.1.1-beta
Reporter: Vinod Kumar Vavilapalli
Assignee: Omkar Vinit Joshi
Priority: Blocker
 Attachments: MAPREDUCE-5480.20130824.1.patch


 See https://builds.apache.org/job/PreCommit-YARN-Build/1755//testReport/.
 {code}
 org.apache.hadoop.yarn.webapp.WebAppException: Error starting http server
   at org.apache.hadoop.yarn.webapp.WebApps$Builder.start(WebApps.java:251)
   at 
 org.apache.hadoop.mapreduce.v2.hs.HistoryClientService.initializeWebApp(HistoryClientService.java:152)
  ---
 Caused by: javax.servlet.ServletException: javax.servlet.ServletException: 
 Principal not defined in configuration
   at 
 org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler.init(KerberosAuthenticationHandler.java:203)
   at 
 org.apache.hadoop.security.authentication.server.AuthenticationFilter.init(AuthenticationFilter.java:146)
  ---
 Caused by: javax.servlet.ServletException: Principal not defined in 
 configuration
   at 
 org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler.init(KerberosAuthenticationHandler.java:164)
   ... 53 more
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5439) mapred-default.xml has missing properties

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5439?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5439:
-

Fix Version/s: (was: 2.1.0-beta)
   2.3.0

 mapred-default.xml has missing properties
 -

 Key: MAPREDUCE-5439
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5439
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv2
Affects Versions: 2.1.0-beta
Reporter: Siddharth Wagle
 Fix For: 2.3.0


 Properties that need to be added:
 mapreduce.map.memory.mb
 mapreduce.map.java.opts
 mapreduce.reduce.memory.mb
 mapreduce.reduce.java.opts
 Properties that need to be fixed:
 mapred.child.java.opts should not be in mapred-default.
 yarn.app.mapreduce.am.command-opts description needs fixing

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-4868) Allow multiple iteration for map

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4868?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-4868:
-

Fix Version/s: (was: 2.1.0-beta)
   2.3.0

 Allow multiple iteration for map
 

 Key: MAPREDUCE-4868
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-4868
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: mrv2
Affects Versions: 3.0.0, 2.0.3-alpha
Reporter: Jerry Chen
 Fix For: 3.0.0, 2.3.0

   Original Estimate: 168h
  Remaining Estimate: 168h

 Currently, the Mapper class allows advanced users to override public void 
 run(Context context) method for more control over the execution of the 
 mapper, while Context interface limit the operations over the data which is 
 the foundation of more control.
 One of use cases is that when I am considering a hive optimziation problem, I 
 want to go two passes over the input data instead of using a another job or 
 task ( which may slower the whole process). Each pass do the same thing but 
 with a different parameters.
 This is a new paradigm of Map Reduce usage and can be archived easily by 
 extend Context interface a little with the more control over the data such as 
 reset the input.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5036) Default shuffle handler port should not be 8080

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5036?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5036:
-

Fix Version/s: (was: 2.1.0-beta)
   2.3.0

 Default shuffle handler port should not be 8080
 ---

 Key: MAPREDUCE-5036
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5036
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
Affects Versions: 2.0.3-alpha
Reporter: Sandy Ryza
Assignee: Sandy Ryza
 Fix For: 2.3.0

 Attachments: MAPREDUCE-5036-13562.patch, MAPREDUCE-5036-2.patch, 
 MAPREDUCE-5036.patch


 The shuffle handler port (mapreduce.shuffle.port) defaults to 8080.  This is 
 a pretty common port for web services, and is likely to cause unnecessary 
 port conflicts.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5028) Maps fail when io.sort.mb is set to high value

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5028?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5028:
-

Fix Version/s: (was: 2.1.0-beta)
   2.3.0

 Maps fail when io.sort.mb is set to high value
 --

 Key: MAPREDUCE-5028
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5028
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Affects Versions: 1.1.1, 2.0.3-alpha, 0.23.5
Reporter: Karthik Kambatla
Assignee: Karthik Kambatla
Priority: Critical
 Fix For: 1.2.0, 2.3.0

 Attachments: mr-5028-branch1.patch, mr-5028-branch1.patch, 
 mr-5028-branch1.patch, MR-5028_testapp.patch, mr-5028-trunk.patch, 
 mr-5028-trunk.patch, mr-5028-trunk.patch, repro-mr-5028.patch


 Verified the problem exists on branch-1 with the following configuration:
 Pseudo-dist mode: 2 maps/ 1 reduce, mapred.child.java.opts=-Xmx2048m, 
 io.sort.mb=1280, dfs.block.size=2147483648
 Run teragen to generate 4 GB data
 Maps fail when you run wordcount on this configuration with the following 
 error: 
 {noformat}
 java.io.IOException: Spill failed
   at 
 org.apache.hadoop.mapred.MapTask$MapOutputBuffer.collect(MapTask.java:1031)
   at 
 org.apache.hadoop.mapred.MapTask$NewOutputCollector.write(MapTask.java:692)
   at 
 org.apache.hadoop.mapreduce.TaskInputOutputContext.write(TaskInputOutputContext.java:80)
   at 
 org.apache.hadoop.examples.WordCount$TokenizerMapper.map(WordCount.java:45)
   at 
 org.apache.hadoop.examples.WordCount$TokenizerMapper.map(WordCount.java:34)
   at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)
   at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:766)
   at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)
   at org.apache.hadoop.mapred.Child$4.run(Child.java:255)
   at java.security.AccessController.doPrivileged(Native Method)
   at javax.security.auth.Subject.doAs(Subject.java:396)
   at 
 org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1149)
   at org.apache.hadoop.mapred.Child.main(Child.java:249)
 Caused by: java.io.EOFException
   at java.io.DataInputStream.readInt(DataInputStream.java:375)
   at org.apache.hadoop.io.IntWritable.readFields(IntWritable.java:38)
   at 
 org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:67)
   at 
 org.apache.hadoop.io.serializer.WritableSerialization$WritableDeserializer.deserialize(WritableSerialization.java:40)
   at 
 org.apache.hadoop.mapreduce.ReduceContext.nextKeyValue(ReduceContext.java:116)
   at 
 org.apache.hadoop.mapreduce.ReduceContext.nextKey(ReduceContext.java:92)
   at org.apache.hadoop.mapreduce.Reducer.run(Reducer.java:175)
   at 
 org.apache.hadoop.mapred.Task$NewCombinerRunner.combine(Task.java:1505)
   at 
 org.apache.hadoop.mapred.MapTask$MapOutputBuffer.sortAndSpill(MapTask.java:1438)
   at 
 org.apache.hadoop.mapred.MapTask$MapOutputBuffer.access$1800(MapTask.java:855)
   at 
 org.apache.hadoop.mapred.MapTask$MapOutputBuffer$SpillThread.run(MapTask.java:1346)
 {noformat}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-4468) Encapsulate FairScheduler preemption logic into helper class

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4468?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-4468:
-

Fix Version/s: (was: 2.1.0-beta)
   2.3.0

 Encapsulate FairScheduler preemption logic into helper class
 

 Key: MAPREDUCE-4468
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-4468
 Project: Hadoop Map/Reduce
  Issue Type: Improvement
  Components: scheduler
Reporter: Ryan Hennig
Priority: Minor
  Labels: refactoring, scheduler
 Fix For: 2.3.0

 Attachments: MAPREDUCE-4468.patch

   Original Estimate: 4h
  Remaining Estimate: 4h

 I've extracted the preemption logic from the Fair Scheduler into a helper 
 class so that FairScheduler is closer to following the Single Responsibility 
 Principle.  This may eventually evolve into a generalized preemption module 
 which could be leveraged by other schedulers.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5170) incorrect exception message if min node size min rack size

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5170?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5170:
-

Fix Version/s: (was: 2.1.0-beta)
   2.3.0

 incorrect exception message if min node size  min rack size
 

 Key: MAPREDUCE-5170
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5170
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv2
Affects Versions: 2.0.3-alpha
Reporter: Sangjin Lee
Priority: Trivial
 Fix For: 2.3.0

 Attachments: MAPREDUCE-5170.patch


 The exception message for CombineFileInputFormat if min node size  min rack 
 size is worded backwards.
 Currently it reads Minimum split size per node... cannot be smaller than the 
 minimum split size per rack...
 It should be Minimum split size per node... cannot be LARGER than the 
 minimum split size per rack...

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-4253) Tests for mapreduce-client-core are lying under mapreduce-client-jobclient

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4253?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-4253:
-

Fix Version/s: (was: 2.1.0-beta)
   2.3.0

 Tests for mapreduce-client-core are lying under mapreduce-client-jobclient
 --

 Key: MAPREDUCE-4253
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-4253
 Project: Hadoop Map/Reduce
  Issue Type: Task
  Components: client
Affects Versions: 2.0.0-alpha
Reporter: Harsh J
Assignee: Tsuyoshi OZAWA
 Fix For: 2.3.0

 Attachments: crossing_project_checker.rb, MR-4253.1.patch, 
 MR-4253.2.patch, result.txt


 Many of the tests for client libs from mapreduce-client-core are lying under 
 mapreduce-client-jobclient.
 We should investigate if this is the right thing to do and if not, move the 
 tests back into client-core.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-4462) Enhance readability of TestFairScheduler.java

2013-08-27 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-4462?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-4462:
-

Fix Version/s: (was: 2.1.0-beta)
   2.3.0

 Enhance readability of TestFairScheduler.java
 -

 Key: MAPREDUCE-4462
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-4462
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: scheduler, test
Reporter: Ryan Hennig
Priority: Minor
  Labels: comments, test
 Fix For: 2.3.0

 Attachments: MAPREDUCE-4462.patch

   Original Estimate: 2h
  Remaining Estimate: 2h

 While reading over the unit tests for the Fair Scheduler introduced by 
 MAPREDUCE-3451, I added comments to make the logic of the test easier to grok 
 quickly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5480) TestJSHSecurity.testDelegationToken is breaking after YARN-1085

2013-08-24 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5480?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5480:
-

Priority: Blocker  (was: Major)

 TestJSHSecurity.testDelegationToken is breaking after YARN-1085
 ---

 Key: MAPREDUCE-5480
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5480
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Affects Versions: 2.1.1-beta
Reporter: Vinod Kumar Vavilapalli
Assignee: Vinod Kumar Vavilapalli
Priority: Blocker

 See https://builds.apache.org/job/PreCommit-YARN-Build/1755//testReport/.
 {code}
 org.apache.hadoop.yarn.webapp.WebAppException: Error starting http server
   at org.apache.hadoop.yarn.webapp.WebApps$Builder.start(WebApps.java:251)
   at 
 org.apache.hadoop.mapreduce.v2.hs.HistoryClientService.initializeWebApp(HistoryClientService.java:152)
  ---
 Caused by: javax.servlet.ServletException: javax.servlet.ServletException: 
 Principal not defined in configuration
   at 
 org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler.init(KerberosAuthenticationHandler.java:203)
   at 
 org.apache.hadoop.security.authentication.server.AuthenticationFilter.init(AuthenticationFilter.java:146)
  ---
 Caused by: javax.servlet.ServletException: Principal not defined in 
 configuration
   at 
 org.apache.hadoop.security.authentication.server.KerberosAuthenticationHandler.init(KerberosAuthenticationHandler.java:164)
   ... 53 more
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5475) MRClientService does not verify ACLs properly

2013-08-21 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13746965#comment-13746965
 ] 

Arun C Murthy commented on MAPREDUCE-5475:
--

+1 pending jenkins, lgtm. Thanks [~jlowe]!

 MRClientService does not verify ACLs properly
 -

 Key: MAPREDUCE-5475
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5475
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mr-am, mrv2
Affects Versions: 2.0.4-alpha, 0.23.9
Reporter: Jason Lowe
Assignee: Jason Lowe
Priority: Blocker
 Attachments: MAPREDUCE-5475.patch


 When MRClientService receives requests, it calls verifyAndGetJob which does 
 not actually validate that the current user has the proper access.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5476) Job can fail when RM restarts after staging dir is cleaned but before MR successfully unregister with RM

2013-08-21 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5476:
-

Target Version/s: 2.1.1-beta

 Job can fail when RM restarts after staging dir is cleaned but before MR 
 successfully unregister with RM
 

 Key: MAPREDUCE-5476
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5476
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Reporter: Jian He
Assignee: Jian He
Priority: Blocker
 Attachments: MAPREDUCE-5476.patch, YARN-917.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5476) Job can fail when RM restarts after staging dir is cleaned but before MR successfully unregister with RM

2013-08-21 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5476?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5476:
-

Priority: Blocker  (was: Major)

 Job can fail when RM restarts after staging dir is cleaned but before MR 
 successfully unregister with RM
 

 Key: MAPREDUCE-5476
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5476
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Reporter: Jian He
Assignee: Jian He
Priority: Blocker
 Attachments: MAPREDUCE-5476.patch, YARN-917.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5468) AM recovery does not work for map only jobs

2013-08-21 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5468?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5468:
-

   Resolution: Fixed
Fix Version/s: 2.1.1-beta
   Status: Resolved  (was: Patch Available)

+1 looks good. Good test too.

I just committed this. Thanks [~vinodkv]!

 AM recovery does not work for map only jobs
 ---

 Key: MAPREDUCE-5468
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5468
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Reporter: yeshavora
Assignee: Vinod Kumar Vavilapalli
Priority: Blocker
 Fix For: 2.1.1-beta

 Attachments: MAPREDUCE-5468-20130821.txt


 Map only job (randomwriter, randomtextwriter) restarts from scratch [0% map 
 0% reduce] after RM restart.
 It should resume from the last state when AM restarted.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5466) Historyserver does not refresh the result of restarted jobs after RM restart

2013-08-20 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5466:
-

Priority: Blocker  (was: Major)

 Historyserver does not refresh the result of restarted jobs after RM restart
 

 Key: MAPREDUCE-5466
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5466
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Reporter: yeshavora
Assignee: Jian He
Priority: Blocker
 Attachments: MAPREDUCE-5466.1.patch, MAPREDUCE-5466.patch


 Restart RM when sort job is running and verify that the job passes 
 successfully after RM restarts. 
 Once the job finishes successfully, run job status command for sort job. It 
 shows Job state =FAILED. Job history server does not update the result for 
 the job which restarted after RM restart.
 hadoop job -status job_1375923346354_0003
 13/08/08 01:24:13 INFO mapred.ClientServiceDelegate: Application state is 
 completed. FinalApplicationStatus=SUCCEEDED. Redirecting to job history server
 Job: job_1375923346354_0003
 Job File: 
 hdfs://host1:port1/history/done/2013/08/08/00/job_1375923346354_0003_conf.xml
 Job Tracking URL : 
 http://historyserver:port2/jobhistory/job/job_1375923346354_0003
 Uber job : false
 Number of maps: 80
 Number of reduces: 1
 map() completion: 0.0
 reduce() completion: 0.0
 Job state: FAILED
 retired: false
 reason for failure: There are no failed tasks for the job. Job is failed due 
 to some other reason and reason can be found in the logs.
 Counters not available. Job is retired.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5468) AM recovery does not work for map only jobs

2013-08-20 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5468?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5468:
-

Priority: Blocker  (was: Major)

 AM recovery does not work for map only jobs
 ---

 Key: MAPREDUCE-5468
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5468
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Reporter: yeshavora
Assignee: Vinod Kumar Vavilapalli
Priority: Blocker

 Map only job (randomwriter, randomtextwriter) restarts from scratch [0% map 
 0% reduce] after RM restart.
 It should resume from the last state when AM restarted.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5466) Historyserver does not refresh the result of restarted jobs after RM restart

2013-08-20 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5466:
-

Target Version/s: 2.1.1-beta  (was: 2.1.0-beta)

 Historyserver does not refresh the result of restarted jobs after RM restart
 

 Key: MAPREDUCE-5466
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5466
 Project: Hadoop Map/Reduce
  Issue Type: Bug
Reporter: yeshavora
Assignee: Jian He
Priority: Blocker
 Attachments: MAPREDUCE-5466.1.patch, MAPREDUCE-5466.patch


 Restart RM when sort job is running and verify that the job passes 
 successfully after RM restarts. 
 Once the job finishes successfully, run job status command for sort job. It 
 shows Job state =FAILED. Job history server does not update the result for 
 the job which restarted after RM restart.
 hadoop job -status job_1375923346354_0003
 13/08/08 01:24:13 INFO mapred.ClientServiceDelegate: Application state is 
 completed. FinalApplicationStatus=SUCCEEDED. Redirecting to job history server
 Job: job_1375923346354_0003
 Job File: 
 hdfs://host1:port1/history/done/2013/08/08/00/job_1375923346354_0003_conf.xml
 Job Tracking URL : 
 http://historyserver:port2/jobhistory/job/job_1375923346354_0003
 Uber job : false
 Number of maps: 80
 Number of reduces: 1
 map() completion: 0.0
 reduce() completion: 0.0
 Job state: FAILED
 retired: false
 reason for failure: There are no failed tasks for the job. Job is failed due 
 to some other reason and reason can be found in the logs.
 Counters not available. Job is retired.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-14 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13739629#comment-13739629
 ] 

Arun C Murthy commented on MAPREDUCE-5311:
--

I'm glad we are coming around to understand we can break existing users of 
SLOT_MILLIS.

I wasn't thrilled about using configs and I'm even more leery of adding new 
configs to track existing configs in RM. This can lead to all sorts of 
configuration-error hell with mismatch in configs b/w YarnConfiguration and 
MRConf. I'm against getting into that sort of situations.

A much simpler approach is to re-introduce a LimitedPrivate 
RegisterApplicationMasterResponse.getMinimumResourceCapability as described 
here (http://bit.ly/18uA8zj). This would revert some part of YARN-787.

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-14 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740390#comment-13740390
 ] 

Arun C Murthy commented on MAPREDUCE-5311:
--

I'll say this again, I'm -1 on adding a new config to track another one in the 
scheduler. Expecting admins to remember another config for the same value (one 
for YARN  another for MR) is not kosher - our configuration is bad enough.

We can use the same one in the scheduler (i.e. don't fix YARN-1004) or we can 
add back RegisterApplicationMasterResponse.getMinimumResourceCapability as 
LimitedPrivate(MapReduce).

I'm ok with either of the above. 

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-14 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740577#comment-13740577
 ] 

Arun C Murthy commented on MAPREDUCE-5311:
--

This isn't a special case for 0.23. 

It will break tools built atop both hadoop-1.x  hadoop-0.23.*.

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-14 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740598#comment-13740598
 ] 

Arun C Murthy commented on MAPREDUCE-5311:
--

I'm ok with Vinod's suggestion too.

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-14 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740613#comment-13740613
 ] 

Arun C Murthy commented on MAPREDUCE-5311:
--

[~tucu00] Can you please decide on one of the 3 options available which are 
reasonable? 

I'm ready to roll 2.1.0-rc2 ATM, this is the only issue I can think of. Thanks.

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-14 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740671#comment-13740671
 ] 

Arun C Murthy commented on MAPREDUCE-5311:
--

I'll repeat that I'm against multiple configs for the same property which leads 
to operator errors (admins setting one but forgetting to set the other) and am 
willing to live with other alternatives which seem better from a technical 
standpoint.

bq. Why we don't use the Hadoop 1 JT properties for the same, in the 
mapred-site.xml documenting how they have to be set to the MIN of the scheduler 
for SLOT_MILLIS counter to kick in? To me this seem a much more correct way of 
doing it.

ATM, I don't see where this argument is going at all - frankly, we seem to be 
making YARN-1004 much bigger deal than it is.

Several folks have expressed concerns about multiple configs and yet we don't 
seem to converge; this is in spite of, IMO, very reasonable alternatives. 

I'm inclined to go ahead with rc2 without this  YARN-1004. We can revisit both 
for 2.1.1 via deprecation of properties in YARN-1004 etc. if necessary.

My plan is to cut an rc2 by noon PST tmrw (9/15). Let's revisit this after if 
we can agree today. Thanks.

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Comment Edited] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-14 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13740671#comment-13740671
 ] 

Arun C Murthy edited comment on MAPREDUCE-5311 at 8/15/13 5:04 AM:
---

I'll repeat that I'm against multiple configs for the same property which leads 
to operator errors (admins setting one but forgetting to set the other) and am 
willing to live with other alternatives which seem better from a technical 
standpoint.

bq. Why we don't use the Hadoop 1 JT properties for the same, in the 
mapred-site.xml documenting how they have to be set to the MIN of the scheduler 
for SLOT_MILLIS counter to kick in? To me this seem a much more correct way of 
doing it.

ATM, I don't see where this argument is going at all - frankly, we seem to be 
making YARN-1004 much bigger deal than it is.

Several folks have expressed concerns about multiple configs and yet we don't 
seem to converge; this is in spite of, IMO, very reasonable alternatives. 

I'm inclined to go ahead with rc2 without MAPREDUCE-5311  YARN-1004. Of 
course, we can revisit both for 2.1.1 via deprecation of yarn.scheduler.min 
being discussed in YARN-1004. 

I'll stress that we can continue talking and get both in once we all agree.

My plan is to cut an rc2 by noon PST tmrw (9/15). Let's revisit this discussion 
after if we can't converge before then. Makes sense? Thanks.

  was (Author: acmurthy):
I'll repeat that I'm against multiple configs for the same property which 
leads to operator errors (admins setting one but forgetting to set the other) 
and am willing to live with other alternatives which seem better from a 
technical standpoint.

bq. Why we don't use the Hadoop 1 JT properties for the same, in the 
mapred-site.xml documenting how they have to be set to the MIN of the scheduler 
for SLOT_MILLIS counter to kick in? To me this seem a much more correct way of 
doing it.

ATM, I don't see where this argument is going at all - frankly, we seem to be 
making YARN-1004 much bigger deal than it is.

Several folks have expressed concerns about multiple configs and yet we don't 
seem to converge; this is in spite of, IMO, very reasonable alternatives. 

I'm inclined to go ahead with rc2 without this  YARN-1004. We can revisit both 
for 2.1.1 via deprecation of properties in YARN-1004 etc. if necessary.

My plan is to cut an rc2 by noon PST tmrw (9/15). Let's revisit this after if 
we can agree today. Thanks.
  
 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-13 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13738236#comment-13738236
 ] 

Arun C Murthy commented on MAPREDUCE-5311:
--

I've repeatedly pointed out that SLOT_MILLIS is something which is very useful 
and is used by a lot of end-users. I know of a lot of users who have built 
rudimentary reporting/charge-back systems using SLOT_MILLIS. Breaking 
compatibility (semantic or api) is not an option. We've done a lot of work on 
MR to keep it compatible so far.

The more I think about it, the more strongly I believe YARN-787 went too far by 
removing 'minimum' from the API. We can/should keep 
RegisterApplicationMasterResponse.getMinimumResourceCapability as a 
LimitedPrivate api and use it only for calculation of SLOT_MILLIS. We don't 
have to expose it to users or other applications. Makes sense?

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-13 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13738528#comment-13738528
 ] 

Arun C Murthy commented on MAPREDUCE-5311:
--

Let's up-level.

We agree we can't break compatibility for existing applications ?

If so, we can figure ways around how to fix YARN-1004, MAPREDUCE-5310 etc. 
without breaking compatibility.

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-13 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13738663#comment-13738663
 ] 

Arun C Murthy commented on MAPREDUCE-5311:
--

bq. As slots no longer exist in MR2, we cannot avoid breaking compatibility for 
some existing applications.

The original intention was to model the old notion of 'slots' as a minimum 
container.

So, we have 2 cases:
# mapred.cluster.(map,reduce).memory.mb set to -1 for the job. In this case, we 
just divide the 'timing' by 1.
# mapred.cluster.(map,reduce).memory.mb set to non-zero then we normalize to 
yarn.scheduler.minimum-allocation-mb and divide 'timing' by that number.

Makes sense?

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-09 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13735003#comment-13735003
 ] 

Arun C Murthy commented on MAPREDUCE-5311:
--

We seem to understand that we can't remove SLOTS_MILLIS_MAPS - that will break 
BC.

For now, let's punt on this for 2.1.0-beta since we can add new counters 
(CONTAINERS_MILLIS_MAPS) after a bit more discussion.

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-09 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5311:
-

Fix Version/s: (was: 2.1.0-beta)

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-09 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5311:
-

Target Version/s:   (was: 2.1.0-beta)

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5450) Unnecessary Configuration instantiation in IFileInputStream slows down merge - Port to branch-1

2013-08-08 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5450?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733826#comment-13733826
 ] 

Arun C Murthy commented on MAPREDUCE-5450:
--

Sorry, I meant for 1.2.2 (if necessary).

 Unnecessary Configuration instantiation in IFileInputStream slows down merge 
 - Port to branch-1
 ---

 Key: MAPREDUCE-5450
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5450
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv1
Affects Versions: 1.1.0
Reporter: Stanislav Barton
Assignee: Stanislav Barton
Priority: Blocker
 Fix For: 1.3.0

 Attachments: MAPREDUCE-5450-1.1.0.txt, mapreduce-5450.txt


 We are using hadoop-2.0.0+1357-1.cdh4.3.0.p0.21 with MRv1. After upgrade from 
 4.1.2 to 4.3.0, I have noticed some performance deterioration in our MR job 
 in the Reduce phase. The MR job has usually 10 000 map tasks (10 000 files on 
 input each about 100MB) and 6 000 reducers (one reducer per table region). I 
 was trying to figure out what at which phase the slow down appears (firstly I 
 suspected that the slow gathering of the 1 map output files is the 
 culprit) and found out that the problem is not reading the map output (the 
 shuffle) but the sort/merge phase that follows - the last and actual reduce 
 phase is fast. I have tried to up the io.sort.factor because I thought the 
 lots of small files are being merged on disk, but again upping that to 1000 
 didnt do any difference. I have then printed the stack trace and found out 
 that the problem is initialization of the 
 org.apache.hadoop.mapred.IFileInputStream namely the creation of the 
 Configuration object which is not propagated along from earlier context, see 
 the stack trace:
 Thread 13332: (state = IN_NATIVE)
  - java.io.UnixFileSystem.getBooleanAttributes0(java.io.File) @bci=0 
 (Compiled frame; information may be imprecise)
  - java.io.UnixFileSystem.getBooleanAttributes(java.io.File) @bci=2, line=228 
 (Compiled frame)
  - java.io.File.exists() @bci=20, line=733 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.getResource(java.lang.String, boolean) 
 @bci=136, line=999 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.findResource(java.lang.String, boolean) 
 @bci=3, line=966 (Compiled frame)
  - sun.misc.URLClassPath.findResource(java.lang.String, boolean) @bci=17, 
 line=146 (Compiled frame)
  - java.net.URLClassLoader$2.run() @bci=12, line=385 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction, 
 java.security.AccessControlContext) @bci=0 (Compiled frame)
  - java.net.URLClassLoader.findResource(java.lang.String) @bci=13, line=382 
 (Compiled frame)
  - java.lang.ClassLoader.getResource(java.lang.String) @bci=30, line=1002 
 (Compiled frame)
  - java.lang.ClassLoader.getResourceAsStream(java.lang.String) @bci=2, 
 line=1192 (Compiled frame)
  - javax.xml.parsers.SecuritySupport$4.run() @bci=26, line=96 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction) 
 @bci=0 (Compiled frame)
  - 
 javax.xml.parsers.SecuritySupport.getResourceAsStream(java.lang.ClassLoader, 
 java.lang.String) @bci=10, line=89 (Compiled frame)
  - javax.xml.parsers.FactoryFinder.findJarServiceProvider(java.lang.String) 
 @bci=38, line=250 (Interpreted frame)
  - javax.xml.parsers.FactoryFinder.find(java.lang.String, java.lang.String) 
 @bci=273, line=223 (Interpreted frame)
  - javax.xml.parsers.DocumentBuilderFactory.newInstance() @bci=4, line=123 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResource(java.util.Properties, 
 org.apache.hadoop.conf.Configuration$Resource, boolean) @bci=16, line=1890 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResources(java.util.Properties, 
 java.util.ArrayList, boolean) @bci=49, line=1867 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getProps() @bci=43, line=1785 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.get(java.lang.String) @bci=35, 
 line=712 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getTrimmed(java.lang.String) @bci=2, 
 line=731 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getBoolean(java.lang.String, boolean) 
 @bci=2, line=1047 (Interpreted frame)
  - org.apache.hadoop.mapred.IFileInputStream.init(java.io.InputStream, 
 long, org.apache.hadoop.conf.Configuration) @bci=111, line=93 (Interpreted 
 frame)
  - 
 org.apache.hadoop.mapred.IFile$Reader.init(org.apache.hadoop.conf.Configuration,
  org.apache.hadoop.fs.FSDataInputStream, long, 
 org.apache.hadoop.io.compress.CompressionCodec, 
 org.apache.hadoop.mapred.Counters$Counter) @bci=60, line=303 (Interpreted 
 frame)
  - 
 

[jira] [Commented] (MAPREDUCE-5450) Unnecessary Configuration instantiation in IFileInputStream slows down merge - Port to branch-1

2013-08-08 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5450?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733823#comment-13733823
 ] 

Arun C Murthy commented on MAPREDUCE-5450:
--

Let's get [~stanislav.barton]'s original patch into branch-1.2 for 1.2.1.

Any objections?


 Unnecessary Configuration instantiation in IFileInputStream slows down merge 
 - Port to branch-1
 ---

 Key: MAPREDUCE-5450
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5450
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv1
Affects Versions: 1.1.0
Reporter: Stanislav Barton
Assignee: Stanislav Barton
Priority: Blocker
 Fix For: 1.3.0

 Attachments: MAPREDUCE-5450-1.1.0.txt, mapreduce-5450.txt


 We are using hadoop-2.0.0+1357-1.cdh4.3.0.p0.21 with MRv1. After upgrade from 
 4.1.2 to 4.3.0, I have noticed some performance deterioration in our MR job 
 in the Reduce phase. The MR job has usually 10 000 map tasks (10 000 files on 
 input each about 100MB) and 6 000 reducers (one reducer per table region). I 
 was trying to figure out what at which phase the slow down appears (firstly I 
 suspected that the slow gathering of the 1 map output files is the 
 culprit) and found out that the problem is not reading the map output (the 
 shuffle) but the sort/merge phase that follows - the last and actual reduce 
 phase is fast. I have tried to up the io.sort.factor because I thought the 
 lots of small files are being merged on disk, but again upping that to 1000 
 didnt do any difference. I have then printed the stack trace and found out 
 that the problem is initialization of the 
 org.apache.hadoop.mapred.IFileInputStream namely the creation of the 
 Configuration object which is not propagated along from earlier context, see 
 the stack trace:
 Thread 13332: (state = IN_NATIVE)
  - java.io.UnixFileSystem.getBooleanAttributes0(java.io.File) @bci=0 
 (Compiled frame; information may be imprecise)
  - java.io.UnixFileSystem.getBooleanAttributes(java.io.File) @bci=2, line=228 
 (Compiled frame)
  - java.io.File.exists() @bci=20, line=733 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.getResource(java.lang.String, boolean) 
 @bci=136, line=999 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.findResource(java.lang.String, boolean) 
 @bci=3, line=966 (Compiled frame)
  - sun.misc.URLClassPath.findResource(java.lang.String, boolean) @bci=17, 
 line=146 (Compiled frame)
  - java.net.URLClassLoader$2.run() @bci=12, line=385 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction, 
 java.security.AccessControlContext) @bci=0 (Compiled frame)
  - java.net.URLClassLoader.findResource(java.lang.String) @bci=13, line=382 
 (Compiled frame)
  - java.lang.ClassLoader.getResource(java.lang.String) @bci=30, line=1002 
 (Compiled frame)
  - java.lang.ClassLoader.getResourceAsStream(java.lang.String) @bci=2, 
 line=1192 (Compiled frame)
  - javax.xml.parsers.SecuritySupport$4.run() @bci=26, line=96 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction) 
 @bci=0 (Compiled frame)
  - 
 javax.xml.parsers.SecuritySupport.getResourceAsStream(java.lang.ClassLoader, 
 java.lang.String) @bci=10, line=89 (Compiled frame)
  - javax.xml.parsers.FactoryFinder.findJarServiceProvider(java.lang.String) 
 @bci=38, line=250 (Interpreted frame)
  - javax.xml.parsers.FactoryFinder.find(java.lang.String, java.lang.String) 
 @bci=273, line=223 (Interpreted frame)
  - javax.xml.parsers.DocumentBuilderFactory.newInstance() @bci=4, line=123 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResource(java.util.Properties, 
 org.apache.hadoop.conf.Configuration$Resource, boolean) @bci=16, line=1890 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResources(java.util.Properties, 
 java.util.ArrayList, boolean) @bci=49, line=1867 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getProps() @bci=43, line=1785 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.get(java.lang.String) @bci=35, 
 line=712 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getTrimmed(java.lang.String) @bci=2, 
 line=731 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getBoolean(java.lang.String, boolean) 
 @bci=2, line=1047 (Interpreted frame)
  - org.apache.hadoop.mapred.IFileInputStream.init(java.io.InputStream, 
 long, org.apache.hadoop.conf.Configuration) @bci=111, line=93 (Interpreted 
 frame)
  - 
 org.apache.hadoop.mapred.IFile$Reader.init(org.apache.hadoop.conf.Configuration,
  org.apache.hadoop.fs.FSDataInputStream, long, 
 org.apache.hadoop.io.compress.CompressionCodec, 
 

[jira] [Updated] (MAPREDUCE-5450) Unnecessary Configuration instantiation in IFileInputStream slows down merge - Port to branch-1

2013-08-08 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5450?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5450:
-

Fix Version/s: (was: 1.3.0)
   1.2.1

 Unnecessary Configuration instantiation in IFileInputStream slows down merge 
 - Port to branch-1
 ---

 Key: MAPREDUCE-5450
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5450
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv1
Affects Versions: 1.1.0
Reporter: Stanislav Barton
Assignee: Stanislav Barton
Priority: Blocker
 Fix For: 1.2.1

 Attachments: MAPREDUCE-5450-1.1.0.txt, mapreduce-5450.txt


 We are using hadoop-2.0.0+1357-1.cdh4.3.0.p0.21 with MRv1. After upgrade from 
 4.1.2 to 4.3.0, I have noticed some performance deterioration in our MR job 
 in the Reduce phase. The MR job has usually 10 000 map tasks (10 000 files on 
 input each about 100MB) and 6 000 reducers (one reducer per table region). I 
 was trying to figure out what at which phase the slow down appears (firstly I 
 suspected that the slow gathering of the 1 map output files is the 
 culprit) and found out that the problem is not reading the map output (the 
 shuffle) but the sort/merge phase that follows - the last and actual reduce 
 phase is fast. I have tried to up the io.sort.factor because I thought the 
 lots of small files are being merged on disk, but again upping that to 1000 
 didnt do any difference. I have then printed the stack trace and found out 
 that the problem is initialization of the 
 org.apache.hadoop.mapred.IFileInputStream namely the creation of the 
 Configuration object which is not propagated along from earlier context, see 
 the stack trace:
 Thread 13332: (state = IN_NATIVE)
  - java.io.UnixFileSystem.getBooleanAttributes0(java.io.File) @bci=0 
 (Compiled frame; information may be imprecise)
  - java.io.UnixFileSystem.getBooleanAttributes(java.io.File) @bci=2, line=228 
 (Compiled frame)
  - java.io.File.exists() @bci=20, line=733 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.getResource(java.lang.String, boolean) 
 @bci=136, line=999 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.findResource(java.lang.String, boolean) 
 @bci=3, line=966 (Compiled frame)
  - sun.misc.URLClassPath.findResource(java.lang.String, boolean) @bci=17, 
 line=146 (Compiled frame)
  - java.net.URLClassLoader$2.run() @bci=12, line=385 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction, 
 java.security.AccessControlContext) @bci=0 (Compiled frame)
  - java.net.URLClassLoader.findResource(java.lang.String) @bci=13, line=382 
 (Compiled frame)
  - java.lang.ClassLoader.getResource(java.lang.String) @bci=30, line=1002 
 (Compiled frame)
  - java.lang.ClassLoader.getResourceAsStream(java.lang.String) @bci=2, 
 line=1192 (Compiled frame)
  - javax.xml.parsers.SecuritySupport$4.run() @bci=26, line=96 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction) 
 @bci=0 (Compiled frame)
  - 
 javax.xml.parsers.SecuritySupport.getResourceAsStream(java.lang.ClassLoader, 
 java.lang.String) @bci=10, line=89 (Compiled frame)
  - javax.xml.parsers.FactoryFinder.findJarServiceProvider(java.lang.String) 
 @bci=38, line=250 (Interpreted frame)
  - javax.xml.parsers.FactoryFinder.find(java.lang.String, java.lang.String) 
 @bci=273, line=223 (Interpreted frame)
  - javax.xml.parsers.DocumentBuilderFactory.newInstance() @bci=4, line=123 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResource(java.util.Properties, 
 org.apache.hadoop.conf.Configuration$Resource, boolean) @bci=16, line=1890 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResources(java.util.Properties, 
 java.util.ArrayList, boolean) @bci=49, line=1867 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getProps() @bci=43, line=1785 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.get(java.lang.String) @bci=35, 
 line=712 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getTrimmed(java.lang.String) @bci=2, 
 line=731 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getBoolean(java.lang.String, boolean) 
 @bci=2, line=1047 (Interpreted frame)
  - org.apache.hadoop.mapred.IFileInputStream.init(java.io.InputStream, 
 long, org.apache.hadoop.conf.Configuration) @bci=111, line=93 (Interpreted 
 frame)
  - 
 org.apache.hadoop.mapred.IFile$Reader.init(org.apache.hadoop.conf.Configuration,
  org.apache.hadoop.fs.FSDataInputStream, long, 
 org.apache.hadoop.io.compress.CompressionCodec, 
 org.apache.hadoop.mapred.Counters$Counter) @bci=60, line=303 (Interpreted 
 frame)
  - 
 

[jira] [Commented] (MAPREDUCE-5450) Unnecessary Configuration instantiation in IFileInputStream slows down merge - Port to branch-1

2013-08-08 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5450?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13733834#comment-13733834
 ] 

Arun C Murthy commented on MAPREDUCE-5450:
--

I merged this to branch-1.2 too. Thanks guys!

 Unnecessary Configuration instantiation in IFileInputStream slows down merge 
 - Port to branch-1
 ---

 Key: MAPREDUCE-5450
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5450
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv1
Affects Versions: 1.1.0
Reporter: Stanislav Barton
Assignee: Stanislav Barton
Priority: Blocker
 Fix For: 1.2.1

 Attachments: MAPREDUCE-5450-1.1.0.txt, mapreduce-5450.txt


 We are using hadoop-2.0.0+1357-1.cdh4.3.0.p0.21 with MRv1. After upgrade from 
 4.1.2 to 4.3.0, I have noticed some performance deterioration in our MR job 
 in the Reduce phase. The MR job has usually 10 000 map tasks (10 000 files on 
 input each about 100MB) and 6 000 reducers (one reducer per table region). I 
 was trying to figure out what at which phase the slow down appears (firstly I 
 suspected that the slow gathering of the 1 map output files is the 
 culprit) and found out that the problem is not reading the map output (the 
 shuffle) but the sort/merge phase that follows - the last and actual reduce 
 phase is fast. I have tried to up the io.sort.factor because I thought the 
 lots of small files are being merged on disk, but again upping that to 1000 
 didnt do any difference. I have then printed the stack trace and found out 
 that the problem is initialization of the 
 org.apache.hadoop.mapred.IFileInputStream namely the creation of the 
 Configuration object which is not propagated along from earlier context, see 
 the stack trace:
 Thread 13332: (state = IN_NATIVE)
  - java.io.UnixFileSystem.getBooleanAttributes0(java.io.File) @bci=0 
 (Compiled frame; information may be imprecise)
  - java.io.UnixFileSystem.getBooleanAttributes(java.io.File) @bci=2, line=228 
 (Compiled frame)
  - java.io.File.exists() @bci=20, line=733 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.getResource(java.lang.String, boolean) 
 @bci=136, line=999 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.findResource(java.lang.String, boolean) 
 @bci=3, line=966 (Compiled frame)
  - sun.misc.URLClassPath.findResource(java.lang.String, boolean) @bci=17, 
 line=146 (Compiled frame)
  - java.net.URLClassLoader$2.run() @bci=12, line=385 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction, 
 java.security.AccessControlContext) @bci=0 (Compiled frame)
  - java.net.URLClassLoader.findResource(java.lang.String) @bci=13, line=382 
 (Compiled frame)
  - java.lang.ClassLoader.getResource(java.lang.String) @bci=30, line=1002 
 (Compiled frame)
  - java.lang.ClassLoader.getResourceAsStream(java.lang.String) @bci=2, 
 line=1192 (Compiled frame)
  - javax.xml.parsers.SecuritySupport$4.run() @bci=26, line=96 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction) 
 @bci=0 (Compiled frame)
  - 
 javax.xml.parsers.SecuritySupport.getResourceAsStream(java.lang.ClassLoader, 
 java.lang.String) @bci=10, line=89 (Compiled frame)
  - javax.xml.parsers.FactoryFinder.findJarServiceProvider(java.lang.String) 
 @bci=38, line=250 (Interpreted frame)
  - javax.xml.parsers.FactoryFinder.find(java.lang.String, java.lang.String) 
 @bci=273, line=223 (Interpreted frame)
  - javax.xml.parsers.DocumentBuilderFactory.newInstance() @bci=4, line=123 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResource(java.util.Properties, 
 org.apache.hadoop.conf.Configuration$Resource, boolean) @bci=16, line=1890 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResources(java.util.Properties, 
 java.util.ArrayList, boolean) @bci=49, line=1867 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getProps() @bci=43, line=1785 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.get(java.lang.String) @bci=35, 
 line=712 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getTrimmed(java.lang.String) @bci=2, 
 line=731 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getBoolean(java.lang.String, boolean) 
 @bci=2, line=1047 (Interpreted frame)
  - org.apache.hadoop.mapred.IFileInputStream.init(java.io.InputStream, 
 long, org.apache.hadoop.conf.Configuration) @bci=111, line=93 (Interpreted 
 frame)
  - 
 org.apache.hadoop.mapred.IFile$Reader.init(org.apache.hadoop.conf.Configuration,
  org.apache.hadoop.fs.FSDataInputStream, long, 
 org.apache.hadoop.io.compress.CompressionCodec, 
 org.apache.hadoop.mapred.Counters$Counter) @bci=60, line=303 (Interpreted 
 frame)
  

[jira] [Updated] (MAPREDUCE-5450) Unnecessary Configuration instantiation in IFileInputStream slows down merge - Port to branch-1

2013-08-08 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5450?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5450:
-

Fix Version/s: (was: 1.2.1)
   1.2.2

 Unnecessary Configuration instantiation in IFileInputStream slows down merge 
 - Port to branch-1
 ---

 Key: MAPREDUCE-5450
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5450
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv1
Affects Versions: 1.1.0
Reporter: Stanislav Barton
Assignee: Stanislav Barton
Priority: Blocker
 Fix For: 1.2.2

 Attachments: MAPREDUCE-5450-1.1.0.txt, mapreduce-5450.txt


 We are using hadoop-2.0.0+1357-1.cdh4.3.0.p0.21 with MRv1. After upgrade from 
 4.1.2 to 4.3.0, I have noticed some performance deterioration in our MR job 
 in the Reduce phase. The MR job has usually 10 000 map tasks (10 000 files on 
 input each about 100MB) and 6 000 reducers (one reducer per table region). I 
 was trying to figure out what at which phase the slow down appears (firstly I 
 suspected that the slow gathering of the 1 map output files is the 
 culprit) and found out that the problem is not reading the map output (the 
 shuffle) but the sort/merge phase that follows - the last and actual reduce 
 phase is fast. I have tried to up the io.sort.factor because I thought the 
 lots of small files are being merged on disk, but again upping that to 1000 
 didnt do any difference. I have then printed the stack trace and found out 
 that the problem is initialization of the 
 org.apache.hadoop.mapred.IFileInputStream namely the creation of the 
 Configuration object which is not propagated along from earlier context, see 
 the stack trace:
 Thread 13332: (state = IN_NATIVE)
  - java.io.UnixFileSystem.getBooleanAttributes0(java.io.File) @bci=0 
 (Compiled frame; information may be imprecise)
  - java.io.UnixFileSystem.getBooleanAttributes(java.io.File) @bci=2, line=228 
 (Compiled frame)
  - java.io.File.exists() @bci=20, line=733 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.getResource(java.lang.String, boolean) 
 @bci=136, line=999 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.findResource(java.lang.String, boolean) 
 @bci=3, line=966 (Compiled frame)
  - sun.misc.URLClassPath.findResource(java.lang.String, boolean) @bci=17, 
 line=146 (Compiled frame)
  - java.net.URLClassLoader$2.run() @bci=12, line=385 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction, 
 java.security.AccessControlContext) @bci=0 (Compiled frame)
  - java.net.URLClassLoader.findResource(java.lang.String) @bci=13, line=382 
 (Compiled frame)
  - java.lang.ClassLoader.getResource(java.lang.String) @bci=30, line=1002 
 (Compiled frame)
  - java.lang.ClassLoader.getResourceAsStream(java.lang.String) @bci=2, 
 line=1192 (Compiled frame)
  - javax.xml.parsers.SecuritySupport$4.run() @bci=26, line=96 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction) 
 @bci=0 (Compiled frame)
  - 
 javax.xml.parsers.SecuritySupport.getResourceAsStream(java.lang.ClassLoader, 
 java.lang.String) @bci=10, line=89 (Compiled frame)
  - javax.xml.parsers.FactoryFinder.findJarServiceProvider(java.lang.String) 
 @bci=38, line=250 (Interpreted frame)
  - javax.xml.parsers.FactoryFinder.find(java.lang.String, java.lang.String) 
 @bci=273, line=223 (Interpreted frame)
  - javax.xml.parsers.DocumentBuilderFactory.newInstance() @bci=4, line=123 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResource(java.util.Properties, 
 org.apache.hadoop.conf.Configuration$Resource, boolean) @bci=16, line=1890 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResources(java.util.Properties, 
 java.util.ArrayList, boolean) @bci=49, line=1867 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getProps() @bci=43, line=1785 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.get(java.lang.String) @bci=35, 
 line=712 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getTrimmed(java.lang.String) @bci=2, 
 line=731 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getBoolean(java.lang.String, boolean) 
 @bci=2, line=1047 (Interpreted frame)
  - org.apache.hadoop.mapred.IFileInputStream.init(java.io.InputStream, 
 long, org.apache.hadoop.conf.Configuration) @bci=111, line=93 (Interpreted 
 frame)
  - 
 org.apache.hadoop.mapred.IFile$Reader.init(org.apache.hadoop.conf.Configuration,
  org.apache.hadoop.fs.FSDataInputStream, long, 
 org.apache.hadoop.io.compress.CompressionCodec, 
 org.apache.hadoop.mapred.Counters$Counter) @bci=60, line=303 (Interpreted 
 frame)
  - 
 

[jira] [Commented] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-06 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731444#comment-13731444
 ] 

Arun C Murthy commented on MAPREDUCE-5311:
--

This will break every single MR app. I'm -1 on this.

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5311) Replace SLOTS_MILLIS counters with MEM_MILLIS

2013-08-06 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5311?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13731445#comment-13731445
 ] 

Arun C Murthy commented on MAPREDUCE-5311:
--

To Sandy's point - several folks do use mapred.cluster.(map,reduce).memory.mb, 
so it will be a big BC break for them.

Also, what does this mean for BC after we spent all the time fixing 
MAPREDUCE-5108.

 Replace SLOTS_MILLIS counters with MEM_MILLIS
 -

 Key: MAPREDUCE-5311
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5311
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: applicationmaster
Affects Versions: 2.0.4-alpha
Reporter: Alejandro Abdelnur
Assignee: Sandy Ryza
Priority: Blocker
 Fix For: 2.1.0-beta

 Attachments: MAPREDUCE-5311-1.patch, MAPREDUCE-5311.patch, 
 MAPREDUCE-5311.patch


 Per discussion in MAPREDUCE-5310 and comments in the code we should remove 
 all the related logic and just leave the counter constant for backwards 
 compatibility and deprecate the counter constants.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (MAPREDUCE-5399) Large number of map tasks cause slow sort at reduce phase, invariant to amount of data to sort

2013-08-05 Thread Arun C Murthy (JIRA)

[ 
https://issues.apache.org/jira/browse/MAPREDUCE-5399?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13730102#comment-13730102
 ] 

Arun C Murthy commented on MAPREDUCE-5399:
--

[~sandyr] Can you pls commit this to trunk, branch-2, branch-2.1-beta and 
branch-2.1.0-beta? Thanks!

 Large number of map tasks cause slow sort at reduce phase, invariant to 
 amount of data to sort
 --

 Key: MAPREDUCE-5399
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5399
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv1, mrv2
Affects Versions: 1.1.0, 2.0.2-alpha
Reporter: Stanislav Barton
Assignee: Stanislav Barton
Priority: Blocker
 Attachments: MAPREDUCE-5399.patch


 We are using hadoop-2.0.0+1357-1.cdh4.3.0.p0.21 with MRv1. After upgrade from 
 4.1.2 to 4.3.0, I have noticed some performance deterioration in our MR job 
 in the Reduce phase. The MR job has usually 10 000 map tasks (10 000 files on 
 input each about 100MB) and 6 000 reducers (one reducer per table region). I 
 was trying to figure out what at which phase the slow down appears (firstly I 
 suspected that the slow gathering of the 1 map output files is the 
 culprit) and found out that the problem is not reading the map output (the 
 shuffle) but the sort/merge phase that follows - the last and actual reduce 
 phase is fast. I have tried to up the io.sort.factor because I thought the 
 lots of small files are being merged on disk, but again upping that to 1000 
 didnt do any difference. I have then printed the stack trace and found out 
 that the problem is initialization of the 
 org.apache.hadoop.mapred.IFileInputStream namely the creation of the 
 Configuration object which is not propagated along from earlier context, see 
 the stack trace:
 Thread 13332: (state = IN_NATIVE)
  - java.io.UnixFileSystem.getBooleanAttributes0(java.io.File) @bci=0 
 (Compiled frame; information may be imprecise)
  - java.io.UnixFileSystem.getBooleanAttributes(java.io.File) @bci=2, line=228 
 (Compiled frame)
  - java.io.File.exists() @bci=20, line=733 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.getResource(java.lang.String, boolean) 
 @bci=136, line=999 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.findResource(java.lang.String, boolean) 
 @bci=3, line=966 (Compiled frame)
  - sun.misc.URLClassPath.findResource(java.lang.String, boolean) @bci=17, 
 line=146 (Compiled frame)
  - java.net.URLClassLoader$2.run() @bci=12, line=385 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction, 
 java.security.AccessControlContext) @bci=0 (Compiled frame)
  - java.net.URLClassLoader.findResource(java.lang.String) @bci=13, line=382 
 (Compiled frame)
  - java.lang.ClassLoader.getResource(java.lang.String) @bci=30, line=1002 
 (Compiled frame)
  - java.lang.ClassLoader.getResourceAsStream(java.lang.String) @bci=2, 
 line=1192 (Compiled frame)
  - javax.xml.parsers.SecuritySupport$4.run() @bci=26, line=96 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction) 
 @bci=0 (Compiled frame)
  - 
 javax.xml.parsers.SecuritySupport.getResourceAsStream(java.lang.ClassLoader, 
 java.lang.String) @bci=10, line=89 (Compiled frame)
  - javax.xml.parsers.FactoryFinder.findJarServiceProvider(java.lang.String) 
 @bci=38, line=250 (Interpreted frame)
  - javax.xml.parsers.FactoryFinder.find(java.lang.String, java.lang.String) 
 @bci=273, line=223 (Interpreted frame)
  - javax.xml.parsers.DocumentBuilderFactory.newInstance() @bci=4, line=123 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResource(java.util.Properties, 
 org.apache.hadoop.conf.Configuration$Resource, boolean) @bci=16, line=1890 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResources(java.util.Properties, 
 java.util.ArrayList, boolean) @bci=49, line=1867 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getProps() @bci=43, line=1785 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.get(java.lang.String) @bci=35, 
 line=712 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getTrimmed(java.lang.String) @bci=2, 
 line=731 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getBoolean(java.lang.String, boolean) 
 @bci=2, line=1047 (Interpreted frame)
  - org.apache.hadoop.mapred.IFileInputStream.init(java.io.InputStream, 
 long, org.apache.hadoop.conf.Configuration) @bci=111, line=93 (Interpreted 
 frame)
  - 
 org.apache.hadoop.mapred.IFile$Reader.init(org.apache.hadoop.conf.Configuration,
  org.apache.hadoop.fs.FSDataInputStream, long, 
 org.apache.hadoop.io.compress.CompressionCodec, 
 org.apache.hadoop.mapred.Counters$Counter) @bci=60, line=303 

[jira] [Updated] (MAPREDUCE-5399) Large number of map tasks cause slow sort at reduce phase, invariant to amount of data to sort

2013-08-05 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5399?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5399:
-

Target Version/s: 2.1.0-beta

 Large number of map tasks cause slow sort at reduce phase, invariant to 
 amount of data to sort
 --

 Key: MAPREDUCE-5399
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5399
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv1, mrv2
Affects Versions: 1.1.0, 2.0.2-alpha
Reporter: Stanislav Barton
Assignee: Stanislav Barton
Priority: Blocker
 Attachments: MAPREDUCE-5399.patch


 We are using hadoop-2.0.0+1357-1.cdh4.3.0.p0.21 with MRv1. After upgrade from 
 4.1.2 to 4.3.0, I have noticed some performance deterioration in our MR job 
 in the Reduce phase. The MR job has usually 10 000 map tasks (10 000 files on 
 input each about 100MB) and 6 000 reducers (one reducer per table region). I 
 was trying to figure out what at which phase the slow down appears (firstly I 
 suspected that the slow gathering of the 1 map output files is the 
 culprit) and found out that the problem is not reading the map output (the 
 shuffle) but the sort/merge phase that follows - the last and actual reduce 
 phase is fast. I have tried to up the io.sort.factor because I thought the 
 lots of small files are being merged on disk, but again upping that to 1000 
 didnt do any difference. I have then printed the stack trace and found out 
 that the problem is initialization of the 
 org.apache.hadoop.mapred.IFileInputStream namely the creation of the 
 Configuration object which is not propagated along from earlier context, see 
 the stack trace:
 Thread 13332: (state = IN_NATIVE)
  - java.io.UnixFileSystem.getBooleanAttributes0(java.io.File) @bci=0 
 (Compiled frame; information may be imprecise)
  - java.io.UnixFileSystem.getBooleanAttributes(java.io.File) @bci=2, line=228 
 (Compiled frame)
  - java.io.File.exists() @bci=20, line=733 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.getResource(java.lang.String, boolean) 
 @bci=136, line=999 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.findResource(java.lang.String, boolean) 
 @bci=3, line=966 (Compiled frame)
  - sun.misc.URLClassPath.findResource(java.lang.String, boolean) @bci=17, 
 line=146 (Compiled frame)
  - java.net.URLClassLoader$2.run() @bci=12, line=385 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction, 
 java.security.AccessControlContext) @bci=0 (Compiled frame)
  - java.net.URLClassLoader.findResource(java.lang.String) @bci=13, line=382 
 (Compiled frame)
  - java.lang.ClassLoader.getResource(java.lang.String) @bci=30, line=1002 
 (Compiled frame)
  - java.lang.ClassLoader.getResourceAsStream(java.lang.String) @bci=2, 
 line=1192 (Compiled frame)
  - javax.xml.parsers.SecuritySupport$4.run() @bci=26, line=96 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction) 
 @bci=0 (Compiled frame)
  - 
 javax.xml.parsers.SecuritySupport.getResourceAsStream(java.lang.ClassLoader, 
 java.lang.String) @bci=10, line=89 (Compiled frame)
  - javax.xml.parsers.FactoryFinder.findJarServiceProvider(java.lang.String) 
 @bci=38, line=250 (Interpreted frame)
  - javax.xml.parsers.FactoryFinder.find(java.lang.String, java.lang.String) 
 @bci=273, line=223 (Interpreted frame)
  - javax.xml.parsers.DocumentBuilderFactory.newInstance() @bci=4, line=123 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResource(java.util.Properties, 
 org.apache.hadoop.conf.Configuration$Resource, boolean) @bci=16, line=1890 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResources(java.util.Properties, 
 java.util.ArrayList, boolean) @bci=49, line=1867 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getProps() @bci=43, line=1785 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.get(java.lang.String) @bci=35, 
 line=712 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getTrimmed(java.lang.String) @bci=2, 
 line=731 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getBoolean(java.lang.String, boolean) 
 @bci=2, line=1047 (Interpreted frame)
  - org.apache.hadoop.mapred.IFileInputStream.init(java.io.InputStream, 
 long, org.apache.hadoop.conf.Configuration) @bci=111, line=93 (Interpreted 
 frame)
  - 
 org.apache.hadoop.mapred.IFile$Reader.init(org.apache.hadoop.conf.Configuration,
  org.apache.hadoop.fs.FSDataInputStream, long, 
 org.apache.hadoop.io.compress.CompressionCodec, 
 org.apache.hadoop.mapred.Counters$Counter) @bci=60, line=303 (Interpreted 
 frame)
  - 
 org.apache.hadoop.mapred.IFile$InMemoryReader.init(org.apache.hadoop.mapred.RamManager,
  

[jira] [Updated] (MAPREDUCE-5450) Unnecessary Configuration instantiation in IFileInputStream slows down merge - Port to branch-1

2013-08-05 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5450?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5450:
-

Component/s: (was: mrv2)

 Unnecessary Configuration instantiation in IFileInputStream slows down merge 
 - Port to branch-1
 ---

 Key: MAPREDUCE-5450
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5450
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv1
Affects Versions: 1.1.0, 2.0.2-alpha
Reporter: Stanislav Barton
Assignee: Stanislav Barton
Priority: Blocker
 Fix For: 2.1.0-beta


 We are using hadoop-2.0.0+1357-1.cdh4.3.0.p0.21 with MRv1. After upgrade from 
 4.1.2 to 4.3.0, I have noticed some performance deterioration in our MR job 
 in the Reduce phase. The MR job has usually 10 000 map tasks (10 000 files on 
 input each about 100MB) and 6 000 reducers (one reducer per table region). I 
 was trying to figure out what at which phase the slow down appears (firstly I 
 suspected that the slow gathering of the 1 map output files is the 
 culprit) and found out that the problem is not reading the map output (the 
 shuffle) but the sort/merge phase that follows - the last and actual reduce 
 phase is fast. I have tried to up the io.sort.factor because I thought the 
 lots of small files are being merged on disk, but again upping that to 1000 
 didnt do any difference. I have then printed the stack trace and found out 
 that the problem is initialization of the 
 org.apache.hadoop.mapred.IFileInputStream namely the creation of the 
 Configuration object which is not propagated along from earlier context, see 
 the stack trace:
 Thread 13332: (state = IN_NATIVE)
  - java.io.UnixFileSystem.getBooleanAttributes0(java.io.File) @bci=0 
 (Compiled frame; information may be imprecise)
  - java.io.UnixFileSystem.getBooleanAttributes(java.io.File) @bci=2, line=228 
 (Compiled frame)
  - java.io.File.exists() @bci=20, line=733 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.getResource(java.lang.String, boolean) 
 @bci=136, line=999 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.findResource(java.lang.String, boolean) 
 @bci=3, line=966 (Compiled frame)
  - sun.misc.URLClassPath.findResource(java.lang.String, boolean) @bci=17, 
 line=146 (Compiled frame)
  - java.net.URLClassLoader$2.run() @bci=12, line=385 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction, 
 java.security.AccessControlContext) @bci=0 (Compiled frame)
  - java.net.URLClassLoader.findResource(java.lang.String) @bci=13, line=382 
 (Compiled frame)
  - java.lang.ClassLoader.getResource(java.lang.String) @bci=30, line=1002 
 (Compiled frame)
  - java.lang.ClassLoader.getResourceAsStream(java.lang.String) @bci=2, 
 line=1192 (Compiled frame)
  - javax.xml.parsers.SecuritySupport$4.run() @bci=26, line=96 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction) 
 @bci=0 (Compiled frame)
  - 
 javax.xml.parsers.SecuritySupport.getResourceAsStream(java.lang.ClassLoader, 
 java.lang.String) @bci=10, line=89 (Compiled frame)
  - javax.xml.parsers.FactoryFinder.findJarServiceProvider(java.lang.String) 
 @bci=38, line=250 (Interpreted frame)
  - javax.xml.parsers.FactoryFinder.find(java.lang.String, java.lang.String) 
 @bci=273, line=223 (Interpreted frame)
  - javax.xml.parsers.DocumentBuilderFactory.newInstance() @bci=4, line=123 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResource(java.util.Properties, 
 org.apache.hadoop.conf.Configuration$Resource, boolean) @bci=16, line=1890 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResources(java.util.Properties, 
 java.util.ArrayList, boolean) @bci=49, line=1867 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getProps() @bci=43, line=1785 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.get(java.lang.String) @bci=35, 
 line=712 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getTrimmed(java.lang.String) @bci=2, 
 line=731 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getBoolean(java.lang.String, boolean) 
 @bci=2, line=1047 (Interpreted frame)
  - org.apache.hadoop.mapred.IFileInputStream.init(java.io.InputStream, 
 long, org.apache.hadoop.conf.Configuration) @bci=111, line=93 (Interpreted 
 frame)
  - 
 org.apache.hadoop.mapred.IFile$Reader.init(org.apache.hadoop.conf.Configuration,
  org.apache.hadoop.fs.FSDataInputStream, long, 
 org.apache.hadoop.io.compress.CompressionCodec, 
 org.apache.hadoop.mapred.Counters$Counter) @bci=60, line=303 (Interpreted 
 frame)
  - 
 org.apache.hadoop.mapred.IFile$InMemoryReader.init(org.apache.hadoop.mapred.RamManager,
  

[jira] [Updated] (MAPREDUCE-5450) Unnecessary Configuration instantiation in IFileInputStream slows down merge - Port to branch-1

2013-08-05 Thread Arun C Murthy (JIRA)

 [ 
https://issues.apache.org/jira/browse/MAPREDUCE-5450?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arun C Murthy updated MAPREDUCE-5450:
-

Affects Version/s: (was: 2.0.2-alpha)

 Unnecessary Configuration instantiation in IFileInputStream slows down merge 
 - Port to branch-1
 ---

 Key: MAPREDUCE-5450
 URL: https://issues.apache.org/jira/browse/MAPREDUCE-5450
 Project: Hadoop Map/Reduce
  Issue Type: Bug
  Components: mrv1
Affects Versions: 1.1.0
Reporter: Stanislav Barton
Assignee: Stanislav Barton
Priority: Blocker
 Fix For: 2.1.0-beta


 We are using hadoop-2.0.0+1357-1.cdh4.3.0.p0.21 with MRv1. After upgrade from 
 4.1.2 to 4.3.0, I have noticed some performance deterioration in our MR job 
 in the Reduce phase. The MR job has usually 10 000 map tasks (10 000 files on 
 input each about 100MB) and 6 000 reducers (one reducer per table region). I 
 was trying to figure out what at which phase the slow down appears (firstly I 
 suspected that the slow gathering of the 1 map output files is the 
 culprit) and found out that the problem is not reading the map output (the 
 shuffle) but the sort/merge phase that follows - the last and actual reduce 
 phase is fast. I have tried to up the io.sort.factor because I thought the 
 lots of small files are being merged on disk, but again upping that to 1000 
 didnt do any difference. I have then printed the stack trace and found out 
 that the problem is initialization of the 
 org.apache.hadoop.mapred.IFileInputStream namely the creation of the 
 Configuration object which is not propagated along from earlier context, see 
 the stack trace:
 Thread 13332: (state = IN_NATIVE)
  - java.io.UnixFileSystem.getBooleanAttributes0(java.io.File) @bci=0 
 (Compiled frame; information may be imprecise)
  - java.io.UnixFileSystem.getBooleanAttributes(java.io.File) @bci=2, line=228 
 (Compiled frame)
  - java.io.File.exists() @bci=20, line=733 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.getResource(java.lang.String, boolean) 
 @bci=136, line=999 (Compiled frame)
  - sun.misc.URLClassPath$FileLoader.findResource(java.lang.String, boolean) 
 @bci=3, line=966 (Compiled frame)
  - sun.misc.URLClassPath.findResource(java.lang.String, boolean) @bci=17, 
 line=146 (Compiled frame)
  - java.net.URLClassLoader$2.run() @bci=12, line=385 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction, 
 java.security.AccessControlContext) @bci=0 (Compiled frame)
  - java.net.URLClassLoader.findResource(java.lang.String) @bci=13, line=382 
 (Compiled frame)
  - java.lang.ClassLoader.getResource(java.lang.String) @bci=30, line=1002 
 (Compiled frame)
  - java.lang.ClassLoader.getResourceAsStream(java.lang.String) @bci=2, 
 line=1192 (Compiled frame)
  - javax.xml.parsers.SecuritySupport$4.run() @bci=26, line=96 (Compiled frame)
  - 
 java.security.AccessController.doPrivileged(java.security.PrivilegedAction) 
 @bci=0 (Compiled frame)
  - 
 javax.xml.parsers.SecuritySupport.getResourceAsStream(java.lang.ClassLoader, 
 java.lang.String) @bci=10, line=89 (Compiled frame)
  - javax.xml.parsers.FactoryFinder.findJarServiceProvider(java.lang.String) 
 @bci=38, line=250 (Interpreted frame)
  - javax.xml.parsers.FactoryFinder.find(java.lang.String, java.lang.String) 
 @bci=273, line=223 (Interpreted frame)
  - javax.xml.parsers.DocumentBuilderFactory.newInstance() @bci=4, line=123 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResource(java.util.Properties, 
 org.apache.hadoop.conf.Configuration$Resource, boolean) @bci=16, line=1890 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.loadResources(java.util.Properties, 
 java.util.ArrayList, boolean) @bci=49, line=1867 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getProps() @bci=43, line=1785 
 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.get(java.lang.String) @bci=35, 
 line=712 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getTrimmed(java.lang.String) @bci=2, 
 line=731 (Compiled frame)
  - org.apache.hadoop.conf.Configuration.getBoolean(java.lang.String, boolean) 
 @bci=2, line=1047 (Interpreted frame)
  - org.apache.hadoop.mapred.IFileInputStream.init(java.io.InputStream, 
 long, org.apache.hadoop.conf.Configuration) @bci=111, line=93 (Interpreted 
 frame)
  - 
 org.apache.hadoop.mapred.IFile$Reader.init(org.apache.hadoop.conf.Configuration,
  org.apache.hadoop.fs.FSDataInputStream, long, 
 org.apache.hadoop.io.compress.CompressionCodec, 
 org.apache.hadoop.mapred.Counters$Counter) @bci=60, line=303 (Interpreted 
 frame)
  - 
 org.apache.hadoop.mapred.IFile$InMemoryReader.init(org.apache.hadoop.mapred.RamManager,
  

  1   2   3   4   5   6   7   8   9   10   >