Any suggestions as to how to track down the root cause of these errors?

1178709 [main] INFO org.apache.hadoop.mapred.JobClient  -  map 6% reduce 0%
1178709 [main] INFO org.apache.hadoop.mapred.JobClient  -  map 6% reduce 0%
11/11/15 00:45:29 INFO mapred.JobClient: Task Id :
attempt_201111150008_0002_r_000000_0, Status : FAILED
1208771 [main] INFO org.apache.hadoop.mapred.JobClient  - Task Id :
attempt_201111150008_0002_r_000000_0, Status : FAILED
1208771 [main] INFO org.apache.hadoop.mapred.JobClient  - Task Id :
attempt_201111150008_0002_r_000000_0, Status : FAILED
Error: java.lang.OutOfMemoryError: Java heap space
    at org.apache.hadoop.mapred.IFile$Reader.readNextBlock(IFile.java:342)
    at org.apache.hadoop.mapred.IFile$Reader.next(IFile.java:404)
    at org.apache.hadoop.mapred.Merger$Segment.next(Merger.java:220)
    at org.apache.hadoop.mapred.Merger$MergeQueue.merge(Merger.java:420)
    at org.apache.hadoop.mapred.Merger$MergeQueue.merge(Merger.java:381)
    at org.apache.hadoop.mapred.Merger.merge(Merger.java:60)
    at 
org.apache.hadoop.mapred.ReduceTask$ReduceCopier$LocalFSMerger.run(ReduceTas
k.java:2651)


On 11/13/11 6:34 PM, "Eric Fiala" <e...@fiala.ca> wrote:

> Hoot, these are big numbers - some thoughts
> 1) does your machine have 1000GB to spare for each java child thread (each
> mapper + each reducer)?  mapred.child.java.opts / -Xmx1048576m
> 2) does each of your daemons need / have 10G? HADOOP_HEAPSIZE=10000
> 
> hth
> EF
>>>>> # The maximum amount of heap to use, in MB. Default is 1000.
>>>>>  export HADOOP_HEAPSIZE=10000
>>>>> <name>mapred.child.java.opts</name>
>>>>> <value>-Xmx1048576m</value>
>>>>> </property>
>>>>> 
> 

Reply via email to