Hi, I will get the information and will reply to you.
Sent from my mobile On 06/11/2010, at 11:16, "Hari Sreekumar" <[email protected]> wrote: > It's an out of mem error, so I feel it has to do with ram rather > tham disk > space. Did you check if it's swapping? (top/htop)... Is your reduce > phase > very mem-intensive? Seems to be a memory leak somewhere.. What does > htop > say? What processes are you running on each node? What does the log > file > that it is showing say? > > On Sat, Nov 6, 2010 at 2:36 PM, Shavit Netzer <[email protected]> > wrote: > >> 7GB >> >> Sent from my mobile >> >> On 06/11/2010, at 11:00, "Hari Sreekumar" <[email protected] >> <mailto:[email protected]>> wrote: >> >> What's the RAM on each node? >> >> On Sat, Nov 6, 2010 at 11:03 AM, Shavit Netzer >> <[email protected]<mailto: >> [email protected]>> wrote: >> >> Hello, >> >> I have a question regarding MapRed jobs. >> >> I have 24 nodes, each node have 4 disks (mnt – mnt3), 500GB each m >> nt. >> >> All balanced ( I used the balancer, except mnt, which have 97% used). >> >> My question is: >> I got the following error and I relate it to the disk space (maybe >> I'm >> wrong). >> >> Maybe there is a configuration that I can add, change in order to >> have few >> more retries on separate disk: >> >> >> 10/10/27 21:59:01 INFO mapred.JobClient: map 100% reduce 26% >> >> 10/10/27 21:59:02 INFO mapred.JobClient: Task Id : >> attempt_201010201240_4059_r_000023_0, Status : FAILED >> >> java.io.IOException: Task process exit with nonzero status of 134. >> >> at >> org.apache.hadoop.mapred.TaskRunner.runChild(TaskRunner.java:462) >> >> at >> org.apache.hadoop.mapred.TaskRunner.run(TaskRunner.java:403) >> >> >> >> attempt_201010201240_4059_r_000023_0: # >> >> attempt_201010201240_4059_r_000023_0: # A fatal error has been >> detected by >> the Java Runtime Environment: >> >> attempt_201010201240_4059_r_000023_0: # >> >> attempt_201010201240_4059_r_000023_0: # java.lang.OutOfMemoryError: >> requested 32744 bytes for ChunkPool::allocate. Out of swap space? >> >> attempt_201010201240_4059_r_000023_0: # >> >> attempt_201010201240_4059_r_000023_0: # Internal Error >> (allocation.cpp:117), pid=15974, tid=1089702224 >> >> attempt_201010201240_4059_r_000023_0: # Error: ChunkPool::allocate >> >> attempt_201010201240_4059_r_000023_0: # >> >> attempt_201010201240_4059_r_000023_0: # JRE version: 6.0_14-b08 >> >> attempt_201010201240_4059_r_000023_0: # Java VM: Java HotSpot(TM) >> 64-Bit >> Server VM (14.0-b16 mixed mode linux-amd64 ) >> >> attempt_201010201240_4059_r_000023_0: # An error report file with >> more >> information is saved as: >> >> attempt_201010201240_4059_r_000023_0: # >> >> /mnt2/hadoop/mapred/local/taskTracker/jobcache/ >> job_201010201240_4059/attempt_201010201240_4059_r_000023_0/work/ >> hs_err_pid15974.log >> >> attempt_201010201240_4059_r_000023_0: # >> >> attempt_201010201240_4059_r_000023_0: # If you would like to submit >> a bug >> report, please visit: >> >> attempt_201010201240_4059_r_000023_0: # >> http://java.sun.com/webapps/bugreport/crash.jsp >> >> attempt_201010201240_4059_r_000023_0: # >> >> Regards, >> Shavit >> >>
