The JDK also provides "jmap -histo PID" which will give you some crude information about where the memory is going.
-Michael On 12/6/07 2:16 PM, "Ted Dunning" <[EMAIL PROTECTED]> wrote: There is a bug in the GZipInputStream on java 1.5 that can cause an out-of-memory error on a malformed gzip input. It is possible that you are trying to treat this input as a splittable file which is causing your maps to be fed from chunks of the gzip file. Those chunks would be ill-formed, of course, and it is possible that this is causing an out-of-memory condition. I am just speculating, however. To confirm or discard this possibility, you should examine the stack traces for the maps that are falling over. On 12/6/07 2:05 PM, "Rui Shi" <[EMAIL PROTECTED]> wrote: > > Hi, > > It is hard to believe that you need to enlarge heap size given the input size > is only 10MB. In particular, you don't load all input at the same time. As for > the program logic, no much fancy stuff, mostly cut and sorting. So GC should > be able to handle... > > Thanks, > > Rui > > > ----- Original Message ---- > From: Joydeep Sen Sarma <[EMAIL PROTECTED]> > To: hadoop-user@lucene.apache.org > Sent: Thursday, December 6, 2007 1:14:51 PM > Subject: RE: Mapper Out of Memory > > > Can control heap size using 'mapred.child.java.opts' option. > > Check ur program logic though. Personal experience is that running out > of heap space in map task usually suggests some runaway logic > somewhere. > > -----Original Message----- > From: Rui Shi [mailto:[EMAIL PROTECTED] > Sent: Thursday, December 06, 2007 12:31 PM > To: hadoop-user@lucene.apache.org > Subject: Mapper Out of Memory > > > Hi, > > I run hadoop on a BSD4 clusters and each map task is a gzip file (about > 10MB). Some tasks finished. But many of them failed due to heap out of > memory. I got the following syslogs: > > 2007-12-06 12:16:50,277 INFO org.apache.hadoop.metrics.jvm.JvmMetrics: > Initializing JVM Metrics with processName=MAP, sessionId= > 2007-12-06 12:16:53,128 INFO org.apache.hadoop.mapred.MapTask: > numReduceTasks: 256 > 2007-12-06 12:16:53,638 WARN org.apache.hadoop.util.NativeCodeLoader: > Unable to load native-hadoop library for your platform... using > builtin-java classes where applicable > 2007-12-06 12:18:19,079 WARN org.apache.hadoop.mapred.TaskTracker: > Error > running child > java.lang.OutOfMemoryError: Java heap space > Does anyone know what is the reason and how should we avoid it? > > Thanks, > > Rui > > > > > > > ________________________________________________________________________ > ____________ > Never miss a thing. Make Yahoo your home page. > http://www.yahoo.com/r/hs > > > > > > > > ______________________________________________________________________________ > ______ > Looking for last minute shopping deals? > Find them fast with Yahoo! Search. > http://tools.search.yahoo.com/newsearch/category.php?category=shopping