Thanks a lot Markus!

Where do we define this parameter, please?

Benjamin


On Tue, Jul 2, 2013 at 4:28 PM, Markus Jelsma <[email protected]>wrote:

> Hi,
>
> Increase your memory in the task trackers by setting your Xmx in
> mapred.map.child.java.opts.
>
> Cheers
>
>
>
> -----Original message-----
> > From:Sznajder ForMailingList <[email protected]>
> > Sent: Tuesday 2nd July 2013 15:25
> > To: [email protected]
> > Subject: Distributed mode and java/lang/OutOfMemoryError
> >
> > Hi,
> >
> > I am running  Nutch 1.7 on a cluster of 6 nodes.
> >
> > I tempted to launch the bin/crawl script in this configuration and I am
> > getting a very strange error (an error I did not get in the local mode):
> >
> > 13/07/02 16:04:23 INFO fetcher.Fetcher: Fetcher Timelimit set for :
> > 1372781063368
> > 13/07/02 16:04:25 INFO mapred.FileInputFormat: Total input paths to
> process
> > : 1
> > 13/07/02 16:04:27 INFO mapred.JobClient: Running job:
> job_201207110955_1173
> > 13/07/02 16:04:28 INFO mapred.JobClient:  map 0% reduce 0%
> > 13/07/02 16:04:46 INFO mapred.JobClient:  map 100% reduce 0%
> > 13/07/02 16:05:01 INFO mapred.JobClient: Task Id :
> > attempt_201207110955_1173_r_000000_0, Status : FAILED
> > Error: null
> > attempt_201207110955_1173_r_000000_0: JVMDUMP006I Processing dump event
> > "systhrow", detail "java/lang/OutOfMemoryError" - please wait.
> > attempt_201207110955_1173_r_000000_0: JVMDUMP032I JVM requested Heap dump
> > using
> >
> '/local/hadoop/tmp/mapred/local/taskTracker/jobcache/job_201207110955_1173/attempt_201207110955_1173_r_000000_0/work/heapdump.20130702.160821.24995.0001.phd'
> > in response to an event
> >
> >
> > Do you have an idea how to solve that?
> >
> > Benjamin
> >
>

Reply via email to