I'm assuming that if your running on an established hadoop cluster you will
wish to keep it over there.
Lewis

On Tuesday, July 2, 2013, Sznajder ForMailingList <[email protected]>
wrote:
> Thanks
> Can I copy this file to my $NUTCH/conf directory, or must I keep it in the
> $HADOOP/conf directory?
>
> Benjamin
>
>
> On Tue, Jul 2, 2013 at 5:10 PM, Lewis John Mcgibbney <
> [email protected]> wrote:
>
>> in mapred-site.xml
>> It is your Mapreduce configuration override.
>> hth
>>
>> On Tuesday, July 2, 2013, Sznajder ForMailingList <
>> [email protected]>
>> wrote:
>> > Thanks a lot Markus!
>> >
>> > Where do we define this parameter, please?
>> >
>> > Benjamin
>> >
>> >
>> > On Tue, Jul 2, 2013 at 4:28 PM, Markus Jelsma <
>> [email protected]
>> >wrote:
>> >
>> >> Hi,
>> >>
>> >> Increase your memory in the task trackers by setting your Xmx in
>> >> mapred.map.child.java.opts.
>> >>
>> >> Cheers
>> >>
>> >>
>> >>
>> >> -----Original message-----
>> >> > From:Sznajder ForMailingList <[email protected]>
>> >> > Sent: Tuesday 2nd July 2013 15:25
>> >> > To: [email protected]
>> >> > Subject: Distributed mode and java/lang/OutOfMemoryError
>> >> >
>> >> > Hi,
>> >> >
>> >> > I am running  Nutch 1.7 on a cluster of 6 nodes.
>> >> >
>> >> > I tempted to launch the bin/crawl script in this configuration and I
>> am
>> >> > getting a very strange error (an error I did not get in the local
>> mode):
>> >> >
>> >> > 13/07/02 16:04:23 INFO fetcher.Fetcher: Fetcher Timelimit set for :
>> >> > 1372781063368
>> >> > 13/07/02 16:04:25 INFO mapred.FileInputFormat: Total input paths to
>> >> process
>> >> > : 1
>> >> > 13/07/02 16:04:27 INFO mapred.JobClient: Running job:
>> >> job_201207110955_1173
>> >> > 13/07/02 16:04:28 INFO mapred.JobClient:  map 0% reduce 0%
>> >> > 13/07/02 16:04:46 INFO mapred.JobClient:  map 100% reduce 0%
>> >> > 13/07/02 16:05:01 INFO mapred.JobClient: Task Id :
>> >> > attempt_201207110955_1173_r_000000_0, Status : FAILED
>> >> > Error: null
>> >> > attempt_201207110955_1173_r_000000_0: JVMDUMP006I Processing dump
>> event
>> >> > "systhrow", detail "java/lang/OutOfMemoryError" - please wait.
>> >> > attempt_201207110955_1173_r_000000_0: JVMDUMP032I JVM requested Heap
>> dump
>> >> > using
>> >> >
>> >>
>>
>>
'/local/hadoop/tmp/mapred/local/taskTracker/jobcache/job_201207110955_1173/attempt_201207110955_1173_r_000000_0/work/heapdump.20130702.160821.24995.0001.phd'
>> >> > in response to an event
>> >> >
>> >> >
>> >> > Do you have an idea how to solve that?
>> >> >
>> >> > Benjamin
>> >> >
>> >>
>> >
>>
>> --
>> *Lewis*
>>
>

-- 
*Lewis*

Reply via email to