Neither. You leave it in $NUTCH/conf and compile a job file with 'ant job'
which gets used from runtime/deploy/bin

BTW new users should at least do the basic Hadoop tutorial.


On 2 July 2013 16:23, Lewis John Mcgibbney <[email protected]>wrote:

> I'm assuming that if your running on an established hadoop cluster you will
> wish to keep it over there.
> Lewis
>
> On Tuesday, July 2, 2013, Sznajder ForMailingList <
> [email protected]>
> wrote:
> > Thanks
> > Can I copy this file to my $NUTCH/conf directory, or must I keep it in
> the
> > $HADOOP/conf directory?
> >
> > Benjamin
> >
> >
> > On Tue, Jul 2, 2013 at 5:10 PM, Lewis John Mcgibbney <
> > [email protected]> wrote:
> >
> >> in mapred-site.xml
> >> It is your Mapreduce configuration override.
> >> hth
> >>
> >> On Tuesday, July 2, 2013, Sznajder ForMailingList <
> >> [email protected]>
> >> wrote:
> >> > Thanks a lot Markus!
> >> >
> >> > Where do we define this parameter, please?
> >> >
> >> > Benjamin
> >> >
> >> >
> >> > On Tue, Jul 2, 2013 at 4:28 PM, Markus Jelsma <
> >> [email protected]
> >> >wrote:
> >> >
> >> >> Hi,
> >> >>
> >> >> Increase your memory in the task trackers by setting your Xmx in
> >> >> mapred.map.child.java.opts.
> >> >>
> >> >> Cheers
> >> >>
> >> >>
> >> >>
> >> >> -----Original message-----
> >> >> > From:Sznajder ForMailingList <[email protected]>
> >> >> > Sent: Tuesday 2nd July 2013 15:25
> >> >> > To: [email protected]
> >> >> > Subject: Distributed mode and java/lang/OutOfMemoryError
> >> >> >
> >> >> > Hi,
> >> >> >
> >> >> > I am running  Nutch 1.7 on a cluster of 6 nodes.
> >> >> >
> >> >> > I tempted to launch the bin/crawl script in this configuration and
> I
> >> am
> >> >> > getting a very strange error (an error I did not get in the local
> >> mode):
> >> >> >
> >> >> > 13/07/02 16:04:23 INFO fetcher.Fetcher: Fetcher Timelimit set for :
> >> >> > 1372781063368
> >> >> > 13/07/02 16:04:25 INFO mapred.FileInputFormat: Total input paths to
> >> >> process
> >> >> > : 1
> >> >> > 13/07/02 16:04:27 INFO mapred.JobClient: Running job:
> >> >> job_201207110955_1173
> >> >> > 13/07/02 16:04:28 INFO mapred.JobClient:  map 0% reduce 0%
> >> >> > 13/07/02 16:04:46 INFO mapred.JobClient:  map 100% reduce 0%
> >> >> > 13/07/02 16:05:01 INFO mapred.JobClient: Task Id :
> >> >> > attempt_201207110955_1173_r_000000_0, Status : FAILED
> >> >> > Error: null
> >> >> > attempt_201207110955_1173_r_000000_0: JVMDUMP006I Processing dump
> >> event
> >> >> > "systhrow", detail "java/lang/OutOfMemoryError" - please wait.
> >> >> > attempt_201207110955_1173_r_000000_0: JVMDUMP032I JVM requested
> Heap
> >> dump
> >> >> > using
> >> >> >
> >> >>
> >>
> >>
>
> '/local/hadoop/tmp/mapred/local/taskTracker/jobcache/job_201207110955_1173/attempt_201207110955_1173_r_000000_0/work/heapdump.20130702.160821.24995.0001.phd'
> >> >> > in response to an event
> >> >> >
> >> >> >
> >> >> > Do you have an idea how to solve that?
> >> >> >
> >> >> > Benjamin
> >> >> >
> >> >>
> >> >
> >>
> >> --
> >> *Lewis*
> >>
> >
>
> --
> *Lewis*
>



-- 
*
*Open Source Solutions for Text Engineering

http://digitalpebble.blogspot.com/
http://www.digitalpebble.com
http://twitter.com/digitalpebble

Reply via email to