Java version: JSDK 1.4.2_08
URL Seed: http://www.math.psu.edu/MathLists/Contents.html

I even tried allocating more stack memory using "-Xss", process memory
"-Xms" option. However, if I run the individual tools (fetchlisttool,
fetcher, updatedb..etc) separately from the shell, it works fine.

Thanks,
--Rajesh



On 4/6/06, Piotr Kosiorowski <[EMAIL PROTECTED]> wrote:
>
> Which Java version do you use?
> Is it the same for all urls or only for specific one?
> If URL you are trying to crawl is public you can send it to me (off list
> if you wish) and I can check it on my machine.
> Regards
> Piotr
>
> Rajesh Munavalli wrote:
> > I had earlier posted this message to the list but havent got any
> response.
> > Here are more details.
> >
> > Nutch versionI: nutch.0.7.2
> > URL File: contains a single URL. File name: "urls"
> > Crawl-url-filter: is set to grab all URLs
> >
> > Command: bin/nutch crawl urls -dir crawl.test -depth 3
> > Error: java.lang.StackOverflowError
> >
> > The error occurrs while it executes the "UpdateDatabaseTool".
> >
> > One solution I can think of is to provide more stack memory. But is
> there a
> > better solution to this?
> >
> > Thanks,
> >
> > Rajesh
> >
>
>
>
>

Reply via email to