Java version: JSDK 1.4.2_08 URL Seed: http://www.math.psu.edu/MathLists/Contents.html
I even tried allocating more stack memory using "-Xss", process memory "-Xms" option. However, if I run the individual tools (fetchlisttool, fetcher, updatedb..etc) separately from the shell, it works fine. Thanks, --Rajesh On 4/6/06, Piotr Kosiorowski <[EMAIL PROTECTED]> wrote: > > Which Java version do you use? > Is it the same for all urls or only for specific one? > If URL you are trying to crawl is public you can send it to me (off list > if you wish) and I can check it on my machine. > Regards > Piotr > > Rajesh Munavalli wrote: > > I had earlier posted this message to the list but havent got any > response. > > Here are more details. > > > > Nutch versionI: nutch.0.7.2 > > URL File: contains a single URL. File name: "urls" > > Crawl-url-filter: is set to grab all URLs > > > > Command: bin/nutch crawl urls -dir crawl.test -depth 3 > > Error: java.lang.StackOverflowError > > > > The error occurrs while it executes the "UpdateDatabaseTool". > > > > One solution I can think of is to provide more stack memory. But is > there a > > better solution to this? > > > > Thanks, > > > > Rajesh > > > > > >
