Forgot to mention one more parameter. Modify the crawl-urlfilter to accept any URL.
On 4/6/06, Rajesh Munavalli <[EMAIL PROTECTED]> wrote: > > Java version: JSDK 1.4.2_08 > URL Seed: http://www.math.psu.edu/MathLists/Contents.html > > I even tried allocating more stack memory using "-Xss", process memory > "-Xms" option. However, if I run the individual tools (fetchlisttool, > fetcher, updatedb..etc) separately from the shell, it works fine. > > Thanks, > --Rajesh > > > > On 4/6/06, Piotr Kosiorowski <[EMAIL PROTECTED]> wrote: > > > > Which Java version do you use? > > Is it the same for all urls or only for specific one? > > If URL you are trying to crawl is public you can send it to me (off list > > > > if you wish) and I can check it on my machine. > > Regards > > Piotr > > > > Rajesh Munavalli wrote: > > > I had earlier posted this message to the list but havent got any > > response. > > > Here are more details. > > > > > > Nutch versionI: nutch.0.7.2 > > > URL File: contains a single URL. File name: "urls" > > > Crawl-url-filter: is set to grab all URLs > > > > > > Command: bin/nutch crawl urls -dir crawl.test -depth 3 > > > Error: java.lang.StackOverflowError > > > > > > The error occurrs while it executes the "UpdateDatabaseTool". > > > > > > One solution I can think of is to provide more stack memory. But is > > there a > > > better solution to this? > > > > > > Thanks, > > > > > > Rajesh > > > > > > > > > > > >
