The crawl command creates a crawlDB for each call. So as Rchard mentioned try a higher depth. In case you like nutch to go deeper with each iteration, try the whole web tutorial but change the url filter in a manner that it only crawls your webpage.
This will go as deep as much iteration you run.

Stefan

In case you like to
Am Mar 4, 2006 um 9:09 PM schrieb Richard Braman:

Try using depth=n when you do the crawl. Post crawl I don't know, but I have the same question. How do you make the index go deeper when you do
your next roudn of fetching is still something I haven't figured out.

-----Original Message-----
From: Peter Swoboda [mailto:[EMAIL PROTECTED]
Sent: Friday, March 03, 2006 4:28 AM
To: [email protected]
Subject: how can i go deep?


Hi.
I've don a whole web crawl like it is shown in the tutorial. There is
just "http://www.kreuztal.de/"; in the urls.txt i did the Fetching three times. But unfortunately the crawl hasn't gone deep. while searching, i can only find keywords from the first(home-)site. for example i couldn't
find anythin on "http://www.kreuztal.de/impressum.php";
How can i configure the depht?
Thanx for helping.

greetings
Peter

--
Bis zu 70% Ihrer Onlinekosten sparen: GMX SmartSurfer! Kostenlos
downloaden: http://www.gmx.net/de/go/smartsurfer



---------------------------------------------
blog: http://www.find23.org
company: http://www.media-style.com




-------------------------------------------------------
This SF.Net email is sponsored by xPML, a groundbreaking scripting language
that extends applications into web and mobile media. Attend the live webcast
and join the prime developer group breaking into this new coding territory!
http://sel.as-us.falkag.net/sel?cmd=lnk&kid=110944&bid=241720&dat=121642
_______________________________________________
Nutch-general mailing list
[email protected]
https://lists.sourceforge.net/lists/listinfo/nutch-general

Reply via email to