That's an awesome explanation Matt... Thanks :)

----- Original Message ----- From: "Matthew Holt" <[EMAIL PROTECTED]>
To: <[email protected]>
Sent: Tuesday, July 11, 2006 1:51 PM
Subject: Re: Adddays confusion - easy question for the experts


Honda-Search Administrator wrote:
Reader's Digest version:
How can I ensure that nutch only crawls the urls I inject into the fetchlist and not recrawl the entire webdb?
Can anyone explain to me (in simple terms) exactly what adddays does?

Long version:
My setup is simple. I crawl a number of internet forums. This requires me to scan new posts every night to stay on top of things.

I crawled all of the older posts on these forums a while ago, and now have to just worry about newer posts. I have written a small script that injects the pages that have changed or the new pages each night.

When I run the recrawl script, I only want to crawl the pages that are injected into the fetchlist (via bin/nutch inject). I have also changed the default nutch recrawl time interval (normally 30 days) to a VERY large number to ensure that nutch will not recrawl old pages for a very long time.

Anyway, back to my original question.

i recrawled today hoping that nutch would ONLY recrawl the 3000 documents I injected (via bin/nutch inject). I used depth of 1 and left the adddays parameter blank (because I really can't get a clear idea of what it does). Depth of 1 is used because I only want to crawl the URLs I have injected into the fetchlist and not have nutch go crazy on other domains, documents, etc. Using the regex-urlfilter I have also ensured that it will only crawl the domains I want it to crawl.

So my command looks something like this:

/home/nutch/recrawl.sh /home/nutch/database 1

my recrawl script can be seen here: http://www.honda-search.com/script.html

Much to my surprised Nutch is recrawling EVERY document in my webdb (plus, I assume, the newly injected documents). Is this because the adddays variable is left blank? Should I set the addays variable really high? How can I ensure that it only crawls the urls that are injected?

Can anyone explain what adddays does (in easy to understand terms?) The wiki isn't very clear for a newbie like myself.

I was looking for similar info. The adddays option advances the clock however many days you specify. The default for page reindexing is 30 days, so every 30 days the page will expire and nutch will reindex it. However, if you pass the param -adddays 31, it will advance the clock 31 days and cause every page to be reindexed.

If you pass the param -adddays 27 and you have the default reindexing set to be 30 days, nutch will reindex all pages older than 3 days. Correct me if I'm wrong.
 Matt



Reply via email to