#2 should be a pluggable/hookable parameter. "high-scoring" sounds like a reasonable default basis for choosing recrawl intervals, but I'm sure that nearly everyone will think of a way to improve upon that for their particular system.

e.g. "high-scoring" ain't gonna cut it for my needs. (0.5 wink ;)

--matt

On Dec 1, 2005, at 2:15 PM, Doug Cutting wrote:

It would be good to improve the support for incremental crawling added to Nutch. Here are some ideas about how we might implement it. Andrzej has posted in the past about this, so he probably has better ideas.

Incremental crawling could proceed as follows:

1. Bootstrap with a batch crawl, using the 'crawl' command. Modify CrawlDatum to store the MD5Hash of the content of fetched urls.

2. Reduce the fetch interval for high-scoring urls. If the default is monthly, then the top-scoring 1% of urls might be set to daily, and the top-scoring 10% of urls might be set to weekly.

3. Generate a fetch list & fetch it. When the url has been previously fetched, and its content is unchanged, increase its fetch interval by an amount, e.g., 50%. If the content is changed, decrease the fetch interval. The percentage of increase and decrease might be influenced by the url's score.

4. Update the crawl db & link db, index the new segment, dedup, etc. When updating the crawl db, scores for existing urls should not change, since the scoring method we're using (OPIC) assumes each page is fetched only once.

Steps 3 & 4 can be packaged as an 'update' command. Step 2 can be included in the 'crawl' command, so that crawled indexes are always ready for update.

Comments?

Doug

--
Matt Kangas / [EMAIL PROTECTED]


Reply via email to