Hello,

One can set generate.max.count to limit number of urls for domain or host
in next fetch step. But is there a way to limit number of fetched urls for
domain/host in a whole crawl process? Supposing I run generate/fetch/update
cycle 6 times and want to limit number of urls per host to 100 urls (pages)
and not more in a whole crawldb. How can I achieve that?

Regards,
Tomasz

Reply via email to