Hello, One can set generate.max.count to limit number of urls for domain or host in next fetch step. But is there a way to limit number of fetched urls for domain/host in a whole crawl process? Supposing I run generate/fetch/update cycle 6 times and want to limit number of urls per host to 100 urls (pages) and not more in a whole crawldb. How can I achieve that?
Regards, Tomasz

