[ http://issues.apache.org/jira/browse/NUTCH-350?page=all ]
Stefan Groschupf updated NUTCH-350:
-----------------------------------
Attachment: protocolRetryV5.patch
This patch will dramatically increase the number of successfully fetched pages
of a intranet crawl over the time.
> urls blocked db.fetch.retry.max * http.max.delays times during fetching are
> marked as STATUS_DB_GONE
> ----------------------------------------------------------------------------------------------------
>
> Key: NUTCH-350
> URL: http://issues.apache.org/jira/browse/NUTCH-350
> Project: Nutch
> Issue Type: Bug
> Reporter: Stefan Groschupf
> Priority: Critical
> Attachments: protocolRetryV5.patch
>
>
> Intranet crawls or focused crawls will fetch many pages from the same host.
> This causes that a thread will be blocked since a other thread already fetch
> from the same host. It is very likely that threads are more often blocked
> than http.max.delays. In such a case the HttpBase.blockAddr method throws a
> HttpException. This will be handled in the fetcher by increment the
> crawlDatum retries and set the status to STATUS_FETCH_RETRY. That means that
> at least you have only db.fetch.retry.max * http.max.delays chances to fetch
> a url. But in intranet or focused crawls it is very likely that this is not
> enough. Increaing one of the involved properties dramatically slow down the
> fetch.
> I suggest to not increase the CrawlDatum RetriesSinceFetch in case the
> problem was caused by a blocked thread.
--
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators:
http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see: http://www.atlassian.com/software/jira