[ http://issues.apache.org/jira/browse/NUTCH-350?page=all ]

Stefan Groschupf updated NUTCH-350:
-----------------------------------

    Attachment: protocolRetryV5.patch

This patch will dramatically increase the number of successfully fetched pages 
of a intranet crawl over the time. 

> urls blocked db.fetch.retry.max * http.max.delays times during fetching are 
> marked as STATUS_DB_GONE
> ----------------------------------------------------------------------------------------------------
>
>                 Key: NUTCH-350
>                 URL: http://issues.apache.org/jira/browse/NUTCH-350
>             Project: Nutch
>          Issue Type: Bug
>            Reporter: Stefan Groschupf
>            Priority: Critical
>         Attachments: protocolRetryV5.patch
>
>
> Intranet crawls or focused crawls will fetch many pages from the same host. 
> This causes that a thread will be blocked since a other thread already fetch 
> from the same host. It is very likely that threads are more often blocked 
> than http.max.delays. In such a case the HttpBase.blockAddr method throws a 
> HttpException. This will be handled in the fetcher by increment the 
> crawlDatum retries and set the status to STATUS_FETCH_RETRY. That means that 
> at least you have only db.fetch.retry.max * http.max.delays chances to fetch 
> a url. But in intranet or focused crawls it is very likely that this is not 
> enough. Increaing one of the involved properties dramatically slow down the 
> fetch. 
> I suggest to not increase the CrawlDatum RetriesSinceFetch in case the 
> problem was caused by a blocked thread.

-- 
This message is automatically generated by JIRA.
-
If you think it was sent incorrectly contact one of the administrators: 
http://issues.apache.org/jira/secure/Administrators.jspa
-
For more information on JIRA, see: http://www.atlassian.com/software/jira

        

Reply via email to