Hello - you mean transient errors such as temporary connection issues, time 
outs etc? That is not possible, but they will be retried one day later. One 
option is to dump the failed URL's via readdb and using the -retry switch. The 
output contains failed URL's. Parse the dump, get the URL's and use the freegen 
tool to force recrawl. You can also patch the generator tool to restrict to 
records with a non-zero retry count.

Markus
 
-----Original message-----
> From:Manish Verma <[email protected]>
> Sent: Thursday 14th January 2016 23:27
> To: [email protected]
> Subject: Need To Crawl Only Failed URLS
> 
> Hi,
> 
> I want to crawl just failed url, I picked failed URL’s from log and want to 
> crawl just these. 
> I believe if I put these as seed and run crawl script , it will generate 
> fetch list containing not only these failed urls rather all present in 
> crawlDB for next fetch cycle ?
> 
> Any Thoughts .
> 
> Thanks Manish
> 
> 

Reply via email to