Thanks Markus, Will Try This .

Thanks
Manish Verma


> On Jan 15, 2016, at 5:28 AM, Markus Jelsma <[email protected]> wrote:
> 
> Hello - you mean transient errors such as temporary connection issues, time 
> outs etc? That is not possible, but they will be retried one day later. One 
> option is to dump the failed URL's via readdb and using the -retry switch. 
> The output contains failed URL's. Parse the dump, get the URL's and use the 
> freegen tool to force recrawl. You can also patch the generator tool to 
> restrict to records with a non-zero retry count.
> 
> Markus
> 
> -----Original message-----
>> From:Manish Verma <[email protected]>
>> Sent: Thursday 14th January 2016 23:27
>> To: [email protected]
>> Subject: Need To Crawl Only Failed URLS
>> 
>> Hi,
>> 
>> I want to crawl just failed url, I picked failed URL’s from log and want to 
>> crawl just these. 
>> I believe if I put these as seed and run crawl script , it will generate 
>> fetch list containing not only these failed urls rather all present in 
>> crawlDB for next fetch cycle ?
>> 
>> Any Thoughts .
>> 
>> Thanks Manish
>> 
>> 

Reply via email to