Hi - there's some setting fetcher queue exceptions something that will dump a 
queue if it had too many exceptions.

 
-----Original message-----
> From:Adam Estrada <[email protected]>
> Sent: Thursday 17th July 2014 16:07
> To: [email protected]
> Subject: Ignoring errors in crawl
> 
> All,
> 
> I am coming across a few pages that are not responsive at all which is
> causing Nutch to #failwhale before finishing the current crawl. I have
> increased http.timeout and it still crashes. How can I get Nutch to
> skip over unresponsive URLs that are causing the entire thing to bail?
> 
> Thanks,
> Adam
> 

Reply via email to