Don't know off the top of my head, but I'd guess no, because Nutch uses 
Hadoop/HDFS.  HDFS files are write-once, so I doubt you can just update a 
single URL's data.  But you could write a MapReduce job that goes over the 
whole CrawlDb and modifies only the records you need modified.  You'll need to 
essentially rewrite the whole CrawlDb and replace the old version.

It would be nice to be able to change specific URL(s) data...


Otis 
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch


----- Original Message ----
> From: Chris Kline <[EMAIL PROTECTED]>
> To: [email protected]
> Sent: Tuesday, June 17, 2008 6:19:51 PM
> Subject: updating retry inteval
> 
> is there a way to update the retry interval for a specific url?
> 
> -Chris

Reply via email to