Don't know off the top of my head, but I'd guess no, because Nutch uses Hadoop/HDFS. HDFS files are write-once, so I doubt you can just update a single URL's data. But you could write a MapReduce job that goes over the whole CrawlDb and modifies only the records you need modified. You'll need to essentially rewrite the whole CrawlDb and replace the old version.
It would be nice to be able to change specific URL(s) data... Otis -- Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch ----- Original Message ---- > From: Chris Kline <[EMAIL PROTECTED]> > To: [email protected] > Sent: Tuesday, June 17, 2008 6:19:51 PM > Subject: updating retry inteval > > is there a way to update the retry interval for a specific url? > > -Chris
