Thanks for your reply. I'm wondering, is it possible to skip this
dedup phase then, or to not acquire a lock? The reason I'd like to
use 0.14 code is that I've instrumented it to add some tracing and
I'd like to collect traces of how Nutch uses Hadoop. It may be
possible to port the changes back to 0.12 but I'd prefer not to
because I may have other apps that use things in 0.14 and because I
want to trace the best-performing Hadoop version possible.
Matei
On Oct 18, 2007, at 12:58 AM, Nguyen Manh Tien wrote:
You should you hadoop 0.12.3 for example to dedup. The current version
0.14.x don't support Lock operation.
2007/10/18, Matei Zaharia <[EMAIL PROTECTED]>:
Hi,
I'm sometimes getting the following error in the dedup 3 job when
running Nutch 0.9 on top of Hadoop 0.14.2:
java.io.IOException: Lock obtain timed out: [EMAIL PROTECTED]://r37:54310/
user/matei/crawl4/indexes/part-00000/write.lock
at org.apache.lucene.store.Lock.obtain(Lock.java:69)
at org.apache.lucene.index.IndexReader.aquireWriteLock
(IndexReader.java:526)
at org.apache.lucene.index.IndexReader.deleteDocument
(IndexReader.java:551)
at org.apache.nutch.indexer.DeleteDuplicates.reduce
(DeleteDuplicates.java:378)
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:
322)
at org.apache.hadoop.mapred.TaskTracker$Child.main(
TaskTracker.java:
1782)
Other times, it works just fine. Do you know why this is happening?
Thanks,
Matei Zaharia