Greeting,
Here is the hadoop.log output from my crash - any ideas?
2007-07-31 19:06:50,702 INFO indexer.IndexingFilters - Adding
org.apache.nutch.indexer.basic.BasicIndexingFilter
2007-07-31 19:06:50,799 INFO indexer.Indexer - Optimizing index.
2007-07-31 19:06:51,497 INFO indexer.Indexer - Indexer: done
2007-07-31 19:06:51,498 INFO indexer.DeleteDuplicates - Dedup: starting
2007-07-31 19:06:51,510 INFO indexer.DeleteDuplicates - Dedup:
adding indexes in: /var/webindex/data/indexes
2007-07-31 19:06:51,733 WARN mapred.LocalJobRunner - job_2xsg2o
java.lang.ArrayIndexOutOfBoundsException: -1
at org.apache.lucene.index.MultiReader.isDeleted
(MultiReader.java:113)
at org.apache.nutch.indexer.DeleteDuplicates$InputFormat
$DDRecordReader.next(DeleteDuplicates.java:176)
at org.apache.hadoop.mapred.MapTask$1.next(MapTask.java:157)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:46)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:175)
at org.apache.hadoop.mapred.LocalJobRunner$Job.run
(LocalJobRunner.java:126)
On Jul 30, 2007, at 2:02 PM, DES wrote:
Look in logs/hadoop.log for the actual reason for this exception. The
console message is not really helpful.
On 7/30/07, Micah Vivion <[EMAIL PROTECTED]> wrote:
Exception in thread "main" java.io.IOException: Job failed!
at org.apache.hadoop.mapred.JobClient.runJob
(JobClient.java:
604)
at org.apache.nutch.indexer.DeleteDuplicates.dedup
(DeleteDuplicates.java:439)
at org.apache.nutch.crawl.Crawl.main(Crawl.java:135)