No, this problem was supposed to be already fixed in nutch-711 in 1.0-dev.  I
tried it in 1.0, and somehow it broke again.  Maybe some changes after 711
broke it again?  Does anyone know how to fix this?  Thanks.


andy2005cst wrote:
> 
> I met the same problem as you, have you find a way to solve it?
> 
> 
> dealmaker wrote:
>> 
>> I am using the nutch nightly build #741 (Mar 3, 2009 4:01:53 AM).  I am
>> at the final phrase of crawling following the tutorial on Nutch.org
>> website.  I ran the following command, and I got exception in Hadoop.  I
>> double checked the folder path in nutch-site.xml, and they are correct. 
>> I tried multiple times, and I got same problem.  I didn't have same
>> problem in 0.9.   What's wrong?
>> 
>> $ bin/nutch index crawl/indexes crawl/crawldb crawl/linkdb
>> crawl/segments/*
>> Indexer: starting
>> Indexer: java.io.IOException: Job failed!
>>      at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1232)
>>      at org.apache.nutch.indexer.Indexer.index(Indexer.java:72)
>>      at org.apache.nutch.indexer.Indexer.run(Indexer.java:92)
>>      at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>>      at org.apache.nutch.indexer.Indexer.main(Indexer.java:101)
>> 
>> 
>> Log from Hadoop:
>> 2009-03-04 14:30:31,531 WARN  mapred.LocalJobRunner - job_local_0001
>> java.lang.IllegalArgumentException: it doesn't make sense to have a field
>> that is neither indexed nor stored
>>      at org.apache.lucene.document.Field.<init>(Field.java:279)
>>      at
>> org.apache.nutch.indexer.lucene.LuceneWriter.createLuceneDoc(LuceneWriter.java:133)
>>      at
>> org.apache.nutch.indexer.lucene.LuceneWriter.write(LuceneWriter.java:239)
>>      at
>> org.apache.nutch.indexer.IndexerOutputFormat$1.write(IndexerOutputFormat.java:50)
>>      at
>> org.apache.nutch.indexer.IndexerOutputFormat$1.write(IndexerOutputFormat.java:40)
>>      at org.apache.hadoop.mapred.ReduceTask$3.collect(ReduceTask.java:410)
>>      at
>> org.apache.nutch.indexer.IndexerMapReduce.reduce(IndexerMapReduce.java:158)
>>      at
>> org.apache.nutch.indexer.IndexerMapReduce.reduce(IndexerMapReduce.java:50)
>>      at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:436)
>>      at
>> org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:170)
>> 2009-03-04 14:30:31,668 FATAL indexer.Indexer - Indexer:
>> java.io.IOException: Job failed!
>>      at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1232)
>>      at org.apache.nutch.indexer.Indexer.index(Indexer.java:72)
>>      at org.apache.nutch.indexer.Indexer.run(Indexer.java:92)
>>      at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
>>      at org.apache.nutch.indexer.Indexer.main(Indexer.java:101)
>> 
>> 
> 
> 

-- 
View this message in context: 
http://www.nabble.com/Hadoop--java.io.IOException%3A-Job-failed%21-at-org.apache.hadoop.mapred.JobClient.runJob%28JobClient.java%3A1232%29-while-indexing.-tp22341554p22983483.html
Sent from the Nutch - User mailing list archive at Nabble.com.

Reply via email to