Hi,

it's better to use
 bin/crawl
instead of the deprecated class o.a.n.crawl.Crawl which
will be removed in 1.8.

To retry to index already crawled data you can also use:
 bin/nutch index
resp.
 bin/nutch dedup
 bin/nutch clean
to remove duplicated and gone documents.

Sebastian

On 02/26/2014 01:17 PM, Yicheng Ye wrote:
> Hi,
> 
> I tried to run Nutch 1.7 on Hadoop 1.2.1. It works fine when I use 
> 
> bin/hadoop jar
> /home/yyc/Desktop/apache-nutch-1.7/runtime/deploy/apache-nutch-1.7.job
> org.apache.nutch.crawl.Crawl urls -dir crawl -depth 3 -topN 5
> 
> But when I tried to index it using solr, I got the following error msg:
> 
> <http://lucene.472066.n3.nabble.com/file/n4119756/QQ%E6%88%AA%E5%9B%BE20140226201459.jpg>
>  
> 
> Can anybody tell me how to resolve this problem? 
> 
> Thanks in advance!
> 
> Yicheng
> 
> 
> 
> --
> View this message in context: 
> http://lucene.472066.n3.nabble.com/org-apache-solr-client-solrj-SolrServerException-on-Nutch-1-7-with-Hadoop-1-2-1-tp4119756.html
> Sent from the Nutch - User mailing list archive at Nabble.com.
> 

Reply via email to