Hi, it's better to use bin/crawl instead of the deprecated class o.a.n.crawl.Crawl which will be removed in 1.8.
To retry to index already crawled data you can also use: bin/nutch index resp. bin/nutch dedup bin/nutch clean to remove duplicated and gone documents. Sebastian On 02/26/2014 01:17 PM, Yicheng Ye wrote: > Hi, > > I tried to run Nutch 1.7 on Hadoop 1.2.1. It works fine when I use > > bin/hadoop jar > /home/yyc/Desktop/apache-nutch-1.7/runtime/deploy/apache-nutch-1.7.job > org.apache.nutch.crawl.Crawl urls -dir crawl -depth 3 -topN 5 > > But when I tried to index it using solr, I got the following error msg: > > <http://lucene.472066.n3.nabble.com/file/n4119756/QQ%E6%88%AA%E5%9B%BE20140226201459.jpg> > > > Can anybody tell me how to resolve this problem? > > Thanks in advance! > > Yicheng > > > > -- > View this message in context: > http://lucene.472066.n3.nabble.com/org-apache-solr-client-solrj-SolrServerException-on-Nutch-1-7-with-Hadoop-1-2-1-tp4119756.html > Sent from the Nutch - User mailing list archive at Nabble.com. >

