Hai, 

bin/nutch crawl urls -dir crawl_NEW1 -depth 3 -topN 50 

I have used the above command to crawl. 

I am getting the following error. 

Dedup: adding indexes in: crawl_NEW1/indexes 
Exception in thread "main" java.io.IOException: Job failed! 
        at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:604) 
        at
org.apache.nutch.indexer.DeleteDuplicates.dedup(DeleteDuplicates.java 
:439) 
        at org.apache.nutch.crawl.Crawl.main(Crawl.java:135) 


can anyone help me to resolve this problem. 

Thank you in advance. 

-- 
View this message in context: 
http://www.nabble.com/NUTCH_CRAWLING-tp25903220p25903220.html
Sent from the Nutch - User mailing list archive at Nabble.com.

Reply via email to