Hi all:

I'd like to run Nutch Crawl in Eclipse. I have followed the
"RunNutchInEclipse" tutorial (http://wiki.apache.org/nutch/RunNutchInEclipse).
However when I tried to run the crawler, the following exception occurred:

==============================================================================
solrUrl is not set, indexing will be skipped...
Exception in thread "main" java.io.IOException: Job failed!
at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:1252)
 at org.apache.nutch.crawl.Injector.inject(Injector.java:217)
at org.apache.nutch.crawl.Crawl.run(Crawl.java:127)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65)
at org.apache.nutch.crawl.Crawl.main(Crawl.java:55)
==============================================================================


The Eclipse Run Configurations are all set according to the tutorial.
==============================================================================

Main Class: org.apache.nutch.crawl.Crawl
Program Arguments: urls -dir crawl -depth 2 -topN 10
VM arguments: -Dhadoop.log.dir=logs -Dhadoop.log.file=hadoop.log
Working directory: Default

==============================================================================
And I have set the classpath, src path, ivy path, etc according to the
tutorial too.


I can build using Ant in Eclipse. Afterwards I can successfully manually
run the crawl script using

$NUTCH_HOME/runtime/local/bin/nutch crawl urls -dir crawl -depth 2 -topN 10


And it does do the crawl process correctly and give the correct result. But
when I try to run it in Eclipse, it always fail.



Does anyone have the similar problem and knows how it can be solved?
Appreciate your time and help.


Andy

Reply via email to