i try to crawl with nutch "bin/nutch crawl urls -dir crawled -depth 3" but
show error that
- parsing
jar:file:/usr/lib/jdk1.6.0_01/jre/lib/ext/hadoop-0.9.2.jar!/hadoop-default.xml
- parsing jar:file:/nutch/search/lib/nutch-0.8.1.jar!/nutch-default.xml
- parsing file:/nutch/search/conf/crawl-tool.xml
- parsing
jar:file:/usr/lib/jdk1.6.0_01/jre/lib/ext/hadoop-0.9.2.jar!/mapred-default.xml
- parsing
jar:file:/usr/lib/jdk1.6.0_01/jre/lib/ext/hadoop-0.9.2.jar!/mapred-default.xml
- org.apache.hadoop.io.ObjectWritableConnection culler maxidletime= 1000ms
- org.apache.hadoop.io.ObjectWritable Connection Culler: starting
- crawl started in: crawled
- rootUrlDir = urls
- threads = 10
- depth = 3
- Injector: starting
- Injector: crawlDb: crawled/crawldb
- Injector: urlDir: urls
- Injector: Converting injected urls to crawl db entries.
- parsing
jar:file:/usr/lib/jdk1.6.0_01/jre/lib/ext/hadoop-0.9.2.jar!/hadoop-default.xml
- parsing jar:file:/nutch/search/lib/nutch-0.8.1.jar!/nutch-default.xml
- parsing file:/nutch/search/conf/crawl-tool.xml
- parsing
jar:file:/usr/lib/jdk1.6.0_01/jre/lib/ext/hadoop-0.9.2.jar!/mapred-default.xml
- parsing
jar:file:/usr/lib/jdk1.6.0_01/jre/lib/ext/hadoop-0.9.2.jar!/mapred-default.xml
- parsing
jar:file:/usr/lib/jdk1.6.0_01/jre/lib/ext/hadoop-0.9.2.jar!/mapred-default.xml
- parsing
jar:file:/usr/lib/jdk1.6.0_01/jre/lib/ext/hadoop-0.9.2.jar!/mapred-default.xml
- parsing
jar:file:/usr/lib/jdk1.6.0_01/jre/lib/ext/hadoop-0.9.2.jar!/hadoop-default.xml
- Running job: job_0005
- map 0% reduce 0%
- map 100% reduce 100%
Exception in thread "main" java.io.IOException: Job failed!
at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:399)
at org.apache.nutch.crawl.Injector.inject(Injector.java:138)
at org.apache.nutch.crawl.Crawl.main(Crawl.java:105)
why? how i solve this problem?
--
View this message in context:
http://www.nabble.com/Error-Nutch-Crawl-tf4546229.html#a12973344
Sent from the Hadoop Users mailing list archive at Nabble.com.