Hi, Kuro
I've fixed my problem, actually the in-dir and out-dir arguments for the
wordcount program should be hadoopDFS instead of the local filesystem, so
that's why it has an exception.
So for your case, are you using Nutch with hadoop or just nutch itself?
William
Teruhiko Kurosaka <[EMAIL PROTECTED]> wrote: I downloaded the latest nightly
build of nutch today and tried to run
"nutch crawl" but
the crawling stopped with the same "Job Failed!" message:
060607 154925 CrawlDb update: Merging segment data into db.
java.io.IOException: Job failed!
at org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:342)
at org.apache.nutch.crawl.CrawlDb.update(CrawlDb.java:54)
at org.apache.nutch.crawl.Crawl.main(Crawl.java:113)
Exception in thread "main"
Is my symptom related to the one that William reports?
-kuro
> -----Original Message-----
> From: William Choi [mailto:[EMAIL PROTECTED]
> Sent: 2006-6-07 16:01
> To: [email protected]
> Subject: Error on running Hadoop examples
>
> Hi all,
>
> I've downloaded last night build of Hadoop, and I want
> to run the wordcount example. However I got an error:
>
> [EMAIL PROTECTED] hadoopBuild]$ bin/hadoop jar
> build/hadoop-0.3-dev-examples.jar sort
> /hadoop/home/testExampleIn/ /hadoop/home/testExampleOut/
> 060607 155753 parsing
> file:/hadoop/hadoopBuild/conf/hadoop-default.xml
> 060607 155753 parsing
> file:/hadoop/hadoopBuild/conf/mapred-default.xml
> 060607 155753 parsing
> file:/hadoop/hadoopBuild/conf/mapred-default.xml
> 060607 155753 parsing file:/hadoop/hadoopBuild/conf/hadoop-site.xml
> 060607 155753 Client connection to 192.168.0.101:9001: starting
> 060607 155753 Client connection to 192.168.0.101:9000: starting
> Running on 3 nodes to sort from /hadoop/home/testExampleIn
> into /hadoop/home/testExampleOut.
> Job started: Wed Jun 07 15:57:53 PDT 2006
> 060607 155753 parsing
> file:/hadoop/hadoopBuild/conf/hadoop-default.xml
> 060607 155754 parsing file:/hadoop/hadoopBuild/conf/hadoop-site.xml
> 060607 155754 Running job: job_0001
> 060607 155755 map 100% reduce 100%
> java.io.IOException: Job failed!
> at
> org.apache.hadoop.mapred.JobClient.runJob(JobClient.java:341)
> at org.apache.hadoop.examples.Sort.main(Sort.java:115)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccess
> orImpl.java:39)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMeth
> odAccessorImpl.java:25)
> at java.lang.reflect.Method.invoke(Method.java:585)
> at
> org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke
> (ProgramDriver.java:69)
> at
> org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:141)
> at
> org.apache.hadoop.examples.ExampleDriver.main(ExampleDriver.java:39)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> at
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccess
> orImpl.java:39)
> at
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMeth
> odAccessorImpl.java:25)
> at java.lang.reflect.Method.invoke(Method.java:585)
> at org.apache.hadoop.util.RunJar.main(RunJar.java:130)
>
>
> ANY idea what's going ON? I set this up on 3 machines, the
> set up is actually pretty similar as what I told by the
> NutchHadoopTutorial. I only modified the hadoop-site.xml,
> hadoop-env.sh and slaves so that it can work on 3 machines.
>
> Any help will be appreciated.
>
>
> William
>
> __________________________________________________
> Do You Yahoo!?
> Tired of spam? Yahoo! Mail has the best spam protection around
> http://mail.yahoo.com
> __________________________________________________
> Do You Yahoo!?
> Tired of spam? Yahoo! Mail has the best spam protection around
> http://mail.yahoo.com
>
__________________________________________________
Do You Yahoo!?
Tired of spam? Yahoo! Mail has the best spam protection around
http://mail.yahoo.com
__________________________________________________
Do You Yahoo!?
Tired of spam? Yahoo! Mail has the best spam protection around
http://mail.yahoo.com