thanks. I think I put the data there in a wrong format. It works now up
until a point but for some reason it fails again:
[tjam...@bunwell ~]$ hadoop jar
/localhome/tjambor/mahout/mahout-core-0.3.job
org.apache.mahout.cf.taste.hadoop.item.RecommenderJob --input
testdata/100k_data.data --output output -t temp --jarFile
/localhome/tjambor/mahout/mahout-core-0.3.jar
10/05/02 19:22:32 WARN mapred.JobClient: Use GenericOptionsParser for
parsing the arguments. Applications should implement Tool for the same.
10/05/02 19:22:33 INFO mapred.FileInputFormat: Total input paths to
process : 1
10/05/02 19:22:33 INFO mapred.JobClient: Running job: job_201004291158_0024
10/05/02 19:22:34 INFO mapred.JobClient: map 0% reduce 0%
10/05/02 19:22:44 INFO mapred.JobClient: map 50% reduce 0%
10/05/02 19:22:45 INFO mapred.JobClient: map 100% reduce 0%
10/05/02 19:22:56 INFO mapred.JobClient: map 100% reduce 100%
10/05/02 19:22:58 INFO mapred.JobClient: Job complete: job_201004291158_0024
10/05/02 19:22:58 INFO mapred.JobClient: Counters: 19
10/05/02 19:22:58 INFO mapred.JobClient: Job Counters
10/05/02 19:22:58 INFO mapred.JobClient: Launched reduce tasks=1
10/05/02 19:22:58 INFO mapred.JobClient: Rack-local map tasks=1
10/05/02 19:22:58 INFO mapred.JobClient: Launched map tasks=2
10/05/02 19:22:58 INFO mapred.JobClient: Data-local map tasks=1
10/05/02 19:22:58 INFO mapred.JobClient: FileSystemCounters
10/05/02 19:22:58 INFO mapred.JobClient: FILE_BYTES_READ=1400006
10/05/02 19:22:58 INFO mapred.JobClient: HDFS_BYTES_READ=981108
10/05/02 19:22:58 INFO mapred.JobClient: FILE_BYTES_WRITTEN=2800082
10/05/02 19:22:58 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=42610
10/05/02 19:22:58 INFO mapred.JobClient: Map-Reduce Framework
10/05/02 19:22:58 INFO mapred.JobClient: Reduce input groups=1682
10/05/02 19:22:58 INFO mapred.JobClient: Combine output records=0
10/05/02 19:22:58 INFO mapred.JobClient: Map input records=100000
10/05/02 19:22:58 INFO mapred.JobClient: Reduce shuffle bytes=691382
10/05/02 19:22:58 INFO mapred.JobClient: Reduce output records=1682
10/05/02 19:22:58 INFO mapred.JobClient: Spilled Records=200000
10/05/02 19:22:58 INFO mapred.JobClient: Map output bytes=1200000
10/05/02 19:22:58 INFO mapred.JobClient: Map input bytes=979173
10/05/02 19:22:58 INFO mapred.JobClient: Combine input records=0
10/05/02 19:22:58 INFO mapred.JobClient: Map output records=100000
10/05/02 19:22:58 INFO mapred.JobClient: Reduce input records=100000
10/05/02 19:22:58 WARN mapred.JobClient: Use GenericOptionsParser for
parsing the arguments. Applications should implement Tool for the same.
10/05/02 19:22:58 INFO mapred.FileInputFormat: Total input paths to
process : 1
10/05/02 19:22:58 INFO mapred.JobClient: Running job: job_201004291158_0025
10/05/02 19:22:59 INFO mapred.JobClient: map 0% reduce 0%
10/05/02 19:23:09 INFO mapred.JobClient: map 100% reduce 0%
10/05/02 19:23:20 INFO mapred.JobClient: Task Id :
attempt_201004291158_0025_r_000000_0, Status : FAILED
Error: java.lang.ClassNotFoundException: org.apache.mahout.math.Vector
at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
at java.lang.ClassLoader.loadClass(ClassLoader.java:307)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
at java.lang.ClassLoader.loadClass(ClassLoader.java:248)
at java.lang.Class.forName0(Native Method)
at java.lang.Class.forName(Class.java:247)
at
org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:762)
at
org.apache.hadoop.conf.Configuration.getClass(Configuration.java:807)
at
org.apache.hadoop.conf.Configuration.getClass(Configuration.java:833)
at
org.apache.hadoop.mapred.JobConf.getReducerClass(JobConf.java:832)
at
org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:426)
at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:411)
at org.apache.hadoop.mapred.Child.main(Child.java:170)
On 02/05/2010 19:02, Sean Owen wrote:
(PS you should really try using the latest code from Subversion --
it's changed a little bit in the arguments, but is much more efficient
and effective. The javadoc explains the new usage.)
On Sun, May 2, 2010 at 7:01 PM, Sean Owen<sro...@gmail.com> wrote:
--input specifies the data to use, and you have done so. It sounds
like it's empty or not in the right format. What is in
testdata/test.txt?