thanks. I think I put the data there in a wrong format. It works now up until a point but for some reason it fails again:

[tjam...@bunwell ~]$ hadoop jar /localhome/tjambor/mahout/mahout-core-0.3.job org.apache.mahout.cf.taste.hadoop.item.RecommenderJob --input testdata/100k_data.data --output output -t temp --jarFile /localhome/tjambor/mahout/mahout-core-0.3.jar 10/05/02 19:22:32 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 10/05/02 19:22:33 INFO mapred.FileInputFormat: Total input paths to process : 1
10/05/02 19:22:33 INFO mapred.JobClient: Running job: job_201004291158_0024
10/05/02 19:22:34 INFO mapred.JobClient:  map 0% reduce 0%
10/05/02 19:22:44 INFO mapred.JobClient:  map 50% reduce 0%
10/05/02 19:22:45 INFO mapred.JobClient:  map 100% reduce 0%
10/05/02 19:22:56 INFO mapred.JobClient:  map 100% reduce 100%
10/05/02 19:22:58 INFO mapred.JobClient: Job complete: job_201004291158_0024
10/05/02 19:22:58 INFO mapred.JobClient: Counters: 19
10/05/02 19:22:58 INFO mapred.JobClient:   Job Counters
10/05/02 19:22:58 INFO mapred.JobClient:     Launched reduce tasks=1
10/05/02 19:22:58 INFO mapred.JobClient:     Rack-local map tasks=1
10/05/02 19:22:58 INFO mapred.JobClient:     Launched map tasks=2
10/05/02 19:22:58 INFO mapred.JobClient:     Data-local map tasks=1
10/05/02 19:22:58 INFO mapred.JobClient:   FileSystemCounters
10/05/02 19:22:58 INFO mapred.JobClient:     FILE_BYTES_READ=1400006
10/05/02 19:22:58 INFO mapred.JobClient:     HDFS_BYTES_READ=981108
10/05/02 19:22:58 INFO mapred.JobClient:     FILE_BYTES_WRITTEN=2800082
10/05/02 19:22:58 INFO mapred.JobClient:     HDFS_BYTES_WRITTEN=42610
10/05/02 19:22:58 INFO mapred.JobClient:   Map-Reduce Framework
10/05/02 19:22:58 INFO mapred.JobClient:     Reduce input groups=1682
10/05/02 19:22:58 INFO mapred.JobClient:     Combine output records=0
10/05/02 19:22:58 INFO mapred.JobClient:     Map input records=100000
10/05/02 19:22:58 INFO mapred.JobClient:     Reduce shuffle bytes=691382
10/05/02 19:22:58 INFO mapred.JobClient:     Reduce output records=1682
10/05/02 19:22:58 INFO mapred.JobClient:     Spilled Records=200000
10/05/02 19:22:58 INFO mapred.JobClient:     Map output bytes=1200000
10/05/02 19:22:58 INFO mapred.JobClient:     Map input bytes=979173
10/05/02 19:22:58 INFO mapred.JobClient:     Combine input records=0
10/05/02 19:22:58 INFO mapred.JobClient:     Map output records=100000
10/05/02 19:22:58 INFO mapred.JobClient:     Reduce input records=100000
10/05/02 19:22:58 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same. 10/05/02 19:22:58 INFO mapred.FileInputFormat: Total input paths to process : 1
10/05/02 19:22:58 INFO mapred.JobClient: Running job: job_201004291158_0025
10/05/02 19:22:59 INFO mapred.JobClient:  map 0% reduce 0%
10/05/02 19:23:09 INFO mapred.JobClient:  map 100% reduce 0%
10/05/02 19:23:20 INFO mapred.JobClient: Task Id : attempt_201004291158_0025_r_000000_0, Status : FAILED
Error: java.lang.ClassNotFoundException: org.apache.mahout.math.Vector
        at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
        at java.security.AccessController.doPrivileged(Native Method)
        at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:307)
        at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:248)
        at java.lang.Class.forName0(Native Method)
        at java.lang.Class.forName(Class.java:247)
at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:762) at org.apache.hadoop.conf.Configuration.getClass(Configuration.java:807) at org.apache.hadoop.conf.Configuration.getClass(Configuration.java:833) at org.apache.hadoop.mapred.JobConf.getReducerClass(JobConf.java:832) at org.apache.hadoop.mapred.ReduceTask.runOldReducer(ReduceTask.java:426)
        at org.apache.hadoop.mapred.ReduceTask.run(ReduceTask.java:411)
        at org.apache.hadoop.mapred.Child.main(Child.java:170)


On 02/05/2010 19:02, Sean Owen wrote:
(PS you should really try using the latest code from Subversion --
it's changed a little bit in the arguments, but is much more efficient
and effective. The javadoc explains the new usage.)

On Sun, May 2, 2010 at 7:01 PM, Sean Owen<sro...@gmail.com>  wrote:
--input specifies the data to use, and you have done so. It sounds
like it's empty or not in the right format. What is in
testdata/test.txt?

Reply via email to