OK, well the underlying problem remains that it cannot find the mahout-math jar file in the classpath. It's not a problem with 0.5 per se, but something to do with finding the jars.
On Fri, Aug 5, 2011 at 12:52 PM, air <[email protected]> wrote: > hi, Sean, for 0.4 and 0.5, I use the same settings ,in fact no settings, I > just download them and unpack them, and then use them to do some test... > > anyone meet this kind of problem ? thank you . > > > 2011/8/5 Sean Owen <[email protected]> > > > I think you have changed your classpath to use 0.5, and in your new > > classpath, you are not including all the jars. You probably need core, > > math, > > and collections at minimum. > > > > On Fri, Aug 5, 2011 at 12:40 PM, air <[email protected]> wrote: > > > > > when I use mahout 0.4 to execute on a hadoop cluster: > > > * > > > ./mahout org.apache.mahout.clustering.syntheticcontrol.canopy.Job > --input > > > /mahout/input --output /mahout/output13 -t1 80.0 -t2 50.0* > > > > > > it works very well...... > > > > > > but when I use mahout 0.5 to execute *the same command(if omitting the > > > output directory...):* > > > > > > *./mahout org.apache.mahout.clustering.syntheticcontrol.canopy.Job > > --input > > > /mahout/input --output /mahout/output14 -t1 80.0 -t2 50.0* > > > > > > it reports CLASSNOTFOUND exception, what cause this happen, how to > solve > > it > > > ... I am really confused....: > > > > > > > > > [root@x06 bin]# ./mahout > > > org.apache.mahout.clustering.syntheticcontrol.canopy.Job --input > > > /mahout/input --output /mahout/output15 -t1 80.0 -t2 50.0 > > > Running on hadoop, using HADOOP_HOME=/usr/lib/hadoop > > > No HADOOP_CONF_DIR set, using /usr/lib/hadoop/src/conf > > > 11/08/05 19:36:43 WARN driver.MahoutDriver: No > > > org.apache.mahout.clustering.syntheticcontrol.canopy.Job.props found on > > > classpath, will use command-line arguments only > > > 11/08/05 19:36:43 INFO canopy.Job: Running with only user-supplied > > > arguments > > > 11/08/05 19:36:43 INFO common.AbstractJob: Command line arguments: > > > > > > > > > {--distanceMeasure=org.apache.mahout.common.distance.SquaredEuclideanDistanceMeasure, > > > --endPhase=2147483647, --input=/mahout/input, > --output=/mahout/output15, > > > --startPhase=0, --t1=80.0, --t2=50.0, --tempDir=temp} > > > 11/08/05 19:37:09 WARN mapred.JobClient: Use GenericOptionsParser for > > > parsing the arguments. Applications should implement Tool for the same. > > > 11/08/05 19:37:43 INFO input.FileInputFormat: Total input paths to > > process > > > : > > > 1 > > > 11/08/05 19:37:43 INFO lzo.GPLNativeCodeLoader: Loaded native gpl > library > > > 11/08/05 19:37:43 INFO lzo.LzoCodec: Successfully loaded & initialized > > > native-lzo library [hadoop-lzo rev > > > 2991c7d00d4b50494958eb1c8e7ef086d24853ea] > > > 11/08/05 19:37:54 INFO mapred.JobClient: Running job: > > job_201108041755_7040 > > > 11/08/05 19:37:55 INFO mapred.JobClient: map 0% reduce 0% > > > 11/08/05 19:38:19 INFO mapred.JobClient: map 100% reduce 0% > > > 11/08/05 19:38:21 INFO mapred.JobClient: Job complete: > > > job_201108041755_7040 > > > 11/08/05 19:38:21 INFO mapred.JobClient: Counters: 13 > > > 11/08/05 19:38:21 INFO mapred.JobClient: Job Counters > > > 11/08/05 19:38:21 INFO mapred.JobClient: SLOTS_MILLIS_MAPS=4911 > > > 11/08/05 19:38:21 INFO mapred.JobClient: Total time spent by all > > > reduces > > > waiting after reserving slots (ms)=0 > > > 11/08/05 19:38:21 INFO mapred.JobClient: Total time spent by all > maps > > > waiting after reserving slots (ms)=0 > > > 11/08/05 19:38:21 INFO mapred.JobClient: Launched map tasks=1 > > > 11/08/05 19:38:21 INFO mapred.JobClient: Data-local map tasks=1 > > > 11/08/05 19:38:21 INFO mapred.JobClient: SLOTS_MILLIS_REDUCES=4017 > > > 11/08/05 19:38:21 INFO mapred.JobClient: FileSystemCounters > > > 11/08/05 19:38:21 INFO mapred.JobClient: HDFS_BYTES_READ=288490 > > > 11/08/05 19:38:21 INFO mapred.JobClient: FILE_BYTES_WRITTEN=49166 > > > 11/08/05 19:38:21 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=335470 > > > 11/08/05 19:38:21 INFO mapred.JobClient: Map-Reduce Framework > > > 11/08/05 19:38:21 INFO mapred.JobClient: Map input records=600 > > > 11/08/05 19:38:21 INFO mapred.JobClient: Spilled Records=0 > > > 11/08/05 19:38:21 INFO mapred.JobClient: Map output records=600 > > > 11/08/05 19:38:21 INFO mapred.JobClient: SPLIT_RAW_BYTES=116 > > > 11/08/05 19:38:21 INFO canopy.CanopyDriver: Build Clusters Input: > > > /mahout/output15/data Out: /mahout/output15 Measure: > > > > > > > > > org.apache.mahout.common.distance.SquaredEuclideanDistanceMeasure@4dc1c92bt1 > > > : > > > 80.0 t2: 50.0 > > > 11/08/05 19:38:21 WARN mapred.JobClient: Use GenericOptionsParser for > > > parsing the arguments. Applications should implement Tool for the same. > > > 11/08/05 19:38:26 INFO input.FileInputFormat: Total input paths to > > process > > > : > > > 1 > > > 11/08/05 19:38:31 INFO mapred.JobClient: Running job: > > job_201108041755_7056 > > > 11/08/05 19:38:32 INFO mapred.JobClient: map 0% reduce 0% > > > 11/08/05 19:38:46 INFO mapred.JobClient: Task Id : > > > attempt_201108041755_7056_m_000000_0, Status : FAILED > > > Error: java.lang.ClassNotFoundException: org.apache.mahout.math.Vector > > > at java.net.URLClassLoader$1.run(URLClassLoader.java:202) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at java.net.URLClassLoader.findClass(URLClassLoader.java:190) > > > at java.lang.ClassLoader.loadClass(ClassLoader.java:306) > > > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) > > > at java.lang.ClassLoader.loadClass(ClassLoader.java:247) > > > at java.lang.Class.forName0(Native Method) > > > at java.lang.Class.forName(Class.java:247) > > > at > > > > > > org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:943) > > > at > > > org.apache.hadoop.conf.Configuration.getClass(Configuration.java:994) > > > at > > > org.apache.hadoop.conf.Configuration.getClass(Configuration.java:1020) > > > at > > > > org.apache.hadoop.mapred.JobConf.getMapOutputValueClass(JobConf.java:747) > > > at > > > > org.apache.hadoop.mapred.MapTask$MapOutputBuffer.<init>(MapTask.java:819) > > > at > > > > > > > > > org.apache.hadoop.mapred.MapTask$NewOutputCollector.<init>(MapTask.java:557) > > > at > org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:639) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:323) > > > at org.apache.hadoop.mapred.Child$4.run(Child.java:270) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:396) > > > at > > > > > > > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1127) > > > at org.apache.hadoop.mapred.Child.main(Child.java:264) > > > > > > 11/08/05 19:38:52 INFO mapred.JobClient: Task Id : > > > attempt_201108041755_7056_m_000000_1, Status : FAILED > > > Error: java.lang.ClassNotFoundException: org.apache.mahout.math.Vector > > > at java.net.URLClassLoader$1.run(URLClassLoader.java:202) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at java.net.URLClassLoader.findClass(URLClassLoader.java:190) > > > at java.lang.ClassLoader.loadClass(ClassLoader.java:306) > > > at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) > > > at java.lang.ClassLoader.loadClass(ClassLoader.java:247) > > > at java.lang.Class.forName0(Native Method) > > > at java.lang.Class.forName(Class.java:247) > > > at > > > > > > org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:943) > > > at > > > org.apache.hadoop.conf.Configuration.getClass(Configuration.java:994) > > > at > > > org.apache.hadoop.conf.Configuration.getClass(Configuration.java:1020) > > > at > > > > org.apache.hadoop.mapred.JobConf.getMapOutputValueClass(JobConf.java:747) > > > at > > > > org.apache.hadoop.mapred.MapTask$MapOutputBuffer.<init>(MapTask.java:819) > > > at > > > > > > > > > org.apache.hadoop.mapred.MapTask$NewOutputCollector.<init>(MapTask.java:557) > > > at > org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:639) > > > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:323) > > > at org.apache.hadoop.mapred.Child$4.run(Child.java:270) > > > at java.security.AccessController.doPrivileged(Native Method) > > > at javax.security.auth.Subject.doAs(Subject.java:396) > > > at > > > > > > > > > org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1127) > > > at org.apache.hadoop.mapred.Child.main(Child.java:264) > > > > > > 11/08/05 19:39:00 INFO mapred.JobClient: Task Id : > > > attempt_201108041755_7056_m_000000_2, Status : FAILED > > > 11/08/05 19:39:12 INFO mapred.JobClient: Job complete: > > > job_201108041755_7056 > > > 11/08/05 19:39:12 INFO mapred.JobClient: Counters: 8 > > > 11/08/05 19:39:12 INFO mapred.JobClient: Job Counters > > > 11/08/05 19:39:12 INFO mapred.JobClient: SLOTS_MILLIS_MAPS=7530 > > > 11/08/05 19:39:12 INFO mapred.JobClient: Total time spent by all > > > reduces > > > waiting after reserving slots (ms)=0 > > > 11/08/05 19:39:12 INFO mapred.JobClient: Total time spent by all > maps > > > waiting after reserving slots (ms)=0 > > > 11/08/05 19:39:12 INFO mapred.JobClient: Rack-local map tasks=3 > > > 11/08/05 19:39:12 INFO mapred.JobClient: Launched map tasks=4 > > > 11/08/05 19:39:12 INFO mapred.JobClient: Data-local map tasks=1 > > > 11/08/05 19:39:12 INFO mapred.JobClient: Failed map tasks=1 > > > 11/08/05 19:39:12 INFO mapred.JobClient: SLOTS_MILLIS_REDUCES=2464 > > > Exception in thread "main" java.lang.InterruptedException: Canopy Job > > > failed > > > processing /mahout/output15/data > > > at > > > > > > > > > org.apache.mahout.clustering.canopy.CanopyDriver.buildClustersMR(CanopyDriver.java:355) > > > at > > > > > > > > > org.apache.mahout.clustering.canopy.CanopyDriver.buildClusters(CanopyDriver.java:246) > > > at > > > > > > org.apache.mahout.clustering.canopy.CanopyDriver.run(CanopyDriver.java:143) > > > at > > > > > > org.apache.mahout.clustering.canopy.CanopyDriver.run(CanopyDriver.java:161) > > > at > > > > org.apache.mahout.clustering.syntheticcontrol.canopy.Job.run(Job.java:92) > > > at > > > > > > org.apache.mahout.clustering.syntheticcontrol.canopy.Job.run(Job.java:127) > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > > > at > > > > > > org.apache.mahout.clustering.syntheticcontrol.canopy.Job.main(Job.java:49) > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > > at > > > > > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > > > at > > > > > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > > > at java.lang.reflect.Method.invoke(Method.java:597) > > > at > > > > > > > > > org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:68) > > > at > > > org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:139) > > > at > > org.apache.mahout.driver.MahoutDriver.main(MahoutDriver.java:187) > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > > at > > > > > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > > > at > > > > > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > > > at java.lang.reflect.Method.invoke(Method.java:597) > > > at org.apache.hadoop.util.RunJar.main(RunJar.java:186) > > > [root@x06 bin]# > > > -- > > > Knowledge Mangement . > > > > > > > > > -- > Knowledge Mangement . >
