Hi,

I am new to the world of Mahout and Hadoop though I have worked with Lucene.

I am trying to run the clustering example as specified here :
http://cwiki.apache.org/MAHOUT/syntheticcontroldata.html

I got the job file for examples from
http://mirrors.ibiblio.org/pub/mirrors/maven2/org/apache/mahout/mahout-examples/0.1/

I started Hadoop (in a single node configuration) and tried to run the
example with following command.

$HADOOP_HOME/bin/hadoop jar
$MAHOUT_HOME/examples/target/mahout-examples-0.1.job
org.apache.mahout.clustering.syntheticcontrol.kmeans.Job

It starts and displays following message.

INFO mapred.FileInputFormat: Total input paths to process : 1
INFO mapred.FileInputFormat: Total input paths to process : 1
mapred.JobClient: Running job: job_200904281825_0005
INFO mapred.JobClient:  map 0% reduce 0%

Then immediately, it throws following exception multiple times and dies.

INFO mapred.JobClient: Task Id : attempt_200904281825_0004_m_000001_2,
Status : FAILED
java.lang.UnsupportedClassVersionError: Bad version number in .class file

Initially, I got the version number error at the beginning. I found
that JDK version was 1.5. It has been upgraded it to 1.6. Now
JAVA_HOME points to /usr/java/jdk1.6.0_13/  and I am using Hadoop
0.18.3.

1. What could possibly be wrong? I checked the Hadoop script. Value of
JAVA_HOME is correct (ie 1.6). Is it possible that somehow it is still
using 1.5?

2. The last step the clustering tutorial says "Get the data out of
HDFS and have a look." Can you please point me to the documentation of
Hadoop about how to read this data?


Thanks,

--shashi

Reply via email to