Is it possible that the Hadoop job jar mechanism is broken? Try disabling the distributed Hadoop feature, and run "pseudo-distributed": unset HADOOP_HOME sh examples/bin/reuters-build.sh
On Tue, Nov 22, 2011 at 8:25 PM, Lance Norskog <[email protected]> wrote: > There is something wrong with how you are building the Mahout source. This > is the sequence that should work: > > First, remove your Maven module download directory. Usually this is > /home/dipesh/.m2. Your build will now download all of the dependencies. (It > is not usually the problem, but it helps to do everything from the > beginning.) > > export MAHOUT_HOME=/your/path/of/source/code > cd $MAHOUT_HOME > mvn clean install > bin/mahout > > This should give you a list of the commands. > > Now run the reuters script. > > > On Tue, Nov 22, 2011 at 1:36 PM, Isabel Drost <[email protected]> wrote: > >> ** >> >> On 22.11.2011 DIPESH KUMAR SINGH wrote: >> >> > I ran the script and i was getting error regarding missing libraries. >> The >> >> > error which i got is attached. >> >> > Then i tried executing the commands in the script, command by command, >> and >> >> > i figured out that error was coming >> >> > in the seq2sparse step. (Prior to this step all the conversions are >> working >> >> > fine) >> >> >> There seem to be problems resolving some of the dependencies used - not >> sure why though. You did compile the project and in that process created a >> job jar? >> >> >> >> > What i exactly want to try is document clustering, i thought it is >> better >> >> > to try first with Reuters dataset to get started. >> >> > Are the source files of kmeans (mapper and reducer etc) are there in >> mahout >> >> > source folder? >> >> >> Sure, look in the maven module core in the o.a.m.clustering package - all >> kmeans related code is in there. >> >> >> Isabel >> > > > > -- > Lance Norskog > [email protected] > > -- Lance Norskog [email protected]
