Is it possible that the Hadoop job jar mechanism is broken? Try disabling
the distributed Hadoop feature, and run "pseudo-distributed":
unset HADOOP_HOME
sh examples/bin/reuters-build.sh

On Tue, Nov 22, 2011 at 8:25 PM, Lance Norskog <[email protected]> wrote:

> There is something wrong with how you are building the Mahout source. This
> is the sequence that should work:
>
> First, remove your Maven module download directory. Usually this is
> /home/dipesh/.m2. Your build will now download all of the dependencies. (It
> is not usually the problem, but it helps to do everything from the
> beginning.)
>
> export MAHOUT_HOME=/your/path/of/source/code
> cd $MAHOUT_HOME
> mvn clean install
> bin/mahout
>
> This should give you a list of the commands.
>
> Now run the reuters script.
>
>
> On Tue, Nov 22, 2011 at 1:36 PM, Isabel Drost <[email protected]> wrote:
>
>> **
>>
>> On 22.11.2011 DIPESH KUMAR SINGH wrote:
>>
>> > I ran the script and i was getting error regarding missing libraries.
>> The
>>
>> > error which i got is attached.
>>
>> > Then i tried executing the commands in the script, command by command,
>> and
>>
>> > i figured out that error was coming
>>
>> > in the seq2sparse step. (Prior to this step all the conversions are
>> working
>>
>> > fine)
>>
>>
>> There seem to be problems resolving some of the dependencies used - not
>> sure why though. You did compile the project and in that process created a
>> job jar?
>>
>>
>>
>> > What i exactly want to try is document clustering, i thought it is
>> better
>>
>> > to try first with Reuters dataset to get started.
>>
>> > Are the source files of kmeans (mapper and reducer etc) are there in
>> mahout
>>
>> > source folder?
>>
>>
>> Sure, look in the maven module core in the o.a.m.clustering package - all
>> kmeans related code is in there.
>>
>>
>> Isabel
>>
>
>
>
> --
> Lance Norskog
> [email protected]
>
>


-- 
Lance Norskog
[email protected]

Reply via email to