You should not need to configure any classpath. Use the "job" files which
contains all dependencies. You can run it locally or in a cluster.

2011/8/10 Clément Notin <[email protected]>

> Ho I'm sorry I thought it wasn't running on HDFS because of the local
> /tmp/... folder. You're right thanks !
>
> But (yes I know...) it's using LocalJobRunner (in the log msg) so I assume
> it doesn't run on the cluster. I have in the classpath the "masters" and
> "slaves" files, and also core-site.xml and hdfs-site.xml so it should run
> there nop ?
>
> Thanks for you help !
>
> 2011/8/10 Sean Owen <[email protected]>
>
> > I don't believe it's actually cleaned out then. Hadoop thinks the temp
> > directory exists from a previous run, which perhaps failed. Make sure it
> is
> > deleted in HDFS. This is, at least, what the error is trying to tell you.
> > Are you running two jobs that might both want this directory?
> >
> > 2011/8/10 Clément Notin <[email protected]>
> >
> > > Yes I agree it's ugly ;)
> > >
> > > I tried with the params
> > > "org.apache.mahout.cf.taste.hadoop.item.RecommenderJob
> > > -Dmapred.input.dir=mb-recouser-input/input.csv
> > > -Dmapred.output.dir=mb-recouser-output/reco.csv --numRecommendations 3
> > > --booleanData true --similarityClassname SIMILARITY_EUCLIDEAN_DISTANCE"
> > (of
> > > course I split them).
> > >
> > > But I'm getting an error :
> > >  INFO [2011-08-10 14:52:05,195] (JobClient.java:871) - Cleaning up the
> > > staging area
> > >
> > >
> >
> file:/tmp/hadoop-clement/mapred/staging/clement1957523084/.staging/job_local_0001
> > > org.apache.hadoop.mapred.FileAlreadyExistsException: Output directory
> > > temp/itemIDIndex already exists
> > >
> > > Even if I clean before the /tmp/hadoop-clement/ folder...
> > > And it don't seems to run on the cluster.
> > >
> > > 2011/8/10 Sean Owen <[email protected]>
> > >
> > > > You could just run the main() method with an array of the same
> > arguments
> > > > you
> > > > passed on the command line. It's a little ugly but entirely works.
> > > >
> > > > 2011/8/10 Clément Notin <[email protected]>
> > > >
> > > > > Hello,
> > > > >
> > > > > I've achieved to run a recommender over hadoop using the command
> line
> > > > > /bin/mahout org.apache.mahout.cf.taste.hadoop.item.RecommenderJob
> > > --input
> > > > > .....
> > > > > I'm happy with it but now I want to launch this using Java.
> > > > >
> > > > > What is the easiest way to do this ? I tried to run the
> MahoutDriver
> > > but
> > > > it
> > > > > runs locally however I want to launch the job on an hadoop cluster.
> > > > >
> > > > > Regards.
> > > > >
> > > > > --
> > > > > *Clément **Notin*
> > > > >
> > > >
> > >
> > >
> > >
> > > --
> > > *Clément **Notin*
> > >
> >
> > --
> > *Clément **Notin*
> >  <http://fr.linkedin.com/in/clementnotin>
> >
>

Reply via email to