You should not need to configure any classpath. Use the "job" files which contains all dependencies. You can run it locally or in a cluster.
2011/8/10 Clément Notin <[email protected]> > Ho I'm sorry I thought it wasn't running on HDFS because of the local > /tmp/... folder. You're right thanks ! > > But (yes I know...) it's using LocalJobRunner (in the log msg) so I assume > it doesn't run on the cluster. I have in the classpath the "masters" and > "slaves" files, and also core-site.xml and hdfs-site.xml so it should run > there nop ? > > Thanks for you help ! > > 2011/8/10 Sean Owen <[email protected]> > > > I don't believe it's actually cleaned out then. Hadoop thinks the temp > > directory exists from a previous run, which perhaps failed. Make sure it > is > > deleted in HDFS. This is, at least, what the error is trying to tell you. > > Are you running two jobs that might both want this directory? > > > > 2011/8/10 Clément Notin <[email protected]> > > > > > Yes I agree it's ugly ;) > > > > > > I tried with the params > > > "org.apache.mahout.cf.taste.hadoop.item.RecommenderJob > > > -Dmapred.input.dir=mb-recouser-input/input.csv > > > -Dmapred.output.dir=mb-recouser-output/reco.csv --numRecommendations 3 > > > --booleanData true --similarityClassname SIMILARITY_EUCLIDEAN_DISTANCE" > > (of > > > course I split them). > > > > > > But I'm getting an error : > > > INFO [2011-08-10 14:52:05,195] (JobClient.java:871) - Cleaning up the > > > staging area > > > > > > > > > file:/tmp/hadoop-clement/mapred/staging/clement1957523084/.staging/job_local_0001 > > > org.apache.hadoop.mapred.FileAlreadyExistsException: Output directory > > > temp/itemIDIndex already exists > > > > > > Even if I clean before the /tmp/hadoop-clement/ folder... > > > And it don't seems to run on the cluster. > > > > > > 2011/8/10 Sean Owen <[email protected]> > > > > > > > You could just run the main() method with an array of the same > > arguments > > > > you > > > > passed on the command line. It's a little ugly but entirely works. > > > > > > > > 2011/8/10 Clément Notin <[email protected]> > > > > > > > > > Hello, > > > > > > > > > > I've achieved to run a recommender over hadoop using the command > line > > > > > /bin/mahout org.apache.mahout.cf.taste.hadoop.item.RecommenderJob > > > --input > > > > > ..... > > > > > I'm happy with it but now I want to launch this using Java. > > > > > > > > > > What is the easiest way to do this ? I tried to run the > MahoutDriver > > > but > > > > it > > > > > runs locally however I want to launch the job on an hadoop cluster. > > > > > > > > > > Regards. > > > > > > > > > > -- > > > > > *Clément **Notin* > > > > > > > > > > > > > > > > > > > > > -- > > > *Clément **Notin* > > > > > > > -- > > *Clément **Notin* > > <http://fr.linkedin.com/in/clementnotin> > > >
