Thanks, that's where I started but #3, which has several methods I
tried, did not work and the reference claims they are not currently
supported in cloudera's implementation. Not sure why so I don't know if
it's a hadoop thing or not.
I ended up putting the classes in the job jar (#2 of cloudera ref),
which isn't ideal since I'm using mahout's jobs and I really didn't want
to rebuild that all the time but oh well it works.
On 5/8/12 11:02 PM, JunYong Li wrote:
hope useful to you:
http://www.cloudera.com/blog/2011/01/how-to-include-third-party-libraries-in-your-map-reduce-job/
2012/5/9 Pat Ferrel<p...@farfetchers.com>
I specify a class at runtime via a parameter to a job. I used to put a jar
with the possible class choices in $HADOOP_HOME/lib on all machines (using
only one in a psuedo cluster for this example) then restart my cluster and
run my job. Then I upgraded from 0.20.203 to 0.20.205, which may have
nothing to do with the problem but there it is.
In any case the job cannot find the class in $HADOOP_HOME/lib. I checked
the jar and its in the right place and has my classes in it.
Did I miss something? Is there a better way?
Sorry ahead of time, this must be a repeat of a repeat but the above is
the best advice I can find.