my admin add some patch to the current hadoop 0.19, and now it works perfectly with mahout 0.7. btw,thank you for all your concernl!
On Fri, Sep 21, 2012 at 10:58 PM, Ted Dunning <[email protected]> wrote: > On the other hand, the only way that I have been able to do a major version > upgrade of Hadoop is to start a new company. > > It is really hard to change code and platform at the same time. If you > don't have enough hardware to have two clusters temporarily, things will be > really hard moving off of 0.19. > > On Fri, Sep 21, 2012 at 9:44 AM, Mat Kelcey <[email protected] > >wrote: > > > I imagine the best use of your time and effort is to convince your admins > > that running a 3 year old version of hadoop is a bad idea. Things are > only > > going to get worse... > > Mat > > On Sep 13, 2012 7:15 PM, "Phoenix Bai" <[email protected]> wrote: > > > > > Hi guys, > > > > > > I am trying to compile my application code using mahout 0.7 and hadoop > > > 0.19. > > > during the compile process, it is throwing errors as below: > > > > > > $ hadoop jar cluster-0.0.1-SNAPSHOT-jar-with-dependencies.jar > > > mahout.sample.ClusterVideos > > > 12/09/13 20:36:18 INFO vectorizer.SparseVectorsFromSequenceFiles: > Maximum > > > n-gram size is: 1 > > > 12/09/13 20:36:31 INFO vectorizer.SparseVectorsFromSequenceFiles: > Minimum > > > LLR value: 1.0 > > > 12/09/13 20:36:31 INFO vectorizer.SparseVectorsFromSequenceFiles: > Number > > of > > > reduce tasks: 1 > > > java.lang.VerifyError: (class: org/apache/hadoop/mapreduce/Job, method: > > > submit signature: ()V) Incompatible argument to function > > > at > > > > > > > > > org.apache.mahout.vectorizer.DocumentProcessor.tokenizeDocuments(DocumentProcessor.java:78) > > > at > > > > > > > > > org.apache.mahout.vectorizer.SparseVectorsFromSequenceFiles.run(SparseVectorsFromSequenceFiles.java:253) > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) > > > at > > > > > > > > > org.apache.mahout.vectorizer.SparseVectorsFromSequenceFiles.main(SparseVectorsFromSequenceFiles.java:55) > > > at mahout.sample.ClusterVideos.runSeq2Sparse(ClusterVideos.java:133) > > > at mahout.sample.ClusterVideos.main(ClusterVideos.java:54) > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > > at > > > > > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > > > at > > > > > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > > > at java.lang.reflect.Method.invoke(Method.java:597) > > > at org.apache.hadoop.util.RunJar.main(RunJar.java:165) > > > at org.apache.hadoop.mapred.JobShell.run(JobShell.java:54) > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > > > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:79) > > > at org.apache.hadoop.mapred.JobShell.main(JobShell.java:68) > > > > > > > > > This is due to incompability between hadoop0.19 and mahout 0.7 right? > > > so, how can I fix it? > > > I can`t upgrade hadoop 0.19 because it is not up to me, > > > and I don`t want to use mahout 0.5 either because, in that case, I > might > > > have to rewrite my application code. > > > > > > so, is there any way to solve this like through a patch or something? > > > > > > Thanks > > > > > >
