It is running on my pseudo-distributed cluster running on my laptop.
________________________________ From: Lance Norskog <[email protected]> To: [email protected]; Krishnanand Khambadkone <[email protected]> Sent: Wednesday, April 18, 2012 8:09 PM Subject: Re: Not able to run Wikipedia Bayes Example Is this on a Hadoop cluster, or running in pseudo-distributed mode (no cluster)? On Wed, Apr 18, 2012 at 8:00 PM, Krishnanand Khambadkone <[email protected]> wrote: > Hi, I am trying to run the Mahout sample in this link, > > https://cwiki.apache.org/MAHOUT/wikipedia-bayes-example.html > > When I try to run this step, > > $MAHOUT_HOME/bin/mahout wikipediaDataSetCreator -i wikipedia/chunks -o > wikipediainput -c $MAHOUT_HOME/examples/src/test/resources/country.txt > > > I get the following exception, I am running it with Mahout distribution from > Cloudera (mahout-0.5-cdh3u3) > > 12/04/17 18:59:13 INFO mapred.JobClient: Task Id : > attempt_201204171311_0005_m_000000_0, Status : FAILED > attempt_201204171311_0005_m_000000_0: 2012-04-17 18:59:09.221 java[4156:1d03] > Unable to load realm info from SCDynamicStore > 12/04/17 18:59:13 INFO mapred.JobClient: Task Id : > attempt_201204171311_0005_m_000001_0, Status : FAILED > java.lang.ArrayStoreException: [C > at java.util.AbstractCollection.toArray(AbstractCollection.java:171) > at > org.apache.mahout.analysis.WikipediaAnalyzer.<init>(WikipediaAnalyzer.java:38) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) > at java.lang.reflect.Constructor.newInstance(Constructor.java:513) > at java.lang.Class.newInstance0(Class.java:355) > at java.lang.Class.newInstance(Class.java:308) > at > org.apache.mahout.classifier.bayes.WikipediaDatasetCreatorMapper.setup(WikipediaDatasetCreatorMapper.java:107) > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:142) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:647) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:323) > at org.apache.hadoop.mapred.Child$4.run(Child.java:270) > at java.security.AccessController.do > attempt_201204171311_0005_m_000001_0: 2012-04-17 18:59:09.208 java[4160:1d03] > Unable to load realm info from SCDynamicStore > 12/04/17 18:59:18 INFO mapred.JobClient: Task Id : > attempt_201204171311_0005_m_000000_1, Status : FAILED > java.lang.ArrayStoreException: [C > at java.util.AbstractCollection.toArray(AbstractCollection.java:171) > at > org.apache.mahout.analysis.WikipediaAnalyzer.<init>(WikipediaAnalyzer.java:38) > at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) > at > sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) > at > sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) > at java.lang.reflect.Constructor.newInstance(Constructor.java:513) > at java.lang.Class.newInstance0(Class.java:355) > at java.lang.Class.newInstance(Class.java:308) > at > org.apache.mahout.classifier.bayes.WikipediaDatasetCreatorMapper.setup(WikipediaDatasetCreatorMapper.java:107) > at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:142) > at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:647) > at org.apache.hadoop.mapred.MapTask.run(MapTask.java:323) > at org.apache.hadoop.mapred.Child$4.run(Child.java:270) > at java.security.AccessController.do -- Lance Norskog [email protected]
