It is running on my pseudo-distributed cluster running on my laptop.


________________________________
 From: Lance Norskog <[email protected]>
To: [email protected]; Krishnanand Khambadkone <[email protected]> 
Sent: Wednesday, April 18, 2012 8:09 PM
Subject: Re: Not able to run Wikipedia Bayes Example
 
Is this on a Hadoop cluster, or running in pseudo-distributed mode (no
cluster)?

On Wed, Apr 18, 2012 at 8:00 PM, Krishnanand Khambadkone
<[email protected]> wrote:
> Hi,  I am trying to run the Mahout sample in this link,
>
> https://cwiki.apache.org/MAHOUT/wikipedia-bayes-example.html
>
> When I try to run this step,
>
> $MAHOUT_HOME/bin/mahout  wikipediaDataSetCreator  -i wikipedia/chunks -o 
> wikipediainput -c $MAHOUT_HOME/examples/src/test/resources/country.txt
>
>
> I get the following exception,  I am running it with Mahout distribution from 
> Cloudera (mahout-0.5-cdh3u3)
>
> 12/04/17 18:59:13 INFO mapred.JobClient: Task Id : 
> attempt_201204171311_0005_m_000000_0, Status : FAILED
> attempt_201204171311_0005_m_000000_0: 2012-04-17 18:59:09.221 java[4156:1d03] 
> Unable to load realm info from SCDynamicStore
> 12/04/17 18:59:13 INFO mapred.JobClient: Task Id : 
> attempt_201204171311_0005_m_000001_0, Status : FAILED
> java.lang.ArrayStoreException: [C
> at java.util.AbstractCollection.toArray(AbstractCollection.java:171)
> at 
> org.apache.mahout.analysis.WikipediaAnalyzer.<init>(WikipediaAnalyzer.java:38)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
> at java.lang.Class.newInstance0(Class.java:355)
> at java.lang.Class.newInstance(Class.java:308)
> at 
> org.apache.mahout.classifier.bayes.WikipediaDatasetCreatorMapper.setup(WikipediaDatasetCreatorMapper.java:107)
> at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:142)
> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:647)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:323)
> at org.apache.hadoop.mapred.Child$4.run(Child.java:270)
> at java.security.AccessController.do
> attempt_201204171311_0005_m_000001_0: 2012-04-17 18:59:09.208 java[4160:1d03] 
> Unable to load realm info from SCDynamicStore
> 12/04/17 18:59:18 INFO mapred.JobClient: Task Id : 
> attempt_201204171311_0005_m_000000_1, Status : FAILED
> java.lang.ArrayStoreException: [C
> at java.util.AbstractCollection.toArray(AbstractCollection.java:171)
> at 
> org.apache.mahout.analysis.WikipediaAnalyzer.<init>(WikipediaAnalyzer.java:38)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:513)
> at java.lang.Class.newInstance0(Class.java:355)
> at java.lang.Class.newInstance(Class.java:308)
> at 
> org.apache.mahout.classifier.bayes.WikipediaDatasetCreatorMapper.setup(WikipediaDatasetCreatorMapper.java:107)
> at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:142)
> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:647)
> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:323)
> at org.apache.hadoop.mapred.Child$4.run(Child.java:270)
> at java.security.AccessController.do



-- 
Lance Norskog
[email protected]

Reply via email to