Hi- The bin/mahout program uses its own Java option: MAHOUT_OPTS. http://www.lucidimagination.com/search/?q=mahout_opts
Accd. to this search, it is not on described on the wiki. It is in the bin/mahout script. Lance On Fri, Apr 20, 2012 at 5:48 PM, Krishnanand Khambadkone <[email protected]> wrote: > Lance, I was able to advance further after installing and building the > mahout-trunk. However when I try to run the last step, > > $MAHOUT_HOME/bin/mahout testclassifier -m wikipediamodel -d wikipediainput > > I get this error. My HADOOP_HEAPSIZE is set to 2000MB. I have 8GB on my > mac book pro. > > > Exception in thread "Thread for syncLogs" java.lang.OutOfMemoryError: Java > heap space > > ________________________________ > From: Lance Norskog <[email protected]> > > To: Krishnanand Khambadkone <[email protected]> > Sent: Thursday, April 19, 2012 6:54 PM > Subject: Re: Fw: Not able to run Wikipedia Bayes Example > > Yikes! You're running Mahout 0.5. This is an old release. We generally > suggest that you upgrade to the Mahout trunk. When you run this > program in pseudo-distributed mode, the Cloudera Hadoop code will not > be used- all Hadoop code comes from the Mahout project. > > Lance > > On Wed, Apr 18, 2012 at 10:51 PM, Krishnanand Khambadkone > <[email protected]> wrote: >> Lance, Will this sample run only on a true cluster? I am running it on a >> pseudo-distributed cluster (cloudera u3) on my laptop. >> >> ----- Forwarded Message ----- >> From: Lance Norskog <[email protected]> >> To: [email protected]; Krishnanand Khambadkone >> <[email protected]> >> Sent: Wednesday, April 18, 2012 8:09 PM >> Subject: Re: Not able to run Wikipedia Bayes Example >> >> Is this on a Hadoop cluster, or running in pseudo-distributed mode (no >> cluster)? >> >> On Wed, Apr 18, 2012 at 8:00 PM, Krishnanand Khambadkone >> <[email protected]> wrote: >>> Hi, I am trying to run the Mahout sample in this link, >>> >>> https://cwiki.apache.org/MAHOUT/wikipedia-bayes-example.html >>> >>> When I try to run this step, >>> >>> $MAHOUT_HOME/bin/mahout wikipediaDataSetCreator -i wikipedia/chunks -o >>> wikipediainput -c $MAHOUT_HOME/examples/src/test/resources/country.txt >>> >>> >>> I get the following exception, I am running it with Mahout distribution >>> from Cloudera (mahout-0.5-cdh3u3) >>> >>> 12/04/17 18:59:13 INFO mapred.JobClient: Task Id : >>> attempt_201204171311_0005_m_000000_0, Status : FAILED >>> attempt_201204171311_0005_m_000000_0: 2012-04-17 18:59:09.221 >>> java[4156:1d03] Unable to load realm info from SCDynamicStore >>> 12/04/17 18:59:13 INFO mapred.JobClient: Task Id : >>> attempt_201204171311_0005_m_000001_0, Status : FAILED >>> java.lang.ArrayStoreException: [C >>> at java.util.AbstractCollection.toArray(AbstractCollection.java:171) >>> at >>> >>> org.apache.mahout.analysis.WikipediaAnalyzer.<init>(WikipediaAnalyzer.java:38) >>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) >>> at >>> >>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) >>> at >>> >>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) >>> at java.lang.reflect.Constructor.newInstance(Constructor.java:513) >>> at java.lang.Class.newInstance0(Class.java:355) >>> at java.lang.Class.newInstance(Class.java:308) >>> at >>> >>> org.apache.mahout.classifier.bayes.WikipediaDatasetCreatorMapper.setup(WikipediaDatasetCreatorMapper.java:107) >>> at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:142) >>> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:647) >>> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:323) >>> at org.apache.hadoop.mapred.Child$4.run(Child.java:270) >>> at java.security.AccessController.do >>> attempt_201204171311_0005_m_000001_0: 2012-04-17 18:59:09.208 >>> java[4160:1d03] Unable to load realm info from SCDynamicStore >>> 12/04/17 18:59:18 INFO mapred.JobClient: Task Id : >>> attempt_201204171311_0005_m_000000_1, Status : FAILED >>> java.lang.ArrayStoreException: [C >>> at java.util.AbstractCollection.toArray(AbstractCollection.java:171) >>> at >>> >>> org.apache.mahout.analysis.WikipediaAnalyzer.<init>(WikipediaAnalyzer.java:38) >>> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) >>> at >>> >>> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:39) >>> at >>> >>> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:27) >>> at java.lang.reflect.Constructor.newInstance(Constructor.java:513) >>> at java.lang.Class.newInstance0(Class.java:355) >>> at java.lang.Class.newInstance(Class.java:308) >>> at >>> >>> org.apache.mahout.classifier.bayes.WikipediaDatasetCreatorMapper.setup(WikipediaDatasetCreatorMapper.java:107) >>> at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:142) >>> at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:647) >>> at org.apache.hadoop.mapred.MapTask.run(MapTask.java:323) >>> at org.apache.hadoop.mapred.Child$4.run(Child.java:270) >>> at java.security.AccessController.do >> >> >> >> -- >> Lance Norskog >> [email protected] >> >> > > > > -- > Lance Norskog > [email protected] > > -- Lance Norskog [email protected]
