Sorry, 3200m On Tue, Jan 25, 2011 at 12:40 AM, Ted Dunning <[email protected]> wrote:
> 3200? > > or 3200m? > > > On Mon, Jan 24, 2011 at 6:54 PM, Jia Rao <[email protected]> wrote: > > > Hi all, > > > > I am having a problem running the 20 newsgroup example in a hadoop > cluster. > > The trainclassifier worked fine but I got "out of memory java heap" > problem > > in the testclassifier. > > > > The following is the configuration of the hadoop cluster. > > > > Physical machines: 4 nodes, each with 6GB memory. > > > > Hadoop: 0.20.2, HADOOP_HEAP_SIZE=3200 in hadoop-env.sh, > > mapred.child.java.opts=-Xmx1024M in mapred-site.xml. > > > > mahout: tried release 0.4 and the latest source, same problem. > > > > Command line arguments used: > > > > $MAHOUT_HOME/bin/mahout testclassifier \ > > -m newsmodel \ > > -d 20news-input \ > > -type bayes \ > > -ng 3 \ > > -source hdfs \ > > -method mapreduce > > > > > > Any ideas ? > > Thanks ! > > >
