I solved my problem, I didnt split the right file.
2014-08-07 10:28 GMT+02:00 xenlee - Zerg <[email protected]>: > Hi, > > I am following the Apache Mahout Cookbook tutorials and I tried to run a > classifier on 20newsGroup.I managed to convert the files in SF then run the > TF-IDF algorithm, and split the Data into Train/test. But when I finally > build my model with trainnb, here is the error I got. > Did this already occur to someone? > > Regards, > xenlee - > > > [mapr@fb-mapr1 new]$ mahout trainnb -i /input/new/20news-train-vectors > -el -o /input/new/model -li /input/new/labelindex -ow > No MAHOUT_CONF_DIR found > MAHOUT_LOCAL is not set; adding HADOOP_CONF_DIR to classpath. > Running on hadoop, using /opt/mapr/hadoop/hadoop-0.20.2/bin/hadoop and > HADOOP_CONF_DIR=/opt/mapr/hadoop/hadoop-0.20.2/conf > MAHOUT-JOB: /opt/mapr/mahout/mahout-0.9/mahout-examples-0.9-mapr-job.jar > 14/08/07 08:22:44 WARN driver.MahoutDriver: No trainnb.props found on > classpath, will use command-line arguments only > 14/08/07 08:22:44 INFO common.AbstractJob: Command line arguments: > {--alphaI=[1.0], --endPhase=[2147483647], --extractLabels=null, > --input=[/input/new/20news-train-vectors], > --labelIndex=[/input/new/labelindex], --output=[/input/new/model], > --overwrite=null, --startPhase=[0], --tempDir=[temp]} > 14/08/07 08:22:44 INFO zlib.ZlibFactory: Successfully loaded & initialized > native-zlib library > 14/08/07 08:22:44 INFO compress.CodecPool: Got brand-new decompressor > Exception in thread "main" java.lang.ArrayIndexOutOfBoundsException: 1 > at > org.apache.mahout.classifier.naivebayes.BayesUtils.writeLabelIndex(BayesUtils.java:123) > at > org.apache.mahout.classifier.naivebayes.training.TrainNaiveBayesJob.createLabelIndex(TrainNaiveBayesJob.java:180) > at > org.apache.mahout.classifier.naivebayes.training.TrainNaiveBayesJob.run(TrainNaiveBayesJob.java:94) > at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:65) > at > org.apache.mahout.classifier.naivebayes.training.TrainNaiveBayesJob.main(TrainNaiveBayesJob.java:64) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at > org.apache.hadoop.util.ProgramDriver$ProgramDescription.invoke(ProgramDriver.java:68) > at > org.apache.hadoop.util.ProgramDriver.driver(ProgramDriver.java:139) > at > org.apache.mahout.driver.MahoutDriver.main(MahoutDriver.java:195) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:606) > at org.apache.hadoop.util.RunJar.main(RunJar.java:197) >
