The evaluations file that is being imported is invalid. What's your command line? are you pointing at the right files?
On Thu, Apr 28, 2011 at 12:20 PM, Jose Fuentes De Frutos <[email protected]>wrote: > Hello, > > I'm trying to execute a genetic algorithm with the Mahout distributed mode. > It works well in my local computer, but when I try to execute it in a > distributed environment something goes wrong. After finish the first > map-reduce iteration the following exception is throwed: > > 11/04/28 10:48:20 INFO mapreduce.Job: map 100% reduce 100% > 11/04/28 10:48:22 INFO mapreduce.Job: Job complete: job_201104281044_0001 > 11/04/28 10:48:22 INFO mapreduce.Job: Counters: 33 > FileInputFormatCounters > BYTES_READ=2004 > FileSystemCounters > FILE_BYTES_READ=222 > FILE_BYTES_WRITTEN=476 > HDFS_BYTES_READ=2155 > HDFS_BYTES_WRITTEN=384 > > Shuffle Errors > BAD_ID=0 > CONNECTION=0 > IO_ERROR=0 > WRONG_LENGTH=0 > WRONG_MAP=0 > WRONG_REDUCE=0 > Job Counters > Data-local map tasks=1 > Total time spent by all maps waiting after reserving slots (ms)=0 > Total time spent by all reduces waiting after reserving slots (ms)=0 > SLOTS_MILLIS_MAPS=3232 > SLOTS_MILLIS_REDUCES=3376 > Launched map tasks=1 > Launched reduce tasks=1 > Map-Reduce Framework > Combine input records=0 > Combine output records=0 > Failed Shuffles=0 > GC time elapsed (ms)=47 > Map input records=12 > Map output bytes=192 > Map output records=12 > Merged Map outputs=1 > Reduce input groups=12 > Reduce input records=12 > Reduce output records=12 > Reduce shuffle bytes=222 > Shuffled Maps =1 > Spilled Records=24 > SPLIT_RAW_BYTES=151 > 11/04/28 10:48:22 WARN conf.Configuration: io.sort.mb is deprecated. > Instead, use mapreduce.task.io.sort.mb > 11/04/28 10:48:22 WARN conf.Configuration: io.sort.factor is deprecated. > Instead, use mapreduce.task.io.sort.factor > Exception in thread "main" java.io.EOFException > at java.io.DataInputStream.readFully(DataInputStream.java:197) > at java.io.DataInputStream.readFully(DataInputStream.java:169) > at org.apache.hadoop.io.SequenceFile$Reader.init(SequenceFile.java:1518) > at org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1483) > at org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1366) > at > org.apache.hadoop.io.SequenceFile$Sorter$SegmentDescriptor.nextRawKey(SequenceFile.java:3163) > at > org.apache.hadoop.io.SequenceFile$Sorter$MergeQueue.merge(SequenceFile.java:2977) > at org.apache.hadoop.io.SequenceFile$Sorter.merge(SequenceFile.java:2706) > at org.apache.hadoop.io.SequenceFile$Sorter.merge(SequenceFile.java:2677) > at org.apache.hadoop.io.SequenceFile$Sorter.merge(SequenceFile.java:2791) > at > org.apache.mahout.ga.watchmaker.OutputUtils.importEvaluations(OutputUtils.java:81) > at > org.apache.mahout.ga.watchmaker.MahoutEvaluator.evaluate(MahoutEvaluator.java:79) > at es.udc.tic.world.Poblacion.actualizarFitness(Poblacion.java:66) > at Principal.main(Principal.java:42) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:613) > at org.apache.hadoop.util.RunJar.main(RunJar.java:192) > > Does anyone know what the problem is? > > Thanks in advance. >
