Are you running on Windows? There are quirks about closing and opening files in read/write mode. You may be getting a sequencing error.
On Wed, Nov 30, 2011 at 9:09 AM, J. Rottinghuis <jrottingh...@gmail.com> wrote: > The error is that you cannot open /tmp/MatrixMultiply/out/_logs > Does the directory exist? > Do you have proper access rights set? > > Joep > > On Wed, Nov 30, 2011 at 3:23 AM, ChWaqas <waqas...@gmail.com> wrote: > >> >> Hi I am trying to run the matrix multiplication example mentioned(with >> source >> code) on the following link: >> >> http://www.norstad.org/matrix-multiply/index.html >> >> I have hadoop setup in pseudodistributed mode and I configured it using >> this >> tutorial: >> >> >> http://hadoop-tutorial.blogspot.com/2010/11/running-hadoop-in-pseudo-distributed.html?showComment=1321528406255#c3661776111033973764 >> >> When I run my jar file then I get the following error: >> >> Identity test >> 11/11/30 10:37:34 INFO input.FileInputFormat: Total input paths to process >> : >> 2 >> 11/11/30 10:37:34 INFO mapred.JobClient: Running job: job_201111291041_0010 >> 11/11/30 10:37:35 INFO mapred.JobClient: map 0% reduce 0% >> 11/11/30 10:37:44 INFO mapred.JobClient: map 100% reduce 0% >> 11/11/30 10:37:56 INFO mapred.JobClient: map 100% reduce 100% >> 11/11/30 10:37:58 INFO mapred.JobClient: Job complete: >> job_201111291041_0010 >> 11/11/30 10:37:58 INFO mapred.JobClient: Counters: 17 >> 11/11/30 10:37:58 INFO mapred.JobClient: Job Counters >> 11/11/30 10:37:58 INFO mapred.JobClient: Launched reduce tasks=1 >> 11/11/30 10:37:58 INFO mapred.JobClient: Launched map tasks=2 >> 11/11/30 10:37:58 INFO mapred.JobClient: Data-local map tasks=2 >> 11/11/30 10:37:58 INFO mapred.JobClient: FileSystemCounters >> 11/11/30 10:37:58 INFO mapred.JobClient: FILE_BYTES_READ=114 >> 11/11/30 10:37:58 INFO mapred.JobClient: HDFS_BYTES_READ=248 >> 11/11/30 10:37:58 INFO mapred.JobClient: FILE_BYTES_WRITTEN=298 >> 11/11/30 10:37:58 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=124 >> 11/11/30 10:37:58 INFO mapred.JobClient: Map-Reduce Framework >> 11/11/30 10:37:58 INFO mapred.JobClient: Reduce input groups=2 >> 11/11/30 10:37:58 INFO mapred.JobClient: Combine output records=0 >> 11/11/30 10:37:58 INFO mapred.JobClient: Map input records=4 >> 11/11/30 10:37:58 INFO mapred.JobClient: Reduce shuffle bytes=60 >> 11/11/30 10:37:58 INFO mapred.JobClient: Reduce output records=2 >> 11/11/30 10:37:58 INFO mapred.JobClient: Spilled Records=8 >> 11/11/30 10:37:58 INFO mapred.JobClient: Map output bytes=100 >> 11/11/30 10:37:58 INFO mapred.JobClient: Combine input records=0 >> 11/11/30 10:37:58 INFO mapred.JobClient: Map output records=4 >> 11/11/30 10:37:58 INFO mapred.JobClient: Reduce input records=4 >> 11/11/30 10:37:58 INFO input.FileInputFormat: Total input paths to process >> : >> 1 >> 11/11/30 10:37:59 INFO mapred.JobClient: Running job: job_201111291041_0011 >> 11/11/30 10:38:00 INFO mapred.JobClient: map 0% reduce 0% >> 11/11/30 10:38:09 INFO mapred.JobClient: map 100% reduce 0% >> 11/11/30 10:38:21 INFO mapred.JobClient: map 100% reduce 100% >> 11/11/30 10:38:23 INFO mapred.JobClient: Job complete: >> job_201111291041_0011 >> 11/11/30 10:38:23 INFO mapred.JobClient: Counters: 17 >> 11/11/30 10:38:23 INFO mapred.JobClient: Job Counters >> 11/11/30 10:38:23 INFO mapred.JobClient: Launched reduce tasks=1 >> 11/11/30 10:38:23 INFO mapred.JobClient: Launched map tasks=1 >> 11/11/30 10:38:23 INFO mapred.JobClient: Data-local map tasks=1 >> 11/11/30 10:38:23 INFO mapred.JobClient: FileSystemCounters >> 11/11/30 10:38:23 INFO mapred.JobClient: FILE_BYTES_READ=34 >> 11/11/30 10:38:23 INFO mapred.JobClient: HDFS_BYTES_READ=124 >> 11/11/30 10:38:23 INFO mapred.JobClient: FILE_BYTES_WRITTEN=100 >> 11/11/30 10:38:23 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=124 >> 11/11/30 10:38:23 INFO mapred.JobClient: Map-Reduce Framework >> 11/11/30 10:38:23 INFO mapred.JobClient: Reduce input groups=2 >> 11/11/30 10:38:23 INFO mapred.JobClient: Combine output records=2 >> 11/11/30 10:38:23 INFO mapred.JobClient: Map input records=2 >> 11/11/30 10:38:23 INFO mapred.JobClient: Reduce shuffle bytes=0 >> 11/11/30 10:38:23 INFO mapred.JobClient: Reduce output records=2 >> 11/11/30 10:38:23 INFO mapred.JobClient: Spilled Records=4 >> 11/11/30 10:38:23 INFO mapred.JobClient: Map output bytes=24 >> 11/11/30 10:38:23 INFO mapred.JobClient: Combine input records=2 >> 11/11/30 10:38:23 INFO mapred.JobClient: Map output records=2 >> 11/11/30 10:38:23 INFO mapred.JobClient: Reduce input records=2 >> Exception in thread "main" java.io.IOException: Cannot open filename >> /tmp/MatrixMultiply/out/_logs >> at >> org.apache.hadoop.hdfs.DFSClient$DFSInputStream.openInfo(DFSClient.ja >> va:1497) >> at >> org.apache.hadoop.hdfs.DFSClient$DFSInputStream.<init>(DFSClient.java >> :1488) >> at org.apache.hadoop.hdfs.DFSClient.open(DFSClient.java:376) >> at >> org.apache.hadoop.hdfs.DistributedFileSystem.open(DistributedFileSyst >> em.java:178) >> at >> org.apache.hadoop.io.SequenceFile$Reader.openFile(SequenceFile.java:1 437) >> at >> org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:142 4) >> at >> org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:141 7) >> at >> org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:141 2) >> at TestMatrixMultiply.fillMatrix(TestMatrixMultiply.java:62) >> at TestMatrixMultiply.readMatrix(TestMatrixMultiply.java:84) >> at TestMatrixMultiply.checkAnswer(TestMatrixMultiply.java:108) >> at TestMatrixMultiply.runOneTest(TestMatrixMultiply.java:144) >> at TestMatrixMultiply.testIdentity(TestMatrixMultiply.java:156) >> at TestMatrixMultiply.main(TestMatrixMultiply.java:258) >> at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) >> at >> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl. >> java:57) >> at >> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAcces >> sorImpl.java:43) >> at java.lang.reflect.Method.invoke(Method.java:601) >> at org.apache.hadoop.util.RunJar.main(RunJar.java:156) >> >> Can someone please suggest me that what I am doing wrong >> >> Ch. >> -- >> View this message in context: >> http://old.nabble.com/mapreduce-matrix-multiplication-on-hadoop-tp32883909p32883909.html >> Sent from the Hadoop core-user mailing list archive at Nabble.com. >> >> -- Lance Norskog goks...@gmail.com