But the thing is, it works with hadoop 0.20. even with 100 x100(and even bigger matrices) but when it comes to hadoop 1.0.3 then even there is a problem with 3x3 matrix.
On Sun, May 27, 2012 at 12:00 PM, Prashant Kommireddi <prash1...@gmail.com>wrote: > I have seen this issue with large file writes using SequenceFile writer. > Not found the same issue when testing with writing fairly small files ( < > 1GB). > > On Fri, May 25, 2012 at 10:33 PM, Kasi Subrahmanyam > <kasisubbu...@gmail.com>wrote: > > > Hi, > > If you are using a custom writable object while passing data from the > > mapper to the reducer make sure that the read fields and the write has > the > > same number of variables. It might be possible that you wrote datavtova > > file using custom writable but later modified the custom writable (like > > adding new attribute to the writable) which the old data doesn't have. > > > > It might be a possibility is please check once > > > > On Friday, May 25, 2012, waqas latif wrote: > > > > > Hi Experts, > > > > > > I am fairly new to hadoop MapR and I was trying to run a matrix > > > multiplication example presented by Mr. Norstadt under following link > > > http://www.norstad.org/matrix-multiply/index.html. I can run it > > > successfully with hadoop 0.20.2 but I tried to run it with hadoop 1.0.3 > > but > > > I am getting following error. Is it the problem with my hadoop > > > configuration or it is compatibility problem in the code which was > > written > > > in hadoop 0.20 by author.Also please guide me that how can I fix this > > error > > > in either case. Here is the error I am getting. > > > > > > in thread "main" java.io.EOFException > > > at java.io.DataInputStream.readFully(DataInputStream.java:180) > > > at java.io.DataInputStream.readFully(DataInputStream.java:152) > > > at > > > org.apache.hadoop.io.SequenceFile$Reader.init(SequenceFile.java:1508) > > > at > > > org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1486) > > > at > > > org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1475) > > > at > > > org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1470) > > > at TestMatrixMultiply.fillMatrix(TestMatrixMultiply.java:60) > > > at TestMatrixMultiply.readMatrix(TestMatrixMultiply.java:87) > > > at TestMatrixMultiply.checkAnswer(TestMatrixMultiply.java:112) > > > at TestMatrixMultiply.runOneTest(TestMatrixMultiply.java:150) > > > at TestMatrixMultiply.testRandom(TestMatrixMultiply.java:278) > > > at TestMatrixMultiply.main(TestMatrixMultiply.java:308) > > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > > > at > > > > > > > > > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > > > at > > > > > > > > > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > > > at java.lang.reflect.Method.invoke(Method.java:597) > > > at org.apache.hadoop.util.RunJar.main(RunJar.java:156) > > > > > > Thanks in advance > > > > > > Regards, > > > waqas > > > > > >