hadoop-common-user mailing list archives

Site index · List index
Message view « Date » · « Thread »
Top « Date » · « Thread »
From waqas latif <waqas...@gmail.com>
Subject Re: EOFException at org.apache.hadoop.io.SequenceFile$Reader.init(SequenceFile.java:1508)......
Date Sun, 27 May 2012 16:15:32 GMT
But the thing is, it works with hadoop 0.20. even with 100 x100(and even
bigger matrices)  but when it comes to hadoop 1.0.3 then even there is a
problem with 3x3 matrix.

On Sun, May 27, 2012 at 12:00 PM, Prashant Kommireddi
<prash1784@gmail.com>wrote:

> I have seen this issue with large file writes using SequenceFile writer.
> Not found the same issue when testing with writing fairly small files ( <
> 1GB).
>
> On Fri, May 25, 2012 at 10:33 PM, Kasi Subrahmanyam
> <kasisubbu440@gmail.com>wrote:
>
> > Hi,
> > If you are using a custom writable object while passing data from the
> > mapper to the reducer make sure that the read fields and the write has
> the
> > same number of variables. It might be possible that you wrote datavtova
> > file using custom writable but later modified the custom writable (like
> > adding new attribute to the writable) which the old data doesn't have.
> >
> > It might be a possibility is please check once
> >
> > On Friday, May 25, 2012, waqas latif wrote:
> >
> > > Hi Experts,
> > >
> > > I am fairly new to hadoop MapR and I was trying to run a matrix
> > > multiplication example presented by Mr. Norstadt under following link
> > > http://www.norstad.org/matrix-multiply/index.html. I can run it
> > > successfully with hadoop 0.20.2 but I tried to run it with hadoop 1.0.3
> > but
> > > I am getting following error. Is it the problem with my hadoop
> > > configuration or it is compatibility problem in the code which was
> > written
> > > in hadoop 0.20 by author.Also please guide me that how can I fix this
> > error
> > > in either case. Here is the error I am getting.
> > >
> > > in thread "main" java.io.EOFException
> > >        at java.io.DataInputStream.readFully(DataInputStream.java:180)
> > >        at java.io.DataInputStream.readFully(DataInputStream.java:152)
> > >        at
> > > org.apache.hadoop.io.SequenceFile$Reader.init(SequenceFile.java:1508)
> > >        at
> > > org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1486)
> > >        at
> > > org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1475)
> > >        at
> > > org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1470)
> > >        at TestMatrixMultiply.fillMatrix(TestMatrixMultiply.java:60)
> > >        at TestMatrixMultiply.readMatrix(TestMatrixMultiply.java:87)
> > >        at TestMatrixMultiply.checkAnswer(TestMatrixMultiply.java:112)
> > >        at TestMatrixMultiply.runOneTest(TestMatrixMultiply.java:150)
> > >        at TestMatrixMultiply.testRandom(TestMatrixMultiply.java:278)
> > >        at TestMatrixMultiply.main(TestMatrixMultiply.java:308)
> > >        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
> > >        at
> > >
> > >
> >
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
> > >        at
> > >
> > >
> >
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
> > >        at java.lang.reflect.Method.invoke(Method.java:597)
> > >        at org.apache.hadoop.util.RunJar.main(RunJar.java:156)
> > >
> > > Thanks in advance
> > >
> > > Regards,
> > > waqas
> > >
> >
>

Mime
  • Unnamed multipart/alternative (inline, None, 0 bytes)
View raw message