Thanks Harsh. I got it running. On Wed, May 30, 2012 at 5:58 PM, Harsh J <ha...@cloudera.com> wrote:
> When your code does a listStatus, you can pass a PathFilter object > along that can do this filtering for you. See > > http://hadoop.apache.org/common/docs/current/api/org/apache/hadoop/fs/FileSystem.html#listStatus(org.apache.hadoop.fs.Path,%20org.apache.hadoop.fs.PathFilter) > for the API javadocs on that. > > On Wed, May 30, 2012 at 7:46 PM, waqas latif <waqas...@gmail.com> wrote: > > I got the problem with I am unable to solve it. I need to apply a filter > > for _SUCCESS file while using FileSystem.listStatus method. Can someone > > please guide me how to filter _SUCCESS files. Thanks > > > > On Tue, May 29, 2012 at 1:42 PM, waqas latif <waqas...@gmail.com> wrote: > > > >> So my question is that do hadoop 0.20 and 1.0.3 differ in their support > of > >> writing or reading sequencefiles? same code works fine with hadoop 0.20 > but > >> problem occurs when run it under hadoop 1.0.3. > >> > >> > >> On Sun, May 27, 2012 at 6:15 PM, waqas latif <waqas...@gmail.com> > wrote: > >> > >>> But the thing is, it works with hadoop 0.20. even with 100 x100(and > even > >>> bigger matrices) but when it comes to hadoop 1.0.3 then even there is > a > >>> problem with 3x3 matrix. > >>> > >>> > >>> On Sun, May 27, 2012 at 12:00 PM, Prashant Kommireddi < > >>> prash1...@gmail.com> wrote: > >>> > >>>> I have seen this issue with large file writes using SequenceFile > writer. > >>>> Not found the same issue when testing with writing fairly small files > ( < > >>>> 1GB). > >>>> > >>>> On Fri, May 25, 2012 at 10:33 PM, Kasi Subrahmanyam > >>>> <kasisubbu...@gmail.com>wrote: > >>>> > >>>> > Hi, > >>>> > If you are using a custom writable object while passing data from > the > >>>> > mapper to the reducer make sure that the read fields and the write > has > >>>> the > >>>> > same number of variables. It might be possible that you wrote > datavtova > >>>> > file using custom writable but later modified the custom writable > (like > >>>> > adding new attribute to the writable) which the old data doesn't > have. > >>>> > > >>>> > It might be a possibility is please check once > >>>> > > >>>> > On Friday, May 25, 2012, waqas latif wrote: > >>>> > > >>>> > > Hi Experts, > >>>> > > > >>>> > > I am fairly new to hadoop MapR and I was trying to run a matrix > >>>> > > multiplication example presented by Mr. Norstadt under following > link > >>>> > > http://www.norstad.org/matrix-multiply/index.html. I can run it > >>>> > > successfully with hadoop 0.20.2 but I tried to run it with hadoop > >>>> 1.0.3 > >>>> > but > >>>> > > I am getting following error. Is it the problem with my hadoop > >>>> > > configuration or it is compatibility problem in the code which was > >>>> > written > >>>> > > in hadoop 0.20 by author.Also please guide me that how can I fix > this > >>>> > error > >>>> > > in either case. Here is the error I am getting. > >>>> > > > >>>> > > in thread "main" java.io.EOFException > >>>> > > at > java.io.DataInputStream.readFully(DataInputStream.java:180) > >>>> > > at > java.io.DataInputStream.readFully(DataInputStream.java:152) > >>>> > > at > >>>> > > > org.apache.hadoop.io.SequenceFile$Reader.init(SequenceFile.java:1508) > >>>> > > at > >>>> > > > >>>> > org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1486) > >>>> > > at > >>>> > > > >>>> > org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1475) > >>>> > > at > >>>> > > > >>>> > org.apache.hadoop.io.SequenceFile$Reader.<init>(SequenceFile.java:1470) > >>>> > > at > TestMatrixMultiply.fillMatrix(TestMatrixMultiply.java:60) > >>>> > > at > TestMatrixMultiply.readMatrix(TestMatrixMultiply.java:87) > >>>> > > at > TestMatrixMultiply.checkAnswer(TestMatrixMultiply.java:112) > >>>> > > at > TestMatrixMultiply.runOneTest(TestMatrixMultiply.java:150) > >>>> > > at > TestMatrixMultiply.testRandom(TestMatrixMultiply.java:278) > >>>> > > at TestMatrixMultiply.main(TestMatrixMultiply.java:308) > >>>> > > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native > Method) > >>>> > > at > >>>> > > > >>>> > > > >>>> > > >>>> > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39) > >>>> > > at > >>>> > > > >>>> > > > >>>> > > >>>> > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25) > >>>> > > at java.lang.reflect.Method.invoke(Method.java:597) > >>>> > > at org.apache.hadoop.util.RunJar.main(RunJar.java:156) > >>>> > > > >>>> > > Thanks in advance > >>>> > > > >>>> > > Regards, > >>>> > > waqas > >>>> > > > >>>> > > >>>> > >>> > >>> > >> > > > > -- > Harsh J >