Re: Fast way to read thousands of double value in hadoop jobs

2016-08-19 Thread Daniel Haviv
That was the idea :) Thanks for the update On Friday, 19 August 2016, Madhav Sharan wrote: > Thanks for your suggestion Daniel. I was already using SequenceFile but my > format was poor. I was storing file contents as Text in my SeqFile, > > So all my map jobs did repeated

Re: Fast way to read thousands of double value in hadoop jobs

2016-08-18 Thread Madhav Sharan
Thanks for your suggestion Daniel. I was already using SequenceFile but my format was poor. I was storing file contents as Text in my SeqFile, So all my map jobs did repeated conversion from Text to double. I resolved this by correcting SequenceFile format. Now I store serialised java object in

Re: Fast way to read thousands of double value in hadoop jobs

2016-08-18 Thread Daniel Haviv
Store them within a sequencefile On Thursday, 18 August 2016, Madhav Sharan wrote: > Hi , can someone please recommend a fast way in hadoop to store and > retrieve matrix of double values? > > As of now we store values in text files and the read it in java using HDFS >

Fast way to read thousands of double value in hadoop jobs

2016-08-17 Thread Madhav Sharan
Hi , can someone please recommend a fast way in hadoop to store and retrieve matrix of double values? As of now we store values in text files and the read it in java using HDFS inputstream and Scanner. *[0]* These files are actually vectors representing a video file. Each vector is 883 X 200 and