Re: Anatomy of read in hdfs

2017-04-09 Thread daemeon reiydelle
Readers ARE parallel processes, one per map task. There are defaults in map phase, about how many readers there are for the input file(s). Default is one mapper task block (or file, where any file is smaller than the hdfs block size). There is no java framework per se for splitting up an file

Re: Anatomy of read in hdfs

2017-04-09 Thread Mohammad Tariq
Hi Sidharth, I'm sorry I didn't quite get the first part your question. What do you mean by real time? Could you please elaborate it a bit? That'll help me answering your question in a better manner. And for your second question, This is how write happens - Suppose your file resides in your

Re: Anatomy of read in hdfs

2017-04-09 Thread Sidharth Kumar
Thanks Tariq, It really helped me to understand but just one another doubt that if reading is not a parallel process then to ready a file of 100GB and hdfs block size is 128MB. It will take lot much to read the complete file but it's not the scenerio in the real time. And second question is write