If we have a 640 MB data file and have 3 Data Nodes in a cluster. The file can be split into 10 Blocks and starts the Mappers M1, M2, M3 first. As each one completes the task M4 and so on will be run. It appears like it is not necessary to run all the 10 Map tasks in parallel at once. Just wondering if this is right assumption. What if we have 10 TB of data file with 3 Data Nodes, how to find the number of mappers that will be created. Thanks Sai
- Re: Who splits the file into blocks Rahul Bhattacharjee
- Re: Who splits the file into blocks Harsh J
- Re: fsImage & editsLog questions Sai Sai
- Re: Reduce starts before map completes (at 23%) Sai Sai
- Re: Reduce starts before map completes (at 23... Kai Voigt
- Re: Reduce starts before map completes (at 23... Stephen Boesch
- Re: Reduce starts before map completes (at 23... shanthi k
- Re: Reduce starts before map completes (a... saigraph
- Re: Does a Map task run 3 times on 3 TTs ... Sai Sai
- Re: 10 TB of a data file. Sai Sai
- Re: How to find the num of Mappers Sai Sai
- Re: Will HDFS refer to the memory of Name... Sai Sai
- Re: Will HDFS refer to the memory of Name... Nitin Pawar
- Re: How to find the num of Mappers Nitin Pawar
- Re: 10 TB of a data file. Nitin Pawar
- Re: 100K Maps scenario Sai Sai
- Re: 100K Maps scenario Kai Voigt
- Re: 100K Maps scenario Sai Sai
- Re: Flume port issue Sai Sai
- Re: Flume port issue Lenin Raj
- Re: Flume port issue Sai Sai