Hi Users,
Hadoop can distribute all the data into HDFS inside MapReduce
tasks can work together. which one is goes to which data node and how it
works all those things it can maintain each task has own JVM in each data
node. JVM can handle hell number of data to process to the all
genrally in hadoop map function will be exeucted by all the data nodes on
the input data set ,against this how can i do the following.
i have some filter programs , and what i want to do is each data
node(slave) has to execute one filter alogrithm simultaneously, diffent
from other data nodes
Not sure what you are trying to do...
You want to pass through the entire data set on all nodes where each node runs
a single filter?
You're thinking is orthogonal to how Hadoop works.
You would be better off letting each node work on a portion of the data which
is local to that node
You can check the value of map.input.file in order to apply a different
logic for each type of files (in the mapper).
More information about your problem/context would help the readers to
provide a more extensive reply.
Regards
Bertrand
On Mon, Sep 3, 2012 at 6:25 PM, Michael Segel
Hi,
Assuming you have to compute these value for every RGB pixel.
Why couldn't you compute all these values at the same time on the same node?
Hadoop let you distribute your computation but it doesn't mean each node
has to compute only a part of the equations.
Each node can compute all equations