Do not worry about this.

My problem is just run an algorithm on all nodes in a grid. So I realized,
hadoop does not serve for this purpose and I am already studying a
alternative. If you have some suggestion I will be grateful.


Luiz Carlos Melo Muniz




2012/3/29 Harsh J <ha...@cloudera.com>

> Luiz,
>
> Though it is possible to 'hint' this by tweaking the InputSplits
> passed from the job, the default schedulers of Hadoop do not make any
> such guarantees and hence this isn't possible unless you write your
> own complete scheduler, an exercise that wouldn't suit production
> deployments unless you also test your scheduler intensively for other
> types of workloads.
>
> Why do you even need such a thing? For processing purposes or
> otherwise? I'm hoping its not a monitoring sort of hack you're trying
> to do.
>
> On Thu, Mar 29, 2012 at 5:55 AM, Luiz Carlos Muniz <lcmu...@gmail.com>
> wrote:
> > Hi,
> >
> > Is there any way to ensure the execution of a map on all nodes of a
> > clusterin a way that each node run the map once and only once. That is, I
> > would use Hadoop to execute a method on all nodes in the cluster. Without
> > the possibility that the method execute twice in the same node even if
> > another node fails.
> >
> > I already set mapred.tasktracker.map.tasks.maximum  to 1 and
> > mapred.max.jobs.per.node to 1 but still, if a node fails, another node
> that
> > has
> > carried out a map before run the map again to meet the absence of which
> > failed.
> >
> > Luiz Carlos Melo Muniz
> >
> > Luiz Carlos Melo Muniz
> >
> >
>
>
>
> --
> Harsh J
>

Reply via email to