Thanks Owen for the suggestion.  I wonder if there would be side effects
from failing the job from the node consistently? Would job tracker black
list the nodes for other jobs as well?

Htin

-----Original Message-----
From: Owen O'Malley [mailto:[EMAIL PROTECTED] 
Sent: Monday, April 21, 2008 10:53 PM
To: core-user@hadoop.apache.org
Subject: Re: How to instruct Job Tracker to use certain hosts only


On Apr 18, 2008, at 1:52 PM, Htin Hlaing wrote:

> I would like to run the first job to run on all the compute hosts  
> in the
> cluster (which is by default) and then, I would like to run the  
> second job
> with only on  a subset of the hosts (due to some licensing issue).

One option would be to set mapred.map.max.attempts and  
mapred.reduce.max.attempts to larger numbers and have the map or  
reduce fail if it is run on a bad node. When the task re-runs, it  
will run on a different node. Eventually it will find a valid node.

-- Owen

Reply via email to