Felix, You can use ChainMapper and ChainReducer to create jobs of the form M+RM*. Is that what you're looking for? I'm not aware of anything that allows you to have multiple reduce functions without the job "barrier".
Ed On Wed, Feb 3, 2010 at 9:41 PM, Felix Halim <felix.ha...@gmail.com> wrote: > Hi all, > > As far as I know, a barrier exists between map and reduce function in > one round of MR. There is another barrier for the reducer to end the > job for that round. However if we want to run in several rounds using > the same map and reduce functions, then the barrier between reduce and > the map of the next round is NOT necessary, right? Since the reducer > only output a single value for each key. This reducer may as well run > a map task for the next round immediately rather than waiting for all > reducer to finish. This way, the utilization of the machines between > rounds can be improved. > > Is there a setting in Hadoop to do that? > > Felix Halim >