Along with partitioner, try to plug in a combiner. It would provide significant 
performance gains. Not sure about the algo you use, but might have to tweak 
that a little to facilitate a combiner.

Thanks,
Amogh

-----Original Message-----
From: Chandraprakash Bhagtani [mailto:cpbhagt...@gmail.com] 
Sent: Tuesday, September 29, 2009 12:25 PM
To: common-user@hadoop.apache.org
Cc: core-u...@hadoop.apache.org
Subject: Re: Best Idea to deal with following situation

you can write your custom partitioner instead of hash partitioner

On Sat, Sep 26, 2009 at 6:18 AM, Pankil Doshi <forpan...@gmail.com> wrote:

> Hello everyone,
>
> I have job whose result has  only 5 keys but but each key has long list of
> values like in 100000's .
> What should be best way to deal with it. I feel few of my reducers get over
> loaded as two or more keys go to same reduce and hence they have lots of
> work to do.
>
> So what should be best way out with this situation?
>
> Pankil
>



-- 
Thanks & Regards,
Chandra Prakash Bhagtani,

Reply via email to