Well, The reducers normally will take much longer than the mappers stage, 
because the copy/shuffle/sort all happened at this time, and they are the hard 
part.
But before we simply say it is part of life, you need to dig into more of your 
MR jobs to find out if you can make it faster.
You are the person most familiar with your data, and you wrote the code to 
group/partition them, and send them to the reducers. Even you set up 255 
reducers, the question is, do each of them get its fair share?You need to read 
the COUNTER information of each reducer, and found out how many reducer groups 
each reducer gets, and how many input bytes it get, etc.
Simple example, if you send 200G data, and group them by DATE, if all the data 
belongs to 2 days, and one of them contains 90% of data, then in this case, 
giving 255 reducers won't help, as only 2 reducers will consume data, and one 
of them will consume 90% of data, and will finish in a very long time, which 
WILL delay the whole MR job, while the rest reducers will finish within 
seconds. In this case, maybe you need to rethink what should be your key, and 
make sure each reducer get its fair share of volume of data.
After the above fix (in fact, normally it will fix 90% of reducer performance 
problems, especially you have 255 reducer tasks available, so each one average 
will only get 1G data, good for your huge cluster only needs to process 256G 
data :-), if you want to make it even faster, then check you code. Do you have 
to use String.compareTo()? Is it slow?  Google hadoop rawcomparator to see if 
you can do something here.
After that, if you still think the reducer stage slow, check you cluster 
system. Does the reducer spend most time on copy stage, or sort, or in your 
reducer class? Find out the where the time spends, then identify the solution.
Yong

Date: Fri, 30 Aug 2013 11:02:05 -0400
Subject: Re: secondary sort - number of reducers
From: adeelmahm...@gmail.com
To: user@hadoop.apache.org



my secondary sort on multiple keys seem to work fine with smaller data sets but 
with bigger data sets (like 256 gig and 800M+ records) the mapper phase gets 
done pretty quick (about 15 mins) but then the reducer phase seem to take 
forever. I am using 255 reducers.

basic idea is that my composite key has both group and sort keys in it which i 
parse in the appropriate comparator classes to perform grouping and sorting .. 
my thinking is that mappers is where most of the work is done 
1. mapper itself (create composite key and value)2. recods sorting3. 
partiotioner
if all this gets done in 15 mins then reducer has the simple task of1. grouping 
comparator
2. reducer itself (simply output records)
should take less time than mappers .. instead it essentially gets stuck in 
reduce phase .. im gonna paste my code here to see if anything stands out as a 
fundamental design issue

//////PARTITIONERpublic int getPartition(Text key, HCatRecord record, int 
numReduceTasks) {             //extract the group key from composite key
                String groupKey = key.toString().split("\\|")[0];               
                return (groupKey.hashCode() & Integer.MAX_VALUE) % 
numReduceTasks;
        }

////////////GROUP COMAPRATORpublic int compare(WritableComparable a, 
WritableComparable b) {            //compare to text objects
                String thisGroupKey = ((Text) a).toString().split("\\|")[0];    
        String otherGroupKey = ((Text) b).toString().split("\\|")[0];
                //extract               return 
thisGroupKey.compareTo(otherGroupKey);   }


////////////SORT COMPARATOR is similar to group comparator and is in map phase 
and gets done quick 


//////////REDUCER
public void reduce(Text key, Iterable<HCatRecord> records, Context context) 
throws IOException, InterruptedException {          log.info("in reducer for 
key " + key.toString());
                Iterator<HCatRecord> recordsIter = records.iterator();          
//we are only interested in the first record after sorting and grouping
                if(recordsIter.hasNext()){                      HCatRecord rec 
= recordsIter.next();                    context.write(nw, rec);
                        log.info("returned record >> " + rec.toString());       
        }       }


On Fri, Aug 30, 2013 at 9:24 AM, Adeel Qureshi <adeelmahm...@gmail.com> wrote:

yup it was negative and by doing this now it seems to be working fine


On Fri, Aug 30, 2013 at 3:09 AM, Shekhar Sharma <shekhar2...@gmail.com> wrote:


Is the hash code of that key  is negative.?

Do something like this



return groupKey.hashCode() & Integer.MAX_VALUE % numParts;



Regards,

Som Shekhar Sharma

+91-8197243810





On Fri, Aug 30, 2013 at 6:25 AM, Adeel Qureshi <adeelmahm...@gmail.com> wrote:

> okay so when i specify the number of reducers e.g. in my example i m using 4

> (for a much smaller data set) it works if I use a single column in my

> composite key .. but if I add multiple columns in the composite key

> separated by a delimi .. it then throws the illegal partition error (keys

> before the pipe are group keys and after the pipe are the sort keys and my

> partioner only uses the group keys

>

> java.io.IOException: Illegal partition for Atlanta:GA|Atlanta:GA:1:Adeel

> (-1)

>         at

> org.apache.hadoop.mapred.MapTask$MapOutputBuffer.collect(MapTask.java:1073)

>         at

> org.apache.hadoop.mapred.MapTask$NewOutputCollector.write(MapTask.java:691)

>         at

> org.apache.hadoop.mapreduce.TaskInputOutputContext.write(TaskInputOutputContext.java:80)

>         at com.att.hadoop.hivesort.HSMapper.map(HSMapper.java:39)

>         at com.att.hadoop.hivesort.HSMapper.map(HSMapper.java:1)

>         at org.apache.hadoop.mapreduce.Mapper.run(Mapper.java:144)

>         at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:764)

>         at org.apache.hadoop.mapred.MapTask.run(MapTask.java:370)

>         at org.apache.hadoop.mapred.Child$4.run(Child.java:255)

>         at java.security.AccessController.doPrivileged(Native Method)

>         at javax.security.auth.Subject.doAs(Subject.java:396)

>         at

> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1136)

>         at org.apache.hadoop.mapred.Child.main(Child.java:249)

>

>

> public int getPartition(Text key, HCatRecord record, int numParts) {

> //extract the group key from composite key

> String groupKey = key.toString().split("\\|")[0];

> return groupKey.hashCode() % numParts;

> }

>

>

> On Thu, Aug 29, 2013 at 8:31 PM, Shekhar Sharma <shekhar2...@gmail.com>

> wrote:

>>

>> No...partitionr decides which keys should go to which reducer...and

>> number of reducers you need to decide...No of reducers depends on

>> factors like number of key value pair, use case etc

>> Regards,

>> Som Shekhar Sharma

>> +91-8197243810

>>

>>

>> On Fri, Aug 30, 2013 at 5:54 AM, Adeel Qureshi <adeelmahm...@gmail.com>

>> wrote:

>> > so it cant figure out an appropriate number of reducers as it does for

>> > mappers .. in my case hadoop is using 2100+ mappers and then only 1

>> > reducer

>> > .. since im overriding the partitioner class shouldnt that decide how

>> > manyredeucers there should be based on how many different partition

>> > values

>> > being returned by the custom partiotioner

>> >

>> >

>> > On Thu, Aug 29, 2013 at 7:38 PM, Ian Wrigley <i...@cloudera.com> wrote:

>> >>

>> >> If you don't specify the number of Reducers, Hadoop will use the

>> >> default

>> >> -- which, unless you've changed it, is 1.

>> >>

>> >> Regards

>> >>

>> >> Ian.

>> >>

>> >> On Aug 29, 2013, at 4:23 PM, Adeel Qureshi <adeelmahm...@gmail.com>

>> >> wrote:

>> >>

>> >> I have implemented secondary sort in my MR job and for some reason if i

>> >> dont specify the number of reducers it uses 1 which doesnt seems right

>> >> because im working with 800M+ records and one reducer slows things down

>> >> significantly. Is this some kind of limitation with the secondary sort

>> >> that

>> >> it has to use a single reducer .. that kind of would defeat the purpose

>> >> of

>> >> having a scalable solution such as secondary sort. I would appreciate

>> >> any

>> >> help.

>> >>

>> >> Thanks

>> >> Adeel

>> >>

>> >>

>> >>

>> >> ---

>> >> Ian Wrigley

>> >> Sr. Curriculum Manager

>> >> Cloudera, Inc

>> >> Cell: (323) 819 4075

>> >>

>> >

>

>




                                          

Reply via email to