On Thu, May 7, 2009 at 8:51 PM, jason hadoop <jason.had...@gmail.com> wrote:
> Most likely the 3rd mapper ran as a speculative execution, and it is
> possible that all of your keys hashed to a single partition. Also, if you
> don't specify the default is to run a single reduce task.

As I mentioned in my first mail, I tried printiing out the hashCode()
for the keys myself in a manner like this:

System.out.println(key.hashCode());

key was of type Text. Some of the keys were even and some of them were
odd. So, I was expecting the odd ones to go to one slave and the even
ones to another. Is my expectation correct? Could you now throw some
light what else might have happened?

>
> From JobConf,
> /**
>   * Get configured the number of reduce tasks for this job. Defaults to
>   * <code>1</code>.
>   *
>   * @return the number of reduce tasks for this job.
>   */
>  public int getNumReduceTasks() { return getInt("mapred.reduce.tasks", *1*);
> }
>

I configured mapred.reduce.tasks as 2. I did this configuration in
hadoop-site.xml of the job-tracker. Is this fine?

Reply via email to