I tried that but it does not seem to work.

Hive> set mapred.map.tasks=100;
Hive> select count(1) from FCT_PRSC where date_key>='2009121600' and
date_key<'2009121700';
Total MapReduce jobs = 1
Number of reduce tasks determined at compile time: 1
In order to change the average load for a reducer (in bytes):
  set hive.exec.reducers.bytes.per.reducer=<number>
In order to limit the maximum number of reducers:
  set hive.exec.reducers.max=<number>
In order to set a constant number of reducers:
  set mapred.reduce.tasks=<number>


And this is what I see:
Jobid                           Priority        User    Name    Map %
Complete        Map Total       Maps Completed  Reduce % Complete
Reduce Total    Reduces Completed       Job Scheduling Information
job_200912162225_0033   NORMAL  dwhs    select count(1) from
...ate_key<'2009121700'(1/1)    1.06%
657     7       0.00%
1        0         

As you can see the "Map Total" is 657.


Lee Sagi | Data Warehouse Tech Lead & Architect | Work: 650-616-6575 |
Cell: 718-930-7947

-----Original Message-----
From: Edward Capriolo [mailto:edlinuxg...@gmail.com] 
Sent: Thursday, December 17, 2009 11:03 AM
To: hive-user@hadoop.apache.org
Subject: Re: Throttling hive queries

You should be able

hive > set mapred.map.tasks=1000
hive > set mapred.reduce.tasks=5

In some cases mappers is controlled by input files (pre hadoop 20)


On Thu, Dec 17, 2009 at 1:58 PM, Sagi, Lee <ls...@shopping.com> wrote:
> Is there a way to throttle hive queries?
>
> For example, I want to tell hive to not use more then 1000 mappers and

> 5 reducers for a particular query (or session).
>

Reply via email to