this may be simple - but is this the right solution? (and i have the same 
concern about hod)

if the cluster is unused - why restrict parallelism? if someone's willing to 
wake up at 4am to beat the crowd - they would just absolutely hate this.


-----Original Message-----
From: Doug Cutting [mailto:[EMAIL PROTECTED]
Sent: Thu 1/10/2008 9:50 AM
To: hadoop-user@lucene.apache.org
Subject: Re: Question on running simultaneous jobs
 
Aaron Kimball wrote:
> Multiple students should be able to submit jobs and if one student's 
> poorly-written task is grinding up a lot of cycles on a shared cluster, 
> other students still need to be able to test their code in the meantime; 

I think a simple approach to address this is to limit the number of 
tasks from a job that are permitted to execute simultaneously.  If, for 
example, you have a cluster of 50 dual-core nodes, with 100 map task 
slots and 100 reduce task slots, and the configured limit is 25 
simultaneous tasks/job, then four or more jobs will be able to run at a 
time.  This will permit faster jobs to pass slower jobs.  This approach 
also avoids some problems we've seen with HOD, where nodes are 
underutilized during the tail of jobs, and with input locality.

The JobTracker already handles simultaneously executing jobs, so the 
primary change required is just to task allocation, and thus should not 
prove intractable.

I've added a Jira issue for this:

   https://issues.apache.org/jira/browse/HADOOP-2573

Please add further comments there.

Doug

Reply via email to