[ 
https://issues.apache.org/jira/browse/HADOOP-4523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12644088#action_12644088
 ] 

Owen O'Malley commented on HADOOP-4523:
---------------------------------------

This jira isn't very clear. What are you proposing changing? Is it to make the 
mapred.tasktracker.tasks.maxmemory pluggable? If so, I'd propose making an 
interface like:

{code}
abstract class MemoryPlugin {
  long getVirtualMemorySize(Configuration conf);
}
{code}

and you configure an implementation of it. (mapred.server.memory.plugin ?)


> Enhance how memory-intensive user tasks are handled
> ---------------------------------------------------
>
>                 Key: HADOOP-4523
>                 URL: https://issues.apache.org/jira/browse/HADOOP-4523
>             Project: Hadoop Core
>          Issue Type: Improvement
>          Components: mapred
>    Affects Versions: 0.19.0
>            Reporter: Vivek Ratan
>            Assignee: Vinod K V
>
> HADOOP-3581 monitors each Hadoop task to see if its memory usage (which 
> includes usage of any tasks spawned by it and so on) is within a per-task 
> limit. If the task's memory usage goes over its limit, the task is killed. 
> This, by itself, is not enough to prevent badly behaving jobs from bringing 
> down nodes. What is also needed is the ability to make sure that the sum 
> total of VM usage of all Hadoop tasks does not exceed a certain limit.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to