[
https://issues.apache.org/jira/browse/HADOOP-5083?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12668103#action_12668103
]
dhruba borthakur commented on HADOOP-5083:
------------------------------------------
We are 0.17 (and 0.19) with long-running JTs and one problem we are seeing is
that the JT sometimes runs out of its 3GB heap space. The system is used by
around 50-80 users. The max jobs per user before retirement is 5. But this
still means that the JT keeps info about 80*5=400 completed jobs in memory.
Sometimes these jobs have huge number of tasks. This eats up most of the memory
in the JT.
Can part of this fix (i.e. purge all jobs from completed queue when there is
memory usage on the JT exceeds a configured threshold) be ported to 0.19. It is
very useful to have it in 0.19, otherwise the JT just hangs up and the cluster
has to be restarted.
> Optionally a separate daemon should serve JobHistory
> ----------------------------------------------------
>
> Key: HADOOP-5083
> URL: https://issues.apache.org/jira/browse/HADOOP-5083
> Project: Hadoop Core
> Issue Type: Improvement
> Components: mapred
> Reporter: Arun C Murthy
> Assignee: Amar Kamat
> Attachments: HADOOP-5083-v1.2.patch, HADOOP-5083-v1.9.patch
>
>
> Currently the JobTracker serves the JobHistory to end-users off files
> local-disk/hdfs. While running very large clusters with a large user-base
> might result in lots of traffic for job-history which needlessly taxes the
> JobTracker. The proposal is to have an optional daemon which handles serving
> of job-history requests.
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.