[ https://issues.apache.org/jira/browse/YARN-1680?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14529654#comment-14529654 ]
Wangda Tan commented on YARN-1680: ---------------------------------- I think we should stop adding such application-specific logic into RM, application can have very varied resource request, for example: - Relax locality, this is very similar to "white list" - Black list. - In the future we can have affinity/anti-affinity/constraints. We cannot do so much expensive calculation in centralized way. RM should only take care of general limits, such as user-limits/queue-limits, like what we have now. I propose to - In short term, treat the headroom just a hint, like what [~kasha] mentioned: https://issues.apache.org/jira/browse/MAPREDUCE-6302?focusedCommentId=14494728&page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14494728. I'm not sure if MAPREDUCE-6302 solved the problem already, I haven't looked at the patch yet. - In longer term, support headroom calculation in client-side utils, maybe AMRMClient is a good place. > availableResources sent to applicationMaster in heartbeat should exclude > blacklistedNodes free memory. > ------------------------------------------------------------------------------------------------------ > > Key: YARN-1680 > URL: https://issues.apache.org/jira/browse/YARN-1680 > Project: Hadoop YARN > Issue Type: Sub-task > Components: capacityscheduler > Affects Versions: 2.2.0, 2.3.0 > Environment: SuSE 11 SP2 + Hadoop-2.3 > Reporter: Rohith > Assignee: Craig Welch > Attachments: YARN-1680-WIP.patch, YARN-1680-v2.patch, > YARN-1680-v2.patch, YARN-1680.patch > > > There are 4 NodeManagers with 8GB each.Total cluster capacity is 32GB.Cluster > slow start is set to 1. > Job is running reducer task occupied 29GB of cluster.One NodeManager(NM-4) is > become unstable(3 Map got killed), MRAppMaster blacklisted unstable > NodeManager(NM-4). All reducer task are running in cluster now. > MRAppMaster does not preempt the reducers because for Reducer preemption > calculation, headRoom is considering blacklisted nodes memory. This makes > jobs to hang forever(ResourceManager does not assing any new containers on > blacklisted nodes but returns availableResouce considers cluster free > memory). -- This message was sent by Atlassian JIRA (v6.3.4#6332)