[ https://issues.apache.org/jira/browse/YARN-3933?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ]
Miklos Szegedi updated YARN-3933: --------------------------------- Attachment: YARN-3933.006.patch [~guoshiwei], thank you for looking into this. I ran into the same issue in YARN-6158 and I had a patch there. I attach the rebased version here for your consideration. It is very similar to the current patch, it just addresses the test issue. I am also okay with 004.patch. It would be nice to get either of the fixes checked in. What do you think? > Race condition when calling AbstractYarnScheduler.completedContainer. > --------------------------------------------------------------------- > > Key: YARN-3933 > URL: https://issues.apache.org/jira/browse/YARN-3933 > Project: Hadoop YARN > Issue Type: Bug > Components: fairscheduler > Affects Versions: 2.7.1 > Reporter: Lavkesh Lahngir > Assignee: Shiwei Guo > Labels: oct16-medium > Attachments: YARN-3933.001.patch, YARN-3933.002.patch, > YARN-3933.003.patch, YARN-3933.004.patch, YARN-3933.005.patch, > YARN-3933.006.patch > > > In our cluster we are seeing available memory and cores being negative. > Initial inspection: > Scenario no. 1: > In capacity scheduler the method allocateContainersToNode() checks if > there are excess reservation of containers for an application, and they are > no longer needed then it calls queue.completedContainer() which causes > resources being negative. And they were never assigned in the first place. > I am still looking through the code. Can somebody suggest how to simulate > excess containers assignments ? -- This message was sent by Atlassian JIRA (v6.3.15#6346) --------------------------------------------------------------------- To unsubscribe, e-mail: yarn-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: yarn-issues-h...@hadoop.apache.org