From: Johannes Weiner <han...@cmpxchg.org> Instead of returning NULL from try_get_mem_cgroup_from_mm() when the mm owner is exiting, just return root_mem_cgroup. This makes sense for all callsites and gets rid of some of them having to fallback manually.
[fengguang...@intel.com: fix warnings] Signed-off-by: Johannes Weiner <han...@cmpxchg.org> Signed-off-by: Fengguang Wu <fengguang...@intel.com> Acked-by: Michal Hocko <mho...@suse.cz> Signed-off-by: Andrew Morton <a...@linux-foundation.org> Signed-off-by: Linus Torvalds <torva...@linux-foundation.org> https://jira.sw.ru/browse/PSBM-51558 (cherry picked from commit df381975463996178d685f6ef7d3555c5f887201) Signed-off-by: Andrey Ryabinin <aryabi...@virtuozzo.com> --- include/linux/memcontrol.h | 12 ++++++------ mm/memcontrol.c | 37 +++++++++++++++++-------------------- mm/oom_kill.c | 4 ++-- net/packet/af_packet.c | 2 +- 4 files changed, 26 insertions(+), 29 deletions(-) diff --git a/include/linux/memcontrol.h b/include/linux/memcontrol.h index f1b599f..bcf7752 100644 --- a/include/linux/memcontrol.h +++ b/include/linux/memcontrol.h @@ -85,7 +85,7 @@ int task_in_mem_cgroup(struct task_struct *task, const struct mem_cgroup *memcg) extern struct mem_cgroup *try_get_mem_cgroup_from_page(struct page *page); extern struct mem_cgroup *mem_cgroup_from_task(struct task_struct *p); -extern struct mem_cgroup *try_get_mem_cgroup_from_mm(struct mm_struct *mm); +extern struct mem_cgroup *get_mem_cgroup_from_mm(struct mm_struct *mm); extern struct mem_cgroup *parent_mem_cgroup(struct mem_cgroup *memcg); extern struct mem_cgroup *mem_cgroup_from_cont(struct cgroup *cont); @@ -304,17 +304,17 @@ static inline struct mem_cgroup *try_get_mem_cgroup_from_page(struct page *page) return NULL; } -static inline struct mem_cgroup *try_get_mem_cgroup_from_mm(struct mm_struct *mm) -{ - return NULL; -} - static inline bool mm_match_cgroup(struct mm_struct *mm, struct mem_cgroup *memcg) { return true; } +static inline struct mem_cgroup *get_mem_cgroup_from_mm(struct mm_struct *mm) +{ + return NULL; +} + static inline int task_in_mem_cgroup(struct task_struct *task, const struct mem_cgroup *memcg) { diff --git a/mm/memcontrol.c b/mm/memcontrol.c index d09d55d9..40ac81b 100644 --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -1166,15 +1166,24 @@ struct mem_cgroup *mem_cgroup_from_task(struct task_struct *p) return mem_cgroup_from_css(task_subsys_state(p, mem_cgroup_subsys_id)); } -struct mem_cgroup *try_get_mem_cgroup_from_mm(struct mm_struct *mm) +struct mem_cgroup *get_mem_cgroup_from_mm(struct mm_struct *mm) { struct mem_cgroup *memcg = NULL; rcu_read_lock(); do { - memcg = mem_cgroup_from_task(rcu_dereference(mm->owner)); - if (unlikely(!memcg)) - break; + /* + * Page cache insertions can happen withou an + * actual mm context, e.g. during disk probing + * on boot, loopback IO, acct() writes etc. + */ + if (unlikely(!mm)) + memcg = root_mem_cgroup; + else { + memcg = mem_cgroup_from_task(rcu_dereference(mm->owner)); + if (unlikely(!memcg)) + memcg = root_mem_cgroup; + } } while (!css_tryget(&memcg->css)); rcu_read_unlock(); return memcg; @@ -1546,7 +1555,7 @@ int task_in_mem_cgroup(struct task_struct *task, const struct mem_cgroup *memcg) p = find_lock_task_mm(task); if (p) { - curr = try_get_mem_cgroup_from_mm(p->mm); + curr = get_mem_cgroup_from_mm(p->mm); task_unlock(p); } else { /* @@ -1560,8 +1569,6 @@ int task_in_mem_cgroup(struct task_struct *task, const struct mem_cgroup *memcg) css_get(&curr->css); task_unlock(task); } - if (!curr) - return 0; /* * We should check use_hierarchy of "memcg" not "curr". Because checking * use_hierarchy of "curr" here make this function true if hierarchy is @@ -1693,7 +1700,7 @@ void mem_cgroup_note_oom_kill(struct mem_cgroup *root_memcg, p = find_lock_task_mm(task); if (p) { - memcg = try_get_mem_cgroup_from_mm(p->mm); + memcg = get_mem_cgroup_from_mm(p->mm); task_unlock(p); } else { rcu_read_lock(); @@ -3458,9 +3465,7 @@ struct kmem_cache *__memcg_kmem_get_cache(struct kmem_cache *cachep, if (!current->mm || current->memcg_kmem_skip_account) return cachep; - memcg = try_get_mem_cgroup_from_mm(current->mm); - if (unlikely(!memcg)) - return cachep; + memcg = get_mem_cgroup_from_mm(current->mm); if (!memcg_kmem_is_active(memcg)) goto out; @@ -3543,15 +3548,7 @@ __memcg_kmem_newpage_charge(struct page *page, gfp_t gfp, int order) if (!current->mm || current->memcg_kmem_skip_account) return true; - memcg = try_get_mem_cgroup_from_mm(current->mm); - - /* - * very rare case described in mem_cgroup_from_task. Unfortunately there - * isn't much we can do without complicating this too much, and it would - * be gfp-dependent anyway. Just let it go - */ - if (unlikely(!memcg)) - return true; + memcg = get_mem_cgroup_from_mm(current->mm); if (!memcg_kmem_is_active(memcg)) { css_put(&memcg->css); diff --git a/mm/oom_kill.c b/mm/oom_kill.c index 6b0542c2..a82433f 100644 --- a/mm/oom_kill.c +++ b/mm/oom_kill.c @@ -182,7 +182,7 @@ static unsigned long mm_overdraft(struct mm_struct *mm) struct oom_context *ctx; unsigned long overdraft; - memcg = try_get_mem_cgroup_from_mm(mm); + memcg = get_mem_cgroup_from_mm(mm); ctx = mem_cgroup_oom_context(memcg); overdraft = ctx->overdraft; if (memcg) @@ -476,7 +476,7 @@ void mark_oom_victim(struct task_struct *tsk) * owner memcg so that others can wait for it to exit. It will * be cleared in exit_oom_victim. */ - memcg = try_get_mem_cgroup_from_mm(tsk->mm); + memcg = get_mem_cgroup_from_mm(tsk->mm); ctx = mem_cgroup_oom_context(memcg); spin_lock(&oom_context_lock); if (!ctx->victim) { diff --git a/net/packet/af_packet.c b/net/packet/af_packet.c index f58fa90..c15e57a 100644 --- a/net/packet/af_packet.c +++ b/net/packet/af_packet.c @@ -2521,7 +2521,7 @@ static struct cg_proto *packet_sk_charge(void) goto out; err = 0; - psc->memcg = try_get_mem_cgroup_from_mm(current->mm); + psc->memcg = get_mem_cgroup_from_mm(current->mm); if (!psc->memcg) goto out_free_psc; if (!memcg_kmem_is_active(psc->memcg)) -- 2.10.2 _______________________________________________ Devel mailing list Devel@openvz.org https://lists.openvz.org/mailman/listinfo/devel