Andrew, I've had some concerns wrt. this patch - especially the additional complexity - and I have to say I am not convinced that this is really needed. Our past experience in this area suggests that more tricky code leads to different corner cases. So I am really reluctant to add more complexity without any real world reports.
On Tue 23-07-19 16:14:29, Andrew Morton wrote: > From: Tetsuo Handa <penguin-ker...@i-love.sakura.ne.jp> > Subject: mm, oom: avoid printk() iteration under RCU > > Currently dump_tasks() might call printk() for many thousands times under > RCU, which might take many minutes for slow consoles. Therefore, split > dump_tasks() into three stages; take a snapshot of possible OOM victim > candidates under RCU, dump the snapshot from reschedulable context, and > destroy the snapshot. > > In a future patch, the first stage would be moved to select_bad_process() > and the third stage would be moved to after oom_kill_process(), and will > simplify refcount handling. > > Link: > http://lkml.kernel.org/r/1563360901-8277-1-git-send-email-penguin-ker...@i-love.sakura.ne.jp > Signed-off-by: Tetsuo Handa <penguin-ker...@i-love.sakura.ne.jp> > Cc: Shakeel Butt <shake...@google.com> > Cc: Michal Hocko <mho...@suse.com> > Cc: Roman Gushchin <g...@fb.com> > Cc: David Rientjes <rient...@google.com> > Signed-off-by: Andrew Morton <a...@linux-foundation.org> > --- > > include/linux/sched.h | 1 > mm/oom_kill.c | 67 +++++++++++++++++++--------------------- > 2 files changed, 34 insertions(+), 34 deletions(-) > > --- a/include/linux/sched.h~mm-oom-avoid-printk-iteration-under-rcu > +++ a/include/linux/sched.h > @@ -1246,6 +1246,7 @@ struct task_struct { > #ifdef CONFIG_MMU > struct task_struct *oom_reaper_list; > #endif > + struct list_head oom_victim_list; > #ifdef CONFIG_VMAP_STACK > struct vm_struct *stack_vm_area; > #endif > --- a/mm/oom_kill.c~mm-oom-avoid-printk-iteration-under-rcu > +++ a/mm/oom_kill.c > @@ -377,36 +377,13 @@ static void select_bad_process(struct oo > } > } > > -static int dump_task(struct task_struct *p, void *arg) > -{ > - struct oom_control *oc = arg; > - struct task_struct *task; > - > - if (oom_unkillable_task(p)) > - return 0; > > - /* p may not have freeable memory in nodemask */ > - if (!is_memcg_oom(oc) && !oom_cpuset_eligible(p, oc)) > - return 0; > - > - task = find_lock_task_mm(p); > - if (!task) { > - /* > - * This is a kthread or all of p's threads have already > - * detached their mm's. There's no need to report > - * them; they can't be oom killed anyway. > - */ > - return 0; > +static int add_candidate_task(struct task_struct *p, void *arg) > +{ > + if (!oom_unkillable_task(p)) { > + get_task_struct(p); > + list_add_tail(&p->oom_victim_list, (struct list_head *) arg); > } > - > - pr_info("[%7d] %5d %5d %8lu %8lu %8ld %8lu %5hd %s\n", > - task->pid, from_kuid(&init_user_ns, task_uid(task)), > - task->tgid, task->mm->total_vm, get_mm_rss(task->mm), > - mm_pgtables_bytes(task->mm), > - get_mm_counter(task->mm, MM_SWAPENTS), > - task->signal->oom_score_adj, task->comm); > - task_unlock(task); > - > return 0; > } > > @@ -422,19 +399,41 @@ static int dump_task(struct task_struct > */ > static void dump_tasks(struct oom_control *oc) > { > - pr_info("Tasks state (memory values in pages):\n"); > - pr_info("[ pid ] uid tgid total_vm rss pgtables_bytes > swapents oom_score_adj name\n"); > + static LIST_HEAD(list); > + struct task_struct *p; > + struct task_struct *t; > > if (is_memcg_oom(oc)) > - mem_cgroup_scan_tasks(oc->memcg, dump_task, oc); > + mem_cgroup_scan_tasks(oc->memcg, add_candidate_task, &list); > else { > - struct task_struct *p; > - > rcu_read_lock(); > for_each_process(p) > - dump_task(p, oc); > + add_candidate_task(p, &list); > rcu_read_unlock(); > } > + pr_info("Tasks state (memory values in pages):\n"); > + pr_info("[ pid ] uid tgid total_vm rss pgtables_bytes > swapents oom_score_adj name\n"); > + list_for_each_entry(p, &list, oom_victim_list) { > + cond_resched(); > + /* p may not have freeable memory in nodemask */ > + if (!is_memcg_oom(oc) && !oom_cpuset_eligible(p, oc)) > + continue; > + /* All of p's threads might have already detached their mm's. */ > + t = find_lock_task_mm(p); > + if (!t) > + continue; > + pr_info("[%7d] %5d %5d %8lu %8lu %8ld %8lu %5hd %s\n", > + t->pid, from_kuid(&init_user_ns, task_uid(t)), > + t->tgid, t->mm->total_vm, get_mm_rss(t->mm), > + mm_pgtables_bytes(t->mm), > + get_mm_counter(t->mm, MM_SWAPENTS), > + t->signal->oom_score_adj, t->comm); > + task_unlock(t); > + } > + list_for_each_entry_safe(p, t, &list, oom_victim_list) { > + list_del(&p->oom_victim_list); > + put_task_struct(p); > + } > } > > static void dump_oom_summary(struct oom_control *oc, struct task_struct > *victim) > _ > > Patches currently in -mm which might be from > penguin-ker...@i-love.sakura.ne.jp are > > mm-oom-avoid-printk-iteration-under-rcu.patch > info-task-hung-in-generic_file_write_iter.patch > info-task-hung-in-generic_file_write-fix.patch > kexec-bail-out-upon-sigkill-when-allocating-memory.patch -- Michal Hocko SUSE Labs