On Tue, Sep 22, 2020 at 09:05:05AM -0300, Jason Gunthorpe wrote:
> On Mon, Sep 21, 2020 at 05:20:31PM -0400, Peter Xu wrote:
> > Pinned pages shouldn't be write-protected when fork() happens, because 
> > follow
> > up copy-on-write on these pages could cause the pinned pages to be replaced 
> > by
> > random newly allocated pages.
> > 
> > For huge PMDs, we split the huge pmd if pinning is detected.  So that future
> > handling will be done by the PTE level (with our latest changes, each of the
> > small pages will be copied).  We can achieve this by let copy_huge_pmd() 
> > return
> > -EAGAIN for pinned pages, so that we'll fallthrough in copy_pmd_range() and
> > finally land the next copy_pte_range() call.
> > 
> > Huge PUDs will be even more special - so far it does not support anonymous
> > pages.  But it can actually be done the same as the huge PMDs even if the 
> > split
> > huge PUDs means to erase the PUD entries.  It'll guarantee the follow up 
> > fault
> > ins will remap the same pages in either parent/child later.
> > 
> > This might not be the most efficient way, but it should be easy and clean
> > enough.  It should be fine, since we're tackling with a very rare case just 
> > to
> > make sure userspaces that pinned some thps will still work even without
> > MADV_DONTFORK and after they fork()ed.
> > 
> > Signed-off-by: Peter Xu <pet...@redhat.com>
> >  mm/huge_memory.c | 26 ++++++++++++++++++++++++++
> >  1 file changed, 26 insertions(+)
> > 
> > diff --git a/mm/huge_memory.c b/mm/huge_memory.c
> > index 7ff29cc3d55c..c40aac0ad87e 100644
> > +++ b/mm/huge_memory.c
> > @@ -1074,6 +1074,23 @@ int copy_huge_pmd(struct mm_struct *dst_mm, struct 
> > mm_struct *src_mm,
> >  
> >     src_page = pmd_page(pmd);
> >     VM_BUG_ON_PAGE(!PageHead(src_page), src_page);
> > +
> > +   /*
> > +    * If this page is a potentially pinned page, split and retry the fault
> > +    * with smaller page size.  Normally this should not happen because the
> > +    * userspace should use MADV_DONTFORK upon pinned regions.  This is a
> > +    * best effort that the pinned pages won't be replaced by another
> > +    * random page during the coming copy-on-write.
> > +    */
> > +   if (unlikely(READ_ONCE(src_mm->has_pinned) &&
> > +                page_maybe_dma_pinned(src_page))) {
> > +           pte_free(dst_mm, pgtable);
> > +           spin_unlock(src_ptl);
> > +           spin_unlock(dst_ptl);
> > +           __split_huge_pmd(vma, src_pmd, addr, false, NULL);
> > +           return -EAGAIN;
> > +   }
> 
> Not sure why, but the PMD stuff here is not calling is_cow_mapping()
> before doing the write protect. Seems like it might be an existing
> bug?

IMHO it's not a bug, because splitting a huge pmd should always be safe.

One thing I can think of that might be special here is when the pmd is
anonymously mapped but also shared (shared, tmpfs thp, I think?), then here
we'll also mark it as wrprotected even if we don't need to (or maybe we need it
for some reason..).  But again I think it's safe anyways - when page fault
happens, wp_huge_pmd() should split it into smaller pages unconditionally.  I
just don't know whether it's the ideal way for the shared case.  Andrea should
definitely know it better (because it is there since the 1st day of thp).

> 
> In any event, the has_pinned logic shouldn't be used without also
> checking is_cow_mapping(), so it should be added to that test. Same
> remarks for PUD

I think the case mentioned above is also the special case here when we didn't
check is_cow_mapping().  The major difference is whether we'll split the page
right now, or postpone it until the next write to each mm.  But I think, yes,
maybe I should better still keep the is_cow_mapping() to be explicit.

Thanks,

-- 
Peter Xu

Reply via email to