Commit-ID:  1a201bacffd9f7cbc0f9a3b790dcb8108336784d
Gitweb:     http://git.kernel.org/tip/1a201bacffd9f7cbc0f9a3b790dcb8108336784d
Author:     Peter Zijlstra <[email protected]>
AuthorDate: Wed, 26 Sep 2012 16:05:01 +0200
Committer:  Ingo Molnar <[email protected]>
CommitDate: Thu, 27 Sep 2012 14:46:02 +0200

mm: Optimize do_prot_none()

Reduces do_prot_none() to a single pte_lock acquisition in the !migrate case.

Also flipping the protection bits back sooner avoids the chance of
other CPUs hitting it and causing a pile-up on the pte_lock.

This significantly reduces contention on the pte_lock in a NUMA page fault
benchmark, system time dropped from around 78% to 35% and __memset_sse2()
is the most expensive function instead of _raw_spin_lock()/handle_pte_fault().

Signed-off-by: Peter Zijlstra <[email protected]>
Cc: Linus Torvalds <[email protected]>
Cc: Andrew Morton <[email protected]>
Cc: Rik van Riel <[email protected]>
Link: http://lkml.kernel.org/n/[email protected]
Signed-off-by: Ingo Molnar <[email protected]>
---
 mm/memory.c |   66 +++++++++++++++++++++++++----------------------------------
 1 files changed, 28 insertions(+), 38 deletions(-)

diff --git a/mm/memory.c b/mm/memory.c
index d896a24..965eeef 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -3438,56 +3438,39 @@ static bool pte_prot_none(struct vm_area_struct *vma, 
pte_t pte)
        return pte_same(pte, pte_modify(pte, vma_prot_none(vma)));
 }
 
-static int do_prot_none(struct mm_struct *mm, struct vm_area_struct *vma,
-                       unsigned long address, pte_t *ptep, pmd_t *pmd,
-                       unsigned int flags, pte_t entry)
+#ifdef CONFIG_NUMA
+static void do_prot_none_numa(struct mm_struct *mm, struct vm_area_struct *vma,
+                             unsigned long address, struct page *page)
 {
-       struct page *page = NULL;
-       spinlock_t *ptl;
        int node;
 
-       ptl = pte_lockptr(mm, pmd);
-       spin_lock(ptl);
-       if (unlikely(!pte_same(*ptep, entry)))
-               goto unlock;
-
-#ifdef CONFIG_NUMA
        /*
         * For NUMA systems we use the special PROT_NONE maps to drive
         * lazy page migration, see MPOL_MF_LAZY and related.
         */
-       page = vm_normal_page(vma, address, entry);
-       if (!page)
-               goto do_fixup_locked;
-
-       get_page(page);
-       pte_unmap_unlock(ptep, ptl);
-
        node = mpol_misplaced(page, vma, address);
-       if (node == -1)
-               goto do_fixup;
+       if (node != -1)
+               migrate_misplaced_page(mm, page, node);
+}
+#else
+static void do_prot_none_numa(struct mm_struct *mm, struct vm_area_struct *vma,
+                             unsigned long address, struct page *page)
+{
+}
+#endif /* CONFIG_NUMA */
 
-       /*
-        * Page migration will install a new pte with vma->vm_page_prot,
-        * otherwise fall-through to the fixup. Next time,.. perhaps.
-        */
-       if (!migrate_misplaced_page(mm, page, node)) {
-               put_page(page);
-               return 0;
-       }
+static int do_prot_none(struct mm_struct *mm, struct vm_area_struct *vma,
+                       unsigned long address, pte_t *ptep, pmd_t *pmd,
+                       unsigned int flags, pte_t entry)
+{
+       struct page *page = NULL;
+       spinlock_t *ptl;
 
-do_fixup:
-       /*
-        * OK, nothing to do,.. change the protection back to what it
-        * ought to be.
-        */
-       ptep = pte_offset_map_lock(mm, pmd, address, &ptl);
+       ptl = pte_lockptr(mm, pmd);
+       spin_lock(ptl);
        if (unlikely(!pte_same(*ptep, entry)))
                goto unlock;
 
-do_fixup_locked:
-#endif /* CONFIG_NUMA */
-
        flush_cache_page(vma, address, pte_pfn(entry));
 
        ptep_modify_prot_start(mm, address, ptep);
@@ -3495,10 +3478,17 @@ do_fixup_locked:
        ptep_modify_prot_commit(mm, address, ptep, entry);
 
        update_mmu_cache(vma, address, ptep);
+
+       page = vm_normal_page(vma, address, entry);
+       if (page)
+               get_page(page);
+
 unlock:
        pte_unmap_unlock(ptep, ptl);
-       if (page)
+       if (page) {
+               do_prot_none_numa(mm, vma, address, page);
                put_page(page);
+       }
        return 0;
 }
 
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [email protected]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to