Author: dougm
Date: Fri May 31 21:02:42 2019
New Revision: 348484
URL: https://svnweb.freebsd.org/changeset/base/348484

Log:
  The function vm_phys_free_contig invokes vm_phys_free_pages for every
  power-of-two page block it frees, launching an unsuccessful search for
  a buddy to pair up with each time.  The only possible buddy-up mergers
  are across the boundaries of the freed region, so change
  vm_phys_free_contig simply to enqueue the freed interior blocks, via a
  new function vm_phys_enqueue_contig, and then call vm_phys_free_pages
  on the bounding blocks to create as big a cross-boundary block as
  possible after buddy-merging.
  
  The only callers of vm_phys_free_contig at the moment call it in
  situations where merging blocks across the boundary is clearly
  impossible, so just call vm_phys_enqueue_contig in those places and
  avoid trying to buddy-up at all.
  
  One beneficiary of this change is in breaking reservations.  For the
  case where memory is freed in breaking a reservation with only the
  first and last pages allocated, the number of cycles consumed by the
  operation drops about 11% with this change.
  
  Suggested by: alc
  Reviewed by: alc
  Approved by: kib, markj (mentors)
  Differential Revision: https://reviews.freebsd.org/D16901

Modified:
  head/sys/vm/vm_page.c
  head/sys/vm/vm_phys.c
  head/sys/vm/vm_phys.h
  head/sys/vm/vm_reserv.c

Modified: head/sys/vm/vm_page.c
==============================================================================
--- head/sys/vm/vm_page.c       Fri May 31 20:36:32 2019        (r348483)
+++ head/sys/vm/vm_page.c       Fri May 31 21:02:42 2019        (r348484)
@@ -833,7 +833,7 @@ vm_page_startup(vm_offset_t vaddr)
 
                        vmd = VM_DOMAIN(seg->domain);
                        vm_domain_free_lock(vmd);
-                       vm_phys_free_contig(m, pagecount);
+                       vm_phys_enqueue_contig(m, pagecount);
                        vm_domain_free_unlock(vmd);
                        vm_domain_freecnt_inc(vmd, pagecount);
                        vm_cnt.v_page_count += (u_int)pagecount;

Modified: head/sys/vm/vm_phys.c
==============================================================================
--- head/sys/vm/vm_phys.c       Fri May 31 20:36:32 2019        (r348483)
+++ head/sys/vm/vm_phys.c       Fri May 31 21:02:42 2019        (r348484)
@@ -1095,14 +1095,35 @@ vm_phys_free_pages(vm_page_t m, int order)
 }
 
 /*
- * Free a contiguous, arbitrarily sized set of physical pages.
+ * Return the largest possible order of a set of pages starting at m.
+ */
+static int
+max_order(vm_page_t m)
+{
+
+       /*
+        * Unsigned "min" is used here so that "order" is assigned
+        * "VM_NFREEORDER - 1" when "m"'s physical address is zero
+        * or the low-order bits of its physical address are zero
+        * because the size of a physical address exceeds the size of
+        * a long.
+        */
+       return (min(ffsl(VM_PAGE_TO_PHYS(m) >> PAGE_SHIFT) - 1,
+           VM_NFREEORDER - 1));
+}
+
+/*
+ * Free a contiguous, arbitrarily sized set of physical pages, without
+ * merging across set boundaries.
  *
  * The free page queues must be locked.
  */
 void
-vm_phys_free_contig(vm_page_t m, u_long npages)
+vm_phys_enqueue_contig(vm_page_t m, u_long npages)
 {
-       u_int n;
+       struct vm_freelist *fl;
+       struct vm_phys_seg *seg;
+       vm_page_t m_end;
        int order;
 
        /*
@@ -1110,29 +1131,68 @@ vm_phys_free_contig(vm_page_t m, u_long npages)
         * possible power-of-two-sized subsets.
         */
        vm_domain_free_assert_locked(vm_pagequeue_domain(m));
-       for (;; npages -= n) {
-               /*
-                * Unsigned "min" is used here so that "order" is assigned
-                * "VM_NFREEORDER - 1" when "m"'s physical address is zero
-                * or the low-order bits of its physical address are zero
-                * because the size of a physical address exceeds the size of
-                * a long.
-                */
-               order = min(ffsl(VM_PAGE_TO_PHYS(m) >> PAGE_SHIFT) - 1,
-                   VM_NFREEORDER - 1);
-               n = 1 << order;
-               if (npages < n)
-                       break;
-               vm_phys_free_pages(m, order);
-               m += n;
+       seg = &vm_phys_segs[m->segind];
+       fl = (*seg->free_queues)[m->pool];
+       m_end = m + npages;
+       /* Free blocks of increasing size. */
+       while ((order = max_order(m)) < VM_NFREEORDER - 1 &&
+           m + (1 << order) <= m_end) {
+               KASSERT(seg == &vm_phys_segs[m->segind],
+                   ("%s: page range [%p,%p) spans multiple segments",
+                   __func__, m_end - npages, m));
+               vm_freelist_add(fl, m, order, 1);
+               m += 1 << order;
        }
-       /* The residual "npages" is less than "1 << (VM_NFREEORDER - 1)". */
-       for (; npages > 0; npages -= n) {
-               order = flsl(npages) - 1;
-               n = 1 << order;
-               vm_phys_free_pages(m, order);
-               m += n;
+       /* Free blocks of maximum size. */
+       while (m + (1 << order) <= m_end) {
+               KASSERT(seg == &vm_phys_segs[m->segind],
+                   ("%s: page range [%p,%p) spans multiple segments",
+                   __func__, m_end - npages, m));
+               vm_freelist_add(fl, m, order, 1);
+               m += 1 << order;
        }
+       /* Free blocks of diminishing size. */
+       while (m < m_end) {
+               KASSERT(seg == &vm_phys_segs[m->segind],
+                   ("%s: page range [%p,%p) spans multiple segments",
+                   __func__, m_end - npages, m));
+               order = flsl(m_end - m) - 1;
+               vm_freelist_add(fl, m, order, 1);
+               m += 1 << order;
+       }
+}
+
+/*
+ * Free a contiguous, arbitrarily sized set of physical pages.
+ *
+ * The free page queues must be locked.
+ */
+void
+vm_phys_free_contig(vm_page_t m, u_long npages)
+{
+       int order_start, order_end;
+       vm_page_t m_start, m_end;
+
+       vm_domain_free_assert_locked(vm_pagequeue_domain(m));
+
+       m_start = m;
+       order_start = max_order(m_start);
+       if (order_start < VM_NFREEORDER - 1)
+               m_start += 1 << order_start;
+       m_end = m + npages;
+       order_end = max_order(m_end);
+       if (order_end < VM_NFREEORDER - 1)
+               m_end -= 1 << order_end;
+       /*
+        * Avoid unnecessary coalescing by freeing the pages at the start and
+        * end of the range last.
+        */
+       if (m_start < m_end)
+               vm_phys_enqueue_contig(m_start, m_end - m_start);
+       if (order_start < VM_NFREEORDER - 1)
+               vm_phys_free_pages(m, order_start);
+       if (order_end < VM_NFREEORDER - 1)
+               vm_phys_free_pages(m_end, order_end);
 }
 
 /*

Modified: head/sys/vm/vm_phys.h
==============================================================================
--- head/sys/vm/vm_phys.h       Fri May 31 20:36:32 2019        (r348483)
+++ head/sys/vm/vm_phys.h       Fri May 31 21:02:42 2019        (r348484)
@@ -84,6 +84,7 @@ vm_page_t vm_phys_alloc_freelist_pages(int domain, int
 int vm_phys_alloc_npages(int domain, int pool, int npages, vm_page_t ma[]);
 vm_page_t vm_phys_alloc_pages(int domain, int pool, int order);
 int vm_phys_domain_match(int prefer, vm_paddr_t low, vm_paddr_t high);
+void vm_phys_enqueue_contig(vm_page_t m, u_long npages);
 int vm_phys_fictitious_reg_range(vm_paddr_t start, vm_paddr_t end,
     vm_memattr_t memattr);
 void vm_phys_fictitious_unreg_range(vm_paddr_t start, vm_paddr_t end);

Modified: head/sys/vm/vm_reserv.c
==============================================================================
--- head/sys/vm/vm_reserv.c     Fri May 31 20:36:32 2019        (r348483)
+++ head/sys/vm/vm_reserv.c     Fri May 31 21:02:42 2019        (r348484)
@@ -1066,7 +1066,7 @@ vm_reserv_break(vm_reserv_t rv)
                        else {
                                hi = NBPOPMAP * i + bitpos;
                                vm_domain_free_lock(VM_DOMAIN(rv->domain));
-                               vm_phys_free_contig(&rv->pages[lo], hi - lo);
+                               vm_phys_enqueue_contig(&rv->pages[lo], hi - lo);
                                vm_domain_free_unlock(VM_DOMAIN(rv->domain));
                                lo = hi;
                        }
_______________________________________________
svn-src-head@freebsd.org mailing list
https://lists.freebsd.org/mailman/listinfo/svn-src-head
To unsubscribe, send any mail to "svn-src-head-unsubscr...@freebsd.org"

Reply via email to