This patch adds the following:
1. Functional skeleton for the guest implementation. It enables the
guest to maintain the PFN of head buddy free pages of order
FREE_PAGE_HINTING_MIN_ORDER (currently defined as MAX_ORDER - 1)
in a per-cpu array.
Guest uses guest_free_page_enqueue() to enqueue the free pages post buddy
merging to the above mentioned per-cpu array.
guest_free_page_try_hinting() is used to initiate hinting operation once
the collected entries of the per-cpu array reaches or exceeds
HINTING_THRESHOLD (128). Having larger array size(MAX_FGPT_ENTRIES = 256)
than HINTING_THRESHOLD allows us to capture more pages specifically when
guest_free_page_enqueue() is called from free_pcppages_bulk().
For now guest_free_page_hinting() just resets the array index to continue
capturing of the freed pages.
2. Enables the support for x86 architecture.

Signed-off-by: Nitesh Narayan Lal <nit...@redhat.com>
---
 arch/x86/Kbuild              |  2 +-
 arch/x86/kvm/Kconfig         |  8 +++
 arch/x86/kvm/Makefile        |  2 +
 include/linux/page_hinting.h | 15 ++++++
 mm/page_alloc.c              |  5 ++
 virt/kvm/page_hinting.c      | 98 ++++++++++++++++++++++++++++++++++++
 6 files changed, 129 insertions(+), 1 deletion(-)
 create mode 100644 include/linux/page_hinting.h
 create mode 100644 virt/kvm/page_hinting.c

diff --git a/arch/x86/Kbuild b/arch/x86/Kbuild
index c625f57472f7..3244df4ee311 100644
--- a/arch/x86/Kbuild
+++ b/arch/x86/Kbuild
@@ -2,7 +2,7 @@ obj-y += entry/
 
 obj-$(CONFIG_PERF_EVENTS) += events/
 
-obj-$(CONFIG_KVM) += kvm/
+obj-$(subst m,y,$(CONFIG_KVM)) += kvm/
 
 # Xen paravirtualization support
 obj-$(CONFIG_XEN) += xen/
diff --git a/arch/x86/kvm/Kconfig b/arch/x86/kvm/Kconfig
index 72fa955f4a15..2fae31459706 100644
--- a/arch/x86/kvm/Kconfig
+++ b/arch/x86/kvm/Kconfig
@@ -96,6 +96,14 @@ config KVM_MMU_AUDIT
         This option adds a R/W kVM module parameter 'mmu_audit', which allows
         auditing of KVM MMU events at runtime.
 
+# KVM_FREE_PAGE_HINTING will allow the guest to report the free pages to the
+# host in regular interval of time.
+config KVM_FREE_PAGE_HINTING
+       def_bool y
+       depends on KVM
+       select VIRTIO
+       select VIRTIO_BALLOON
+
 # OK, it's a little counter-intuitive to do this, but it puts it neatly under
 # the virtualization menu.
 source "drivers/vhost/Kconfig"
diff --git a/arch/x86/kvm/Makefile b/arch/x86/kvm/Makefile
index 69b3a7c30013..78640a80501e 100644
--- a/arch/x86/kvm/Makefile
+++ b/arch/x86/kvm/Makefile
@@ -16,6 +16,8 @@ kvm-y                 += x86.o mmu.o emulate.o i8259.o irq.o 
lapic.o \
                           i8254.o ioapic.o irq_comm.o cpuid.o pmu.o mtrr.o \
                           hyperv.o page_track.o debugfs.o
 
+obj-$(CONFIG_KVM_FREE_PAGE_HINTING)    += $(KVM)/page_hinting.o
+
 kvm-intel-y            += vmx/vmx.o vmx/vmenter.o vmx/pmu_intel.o vmx/vmcs12.o 
vmx/evmcs.o vmx/nested.o
 kvm-amd-y              += svm.o pmu_amd.o
 
diff --git a/include/linux/page_hinting.h b/include/linux/page_hinting.h
new file mode 100644
index 000000000000..90254c582789
--- /dev/null
+++ b/include/linux/page_hinting.h
@@ -0,0 +1,15 @@
+#include <linux/gfp.h>
+/*
+ * Size of the array which is used to store the freed pages is defined by
+ * MAX_FGPT_ENTRIES.
+ */
+#define MAX_FGPT_ENTRIES       256
+/*
+ * Threshold value after which hinting needs to be initiated on the captured
+ * free pages.
+ */
+#define HINTING_THRESHOLD      128
+#define FREE_PAGE_HINTING_MIN_ORDER    (MAX_ORDER - 1)
+
+void guest_free_page_enqueue(struct page *page, int order);
+void guest_free_page_try_hinting(void);
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index d295c9bc01a8..684d047f33ee 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -67,6 +67,7 @@
 #include <linux/lockdep.h>
 #include <linux/nmi.h>
 #include <linux/psi.h>
+#include <linux/page_hinting.h>
 
 #include <asm/sections.h>
 #include <asm/tlbflush.h>
@@ -1194,9 +1195,11 @@ static void free_pcppages_bulk(struct zone *zone, int 
count,
                        mt = get_pageblock_migratetype(page);
 
                __free_one_page(page, page_to_pfn(page), zone, 0, mt);
+               guest_free_page_enqueue(page, 0);
                trace_mm_page_pcpu_drain(page, 0, mt);
        }
        spin_unlock(&zone->lock);
+       guest_free_page_try_hinting();
 }
 
 static void free_one_page(struct zone *zone,
@@ -1210,7 +1213,9 @@ static void free_one_page(struct zone *zone,
                migratetype = get_pfnblock_migratetype(page, pfn);
        }
        __free_one_page(page, pfn, zone, order, migratetype);
+       guest_free_page_enqueue(page, order);
        spin_unlock(&zone->lock);
+       guest_free_page_try_hinting();
 }
 
 static void __meminit __init_single_page(struct page *page, unsigned long pfn,
diff --git a/virt/kvm/page_hinting.c b/virt/kvm/page_hinting.c
new file mode 100644
index 000000000000..48b4b5e796b0
--- /dev/null
+++ b/virt/kvm/page_hinting.c
@@ -0,0 +1,98 @@
+#include <linux/mm.h>
+#include <linux/page_hinting.h>
+
+/*
+ * struct guest_free_pages- holds array of guest freed PFN's along with an
+ * index variable to track total freed PFN's.
+ * @free_pfn_arr: array to store the page frame number of all the pages which
+ * are freed by the guest.
+ * @guest_free_pages_idx: index to track the number entries stored in
+ * free_pfn_arr.
+ */
+struct guest_free_pages {
+       unsigned long free_page_arr[MAX_FGPT_ENTRIES];
+       int free_pages_idx;
+};
+
+DEFINE_PER_CPU(struct guest_free_pages, free_pages_obj);
+
+struct page *get_buddy_page(struct page *page)
+{
+       unsigned long pfn = page_to_pfn(page);
+       unsigned int order;
+
+       for (order = 0; order < MAX_ORDER; order++) {
+               struct page *page_head = page - (pfn & ((1 << order) - 1));
+
+               if (PageBuddy(page_head) && page_private(page_head) >= order)
+                       return page_head;
+       }
+       return NULL;
+}
+
+static void guest_free_page_hinting(void)
+{
+       struct guest_free_pages *hinting_obj = &get_cpu_var(free_pages_obj);
+
+       hinting_obj->free_pages_idx = 0;
+       put_cpu_var(hinting_obj);
+}
+
+int if_exist(struct page *page)
+{
+       int i = 0;
+       struct guest_free_pages *hinting_obj = this_cpu_ptr(&free_pages_obj);
+
+       while (i < MAX_FGPT_ENTRIES) {
+               if (page_to_pfn(page) == hinting_obj->free_page_arr[i])
+                       return 1;
+               i++;
+       }
+       return 0;
+}
+
+void guest_free_page_enqueue(struct page *page, int order)
+{
+       unsigned long flags;
+       struct guest_free_pages *hinting_obj;
+       int l_idx;
+
+       /*
+        * use of global variables may trigger a race condition between irq and
+        * process context causing unwanted overwrites. This will be replaced
+        * with a better solution to prevent such race conditions.
+        */
+       local_irq_save(flags);
+       hinting_obj = this_cpu_ptr(&free_pages_obj);
+       l_idx = hinting_obj->free_pages_idx;
+       if (l_idx != MAX_FGPT_ENTRIES) {
+               if (PageBuddy(page) && page_private(page) >=
+                   FREE_PAGE_HINTING_MIN_ORDER) {
+                       hinting_obj->free_page_arr[l_idx] = page_to_pfn(page);
+                       hinting_obj->free_pages_idx += 1;
+               } else {
+                       struct page *buddy_page = get_buddy_page(page);
+
+                       if (buddy_page && page_private(buddy_page) >=
+                           FREE_PAGE_HINTING_MIN_ORDER &&
+                           !if_exist(buddy_page)) {
+                               unsigned long buddy_pfn =
+                                       page_to_pfn(buddy_page);
+
+                               hinting_obj->free_page_arr[l_idx] =
+                                                       buddy_pfn;
+                               hinting_obj->free_pages_idx += 1;
+                       }
+               }
+       }
+       local_irq_restore(flags);
+}
+
+void guest_free_page_try_hinting(void)
+{
+       struct guest_free_pages *hinting_obj;
+
+       hinting_obj = this_cpu_ptr(&free_pages_obj);
+       if (hinting_obj->free_pages_idx >= HINTING_THRESHOLD)
+               guest_free_page_hinting();
+}
-- 
2.17.2

Reply via email to