Process-local memory uses a dedicated pgd entry in kernel space and its
own page table structure. Hook mm exit functions to cleanup those
dedicated page tables. As a preparation, release any left-over
process-local allocations in the address space.

Signed-off-by: Marius Hillenbrand <mhill...@amazon.de>
Cc: Alexander Graf <g...@amazon.de>
Cc: David Woodhouse <d...@amazon.co.uk>
---
 arch/x86/include/asm/proclocal.h |  11 +++
 arch/x86/mm/Makefile             |   1 +
 arch/x86/mm/proclocal.c          | 131 +++++++++++++++++++++++++++++++
 include/linux/mm_types.h         |   9 +++
 include/linux/proclocal.h        |  16 ++++
 kernel/fork.c                    |   6 ++
 mm/Makefile                      |   1 +
 mm/proclocal.c                   |  35 +++++++++
 8 files changed, 210 insertions(+)
 create mode 100644 arch/x86/include/asm/proclocal.h
 create mode 100644 arch/x86/mm/proclocal.c
 create mode 100644 include/linux/proclocal.h
 create mode 100644 mm/proclocal.c

diff --git a/arch/x86/include/asm/proclocal.h b/arch/x86/include/asm/proclocal.h
new file mode 100644
index 000000000000..a66983e49209
--- /dev/null
+++ b/arch/x86/include/asm/proclocal.h
@@ -0,0 +1,11 @@
+/*
+ * Copyright 2019 Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ */
+#ifndef _ASM_X86_PROCLOCAL_H
+#define _ASM_X86_PROCLOCAL_H
+
+struct mm_struct;
+
+void arch_proclocal_teardown_pages_and_pt(struct mm_struct *mm);
+
+#endif /* _ASM_X86_PROCLOCAL_H */
diff --git a/arch/x86/mm/Makefile b/arch/x86/mm/Makefile
index e4ffcf74770e..a7c7111eb8f0 100644
--- a/arch/x86/mm/Makefile
+++ b/arch/x86/mm/Makefile
@@ -56,3 +56,4 @@ obj-$(CONFIG_AMD_MEM_ENCRYPT) += mem_encrypt_boot.o
 
 obj-$(CONFIG_XPFO)             += xpfo.o
 
+obj-$(CONFIG_PROCLOCAL) += proclocal.o
diff --git a/arch/x86/mm/proclocal.c b/arch/x86/mm/proclocal.c
new file mode 100644
index 000000000000..c64a8ea6360d
--- /dev/null
+++ b/arch/x86/mm/proclocal.c
@@ -0,0 +1,131 @@
+/*
+ * Architecture-specific code for handling process-local memory on x86-64.
+ *
+ * Copyright 2019 Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ */
+
+#include <linux/list.h>
+#include <linux/mm.h>
+#include <linux/proclocal.h>
+#include <linux/set_memory.h>
+#include <asm/tlb.h>
+
+
+static void unmap_leftover_pages_pte(struct mm_struct *mm, pmd_t *pmd,
+                                    unsigned long addr, unsigned long end,
+                                    struct list_head *page_list)
+{
+       pte_t *pte;
+       struct page *page;
+
+       for (pte = pte_offset_map(pmd, addr);
+            addr < end; addr += PAGE_SIZE, pte++) {
+               if (!pte_present(*pte))
+                       continue;
+
+               page = pte_page(*pte);
+               pte_clear(mm, addr, pte);
+               set_direct_map_default_noflush(page);
+
+               /*
+                * scrub page contents. since mm teardown happens from a
+                * different mm, we cannot just use the process-local virtual
+                * address; access the page via the physmap instead. note that
+                * there is a small time frame where leftover data is globally
+                * visible in the kernel address space.
+                *
+                * tbd in later commit: scrub the page via a temporary mapping
+                * in process-local memory area before re-attaching it to the
+                * physmap.
+                */
+               memset(page_to_virt(page), 0, PAGE_SIZE);
+
+               /*
+                * track page for cleanup later;
+                * note that the proclocal_next list is used only for regular
+                * kfree_proclocal, so ripping pages out now is fine.
+                */
+               INIT_LIST_HEAD(&page->proclocal_next);
+               list_add_tail(&page->proclocal_next, page_list);
+       }
+}
+
+/*
+ * Walk through process-local mappings on each page table level. Avoid code
+ * duplication and use a macro to generate one function for each level.
+ *
+ * The macro generates a function for page table level LEVEL. The function is
+ * passed a pointer to the entry in the page table level ABOVE and recurses 
into
+ * the page table level BELOW.
+ */
+#define UNMAP_LEFTOVER_LEVEL(LEVEL, ABOVE, BELOW) \
+       static void unmap_leftover_pages_ ## LEVEL (struct mm_struct *mm, ABOVE 
## _t *ABOVE,   \
+                                                   unsigned long addr, 
unsigned long end,      \
+                                                   struct list_head 
*page_list)                \
+       {                                                                       
        \
+               LEVEL ## _t *LEVEL  = LEVEL ## _offset(ABOVE, addr);            
        \
+               unsigned long next;                                             
        \
+               do {                                                            
        \
+                       next = LEVEL ## _addr_end(addr, end);                   
        \
+                       if (LEVEL ## _present(*LEVEL))                          
        \
+                               unmap_leftover_pages_## BELOW (mm, LEVEL, addr, 
next, page_list); \
+               } while (LEVEL++, addr = next, addr < end);                     
        \
+       }
+
+UNMAP_LEFTOVER_LEVEL(pmd, pud, pte)
+UNMAP_LEFTOVER_LEVEL(pud, p4d, pmd)
+UNMAP_LEFTOVER_LEVEL(p4d, pgd, pud)
+#undef UNMAP_LEFTOVER_LEVEL
+
+extern void proclocal_release_pages(struct list_head *pages);
+
+static void unmap_free_leftover_proclocal_pages(struct mm_struct *mm)
+{
+       LIST_HEAD(page_list);
+       unsigned long addr = PROCLOCAL_START, next;
+       unsigned long end = PROCLOCAL_START + PROCLOCAL_SIZE;
+
+       /*
+        * Walk page tables in process-local memory area and handle leftover
+        * process-local pages. Note that we cannot use the kernel's
+        * walk_page_range, because that function assumes walking across vmas.
+        */
+       spin_lock(&mm->page_table_lock);
+       do {
+               pgd_t *pgd = pgd_offset(mm, addr);
+               next = pgd_addr_end(addr, end);
+
+               if (pgd_present(*pgd)) {
+                       unmap_leftover_pages_p4d(mm, pgd, addr, next, 
&page_list);
+               }
+               addr = next;
+       } while (addr < end);
+       spin_unlock(&mm->page_table_lock);
+       /*
+        * Flush any mappings of process-local pages from the TLBs, so that we
+        * can release the pages afterwards.
+        */
+       flush_tlb_mm_range(mm, PROCLOCAL_START, end, PAGE_SHIFT, false);
+       proclocal_release_pages(&page_list);
+}
+
+static void arch_proclocal_teardown_pt(struct mm_struct *mm)
+{
+       struct mmu_gather tlb;
+       /*
+        * clean up page tables for the whole pgd used exclusively by
+        * process-local memory.
+        */
+       unsigned long proclocal_base_pgd = PROCLOCAL_START & PGDIR_MASK;
+       unsigned long proclocal_end_pgd = proclocal_base_pgd + PGDIR_SIZE;
+
+       tlb_gather_mmu(&tlb, mm, proclocal_base_pgd, proclocal_end_pgd);
+       free_pgd_range(&tlb, proclocal_base_pgd, proclocal_end_pgd, 0, 0);
+       tlb_finish_mmu(&tlb, proclocal_base_pgd, proclocal_end_pgd);
+}
+
+void arch_proclocal_teardown_pages_and_pt(struct mm_struct *mm)
+{
+       unmap_free_leftover_proclocal_pages(mm);
+       arch_proclocal_teardown_pt(mm);
+}
diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h
index 2fe4dbfcdebd..1cb9243dd299 100644
--- a/include/linux/mm_types.h
+++ b/include/linux/mm_types.h
@@ -158,6 +158,15 @@ struct page {
 
                /** @rcu_head: You can use this to free a page by RCU. */
                struct rcu_head rcu_head;
+
+#ifdef CONFIG_PROCLOCAL
+               struct {        /* PROCLOCAL pages */
+                       struct list_head proclocal_next; /* track pages in one 
allocation */
+                       unsigned long _proclocal_pad_1; /* mapping */
+                       /* head page of an allocation stores its length */
+                       size_t proclocal_nr_pages;
+               };
+#endif
        };
 
        union {         /* This union is 4 bytes in size. */
diff --git a/include/linux/proclocal.h b/include/linux/proclocal.h
new file mode 100644
index 000000000000..9dae140c0796
--- /dev/null
+++ b/include/linux/proclocal.h
@@ -0,0 +1,16 @@
+/*
+ * Copyright 2019 Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ */
+#ifndef _PROCLOCAL_H
+#define _PROCLOCAL_H
+
+#ifdef CONFIG_PROCLOCAL
+
+struct mm_struct;
+
+void proclocal_mm_exit(struct mm_struct *mm);
+#else  /* !CONFIG_PROCLOCAL */
+static inline void proclocal_mm_exit(struct mm_struct *mm) { }
+#endif
+
+#endif /* _PROCLOCAL_H */
diff --git a/kernel/fork.c b/kernel/fork.c
index 6e37f5626417..caca6b16ee1e 100644
--- a/kernel/fork.c
+++ b/kernel/fork.c
@@ -92,6 +92,7 @@
 #include <linux/livepatch.h>
 #include <linux/thread_info.h>
 #include <linux/stackleak.h>
+#include <linux/proclocal.h>
 
 #include <asm/pgtable.h>
 #include <asm/pgalloc.h>
@@ -1051,6 +1052,11 @@ static inline void __mmput(struct mm_struct *mm)
        exit_mmap(mm);
        mm_put_huge_zero_page(mm);
        set_mm_exe_file(mm, NULL);
+       /*
+        * No real users of this address space left, dropping process-local
+        * mappings.
+        */
+       proclocal_mm_exit(mm);
        if (!list_empty(&mm->mmlist)) {
                spin_lock(&mmlist_lock);
                list_del(&mm->mmlist);
diff --git a/mm/Makefile b/mm/Makefile
index e99e1e6ae5ae..029d7e2ee80b 100644
--- a/mm/Makefile
+++ b/mm/Makefile
@@ -100,3 +100,4 @@ obj-$(CONFIG_PERCPU_STATS) += percpu-stats.o
 obj-$(CONFIG_HMM) += hmm.o
 obj-$(CONFIG_MEMFD_CREATE) += memfd.o
 obj-$(CONFIG_XPFO) += xpfo.o
+obj-$(CONFIG_PROCLOCAL) += proclocal.o
diff --git a/mm/proclocal.c b/mm/proclocal.c
new file mode 100644
index 000000000000..72c485c450bf
--- /dev/null
+++ b/mm/proclocal.c
@@ -0,0 +1,35 @@
+/*
+ * mm/proclocal.c
+ *
+ * The code in this file implements process-local mappings in the Linux kernel
+ * address space. This memory is only usable in the process context. With 
memory
+ * not globally visible in the kernel, it cannot easily be prefetched and 
leaked
+ * via L1TF.
+ *
+ * Copyright 2019 Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ */
+#include <linux/mm.h>
+#include <linux/proclocal.h>
+#include <linux/set_memory.h>
+#include <linux/slab.h>
+#include <linux/spinlock.h>
+
+#include <asm/proclocal.h>
+#include <asm/pgalloc.h>
+#include <asm/tlb.h>
+
+void proclocal_release_pages(struct list_head *pages)
+{
+       struct page *pos, *n;
+       list_for_each_entry_safe(pos, n, pages, proclocal_next) {
+               list_del(&pos->proclocal_next);
+               __free_page(pos);
+       }
+}
+
+void proclocal_mm_exit(struct mm_struct *mm)
+{
+       pr_debug("proclocal_mm_exit for mm %p pgd %p (current is %p)\n", mm, 
mm->pgd, current);
+
+       arch_proclocal_teardown_pages_and_pt(mm);
+}
-- 
2.21.0

Reply via email to