new file mode 100644
@@ -0,0 +1,11 @@
+/*
+ * Copyright 2019 Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ */
+#ifndef _ASM_X86_PROCLOCAL_H
+#define _ASM_X86_PROCLOCAL_H
+
+struct mm_struct;
+
+void arch_proclocal_teardown_pages_and_pt(struct mm_struct *mm);
+
+#endif /* _ASM_X86_PROCLOCAL_H */
@@ -56,3 +56,4 @@ obj-$(CONFIG_AMD_MEM_ENCRYPT) += mem_encrypt_boot.o
obj-$(CONFIG_XPFO) += xpfo.o
+obj-$(CONFIG_PROCLOCAL) += proclocal.o
new file mode 100644
@@ -0,0 +1,131 @@
+/*
+ * Architecture-specific code for handling process-local memory on x86-64.
+ *
+ * Copyright 2019 Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ */
+
+#include <linux/list.h>
+#include <linux/mm.h>
+#include <linux/proclocal.h>
+#include <linux/set_memory.h>
+#include <asm/tlb.h>
+
+
+static void unmap_leftover_pages_pte(struct mm_struct *mm, pmd_t *pmd,
+ unsigned long addr, unsigned long end,
+ struct list_head *page_list)
+{
+ pte_t *pte;
+ struct page *page;
+
+ for (pte = pte_offset_map(pmd, addr);
+ addr < end; addr += PAGE_SIZE, pte++) {
+ if (!pte_present(*pte))
+ continue;
+
+ page = pte_page(*pte);
+ pte_clear(mm, addr, pte);
+ set_direct_map_default_noflush(page);
+
+ /*
+ * scrub page contents. since mm teardown happens from a
+ * different mm, we cannot just use the process-local virtual
+ * address; access the page via the physmap instead. note that
+ * there is a small time frame where leftover data is globally
+ * visible in the kernel address space.
+ *
+ * tbd in later commit: scrub the page via a temporary mapping
+ * in process-local memory area before re-attaching it to the
+ * physmap.
+ */
+ memset(page_to_virt(page), 0, PAGE_SIZE);
+
+ /*
+ * track page for cleanup later;
+ * note that the proclocal_next list is used only for regular
+ * kfree_proclocal, so ripping pages out now is fine.
+ */
+ INIT_LIST_HEAD(&page->proclocal_next);
+ list_add_tail(&page->proclocal_next, page_list);
+ }
+}
+
+/*
+ * Walk through process-local mappings on each page table level. Avoid code
+ * duplication and use a macro to generate one function for each level.
+ *
+ * The macro generates a function for page table level LEVEL. The function is
+ * passed a pointer to the entry in the page table level ABOVE and recurses into
+ * the page table level BELOW.
+ */
+#define UNMAP_LEFTOVER_LEVEL(LEVEL, ABOVE, BELOW) \
+ static void unmap_leftover_pages_ ## LEVEL (struct mm_struct *mm, ABOVE ## _t *ABOVE, \
+ unsigned long addr, unsigned long end, \
+ struct list_head *page_list) \
+ { \
+ LEVEL ## _t *LEVEL = LEVEL ## _offset(ABOVE, addr); \
+ unsigned long next; \
+ do { \
+ next = LEVEL ## _addr_end(addr, end); \
+ if (LEVEL ## _present(*LEVEL)) \
+ unmap_leftover_pages_## BELOW (mm, LEVEL, addr, next, page_list); \
+ } while (LEVEL++, addr = next, addr < end); \
+ }
+
+UNMAP_LEFTOVER_LEVEL(pmd, pud, pte)
+UNMAP_LEFTOVER_LEVEL(pud, p4d, pmd)
+UNMAP_LEFTOVER_LEVEL(p4d, pgd, pud)
+#undef UNMAP_LEFTOVER_LEVEL
+
+extern void proclocal_release_pages(struct list_head *pages);
+
+static void unmap_free_leftover_proclocal_pages(struct mm_struct *mm)
+{
+ LIST_HEAD(page_list);
+ unsigned long addr = PROCLOCAL_START, next;
+ unsigned long end = PROCLOCAL_START + PROCLOCAL_SIZE;
+
+ /*
+ * Walk page tables in process-local memory area and handle leftover
+ * process-local pages. Note that we cannot use the kernel's
+ * walk_page_range, because that function assumes walking across vmas.
+ */
+ spin_lock(&mm->page_table_lock);
+ do {
+ pgd_t *pgd = pgd_offset(mm, addr);
+ next = pgd_addr_end(addr, end);
+
+ if (pgd_present(*pgd)) {
+ unmap_leftover_pages_p4d(mm, pgd, addr, next, &page_list);
+ }
+ addr = next;
+ } while (addr < end);
+ spin_unlock(&mm->page_table_lock);
+ /*
+ * Flush any mappings of process-local pages from the TLBs, so that we
+ * can release the pages afterwards.
+ */
+ flush_tlb_mm_range(mm, PROCLOCAL_START, end, PAGE_SHIFT, false);
+ proclocal_release_pages(&page_list);
+}
+
+static void arch_proclocal_teardown_pt(struct mm_struct *mm)
+{
+ struct mmu_gather tlb;
+ /*
+ * clean up page tables for the whole pgd used exclusively by
+ * process-local memory.
+ */
+ unsigned long proclocal_base_pgd = PROCLOCAL_START & PGDIR_MASK;
+ unsigned long proclocal_end_pgd = proclocal_base_pgd + PGDIR_SIZE;
+
+ tlb_gather_mmu(&tlb, mm, proclocal_base_pgd, proclocal_end_pgd);
+ free_pgd_range(&tlb, proclocal_base_pgd, proclocal_end_pgd, 0, 0);
+ tlb_finish_mmu(&tlb, proclocal_base_pgd, proclocal_end_pgd);
+}
+
+void arch_proclocal_teardown_pages_and_pt(struct mm_struct *mm)
+{
+ unmap_free_leftover_proclocal_pages(mm);
+ arch_proclocal_teardown_pt(mm);
+}
@@ -158,6 +158,15 @@ struct page {
/** @rcu_head: You can use this to free a page by RCU. */
struct rcu_head rcu_head;
+
+#ifdef CONFIG_PROCLOCAL
+ struct { /* PROCLOCAL pages */
+ struct list_head proclocal_next; /* track pages in one allocation */
+ unsigned long _proclocal_pad_1; /* mapping */
+ /* head page of an allocation stores its length */
+ size_t proclocal_nr_pages;
+ };
+#endif
};
union { /* This union is 4 bytes in size. */
new file mode 100644
@@ -0,0 +1,16 @@
+/*
+ * Copyright 2019 Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ */
+#ifndef _PROCLOCAL_H
+#define _PROCLOCAL_H
+
+#ifdef CONFIG_PROCLOCAL
+
+struct mm_struct;
+
+void proclocal_mm_exit(struct mm_struct *mm);
+#else /* !CONFIG_PROCLOCAL */
+static inline void proclocal_mm_exit(struct mm_struct *mm) { }
+#endif
+
+#endif /* _PROCLOCAL_H */
@@ -92,6 +92,7 @@
#include <linux/livepatch.h>
#include <linux/thread_info.h>
#include <linux/stackleak.h>
+#include <linux/proclocal.h>
#include <asm/pgtable.h>
#include <asm/pgalloc.h>
@@ -1051,6 +1052,11 @@ static inline void __mmput(struct mm_struct *mm)
exit_mmap(mm);
mm_put_huge_zero_page(mm);
set_mm_exe_file(mm, NULL);
+ /*
+ * No real users of this address space left, dropping process-local
+ * mappings.
+ */
+ proclocal_mm_exit(mm);
if (!list_empty(&mm->mmlist)) {
spin_lock(&mmlist_lock);
list_del(&mm->mmlist);
@@ -100,3 +100,4 @@ obj-$(CONFIG_PERCPU_STATS) += percpu-stats.o
obj-$(CONFIG_HMM) += hmm.o
obj-$(CONFIG_MEMFD_CREATE) += memfd.o
obj-$(CONFIG_XPFO) += xpfo.o
+obj-$(CONFIG_PROCLOCAL) += proclocal.o
new file mode 100644
@@ -0,0 +1,35 @@
+/*
+ * mm/proclocal.c
+ *
+ * The code in this file implements process-local mappings in the Linux kernel
+ * address space. This memory is only usable in the process context. With memory
+ * not globally visible in the kernel, it cannot easily be prefetched and leaked
+ * via L1TF.
+ *
+ * Copyright 2019 Amazon.com, Inc. or its affiliates. All Rights Reserved.
+ */
+#include <linux/mm.h>
+#include <linux/proclocal.h>
+#include <linux/set_memory.h>
+#include <linux/slab.h>
+#include <linux/spinlock.h>
+
+#include <asm/proclocal.h>
+#include <asm/pgalloc.h>
+#include <asm/tlb.h>
+
+void proclocal_release_pages(struct list_head *pages)
+{
+ struct page *pos, *n;
+ list_for_each_entry_safe(pos, n, pages, proclocal_next) {
+ list_del(&pos->proclocal_next);
+ __free_page(pos);
+ }
+}
+
+void proclocal_mm_exit(struct mm_struct *mm)
+{
+ pr_debug("proclocal_mm_exit for mm %p pgd %p (current is %p)\n", mm, mm->pgd, current);
+
+ arch_proclocal_teardown_pages_and_pt(mm);
+}
Process-local memory uses a dedicated pgd entry in kernel space and its own page table structure. Hook mm exit functions to cleanup those dedicated page tables. As a preparation, release any left-over process-local allocations in the address space. Signed-off-by: Marius Hillenbrand <mhillenb@amazon.de> Cc: Alexander Graf <graf@amazon.de> Cc: David Woodhouse <dwmw@amazon.co.uk> --- arch/x86/include/asm/proclocal.h | 11 +++ arch/x86/mm/Makefile | 1 + arch/x86/mm/proclocal.c | 131 +++++++++++++++++++++++++++++++ include/linux/mm_types.h | 9 +++ include/linux/proclocal.h | 16 ++++ kernel/fork.c | 6 ++ mm/Makefile | 1 + mm/proclocal.c | 35 +++++++++ 8 files changed, 210 insertions(+) create mode 100644 arch/x86/include/asm/proclocal.h create mode 100644 arch/x86/mm/proclocal.c create mode 100644 include/linux/proclocal.h create mode 100644 mm/proclocal.c