Message ID | 20190129003422.9328-15-rick.p.edgecombe@intel.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | Merge text_poke fixes and executable lockdowns | expand |
On Mon, Jan 28, 2019 at 04:34:16PM -0800, Rick Edgecombe wrote: > For architectures with CONFIG_ARCH_HAS_SET_ALIAS, pages can be unmapped > briefly on the directmap, even when CONFIG_DEBUG_PAGEALLOC is not > configured. So this changes kernel_map_pages and kernel_page_present to be s/this changes/change/ From Documentation/process/submitting-patches.rst: "Describe your changes in imperative mood, e.g. "make xyzzy do frotz" instead of "[This patch] makes xyzzy do frotz" or "[I] changed xyzzy to do frotz", as if you are giving orders to the codebase to change its behaviour." Also, please end function names with parentheses. > defined when CONFIG_ARCH_HAS_SET_ALIAS is defined as well. It also changes > places (page_alloc.c) where those functions are assumed to only be > implemented when CONFIG_DEBUG_PAGEALLOC is defined. The commit message doesn't need to say "what" you're doing - that should be obvious from the diff below. It should rather say "why" you're doing it. > So now when CONFIG_ARCH_HAS_SET_ALIAS=y, hibernate will handle not present > page when saving. Previously this was already done when pages > CONFIG_DEBUG_PAGEALLOC was configured. It does not appear to have a big > hibernating performance impact. Comment over safe_copy_page() needs updating I guess. > Before: > [ 4.670938] PM: Wrote 171996 kbytes in 0.21 seconds (819.02 MB/s) > > After: > [ 4.504714] PM: Wrote 178932 kbytes in 0.22 seconds (813.32 MB/s) IINM, that's like 1734 pages more. How am I to understand this number? Code has called set_alias_nv_noflush() on them and safe_copy_page() now maps them one by one to copy them to the hibernation image? Thx.
On Tue, 2019-02-19 at 12:04 +0100, Borislav Petkov wrote: > On Mon, Jan 28, 2019 at 04:34:16PM -0800, Rick Edgecombe wrote: > > For architectures with CONFIG_ARCH_HAS_SET_ALIAS, pages can be unmapped > > briefly on the directmap, even when CONFIG_DEBUG_PAGEALLOC is not > > configured. So this changes kernel_map_pages and kernel_page_present to be > > s/this changes/change/ > > From Documentation/process/submitting-patches.rst: > > "Describe your changes in imperative mood, e.g. "make xyzzy do frotz" > instead of "[This patch] makes xyzzy do frotz" or "[I] changed xyzzy > to do frotz", as if you are giving orders to the codebase to change > its behaviour." > > Also, please end function names with parentheses. Yes, gotcha. > > defined when CONFIG_ARCH_HAS_SET_ALIAS is defined as well. It also changes > > places (page_alloc.c) where those functions are assumed to only be > > implemented when CONFIG_DEBUG_PAGEALLOC is defined. > > The commit message doesn't need to say "what" you're doing - that should > be obvious from the diff below. It should rather say "why" you're doing > it. Ok, sorry. I'll change this to be more concise. > > So now when CONFIG_ARCH_HAS_SET_ALIAS=y, hibernate will handle not present > > page when saving. Previously this was already done when > > pages > > > CONFIG_DEBUG_PAGEALLOC was configured. It does not appear to have a big > > hibernating performance impact. > > Comment over safe_copy_page Oh, yes you are right. > > Before: > > [ 4.670938] PM: Wrote 171996 kbytes in 0.21 seconds (819.02 MB/s) > > > > After: > > [ 4.504714] PM: Wrote 178932 kbytes in 0.22 seconds (813.32 MB/s) > > IINM, that's like 1734 pages more. How am I to understand this number? > > Code has called set_alias_nv_noflush() on them and safe_copy_page() now > maps them one by one to copy them to the hibernation image? > > Thx. > These are from logs hibernate generates. The concern was that hibernate could be slightly slower because of the checking of whether the pages are mapped. The bandwidth number can be used to compare, 819.02->813.32 MB/s. Some randomness must have resulted in different amounts of memory used between tests. I can just remove the log lines and include the bandwidth numbers. Thanks, Rick
On Tue, Feb 19, 2019 at 09:28:55PM +0000, Edgecombe, Rick P wrote: > These are from logs hibernate generates. The concern was that hibernate could be > slightly slower because of the checking of whether the pages are mapped. The > bandwidth number can be used to compare, 819.02->813.32 MB/s. Some randomness > must have resulted in different amounts of memory used between tests. I can just > remove the log lines and include the bandwidth numbers. Nah, I'm just trying to get an idea of the slowdown it would cause. I'm thinking these pages are, as you call them "special" so they should not be a huge chunk of all the system's pages, even if it is a machine with a lot of memory so I guess it ain't that bad. We should keep an eye on it though... :-) Thx.
diff --git a/arch/x86/mm/pageattr.c b/arch/x86/mm/pageattr.c index 3a51915a1410..717bdc188aab 100644 --- a/arch/x86/mm/pageattr.c +++ b/arch/x86/mm/pageattr.c @@ -2257,7 +2257,6 @@ int set_alias_default_noflush(struct page *page) return __set_pages_p(page, 1); } -#ifdef CONFIG_DEBUG_PAGEALLOC void __kernel_map_pages(struct page *page, int numpages, int enable) { if (PageHighMem(page)) @@ -2302,11 +2301,8 @@ bool kernel_page_present(struct page *page) pte = lookup_address((unsigned long)page_address(page), &level); return (pte_val(*pte) & _PAGE_PRESENT); } - #endif /* CONFIG_HIBERNATION */ -#endif /* CONFIG_DEBUG_PAGEALLOC */ - int __init kernel_map_pages_in_pgd(pgd_t *pgd, u64 pfn, unsigned long address, unsigned numpages, unsigned long page_flags) { diff --git a/include/linux/mm.h b/include/linux/mm.h index 80bb6408fe73..b362a280a919 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h @@ -2642,37 +2642,31 @@ static inline void kernel_poison_pages(struct page *page, int numpages, int enable) { } #endif -#ifdef CONFIG_DEBUG_PAGEALLOC extern bool _debug_pagealloc_enabled; -extern void __kernel_map_pages(struct page *page, int numpages, int enable); static inline bool debug_pagealloc_enabled(void) { - return _debug_pagealloc_enabled; + return IS_ENABLED(CONFIG_DEBUG_PAGEALLOC) && _debug_pagealloc_enabled; } +#if defined(CONFIG_DEBUG_PAGEALLOC) || defined(CONFIG_ARCH_HAS_SET_ALIAS) +extern void __kernel_map_pages(struct page *page, int numpages, int enable); + static inline void kernel_map_pages(struct page *page, int numpages, int enable) { - if (!debug_pagealloc_enabled()) - return; - __kernel_map_pages(page, numpages, enable); } #ifdef CONFIG_HIBERNATION extern bool kernel_page_present(struct page *page); #endif /* CONFIG_HIBERNATION */ -#else /* CONFIG_DEBUG_PAGEALLOC */ +#else /* CONFIG_DEBUG_PAGEALLOC || CONFIG_ARCH_HAS_SET_ALIAS */ static inline void kernel_map_pages(struct page *page, int numpages, int enable) {} #ifdef CONFIG_HIBERNATION static inline bool kernel_page_present(struct page *page) { return true; } #endif /* CONFIG_HIBERNATION */ -static inline bool debug_pagealloc_enabled(void) -{ - return false; -} -#endif /* CONFIG_DEBUG_PAGEALLOC */ +#endif /* CONFIG_DEBUG_PAGEALLOC || CONFIG_ARCH_HAS_SET_ALIAS */ #ifdef __HAVE_ARCH_GATE_AREA extern struct vm_area_struct *get_gate_vma(struct mm_struct *mm); diff --git a/mm/page_alloc.c b/mm/page_alloc.c index d295c9bc01a8..92d0a0934274 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -1074,7 +1074,9 @@ static __always_inline bool free_pages_prepare(struct page *page, } arch_free_page(page, order); kernel_poison_pages(page, 1 << order, 0); - kernel_map_pages(page, 1 << order, 0); + if (debug_pagealloc_enabled()) + kernel_map_pages(page, 1 << order, 0); + kasan_free_nondeferred_pages(page, order); return true; @@ -1944,7 +1946,8 @@ inline void post_alloc_hook(struct page *page, unsigned int order, set_page_refcounted(page); arch_alloc_page(page, order); - kernel_map_pages(page, 1 << order, 1); + if (debug_pagealloc_enabled()) + kernel_map_pages(page, 1 << order, 1); kernel_poison_pages(page, 1 << order, 1); kasan_alloc_pages(page, order); set_page_owner(page, order, gfp_flags);