@@ -4318,10 +4318,8 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
* that first
*/
page = get_page_from_freelist(gfp_mask, order, alloc_flags, ac);
- if (page) {
- set_page_refcounted(page);
+ if (page)
goto got_pg;
- }
/*
* For costly allocations, try direct compaction first, as it's likely
@@ -4340,10 +4338,8 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
alloc_flags, ac,
INIT_COMPACT_PRIORITY,
&compact_result);
- if (page) {
- set_page_refcounted(page);
+ if (page)
goto got_pg;
- }
/*
* Checks for costly allocations with __GFP_NORETRY, which
@@ -4403,10 +4399,8 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
/* Attempt with potentially adjusted zonelist and alloc_flags */
page = get_page_from_freelist(gfp_mask, order, alloc_flags, ac);
- if (page) {
- set_page_refcounted(page);
+ if (page)
goto got_pg;
- }
/* Caller is not willing to reclaim, we can't balance anything */
if (!can_direct_reclaim)
@@ -4419,18 +4413,14 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
/* Try direct reclaim and then allocating */
page = __alloc_pages_direct_reclaim(gfp_mask, order, alloc_flags, ac,
&did_some_progress);
- if (page) {
- set_page_refcounted(page);
+ if (page)
goto got_pg;
- }
/* Try direct compaction and then allocating */
page = __alloc_pages_direct_compact(gfp_mask, order, alloc_flags, ac,
compact_priority, &compact_result);
- if (page) {
- set_page_refcounted(page);
+ if (page)
goto got_pg;
- }
/* Do not loop if specifically requested */
if (gfp_mask & __GFP_NORETRY)
@@ -4471,10 +4461,8 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
/* Reclaim has failed us, start killing things */
page = __alloc_pages_may_oom(gfp_mask, order, ac, &did_some_progress);
- if (page) {
- set_page_refcounted(page);
+ if (page)
goto got_pg;
- }
/* Avoid allocations with no watermarks from looping endlessly */
if (tsk_is_oom_victim(current) &&
@@ -4518,10 +4506,8 @@ __alloc_pages_slowpath(gfp_t gfp_mask, unsigned int order,
* the situation worse.
*/
page = __alloc_pages_cpuset_fallback(gfp_mask, order, ALLOC_MIN_RESERVE, ac);
- if (page) {
- set_page_refcounted(page);
+ if (page)
goto got_pg;
- }
cond_resched();
goto retry;
@@ -4813,6 +4799,8 @@ struct page *__alloc_pages_noprof(gfp_t gfp, unsigned int order,
ac.nodemask = nodemask;
page = __alloc_pages_slowpath(alloc_gfp, order, &ac);
+ if (page)
+ set_page_refcounted(page);
out:
if (memcg_kmem_online() && (gfp & __GFP_ACCOUNT) && page &&
In preparation for allocating frozen pages, stop initialising the page refcount in __alloc_pages_slowpath(). Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org> --- mm/page_alloc.c | 30 +++++++++--------------------- 1 file changed, 9 insertions(+), 21 deletions(-)