@@ -2343,7 +2343,7 @@ EXPORT_SYMBOL(alloc_pages_vma);
struct page *alloc_pages_current(gfp_t gfp, unsigned order)
{
struct mempolicy *pol = &default_policy;
- struct page *page;
+ int nid = NUMA_NO_NODE;
if (!in_interrupt() && !(gfp & __GFP_THISNODE))
pol = get_task_policy(current);
@@ -2353,14 +2353,9 @@ struct page *alloc_pages_current(gfp_t gfp, unsigned order)
* nor system default_policy
*/
if (pol->mode == MPOL_INTERLEAVE)
- page = alloc_pages_policy(pol, gfp, order,
- interleave_nodes(pol));
- else
- page = __alloc_pages_nodemask(gfp, order,
- policy_node(gfp, pol, numa_node_id()),
- policy_nodemask(gfp, pol));
+ nid = interleave_nodes(pol);
- return page;
+ return alloc_pages_policy(pol, gfp, order, nid);
}
EXPORT_SYMBOL(alloc_pages_current);
In order to support MPOL_PREFERRED_MANY as the mode used by set_mempolicy(2), alloc_pages_current() needs to support it. This patch does that by using the new helper function to allocate properly based on policy. All the actual machinery to make this work was part of ("mm/mempolicy: Create a page allocator for policy") Link: https://lore.kernel.org/r/20200630212517.308045-10-ben.widawsky@intel.com Signed-off-by: Ben Widawsky <ben.widawsky@intel.com> --- mm/mempolicy.c | 11 +++-------- 1 file changed, 3 insertions(+), 8 deletions(-)