From patchwork Mon Aug 23 13:01:54 2021 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: yaozhenguo X-Patchwork-Id: 12452813 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-13.5 required=3.0 tests=BAYES_00, DKIM_ADSP_CUSTOM_MED,DKIM_INVALID,DKIM_SIGNED,FREEMAIL_FORGED_FROMDOMAIN, FREEMAIL_FROM,HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER, INCLUDES_PATCH,MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,USER_AGENT_GIT autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id C0B7BC432BE for ; Mon, 23 Aug 2021 13:02:11 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 4E15E61214 for ; Mon, 23 Aug 2021 13:02:11 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.4.1 mail.kernel.org 4E15E61214 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=gmail.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=kvack.org Received: by kanga.kvack.org (Postfix) id A80296B006C; Mon, 23 Aug 2021 09:02:10 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id A30016B0072; Mon, 23 Aug 2021 09:02:10 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 91FD38D0001; Mon, 23 Aug 2021 09:02:10 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0189.hostedemail.com [216.40.44.189]) by kanga.kvack.org (Postfix) with ESMTP id 771896B006C for ; Mon, 23 Aug 2021 09:02:10 -0400 (EDT) Received: from smtpin37.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay03.hostedemail.com (Postfix) with ESMTP id 1A4978249980 for ; Mon, 23 Aug 2021 13:02:10 +0000 (UTC) X-FDA: 78506358420.37.4E6076E Received: from mail-pj1-f68.google.com (mail-pj1-f68.google.com [209.85.216.68]) by imf19.hostedemail.com (Postfix) with ESMTP id C49EFB0000BA for ; Mon, 23 Aug 2021 13:02:09 +0000 (UTC) Received: by mail-pj1-f68.google.com with SMTP id z24-20020a17090acb1800b0018e87a24300so1745285pjt.0 for ; Mon, 23 Aug 2021 06:02:09 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=from:to:cc:subject:date:message-id:mime-version :content-transfer-encoding; bh=FzAazooiiJYmFS22cSG4nL3NYpAxYHxNwYwzLrIUQAY=; b=e+KR5PgkrcwTs6bn8OqC9Lvg8zFvcbfgXDhXFOtygO9/DN/8kjBbHGtLlNrKm8zmLN oCOPmBn9hmS4t8g2uDW1la/XJ0N0VnHauGAjR1ipW4aSyIk4SikZNLN2eATR/cFiH0hw bSQaUYCN9DouXUvk+O+zMF0NS4NDDq9yMogoEdbChNhVYvaIaiMujLxa+ZkhVwsGNr0p 6g2tmtNFiJFQbYXolv0Dgnfa4Nz09eKZDeex/txykqVVMyYRn7L7Oaa0GUQCaABr/jlr oDUsHnbAoeNXOP6JG/+y849KP2Q15mAWBakhgp3dYpKHyOzdYEcftw7DsFz0Wm5zT8Le 43Jg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:to:cc:subject:date:message-id:mime-version :content-transfer-encoding; bh=FzAazooiiJYmFS22cSG4nL3NYpAxYHxNwYwzLrIUQAY=; b=CS+g0A3JVcAWpJgp0JG0qpqAAmK5mN4CtJ625kME8i+7+CKhggw+YfFS4qSEFwK9QE u3W9nQupCJUIBJXFx0WyWIvGpUlyn01Net/xx/nCjW8drt+vVZX6XITl+BMjlxMjSXY6 5it5PrWrY5NyzltstAnQl9vu1biM58jc2fiYWUKoO3LSBXjki82zgBJkBunHAEcCuqqS 56q4TzAQ89u9Q0ZcrXPEs1IKwVdt5X++XJkDmYvnZvvs2Yj8wF1TThslq8PAoTNVQcKG QFzPbxUbvwy/x2YB/bqJ0bRy4SGh4aoqph0/YhCV7XOs+iGi6Dae18eyCMtg8zDy7fjt z4WA== X-Gm-Message-State: AOAM531m1cSF9J+FM7tgd+MDwrfzD/zyjwf8HrpNzYWRfkSCP+d2/EZf w/OkdvOClmvBt3LkrH9/jRw= X-Google-Smtp-Source: ABdhPJxyAhjyxa7X7HpxLpaN/GLp3/A/VE9grFMkAcbX/vfXa8QilJZXsL1zQOa4RvSECKVQzggh9w== X-Received: by 2002:a17:90a:b015:: with SMTP id x21mr19551722pjq.26.1629723728579; Mon, 23 Aug 2021 06:02:08 -0700 (PDT) Received: from localhost.localdomain (5e.8a.38a9.ip4.static.sl-reverse.com. [169.56.138.94]) by smtp.gmail.com with ESMTPSA id t10sm19217079pji.30.2021.08.23.06.02.05 (version=TLS1_2 cipher=ECDHE-ECDSA-AES128-GCM-SHA256 bits=128/128); Mon, 23 Aug 2021 06:02:08 -0700 (PDT) From: yaozhenguo To: mike.kravetz@oracle.com, corbet@lwn.net, akpm@linux-foundation.org Cc: yaozhenguo@jd.com, willy@infradead.org, linux-kernel@vger.kernel.org, linux-doc@vger.kernel.org, linux-mm@kvack.org, yaozhenguo Subject: [PATCH V1] hugetlbfs: Extend the definition of hugepages parameter to support node allocation Date: Mon, 23 Aug 2021 21:01:54 +0800 Message-Id: <20210823130154.75070-1-yaozhenguo1@gmail.com> X-Mailer: git-send-email 2.32.0 MIME-Version: 1.0 Authentication-Results: imf19.hostedemail.com; dkim=pass header.d=gmail.com header.s=20161025 header.b=e+KR5Pgk; spf=pass (imf19.hostedemail.com: domain of yaozhenguo1@gmail.com designates 209.85.216.68 as permitted sender) smtp.mailfrom=yaozhenguo1@gmail.com; dmarc=pass (policy=none) header.from=gmail.com X-Stat-Signature: ya3pwq41tyz9h1x8mdrzskwfpoigtcpk X-Rspamd-Queue-Id: C49EFB0000BA X-Rspamd-Server: rspam04 X-HE-Tag: 1629723729-674563 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: We can specify the number of hugepages to allocate at boot. But the hugepages is balanced in all nodes at present. In some scenarios, we only need hugepags in one node. For example: DPDK needs hugepages which is in the same node as NIC. if DPDK needs four hugepags of 1G size in node1 and system has 16 numa nodes. We must reserve 64 hugepags in kernel cmdline. But, only four hugepages is used. The others should be free after boot.If the system memory is low(for example: 64G), it will be an impossible task. So, extend hugepages kernel parameter to specify node number of hugepages to allocate at boot. For example add following parameter: hugepagesz=1G hugepages=0:1,1:3 It will allocate 1 hugepags in node0 and 3 hugepages in node1. Signed-off-by: yaozhenguo --- .../admin-guide/kernel-parameters.txt | 8 +- include/linux/hugetlb.h | 1 + mm/hugetlb.c | 122 ++++++++++++++++-- 3 files changed, 115 insertions(+), 16 deletions(-) diff --git a/Documentation/admin-guide/kernel-parameters.txt b/Documentation/admin-guide/kernel-parameters.txt index bdb22006f..64a128924 100644 --- a/Documentation/admin-guide/kernel-parameters.txt +++ b/Documentation/admin-guide/kernel-parameters.txt @@ -1588,9 +1588,11 @@ the number of pages of hugepagesz to be allocated. If this is the first HugeTLB parameter on the command line, it specifies the number of pages to allocate for - the default huge page size. See also - Documentation/admin-guide/mm/hugetlbpage.rst. - Format: + the default huge page size. If using node format, It + specifies numbers of hugepage in a specific node. + See also Documentation/admin-guide/mm/hugetlbpage.rst. + Format: or (node format) + :[,:] hugepagesz= [HW] The size of the HugeTLB pages. This is used in diff --git a/include/linux/hugetlb.h b/include/linux/hugetlb.h index f7ca1a387..5939ecd4f 100644 --- a/include/linux/hugetlb.h +++ b/include/linux/hugetlb.h @@ -605,6 +605,7 @@ struct hstate { unsigned long nr_overcommit_huge_pages; struct list_head hugepage_activelist; struct list_head hugepage_freelists[MAX_NUMNODES]; + unsigned int max_huge_pages_node[MAX_NUMNODES]; unsigned int nr_huge_pages_node[MAX_NUMNODES]; unsigned int free_huge_pages_node[MAX_NUMNODES]; unsigned int surplus_huge_pages_node[MAX_NUMNODES]; diff --git a/mm/hugetlb.c b/mm/hugetlb.c index dfc940d52..b898cddcd 100644 --- a/mm/hugetlb.c +++ b/mm/hugetlb.c @@ -66,6 +66,7 @@ static struct hstate * __initdata parsed_hstate; static unsigned long __initdata default_hstate_max_huge_pages; static bool __initdata parsed_valid_hugepagesz = true; static bool __initdata parsed_default_hugepagesz; +static unsigned int default_hugepages_in_node[MAX_NUMNODES] __initdata; /* * Protects updates to hugepage_freelists, hugepage_activelist, nr_huge_pages, @@ -2842,10 +2843,56 @@ static void __init gather_bootmem_prealloc(void) } } +static void __init hugetlb_hstate_alloc_pages_onenode(struct hstate *h, int nid) +{ + unsigned long i; + + for (i = 0; i < h->max_huge_pages_node[nid]; i++) { + if (hstate_is_gigantic(h)) { + struct huge_bootmem_page *m; + void *addr; + + addr = memblock_alloc_try_nid_raw( + huge_page_size(h), huge_page_size(h), + 0, MEMBLOCK_ALLOC_ACCESSIBLE, nid); + if (!addr) + break; + m = addr; + BUG_ON(!IS_ALIGNED(virt_to_phys(m), huge_page_size(h))); + /* Put them into a private list first because mem_map is not up yet */ + INIT_LIST_HEAD(&m->list); + list_add(&m->list, &huge_boot_pages); + m->hstate = h; + } else { + struct page *page; + + gfp_t gfp_mask = htlb_alloc_mask(h) | __GFP_THISNODE; + + page = alloc_fresh_huge_page(h, gfp_mask, nid, + &node_states[N_MEMORY], NULL); + if (page) + put_page(page); /* free it into the hugepage allocator */ + + } + } +} + static void __init hugetlb_hstate_alloc_pages(struct hstate *h) { unsigned long i; nodemask_t *node_alloc_noretry; + bool hugetlb_node_set = false; + + /* do node alloc */ + for (i = 0; i < nodes_weight(node_states[N_MEMORY]); i++) { + if (h->max_huge_pages_node[i] > 0) { + hugetlb_hstate_alloc_pages_onenode(h, i); + hugetlb_node_set = true; + } + } + + if (hugetlb_node_set) + return; if (!hstate_is_gigantic(h)) { /* @@ -3580,6 +3627,9 @@ static int __init hugetlb_init(void) default_hstate_max_huge_pages; } } + for (i = 0; i < nodes_weight(node_states[N_MEMORY]); i++) + if (default_hugepages_in_node[i] > 0) + default_hstate.max_huge_pages_node[i] = default_hugepages_in_node[i]; hugetlb_cma_check(); hugetlb_init_hstates(); @@ -3649,6 +3699,11 @@ static int __init hugepages_setup(char *s) { unsigned long *mhp; static unsigned long *last_mhp; + unsigned int node = NUMA_NO_NODE; + int ret; + int count; + unsigned long tmp; + char *p = s; if (!parsed_valid_hugepagesz) { pr_warn("HugeTLB: hugepages=%s does not follow a valid hugepagesz, ignoring\n", s); @@ -3656,25 +3711,66 @@ static int __init hugepages_setup(char *s) return 0; } - /* - * !hugetlb_max_hstate means we haven't parsed a hugepagesz= parameter - * yet, so this hugepages= parameter goes to the "default hstate". - * Otherwise, it goes with the previously parsed hugepagesz or - * default_hugepagesz. - */ - else if (!hugetlb_max_hstate) - mhp = &default_hstate_max_huge_pages; - else - mhp = &parsed_hstate->max_huge_pages; + while (*p) { + count = 0; + ret = sscanf(p, "%lu%n", &tmp, &count); + if (ret != 1) { + pr_warn("HugeTLB: Invalid hugepages parameter %s\n", p); + break; + } + /* Parameter is not node format */ + if (p[count] != ':') { + /* + * !hugetlb_max_hstate means we haven't parsed a hugepagesz= parameter + * yet, so this hugepages= parameter goes to the "default hstate". + * Otherwise, it goes with the previously parsed hugepagesz or + * default_hugepagesz. + */ + if (!hugetlb_max_hstate) { + default_hstate_max_huge_pages = tmp; + mhp = &default_hstate_max_huge_pages; + } else { + parsed_hstate->max_huge_pages = tmp; + mhp = &parsed_hstate->max_huge_pages; + } + break; + } + /* Parameter is node format */ + node = tmp; + p += count + 1; + if (node < 0) { + pr_warn("HugeTLB: Invalid hugepages parameter node:%d\n", node); + break; + } + if (!hugetlb_max_hstate) + mhp = (unsigned long *) + &(default_hugepages_in_node[node]); + else + mhp = (unsigned long *) + &(parsed_hstate->max_huge_pages_node[node]); + /* Parse hugepages */ + ret = sscanf(p, "%lu%n", mhp, &count); + if (ret != 1) { + pr_warn("HugeTLB: Invalid parameter %s\n", p); + *mhp = 0; + break; + } + if (!hugetlb_max_hstate) + default_hstate_max_huge_pages += *mhp; + else + parsed_hstate->max_huge_pages += *mhp; + /* Go to parse next node*/ + if (p[count] == ',') + p += count + 1; + else + break; + } if (mhp == last_mhp) { pr_warn("HugeTLB: hugepages= specified twice without interleaving hugepagesz=, ignoring hugepages=%s\n", s); return 0; } - if (sscanf(s, "%lu", mhp) <= 0) - *mhp = 0; - /* * Global state is always initialized later in hugetlb_init. * But we need to allocate gigantic hstates here early to still