From patchwork Fri Oct 27 03:38:45 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Peng Zhang X-Patchwork-Id: 13438103 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id BC58DC25B47 for ; Fri, 27 Oct 2023 03:40:19 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 3661B6B0399; Thu, 26 Oct 2023 23:40:19 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 313996B039A; Thu, 26 Oct 2023 23:40:19 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 1B4D26B039B; Thu, 26 Oct 2023 23:40:19 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0011.hostedemail.com [216.40.44.11]) by kanga.kvack.org (Postfix) with ESMTP id 0ADA26B0399 for ; Thu, 26 Oct 2023 23:40:19 -0400 (EDT) Received: from smtpin09.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay04.hostedemail.com (Postfix) with ESMTP id DEEA91A0FF7 for ; Fri, 27 Oct 2023 03:40:18 +0000 (UTC) X-FDA: 81389838516.09.F420C5C Received: from mail-pj1-f50.google.com (mail-pj1-f50.google.com [209.85.216.50]) by imf17.hostedemail.com (Postfix) with ESMTP id 0C85340007 for ; Fri, 27 Oct 2023 03:40:16 +0000 (UTC) Authentication-Results: imf17.hostedemail.com; dkim=pass header.d=bytedance.com header.s=google header.b="gjD8HFd/"; dmarc=pass (policy=quarantine) header.from=bytedance.com; spf=pass (imf17.hostedemail.com: domain of zhangpeng.00@bytedance.com designates 209.85.216.50 as permitted sender) smtp.mailfrom=zhangpeng.00@bytedance.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1698378017; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=Ky4w3uAtFSkvxKCreCTuy48cfBYAXj8C3Bu1f6A1ISE=; b=2iXDswc3Ey56BCMPwbnNcBM3T0TigXDenTX8Q5dDiRPsqbEz5s0nzcFwHiSYPN1z85dUNq 7qi65ijRlgoEmCZsz7dd3mzAECwtK0Sfkw8TY/y3GBQ9AcdX6ArY1dmBmSHD7monXAPjvk 4oU484pLV1d/Vs7yxqvisMulabQMWKg= ARC-Authentication-Results: i=1; imf17.hostedemail.com; dkim=pass header.d=bytedance.com header.s=google header.b="gjD8HFd/"; dmarc=pass (policy=quarantine) header.from=bytedance.com; spf=pass (imf17.hostedemail.com: domain of zhangpeng.00@bytedance.com designates 209.85.216.50 as permitted sender) smtp.mailfrom=zhangpeng.00@bytedance.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1698378017; a=rsa-sha256; cv=none; b=sux4YBVvnFJySsPD30fUhJ4kMBtJeADHvAod/bEx6voPzFEc66+5DhiH6CDkApMuRiCEN9 dXpCJ2JXxhCZi4DQNiSLjD0pPxA8l1IlDgYr0jExxRlB2eUuxOZwuNMjwLIznend0HKLCv 2EvaNUhahlOZqUJYf7BKABKGFnsk/gc= Received: by mail-pj1-f50.google.com with SMTP id 98e67ed59e1d1-27d17f5457fso2115861a91.0 for ; Thu, 26 Oct 2023 20:40:16 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=bytedance.com; s=google; t=1698378016; x=1698982816; darn=kvack.org; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=Ky4w3uAtFSkvxKCreCTuy48cfBYAXj8C3Bu1f6A1ISE=; b=gjD8HFd/3WOtI9QJP3TU+g4B7W5NTroBOuvkcJOWGBo9g14UW1TZxOiUw6L6E3fy4S eVabn8gUqptwMb+qcIyMHqSlvBAHlN/WCPGeGjg2q1ARxLCXaJ1WriJMVekaoxcDldb+ XWVVXOGGFA+q4YvpOQNYhSLOwmVs8G3cv/2eKSzDAUn8p4z1nxXw3wGIbzQA6rYkp0wz Oj+pGqVDSlhuEUu8xsqRh0B/4wLDL7QBGRpe5EY3z7HGxwpDUUp7VPm6AlXMKU3i1hv8 qDU40YDSimSC9m5E04i4joWwZyO42LlqagFWUFGpvm56AwmgMhw1nzW46xMmVkAAJ7ve ueNw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1698378016; x=1698982816; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=Ky4w3uAtFSkvxKCreCTuy48cfBYAXj8C3Bu1f6A1ISE=; b=LR1OkpBvEUlj8sOunHllGZ+poXcZISdf66YD973W0kpnDRO8OCZqKKZKG6V7cXrlkM BeRQfDWYe2sMm6T7EfmrfGnsk5FeGJ3WfrHumElsSIzkhHcnX0j3vCrJxBdPRiN/a+K2 TZQU+YTbXGhIl+bveQbi6MV1/Kancq8vLL0aRiOif4eBpABRWErPkgqT5GXl5fVmmO+3 TtoYtuH63HBQC/n3s5BN1ZKM5ZsR6HzCmEauguWU62zTCyzXikMbGczu5vEqzuv4taSQ 6hsrvnsU+ZvJupjs9Hqut3m1cH0+N6dLeUu1faYN1xOfLABEcKKfHxGUPHQTMhJ2+xfj /XSg== X-Gm-Message-State: AOJu0YxdMHqI3OZVqfx/zZnji8Om8opDysKFglkMyxasUCfS04DgtJ62 D45cS6PbBMc+iSB/ioqq+ThRLg== X-Google-Smtp-Source: AGHT+IFJRCeA+OUFMxyndHyvFf1XFZD0IulFfw6dGetvqF75UjjqNY1IEWlGGaZ1U8Eieo7P5Fin/w== X-Received: by 2002:a17:90b:1bc6:b0:27d:4935:8d9a with SMTP id oa6-20020a17090b1bc600b0027d49358d9amr5489370pjb.4.1698378015924; Thu, 26 Oct 2023 20:40:15 -0700 (PDT) Received: from GL4FX4PXWL.bytedance.net ([203.208.189.7]) by smtp.gmail.com with ESMTPSA id ms19-20020a17090b235300b00267d9f4d340sm2345676pjb.44.2023.10.26.20.40.08 (version=TLS1_3 cipher=TLS_CHACHA20_POLY1305_SHA256 bits=256/256); Thu, 26 Oct 2023 20:40:15 -0700 (PDT) From: Peng Zhang To: Liam.Howlett@oracle.com, corbet@lwn.net, akpm@linux-foundation.org, willy@infradead.org, brauner@kernel.org, surenb@google.com, michael.christie@oracle.com, mjguzik@gmail.com, mathieu.desnoyers@efficios.com, npiggin@gmail.com, peterz@infradead.org, oliver.sang@intel.com, mst@redhat.com Cc: zhangpeng.00@bytedance.com, maple-tree@lists.infradead.org, linux-mm@kvack.org, linux-doc@vger.kernel.org, linux-kernel@vger.kernel.org, linux-fsdevel@vger.kernel.org Subject: [PATCH v7 10/10] fork: Use __mt_dup() to duplicate maple tree in dup_mmap() Date: Fri, 27 Oct 2023 11:38:45 +0800 Message-Id: <20231027033845.90608-11-zhangpeng.00@bytedance.com> X-Mailer: git-send-email 2.39.3 (Apple Git-145) In-Reply-To: <20231027033845.90608-1-zhangpeng.00@bytedance.com> References: <20231027033845.90608-1-zhangpeng.00@bytedance.com> MIME-Version: 1.0 X-Rspamd-Queue-Id: 0C85340007 X-Rspam-User: X-Rspamd-Server: rspam02 X-Stat-Signature: 81s9uda4nkwg6rkz41if6qm13gw4bgtr X-HE-Tag: 1698378016-579926 X-HE-Meta: U2FsdGVkX19xcc3iKL6XF+hcCJ0EcpVljhwzt/IdgEha7WjXmiwM+VGOnDhKbX+/E+ANkh+kNjOMTuLcCyScuvTpvJyl4uqTrmERx7vWX/N9smSNQ2qwqj8FIK1tFKanAaBFN4yJAYBdMFuqU0gP0Cxamrm00QRF2XpwYMolF1xDiJukya1B9zWFjRFU2ucqfbeBxAAXkmTfFrM3/WVqUiUu1xPPOJuYiMCS0V0XU2xD5uZgikGuF9C3JGJQ9LRtY+W2W4JUIQXntRQlo4dKsJ7tPSzOHx8x6E0E1LEsnfSD7qAChsLgPvBzOzruvsxA77AtdHweVHd0HW23aKyaM4coIfb719P0jyN7j6xEt/DFb4cfOrQWiWV1727nY3Z3MlSReAvrt8hgaIpW/H0G8MwmvBcG/UP3ngyU5+oRgocSiV44hkRd5fK39kV9CfiL7eIGMFd5UvNOtw6t6wbccHjgZinHkvUvTIiFwDWaLs4AtjpZmlfB7Sukzd4gL2Ny3isk16X9juryMjraq59wPY/GXTR3bcyuhmQ0zVSkeE7kFtpYSfeiv6T8OP+drf4n1Hbjh6MAKe2uLLJ/psgvPNdkCoaJ00GxkATRNv33LWKymSehDhL8HOzeRTNqEqk6q7GiXjVzEMv6bsbEUTP2SJjuXypX3R5uQol9NxHev9jeAcWh+ghG+iNE6bj+97UzzIXVJ1t4eeeTeuloAFeeRwuBWJbUysX9Q//6IMJph8Bqc3Mwnscoautqs7h0S2Hi6fKAhHbeenn1BOAreYwc4EHRox7hrOJvVzBL6w5N3/Rfch+3+KO5sf/9OTWeWe8Q5XPbOqarHFQ2kMDUYg4wFeeLEZx2lgfqdSbaorTfGwQZ1kefI362xrj8OKBTmidzij1R7fmlTjEI/kIDTWIOiM+wKCAIbzJm9NESGUfXFrYeRxM5PTWp7Dl+gJAWsdnM5bPBYDl4514WIAXOdbs xnMl7nKs WLBCoLKZIH/O78fOmeUmpRvslsgPnJrbDwVt96niT6udglpaU6ppv1Nu+ZaL60o4m5rN7ESTaRB76ZBlOp7DA+1UTKMQkSTgj/+YFKIxXJ4aJ/wJ8BdugNMNw8b2ZcROhqB1DY3nMc+3hQKRE0Nky6mohIeB2TOqZaU+cdee1DybCM1a7jfaH2SB6EwtBThd/Ji/mwESM8svISc6ua9h8dZBCqWPAbPgEQXg/xu4geJaLs3W/nSiWaEWYc558928ogjWgFrCApf2kKxxo4JFCM6ocobMJhn4TmzA17Us+/naXeJvgPlaXf6Rb+26uzP724Rv2vMmRQtn4YJjApYhaUDpJDo/gBX/8D/x8QSNfyR7u8aO39Sz1uuk2fEUTk8nJd1TSKy/GcCEsRpkYm2RK4dZPb5c4RkWdHmfi3In1zmDlc6S5dESzwYszwpJEzHZ8bDMupbnQZZTNun+nop7H/KushVwjiXHVCynOHbzW2s2jtMDvPXp1TWdAQiLEexhg0LRIoRW0wIH3Iasc6vmkapipMuseBEqKI5Hs8o/8C1Wx5gs= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: In dup_mmap(), using __mt_dup() to duplicate the old maple tree and then directly replacing the entries of VMAs in the new maple tree can result in better performance. __mt_dup() uses DFS pre-order to duplicate the maple tree, so it is efficient. The average time complexity of __mt_dup() is O(n), where n is the number of VMAs. The proof of the time complexity is provided in the commit log that introduces __mt_dup(). After duplicating the maple tree, each element is traversed and replaced (ignoring the cases of deletion, which are rare). Since it is only a replacement operation for each element, this process is also O(n). Analyzing the exact time complexity of the previous algorithm is challenging because each insertion can involve appending to a node, pushing data to adjacent nodes, or even splitting nodes. The frequency of each action is difficult to calculate. The worst-case scenario for a single insertion is when the tree undergoes splitting at every level. If we consider each insertion as the worst-case scenario, we can determine that the upper bound of the time complexity is O(n*log(n)), although this is a loose upper bound. However, based on the test data, it appears that the actual time complexity is likely to be O(n). As the entire maple tree is duplicated using __mt_dup(), if dup_mmap() fails, there will be a portion of VMAs that have not been duplicated in the maple tree. To handle this, we mark the failure point with XA_ZERO_ENTRY. In exit_mmap(), if this marker is encountered, stop releasing VMAs that have not been duplicated after this point. There is a "spawn" in byte-unixbench[1], which can be used to test the performance of fork(). I modified it slightly to make it work with different number of VMAs. Below are the test results. The first row shows the number of VMAs. The second and third rows show the number of fork() calls per ten seconds, corresponding to next-20231006 and the this patchset, respectively. The test results were obtained with CPU binding to avoid scheduler load balancing that could cause unstable results. There are still some fluctuations in the test results, but at least they are better than the original performance. 21 121 221 421 821 1621 3221 6421 12821 25621 51221 112100 76261 54227 34035 20195 11112 6017 3161 1606 802 393 114558 83067 65008 45824 28751 16072 8922 4747 2436 1233 599 2.19% 8.92% 19.88% 34.64% 42.37% 44.64% 48.28% 50.17% 51.68% 53.74% 52.42% [1] https://github.com/kdlucas/byte-unixbench/tree/master Signed-off-by: Peng Zhang Suggested-by: Liam R. Howlett Reviewed-by: Liam R. Howlett --- include/linux/mm.h | 11 +++++++++++ kernel/fork.c | 40 +++++++++++++++++++++++++++++----------- mm/internal.h | 11 ----------- mm/memory.c | 7 ++++++- mm/mmap.c | 9 ++++++--- 5 files changed, 52 insertions(+), 26 deletions(-) diff --git a/include/linux/mm.h b/include/linux/mm.h index 14d5aaff96d0..e9111ec5808c 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h @@ -996,6 +996,17 @@ static inline int vma_iter_bulk_alloc(struct vma_iterator *vmi, return mas_expected_entries(&vmi->mas, count); } +static inline int vma_iter_clear_gfp(struct vma_iterator *vmi, + unsigned long start, unsigned long end, gfp_t gfp) +{ + __mas_set_range(&vmi->mas, start, end - 1); + mas_store_gfp(&vmi->mas, NULL, gfp); + if (unlikely(mas_is_err(&vmi->mas))) + return -ENOMEM; + + return 0; +} + /* Free any unused preallocations */ static inline void vma_iter_free(struct vma_iterator *vmi) { diff --git a/kernel/fork.c b/kernel/fork.c index 1e6c656e0857..1552ee66517b 100644 --- a/kernel/fork.c +++ b/kernel/fork.c @@ -650,7 +650,6 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm, int retval; unsigned long charge = 0; LIST_HEAD(uf); - VMA_ITERATOR(old_vmi, oldmm, 0); VMA_ITERATOR(vmi, mm, 0); uprobe_start_dup_mmap(); @@ -678,16 +677,22 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm, goto out; khugepaged_fork(mm, oldmm); - retval = vma_iter_bulk_alloc(&vmi, oldmm->map_count); - if (retval) + /* Use __mt_dup() to efficiently build an identical maple tree. */ + retval = __mt_dup(&oldmm->mm_mt, &mm->mm_mt, GFP_KERNEL); + if (unlikely(retval)) goto out; mt_clear_in_rcu(vmi.mas.tree); - for_each_vma(old_vmi, mpnt) { + for_each_vma(vmi, mpnt) { struct file *file; vma_start_write(mpnt); if (mpnt->vm_flags & VM_DONTCOPY) { + retval = vma_iter_clear_gfp(&vmi, mpnt->vm_start, + mpnt->vm_end, GFP_KERNEL); + if (retval) + goto loop_out; + vm_stat_account(mm, mpnt->vm_flags, -vma_pages(mpnt)); continue; } @@ -749,9 +754,11 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm, if (is_vm_hugetlb_page(tmp)) hugetlb_dup_vma_private(tmp); - /* Link the vma into the MT */ - if (vma_iter_bulk_store(&vmi, tmp)) - goto fail_nomem_vmi_store; + /* + * Link the vma into the MT. After using __mt_dup(), memory + * allocation is not necessary here, so it cannot fail. + */ + vma_iter_bulk_store(&vmi, tmp); mm->map_count++; if (!(tmp->vm_flags & VM_WIPEONFORK)) @@ -760,15 +767,28 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm, if (tmp->vm_ops && tmp->vm_ops->open) tmp->vm_ops->open(tmp); - if (retval) + if (retval) { + mpnt = vma_next(&vmi); goto loop_out; + } } /* a new mm has just been created */ retval = arch_dup_mmap(oldmm, mm); loop_out: vma_iter_free(&vmi); - if (!retval) + if (!retval) { mt_set_in_rcu(vmi.mas.tree); + } else if (mpnt) { + /* + * The entire maple tree has already been duplicated. If the + * mmap duplication fails, mark the failure point with + * XA_ZERO_ENTRY. In exit_mmap(), if this marker is encountered, + * stop releasing VMAs that have not been duplicated after this + * point. + */ + mas_set_range(&vmi.mas, mpnt->vm_start, mpnt->vm_end - 1); + mas_store(&vmi.mas, XA_ZERO_ENTRY); + } out: mmap_write_unlock(mm); flush_tlb_mm(oldmm); @@ -778,8 +798,6 @@ static __latent_entropy int dup_mmap(struct mm_struct *mm, uprobe_end_dup_mmap(); return retval; -fail_nomem_vmi_store: - unlink_anon_vmas(tmp); fail_nomem_anon_vma_fork: mpol_put(vma_policy(tmp)); fail_nomem_policy: diff --git a/mm/internal.h b/mm/internal.h index b61034bd50f5..89a5a794d68f 100644 --- a/mm/internal.h +++ b/mm/internal.h @@ -1154,17 +1154,6 @@ static inline void vma_iter_clear(struct vma_iterator *vmi) mas_store_prealloc(&vmi->mas, NULL); } -static inline int vma_iter_clear_gfp(struct vma_iterator *vmi, - unsigned long start, unsigned long end, gfp_t gfp) -{ - __mas_set_range(&vmi->mas, start, end - 1); - mas_store_gfp(&vmi->mas, NULL, gfp); - if (unlikely(mas_is_err(&vmi->mas))) - return -ENOMEM; - - return 0; -} - static inline struct vm_area_struct *vma_iter_load(struct vma_iterator *vmi) { return mas_walk(&vmi->mas); diff --git a/mm/memory.c b/mm/memory.c index 1f18ed4a5497..20cc6e3586e7 100644 --- a/mm/memory.c +++ b/mm/memory.c @@ -374,6 +374,8 @@ void free_pgtables(struct mmu_gather *tlb, struct ma_state *mas, * be 0. This will underflow and is okay. */ next = mas_find(mas, ceiling - 1); + if (unlikely(xa_is_zero(next))) + next = NULL; /* * Hide vma from rmap and truncate_pagecache before freeing @@ -395,6 +397,8 @@ void free_pgtables(struct mmu_gather *tlb, struct ma_state *mas, && !is_vm_hugetlb_page(next)) { vma = next; next = mas_find(mas, ceiling - 1); + if (unlikely(xa_is_zero(next))) + next = NULL; if (mm_wr_locked) vma_start_write(vma); unlink_anon_vmas(vma); @@ -1743,7 +1747,8 @@ void unmap_vmas(struct mmu_gather *tlb, struct ma_state *mas, unmap_single_vma(tlb, vma, start, end, &details, mm_wr_locked); hugetlb_zap_end(vma, &details); - } while ((vma = mas_find(mas, tree_end - 1)) != NULL); + vma = mas_find(mas, tree_end - 1); + } while (vma && likely(!xa_is_zero(vma))); mmu_notifier_invalidate_range_end(&range); } diff --git a/mm/mmap.c b/mm/mmap.c index 984804d77ae1..e98e9715afb2 100644 --- a/mm/mmap.c +++ b/mm/mmap.c @@ -3300,10 +3300,11 @@ void exit_mmap(struct mm_struct *mm) arch_exit_mmap(mm); vma = mas_find(&mas, ULONG_MAX); - if (!vma) { + if (!vma || unlikely(xa_is_zero(vma))) { /* Can happen if dup_mmap() received an OOM */ mmap_read_unlock(mm); - return; + mmap_write_lock(mm); + goto destroy; } lru_add_drain(); @@ -3338,11 +3339,13 @@ void exit_mmap(struct mm_struct *mm) remove_vma(vma, true); count++; cond_resched(); - } while ((vma = mas_find(&mas, ULONG_MAX)) != NULL); + vma = mas_find(&mas, ULONG_MAX); + } while (vma && likely(!xa_is_zero(vma))); BUG_ON(count != mm->map_count); trace_exit_mmap(mm); +destroy: __mt_destroy(&mm->mm_mt); mmap_write_unlock(mm); vm_unacct_memory(nr_accounted);