From patchwork Fri Jun 14 19:06:36 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: "Jason A. Donenfeld" X-Patchwork-Id: 13699082 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 00898C27C6E for ; Fri, 14 Jun 2024 19:07:26 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 854D06B00FD; Fri, 14 Jun 2024 15:07:26 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 804736B010C; Fri, 14 Jun 2024 15:07:26 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 6CC996B026A; Fri, 14 Jun 2024 15:07:26 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0014.hostedemail.com [216.40.44.14]) by kanga.kvack.org (Postfix) with ESMTP id 4E9666B00FD for ; Fri, 14 Jun 2024 15:07:26 -0400 (EDT) Received: from smtpin24.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay09.hostedemail.com (Postfix) with ESMTP id 1A1D8808C0 for ; Fri, 14 Jun 2024 19:07:25 +0000 (UTC) X-FDA: 82230427650.24.0110641 Received: from sin.source.kernel.org (sin.source.kernel.org [145.40.73.55]) by imf25.hostedemail.com (Postfix) with ESMTP id 96EF8A0018 for ; Fri, 14 Jun 2024 19:07:21 +0000 (UTC) Authentication-Results: imf25.hostedemail.com; dkim=pass header.d=zx2c4.com header.s=20210105 header.b=GTza31jG; dmarc=pass (policy=quarantine) header.from=zx2c4.com; spf=pass (imf25.hostedemail.com: domain of "SRS0=OOQL=NQ=zx2c4.com=Jason@kernel.org" designates 145.40.73.55 as permitted sender) smtp.mailfrom="SRS0=OOQL=NQ=zx2c4.com=Jason@kernel.org" ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1718392040; a=rsa-sha256; cv=none; b=4S4IIBwkHya3HUrNWYbxhTAJ7ZgMdb0XBMw2dwdBfC4601xhpaF8P6+Hn9vvWVytNgC7eG l8ySwN1DsyHXml8PQey3t4oPnOe5QKXaEKcZULderhv7s44mxEGFsVsy5OKgo1lU2tpxFU L/3dzpo86C4/W7GPfcrm5npufIrvWlM= ARC-Authentication-Results: i=1; imf25.hostedemail.com; dkim=pass header.d=zx2c4.com header.s=20210105 header.b=GTza31jG; dmarc=pass (policy=quarantine) header.from=zx2c4.com; spf=pass (imf25.hostedemail.com: domain of "SRS0=OOQL=NQ=zx2c4.com=Jason@kernel.org" designates 145.40.73.55 as permitted sender) smtp.mailfrom="SRS0=OOQL=NQ=zx2c4.com=Jason@kernel.org" ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1718392040; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=HIMpXESIZrvLDlNkbefExpVi2+GH2WoPeJxzWPkZD3s=; b=FFLSoU9IAhR2m3fHNXnGELNvQFvy5TArkMY1wpQZj21NIiPE7ZCMcTOjcwPrKkxlezwmHA yXYrnDplrjOfvT5N/5+VXzK/hsuH4AAiYVYQ96r2aIFAPsFX56SSdqrfawkNc30M230WBJ B50+Znpav1HdZ9LME94zXtab/a+qvSI= Received: from smtp.kernel.org (transwarp.subspace.kernel.org [100.75.92.58]) by sin.source.kernel.org (Postfix) with ESMTP id 40C37CE1CAA; Fri, 14 Jun 2024 19:07:18 +0000 (UTC) Received: by smtp.kernel.org (Postfix) with ESMTPSA id 4609FC4AF1C; Fri, 14 Jun 2024 19:07:16 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=zx2c4.com; s=20210105; t=1718392035; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=HIMpXESIZrvLDlNkbefExpVi2+GH2WoPeJxzWPkZD3s=; b=GTza31jGClIfqM18ok0B3ynbc5mdcrGNhdEmG0sGDLpSYwGMIOtOn3D38xm+S3dXSpvq5f LS03m1pltJxTiKMqrKANcsk6x4FszFBJ4y+cWiY7x8Xyb+bBcoq7U3sSvajk0JY1kD3p5b a1KIAh4cVJ/u24Wx72ijoNLCne3bwBo= Received: by mail.zx2c4.com (ZX2C4 Mail Server) with ESMTPSA id b3de6dd9 (TLSv1.3:TLS_AES_256_GCM_SHA384:256:NO); Fri, 14 Jun 2024 19:07:14 +0000 (UTC) From: "Jason A. Donenfeld" To: linux-kernel@vger.kernel.org, patches@lists.linux.dev, tglx@linutronix.de Cc: "Jason A. Donenfeld" , linux-crypto@vger.kernel.org, linux-api@vger.kernel.org, x86@kernel.org, Greg Kroah-Hartman , Adhemerval Zanella Netto , Carlos O'Donell , Florian Weimer , Arnd Bergmann , Jann Horn , Christian Brauner , David Hildenbrand , linux-mm@kvack.org Subject: [PATCH v17 1/5] mm: add VM_DROPPABLE for designating always lazily freeable mappings Date: Fri, 14 Jun 2024 21:06:36 +0200 Message-ID: <20240614190646.2081057-2-Jason@zx2c4.com> In-Reply-To: <20240614190646.2081057-1-Jason@zx2c4.com> References: <20240614190646.2081057-1-Jason@zx2c4.com> MIME-Version: 1.0 X-Rspamd-Queue-Id: 96EF8A0018 X-Rspam-User: X-Rspamd-Server: rspam05 X-Stat-Signature: rzxw3u7aua8z34kmcp6meiffzaw4dop5 X-HE-Tag: 1718392041-530556 X-HE-Meta: U2FsdGVkX1+lQYOUOOMBa+1kD0E08i8URjM/uzryzQFM2CLsIqpOJafASS+mIYUMKzNBi3OMjMnYaPPZUxPIDyZF/8OEKQkN6lPLHh/B0Jtar9j+n7fpHafsRVb6PodaTvAk4deWzykDSOxCDVwfYswVJHla+Sg14X3evcUo+GSkkXuHHnI6QD6Nr5JCxkvMZbIREjsRckF1/cVZxvoyG0FjliVO+eTRjv+td7Ix5bi7kj5f8/43aDZZp2Luh42zzUWcK6VFrg5u8LiGA2MJR/n3ksMOdMCmuFMUA5MjplsN+H9S/sHdc2/tt50g916qLZ3z3Vgrybh0pawfjRu6Sbl1H1VCCsZUrOOSj391PCdGVHwVq+OG08arMt6KWNsIfGcGffc1yLUGRHgQk0Jt8V8VDr8kbiDNT4eOYl7f7VxqlsfOwpF/IJ1Mh9OldCoOMQVaQNTSrNU0TRfTzncx6kEwg2e55PcwNmyIuVwpSnYJ/2Yz4Fnyb+7yGqZG/xk/ssTmH4i4CrfA0Qg9HfUIxsMEcZeLVwbtPnhDQOvo+3WeTfD1LR1erjDgJsfzGt5qxYvXcXkl3wr+rBcwplQzHSeU0GBg7DZG0ItUhdr84bFAw/YHp/qGs3LZGbpyQ3hLi+YTr1tky4qoab6VUEsEteb7z8k5VX0dPB4cwaeQXvfKIMxrORDJGxl34lSarNrwwPpqf8gUA8TxGRd6RkdPf3kVMHK8ucGlqi3nEc38C55W8/YmVH99IEOLnumPG/WcpMvpk9yX5Oa2LDCqBI2H2vLWpT1GnSqhgTPKKzjqgVKJKe/qa1mmcnq7WdX+L8pMfLKrodLlZ8BNSq7aEs+e8m802fKZbjoiiAEu367+QBnC+ks7AflLJMChvNYQxnREWpRv8OjnVteuuxoH78fuLfUUsyHkke0gJ7yatjFqdf4sOtx8Az1TJSQoz2JZdHmx4ovHlCxjBcJV9ndUJPp lQqZU7tC 4sC5oNj4BT2zajKJjcbbiDV/wmG6jAB7k+vpri1D27AFlx6QyaG98B4IpzM+n99E0ZBXzOgIw6DjVS53bs3aoHdlqrmNZFSHnhqmvlcUj9UhbVA7qbk4MjNMXxPr9JwasOd9LYhSWRHiPHVBPF9Bbrry6T8+lkyCaYm0/10LP+mnkeWH+yM5TJMvZsF1AFuYtd+lQSKFrIVMaRlhpamPp0bFHWCgDnFGqtSOCQXNbz63X+vgoVDfsBQBE0fiS26aRy08qM707ywhkRHSmAgSlnkpPGKLfV9bxIVgDQyjzG2cCe6OV5pU5vxUgrHzOMjL9vNusze/1GHb8WRNHsS9Lh32pfrW9mGGaEcQ1Wx3xTgy7DF+X1YEZIO24dH5aVKiNXOSldpfOGzewI+U/2iBhsxaHBltODIBRP/g/mwiC0SUk799sXzEjDWBSaVNeO3rzjqNbCpG0g9eXfJlpBEvtHJws8wY6fJRYRyQnUe/RUN8lMBuEC5CmQ/YGMeeJbHWvP/aL7sjCZMqJRf2+0qYWZOW61YFopPzA/J70K6d8+Ctm1NNEezY0hKN0tg== X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: The vDSO getrandom() implementation works with a buffer allocated with a new system call that has certain requirements: - It shouldn't be written to core dumps. * Easy: VM_DONTDUMP. - It should be zeroed on fork. * Easy: VM_WIPEONFORK. - It shouldn't be written to swap. * Uh-oh: mlock is rlimited. * Uh-oh: mlock isn't inherited by forks. It turns out that the vDSO getrandom() function has three really nice characteristics that we can exploit to solve this problem: 1) Due to being wiped during fork(), the vDSO code is already robust to having the contents of the pages it reads zeroed out midway through the function's execution. 2) In the absolute worst case of whatever contingency we're coding for, we have the option to fallback to the getrandom() syscall, and everything is fine. 3) The buffers the function uses are only ever useful for a maximum of 60 seconds -- a sort of cache, rather than a long term allocation. These characteristics mean that we can introduce VM_DROPPABLE, which has the following semantics: a) It never is written out to swap. b) Under memory pressure, mm can just drop the pages (so that they're zero when read back again). c) It is inherited by fork. d) It doesn't count against the mlock budget, since nothing is locked. This is fairly simple to implement, with the one snag that we have to use 64-bit VM_* flags, but this shouldn't be a problem, since the only consumers will probably be 64-bit anyway. This way, allocations used by vDSO getrandom() can use: VM_DROPPABLE | VM_DONTDUMP | VM_WIPEONFORK | VM_NORESERVE And there will be no problem with using memory when not in use, not wiping on fork(), coredumps, or writing out to swap. Cc: linux-mm@kvack.org Signed-off-by: Jason A. Donenfeld --- fs/proc/task_mmu.c | 3 +++ include/linux/mm.h | 8 ++++++++ include/trace/events/mmflags.h | 7 +++++++ mm/Kconfig | 3 +++ mm/mprotect.c | 2 +- mm/rmap.c | 8 +++++--- 6 files changed, 27 insertions(+), 4 deletions(-) diff --git a/fs/proc/task_mmu.c b/fs/proc/task_mmu.c index f8d35f993fe5..1883d6462ca8 100644 --- a/fs/proc/task_mmu.c +++ b/fs/proc/task_mmu.c @@ -706,6 +706,9 @@ static void show_smap_vma_flags(struct seq_file *m, struct vm_area_struct *vma) #endif /* CONFIG_HAVE_ARCH_USERFAULTFD_MINOR */ #ifdef CONFIG_X86_USER_SHADOW_STACK [ilog2(VM_SHADOW_STACK)] = "ss", +#endif +#ifdef CONFIG_NEED_VM_DROPPABLE + [ilog2(VM_DROPPABLE)] = "dp", #endif }; size_t i; diff --git a/include/linux/mm.h b/include/linux/mm.h index 9849dfda44d4..5978cb4cc21c 100644 --- a/include/linux/mm.h +++ b/include/linux/mm.h @@ -321,12 +321,14 @@ extern unsigned int kobjsize(const void *objp); #define VM_HIGH_ARCH_BIT_3 35 /* bit only usable on 64-bit architectures */ #define VM_HIGH_ARCH_BIT_4 36 /* bit only usable on 64-bit architectures */ #define VM_HIGH_ARCH_BIT_5 37 /* bit only usable on 64-bit architectures */ +#define VM_HIGH_ARCH_BIT_6 38 /* bit only usable on 64-bit architectures */ #define VM_HIGH_ARCH_0 BIT(VM_HIGH_ARCH_BIT_0) #define VM_HIGH_ARCH_1 BIT(VM_HIGH_ARCH_BIT_1) #define VM_HIGH_ARCH_2 BIT(VM_HIGH_ARCH_BIT_2) #define VM_HIGH_ARCH_3 BIT(VM_HIGH_ARCH_BIT_3) #define VM_HIGH_ARCH_4 BIT(VM_HIGH_ARCH_BIT_4) #define VM_HIGH_ARCH_5 BIT(VM_HIGH_ARCH_BIT_5) +#define VM_HIGH_ARCH_6 BIT(VM_HIGH_ARCH_BIT_6) #endif /* CONFIG_ARCH_USES_HIGH_VMA_FLAGS */ #ifdef CONFIG_ARCH_HAS_PKEYS @@ -357,6 +359,12 @@ extern unsigned int kobjsize(const void *objp); # define VM_SHADOW_STACK VM_NONE #endif +#ifdef CONFIG_NEED_VM_DROPPABLE +# define VM_DROPPABLE VM_HIGH_ARCH_6 +#else +# define VM_DROPPABLE VM_NONE +#endif + #if defined(CONFIG_X86) # define VM_PAT VM_ARCH_1 /* PAT reserves whole VMA at once (x86) */ #elif defined(CONFIG_PPC) diff --git a/include/trace/events/mmflags.h b/include/trace/events/mmflags.h index e46d6e82765e..fab7848df50a 100644 --- a/include/trace/events/mmflags.h +++ b/include/trace/events/mmflags.h @@ -165,6 +165,12 @@ IF_HAVE_PG_ARCH_X(arch_3) # define IF_HAVE_UFFD_MINOR(flag, name) #endif +#ifdef CONFIG_NEED_VM_DROPPABLE +# define IF_HAVE_VM_DROPPABLE(flag, name) {flag, name}, +#else +# define IF_HAVE_VM_DROPPABLE(flag, name) +#endif + #define __def_vmaflag_names \ {VM_READ, "read" }, \ {VM_WRITE, "write" }, \ @@ -197,6 +203,7 @@ IF_HAVE_VM_SOFTDIRTY(VM_SOFTDIRTY, "softdirty" ) \ {VM_MIXEDMAP, "mixedmap" }, \ {VM_HUGEPAGE, "hugepage" }, \ {VM_NOHUGEPAGE, "nohugepage" }, \ +IF_HAVE_VM_DROPPABLE(VM_DROPPABLE, "droppable" ) \ {VM_MERGEABLE, "mergeable" } \ #define show_vma_flags(flags) \ diff --git a/mm/Kconfig b/mm/Kconfig index b4cb45255a54..6cd65ea4b3ad 100644 --- a/mm/Kconfig +++ b/mm/Kconfig @@ -1056,6 +1056,9 @@ config ARCH_USES_HIGH_VMA_FLAGS bool config ARCH_HAS_PKEYS bool +config NEED_VM_DROPPABLE + select ARCH_USES_HIGH_VMA_FLAGS + bool config ARCH_USES_PG_ARCH_X bool diff --git a/mm/mprotect.c b/mm/mprotect.c index 8c6cd8825273..57b8dad9adcc 100644 --- a/mm/mprotect.c +++ b/mm/mprotect.c @@ -623,7 +623,7 @@ mprotect_fixup(struct vma_iterator *vmi, struct mmu_gather *tlb, may_expand_vm(mm, oldflags, nrpages)) return -ENOMEM; if (!(oldflags & (VM_ACCOUNT|VM_WRITE|VM_HUGETLB| - VM_SHARED|VM_NORESERVE))) { + VM_SHARED|VM_NORESERVE|VM_DROPPABLE))) { charged = nrpages; if (security_vm_enough_memory_mm(mm, charged)) return -ENOMEM; diff --git a/mm/rmap.c b/mm/rmap.c index e8fc5ecb59b2..d873a3f06506 100644 --- a/mm/rmap.c +++ b/mm/rmap.c @@ -1397,7 +1397,8 @@ void folio_add_new_anon_rmap(struct folio *folio, struct vm_area_struct *vma, VM_WARN_ON_FOLIO(folio_test_hugetlb(folio), folio); VM_BUG_ON_VMA(address < vma->vm_start || address + (nr << PAGE_SHIFT) > vma->vm_end, vma); - __folio_set_swapbacked(folio); + if (!(vma->vm_flags & VM_DROPPABLE)) + __folio_set_swapbacked(folio); __folio_set_anon(folio, vma, address, true); if (likely(!folio_test_large(folio))) { @@ -1841,7 +1842,7 @@ static bool try_to_unmap_one(struct folio *folio, struct vm_area_struct *vma, * plus the rmap(s) (dropped by discard:). */ if (ref_count == 1 + map_count && - !folio_test_dirty(folio)) { + (!folio_test_dirty(folio) || (vma->vm_flags & VM_DROPPABLE))) { dec_mm_counter(mm, MM_ANONPAGES); goto discard; } @@ -1851,7 +1852,8 @@ static bool try_to_unmap_one(struct folio *folio, struct vm_area_struct *vma, * discarded. Remap the page to page table. */ set_pte_at(mm, address, pvmw.pte, pteval); - folio_set_swapbacked(folio); + if (!(vma->vm_flags & VM_DROPPABLE)) + folio_set_swapbacked(folio); ret = false; page_vma_mapped_walk_done(&pvmw); break;