From patchwork Wed Feb 14 20:44:34 2024 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: David Hildenbrand X-Patchwork-Id: 13557012 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id DF927C48BC3 for ; Wed, 14 Feb 2024 20:45:26 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 73A746B00B8; Wed, 14 Feb 2024 15:45:26 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id 6E3AF6B00B9; Wed, 14 Feb 2024 15:45:26 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 55AB76B00BA; Wed, 14 Feb 2024 15:45:26 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0010.hostedemail.com [216.40.44.10]) by kanga.kvack.org (Postfix) with ESMTP id 422866B00B8 for ; Wed, 14 Feb 2024 15:45:26 -0500 (EST) Received: from smtpin13.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay02.hostedemail.com (Postfix) with ESMTP id 24134120ECE for ; Wed, 14 Feb 2024 20:45:26 +0000 (UTC) X-FDA: 81791589852.13.4DAE99D Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.129.124]) by imf12.hostedemail.com (Postfix) with ESMTP id 6D66240010 for ; Wed, 14 Feb 2024 20:45:24 +0000 (UTC) Authentication-Results: imf12.hostedemail.com; dkim=pass header.d=redhat.com header.s=mimecast20190719 header.b=C7sTOd0A; dmarc=pass (policy=none) header.from=redhat.com; spf=pass (imf12.hostedemail.com: domain of david@redhat.com designates 170.10.129.124 as permitted sender) smtp.mailfrom=david@redhat.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1707943524; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=pRS8kYaXs4+JoUa7KvLCUnXcrqNWL6ZvqI66J3j2gDY=; b=MXmLXtg2nDXF7ryCA+MJuQjj18t6cgEheBCmpdDc8Cqf7tl/F7TZY0vdPS8sqJWbff0log J+tOxOBf4Vz3UeWrq97wsEZ7HjlUAoJnQ+96UBYjSnexQcA8s6uqq3WfwLYIC6EIqxs5Tx Z67svbfbqKaqZvTM+52XhCIkHUC9fMQ= ARC-Authentication-Results: i=1; imf12.hostedemail.com; dkim=pass header.d=redhat.com header.s=mimecast20190719 header.b=C7sTOd0A; dmarc=pass (policy=none) header.from=redhat.com; spf=pass (imf12.hostedemail.com: domain of david@redhat.com designates 170.10.129.124 as permitted sender) smtp.mailfrom=david@redhat.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1707943524; a=rsa-sha256; cv=none; b=0cTg1c+2+XPn4SaRt/UvqYU0fKcJF+/UBZb/yfG3A97/kDJsljLGiuKfUc1ZiGsX1XNl6J MD6lyctb+0kAZ3+nE+kuzArXuYMTQIbvMLQsTg/vOJDASnN+T+6gMQ8G+rFhc8EWbfVPPi 8yhlP4v21Xx6OUS8QA9TRpdLVYIA3e8= DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1707943523; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=pRS8kYaXs4+JoUa7KvLCUnXcrqNWL6ZvqI66J3j2gDY=; b=C7sTOd0AAltuL91GX3JE4gHANjZjf16OKvvlqXhnf/reqcAt6AhGu7NcH96krPA360KbI4 dibbLpeBLdLiKwZrWK6Bg8T77CpKjZmvPzPA6OX66WzYEX+5vlcWTVQaEUo0JwFTLtnK+7 JtjQEVcp+ED+cuKf9H/GRZ//Bg3PqOQ= Received: from mimecast-mx02.redhat.com (mimecast-mx02.redhat.com [66.187.233.88]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.3, cipher=TLS_AES_256_GCM_SHA384) id us-mta-73-lVXRetU-PDCixQSdYcx5pA-1; Wed, 14 Feb 2024 15:45:18 -0500 X-MC-Unique: lVXRetU-PDCixQSdYcx5pA-1 Received: from smtp.corp.redhat.com (int-mx07.intmail.prod.int.rdu2.redhat.com [10.11.54.7]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (2048 bits) server-digest SHA256) (No client certificate requested) by mimecast-mx02.redhat.com (Postfix) with ESMTPS id 503D9101A52A; Wed, 14 Feb 2024 20:45:17 +0000 (UTC) Received: from t14s.fritz.box (unknown [10.39.194.174]) by smtp.corp.redhat.com (Postfix) with ESMTP id 8ECBC1C05E0F; Wed, 14 Feb 2024 20:45:13 +0000 (UTC) From: David Hildenbrand To: linux-kernel@vger.kernel.org Cc: linux-mm@kvack.org, David Hildenbrand , Andrew Morton , Matthew Wilcox , Ryan Roberts , Catalin Marinas , Yin Fengwei , Michal Hocko , Will Deacon , "Aneesh Kumar K.V" , Nick Piggin , Peter Zijlstra , Michael Ellerman , Christophe Leroy , "Naveen N. Rao" , Heiko Carstens , Vasily Gorbik , Alexander Gordeev , Christian Borntraeger , Sven Schnelle , Arnd Bergmann , linux-arch@vger.kernel.org, linuxppc-dev@lists.ozlabs.org, linux-s390@vger.kernel.org Subject: [PATCH v3 09/10] mm/mmu_gather: improve cond_resched() handling with large folios and expensive page freeing Date: Wed, 14 Feb 2024 21:44:34 +0100 Message-ID: <20240214204435.167852-10-david@redhat.com> In-Reply-To: <20240214204435.167852-1-david@redhat.com> References: <20240214204435.167852-1-david@redhat.com> MIME-Version: 1.0 X-Scanned-By: MIMEDefang 3.4.1 on 10.11.54.7 X-Rspamd-Queue-Id: 6D66240010 X-Rspam-User: X-Rspamd-Server: rspam02 X-Stat-Signature: 6qg48s5kkmhgc6p1j38qcwb3w8yf6q5h X-HE-Tag: 1707943524-436533 X-HE-Meta: U2FsdGVkX1970syn9K2StVtmRfuEE3DZBFWmzsvugYBF21oZcwIC87AIXGEm4DUnDaLCJehOQyZPV4N2Ua9NG6qSUvOqZgkXAPH/+FLUh1fdiah+LawRS80LFNrzI9Nlf9f6WeFbwXvhNQXiLC/u5to0ASWhxnxcYsLn/aYC4zYfwLCZaGD5VP1aIrShmjhe7vDjwFiaA9iYUY/p42m5kZlo1CRxeyFlyAuBzuU1iI8kyeQY1n+JP90tlpbVz8pKJE/rteVQgcxPV52WVTHWOkq2i/vIXLGQTssZc4aVWpVzhkDDOlT/Pkl3pz7HZXJjySZ3zA1SKyKvk+DM4dOsH4qNbHl5JwQT9P7s+sphhi53dCEoN0jBFsKL8BS+S9bYiX1ukNS6aW2LqWCay1AUnAurwV5A7h2ipH+vVQrwRv2XUPnRdJjCfZq0Z6UyKK+Fll4BI4Y36DIDO8srBorc7Vlnv7If7E3Bx31b9qBfpjupXk9hxxWdPHRjklNmcGe6rqTTKXPRiQwzHmCx+t4SF7Zzyow7EiA9qxHh4xGILKQpMyKBVE6dSh4tW88N9Seh5T9fLY9gEbZy63GAIozuomppZaZR+OfqZcSbDdvTK0W9dcVh3FKZkPqZNRwim0c4lD0Ncm7Ee1pkvQuqRmMAQxSSOAHPo4HHy34SPNaUoCBJYoARre0HcycJAQmwuy0qEvWyRfHRwazwJ9UmpVuBFJgT2ZFvH/Nhq7XC78zYs/Zv9qJN+OQ8xCIzzdFa53Vyj42YeCQR5oKqZI2mRuCQa5agjvLAilLGXUaRAW4mzmCBkZpSRoCaBVl+NvksB293Lb5ohlGopO+i7i88DMQxwluTLbWQVOYmKqlOQklUBm/4xoxMR8HGA2ucdFkWRTvJnApDqXhQAq6m0bFdQBZwqPmx63MRi6b0ExJhureGqFw2nXro/zAjvNiyHeZzMz66ebVnffr12JN1RzT1owV YtaBLyUn GbfQk1MqTynaXqU1h8Sa+F503zI/I63DA+S1GEZAnqCzbOZyN3sVVQIrKn13wpDk11F3epquSO5cHnztf8pC3jw5Ud9fZFXLv9A3xek3x80rizgvRaUlX5kX7KJtBCpqDTHNoLd+NgBGfXeR78z/M6ih+CtJI6f//xHqSXbkZRM6RgP8P+Gcf00RE8YaAElAiRYrOq0uMDnsx+RbEhzRpIGnVXvBknGahe54zJQgM8sFJVUpY2SiaWyuaQnwYBSjf0M1CPcq9SUe64V4pUB0AzoPB3A== X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: In tlb_batch_pages_flush(), we can end up freeing up to 512 pages or now up to 256 folio fragments that span more than one page, before we conditionally reschedule. It's a pain that we have to handle cond_resched() in tlb_batch_pages_flush() manually and cannot simply handle it in release_pages() -- release_pages() can be called from atomic context. Well, in a perfect world we wouldn't have to make our code more complicated at all. With page poisoning and init_on_free, we might now run into soft lockups when we free a lot of rather large folio fragments, because page freeing time then depends on the actual memory size we are freeing instead of on the number of folios that are involved. In the absolute (unlikely) worst case, on arm64 with 64k we will be able to free up to 256 folio fragments that each span 512 MiB: zeroing out 128 GiB does sound like it might take a while. But instead of ignoring this unlikely case, let's just handle it. So, let's teach tlb_batch_pages_flush() that there are some configurations where page freeing is horribly slow, and let's reschedule more frequently -- similarly like we did for now before we had large folio fragments in there. Avoid yet another loop over all encoded pages in the common case by handling that separately. Note that with page poisoning/zeroing, we might now end up freeing only a single folio fragment at a time that might exceed the old 512 pages limit: but if we cannot even free a single MAX_ORDER page on a system without running into soft lockups, something else is already completely bogus. Freeing a PMD-mapped THP would similarly cause trouble. In theory, we might even free 511 order-0 pages + a single MAX_ORDER page, effectively having to zero out 8703 pages on arm64 with 64k, translating to ~544 MiB of memory: however, if 512 MiB doesn't result in soft lockups, 544 MiB is unlikely to result in soft lockups, so we won't care about that for the time being. In the future, we might want to detect if handling cond_resched() is required at all, and just not do any of that with full preemption enabled. Reviewed-by: Ryan Roberts Signed-off-by: David Hildenbrand --- mm/mmu_gather.c | 58 ++++++++++++++++++++++++++++++++++++------------- 1 file changed, 43 insertions(+), 15 deletions(-) diff --git a/mm/mmu_gather.c b/mm/mmu_gather.c index d175c0f1e2c8..99b3e9408aa0 100644 --- a/mm/mmu_gather.c +++ b/mm/mmu_gather.c @@ -91,18 +91,21 @@ void tlb_flush_rmaps(struct mmu_gather *tlb, struct vm_area_struct *vma) } #endif -static void tlb_batch_pages_flush(struct mmu_gather *tlb) -{ - struct mmu_gather_batch *batch; +/* + * We might end up freeing a lot of pages. Reschedule on a regular + * basis to avoid soft lockups in configurations without full + * preemption enabled. The magic number of 512 folios seems to work. + */ +#define MAX_NR_FOLIOS_PER_FREE 512 - for (batch = &tlb->local; batch && batch->nr; batch = batch->next) { - struct encoded_page **pages = batch->encoded_pages; +static void __tlb_batch_free_encoded_pages(struct mmu_gather_batch *batch) +{ + struct encoded_page **pages = batch->encoded_pages; + unsigned int nr, nr_pages; - while (batch->nr) { - /* - * limit free batch count when PAGE_SIZE > 4K - */ - unsigned int nr = min(512U, batch->nr); + while (batch->nr) { + if (!page_poisoning_enabled_static() && !want_init_on_free()) { + nr = min(MAX_NR_FOLIOS_PER_FREE, batch->nr); /* * Make sure we cover page + nr_pages, and don't leave @@ -111,14 +114,39 @@ static void tlb_batch_pages_flush(struct mmu_gather *tlb) if (unlikely(encoded_page_flags(pages[nr - 1]) & ENCODED_PAGE_BIT_NR_PAGES_NEXT)) nr++; + } else { + /* + * With page poisoning and init_on_free, the time it + * takes to free memory grows proportionally with the + * actual memory size. Therefore, limit based on the + * actual memory size and not the number of involved + * folios. + */ + for (nr = 0, nr_pages = 0; + nr < batch->nr && nr_pages < MAX_NR_FOLIOS_PER_FREE; + nr++) { + if (unlikely(encoded_page_flags(pages[nr]) & + ENCODED_PAGE_BIT_NR_PAGES_NEXT)) + nr_pages += encoded_nr_pages(pages[++nr]); + else + nr_pages++; + } + } - free_pages_and_swap_cache(pages, nr); - pages += nr; - batch->nr -= nr; + free_pages_and_swap_cache(pages, nr); + pages += nr; + batch->nr -= nr; - cond_resched(); - } + cond_resched(); } +} + +static void tlb_batch_pages_flush(struct mmu_gather *tlb) +{ + struct mmu_gather_batch *batch; + + for (batch = &tlb->local; batch && batch->nr; batch = batch->next) + __tlb_batch_free_encoded_pages(batch); tlb->active = &tlb->local; }