From patchwork Wed Feb 5 01:39:52 2025 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Rik van Riel X-Patchwork-Id: 13960415 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 073C7C02196 for ; Wed, 5 Feb 2025 01:42:27 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 31C3A280033; Tue, 4 Feb 2025 20:42:05 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id 2CCBA280029; Tue, 4 Feb 2025 20:42:05 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 16C6E280033; Tue, 4 Feb 2025 20:42:05 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0011.hostedemail.com [216.40.44.11]) by kanga.kvack.org (Postfix) with ESMTP id EB8C7280029 for ; Tue, 4 Feb 2025 20:42:04 -0500 (EST) Received: from smtpin05.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay08.hostedemail.com (Postfix) with ESMTP id AB3E9140431 for ; Wed, 5 Feb 2025 01:42:04 +0000 (UTC) X-FDA: 83084190168.05.511F46E Received: from shelob.surriel.com (shelob.surriel.com [96.67.55.147]) by imf20.hostedemail.com (Postfix) with ESMTP id 260D51C000A for ; Wed, 5 Feb 2025 01:42:02 +0000 (UTC) Authentication-Results: imf20.hostedemail.com; dkim=none; spf=pass (imf20.hostedemail.com: domain of riel@shelob.surriel.com designates 96.67.55.147 as permitted sender) smtp.mailfrom=riel@shelob.surriel.com; dmarc=none ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1738719723; h=from:from:sender:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=iAmQmZCZMVLJeE+FEOFdmABFJ0iOT1gtVlpDsFG+APg=; b=1ki6TA5LBir0FHSPtxXQF/vs6GDG4I7OzZqGJRx2tL1of6lyR8khwZdKCRY0vGbrf/ZFhq ZX+RqoR2Ka++hG3FUv0k/77+c8mW+IRKMwhB3Y5NGTLH5RaUVtC6w8/1IFrFffZRy868As NDKeuNdEOdjbISl2ogwgEem+8P5vj+U= ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1738719723; a=rsa-sha256; cv=none; b=FcNJSDxQyvP4xxumX0XUxt1gEoaKLYX6+0P6H/5J0UujfzZK5lbvDmYLCe74NorH7cT4gh /USyvUGwXkDZ5qBXTfTHa9P99emq0bciXn6tV3Mcr0HDgiX1IuX+d5wHADR0y/ENY34+1W j/vIvH39kLMq94hQ6u+SuoBuwdnMUE0= ARC-Authentication-Results: i=1; imf20.hostedemail.com; dkim=none; spf=pass (imf20.hostedemail.com: domain of riel@shelob.surriel.com designates 96.67.55.147 as permitted sender) smtp.mailfrom=riel@shelob.surriel.com; dmarc=none Received: from fangorn.home.surriel.com ([10.0.13.7]) by shelob.surriel.com with esmtpsa (TLS1.2) tls TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384 (Exim 4.97.1) (envelope-from ) id 1tfUP4-000000004Cs-2eJa; Tue, 04 Feb 2025 20:40:34 -0500 From: Rik van Riel To: x86@kernel.org Cc: linux-kernel@vger.kernel.org, bp@alien8.de, peterz@infradead.org, dave.hansen@linux.intel.com, zhengqi.arch@bytedance.com, nadav.amit@gmail.com, thomas.lendacky@amd.com, kernel-team@meta.com, linux-mm@kvack.org, akpm@linux-foundation.org, jannh@google.com, mhklinux@outlook.com, andrew.cooper3@citrix.com, Rik van Riel , Dave Hansen Subject: [PATCH v8 03/12] x86/mm: consolidate full flush threshold decision Date: Tue, 4 Feb 2025 20:39:52 -0500 Message-ID: <20250205014033.3626204-4-riel@surriel.com> X-Mailer: git-send-email 2.47.1 In-Reply-To: <20250205014033.3626204-1-riel@surriel.com> References: <20250205014033.3626204-1-riel@surriel.com> MIME-Version: 1.0 X-Stat-Signature: rwmabkwsghkbetkfeqit3qp3jadkqrzc X-Rspamd-Queue-Id: 260D51C000A X-Rspam-User: X-Rspamd-Server: rspam06 X-HE-Tag: 1738719722-735571 X-HE-Meta: U2FsdGVkX18DrqDoZmShxNfypi88P2W2rKXF3AgAMP6BCjGQK3NQs79xAVrbXBgjK5gUczqk6TT51WIeXiynsmwIrm/SebtTdEtavbH63cqmRiOUwp7blOnO5AKjKGZpK+cHI1RECwEa3BmmiMSEKmlYIiS5gszVWp6nns9ZPDudMB9SWL6OZpuYIfQUuH54wo6jmkBb4xNKwsjTi9vdoP5/AByPp6kcrN27Cd/QrnciwhJghT3R6DjkiUxIVE/9ZuPny9+4XEfdLTXezZhvz+pG43NhtaRTevOUOmaPIpY5E/cYmCppr0SwYKQzzlis3mWT+7RsWdovBba40Vz8ottish0Z/WdkRPnIIWnj0a8h8clyxFDLx+ZUsWt9GoBY72vK5PdOvLtmiyVwGCg6L2zHiQC6TU9cV+kVbKMH0N5PZm38C6+tPd4O+F8g4aagJxnmAlVuxz7jVhXGIYWF7WF2lBUNLF1XK4JpAfIPROrn//VbacDg8PPbCayhejS1EvbW/OYDsCTZTNXlEbnbL2jtX0eOG5/5RoYRi3Or9VH967kWOKcRYYE/cWuytw+w3IEvVyH/81gr5CsYdm2D2znlrPq3K6xisj92iafGgn4rJE2Gg1viGbE+TwljbULH4LUgCbztOwCw7XknU66TStVa9lMjp18Q5Yts0Uv5VdNAUGlUJNXu8vN72+8mKFTwtxD85fV4y2wWrdiuGwLCpgP2B3vsqZc8JAQHDzOgZDcMUu39JQe1BnVfzECeSX33QtRENM9KcQw2+TEmBOxOiHiH+NnnQ2aULav7TAIsmK9WpvYrgxr1jlyfmDEmtY8s3KnP6P5w/DUsfsBHG273zKtbNmcNFFU5sGM82AfS5tcJsIlrJpHqxUIVrlhYElScLzOtBpFWtk4D6L9iL54osK/MwMjrULx/NXNR+ltwkQxGYw1gw659Bwk3GJZ1IRg/GhA6Vk2FZZ1sVtfJtyu mOC0N6i5 K3RaK+uu79SCdFBlilEcjBsJzUNB1I5GBBTp13KCaFOuvj6h0w1HA61cmReKdKyHPymXyRvD/7Z9i4Z2LBs+v4BS+Z/WLebULj3kJ2Rldn1YrGUmkQUGsVsY0NfJtzFRvvyuiZ5HVT1plPQIfeFyE773+klnXPufrWtWCrb7DvqgPLf0WmRnuoVySMi107aBXRb5vCXmTTY5i+3XeLCzFslVSiGZ1Gqz2ZEkRGbP/3YBgmRjL+Ja8P1FFr+zwptEthwV7ABb54WK5xv89tBbd3qr1CepSY4u06eok7Zily7WfsC3sOCGi095iThcU/frBqo3es4G3nsPfwBQIWhKVBV0iyiyMwqcF/nD3wanCvbwBJ8E7RJmjzJ2r7z/LXpnupX9ke/GLY0ugBqEH/zhSA4x/1Qth6I8wZQ/9cS97z+H2dYQdoJzyqinQEHoWyuBqzZemZ6Ze5bhOt8TyOsdzZolKlpa5kINutAeXugHaMm9bKyPAHelIm2YNlA== X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: Reduce code duplication by consolidating the decision point for whether to do individual invalidations or a full flush inside get_flush_tlb_info. Signed-off-by: Rik van Riel Suggested-by: Dave Hansen --- arch/x86/mm/tlb.c | 52 ++++++++++++++++++++++++----------------------- 1 file changed, 27 insertions(+), 25 deletions(-) diff --git a/arch/x86/mm/tlb.c b/arch/x86/mm/tlb.c index 6cf881a942bb..02e1f5c5bca3 100644 --- a/arch/x86/mm/tlb.c +++ b/arch/x86/mm/tlb.c @@ -1000,8 +1000,13 @@ static struct flush_tlb_info *get_flush_tlb_info(struct mm_struct *mm, BUG_ON(this_cpu_inc_return(flush_tlb_info_idx) != 1); #endif - info->start = start; - info->end = end; + /* + * Round the start and end addresses to the page size specified + * by the stride shift. This ensures partial pages at the end of + * a range get fully invalidated. + */ + info->start = round_down(start, 1 << stride_shift); + info->end = round_up(end, 1 << stride_shift); info->mm = mm; info->stride_shift = stride_shift; info->freed_tables = freed_tables; @@ -1009,6 +1014,15 @@ static struct flush_tlb_info *get_flush_tlb_info(struct mm_struct *mm, info->initiating_cpu = smp_processor_id(); info->trim_cpumask = 0; + /* + * If the number of flushes is so large that a full flush + * would be faster, do a full flush. + */ + if ((end - start) >> stride_shift > tlb_single_page_flush_ceiling) { + info->start = 0; + info->end = TLB_FLUSH_ALL; + } + return info; } @@ -1026,17 +1040,8 @@ void flush_tlb_mm_range(struct mm_struct *mm, unsigned long start, bool freed_tables) { struct flush_tlb_info *info; + int cpu = get_cpu(); u64 new_tlb_gen; - int cpu; - - cpu = get_cpu(); - - /* Should we flush just the requested range? */ - if ((end == TLB_FLUSH_ALL) || - ((end - start) >> stride_shift) > tlb_single_page_flush_ceiling) { - start = 0; - end = TLB_FLUSH_ALL; - } /* This is also a barrier that synchronizes with switch_mm(). */ new_tlb_gen = inc_mm_tlb_gen(mm); @@ -1089,22 +1094,19 @@ static void do_kernel_range_flush(void *info) void flush_tlb_kernel_range(unsigned long start, unsigned long end) { - /* Balance as user space task's flush, a bit conservative */ - if (end == TLB_FLUSH_ALL || - (end - start) > tlb_single_page_flush_ceiling << PAGE_SHIFT) { - on_each_cpu(do_flush_tlb_all, NULL, 1); - } else { - struct flush_tlb_info *info; + struct flush_tlb_info *info; - preempt_disable(); - info = get_flush_tlb_info(NULL, start, end, 0, false, - TLB_GENERATION_INVALID); + guard(preempt)(); + + info = get_flush_tlb_info(NULL, start, end, PAGE_SHIFT, false, + TLB_GENERATION_INVALID); + if (info->end == TLB_FLUSH_ALL) + on_each_cpu(do_flush_tlb_all, NULL, 1); + else on_each_cpu(do_kernel_range_flush, info, 1); - put_flush_tlb_info(); - preempt_enable(); - } + put_flush_tlb_info(); } /* @@ -1276,7 +1278,7 @@ void arch_tlbbatch_flush(struct arch_tlbflush_unmap_batch *batch) int cpu = get_cpu(); - info = get_flush_tlb_info(NULL, 0, TLB_FLUSH_ALL, 0, false, + info = get_flush_tlb_info(NULL, 0, TLB_FLUSH_ALL, PAGE_SHIFT, false, TLB_GENERATION_INVALID); /* * flush_tlb_multi() is not optimized for the common case in which only