From patchwork Wed Feb 19 06:06:52 2025 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: mawupeng X-Patchwork-Id: 13981560 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id E647BC021AA for ; Wed, 19 Feb 2025 06:16:34 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 54B88280129; Wed, 19 Feb 2025 01:16:34 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id 4D2ED280122; Wed, 19 Feb 2025 01:16:34 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 2B16A280129; Wed, 19 Feb 2025 01:16:34 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0016.hostedemail.com [216.40.44.16]) by kanga.kvack.org (Postfix) with ESMTP id E9A9B280122 for ; Wed, 19 Feb 2025 01:16:33 -0500 (EST) Received: from smtpin26.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay05.hostedemail.com (Postfix) with ESMTP id 46AEF5071B for ; Wed, 19 Feb 2025 06:16:33 +0000 (UTC) X-FDA: 83135685066.26.7B0BA17 Received: from szxga04-in.huawei.com (szxga04-in.huawei.com [45.249.212.190]) by imf16.hostedemail.com (Postfix) with ESMTP id D26CB18000B for ; Wed, 19 Feb 2025 06:16:30 +0000 (UTC) Authentication-Results: imf16.hostedemail.com; dkim=none; spf=pass (imf16.hostedemail.com: domain of mawupeng1@huawei.com designates 45.249.212.190 as permitted sender) smtp.mailfrom=mawupeng1@huawei.com; dmarc=pass (policy=quarantine) header.from=huawei.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1739945791; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=uaDgv/un/0qcYLaOCZqzL3T1ibJRNOsnrnAmdzKAXm8=; b=eQJA76fDJmcD7dhGAanArIA1InRrC0DobJZNoQHQVkBGWLNd2vVkTLKGYjiHwW/qcoojug nna2MielZ/0keBeo3kNNrpslNlwXz1JxzVoY4H4ASeKf6hGTO0uJp1lsIeCfNs8fGyI8yQ AeV25fHcnJvjPLIs7Gl2horUPRvBMo8= ARC-Authentication-Results: i=1; imf16.hostedemail.com; dkim=none; spf=pass (imf16.hostedemail.com: domain of mawupeng1@huawei.com designates 45.249.212.190 as permitted sender) smtp.mailfrom=mawupeng1@huawei.com; dmarc=pass (policy=quarantine) header.from=huawei.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1739945791; a=rsa-sha256; cv=none; b=316/oyk55yEeXLjHSO6Gtobk2qSNHRs2ejGqAti8pKPFq8a3X6dVLqe30R8XedMMB1m0ay W5BxtAsmkCJKsvxoSnMZoF5WnwWdGmlqXR5SCwBBXcyH66afB52DNkRxdCKHhMTr2Hv/8Q KH5tb0kJocSlSWh0kMObWAVkZW9Rozg= Received: from mail.maildlp.com (unknown [172.19.88.163]) by szxga04-in.huawei.com (SkyGuard) with ESMTP id 4YyQzM71BBz2JYb4; Wed, 19 Feb 2025 14:12:31 +0800 (CST) Received: from kwepemg100017.china.huawei.com (unknown [7.202.181.58]) by mail.maildlp.com (Postfix) with ESMTPS id 865B8180214; Wed, 19 Feb 2025 14:16:26 +0800 (CST) Received: from huawei.com (10.175.124.71) by kwepemg100017.china.huawei.com (7.202.181.58) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384) id 15.2.1544.11; Wed, 19 Feb 2025 14:16:25 +0800 From: Wupeng Ma To: , , , , , CC: , , Subject: [PATCH v3] mm: memory-failure: update ttu flag inside unmap_poisoned_folio Date: Wed, 19 Feb 2025 14:06:52 +0800 Message-ID: <20250219060653.3849083-1-mawupeng1@huawei.com> X-Mailer: git-send-email 2.43.0 In-Reply-To: <55e4ad74-752b-65c6-5ceb-b3a7fd7959a1@huawei.com> References: <55e4ad74-752b-65c6-5ceb-b3a7fd7959a1@huawei.com> MIME-Version: 1.0 X-Originating-IP: [10.175.124.71] X-ClientProxiedBy: dggems705-chm.china.huawei.com (10.3.19.182) To kwepemg100017.china.huawei.com (7.202.181.58) X-Rspamd-Server: rspam04 X-Rspamd-Queue-Id: D26CB18000B X-Stat-Signature: 4akhxg84bz91kmuacu34b83bt6rzojj8 X-Rspam-User: X-HE-Tag: 1739945790-420268 X-HE-Meta: U2FsdGVkX190fpI19GpG03UzfzCVDTz7r2zJnYDaLnWUKVEGYhjNsNiowLy/13ozL2nsjzRjLN2AgYeLy/X4aemevU4Y5/K30l0FI1jchCjkrXcyR5yWUo8woAA6eujxiMhu6GoWYljIMdHPPfHFTWtRV7N/6h6yYziW7RsQXLTqgY/YLbXyJ+eAF8WhYJRzq8txGIGyf8sqnuxcYK4bXyydqLTC/Ry1/tCjGQ39+BaHufAcEo3hjCvJ7OzIqGa2LF6Pz6WfzXv66gH/Oi+a2hMV7njwm03bl601+om+mAS7MwqA7g3E1LcYZOINttRtlWYZBpEJNE2rv5Pbzf6AzJutrSyV34Sgwccac4qMz0+3ud2fAqK609DueLiLvfns/u5lzrMfDj+NDR+lrI9BlVQ/hPJmsVCC+q53hWhFUgLxdWRm5NSCQy+qpz2w8804V43C4Ldvyb9dv/l1Yng/k4+7qUBRN1I/qYBeYehKwNhodYqI4djff9GSAFNOoF0je8oBB3509IfU7IVT4iQzP+eVc1SrOot2K/NoLMmffAVbTDtukWri9B+BUEQrHaVn6nxcxyuD9cCiPB2FBNVCSi20mcEmHcQLyWfztUVlYY8S24jeQVE5Dr174K1zIqq13BLRTGzx9NkX+X7FoNr1Duw3BRqpYOw4Cf3JLTd1rPm72n4BFQxSNYQ/iAG3HStEYt0uIDIXUU1yLHdH8WWdwHhDqDWNpB9Imzb9VbTIorWcVizV6lZNW/JnRkqeKOmzEfiRAZlsQA8bs/ChqI8lamNdROdSved06VbyuPTPcFrbY4zqO34bRVenl8EuovvY2kyjoZfG/WyiX8bvDw4xojskk+6GbWJiFrADYwepsP9ln9pq7shXXe4X21vR73wRXqMR8GH1bQ6GHkqMHEZOxwwQTLJhrStiy+E52U/82CL6u99m2OD/gkh3wzSDsLAmDypFE3pL5isOohVFVJ4 oeDni4Rr +sU2UQaEeSk2DhxS3516M150OWh9Up1iTGQg/abkI7l5HN7lUtG4Ha5B/aX42ITS0fBP/ICtlSmSwiLF38dKPZaIc/h6sSap+bovfjcRXNAcWFyCm3nnXfiADW84bAozNMBA1B3XALU9J+igKSa2r62Q4Kyx37z8zbh0F3BX5zPbY5EalYWma+VyyJF1a7V7qZSq6aHSzYppIRzaaempKHYGMFAgOd4hyDjK0Z9IocYf5YSxE4MSLqeRFHtLxMp0QNlWyubLukbovh7ROaT5H2k4PQZkxIasvm5EO X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: From: Ma Wupeng Commit 6da6b1d4a7df ("mm/hwpoison: convert TTU_IGNORE_HWPOISON to TTU_HWPOISON") introduce TTU_HWPOISON to replace TTU_IGNORE_HWPOISON in order to stop send SIGBUS signal when accessing an error page after a memory error on a clean folio. However during page migration, anon folio must be set with TTU_HWPOISON during unmap_*(). For pagecache we need some policy just like the one in hwpoison_user_mappings to set this flag. So move this policy from hwpoison_user_mappings to unmap_poisoned_folio to handle this warning properly. Warning will be produced during unamp poison folio with the following log: ------------[ cut here ]------------ WARNING: CPU: 1 PID: 365 at mm/rmap.c:1847 try_to_unmap_one+0x8fc/0xd3c Modules linked in: CPU: 1 UID: 0 PID: 365 Comm: bash Tainted: G W 6.13.0-rc1-00018-gacdb4bbda7ab #42 Tainted: [W]=WARN Hardware name: QEMU QEMU Virtual Machine, BIOS 0.0.0 02/06/2015 pstate: 20400005 (nzCv daif +PAN -UAO -TCO -DIT -SSBS BTYPE=--) pc : try_to_unmap_one+0x8fc/0xd3c lr : try_to_unmap_one+0x3dc/0xd3c Call trace: try_to_unmap_one+0x8fc/0xd3c (P) try_to_unmap_one+0x3dc/0xd3c (L) rmap_walk_anon+0xdc/0x1f8 rmap_walk+0x3c/0x58 try_to_unmap+0x88/0x90 unmap_poisoned_folio+0x30/0xa8 do_migrate_range+0x4a0/0x568 offline_pages+0x5a4/0x670 memory_block_action+0x17c/0x374 memory_subsys_offline+0x3c/0x78 device_offline+0xa4/0xd0 state_store+0x8c/0xf0 dev_attr_store+0x18/0x2c sysfs_kf_write+0x44/0x54 kernfs_fop_write_iter+0x118/0x1a8 vfs_write+0x3a8/0x4bc ksys_write+0x6c/0xf8 __arm64_sys_write+0x1c/0x28 invoke_syscall+0x44/0x100 el0_svc_common.constprop.0+0x40/0xe0 do_el0_svc+0x1c/0x28 el0_svc+0x30/0xd0 el0t_64_sync_handler+0xc8/0xcc el0t_64_sync+0x198/0x19c ---[ end trace 0000000000000000 ]--- Fixes: 6da6b1d4a7df ("mm/hwpoison: convert TTU_IGNORE_HWPOISON to TTU_HWPOISON") Suggested-by: David Hildenbrand Signed-off-by: Ma Wupeng Acked-by: David Hildenbrand Acked-by: Miaohe Lin Signed-off-by: Ma Wupeng Acked-by: David Hildenbrand Acked-by: Miaohe Lin Signed-off-by: Andrew Morton --- mm/internal.h | 5 ++-- mm/memory-failure.c | 63 ++++++++++++++++++++++----------------------- mm/memory_hotplug.c | 3 ++- 3 files changed, 36 insertions(+), 35 deletions(-) diff --git a/mm/internal.h b/mm/internal.h index 109ef30fee11..20b3535935a3 100644 --- a/mm/internal.h +++ b/mm/internal.h @@ -1115,7 +1115,7 @@ static inline int find_next_best_node(int node, nodemask_t *used_node_mask) * mm/memory-failure.c */ #ifdef CONFIG_MEMORY_FAILURE -void unmap_poisoned_folio(struct folio *folio, enum ttu_flags ttu); +int unmap_poisoned_folio(struct folio *folio, unsigned long pfn, bool must_kill); void shake_folio(struct folio *folio); extern int hwpoison_filter(struct page *p); @@ -1138,8 +1138,9 @@ unsigned long page_mapped_in_vma(const struct page *page, struct vm_area_struct *vma); #else -static inline void unmap_poisoned_folio(struct folio *folio, enum ttu_flags ttu) +static inline int unmap_poisoned_folio(struct folio *folio, unsigned long pfn, bool must_kill) { + return -EBUSY; } #endif diff --git a/mm/memory-failure.c b/mm/memory-failure.c index 995a15eb67e2..327e02fdc029 100644 --- a/mm/memory-failure.c +++ b/mm/memory-failure.c @@ -1556,11 +1556,35 @@ static int get_hwpoison_page(struct page *p, unsigned long flags) return ret; } -void unmap_poisoned_folio(struct folio *folio, enum ttu_flags ttu) +int unmap_poisoned_folio(struct folio *folio, unsigned long pfn, bool must_kill) { - if (folio_test_hugetlb(folio) && !folio_test_anon(folio)) { - struct address_space *mapping; + enum ttu_flags ttu = TTU_IGNORE_MLOCK | TTU_SYNC | TTU_HWPOISON; + struct address_space *mapping; + + if (folio_test_swapcache(folio)) { + pr_err("%#lx: keeping poisoned page in swap cache\n", pfn); + ttu &= ~TTU_HWPOISON; + } + /* + * Propagate the dirty bit from PTEs to struct page first, because we + * need this to decide if we should kill or just drop the page. + * XXX: the dirty test could be racy: set_page_dirty() may not always + * be called inside page lock (it's recommended but not enforced). + */ + mapping = folio_mapping(folio); + if (!must_kill && !folio_test_dirty(folio) && mapping && + mapping_can_writeback(mapping)) { + if (folio_mkclean(folio)) { + folio_set_dirty(folio); + } else { + ttu &= ~TTU_HWPOISON; + pr_info("%#lx: corrupted page was clean: dropped without side effects\n", + pfn); + } + } + + if (folio_test_hugetlb(folio) && !folio_test_anon(folio)) { /* * For hugetlb folios in shared mappings, try_to_unmap * could potentially call huge_pmd_unshare. Because of @@ -1572,7 +1596,7 @@ void unmap_poisoned_folio(struct folio *folio, enum ttu_flags ttu) if (!mapping) { pr_info("%#lx: could not lock mapping for mapped hugetlb folio\n", folio_pfn(folio)); - return; + return -EBUSY; } try_to_unmap(folio, ttu|TTU_RMAP_LOCKED); @@ -1580,6 +1604,8 @@ void unmap_poisoned_folio(struct folio *folio, enum ttu_flags ttu) } else { try_to_unmap(folio, ttu); } + + return folio_mapped(folio) ? -EBUSY : 0; } /* @@ -1589,8 +1615,6 @@ void unmap_poisoned_folio(struct folio *folio, enum ttu_flags ttu) static bool hwpoison_user_mappings(struct folio *folio, struct page *p, unsigned long pfn, int flags) { - enum ttu_flags ttu = TTU_IGNORE_MLOCK | TTU_SYNC | TTU_HWPOISON; - struct address_space *mapping; LIST_HEAD(tokill); bool unmap_success; int forcekill; @@ -1613,29 +1637,6 @@ static bool hwpoison_user_mappings(struct folio *folio, struct page *p, if (!folio_mapped(folio)) return true; - if (folio_test_swapcache(folio)) { - pr_err("%#lx: keeping poisoned page in swap cache\n", pfn); - ttu &= ~TTU_HWPOISON; - } - - /* - * Propagate the dirty bit from PTEs to struct page first, because we - * need this to decide if we should kill or just drop the page. - * XXX: the dirty test could be racy: set_page_dirty() may not always - * be called inside page lock (it's recommended but not enforced). - */ - mapping = folio_mapping(folio); - if (!(flags & MF_MUST_KILL) && !folio_test_dirty(folio) && mapping && - mapping_can_writeback(mapping)) { - if (folio_mkclean(folio)) { - folio_set_dirty(folio); - } else { - ttu &= ~TTU_HWPOISON; - pr_info("%#lx: corrupted page was clean: dropped without side effects\n", - pfn); - } - } - /* * First collect all the processes that have the page * mapped in dirty form. This has to be done before try_to_unmap, @@ -1643,9 +1644,7 @@ static bool hwpoison_user_mappings(struct folio *folio, struct page *p, */ collect_procs(folio, p, &tokill, flags & MF_ACTION_REQUIRED); - unmap_poisoned_folio(folio, ttu); - - unmap_success = !folio_mapped(folio); + unmap_success = !unmap_poisoned_folio(folio, pfn, flags & MF_MUST_KILL); if (!unmap_success) pr_err("%#lx: failed to unmap page (folio mapcount=%d)\n", pfn, folio_mapcount(folio)); diff --git a/mm/memory_hotplug.c b/mm/memory_hotplug.c index e3655f07dd6e..e7e47838fd49 100644 --- a/mm/memory_hotplug.c +++ b/mm/memory_hotplug.c @@ -1833,7 +1833,8 @@ static void do_migrate_range(unsigned long start_pfn, unsigned long end_pfn) if (WARN_ON(folio_test_lru(folio))) folio_isolate_lru(folio); if (folio_mapped(folio)) - unmap_poisoned_folio(folio, TTU_IGNORE_MLOCK); + unmap_poisoned_folio(folio, pfn, false); + continue; }