From patchwork Fri Apr 1 13:46:49 2022 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Zqiang X-Patchwork-Id: 12798314 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id D0188C433EF for ; Fri, 1 Apr 2022 13:47:05 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 1654C8D0002; Fri, 1 Apr 2022 09:46:55 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 1160D8D0001; Fri, 1 Apr 2022 09:46:55 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id EF8F78D0002; Fri, 1 Apr 2022 09:46:54 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0180.hostedemail.com [216.40.44.180]) by kanga.kvack.org (Postfix) with ESMTP id DCDE48D0001 for ; Fri, 1 Apr 2022 09:46:54 -0400 (EDT) Received: from smtpin27.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay05.hostedemail.com (Postfix) with ESMTP id 980C81831AD66 for ; Fri, 1 Apr 2022 13:46:44 +0000 (UTC) X-FDA: 79308435528.27.7944AED Received: from mga04.intel.com (mga04.intel.com [192.55.52.120]) by imf20.hostedemail.com (Postfix) with ESMTP id 9D4FC1C0017 for ; Fri, 1 Apr 2022 13:46:43 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1648820803; x=1680356803; h=from:to:cc:subject:date:message-id:mime-version: content-transfer-encoding; bh=Dq4ZeuHSuqCc8HBShNDGOUX84T9yRxGqXlCI/ZSyqZE=; b=JRxvMLv+vPjumTUfbQiL+AJ+ZL3tP0OL2WQ8GpPQBVTRFGw5BXIQOe5e phozpeBXlE4bCTTh67c5LsGOoi/V8sEA9/aPIMtbNf7aeBIFyf4ghC+0/ 6Af5gQz3bum9XLjh2/dK+EDdVy6NBOUfZ+RqWlMhOSMpJpfLLXEEvrQw2 16ThuEfonzdu4F4QHqK1aMB0oPy+kgrnETq4uRIS20yFYUe2Bqn749rsJ RXhEi22OHSnZi4uFtJ/jVNIB51zEbylryLldidgVsVZRNVP1lhO6oUSjp kZRLm4E6oQ/vOTEpQ0yviD830bixJvai0+fBoN2QYoQNcLhN/eBhbufsP Q==; X-IronPort-AV: E=McAfee;i="6200,9189,10304"; a="258970154" X-IronPort-AV: E=Sophos;i="5.90,227,1643702400"; d="scan'208";a="258970154" Received: from orsmga005.jf.intel.com ([10.7.209.41]) by fmsmga104.fm.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 01 Apr 2022 06:46:21 -0700 X-IronPort-AV: E=Sophos;i="5.90,227,1643702400"; d="scan'208";a="720888741" Received: from zq-optiplex-7090.bj.intel.com ([10.238.156.125]) by orsmga005-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 01 Apr 2022 06:46:18 -0700 From: Zqiang To: ryabinin.a.a@gmail.com, glider@google.com, andreyknvl@gmail.com, dvyukov@google.com, bigeasy@linutronix.de Cc: linux-kernel@vger.kernel.org, linux-mm@kvack.org Subject: [PATCH v2] kasan: Fix sleeping function called from invalid context on RT kernel Date: Fri, 1 Apr 2022 21:46:49 +0800 Message-Id: <20220401134649.2222485-1-qiang1.zhang@intel.com> X-Mailer: git-send-email 2.25.1 MIME-Version: 1.0 X-Stat-Signature: yhgh4ws9r1ih9d5w9kpjg4kbnmawniug Authentication-Results: imf20.hostedemail.com; dkim=pass header.d=intel.com header.s=Intel header.b=JRxvMLv+; spf=none (imf20.hostedemail.com: domain of qiang1.zhang@intel.com has no SPF policy when checking 192.55.52.120) smtp.mailfrom=qiang1.zhang@intel.com; dmarc=pass (policy=none) header.from=intel.com X-Rspam-User: X-Rspamd-Server: rspam02 X-Rspamd-Queue-Id: 9D4FC1C0017 X-HE-Tag: 1648820803-617920 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: BUG: sleeping function called from invalid context at kernel/locking/spinlock_rt.c:46 in_atomic(): 1, irqs_disabled(): 1, non_block: 0, pid: 1, name: swapper/0 preempt_count: 1, expected: 0 ........... CPU: 0 PID: 1 Comm: swapper/0 Not tainted 5.17.1-rt16-yocto-preempt-rt #22 Hardware name: QEMU Standard PC (Q35 + ICH9, 2009), BIOS rel-1.15.0-0-g2dd4b9b3f840-prebuilt.qemu.org 04/01/2014 Call Trace: dump_stack_lvl+0x60/0x8c dump_stack+0x10/0x12 __might_resched.cold+0x13b/0x173 rt_spin_lock+0x5b/0xf0 ___cache_free+0xa5/0x180 qlist_free_all+0x7a/0x160 per_cpu_remove_cache+0x5f/0x70 smp_call_function_many_cond+0x4c4/0x4f0 on_each_cpu_cond_mask+0x49/0xc0 kasan_quarantine_remove_cache+0x54/0xf0 kasan_cache_shrink+0x9/0x10 kmem_cache_shrink+0x13/0x20 acpi_os_purge_cache+0xe/0x20 acpi_purge_cached_objects+0x21/0x6d acpi_initialize_objects+0x15/0x3b acpi_init+0x130/0x5ba do_one_initcall+0xe5/0x5b0 kernel_init_freeable+0x34f/0x3ad kernel_init+0x1e/0x140 ret_from_fork+0x22/0x30 When the kmem_cache_shrink() be called, the IPI was triggered, the ___cache_free() is called in IPI interrupt context, the local-lock or spin-lock will be acquired. on PREEMPT_RT kernel, these lock is replaced with sleepbale rt-spinlock, so the above problem is triggered. fix it by move the qlist_free_allfrom() the IPI interrupt context to the task context when PREEMPT_RT is enabled. Signed-off-by: Zqiang Acked-by: Dmitry Vyukov --- v1->v2: Add raw_spinlock protect per-cpu shrink qlist. mm/kasan/quarantine.c | 40 ++++++++++++++++++++++++++++++++++++++-- 1 file changed, 38 insertions(+), 2 deletions(-) diff --git a/mm/kasan/quarantine.c b/mm/kasan/quarantine.c index 08291ed33e93..0e33d30abb8d 100644 --- a/mm/kasan/quarantine.c +++ b/mm/kasan/quarantine.c @@ -99,6 +99,17 @@ static unsigned long quarantine_size; static DEFINE_RAW_SPINLOCK(quarantine_lock); DEFINE_STATIC_SRCU(remove_cache_srcu); +#ifdef CONFIG_PREEMPT_RT +struct cpu_shrink_qlist { + raw_spinlock_t lock; + struct qlist_head qlist; +}; + +static DEFINE_PER_CPU(struct cpu_shrink_qlist, shrink_qlist) = { + .lock = __RAW_SPIN_LOCK_UNLOCKED(shrink_qlist.lock), +}; +#endif + /* Maximum size of the global queue. */ static unsigned long quarantine_max_size; @@ -311,12 +322,23 @@ static void qlist_move_cache(struct qlist_head *from, static void per_cpu_remove_cache(void *arg) { struct kmem_cache *cache = arg; - struct qlist_head to_free = QLIST_INIT; struct qlist_head *q; - +#ifndef CONFIG_PREEMPT_RT + struct qlist_head to_free = QLIST_INIT; +#else + unsigned long flags; + struct cpu_shrink_qlist *sq; +#endif q = this_cpu_ptr(&cpu_quarantine); +#ifndef CONFIG_PREEMPT_RT qlist_move_cache(q, &to_free, cache); qlist_free_all(&to_free, cache); +#else + sq = this_cpu_ptr(&shrink_qlist); + raw_spin_lock_irqsave(&sq->lock, flags); + qlist_move_cache(q, &sq->qlist, cache); + raw_spin_unlock_irqrestore(&sq->lock, flags); +#endif } /* Free all quarantined objects belonging to cache. */ @@ -324,6 +346,10 @@ void kasan_quarantine_remove_cache(struct kmem_cache *cache) { unsigned long flags, i; struct qlist_head to_free = QLIST_INIT; +#ifdef CONFIG_PREEMPT_RT + int cpu; + struct cpu_shrink_qlist *sq; +#endif /* * Must be careful to not miss any objects that are being moved from @@ -334,6 +360,16 @@ void kasan_quarantine_remove_cache(struct kmem_cache *cache) */ on_each_cpu(per_cpu_remove_cache, cache, 1); +#ifdef CONFIG_PREEMPT_RT + for_each_online_cpu(cpu) { + sq = per_cpu_ptr(&shrink_qlist, cpu); + raw_spin_lock_irqsave(&sq->lock, flags); + qlist_move_cache(&sq->qlist, &to_free, cache); + raw_spin_unlock_irqrestore(&sq->lock, flags); + } + qlist_free_all(&to_free, cache); +#endif + raw_spin_lock_irqsave(&quarantine_lock, flags); for (i = 0; i < QUARANTINE_BATCHES; i++) { if (qlist_empty(&global_quarantine[i]))