From patchwork Fri Feb 14 04:50:14 2025 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Sergey Senozhatsky X-Patchwork-Id: 13974464 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 7F44DC021A4 for ; Fri, 14 Feb 2025 04:52:34 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 16B0B6B0085; Thu, 13 Feb 2025 23:52:34 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id 0F4276B0089; Thu, 13 Feb 2025 23:52:34 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id E6173280001; Thu, 13 Feb 2025 23:52:33 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0014.hostedemail.com [216.40.44.14]) by kanga.kvack.org (Postfix) with ESMTP id C5D7D6B0085 for ; Thu, 13 Feb 2025 23:52:33 -0500 (EST) Received: from smtpin18.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay01.hostedemail.com (Postfix) with ESMTP id 3EC1D1C93E8 for ; Fri, 14 Feb 2025 04:52:33 +0000 (UTC) X-FDA: 83117329386.18.F0D447F Received: from mail-pl1-f181.google.com (mail-pl1-f181.google.com [209.85.214.181]) by imf30.hostedemail.com (Postfix) with ESMTP id 68A298000B for ; Fri, 14 Feb 2025 04:52:31 +0000 (UTC) Authentication-Results: imf30.hostedemail.com; dkim=pass header.d=chromium.org header.s=google header.b=lhYjrthL; spf=pass (imf30.hostedemail.com: domain of senozhatsky@chromium.org designates 209.85.214.181 as permitted sender) smtp.mailfrom=senozhatsky@chromium.org; dmarc=pass (policy=none) header.from=chromium.org ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1739508751; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=F9dA9jZ1k0KVleOPZIXSXpGbAAlfZaWcl2QUySckvQY=; b=P0VTch20/XP3ZbhUSxqMxcSHr9j/gEzmj63PZfgbwZgwF9rJC9lF7FiNXkpxbYwuEx2k66 jQkcVZ0JinxooTyKbvtAQ9ga8ElH24Jen6Ie/bufmIkaS17QM/JNDVkxzlXQ3zCtY46dsU KY+jdT7gGc+vgEPTz4Ge28cEX4ZyVFU= ARC-Authentication-Results: i=1; imf30.hostedemail.com; dkim=pass header.d=chromium.org header.s=google header.b=lhYjrthL; spf=pass (imf30.hostedemail.com: domain of senozhatsky@chromium.org designates 209.85.214.181 as permitted sender) smtp.mailfrom=senozhatsky@chromium.org; dmarc=pass (policy=none) header.from=chromium.org ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1739508751; a=rsa-sha256; cv=none; b=N/8BFfa/bndWA6dvDuj2CTT+uXPFg9cP4wzq3EY4s4tcPtIM6yuRDFt6jC875uy+hSBy/F eYPwaDjGP9BkCeD0B+ZyiHVh2HnRiEa6DL5M+tJ6j9Z0ngXmM+gxs1QipsQvpXn2tUqZkc +Wp2QdShFr4qwPo3CPLXTofEdwXq+F0= Received: by mail-pl1-f181.google.com with SMTP id d9443c01a7336-220d398bea9so22388075ad.3 for ; Thu, 13 Feb 2025 20:52:31 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=chromium.org; s=google; t=1739508750; x=1740113550; darn=kvack.org; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:from:to:cc:subject:date :message-id:reply-to; bh=F9dA9jZ1k0KVleOPZIXSXpGbAAlfZaWcl2QUySckvQY=; b=lhYjrthLt/vxssY/KKlxqmhBMjfItV47A02v7rFCK6psbBoHDquHi7bux/o6pWzHJ/ Px7sbAcuV3JInFQ1Tyw2lqqP9fFu3Hu6mHRHSX5f0owQU1yfzGFH1nbHlFlfwQ498uoC uDIE94qOYjTz8h6DDc1vMsvsOXKG2K8D0Ym9M= X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1739508750; x=1740113550; h=content-transfer-encoding:mime-version:references:in-reply-to :message-id:date:subject:cc:to:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=F9dA9jZ1k0KVleOPZIXSXpGbAAlfZaWcl2QUySckvQY=; b=U29fhppXLPiVQFXzxEM0FCYAlcnpdiWcP8EhdwnZzlMOFLdKZeF9ybgL4PbdTRLgf6 cvTFGgwQrE3gaYYcnqAR0NVWVcmfFBYDSZhU3ePNCV68MblT4ejATUiFJpd7mW679FwI 26xL1qMcYcM3XEdh0dJMA6/FzDZUVfx5a8F4FJb037x2oA/0FQoJDXgBq4L95dbxRcg0 /KVoNtmi+8CkVmmQGdj9h8EFln53ECawLf7kdlk/N40dxsE8F2+WH0umHcrTqHT9kT1m ixsyDxnpkzAYzu6hQXzogyff2YU9z1GZsFwk9YYLkMAbdESxXsDgeJymv3btBLeVI/Az UTcA== X-Forwarded-Encrypted: i=1; AJvYcCXW/bT6Ld1A/KGjAABRfWsH4Jcr/YxHSrMjQ30Vv3eg0DR7YBtKx9LLowp9PvyIgTDybfBKB4Mf8g==@kvack.org X-Gm-Message-State: AOJu0YwH7Kbvt/qSSaxmvVA/z5xvW2TGsNeZxCtoDNDN4zZfpNWlZDtC aGr1bqN+YVHikLk9OjQ1oepNjywraN333CmHKfJ9gOYiGrLqV65rxpHiCAufHg== X-Gm-Gg: ASbGncvZ7bsu2Qs3YYvg3/ZAH3Bx8uOaYDj4FYvj7sk3J7qh3moihtzhAAfjSu8HP3H 0MM3NSLjHQAK4z0UQax9BUyUpyiJwnDvT7szhflxoURCDO1a2JJToWTxJ0UZuhftKV/DLlFwZ2X c/tejmYn8E9zirboiUPtFJkHCYMxzE9kBXrHltqUvBkUIsP7S7h75zwwqA49h5NPOglmxxALan/ j84rH2TB5Zmx/YuwCS4LRLR+8RUdk3Fky5pP81owscWIDsCYOmdiNFnasTxJnfCchMGub/PPRIR 6jSoXUefhr8YxLdPlA== X-Google-Smtp-Source: AGHT+IHeCccqw7BKz3Nps1nE3WUkuQd67prY/a7knQl57N4aYVrwMvzOwtITAs71quYBmeF0jfumUw== X-Received: by 2002:a05:6a21:898a:b0:1ee:5d05:a197 with SMTP id adf61e73a8af0-1ee5d05a2b1mr13268542637.35.1739508750392; Thu, 13 Feb 2025 20:52:30 -0800 (PST) Received: from localhost ([2401:fa00:8f:203:942d:9291:22aa:8126]) by smtp.gmail.com with UTF8SMTPSA id d2e1a72fcca58-73242761c13sm2252705b3a.136.2025.02.13.20.52.27 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Thu, 13 Feb 2025 20:52:30 -0800 (PST) From: Sergey Senozhatsky To: Andrew Morton Cc: Yosry Ahmed , Hillf Danton , Kairui Song , Minchan Kim , linux-mm@kvack.org, linux-kernel@vger.kernel.org, Sergey Senozhatsky Subject: [PATCH v6 02/17] zram: permit preemption with active compression stream Date: Fri, 14 Feb 2025 13:50:14 +0900 Message-ID: <20250214045208.1388854-3-senozhatsky@chromium.org> X-Mailer: git-send-email 2.48.1.601.g30ceb7b040-goog In-Reply-To: <20250214045208.1388854-1-senozhatsky@chromium.org> References: <20250214045208.1388854-1-senozhatsky@chromium.org> MIME-Version: 1.0 X-Rspam-User: X-Rspamd-Server: rspam11 X-Rspamd-Queue-Id: 68A298000B X-Stat-Signature: mwde4jzw91mjcxngk7ugibrjamdnn8rj X-HE-Tag: 1739508751-631304 X-HE-Meta: U2FsdGVkX1/1YqLSvxRPJSgNZVWT3D2tzztg+hBOGUqjVlmj2taEMdhwtHNkW3nr/a7b9qSKFGAIkIPSMX77Ta07eKjyZuOfqSavXX2iXqgrPoAypiU0YiAhJZ1Q8HRTjKB1DfHvNHbEgck1q/eDEaa0AW7eXMFISVvYR0AtHXq4u+MAx+dspBiweGn0+zm6bRmOfz+/yVj8TQJXMQKo3CppxXpx6AeIXbwtSCo0OuFeXDkvFn62AWYdkq7fvwwJhO/n2y/P6AblPudHN3WX7qkX/K0oLYTsPJ+N2Dg992ythwyrsJoRynWCeOOKZ/jE8/Fiqhx1eW1rAY/hEs1kwZ1bV2ypWR6CWB8lMvCohsUavX54yZiLIEYIXX2xH/PqNaKwOJWefJPJYsgNanUnCCshQZ0jfl/2yUeoH7bvq0Sci9D2H9iy7oDEbO+bo1+5ZzpTrrP8SR8gjTpDfU0NRBgrd9WQdUuMfL5T1p2u3/q4I9nidfj2rMi8QN0KAqG8BZNZ+1H0uIry6/a/5/nCDlzvaEsP7b4lEpk0mfppZ0saISrTxCONMB9d6tZEVh04T/kVmgHxXU9rVVUySsC26N+2Eq+joR/ANXQqrtm+pREBsnFaiXK75rHYl/mI7yvKgAN9RGjqeQBImvWr/38pmtsvbat0NwPHT7GzzfqYdeyXwS2Y3jkHrW1dlkkhSXvuzGLsM4p7131WcDDW7/pxJpcBBqUtXWlOS02QUtHYYZRwFK4RYlQoI8RS4wm8Etp6yEJ0jBl/RoDTpTfjeOyw3NnCKU1EZd5BngHXCOZzg0FoAEhDF8fQ/bK9RDgoqFZdom3morzL9nn+2g9H5dCvsNvrlD0rJjKmFiC85FFJNxsHz3t7HK0q8NsQd/0W+dKbw/NN2KKoguBTofg/hmpDBXSzyzkH4No9T2eer7Sr4MmHXM7WS/UdTZp1r0LaeKDuYjbb9nl5VrWpa933eDZ /DN/xlX6 nV83a/7gcHkPEE7LyZ97PiQVDw+bvj+pHKgGZmGFDI/ZOr3695ruRNucIS3rH9Phmmndhz6ne8NM3flzldHg4VlyjYGdmbVJJrzO47RftuA7IcxCiYeA7xHIPcBIBxJU/L9TgrpcoQsNOkQo4UmYsbNHKCDGSGhxCHt9nPg8FbLgcGbm8ZeYRP2vDC2q7P3xYz9DAyF7s/oS4N5jhVlXrsduxF96NXfsvuHlsG3PLr90GGiI1mv0BikdB8nlfmbCaygnRdO2T63EPB/tRYnpVteL2WyENum8gANMB9T0oEpBWL6GUvHjMiZY88w== X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: Currently, per-CPU stream access is done from a non-preemptible (atomic) section, which imposes the same atomicity requirements on compression backends as entry spin-lock, and makes it impossible to use algorithms that can schedule/wait/sleep during compression and decompression. Switch to preemptible per-CPU model, similar to the one used in zswap. Instead of a per-CPU local lock, each stream carries a mutex which is locked throughout entire time zram uses it for compression or decompression, so that cpu-dead event waits for zram to stop using a particular per-CPU stream and release it. Suggested-by: Yosry Ahmed Signed-off-by: Sergey Senozhatsky Reviewed-by: Yosry Ahmed --- drivers/block/zram/zcomp.c | 41 +++++++++++++++++++++++++---------- drivers/block/zram/zcomp.h | 6 ++--- drivers/block/zram/zram_drv.c | 20 ++++++++--------- 3 files changed, 42 insertions(+), 25 deletions(-) diff --git a/drivers/block/zram/zcomp.c b/drivers/block/zram/zcomp.c index bb514403e305..53e4c37441be 100644 --- a/drivers/block/zram/zcomp.c +++ b/drivers/block/zram/zcomp.c @@ -6,7 +6,7 @@ #include #include #include -#include +#include #include #include @@ -109,13 +109,29 @@ ssize_t zcomp_available_show(const char *comp, char *buf) struct zcomp_strm *zcomp_stream_get(struct zcomp *comp) { - local_lock(&comp->stream->lock); - return this_cpu_ptr(comp->stream); + for (;;) { + struct zcomp_strm *zstrm = raw_cpu_ptr(comp->stream); + + /* + * Inspired by zswap + * + * stream is returned with ->mutex locked which prevents + * cpu_dead() from releasing this stream under us, however + * there is still a race window between raw_cpu_ptr() and + * mutex_lock(), during which we could have been migrated + * from a CPU that has already destroyed its stream. If + * so then unlock and re-try on the current CPU. + */ + mutex_lock(&zstrm->lock); + if (likely(zstrm->buffer)) + return zstrm; + mutex_unlock(&zstrm->lock); + } } -void zcomp_stream_put(struct zcomp *comp) +void zcomp_stream_put(struct zcomp_strm *zstrm) { - local_unlock(&comp->stream->lock); + mutex_unlock(&zstrm->lock); } int zcomp_compress(struct zcomp *comp, struct zcomp_strm *zstrm, @@ -151,12 +167,9 @@ int zcomp_decompress(struct zcomp *comp, struct zcomp_strm *zstrm, int zcomp_cpu_up_prepare(unsigned int cpu, struct hlist_node *node) { struct zcomp *comp = hlist_entry(node, struct zcomp, node); - struct zcomp_strm *zstrm; + struct zcomp_strm *zstrm = per_cpu_ptr(comp->stream, cpu); int ret; - zstrm = per_cpu_ptr(comp->stream, cpu); - local_lock_init(&zstrm->lock); - ret = zcomp_strm_init(comp, zstrm); if (ret) pr_err("Can't allocate a compression stream\n"); @@ -166,16 +179,17 @@ int zcomp_cpu_up_prepare(unsigned int cpu, struct hlist_node *node) int zcomp_cpu_dead(unsigned int cpu, struct hlist_node *node) { struct zcomp *comp = hlist_entry(node, struct zcomp, node); - struct zcomp_strm *zstrm; + struct zcomp_strm *zstrm = per_cpu_ptr(comp->stream, cpu); - zstrm = per_cpu_ptr(comp->stream, cpu); + mutex_lock(&zstrm->lock); zcomp_strm_free(comp, zstrm); + mutex_unlock(&zstrm->lock); return 0; } static int zcomp_init(struct zcomp *comp, struct zcomp_params *params) { - int ret; + int ret, cpu; comp->stream = alloc_percpu(struct zcomp_strm); if (!comp->stream) @@ -186,6 +200,9 @@ static int zcomp_init(struct zcomp *comp, struct zcomp_params *params) if (ret) goto cleanup; + for_each_possible_cpu(cpu) + mutex_init(&per_cpu_ptr(comp->stream, cpu)->lock); + ret = cpuhp_state_add_instance(CPUHP_ZCOMP_PREPARE, &comp->node); if (ret < 0) goto cleanup; diff --git a/drivers/block/zram/zcomp.h b/drivers/block/zram/zcomp.h index ad5762813842..23b8236b9090 100644 --- a/drivers/block/zram/zcomp.h +++ b/drivers/block/zram/zcomp.h @@ -3,7 +3,7 @@ #ifndef _ZCOMP_H_ #define _ZCOMP_H_ -#include +#include #define ZCOMP_PARAM_NO_LEVEL INT_MIN @@ -31,7 +31,7 @@ struct zcomp_ctx { }; struct zcomp_strm { - local_lock_t lock; + struct mutex lock; /* compression buffer */ void *buffer; struct zcomp_ctx ctx; @@ -77,7 +77,7 @@ struct zcomp *zcomp_create(const char *alg, struct zcomp_params *params); void zcomp_destroy(struct zcomp *comp); struct zcomp_strm *zcomp_stream_get(struct zcomp *comp); -void zcomp_stream_put(struct zcomp *comp); +void zcomp_stream_put(struct zcomp_strm *zstrm); int zcomp_compress(struct zcomp *comp, struct zcomp_strm *zstrm, const void *src, unsigned int *dst_len); diff --git a/drivers/block/zram/zram_drv.c b/drivers/block/zram/zram_drv.c index 65e16117f2db..ca439f3b1b9a 100644 --- a/drivers/block/zram/zram_drv.c +++ b/drivers/block/zram/zram_drv.c @@ -1650,7 +1650,7 @@ static int read_compressed_page(struct zram *zram, struct page *page, u32 index) ret = zcomp_decompress(zram->comps[prio], zstrm, src, size, dst); kunmap_local(dst); zs_unmap_object(zram->mem_pool, handle); - zcomp_stream_put(zram->comps[prio]); + zcomp_stream_put(zstrm); return ret; } @@ -1811,14 +1811,14 @@ static int zram_write_page(struct zram *zram, struct page *page, u32 index) kunmap_local(mem); if (unlikely(ret)) { - zcomp_stream_put(zram->comps[ZRAM_PRIMARY_COMP]); + zcomp_stream_put(zstrm); pr_err("Compression failed! err=%d\n", ret); zs_free(zram->mem_pool, handle); return ret; } if (comp_len >= huge_class_size) { - zcomp_stream_put(zram->comps[ZRAM_PRIMARY_COMP]); + zcomp_stream_put(zstrm); return write_incompressible_page(zram, page, index); } @@ -1842,7 +1842,7 @@ static int zram_write_page(struct zram *zram, struct page *page, u32 index) __GFP_HIGHMEM | __GFP_MOVABLE); if (IS_ERR_VALUE(handle)) { - zcomp_stream_put(zram->comps[ZRAM_PRIMARY_COMP]); + zcomp_stream_put(zstrm); atomic64_inc(&zram->stats.writestall); handle = zs_malloc(zram->mem_pool, comp_len, GFP_NOIO | __GFP_HIGHMEM | @@ -1854,7 +1854,7 @@ static int zram_write_page(struct zram *zram, struct page *page, u32 index) } if (!zram_can_store_page(zram)) { - zcomp_stream_put(zram->comps[ZRAM_PRIMARY_COMP]); + zcomp_stream_put(zstrm); zs_free(zram->mem_pool, handle); return -ENOMEM; } @@ -1862,7 +1862,7 @@ static int zram_write_page(struct zram *zram, struct page *page, u32 index) dst = zs_map_object(zram->mem_pool, handle, ZS_MM_WO); memcpy(dst, zstrm->buffer, comp_len); - zcomp_stream_put(zram->comps[ZRAM_PRIMARY_COMP]); + zcomp_stream_put(zstrm); zs_unmap_object(zram->mem_pool, handle); zram_slot_lock(zram, index); @@ -2021,7 +2021,7 @@ static int recompress_slot(struct zram *zram, u32 index, struct page *page, kunmap_local(src); if (ret) { - zcomp_stream_put(zram->comps[prio]); + zcomp_stream_put(zstrm); return ret; } @@ -2031,7 +2031,7 @@ static int recompress_slot(struct zram *zram, u32 index, struct page *page, /* Continue until we make progress */ if (class_index_new >= class_index_old || (threshold && comp_len_new >= threshold)) { - zcomp_stream_put(zram->comps[prio]); + zcomp_stream_put(zstrm); continue; } @@ -2089,13 +2089,13 @@ static int recompress_slot(struct zram *zram, u32 index, struct page *page, __GFP_HIGHMEM | __GFP_MOVABLE); if (IS_ERR_VALUE(handle_new)) { - zcomp_stream_put(zram->comps[prio]); + zcomp_stream_put(zstrm); return PTR_ERR((void *)handle_new); } dst = zs_map_object(zram->mem_pool, handle_new, ZS_MM_WO); memcpy(dst, zstrm->buffer, comp_len_new); - zcomp_stream_put(zram->comps[prio]); + zcomp_stream_put(zstrm); zs_unmap_object(zram->mem_pool, handle_new);