From patchwork Fri Mar 28 12:19:08 2025 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 8bit X-Patchwork-Submitter: =?utf-8?q?Toke_H=C3=B8iland-J=C3=B8rgensen?= X-Patchwork-Id: 14031873 X-Patchwork-Delegate: kuba@kernel.org Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by smtp.subspace.kernel.org (Postfix) with ESMTPS id 3F37C21ADC3 for ; Fri, 28 Mar 2025 12:19:24 +0000 (UTC) Authentication-Results: smtp.subspace.kernel.org; arc=none smtp.client-ip=170.10.133.124 ARC-Seal: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1743164366; cv=none; b=DTu9jPauS2ii7nsD3wS0irQa3Zc7e27KEV5T81xKJ1PvlWuiUThtU2VomInCsw1DbH8y2xg7/FWE2484cEC0MWD1cqB4bNY8hzAZsq4RA5sIw5wWTYW2iK22Q/Riv1e3U5SFOvCmlAHyccR+U1+OZ58rcwgmot9TAujgkzZPmoc= ARC-Message-Signature: i=1; a=rsa-sha256; d=subspace.kernel.org; s=arc-20240116; t=1743164366; c=relaxed/simple; bh=tBaUVOmuwlUC0F0gAM0GQcCY9DRAAjxRSbfYpkziV60=; h=From:Date:Subject:MIME-Version:Content-Type:Message-Id:References: In-Reply-To:To:Cc; b=AzUY07mCg/L3cnDT/IDTCANDhG0Og56lDNJ4dewx/SAL0vJ7Rf1vWRRrfiguyWgFq5sfiWtMrEA3fDCKvVEDlwBKXa15T6sriIROsoNA6R49T35aI5Q+Jd6z/kcSH9qTSortobOnQS9CYpGKbOIPxbgsh8PatTHOeaQnr7pqFe4= ARC-Authentication-Results: i=1; smtp.subspace.kernel.org; dmarc=pass (p=quarantine dis=none) header.from=redhat.com; spf=pass smtp.mailfrom=redhat.com; dkim=pass (1024-bit key) header.d=redhat.com header.i=@redhat.com header.b=iN3RsIT5; arc=none smtp.client-ip=170.10.133.124 Authentication-Results: smtp.subspace.kernel.org; dmarc=pass (p=quarantine dis=none) header.from=redhat.com Authentication-Results: smtp.subspace.kernel.org; spf=pass smtp.mailfrom=redhat.com Authentication-Results: smtp.subspace.kernel.org; dkim=pass (1024-bit key) header.d=redhat.com header.i=@redhat.com header.b="iN3RsIT5" DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1743164363; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=DD6/BKzSTAkniyKROx0SYrtrdsrFKedUxdUETO2d2g0=; b=iN3RsIT5c0YHwGDxj1+1WXAFfTbkeW7nNyr+IfGG+LljXiDugcpuKcK37y0NY46YNT3Ae5 jId1MJtUQLaTduoY3oetqY/4aRpbiq9n0Xz9rAKvP3pRZLaH3cHOnGmv8ZWLeq1wKQNddW o7HA607vKu4TbhyJUNa2F+O285DiHiU= Received: from mail-ej1-f71.google.com (mail-ej1-f71.google.com [209.85.218.71]) by relay.mimecast.com with ESMTP with STARTTLS (version=TLSv1.3, cipher=TLS_AES_256_GCM_SHA384) id us-mta-260-Nt_t1fdlPD-U1Df7h4sGGA-1; Fri, 28 Mar 2025 08:19:22 -0400 X-MC-Unique: Nt_t1fdlPD-U1Df7h4sGGA-1 X-Mimecast-MFC-AGG-ID: Nt_t1fdlPD-U1Df7h4sGGA_1743164361 Received: by mail-ej1-f71.google.com with SMTP id a640c23a62f3a-ac25852291cso191305966b.2 for ; Fri, 28 Mar 2025 05:19:21 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1743164361; x=1743769161; h=cc:to:in-reply-to:references:message-id:content-transfer-encoding :mime-version:subject:date:from:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=DD6/BKzSTAkniyKROx0SYrtrdsrFKedUxdUETO2d2g0=; b=D39Sq5T9PpuSvbdbnLKl0mkGb6GW9I7mv4RM0MGRjYQEp/PAGpE7V4fOF6qX7DhG/d pkoE5fo3qdY75NcVn3/LmvHlm2CKFjCmoEdAXlzwYTZX3kik/HNxe/IO3nC+6SZsg7cd 3R6Ay+PYEwQNji5zZ5vQ7zMtyoYTYf2I0gVO9E/RAvO4Y1+wRStXhfqeQ+mDVot2w4w6 VQlnItcyI+wi4l3hg3CEojvLEo6gzpYLFt4oTA/vu2sFuBW7EdgDZcxIFOSFHHSRYYCM pP4zgCorCNKnBjgDkSgVbH9aAYciM1kvfXABd3vLD/GpMD4z0JNnMhrzr78IbQ9dHw3g BLJw== X-Forwarded-Encrypted: i=1; AJvYcCXWodOGo4i7hXFWLlJbQ9UXexF9uwbIN5pPHYBqDJrlq6ZcPuN3oGL/CZUHQD2us89ElNw=@vger.kernel.org X-Gm-Message-State: AOJu0YzillzJYpQm48QVeb74Nkmv6wX2CA9Hi0zd4C9yTN7j1n/x9ple AEC7Evd0K+0AIsVcKC3HyZsTHyxwifNMX/3hi2yRLb4oEe+0DubxVuDqHF7V4eCc3nQz1jXzwlI Lg573hkxgBIO4lH3y6JDKOD+/Lz+P7Xlgi07wpyVCMnloEFcU5Q== X-Gm-Gg: ASbGncvS1xu+3KvEh5buCbe5sTAcaKKywYy4sM1uakIFyJepOz9Y/E8IDyts/xP7ygd A2OcEpCAfcW2T5bsjMXJhUpRx4ydW4hVuvi1NM0v/ulidhip+enMAuLk3cECur5g6kPKwPRkznw aRWLA1cMqpg1T3QAA42sgmM6vQrKKbjFcLSqdTJJowRPwcMZpGxbiwagqgM6fyGnGyqF7hE+DTg RzUyMJgkMlAPO0X/IZIt1uslaLICKpcYvg0ZTrm0q6wNSaITre8k2Klw4PGPXxU5Q/DuuwZ+cKf LAWxOJYCcR8H X-Received: by 2002:a17:907:f213:b0:ac7:150e:8013 with SMTP id a640c23a62f3a-ac7150e8725mr465947366b.15.1743164360617; Fri, 28 Mar 2025 05:19:20 -0700 (PDT) X-Google-Smtp-Source: AGHT+IH2sH2HU5ArIiKzX+0LiYkdqwF13nh1l17P2mIqzHoh6BBNKHd8I1cIquZFx0QCVq5kjgQpAg== X-Received: by 2002:a17:907:f213:b0:ac7:150e:8013 with SMTP id a640c23a62f3a-ac7150e8725mr465941466b.15.1743164360121; Fri, 28 Mar 2025 05:19:20 -0700 (PDT) Received: from alrua-x1.borgediget.toke.dk ([2a0c:4d80:42:443::2]) by smtp.gmail.com with ESMTPSA id a640c23a62f3a-ac71922bb65sm152046166b.34.2025.03.28.05.19.18 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 28 Mar 2025 05:19:18 -0700 (PDT) Received: by alrua-x1.borgediget.toke.dk (Postfix, from userid 1000) id 8A09B18FCDC5; Fri, 28 Mar 2025 13:19:17 +0100 (CET) From: =?utf-8?q?Toke_H=C3=B8iland-J=C3=B8rgensen?= Date: Fri, 28 Mar 2025 13:19:08 +0100 Subject: [PATCH net-next v5 1/2] page_pool: Move pp_magic check into helper functions Precedence: bulk X-Mailing-List: bpf@vger.kernel.org List-Id: List-Subscribe: List-Unsubscribe: MIME-Version: 1.0 Message-Id: <20250328-page-pool-track-dma-v5-1-55002af683ad@redhat.com> References: <20250328-page-pool-track-dma-v5-0-55002af683ad@redhat.com> In-Reply-To: <20250328-page-pool-track-dma-v5-0-55002af683ad@redhat.com> To: "David S. Miller" , Jakub Kicinski , Jesper Dangaard Brouer , Saeed Mahameed , Leon Romanovsky , Tariq Toukan , Andrew Lunn , Eric Dumazet , Paolo Abeni , Ilias Apalodimas , Simon Horman , Andrew Morton , Mina Almasry , Yonglong Liu , Yunsheng Lin , Pavel Begunkov , Matthew Wilcox Cc: netdev@vger.kernel.org, bpf@vger.kernel.org, linux-rdma@vger.kernel.org, linux-mm@kvack.org, =?utf-8?q?Toke_H=C3=B8iland-J=C3=B8rgensen?= X-Mailer: b4 0.14.2 X-Patchwork-Delegate: kuba@kernel.org Since we are about to stash some more information into the pp_magic field, let's move the magic signature checks into a pair of helper functions so it can be changed in one place. Reviewed-by: Mina Almasry Tested-by: Yonglong Liu Acked-by: Jesper Dangaard Brouer Reviewed-by: Ilias Apalodimas Signed-off-by: Toke Høiland-Jørgensen --- drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c | 4 ++-- include/net/page_pool/types.h | 18 ++++++++++++++++++ mm/page_alloc.c | 9 +++------ net/core/netmem_priv.h | 5 +++++ net/core/skbuff.c | 16 ++-------------- net/core/xdp.c | 4 ++-- 6 files changed, 32 insertions(+), 24 deletions(-) diff --git a/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c b/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c index f803e1c93590068d3a7829b0683be4af019266d1..5ce1b463b7a8dd7969e391618658d66f6e836cc1 100644 --- a/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c +++ b/drivers/net/ethernet/mellanox/mlx5/core/en/xdp.c @@ -707,8 +707,8 @@ static void mlx5e_free_xdpsq_desc(struct mlx5e_xdpsq *sq, xdpi = mlx5e_xdpi_fifo_pop(xdpi_fifo); page = xdpi.page.page; - /* No need to check ((page->pp_magic & ~0x3UL) == PP_SIGNATURE) - * as we know this is a page_pool page. + /* No need to check page_pool_page_is_pp() as we + * know this is a page_pool page. */ page_pool_recycle_direct(page->pp, page); } while (++n < num); diff --git a/include/net/page_pool/types.h b/include/net/page_pool/types.h index 36eb57d73abc6cfc601e700ca08be20fb8281055..df0d3c1608929605224feb26173135ff37951ef8 100644 --- a/include/net/page_pool/types.h +++ b/include/net/page_pool/types.h @@ -54,6 +54,14 @@ struct pp_alloc_cache { netmem_ref cache[PP_ALLOC_CACHE_SIZE]; }; +/* Mask used for checking in page_pool_page_is_pp() below. page->pp_magic is + * OR'ed with PP_SIGNATURE after the allocation in order to preserve bit 0 for + * the head page of compound page and bit 1 for pfmemalloc page. + * page_is_pfmemalloc() is checked in __page_pool_put_page() to avoid recycling + * the pfmemalloc page. + */ +#define PP_MAGIC_MASK ~0x3UL + /** * struct page_pool_params - page pool parameters * @fast: params accessed frequently on hotpath @@ -264,6 +272,11 @@ void page_pool_destroy(struct page_pool *pool); void page_pool_use_xdp_mem(struct page_pool *pool, void (*disconnect)(void *), const struct xdp_mem_info *mem); void page_pool_put_netmem_bulk(netmem_ref *data, u32 count); + +static inline bool page_pool_page_is_pp(struct page *page) +{ + return (page->pp_magic & PP_MAGIC_MASK) == PP_SIGNATURE; +} #else static inline void page_pool_destroy(struct page_pool *pool) { @@ -278,6 +291,11 @@ static inline void page_pool_use_xdp_mem(struct page_pool *pool, static inline void page_pool_put_netmem_bulk(netmem_ref *data, u32 count) { } + +static inline bool page_pool_page_is_pp(struct page *page) +{ + return false; +} #endif void page_pool_put_unrefed_netmem(struct page_pool *pool, netmem_ref netmem, diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 542d25f77be80304b731411ffd29b276ee13be0c..3535ee76afe946cbb042ecbce603bdbedc9233b9 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c @@ -55,6 +55,7 @@ #include #include #include +#include #include #include "internal.h" #include "shuffle.h" @@ -872,9 +873,7 @@ static inline bool page_expected_state(struct page *page, #ifdef CONFIG_MEMCG page->memcg_data | #endif -#ifdef CONFIG_PAGE_POOL - ((page->pp_magic & ~0x3UL) == PP_SIGNATURE) | -#endif + page_pool_page_is_pp(page) | (page->flags & check_flags))) return false; @@ -901,10 +900,8 @@ static const char *page_bad_reason(struct page *page, unsigned long flags) if (unlikely(page->memcg_data)) bad_reason = "page still charged to cgroup"; #endif -#ifdef CONFIG_PAGE_POOL - if (unlikely((page->pp_magic & ~0x3UL) == PP_SIGNATURE)) + if (unlikely(page_pool_page_is_pp(page))) bad_reason = "page_pool leak"; -#endif return bad_reason; } diff --git a/net/core/netmem_priv.h b/net/core/netmem_priv.h index 7eadb8393e002fd1cc2cef8a313d2ea7df76f301..f33162fd281c23e109273ba09950c5d0a2829bc9 100644 --- a/net/core/netmem_priv.h +++ b/net/core/netmem_priv.h @@ -18,6 +18,11 @@ static inline void netmem_clear_pp_magic(netmem_ref netmem) __netmem_clear_lsb(netmem)->pp_magic = 0; } +static inline bool netmem_is_pp(netmem_ref netmem) +{ + return (netmem_get_pp_magic(netmem) & PP_MAGIC_MASK) == PP_SIGNATURE; +} + static inline void netmem_set_pp(netmem_ref netmem, struct page_pool *pool) { __netmem_clear_lsb(netmem)->pp = pool; diff --git a/net/core/skbuff.c b/net/core/skbuff.c index 6cbf77bc61fce74c934628fd74b3a2cb7809e464..74a2d886a35b518d55b6d3cafcb8442212f9beee 100644 --- a/net/core/skbuff.c +++ b/net/core/skbuff.c @@ -893,11 +893,6 @@ static void skb_clone_fraglist(struct sk_buff *skb) skb_get(list); } -static bool is_pp_netmem(netmem_ref netmem) -{ - return (netmem_get_pp_magic(netmem) & ~0x3UL) == PP_SIGNATURE; -} - int skb_pp_cow_data(struct page_pool *pool, struct sk_buff **pskb, unsigned int headroom) { @@ -995,14 +990,7 @@ bool napi_pp_put_page(netmem_ref netmem) { netmem = netmem_compound_head(netmem); - /* page->pp_magic is OR'ed with PP_SIGNATURE after the allocation - * in order to preserve any existing bits, such as bit 0 for the - * head page of compound page and bit 1 for pfmemalloc page, so - * mask those bits for freeing side when doing below checking, - * and page_is_pfmemalloc() is checked in __page_pool_put_page() - * to avoid recycling the pfmemalloc page. - */ - if (unlikely(!is_pp_netmem(netmem))) + if (unlikely(!netmem_is_pp(netmem))) return false; page_pool_put_full_netmem(netmem_get_pp(netmem), netmem, false); @@ -1042,7 +1030,7 @@ static int skb_pp_frag_ref(struct sk_buff *skb) for (i = 0; i < shinfo->nr_frags; i++) { head_netmem = netmem_compound_head(shinfo->frags[i].netmem); - if (likely(is_pp_netmem(head_netmem))) + if (likely(netmem_is_pp(head_netmem))) page_pool_ref_netmem(head_netmem); else page_ref_inc(netmem_to_page(head_netmem)); diff --git a/net/core/xdp.c b/net/core/xdp.c index f86eedad586a77eb63a96a85aa6d068d3e94f077..0ba73943c6eed873b3d1c681b3b9a802b590f2d9 100644 --- a/net/core/xdp.c +++ b/net/core/xdp.c @@ -437,8 +437,8 @@ void __xdp_return(netmem_ref netmem, enum xdp_mem_type mem_type, netmem = netmem_compound_head(netmem); if (napi_direct && xdp_return_frame_no_direct()) napi_direct = false; - /* No need to check ((page->pp_magic & ~0x3UL) == PP_SIGNATURE) - * as mem->type knows this a page_pool page + /* No need to check netmem_is_pp() as mem->type knows this a + * page_pool page */ page_pool_put_full_netmem(netmem_get_pp(netmem), netmem, napi_direct);