Message ID | 20210125164612.243838-4-alobakin@pm.me (mailing list archive) |
---|---|
State | Superseded |
Delegated to: | Netdev Maintainers |
Headers | show |
Series | net: constify page_is_pfmemalloc() and its users | expand |
Context | Check | Description |
---|---|---|
netdev/cover_letter | success | Link |
netdev/fixes_present | success | Link |
netdev/patch_count | success | Link |
netdev/tree_selection | success | Clearly marked for net-next |
netdev/subject_prefix | success | Link |
netdev/cc_maintainers | success | CCed 5 of 5 maintainers |
netdev/source_inline | success | Was 0 now: 0 |
netdev/verify_signedoff | success | Link |
netdev/module_param | success | Was 0 now: 0 |
netdev/build_32bit | success | Errors and warnings before: 1 this patch: 1 |
netdev/kdoc | success | Errors and warnings before: 0 this patch: 0 |
netdev/verify_fixes | success | Link |
netdev/checkpatch | success | total: 0 errors, 0 warnings, 0 checks, 27 lines checked |
netdev/build_allmodconfig_warn | success | Errors and warnings before: 1 this patch: 1 |
netdev/header_inline | success | Link |
netdev/stable | success | Stable not CCed |
On Mon, 25 Jan 2021 16:47:20 +0000 Alexander Lobakin <alobakin@pm.me> wrote: > pool_page_reusable() is a leftover from pre-NUMA-aware times. For now, > this function is just a redundant wrapper over page_is_pfmemalloc(), > so Inline it into its sole call site. > > Signed-off-by: Alexander Lobakin <alobakin@pm.me> > --- > net/core/page_pool.c | 14 ++++---------- > 1 file changed, 4 insertions(+), 10 deletions(-) Acked-by: Jesper Dangaard Brouer <brouer@redhat.com> > > diff --git a/net/core/page_pool.c b/net/core/page_pool.c > index f3c690b8c8e3..ad8b0707af04 100644 > --- a/net/core/page_pool.c > +++ b/net/core/page_pool.c > @@ -350,14 +350,6 @@ static bool page_pool_recycle_in_cache(struct page *page, > return true; > } > > -/* page is NOT reusable when: > - * 1) allocated when system is under some pressure. (page_is_pfmemalloc) > - */ > -static bool pool_page_reusable(struct page_pool *pool, struct page *page) > -{ > - return !page_is_pfmemalloc(page); > -} > - > /* If the page refcnt == 1, this will try to recycle the page. > * if PP_FLAG_DMA_SYNC_DEV is set, we'll try to sync the DMA area for > * the configured size min(dma_sync_size, pool->max_len). > @@ -373,9 +365,11 @@ __page_pool_put_page(struct page_pool *pool, struct page *page, > * regular page allocator APIs. > * > * refcnt == 1 means page_pool owns page, and can recycle it. > + * > + * page is NOT reusable when allocated when system is under > + * some pressure. (page_is_pfmemalloc) > */ > - if (likely(page_ref_count(page) == 1 && > - pool_page_reusable(pool, page))) { > + if (likely(page_ref_count(page) == 1 && !page_is_pfmemalloc(page))) { > /* Read barrier done in page_ref_count / READ_ONCE */ > > if (pool->p.flags & PP_FLAG_DMA_SYNC_DEV)
On Mon, Jan 25, 2021 at 04:47:20PM +0000, Alexander Lobakin wrote: > pool_page_reusable() is a leftover from pre-NUMA-aware times. For now, > this function is just a redundant wrapper over page_is_pfmemalloc(), > so Inline it into its sole call site. > > Signed-off-by: Alexander Lobakin <alobakin@pm.me> > --- > net/core/page_pool.c | 14 ++++---------- > 1 file changed, 4 insertions(+), 10 deletions(-) > > diff --git a/net/core/page_pool.c b/net/core/page_pool.c > index f3c690b8c8e3..ad8b0707af04 100644 > --- a/net/core/page_pool.c > +++ b/net/core/page_pool.c > @@ -350,14 +350,6 @@ static bool page_pool_recycle_in_cache(struct page *page, > return true; > } > > -/* page is NOT reusable when: > - * 1) allocated when system is under some pressure. (page_is_pfmemalloc) > - */ > -static bool pool_page_reusable(struct page_pool *pool, struct page *page) > -{ > - return !page_is_pfmemalloc(page); > -} > - > /* If the page refcnt == 1, this will try to recycle the page. > * if PP_FLAG_DMA_SYNC_DEV is set, we'll try to sync the DMA area for > * the configured size min(dma_sync_size, pool->max_len). > @@ -373,9 +365,11 @@ __page_pool_put_page(struct page_pool *pool, struct page *page, > * regular page allocator APIs. > * > * refcnt == 1 means page_pool owns page, and can recycle it. > + * > + * page is NOT reusable when allocated when system is under > + * some pressure. (page_is_pfmemalloc) > */ > - if (likely(page_ref_count(page) == 1 && > - pool_page_reusable(pool, page))) { > + if (likely(page_ref_count(page) == 1 && !page_is_pfmemalloc(page))) { > /* Read barrier done in page_ref_count / READ_ONCE */ > > if (pool->p.flags & PP_FLAG_DMA_SYNC_DEV) > -- > 2.30.0 > > Reviewed-by: Ilias Apalodimas <ilias.apalodimas@linaro.org>
diff --git a/net/core/page_pool.c b/net/core/page_pool.c index f3c690b8c8e3..ad8b0707af04 100644 --- a/net/core/page_pool.c +++ b/net/core/page_pool.c @@ -350,14 +350,6 @@ static bool page_pool_recycle_in_cache(struct page *page, return true; } -/* page is NOT reusable when: - * 1) allocated when system is under some pressure. (page_is_pfmemalloc) - */ -static bool pool_page_reusable(struct page_pool *pool, struct page *page) -{ - return !page_is_pfmemalloc(page); -} - /* If the page refcnt == 1, this will try to recycle the page. * if PP_FLAG_DMA_SYNC_DEV is set, we'll try to sync the DMA area for * the configured size min(dma_sync_size, pool->max_len). @@ -373,9 +365,11 @@ __page_pool_put_page(struct page_pool *pool, struct page *page, * regular page allocator APIs. * * refcnt == 1 means page_pool owns page, and can recycle it. + * + * page is NOT reusable when allocated when system is under + * some pressure. (page_is_pfmemalloc) */ - if (likely(page_ref_count(page) == 1 && - pool_page_reusable(pool, page))) { + if (likely(page_ref_count(page) == 1 && !page_is_pfmemalloc(page))) { /* Read barrier done in page_ref_count / READ_ONCE */ if (pool->p.flags & PP_FLAG_DMA_SYNC_DEV)
pool_page_reusable() is a leftover from pre-NUMA-aware times. For now, this function is just a redundant wrapper over page_is_pfmemalloc(), so Inline it into its sole call site. Signed-off-by: Alexander Lobakin <alobakin@pm.me> --- net/core/page_pool.c | 14 ++++---------- 1 file changed, 4 insertions(+), 10 deletions(-)