diff mbox series

[RFC,14/14] mm: Remove free_unref_page_list()

Message ID 20230825135918.4164671-15-willy@infradead.org (mailing list archive)
State New
Headers show
Series Rearrange batched folio freeing | expand

Commit Message

Matthew Wilcox Aug. 25, 2023, 1:59 p.m. UTC
All callers now use free_unref_folios() so we can delete this function.

Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
---
 mm/internal.h   |  1 -
 mm/page_alloc.c | 18 ------------------
 2 files changed, 19 deletions(-)

Comments

Ryan Roberts Aug. 31, 2023, 6:27 p.m. UTC | #1
On 25/08/2023 14:59, Matthew Wilcox (Oracle) wrote:
> All callers now use free_unref_folios() so we can delete this function.
> 
> Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>

Reviewed-by: Ryan Roberts <ryan.roberts@arm.com>

> ---
>  mm/internal.h   |  1 -
>  mm/page_alloc.c | 18 ------------------
>  2 files changed, 19 deletions(-)
> 
> diff --git a/mm/internal.h b/mm/internal.h
> index 5c6a53371aeb..42feb3f59ee5 100644
> --- a/mm/internal.h
> +++ b/mm/internal.h
> @@ -443,7 +443,6 @@ extern int user_min_free_kbytes;
>  
>  void free_unref_page(struct page *page, unsigned int order);
>  void free_unref_folios(struct folio_batch *fbatch);
> -void free_unref_page_list(struct list_head *list);
>  
>  extern void zone_pcp_reset(struct zone *zone);
>  extern void zone_pcp_disable(struct zone *zone);
> diff --git a/mm/page_alloc.c b/mm/page_alloc.c
> index e586d17fb7f2..496304014ec0 100644
> --- a/mm/page_alloc.c
> +++ b/mm/page_alloc.c
> @@ -2554,24 +2554,6 @@ void free_unref_folios(struct folio_batch *folios)
>  	folios->nr = 0;
>  }
>  
> -void free_unref_page_list(struct list_head *list)
> -{
> -	struct folio_batch fbatch;
> -
> -	folio_batch_init(&fbatch);
> -	while (!list_empty(list)) {
> -		struct folio *folio = list_first_entry(list, struct folio, lru);
> -
> -		list_del(&folio->lru);
> -		if (folio_batch_add(&fbatch, folio) > 0)
> -			continue;
> -		free_unref_folios(&fbatch);
> -	}
> -
> -	if (fbatch.nr)
> -		free_unref_folios(&fbatch);
> -}
> -
>  /*
>   * split_page takes a non-compound higher-order page, and splits it into
>   * n (1<<order) sub-pages: page[0..n]
diff mbox series

Patch

diff --git a/mm/internal.h b/mm/internal.h
index 5c6a53371aeb..42feb3f59ee5 100644
--- a/mm/internal.h
+++ b/mm/internal.h
@@ -443,7 +443,6 @@  extern int user_min_free_kbytes;
 
 void free_unref_page(struct page *page, unsigned int order);
 void free_unref_folios(struct folio_batch *fbatch);
-void free_unref_page_list(struct list_head *list);
 
 extern void zone_pcp_reset(struct zone *zone);
 extern void zone_pcp_disable(struct zone *zone);
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index e586d17fb7f2..496304014ec0 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -2554,24 +2554,6 @@  void free_unref_folios(struct folio_batch *folios)
 	folios->nr = 0;
 }
 
-void free_unref_page_list(struct list_head *list)
-{
-	struct folio_batch fbatch;
-
-	folio_batch_init(&fbatch);
-	while (!list_empty(list)) {
-		struct folio *folio = list_first_entry(list, struct folio, lru);
-
-		list_del(&folio->lru);
-		if (folio_batch_add(&fbatch, folio) > 0)
-			continue;
-		free_unref_folios(&fbatch);
-	}
-
-	if (fbatch.nr)
-		free_unref_folios(&fbatch);
-}
-
 /*
  * split_page takes a non-compound higher-order page, and splits it into
  * n (1<<order) sub-pages: page[0..n]