diff mbox series

[08/10] mm: Turn deactivate_file_page() into deactivate_file_folio()

Message ID 20220214200017.3150590-9-willy@infradead.org (mailing list archive)
State New, archived
Headers show
Series Various fixes around invalidate_page() | expand

Commit Message

Matthew Wilcox Feb. 14, 2022, 8 p.m. UTC
This function has one caller which already has a reference to the
page, so we don't need to use get_page_unless_zero().  Also move the
prototype to mm/internal.h.

Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
---
 include/linux/swap.h |  1 -
 mm/internal.h        |  1 +
 mm/swap.c            | 33 ++++++++++++++++-----------------
 mm/truncate.c        |  2 +-
 4 files changed, 18 insertions(+), 19 deletions(-)

Comments

Christoph Hellwig Feb. 15, 2022, 7:25 a.m. UTC | #1
On Mon, Feb 14, 2022 at 08:00:15PM +0000, Matthew Wilcox (Oracle) wrote:
> This function has one caller which already has a reference to the
> page, so we don't need to use get_page_unless_zero().  Also move the
> prototype to mm/internal.h.

Looks good:

Reviewed-by: Christoph Hellwig <hch@lst.de>
Miaohe Lin Feb. 15, 2022, 8:26 a.m. UTC | #2
On 2022/2/15 4:00, Matthew Wilcox (Oracle) wrote:
> This function has one caller which already has a reference to the
> page, so we don't need to use get_page_unless_zero().  Also move the
> prototype to mm/internal.h.
> 
> Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
> ---
>  include/linux/swap.h |  1 -
>  mm/internal.h        |  1 +
>  mm/swap.c            | 33 ++++++++++++++++-----------------
>  mm/truncate.c        |  2 +-
>  4 files changed, 18 insertions(+), 19 deletions(-)
> 
> diff --git a/include/linux/swap.h b/include/linux/swap.h
> index 304f174b4d31..064e60e9f63f 100644
> --- a/include/linux/swap.h
> +++ b/include/linux/swap.h
> @@ -372,7 +372,6 @@ extern void lru_add_drain(void);
>  extern void lru_add_drain_cpu(int cpu);
>  extern void lru_add_drain_cpu_zone(struct zone *zone);
>  extern void lru_add_drain_all(void);
> -extern void deactivate_file_page(struct page *page);
>  extern void deactivate_page(struct page *page);
>  extern void mark_page_lazyfree(struct page *page);
>  extern void swap_setup(void);
> diff --git a/mm/internal.h b/mm/internal.h
> index 927a17d58b85..d886b87b1294 100644
> --- a/mm/internal.h
> +++ b/mm/internal.h
> @@ -66,6 +66,7 @@ static inline void wake_throttle_isolated(pg_data_t *pgdat)
>  vm_fault_t do_swap_page(struct vm_fault *vmf);
>  void folio_rotate_reclaimable(struct folio *folio);
>  bool __folio_end_writeback(struct folio *folio);
> +void deactivate_file_folio(struct folio *folio);
>  
>  void free_pgtables(struct mmu_gather *tlb, struct vm_area_struct *start_vma,
>  		unsigned long floor, unsigned long ceiling);
> diff --git a/mm/swap.c b/mm/swap.c
> index bcf3ac288b56..745915127b1f 100644
> --- a/mm/swap.c
> +++ b/mm/swap.c
> @@ -639,32 +639,31 @@ void lru_add_drain_cpu(int cpu)
>  }
>  
>  /**
> - * deactivate_file_page - forcefully deactivate a file page
> - * @page: page to deactivate
> + * deactivate_file_folio() - Forcefully deactivate a file folio.
> + * @folio: Folio to deactivate.
>   *
> - * This function hints the VM that @page is a good reclaim candidate,
> - * for example if its invalidation fails due to the page being dirty
> + * This function hints to the VM that @folio is a good reclaim candidate,
> + * for example if its invalidation fails due to the folio being dirty
>   * or under writeback.
>   */
> -void deactivate_file_page(struct page *page)
> +void deactivate_file_folio(struct folio *folio)
>  {
> +	struct pagevec *pvec;
> +
>  	/*
> -	 * In a workload with many unevictable page such as mprotect,
> -	 * unevictable page deactivation for accelerating reclaim is pointless.
> +	 * In a workload with many unevictable pages such as mprotect,
> +	 * unevictable folio deactivation for accelerating reclaim is pointless.
>  	 */
> -	if (PageUnevictable(page))
> +	if (folio_test_unevictable(folio))
>  		return;
>  
> -	if (likely(get_page_unless_zero(page))) {
> -		struct pagevec *pvec;
> -
> -		local_lock(&lru_pvecs.lock);
> -		pvec = this_cpu_ptr(&lru_pvecs.lru_deactivate_file);
> +	folio_get(folio);

Should we comment the assumption that caller already hold the refcnt?

Anyway, this patch looks good to me. Thanks.

Reviewed-by: Miaohe Lin <linmiaohe@huawei.com>

> +	local_lock(&lru_pvecs.lock);
> +	pvec = this_cpu_ptr(&lru_pvecs.lru_deactivate_file);
>  
> -		if (pagevec_add_and_need_flush(pvec, page))
> -			pagevec_lru_move_fn(pvec, lru_deactivate_file_fn);
> -		local_unlock(&lru_pvecs.lock);
> -	}
> +	if (pagevec_add_and_need_flush(pvec, &folio->page))
> +		pagevec_lru_move_fn(pvec, lru_deactivate_file_fn);
> +	local_unlock(&lru_pvecs.lock);
>  }
>  
>  /*
> diff --git a/mm/truncate.c b/mm/truncate.c
> index 567557c36d45..14486e75ec28 100644
> --- a/mm/truncate.c
> +++ b/mm/truncate.c
> @@ -525,7 +525,7 @@ static unsigned long __invalidate_mapping_pages(struct address_space *mapping,
>  			 * of interest and try to speed up its reclaim.
>  			 */
>  			if (!ret) {
> -				deactivate_file_page(&folio->page);
> +				deactivate_file_folio(folio);
>  				/* It is likely on the pagevec of a remote CPU */
>  				if (nr_pagevec)
>  					(*nr_pagevec)++;
>
Matthew Wilcox Feb. 15, 2022, 8:33 p.m. UTC | #3
On Tue, Feb 15, 2022 at 04:26:22PM +0800, Miaohe Lin wrote:
> > +	folio_get(folio);
> 
> Should we comment the assumption that caller already hold the refcnt?

Added to the kernel-doc:
+ * Context: Caller holds a reference on the page.

> Anyway, this patch looks good to me. Thanks.
> 
> Reviewed-by: Miaohe Lin <linmiaohe@huawei.com>
> 
> > +	local_lock(&lru_pvecs.lock);
> > +	pvec = this_cpu_ptr(&lru_pvecs.lru_deactivate_file);
> >  
> > -		if (pagevec_add_and_need_flush(pvec, page))
> > -			pagevec_lru_move_fn(pvec, lru_deactivate_file_fn);
> > -		local_unlock(&lru_pvecs.lock);
> > -	}
> > +	if (pagevec_add_and_need_flush(pvec, &folio->page))
> > +		pagevec_lru_move_fn(pvec, lru_deactivate_file_fn);
> > +	local_unlock(&lru_pvecs.lock);
> >  }
> >  
> >  /*
> > diff --git a/mm/truncate.c b/mm/truncate.c
> > index 567557c36d45..14486e75ec28 100644
> > --- a/mm/truncate.c
> > +++ b/mm/truncate.c
> > @@ -525,7 +525,7 @@ static unsigned long __invalidate_mapping_pages(struct address_space *mapping,
> >  			 * of interest and try to speed up its reclaim.
> >  			 */
> >  			if (!ret) {
> > -				deactivate_file_page(&folio->page);
> > +				deactivate_file_folio(folio);
> >  				/* It is likely on the pagevec of a remote CPU */
> >  				if (nr_pagevec)
> >  					(*nr_pagevec)++;
> > 
> 
>
Miaohe Lin Feb. 16, 2022, 2:45 a.m. UTC | #4
On 2022/2/16 4:33, Matthew Wilcox wrote:
> On Tue, Feb 15, 2022 at 04:26:22PM +0800, Miaohe Lin wrote:
>>> +	folio_get(folio);
>>
>> Should we comment the assumption that caller already hold the refcnt?
> 
> Added to the kernel-doc:
> + * Context: Caller holds a reference on the page.
> 

I see. Thanks.

>> Anyway, this patch looks good to me. Thanks.
>>
>> Reviewed-by: Miaohe Lin <linmiaohe@huawei.com>
>>
>>> +	local_lock(&lru_pvecs.lock);
>>> +	pvec = this_cpu_ptr(&lru_pvecs.lru_deactivate_file);
>>>  
>>> -		if (pagevec_add_and_need_flush(pvec, page))
>>> -			pagevec_lru_move_fn(pvec, lru_deactivate_file_fn);
>>> -		local_unlock(&lru_pvecs.lock);
>>> -	}
>>> +	if (pagevec_add_and_need_flush(pvec, &folio->page))
>>> +		pagevec_lru_move_fn(pvec, lru_deactivate_file_fn);
>>> +	local_unlock(&lru_pvecs.lock);
>>>  }
>>>  
>>>  /*
>>> diff --git a/mm/truncate.c b/mm/truncate.c
>>> index 567557c36d45..14486e75ec28 100644
>>> --- a/mm/truncate.c
>>> +++ b/mm/truncate.c
>>> @@ -525,7 +525,7 @@ static unsigned long __invalidate_mapping_pages(struct address_space *mapping,
>>>  			 * of interest and try to speed up its reclaim.
>>>  			 */
>>>  			if (!ret) {
>>> -				deactivate_file_page(&folio->page);
>>> +				deactivate_file_folio(folio);
>>>  				/* It is likely on the pagevec of a remote CPU */
>>>  				if (nr_pagevec)
>>>  					(*nr_pagevec)++;
>>>
>>
>>
> .
>
diff mbox series

Patch

diff --git a/include/linux/swap.h b/include/linux/swap.h
index 304f174b4d31..064e60e9f63f 100644
--- a/include/linux/swap.h
+++ b/include/linux/swap.h
@@ -372,7 +372,6 @@  extern void lru_add_drain(void);
 extern void lru_add_drain_cpu(int cpu);
 extern void lru_add_drain_cpu_zone(struct zone *zone);
 extern void lru_add_drain_all(void);
-extern void deactivate_file_page(struct page *page);
 extern void deactivate_page(struct page *page);
 extern void mark_page_lazyfree(struct page *page);
 extern void swap_setup(void);
diff --git a/mm/internal.h b/mm/internal.h
index 927a17d58b85..d886b87b1294 100644
--- a/mm/internal.h
+++ b/mm/internal.h
@@ -66,6 +66,7 @@  static inline void wake_throttle_isolated(pg_data_t *pgdat)
 vm_fault_t do_swap_page(struct vm_fault *vmf);
 void folio_rotate_reclaimable(struct folio *folio);
 bool __folio_end_writeback(struct folio *folio);
+void deactivate_file_folio(struct folio *folio);
 
 void free_pgtables(struct mmu_gather *tlb, struct vm_area_struct *start_vma,
 		unsigned long floor, unsigned long ceiling);
diff --git a/mm/swap.c b/mm/swap.c
index bcf3ac288b56..745915127b1f 100644
--- a/mm/swap.c
+++ b/mm/swap.c
@@ -639,32 +639,31 @@  void lru_add_drain_cpu(int cpu)
 }
 
 /**
- * deactivate_file_page - forcefully deactivate a file page
- * @page: page to deactivate
+ * deactivate_file_folio() - Forcefully deactivate a file folio.
+ * @folio: Folio to deactivate.
  *
- * This function hints the VM that @page is a good reclaim candidate,
- * for example if its invalidation fails due to the page being dirty
+ * This function hints to the VM that @folio is a good reclaim candidate,
+ * for example if its invalidation fails due to the folio being dirty
  * or under writeback.
  */
-void deactivate_file_page(struct page *page)
+void deactivate_file_folio(struct folio *folio)
 {
+	struct pagevec *pvec;
+
 	/*
-	 * In a workload with many unevictable page such as mprotect,
-	 * unevictable page deactivation for accelerating reclaim is pointless.
+	 * In a workload with many unevictable pages such as mprotect,
+	 * unevictable folio deactivation for accelerating reclaim is pointless.
 	 */
-	if (PageUnevictable(page))
+	if (folio_test_unevictable(folio))
 		return;
 
-	if (likely(get_page_unless_zero(page))) {
-		struct pagevec *pvec;
-
-		local_lock(&lru_pvecs.lock);
-		pvec = this_cpu_ptr(&lru_pvecs.lru_deactivate_file);
+	folio_get(folio);
+	local_lock(&lru_pvecs.lock);
+	pvec = this_cpu_ptr(&lru_pvecs.lru_deactivate_file);
 
-		if (pagevec_add_and_need_flush(pvec, page))
-			pagevec_lru_move_fn(pvec, lru_deactivate_file_fn);
-		local_unlock(&lru_pvecs.lock);
-	}
+	if (pagevec_add_and_need_flush(pvec, &folio->page))
+		pagevec_lru_move_fn(pvec, lru_deactivate_file_fn);
+	local_unlock(&lru_pvecs.lock);
 }
 
 /*
diff --git a/mm/truncate.c b/mm/truncate.c
index 567557c36d45..14486e75ec28 100644
--- a/mm/truncate.c
+++ b/mm/truncate.c
@@ -525,7 +525,7 @@  static unsigned long __invalidate_mapping_pages(struct address_space *mapping,
 			 * of interest and try to speed up its reclaim.
 			 */
 			if (!ret) {
-				deactivate_file_page(&folio->page);
+				deactivate_file_folio(folio);
 				/* It is likely on the pagevec of a remote CPU */
 				if (nr_pagevec)
 					(*nr_pagevec)++;