diff mbox series

[4/9] mm: Remove some calls to page_add_new_anon_rmap()

Message ID 20231211162214.2146080-5-willy@infradead.org (mailing list archive)
State New
Headers show
Series Finish two folio conversions | expand

Commit Message

Matthew Wilcox Dec. 11, 2023, 4:22 p.m. UTC
We already have the folio in these functions, we just need to use it.
folio_add_new_anon_rmap() didn't exist at the time they were converted
to folios.

Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
---
 kernel/events/uprobes.c | 2 +-
 mm/memory.c             | 2 +-
 mm/userfaultfd.c        | 2 +-
 3 files changed, 3 insertions(+), 3 deletions(-)

Comments

David Hildenbrand Dec. 12, 2023, 1:20 p.m. UTC | #1
On 11.12.23 17:22, Matthew Wilcox (Oracle) wrote:
> We already have the folio in these functions, we just need to use it.
> folio_add_new_anon_rmap() didn't exist at the time they were converted
> to folios.
> 
> Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
> ---
>   kernel/events/uprobes.c | 2 +-
>   mm/memory.c             | 2 +-
>   mm/userfaultfd.c        | 2 +-
>   3 files changed, 3 insertions(+), 3 deletions(-)
> 
> diff --git a/kernel/events/uprobes.c b/kernel/events/uprobes.c
> index 435aac1d8c27..8b115fc43f04 100644
> --- a/kernel/events/uprobes.c
> +++ b/kernel/events/uprobes.c
> @@ -181,7 +181,7 @@ static int __replace_page(struct vm_area_struct *vma, unsigned long addr,
>   
>   	if (new_page) {
>   		folio_get(new_folio);
> -		page_add_new_anon_rmap(new_page, vma, addr);
> +		folio_add_new_anon_rmap(new_folio, vma, addr);
>   		folio_add_lru_vma(new_folio, vma);
>   	} else
>   		/* no new page, just dec_mm_counter for old_page */
> diff --git a/mm/memory.c b/mm/memory.c
> index 318f923134e4..c03a7729d5b4 100644
> --- a/mm/memory.c
> +++ b/mm/memory.c
> @@ -4068,7 +4068,7 @@ vm_fault_t do_swap_page(struct vm_fault *vmf)
>   
>   	/* ksm created a completely new copy */
>   	if (unlikely(folio != swapcache && swapcache)) {
> -		page_add_new_anon_rmap(page, vma, vmf->address);
> +		folio_add_new_anon_rmap(folio, vma, vmf->address);
>   		folio_add_lru_vma(folio, vma);
>   	} else {
>   		page_add_anon_rmap(page, vma, vmf->address, rmap_flags);
> diff --git a/mm/userfaultfd.c b/mm/userfaultfd.c
> index 71d0281f1162..2d8b03a009b4 100644
> --- a/mm/userfaultfd.c
> +++ b/mm/userfaultfd.c
> @@ -116,7 +116,7 @@ int mfill_atomic_install_pte(pmd_t *dst_pmd,
>   			folio_add_lru(folio);
>   		page_add_file_rmap(page, dst_vma, false);
>   	} else {
> -		page_add_new_anon_rmap(page, dst_vma, dst_addr);
> +		folio_add_new_anon_rmap(folio, dst_vma, dst_addr);
>   		folio_add_lru_vma(folio, dst_vma);
>   	}
>   

Reviewed-by: David Hildenbrand <david@redhat.com>
diff mbox series

Patch

diff --git a/kernel/events/uprobes.c b/kernel/events/uprobes.c
index 435aac1d8c27..8b115fc43f04 100644
--- a/kernel/events/uprobes.c
+++ b/kernel/events/uprobes.c
@@ -181,7 +181,7 @@  static int __replace_page(struct vm_area_struct *vma, unsigned long addr,
 
 	if (new_page) {
 		folio_get(new_folio);
-		page_add_new_anon_rmap(new_page, vma, addr);
+		folio_add_new_anon_rmap(new_folio, vma, addr);
 		folio_add_lru_vma(new_folio, vma);
 	} else
 		/* no new page, just dec_mm_counter for old_page */
diff --git a/mm/memory.c b/mm/memory.c
index 318f923134e4..c03a7729d5b4 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -4068,7 +4068,7 @@  vm_fault_t do_swap_page(struct vm_fault *vmf)
 
 	/* ksm created a completely new copy */
 	if (unlikely(folio != swapcache && swapcache)) {
-		page_add_new_anon_rmap(page, vma, vmf->address);
+		folio_add_new_anon_rmap(folio, vma, vmf->address);
 		folio_add_lru_vma(folio, vma);
 	} else {
 		page_add_anon_rmap(page, vma, vmf->address, rmap_flags);
diff --git a/mm/userfaultfd.c b/mm/userfaultfd.c
index 71d0281f1162..2d8b03a009b4 100644
--- a/mm/userfaultfd.c
+++ b/mm/userfaultfd.c
@@ -116,7 +116,7 @@  int mfill_atomic_install_pte(pmd_t *dst_pmd,
 			folio_add_lru(folio);
 		page_add_file_rmap(page, dst_vma, false);
 	} else {
-		page_add_new_anon_rmap(page, dst_vma, dst_addr);
+		folio_add_new_anon_rmap(folio, dst_vma, dst_addr);
 		folio_add_lru_vma(folio, dst_vma);
 	}