diff mbox series

[mmotm] mm/thp: refix __split_huge_pmd_locked() for migration PMD

Message ID 84792468-f512-e48f-378c-e34c3641e97@google.com (mailing list archive)
State New
Headers show
Series [mmotm] mm/thp: refix __split_huge_pmd_locked() for migration PMD | expand

Commit Message

Hugh Dickins March 3, 2022, 1:43 a.m. UTC
Migration entries do not contribute to a page's reference count: move
__split_huge_pmd_locked()'s page_ref_add() into pmd_migration's else
block (along with the page_count() check - a page is quite likely to
to have reference count frozen to 0 when a migration entry is found).

This will fix a very rare anonymous memory leak, after a split_huge_pmd()
raced with an anon split_huge_page() or an anon THP migrate_pages(): since
the wrongly raised refcount stopped the page (perhaps small, perhaps huge,
depending on when the race hit) from ever being freed.  At first I thought
there were worse risks, from prematurely unfreezing a frozen page: but now
think that would only affect page cache pages, which do not come this way
(except for anonymous pages in swap cache, perhaps).

Fixes: ec0abae6dcdf ("mm/thp: fix __split_huge_pmd_locked() for migration PMD")
Signed-off-by: Hugh Dickins <hughd@google.com>
---
That's an unfair "Fixes": it did not introduce the problem, but it
missed this aspect of the problem; and will be a good guide to where this
refix should go if stable backports are asked for.

 mm/huge_memory.c | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

Comments

Yang Shi March 3, 2022, 7:51 p.m. UTC | #1
On Wed, Mar 2, 2022 at 5:43 PM Hugh Dickins <hughd@google.com> wrote:
>
> Migration entries do not contribute to a page's reference count: move
> __split_huge_pmd_locked()'s page_ref_add() into pmd_migration's else
> block (along with the page_count() check - a page is quite likely to
> to have reference count frozen to 0 when a migration entry is found).
>
> This will fix a very rare anonymous memory leak, after a split_huge_pmd()
> raced with an anon split_huge_page() or an anon THP migrate_pages(): since
> the wrongly raised refcount stopped the page (perhaps small, perhaps huge,
> depending on when the race hit) from ever being freed.  At first I thought
> there were worse risks, from prematurely unfreezing a frozen page: but now
> think that would only affect page cache pages, which do not come this way
> (except for anonymous pages in swap cache, perhaps).

Thanks for catching this. I agree there may be anon memory leak due to
bumped refcount. But I don't think it could affect page cache page
since that code (bumping refcount) is never called for page cache page
IIUC.

The patch looks good to me. Reviewed-by: Yang Shi <shy828301@gmail.com>

>
> Fixes: ec0abae6dcdf ("mm/thp: fix __split_huge_pmd_locked() for migration PMD")
> Signed-off-by: Hugh Dickins <hughd@google.com>
> ---
> That's an unfair "Fixes": it did not introduce the problem, but it
> missed this aspect of the problem; and will be a good guide to where this
> refix should go if stable backports are asked for.
>
>  mm/huge_memory.c | 4 ++--
>  1 file changed, 2 insertions(+), 2 deletions(-)
>
> --- a/mm/huge_memory.c
> +++ b/mm/huge_memory.c
> @@ -2039,9 +2039,9 @@ static void __split_huge_pmd_locked(struct vm_area_struct *vma, pmd_t *pmd,
>                 young = pmd_young(old_pmd);
>                 soft_dirty = pmd_soft_dirty(old_pmd);
>                 uffd_wp = pmd_uffd_wp(old_pmd);
> +               VM_BUG_ON_PAGE(!page_count(page), page);
> +               page_ref_add(page, HPAGE_PMD_NR - 1);
>         }
> -       VM_BUG_ON_PAGE(!page_count(page), page);
> -       page_ref_add(page, HPAGE_PMD_NR - 1);
>
>         /*
>          * Withdraw the table only after we mark the pmd entry invalid.
Hugh Dickins March 3, 2022, 10:45 p.m. UTC | #2
On Thu, 3 Mar 2022, Yang Shi wrote:
> On Wed, Mar 2, 2022 at 5:43 PM Hugh Dickins <hughd@google.com> wrote:
> >
> > Migration entries do not contribute to a page's reference count: move
> > __split_huge_pmd_locked()'s page_ref_add() into pmd_migration's else
> > block (along with the page_count() check - a page is quite likely to
> > to have reference count frozen to 0 when a migration entry is found).
> >
> > This will fix a very rare anonymous memory leak, after a split_huge_pmd()
> > raced with an anon split_huge_page() or an anon THP migrate_pages(): since
> > the wrongly raised refcount stopped the page (perhaps small, perhaps huge,
> > depending on when the race hit) from ever being freed.  At first I thought
> > there were worse risks, from prematurely unfreezing a frozen page: but now
> > think that would only affect page cache pages, which do not come this way
> > (except for anonymous pages in swap cache, perhaps).
> 
> Thanks for catching this. I agree there may be anon memory leak due to
> bumped refcount. But I don't think it could affect page cache page
> since that code (bumping refcount) is never called for page cache page
> IIUC.

Yes, that's what I meant by unfreezing "would only affect page cache pages,
which do not come this way".  But then remembered that anonymous pages in
swap cache also have non-NULL page_mapping(), and involve freezing in the
same way that page cache pages do.  Now, offhand I forget the status of
hugepage swap, so maybe that's impossible, hence "perhaps".  Easier to
fix than to think through all the ramifications of not fixing, as usual.

> 
> The patch looks good to me. Reviewed-by: Yang Shi <shy828301@gmail.com>

Thanks,
Hugh
diff mbox series

Patch

--- a/mm/huge_memory.c
+++ b/mm/huge_memory.c
@@ -2039,9 +2039,9 @@  static void __split_huge_pmd_locked(struct vm_area_struct *vma, pmd_t *pmd,
 		young = pmd_young(old_pmd);
 		soft_dirty = pmd_soft_dirty(old_pmd);
 		uffd_wp = pmd_uffd_wp(old_pmd);
+		VM_BUG_ON_PAGE(!page_count(page), page);
+		page_ref_add(page, HPAGE_PMD_NR - 1);
 	}
-	VM_BUG_ON_PAGE(!page_count(page), page);
-	page_ref_add(page, HPAGE_PMD_NR - 1);
 
 	/*
 	 * Withdraw the table only after we mark the pmd entry invalid.