diff mbox series

[v4,1/2] mm: migration: fix the FOLL_GET failure on following huge page

Message ID 20220815015909.439623-2-haiyue.wang@intel.com (mailing list archive)
State Superseded
Headers show
Series fix follow_page related issues | expand

Commit Message

Haiyue Wang Aug. 15, 2022, 1:59 a.m. UTC
Not all huge page APIs support FOLL_GET option, so the __NR_move_pages
will fail to get the page node information for huge page.

This is an temporary solution to mitigate the racing fix.

After supporting follow huge page by FOLL_GET is done, this fix can be
reverted safely.

Fixes: 4cd614841c06 ("mm: migration: fix possible do_pages_stat_array racing with memory offline")
Signed-off-by: Haiyue Wang <haiyue.wang@intel.com>
---
 mm/migrate.c | 10 ++++++++--
 1 file changed, 8 insertions(+), 2 deletions(-)

Comments

Alistair Popple Aug. 15, 2022, 4:28 a.m. UTC | #1
On Monday, 15 August 2022 11:59:08 AM AEST Haiyue Wang wrote:
> Not all huge page APIs support FOLL_GET option, so the __NR_move_pages
> will fail to get the page node information for huge page.

I think you should be explicit in the commit message about which functions do 
not support FOLL_GET as it's not obvious what support needs to be added before 
this fix can be reverted.

Thanks.

 - Alistair

> This is an temporary solution to mitigate the racing fix.
> 
> After supporting follow huge page by FOLL_GET is done, this fix can be
> reverted safely.
> 
> Fixes: 4cd614841c06 ("mm: migration: fix possible do_pages_stat_array racing 
with memory offline")
> Signed-off-by: Haiyue Wang <haiyue.wang@intel.com>
> ---
>  mm/migrate.c | 10 ++++++++--
>  1 file changed, 8 insertions(+), 2 deletions(-)
> 
> diff --git a/mm/migrate.c b/mm/migrate.c
> index 6a1597c92261..581dfaad9257 100644
> --- a/mm/migrate.c
> +++ b/mm/migrate.c
> @@ -1848,6 +1848,7 @@ static void do_pages_stat_array(struct mm_struct *mm, 
unsigned long nr_pages,
>  
>  	for (i = 0; i < nr_pages; i++) {
>  		unsigned long addr = (unsigned long)(*pages);
> +		unsigned int foll_flags = FOLL_DUMP;
>  		struct vm_area_struct *vma;
>  		struct page *page;
>  		int err = -EFAULT;
> @@ -1856,8 +1857,12 @@ static void do_pages_stat_array(struct mm_struct *mm, 
unsigned long nr_pages,
>  		if (!vma)
>  			goto set_status;
>  
> +		/* Not all huge page follow APIs support 'FOLL_GET' */
> +		if (!is_vm_hugetlb_page(vma))
> +			foll_flags |= FOLL_GET;
> +
>  		/* FOLL_DUMP to ignore special (like zero) pages */
> -		page = follow_page(vma, addr, FOLL_GET | FOLL_DUMP);
> +		page = follow_page(vma, addr, foll_flags);
>  
>  		err = PTR_ERR(page);
>  		if (IS_ERR(page))
> @@ -1865,7 +1870,8 @@ static void do_pages_stat_array(struct mm_struct *mm, 
unsigned long nr_pages,
>  
>  		if (page && !is_zone_device_page(page)) {
>  			err = page_to_nid(page);
> -			put_page(page);
> +			if (foll_flags & FOLL_GET)
> +				put_page(page);
>  		} else {
>  			err = -ENOENT;
>  		}
>
Haiyue Wang Aug. 15, 2022, 4:40 a.m. UTC | #2
> -----Original Message-----
> From: Alistair Popple <apopple@nvidia.com>
> Sent: Monday, August 15, 2022 12:29
> To: linux-mm@kvack.org; linux-kernel@vger.kernel.org; Wang, Haiyue <haiyue.wang@intel.com>
> Cc: akpm@linux-foundation.org; david@redhat.com; linmiaohe@huawei.com; Huang, Ying
> <ying.huang@intel.com>; songmuchun@bytedance.com; naoya.horiguchi@linux.dev; alex.sierra@amd.com; Wang,
> Haiyue <haiyue.wang@intel.com>
> Subject: Re: [PATCH v4 1/2] mm: migration: fix the FOLL_GET failure on following huge page
> 
> On Monday, 15 August 2022 11:59:08 AM AEST Haiyue Wang wrote:
> > Not all huge page APIs support FOLL_GET option, so the __NR_move_pages
> > will fail to get the page node information for huge page.
> 
> I think you should be explicit in the commit message about which functions do
> not support FOLL_GET as it's not obvious what support needs to be added before
> this fix can be reverted.

Yes, make sense, will add them in new patch.

> 
> Thanks.
> 
>  - Alistair
> 
> > This is an temporary solution to mitigate the racing fix.
> >
> > After supporting follow huge page by FOLL_GET is done, this fix can be
> > reverted safely.
> >
> > Fixes: 4cd614841c06 ("mm: migration: fix possible do_pages_stat_array racing
> with memory offline")
> > Signed-off-by: Haiyue Wang <haiyue.wang@intel.com>
> > ---
> >  mm/migrate.c | 10 ++++++++--
> >  1 file changed, 8 insertions(+), 2 deletions(-)
> >
> > diff --git a/mm/migrate.c b/mm/migrate.c
> > index 6a1597c92261..581dfaad9257 100644
> > --- a/mm/migrate.c
> > +++ b/mm/migrate.c
> > @@ -1848,6 +1848,7 @@ static void do_pages_stat_array(struct mm_struct *mm,
> unsigned long nr_pages,
> >
> >  	for (i = 0; i < nr_pages; i++) {
> >  		unsigned long addr = (unsigned long)(*pages);
> > +		unsigned int foll_flags = FOLL_DUMP;
> >  		struct vm_area_struct *vma;
> >  		struct page *page;
> >  		int err = -EFAULT;
> > @@ -1856,8 +1857,12 @@ static void do_pages_stat_array(struct mm_struct *mm,
> unsigned long nr_pages,
> >  		if (!vma)
> >  			goto set_status;
> >
> > +		/* Not all huge page follow APIs support 'FOLL_GET' */
> > +		if (!is_vm_hugetlb_page(vma))
> > +			foll_flags |= FOLL_GET;
> > +
> >  		/* FOLL_DUMP to ignore special (like zero) pages */
> > -		page = follow_page(vma, addr, FOLL_GET | FOLL_DUMP);
> > +		page = follow_page(vma, addr, foll_flags);
> >
> >  		err = PTR_ERR(page);
> >  		if (IS_ERR(page))
> > @@ -1865,7 +1870,8 @@ static void do_pages_stat_array(struct mm_struct *mm,
> unsigned long nr_pages,
> >
> >  		if (page && !is_zone_device_page(page)) {
> >  			err = page_to_nid(page);
> > -			put_page(page);
> > +			if (foll_flags & FOLL_GET)
> > +				put_page(page);
> >  		} else {
> >  			err = -ENOENT;
> >  		}
> >
> 
> 
>
Alistair Popple Aug. 15, 2022, 5:16 a.m. UTC | #3
On Monday, 15 August 2022 2:40:48 PM AEST Wang, Haiyue wrote:
> > -----Original Message-----
> > From: Alistair Popple <apopple@nvidia.com>
> > Sent: Monday, August 15, 2022 12:29
> > To: linux-mm@kvack.org; linux-kernel@vger.kernel.org; Wang, Haiyue 
<haiyue.wang@intel.com>
> > Cc: akpm@linux-foundation.org; david@redhat.com; linmiaohe@huawei.com; 
Huang, Ying
> > <ying.huang@intel.com>; songmuchun@bytedance.com; 
naoya.horiguchi@linux.dev; alex.sierra@amd.com; Wang,
> > Haiyue <haiyue.wang@intel.com>
> > Subject: Re: [PATCH v4 1/2] mm: migration: fix the FOLL_GET failure on 
following huge page
> > 
> > On Monday, 15 August 2022 11:59:08 AM AEST Haiyue Wang wrote:
> > > Not all huge page APIs support FOLL_GET option, so the __NR_move_pages
> > > will fail to get the page node information for huge page.
> > 
> > I think you should be explicit in the commit message about which functions 
do
> > not support FOLL_GET as it's not obvious what support needs to be added 
before
> > this fix can be reverted.
> 
> Yes, make sense, will add them in new patch.

Actually while you're at it I think it would be good to include a description 
of the impact of this failure in the commit message. Ie. You're answer to:

> What are the user-visible runtime effects of this bug?

As it documents what should be tested if this fix does actually ever get 
reverted.

> > 
> > Thanks.
> > 
> >  - Alistair
> > 
> > > This is an temporary solution to mitigate the racing fix.
> > >
> > > After supporting follow huge page by FOLL_GET is done, this fix can be
> > > reverted safely.
> > >
> > > Fixes: 4cd614841c06 ("mm: migration: fix possible do_pages_stat_array 
racing
> > with memory offline")
> > > Signed-off-by: Haiyue Wang <haiyue.wang@intel.com>
> > > ---
> > >  mm/migrate.c | 10 ++++++++--
> > >  1 file changed, 8 insertions(+), 2 deletions(-)
> > >
> > > diff --git a/mm/migrate.c b/mm/migrate.c
> > > index 6a1597c92261..581dfaad9257 100644
> > > --- a/mm/migrate.c
> > > +++ b/mm/migrate.c
> > > @@ -1848,6 +1848,7 @@ static void do_pages_stat_array(struct mm_struct 
*mm,
> > unsigned long nr_pages,
> > >
> > >  	for (i = 0; i < nr_pages; i++) {
> > >  		unsigned long addr = (unsigned long)(*pages);
> > > +		unsigned int foll_flags = FOLL_DUMP;
> > >  		struct vm_area_struct *vma;
> > >  		struct page *page;
> > >  		int err = -EFAULT;
> > > @@ -1856,8 +1857,12 @@ static void do_pages_stat_array(struct mm_struct 
*mm,
> > unsigned long nr_pages,
> > >  		if (!vma)
> > >  			goto set_status;
> > >
> > > +		/* Not all huge page follow APIs support 'FOLL_GET' */
> > > +		if (!is_vm_hugetlb_page(vma))
> > > +			foll_flags |= FOLL_GET;
> > > +
> > >  		/* FOLL_DUMP to ignore special (like zero) pages */
> > > -		page = follow_page(vma, addr, FOLL_GET | FOLL_DUMP);
> > > +		page = follow_page(vma, addr, foll_flags);
> > >
> > >  		err = PTR_ERR(page);
> > >  		if (IS_ERR(page))
> > > @@ -1865,7 +1870,8 @@ static void do_pages_stat_array(struct mm_struct 
*mm,
> > unsigned long nr_pages,
> > >
> > >  		if (page && !is_zone_device_page(page)) {
> > >  			err = page_to_nid(page);
> > > -			put_page(page);
> > > +			if (foll_flags & FOLL_GET)
> > > +				put_page(page);
> > >  		} else {
> > >  			err = -ENOENT;
> > >  		}
> > >
> > 
> > 
> > 
> 
>
Haiyue Wang Aug. 15, 2022, 5:20 a.m. UTC | #4
> -----Original Message-----
> From: Alistair Popple <apopple@nvidia.com>
> Sent: Monday, August 15, 2022 13:17
> To: linux-mm@kvack.org; linux-kernel@vger.kernel.org; Wang, Haiyue <haiyue.wang@intel.com>
> Cc: akpm@linux-foundation.org; david@redhat.com; linmiaohe@huawei.com; Huang, Ying
> <ying.huang@intel.com>; songmuchun@bytedance.com; naoya.horiguchi@linux.dev; alex.sierra@amd.com
> Subject: Re: [PATCH v4 1/2] mm: migration: fix the FOLL_GET failure on following huge page
> 
> On Monday, 15 August 2022 2:40:48 PM AEST Wang, Haiyue wrote:
> > > -----Original Message-----
> > > From: Alistair Popple <apopple@nvidia.com>
> > > Sent: Monday, August 15, 2022 12:29
> > > To: linux-mm@kvack.org; linux-kernel@vger.kernel.org; Wang, Haiyue
> <haiyue.wang@intel.com>
> > > Cc: akpm@linux-foundation.org; david@redhat.com; linmiaohe@huawei.com;
> Huang, Ying
> > > <ying.huang@intel.com>; songmuchun@bytedance.com;
> naoya.horiguchi@linux.dev; alex.sierra@amd.com; Wang,
> > > Haiyue <haiyue.wang@intel.com>
> > > Subject: Re: [PATCH v4 1/2] mm: migration: fix the FOLL_GET failure on
> following huge page
> > >
> > > On Monday, 15 August 2022 11:59:08 AM AEST Haiyue Wang wrote:
> > > > Not all huge page APIs support FOLL_GET option, so the __NR_move_pages
> > > > will fail to get the page node information for huge page.
> > >
> > > I think you should be explicit in the commit message about which functions
> do
> > > not support FOLL_GET as it's not obvious what support needs to be added
> before
> > > this fix can be reverted.
> >
> > Yes, make sense, will add them in new patch.
> 
> Actually while you're at it I think it would be good to include a description
> of the impact of this failure in the commit message. Ie. You're answer to:
> 
> > What are the user-visible runtime effects of this bug?
> 
> As it documents what should be tested if this fix does actually ever get
> reverted.

An short example *.c code to capture the bug in commit message ?

> 
> > >
> > > Thanks.
> > >
> > >  - Alistair
> > >
> > > > This is an temporary solution to mitigate the racing fix.
> > > >
> > > > After supporting follow huge page by FOLL_GET is done, this fix can be
> > > > reverted safely.
> > > >
> > > > Fixes: 4cd614841c06 ("mm: migration: fix possible do_pages_stat_array
> racing
> > > with memory offline")
> > > > Signed-off-by: Haiyue Wang <haiyue.wang@intel.com>
> > > > ---
> > > >  mm/migrate.c | 10 ++++++++--
> > > >  1 file changed, 8 insertions(+), 2 deletions(-)
> > > >
> > > > diff --git a/mm/migrate.c b/mm/migrate.c
> > > > index 6a1597c92261..581dfaad9257 100644
> > > > --- a/mm/migrate.c
> > > > +++ b/mm/migrate.c
> > > > @@ -1848,6 +1848,7 @@ static void do_pages_stat_array(struct mm_struct
> *mm,
> > > unsigned long nr_pages,
> > > >
> > > >  	for (i = 0; i < nr_pages; i++) {
> > > >  		unsigned long addr = (unsigned long)(*pages);
> > > > +		unsigned int foll_flags = FOLL_DUMP;
> > > >  		struct vm_area_struct *vma;
> > > >  		struct page *page;
> > > >  		int err = -EFAULT;
> > > > @@ -1856,8 +1857,12 @@ static void do_pages_stat_array(struct mm_struct
> *mm,
> > > unsigned long nr_pages,
> > > >  		if (!vma)
> > > >  			goto set_status;
> > > >
> > > > +		/* Not all huge page follow APIs support 'FOLL_GET' */
> > > > +		if (!is_vm_hugetlb_page(vma))
> > > > +			foll_flags |= FOLL_GET;
> > > > +
> > > >  		/* FOLL_DUMP to ignore special (like zero) pages */
> > > > -		page = follow_page(vma, addr, FOLL_GET | FOLL_DUMP);
> > > > +		page = follow_page(vma, addr, foll_flags);
> > > >
> > > >  		err = PTR_ERR(page);
> > > >  		if (IS_ERR(page))
> > > > @@ -1865,7 +1870,8 @@ static void do_pages_stat_array(struct mm_struct
> *mm,
> > > unsigned long nr_pages,
> > > >
> > > >  		if (page && !is_zone_device_page(page)) {
> > > >  			err = page_to_nid(page);
> > > > -			put_page(page);
> > > > +			if (foll_flags & FOLL_GET)
> > > > +				put_page(page);
> > > >  		} else {
> > > >  			err = -ENOENT;
> > > >  		}
> > > >
> > >
> > >
> > >
> >
> >
> 
> 
>
Alistair Popple Aug. 15, 2022, 5:35 a.m. UTC | #5
On Monday, 15 August 2022 3:20:28 PM AEST Wang, Haiyue wrote:
> > -----Original Message-----
> > From: Alistair Popple <apopple@nvidia.com>
> > Sent: Monday, August 15, 2022 13:17
> > To: linux-mm@kvack.org; linux-kernel@vger.kernel.org; Wang, Haiyue 
<haiyue.wang@intel.com>
> > Cc: akpm@linux-foundation.org; david@redhat.com; linmiaohe@huawei.com; 
Huang, Ying
> > <ying.huang@intel.com>; songmuchun@bytedance.com; 
naoya.horiguchi@linux.dev; alex.sierra@amd.com
> > Subject: Re: [PATCH v4 1/2] mm: migration: fix the FOLL_GET failure on 
following huge page
> > 
> > On Monday, 15 August 2022 2:40:48 PM AEST Wang, Haiyue wrote:
> > > > -----Original Message-----
> > > > From: Alistair Popple <apopple@nvidia.com>
> > > > Sent: Monday, August 15, 2022 12:29
> > > > To: linux-mm@kvack.org; linux-kernel@vger.kernel.org; Wang, Haiyue
> > <haiyue.wang@intel.com>
> > > > Cc: akpm@linux-foundation.org; david@redhat.com; linmiaohe@huawei.com;
> > Huang, Ying
> > > > <ying.huang@intel.com>; songmuchun@bytedance.com;
> > naoya.horiguchi@linux.dev; alex.sierra@amd.com; Wang,
> > > > Haiyue <haiyue.wang@intel.com>
> > > > Subject: Re: [PATCH v4 1/2] mm: migration: fix the FOLL_GET failure on
> > following huge page
> > > >
> > > > On Monday, 15 August 2022 11:59:08 AM AEST Haiyue Wang wrote:
> > > > > Not all huge page APIs support FOLL_GET option, so the 
__NR_move_pages
> > > > > will fail to get the page node information for huge page.
> > > >
> > > > I think you should be explicit in the commit message about which 
functions
> > do
> > > > not support FOLL_GET as it's not obvious what support needs to be 
added
> > before
> > > > this fix can be reverted.
> > >
> > > Yes, make sense, will add them in new patch.
> > 
> > Actually while you're at it I think it would be good to include a 
description
> > of the impact of this failure in the commit message. Ie. You're answer to:
> > 
> > > What are the user-visible runtime effects of this bug?
> > 
> > As it documents what should be tested if this fix does actually ever get
> > reverted.
> 
> An short example *.c code to capture the bug in commit message ?

That's probably overkill. Just being a bit more explicit about the 
circumstances in which sys_move_pages() actually fails would be good. Eg. 
something like this:

"Without this sys_move_pages() will return -ENOENT for 1GB huge page
memory map when dumping the page node information for nodes != NULL"

> > 
> > > >
> > > > Thanks.
> > > >
> > > >  - Alistair
> > > >
> > > > > This is an temporary solution to mitigate the racing fix.
> > > > >
> > > > > After supporting follow huge page by FOLL_GET is done, this fix can 
be
> > > > > reverted safely.
> > > > >
> > > > > Fixes: 4cd614841c06 ("mm: migration: fix possible 
do_pages_stat_array
> > racing
> > > > with memory offline")
> > > > > Signed-off-by: Haiyue Wang <haiyue.wang@intel.com>
> > > > > ---
> > > > >  mm/migrate.c | 10 ++++++++--
> > > > >  1 file changed, 8 insertions(+), 2 deletions(-)
> > > > >
> > > > > diff --git a/mm/migrate.c b/mm/migrate.c
> > > > > index 6a1597c92261..581dfaad9257 100644
> > > > > --- a/mm/migrate.c
> > > > > +++ b/mm/migrate.c
> > > > > @@ -1848,6 +1848,7 @@ static void do_pages_stat_array(struct 
mm_struct
> > *mm,
> > > > unsigned long nr_pages,
> > > > >
> > > > >  	for (i = 0; i < nr_pages; i++) {
> > > > >  		unsigned long addr = (unsigned long)(*pages);
> > > > > +		unsigned int foll_flags = FOLL_DUMP;
> > > > >  		struct vm_area_struct *vma;
> > > > >  		struct page *page;
> > > > >  		int err = -EFAULT;
> > > > > @@ -1856,8 +1857,12 @@ static void do_pages_stat_array(struct 
mm_struct
> > *mm,
> > > > unsigned long nr_pages,
> > > > >  		if (!vma)
> > > > >  			goto set_status;
> > > > >
> > > > > +		/* Not all huge page follow APIs support 'FOLL_GET' 
*/
> > > > > +		if (!is_vm_hugetlb_page(vma))
> > > > > +			foll_flags |= FOLL_GET;
> > > > > +
> > > > >  		/* FOLL_DUMP to ignore special (like zero) pages */
> > > > > -		page = follow_page(vma, addr, FOLL_GET | FOLL_DUMP);
> > > > > +		page = follow_page(vma, addr, foll_flags);
> > > > >
> > > > >  		err = PTR_ERR(page);
> > > > >  		if (IS_ERR(page))
> > > > > @@ -1865,7 +1870,8 @@ static void do_pages_stat_array(struct 
mm_struct
> > *mm,
> > > > unsigned long nr_pages,
> > > > >
> > > > >  		if (page && !is_zone_device_page(page)) {
> > > > >  			err = page_to_nid(page);
> > > > > -			put_page(page);
> > > > > +			if (foll_flags & FOLL_GET)
> > > > > +				put_page(page);
> > > > >  		} else {
> > > > >  			err = -ENOENT;
> > > > >  		}
> > > > >
> > > >
> > > >
> > > >
> > >
> > >
> > 
> > 
> > 
> 
>
Haiyue Wang Aug. 15, 2022, 5:37 a.m. UTC | #6
> -----Original Message-----
> From: Alistair Popple <apopple@nvidia.com>
> Sent: Monday, August 15, 2022 13:35
> To: linux-mm@kvack.org; linux-kernel@vger.kernel.org; Wang, Haiyue <haiyue.wang@intel.com>
> Cc: akpm@linux-foundation.org; david@redhat.com; linmiaohe@huawei.com; Huang, Ying
> <ying.huang@intel.com>; songmuchun@bytedance.com; naoya.horiguchi@linux.dev; alex.sierra@amd.com
> Subject: Re: [PATCH v4 1/2] mm: migration: fix the FOLL_GET failure on following huge page
> 
> On Monday, 15 August 2022 3:20:28 PM AEST Wang, Haiyue wrote:
> > > -----Original Message-----
> > > From: Alistair Popple <apopple@nvidia.com>
> > > Sent: Monday, August 15, 2022 13:17
> > > To: linux-mm@kvack.org; linux-kernel@vger.kernel.org; Wang, Haiyue
> <haiyue.wang@intel.com>
> > > Cc: akpm@linux-foundation.org; david@redhat.com; linmiaohe@huawei.com;
> Huang, Ying
> > > <ying.huang@intel.com>; songmuchun@bytedance.com;
> naoya.horiguchi@linux.dev; alex.sierra@amd.com
> > > Subject: Re: [PATCH v4 1/2] mm: migration: fix the FOLL_GET failure on
> following huge page
> > >
> > > On Monday, 15 August 2022 2:40:48 PM AEST Wang, Haiyue wrote:
> > > > > -----Original Message-----
> > > > > From: Alistair Popple <apopple@nvidia.com>
> > > > > Sent: Monday, August 15, 2022 12:29
> > > > > To: linux-mm@kvack.org; linux-kernel@vger.kernel.org; Wang, Haiyue
> > > <haiyue.wang@intel.com>
> > > > > Cc: akpm@linux-foundation.org; david@redhat.com; linmiaohe@huawei.com;
> > > Huang, Ying
> > > > > <ying.huang@intel.com>; songmuchun@bytedance.com;
> > > naoya.horiguchi@linux.dev; alex.sierra@amd.com; Wang,
> > > > > Haiyue <haiyue.wang@intel.com>
> > > > > Subject: Re: [PATCH v4 1/2] mm: migration: fix the FOLL_GET failure on
> > > following huge page
> > > > >
> > > > > On Monday, 15 August 2022 11:59:08 AM AEST Haiyue Wang wrote:
> > > > > > Not all huge page APIs support FOLL_GET option, so the
> __NR_move_pages
> > > > > > will fail to get the page node information for huge page.
> > > > >
> > > > > I think you should be explicit in the commit message about which
> functions
> > > do
> > > > > not support FOLL_GET as it's not obvious what support needs to be
> added
> > > before
> > > > > this fix can be reverted.
> > > >
> > > > Yes, make sense, will add them in new patch.
> > >
> > > Actually while you're at it I think it would be good to include a
> description
> > > of the impact of this failure in the commit message. Ie. You're answer to:
> > >
> > > > What are the user-visible runtime effects of this bug?
> > >
> > > As it documents what should be tested if this fix does actually ever get
> > > reverted.
> >
> > An short example *.c code to capture the bug in commit message ?
> 
> That's probably overkill. Just being a bit more explicit about the
> circumstances in which sys_move_pages() actually fails would be good. Eg.
> something like this:
> 
> "Without this sys_move_pages() will return -ENOENT for 1GB huge page
> memory map when dumping the page node information for nodes != NULL"

OK, understood, will try to add more information about the error. ;-)

> 
> > >
> > > > >
> > > > > Thanks.
> > > > >
> > > > >  - Alistair
> > > > >
> > > > > > This is an temporary solution to mitigate the racing fix.
> > > > > >
> > > > > > After supporting follow huge page by FOLL_GET is done, this fix can
> be
> > > > > > reverted safely.
> > > > > >
> > > > > > Fixes: 4cd614841c06 ("mm: migration: fix possible
> do_pages_stat_array
> > > racing
> > > > > with memory offline")
> > > > > > Signed-off-by: Haiyue Wang <haiyue.wang@intel.com>
> > > > > > ---
> > > > > >  mm/migrate.c | 10 ++++++++--
> > > > > >  1 file changed, 8 insertions(+), 2 deletions(-)
> > > > > >
> > > > > > diff --git a/mm/migrate.c b/mm/migrate.c
> > > > > > index 6a1597c92261..581dfaad9257 100644
> > > > > > --- a/mm/migrate.c
> > > > > > +++ b/mm/migrate.c
> > > > > > @@ -1848,6 +1848,7 @@ static void do_pages_stat_array(struct
> mm_struct
> > > *mm,
> > > > > unsigned long nr_pages,
> > > > > >
> > > > > >  	for (i = 0; i < nr_pages; i++) {
> > > > > >  		unsigned long addr = (unsigned long)(*pages);
> > > > > > +		unsigned int foll_flags = FOLL_DUMP;
> > > > > >  		struct vm_area_struct *vma;
> > > > > >  		struct page *page;
> > > > > >  		int err = -EFAULT;
> > > > > > @@ -1856,8 +1857,12 @@ static void do_pages_stat_array(struct
> mm_struct
> > > *mm,
> > > > > unsigned long nr_pages,
> > > > > >  		if (!vma)
> > > > > >  			goto set_status;
> > > > > >
> > > > > > +		/* Not all huge page follow APIs support 'FOLL_GET'
> */
> > > > > > +		if (!is_vm_hugetlb_page(vma))
> > > > > > +			foll_flags |= FOLL_GET;
> > > > > > +
> > > > > >  		/* FOLL_DUMP to ignore special (like zero) pages */
> > > > > > -		page = follow_page(vma, addr, FOLL_GET | FOLL_DUMP);
> > > > > > +		page = follow_page(vma, addr, foll_flags);
> > > > > >
> > > > > >  		err = PTR_ERR(page);
> > > > > >  		if (IS_ERR(page))
> > > > > > @@ -1865,7 +1870,8 @@ static void do_pages_stat_array(struct
> mm_struct
> > > *mm,
> > > > > unsigned long nr_pages,
> > > > > >
> > > > > >  		if (page && !is_zone_device_page(page)) {
> > > > > >  			err = page_to_nid(page);
> > > > > > -			put_page(page);
> > > > > > +			if (foll_flags & FOLL_GET)
> > > > > > +				put_page(page);
> > > > > >  		} else {
> > > > > >  			err = -ENOENT;
> > > > > >  		}
> > > > > >
> > > > >
> > > > >
> > > > >
> > > >
> > > >
> > >
> > >
> > >
> >
> >
> 
> 
>
diff mbox series

Patch

diff --git a/mm/migrate.c b/mm/migrate.c
index 6a1597c92261..581dfaad9257 100644
--- a/mm/migrate.c
+++ b/mm/migrate.c
@@ -1848,6 +1848,7 @@  static void do_pages_stat_array(struct mm_struct *mm, unsigned long nr_pages,
 
 	for (i = 0; i < nr_pages; i++) {
 		unsigned long addr = (unsigned long)(*pages);
+		unsigned int foll_flags = FOLL_DUMP;
 		struct vm_area_struct *vma;
 		struct page *page;
 		int err = -EFAULT;
@@ -1856,8 +1857,12 @@  static void do_pages_stat_array(struct mm_struct *mm, unsigned long nr_pages,
 		if (!vma)
 			goto set_status;
 
+		/* Not all huge page follow APIs support 'FOLL_GET' */
+		if (!is_vm_hugetlb_page(vma))
+			foll_flags |= FOLL_GET;
+
 		/* FOLL_DUMP to ignore special (like zero) pages */
-		page = follow_page(vma, addr, FOLL_GET | FOLL_DUMP);
+		page = follow_page(vma, addr, foll_flags);
 
 		err = PTR_ERR(page);
 		if (IS_ERR(page))
@@ -1865,7 +1870,8 @@  static void do_pages_stat_array(struct mm_struct *mm, unsigned long nr_pages,
 
 		if (page && !is_zone_device_page(page)) {
 			err = page_to_nid(page);
-			put_page(page);
+			if (foll_flags & FOLL_GET)
+				put_page(page);
 		} else {
 			err = -ENOENT;
 		}