diff mbox series

[v2] mm/mempolicy,hugetlb: Checking hstate for hugetlbfs page in vma_migratable

Message ID 1579070637-26945-1-git-send-email-lixinhai.lxh@gmail.com (mailing list archive)
State New, archived
Headers show
Series [v2] mm/mempolicy,hugetlb: Checking hstate for hugetlbfs page in vma_migratable | expand

Commit Message

Li Xinhai Jan. 15, 2020, 6:43 a.m. UTC
Checking hstate at early phase when isolating page, instead of during
unmap and move phase, to avoid useless isolation.

Signed-off-by: Li Xinhai <lixinhai.lxh@gmail.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Mike Kravetz <mike.kravetz@oracle.com>
---
v1->v2:
New function
bool vm_hugepage_migration_supported(struct vm_area_struct *vma)
is introduced to simplify the inter dependency of 
include/linux/mempolicy.h and include/linux/hugetlb.h, and could be
useful for other caller.

 include/linux/hugetlb.h   |  2 ++
 include/linux/mempolicy.h |  6 +++---
 mm/hugetlb.c              | 10 ++++++++++
 3 files changed, 15 insertions(+), 3 deletions(-)

Comments

Li Xinhai Jan. 15, 2020, 9:04 a.m. UTC | #1
On 2020-01-15 at 14:43 Li Xinhai wrote:
>Checking hstate at early phase when isolating page, instead of during
>unmap and move phase, to avoid useless isolation.
>
>Signed-off-by: Li Xinhai <lixinhai.lxh@gmail.com>
>Cc: Michal Hocko <mhocko@suse.com>
>Cc: Mike Kravetz <mike.kravetz@oracle.com> 

It is bad to declare the function in two files, would it be better to declare
the interface in hugetlb_inline.h as this file is created in purpose for break
circular reference between hugetlb and other components.

>---
>v1->v2:
>New function
>bool vm_hugepage_migration_supported(struct vm_area_struct *vma)
>is introduced to simplify the inter dependency of
>include/linux/mempolicy.h and include/linux/hugetlb.h, and could be
>useful for other caller.
>
> include/linux/hugetlb.h   |  2 ++
> include/linux/mempolicy.h |  6 +++---
> mm/hugetlb.c              | 10 ++++++++++
> 3 files changed, 15 insertions(+), 3 deletions(-)
>
>diff --git a/include/linux/hugetlb.h b/include/linux/hugetlb.h
>index 31d4920..52fc034 100644
>--- a/include/linux/hugetlb.h
>+++ b/include/linux/hugetlb.h
>@@ -834,6 +834,8 @@ static inline void set_huge_swap_pte_at(struct mm_struct *mm, unsigned long addr
> }
> #endif	/* CONFIG_HUGETLB_PAGE */
>
>+extern bool vm_hugepage_migration_supported(struct vm_area_struct *vma);
>+
> static inline spinlock_t *huge_pte_lock(struct hstate *h,
> struct mm_struct *mm, pte_t *pte)
> {
>diff --git a/include/linux/mempolicy.h b/include/linux/mempolicy.h
>index 5228c62..6637166 100644
>--- a/include/linux/mempolicy.h
>+++ b/include/linux/mempolicy.h
>@@ -172,6 +172,7 @@ int do_migrate_pages(struct mm_struct *mm, const nodemask_t *from,
>
> extern void mpol_to_str(char *buffer, int maxlen, struct mempolicy *pol);
>
>+extern bool vm_hugepage_migration_supported(struct vm_area_struct *vma);
> /* Check if a vma is migratable */
> static inline bool vma_migratable(struct vm_area_struct *vma)
> {
>@@ -185,10 +186,9 @@ static inline bool vma_migratable(struct vm_area_struct *vma)
> if (vma_is_dax(vma))
> return false;
>
>-#ifndef CONFIG_ARCH_ENABLE_HUGEPAGE_MIGRATION
>-	if (vma->vm_flags & VM_HUGETLB)
>+	if (is_vm_hugetlb_page(vma) &&
>+	!vm_hugepage_migration_supported(vma))
> return false;
>-#endif
>
> /*
> * Migration allocates pages in the highest zone. If we cannot
>diff --git a/mm/hugetlb.c b/mm/hugetlb.c
>index dd8737a..fce149c 100644
>--- a/mm/hugetlb.c
>+++ b/mm/hugetlb.c
>@@ -1316,6 +1316,16 @@ int PageHeadHuge(struct page *page_head)
> return get_compound_page_dtor(page_head) == free_huge_page;
> }
>
>+bool vm_hugepage_migration_supported(struct vm_area_struct *vma)
>+{
>+#ifdef CONFIG_HUGETLB_PAGE
>+	VM_BUG_ON(!is_vm_hugetlb_page(vma));
>+	if (hugepage_migration_supported(hstate_vma(vma)))
>+	return true;
>+#endif
>+	return false;
>+}
>+
> pgoff_t __basepage_index(struct page *page)
> {
> struct page *page_head = compound_head(page);
>--
>1.8.3.1
>
diff mbox series

Patch

diff --git a/include/linux/hugetlb.h b/include/linux/hugetlb.h
index 31d4920..52fc034 100644
--- a/include/linux/hugetlb.h
+++ b/include/linux/hugetlb.h
@@ -834,6 +834,8 @@  static inline void set_huge_swap_pte_at(struct mm_struct *mm, unsigned long addr
 }
 #endif	/* CONFIG_HUGETLB_PAGE */
 
+extern bool vm_hugepage_migration_supported(struct vm_area_struct *vma);
+
 static inline spinlock_t *huge_pte_lock(struct hstate *h,
 					struct mm_struct *mm, pte_t *pte)
 {
diff --git a/include/linux/mempolicy.h b/include/linux/mempolicy.h
index 5228c62..6637166 100644
--- a/include/linux/mempolicy.h
+++ b/include/linux/mempolicy.h
@@ -172,6 +172,7 @@  int do_migrate_pages(struct mm_struct *mm, const nodemask_t *from,
 
 extern void mpol_to_str(char *buffer, int maxlen, struct mempolicy *pol);
 
+extern bool vm_hugepage_migration_supported(struct vm_area_struct *vma);
 /* Check if a vma is migratable */
 static inline bool vma_migratable(struct vm_area_struct *vma)
 {
@@ -185,10 +186,9 @@  static inline bool vma_migratable(struct vm_area_struct *vma)
 	if (vma_is_dax(vma))
 		return false;
 
-#ifndef CONFIG_ARCH_ENABLE_HUGEPAGE_MIGRATION
-	if (vma->vm_flags & VM_HUGETLB)
+	if (is_vm_hugetlb_page(vma) &&
+		!vm_hugepage_migration_supported(vma))
 		return false;
-#endif
 
 	/*
 	 * Migration allocates pages in the highest zone. If we cannot
diff --git a/mm/hugetlb.c b/mm/hugetlb.c
index dd8737a..fce149c 100644
--- a/mm/hugetlb.c
+++ b/mm/hugetlb.c
@@ -1316,6 +1316,16 @@  int PageHeadHuge(struct page *page_head)
 	return get_compound_page_dtor(page_head) == free_huge_page;
 }
 
+bool vm_hugepage_migration_supported(struct vm_area_struct *vma)
+{
+#ifdef CONFIG_HUGETLB_PAGE
+	VM_BUG_ON(!is_vm_hugetlb_page(vma));
+	if (hugepage_migration_supported(hstate_vma(vma)))
+		return true;
+#endif
+	return false;
+}
+
 pgoff_t __basepage_index(struct page *page)
 {
 	struct page *page_head = compound_head(page);