@@ -258,11 +258,14 @@ int huge_ptep_set_access_flags(struct vm_area_struct *vma,
#ifdef CONFIG_PPC_BOOK3S_64
struct hstate *h = hstate_vma(vma);
+ struct hugetlb_pte hpte;
psize = hstate_get_psize(h);
#ifdef CONFIG_DEBUG_VM
- assert_spin_locked(huge_pte_lockptr(huge_page_shift(h),
- vma->vm_mm, ptep));
+ /* HGM is not supported for powerpc yet. */
+ hugetlb_pte_init(&hpte, ptep, huge_page_shift(h),
+ hpage_size_to_level(psize));
+ assert_spin_locked(hpte.ptl);
#endif
#else
@@ -1012,14 +1012,6 @@ static inline gfp_t htlb_modify_alloc_mask(struct hstate *h, gfp_t gfp_mask)
return modified_mask;
}
-static inline spinlock_t *huge_pte_lockptr(unsigned int shift,
- struct mm_struct *mm, pte_t *pte)
-{
- if (shift == PMD_SHIFT)
- return pmd_lockptr(mm, (pmd_t *) pte);
- return &mm->page_table_lock;
-}
-
#ifndef hugepages_supported
/*
* Some platform decide whether they support huge pages at boot
@@ -1228,12 +1220,6 @@ static inline gfp_t htlb_modify_alloc_mask(struct hstate *h, gfp_t gfp_mask)
return 0;
}
-static inline spinlock_t *huge_pte_lockptr(unsigned int shift,
- struct mm_struct *mm, pte_t *pte)
-{
- return &mm->page_table_lock;
-}
-
static inline void hugetlb_count_init(struct mm_struct *mm)
{
}
@@ -1308,16 +1294,6 @@ int hugetlb_collapse(struct mm_struct *mm, unsigned long start,
}
#endif
-static inline spinlock_t *huge_pte_lock(struct hstate *h,
- struct mm_struct *mm, pte_t *pte)
-{
- spinlock_t *ptl;
-
- ptl = huge_pte_lockptr(huge_page_shift(h), mm, pte);
- spin_lock(ptl);
- return ptl;
-}
-
static inline
spinlock_t *hugetlb_pte_lockptr(struct hugetlb_pte *hpte)
{
@@ -1353,8 +1329,22 @@ void hugetlb_pte_init(struct mm_struct *mm, struct hugetlb_pte *hpte,
pte_t *ptep, unsigned int shift,
enum hugetlb_level level)
{
- __hugetlb_pte_init(hpte, ptep, shift, level,
- huge_pte_lockptr(shift, mm, ptep));
+ spinlock_t *ptl;
+
+ /*
+ * For contiguous HugeTLB PTEs that can contain other HugeTLB PTEs
+ * on the same level, the same PTL for both must be used.
+ *
+ * For some architectures that implement hugetlb_walk_step, this
+ * version of hugetlb_pte_populate() may not be correct to use for
+ * high-granularity PTEs. Instead, call __hugetlb_pte_populate()
+ * directly.
+ */
+ if (level == HUGETLB_LEVEL_PMD)
+ ptl = pmd_lockptr(mm, (pmd_t *) ptep);
+ else
+ ptl = &mm->page_table_lock;
+ __hugetlb_pte_init(hpte, ptep, shift, level, ptl);
}
#if defined(CONFIG_HUGETLB_PAGE) && defined(CONFIG_CMA)
@@ -5454,9 +5454,8 @@ int copy_hugetlb_page_range(struct mm_struct *dst, struct mm_struct *src,
put_page(hpage);
/* Install the new hugetlb folio if src pte stable */
- dst_ptl = huge_pte_lock(h, dst, dst_pte);
- src_ptl = huge_pte_lockptr(huge_page_shift(h),
- src, src_pte);
+ dst_ptl = hugetlb_pte_lock(&dst_hpte);
+ src_ptl = hugetlb_pte_lockptr(&src_hpte);
spin_lock_nested(src_ptl, SINGLE_DEPTH_NESTING);
entry = huge_ptep_get(src_pte);
if (!pte_same(src_pte_old, entry)) {
@@ -7582,7 +7581,8 @@ pte_t *huge_pmd_share(struct mm_struct *mm, struct vm_area_struct *vma,
unsigned long saddr;
pte_t *spte = NULL;
pte_t *pte;
- spinlock_t *ptl;
+ struct hugetlb_pte hpte;
+ struct hstate *shstate;
i_mmap_lock_read(mapping);
vma_interval_tree_foreach(svma, &mapping->i_mmap, idx, idx) {
@@ -7603,7 +7603,11 @@ pte_t *huge_pmd_share(struct mm_struct *mm, struct vm_area_struct *vma,
if (!spte)
goto out;
- ptl = huge_pte_lock(hstate_vma(vma), mm, spte);
+ shstate = hstate_vma(svma);
+
+ hugetlb_pte_init(mm, &hpte, spte, huge_page_shift(shstate),
+ hpage_size_to_level(huge_page_size(shstate)));
+ spin_lock(hpte.ptl);
if (pud_none(*pud)) {
pud_populate(mm, pud,
(pmd_t *)((unsigned long)spte & PAGE_MASK));
@@ -7611,7 +7615,7 @@ pte_t *huge_pmd_share(struct mm_struct *mm, struct vm_area_struct *vma,
} else {
put_page(virt_to_page(spte));
}
- spin_unlock(ptl);
+ spin_unlock(hpte.ptl);
out:
pte = (pte_t *)pmd_alloc(mm, pud, addr);
i_mmap_unlock_read(mapping);
@@ -8315,6 +8319,7 @@ static void hugetlb_unshare_pmds(struct vm_area_struct *vma,
unsigned long address;
spinlock_t *ptl;
pte_t *ptep;
+ struct hugetlb_pte hpte;
if (!(vma->vm_flags & VM_MAYSHARE))
return;
@@ -8336,7 +8341,10 @@ static void hugetlb_unshare_pmds(struct vm_area_struct *vma,
ptep = hugetlb_walk(vma, address, sz);
if (!ptep)
continue;
- ptl = huge_pte_lock(h, mm, ptep);
+
+ hugetlb_pte_init(mm, &hpte, ptep, huge_page_shift(h),
+ hpage_size_to_level(sz));
+ ptl = hugetlb_pte_lock(&hpte);
huge_pmd_unshare(mm, vma, address, ptep);
spin_unlock(ptl);
}
They are replaced with hugetlb_pte_lock{,ptr}. All callers that haven't already been replaced don't get called when using HGM, so we handle them by populating hugetlb_ptes with the standard, hstate-sized huge PTEs. Signed-off-by: James Houghton <jthoughton@google.com>