diff mbox series

[v3] mm/debug: use valid physical memory for pmd/pud tests

Message ID 20230110181208.1633879-1-fvdl@google.com (mailing list archive)
State New
Headers show
Series [v3] mm/debug: use valid physical memory for pmd/pud tests | expand

Commit Message

Frank van der Linden Jan. 10, 2023, 6:12 p.m. UTC
The page table debug tests need a physical address to validate
low-level page table manipulation with. The memory at this address
is not actually touched, it just encoded in the page table entries
at various levels during the tests only.

Since the memory is not used, the code just picks the physical
address of the start_kernel symbol. This value is then truncated
to get a properly aligned address that is to be used for various
tests. Because of the truncation, the address might not actually
exist, or might not describe a complete huge page. That's not a
problem for most tests, but the arch-specific code may check
for attribute validity and consistency. The x86 version of
{pud,pmd}_set_huge actually validates the MTRRs for the PMD/PUD
range. This may fail with an address derived from start_kernel,
depending on where the kernel was loaded and what the physical
memory layout of the system is. This then leads to false negatives
for the {pud,pmd}_set_huge tests.

Avoid this by finding a properly aligned memory range that exists
and is usable. If such a range is not found, skip the tests that
needed it.

Fixes: 399145f9eb6c ("mm/debug: add tests validating architecture page table helpers")
Cc: Anshuman Khandual <anshuman.khandual@arm.com>
Signed-off-by: Frank van der Linden <fvdl@google.com>
---
 mm/debug_vm_pgtable.c | 102 ++++++++++++++++++++++++++++++++++--------
 1 file changed, 83 insertions(+), 19 deletions(-)

Comments

Anshuman Khandual Jan. 11, 2023, 4:38 a.m. UTC | #1
On 1/10/23 23:42, Frank van der Linden wrote:
> The page table debug tests need a physical address to validate
> low-level page table manipulation with. The memory at this address
> is not actually touched, it just encoded in the page table entries
> at various levels during the tests only.
> 
> Since the memory is not used, the code just picks the physical
> address of the start_kernel symbol. This value is then truncated
> to get a properly aligned address that is to be used for various
> tests. Because of the truncation, the address might not actually
> exist, or might not describe a complete huge page. That's not a
> problem for most tests, but the arch-specific code may check
> for attribute validity and consistency. The x86 version of
> {pud,pmd}_set_huge actually validates the MTRRs for the PMD/PUD
> range. This may fail with an address derived from start_kernel,
> depending on where the kernel was loaded and what the physical
> memory layout of the system is. This then leads to false negatives
> for the {pud,pmd}_set_huge tests.
> 
> Avoid this by finding a properly aligned memory range that exists
> and is usable. If such a range is not found, skip the tests that
> needed it.
> 
> Fixes: 399145f9eb6c ("mm/debug: add tests validating architecture page table helpers")
> Cc: Anshuman Khandual <anshuman.khandual@arm.com>

Reviewed-by: Anshuman Khandual <anshuman.khandual@arm.com>

> Signed-off-by: Frank van der Linden <fvdl@google.com>
> ---
>  mm/debug_vm_pgtable.c | 102 ++++++++++++++++++++++++++++++++++--------
>  1 file changed, 83 insertions(+), 19 deletions(-)
> 
> diff --git a/mm/debug_vm_pgtable.c b/mm/debug_vm_pgtable.c
> index c631ade3f1d2..bb3328f46126 100644
> --- a/mm/debug_vm_pgtable.c
> +++ b/mm/debug_vm_pgtable.c
> @@ -15,6 +15,7 @@
>  #include <linux/hugetlb.h>
>  #include <linux/kernel.h>
>  #include <linux/kconfig.h>
> +#include <linux/memblock.h>
>  #include <linux/mm.h>
>  #include <linux/mman.h>
>  #include <linux/mm_types.h>
> @@ -80,6 +81,7 @@ struct pgtable_debug_args {
>  	unsigned long		pmd_pfn;
>  	unsigned long		pte_pfn;
>  
> +	unsigned long		fixed_alignment;
>  	unsigned long		fixed_pgd_pfn;
>  	unsigned long		fixed_p4d_pfn;
>  	unsigned long		fixed_pud_pfn;
> @@ -430,7 +432,8 @@ static void __init pmd_huge_tests(struct pgtable_debug_args *args)
>  {
>  	pmd_t pmd;
>  
> -	if (!arch_vmap_pmd_supported(args->page_prot))
> +	if (!arch_vmap_pmd_supported(args->page_prot) ||
> +	    args->fixed_alignment < PMD_SIZE)
>  		return;
>  
>  	pr_debug("Validating PMD huge\n");
> @@ -449,7 +452,8 @@ static void __init pud_huge_tests(struct pgtable_debug_args *args)
>  {
>  	pud_t pud;
>  
> -	if (!arch_vmap_pud_supported(args->page_prot))
> +	if (!arch_vmap_pud_supported(args->page_prot) ||
> +	    args->fixed_alignment < PUD_SIZE)
>  		return;
>  
>  	pr_debug("Validating PUD huge\n");
> @@ -1077,10 +1081,85 @@ debug_vm_pgtable_alloc_huge_page(struct pgtable_debug_args *args, int order)
>  	return page;
>  }
>  
> +/*
> + * Check if a physical memory range described by <pstart, pend> contains
> + * an area that is of size psize, and aligned to psize.
> + *
> + * Don't use address 0, an all-zeroes physical address might mask bugs, and
> + * it's not used on x86.
> + */
> +static void  __init phys_align_check(phys_addr_t pstart,
> +				     phys_addr_t pend, unsigned long psize,
> +				     phys_addr_t *physp, unsigned long *alignp)
> +{
> +	phys_addr_t aligned_start, aligned_end;
> +
> +	if (pstart == 0)
> +		pstart = PAGE_SIZE;
> +
> +	aligned_start = ALIGN(pstart, psize);
> +	aligned_end = aligned_start + psize;
> +
> +	if (aligned_end > aligned_start && aligned_end <= pend) {
> +		*alignp = psize;
> +		*physp = aligned_start;
> +	}
> +}
> +
> +static void __init init_fixed_pfns(struct pgtable_debug_args *args)
> +{
> +	u64 idx;
> +	phys_addr_t phys, pstart, pend;
> +
> +	/*
> +	 * Initialize the fixed pfns. To do this, try to find a
> +	 * valid physical range, preferably aligned to PUD_SIZE,
> +	 * but settling for aligned to PMD_SIZE as a fallback. If
> +	 * neither of those is found, use the physical address of
> +	 * the start_kernel symbol.
> +	 *
> +	 * The memory doesn't need to be allocated, it just needs to exist
> +	 * as usable memory. It won't be touched.
> +	 *
> +	 * The alignment is recorded, and can be checked to see if we
> +	 * can run the tests that require an actual valid physical
> +	 * address range on some architectures ({pmd,pud}_huge_test
> +	 * on x86).
> +	 */
> +
> +	phys = __pa_symbol(&start_kernel);
> +	args->fixed_alignment = PAGE_SIZE;
> +
> +	for_each_mem_range(idx, &pstart, &pend) {
> +		/* First check for a PUD-aligned area */
> +		phys_align_check(pstart, pend, PUD_SIZE, &phys,
> +				 &args->fixed_alignment);
> +
> +		/* If a PUD-aligned area is found, we're done */
> +		if (args->fixed_alignment == PUD_SIZE)
> +			break;
> +
> +		/*
> +		 * If no PMD-aligned area found yet, check for one,
> +		 * but continue the loop to look for a PUD-aligned area.
> +		 */
> +		if (args->fixed_alignment < PMD_SIZE)
> +			phys_align_check(pstart, pend, PMD_SIZE, &phys,
> +					 &args->fixed_alignment);
> +	}
> +
> +	args->fixed_pgd_pfn = __phys_to_pfn(phys & PGDIR_MASK);
> +	args->fixed_p4d_pfn = __phys_to_pfn(phys & P4D_MASK);
> +	args->fixed_pud_pfn = __phys_to_pfn(phys & PUD_MASK);
> +	args->fixed_pmd_pfn = __phys_to_pfn(phys & PMD_MASK);
> +	args->fixed_pte_pfn = __phys_to_pfn(phys & PAGE_MASK);
> +	WARN_ON(!pfn_valid(args->fixed_pte_pfn));
> +}
> +
> +
>  static int __init init_args(struct pgtable_debug_args *args)
>  {
>  	struct page *page = NULL;
> -	phys_addr_t phys;
>  	int ret = 0;
>  
>  	/*
> @@ -1160,22 +1239,7 @@ static int __init init_args(struct pgtable_debug_args *args)
>  	args->start_ptep = pmd_pgtable(READ_ONCE(*args->pmdp));
>  	WARN_ON(!args->start_ptep);
>  
> -	/*
> -	 * PFN for mapping at PTE level is determined from a standard kernel
> -	 * text symbol. But pfns for higher page table levels are derived by
> -	 * masking lower bits of this real pfn. These derived pfns might not
> -	 * exist on the platform but that does not really matter as pfn_pxx()
> -	 * helpers will still create appropriate entries for the test. This
> -	 * helps avoid large memory block allocations to be used for mapping
> -	 * at higher page table levels in some of the tests.
> -	 */
> -	phys = __pa_symbol(&start_kernel);
> -	args->fixed_pgd_pfn = __phys_to_pfn(phys & PGDIR_MASK);
> -	args->fixed_p4d_pfn = __phys_to_pfn(phys & P4D_MASK);
> -	args->fixed_pud_pfn = __phys_to_pfn(phys & PUD_MASK);
> -	args->fixed_pmd_pfn = __phys_to_pfn(phys & PMD_MASK);
> -	args->fixed_pte_pfn = __phys_to_pfn(phys & PAGE_MASK);
> -	WARN_ON(!pfn_valid(args->fixed_pte_pfn));
> +	init_fixed_pfns(args);
>  
>  	/*
>  	 * Allocate (huge) pages because some of the tests need to access
diff mbox series

Patch

diff --git a/mm/debug_vm_pgtable.c b/mm/debug_vm_pgtable.c
index c631ade3f1d2..bb3328f46126 100644
--- a/mm/debug_vm_pgtable.c
+++ b/mm/debug_vm_pgtable.c
@@ -15,6 +15,7 @@ 
 #include <linux/hugetlb.h>
 #include <linux/kernel.h>
 #include <linux/kconfig.h>
+#include <linux/memblock.h>
 #include <linux/mm.h>
 #include <linux/mman.h>
 #include <linux/mm_types.h>
@@ -80,6 +81,7 @@  struct pgtable_debug_args {
 	unsigned long		pmd_pfn;
 	unsigned long		pte_pfn;
 
+	unsigned long		fixed_alignment;
 	unsigned long		fixed_pgd_pfn;
 	unsigned long		fixed_p4d_pfn;
 	unsigned long		fixed_pud_pfn;
@@ -430,7 +432,8 @@  static void __init pmd_huge_tests(struct pgtable_debug_args *args)
 {
 	pmd_t pmd;
 
-	if (!arch_vmap_pmd_supported(args->page_prot))
+	if (!arch_vmap_pmd_supported(args->page_prot) ||
+	    args->fixed_alignment < PMD_SIZE)
 		return;
 
 	pr_debug("Validating PMD huge\n");
@@ -449,7 +452,8 @@  static void __init pud_huge_tests(struct pgtable_debug_args *args)
 {
 	pud_t pud;
 
-	if (!arch_vmap_pud_supported(args->page_prot))
+	if (!arch_vmap_pud_supported(args->page_prot) ||
+	    args->fixed_alignment < PUD_SIZE)
 		return;
 
 	pr_debug("Validating PUD huge\n");
@@ -1077,10 +1081,85 @@  debug_vm_pgtable_alloc_huge_page(struct pgtable_debug_args *args, int order)
 	return page;
 }
 
+/*
+ * Check if a physical memory range described by <pstart, pend> contains
+ * an area that is of size psize, and aligned to psize.
+ *
+ * Don't use address 0, an all-zeroes physical address might mask bugs, and
+ * it's not used on x86.
+ */
+static void  __init phys_align_check(phys_addr_t pstart,
+				     phys_addr_t pend, unsigned long psize,
+				     phys_addr_t *physp, unsigned long *alignp)
+{
+	phys_addr_t aligned_start, aligned_end;
+
+	if (pstart == 0)
+		pstart = PAGE_SIZE;
+
+	aligned_start = ALIGN(pstart, psize);
+	aligned_end = aligned_start + psize;
+
+	if (aligned_end > aligned_start && aligned_end <= pend) {
+		*alignp = psize;
+		*physp = aligned_start;
+	}
+}
+
+static void __init init_fixed_pfns(struct pgtable_debug_args *args)
+{
+	u64 idx;
+	phys_addr_t phys, pstart, pend;
+
+	/*
+	 * Initialize the fixed pfns. To do this, try to find a
+	 * valid physical range, preferably aligned to PUD_SIZE,
+	 * but settling for aligned to PMD_SIZE as a fallback. If
+	 * neither of those is found, use the physical address of
+	 * the start_kernel symbol.
+	 *
+	 * The memory doesn't need to be allocated, it just needs to exist
+	 * as usable memory. It won't be touched.
+	 *
+	 * The alignment is recorded, and can be checked to see if we
+	 * can run the tests that require an actual valid physical
+	 * address range on some architectures ({pmd,pud}_huge_test
+	 * on x86).
+	 */
+
+	phys = __pa_symbol(&start_kernel);
+	args->fixed_alignment = PAGE_SIZE;
+
+	for_each_mem_range(idx, &pstart, &pend) {
+		/* First check for a PUD-aligned area */
+		phys_align_check(pstart, pend, PUD_SIZE, &phys,
+				 &args->fixed_alignment);
+
+		/* If a PUD-aligned area is found, we're done */
+		if (args->fixed_alignment == PUD_SIZE)
+			break;
+
+		/*
+		 * If no PMD-aligned area found yet, check for one,
+		 * but continue the loop to look for a PUD-aligned area.
+		 */
+		if (args->fixed_alignment < PMD_SIZE)
+			phys_align_check(pstart, pend, PMD_SIZE, &phys,
+					 &args->fixed_alignment);
+	}
+
+	args->fixed_pgd_pfn = __phys_to_pfn(phys & PGDIR_MASK);
+	args->fixed_p4d_pfn = __phys_to_pfn(phys & P4D_MASK);
+	args->fixed_pud_pfn = __phys_to_pfn(phys & PUD_MASK);
+	args->fixed_pmd_pfn = __phys_to_pfn(phys & PMD_MASK);
+	args->fixed_pte_pfn = __phys_to_pfn(phys & PAGE_MASK);
+	WARN_ON(!pfn_valid(args->fixed_pte_pfn));
+}
+
+
 static int __init init_args(struct pgtable_debug_args *args)
 {
 	struct page *page = NULL;
-	phys_addr_t phys;
 	int ret = 0;
 
 	/*
@@ -1160,22 +1239,7 @@  static int __init init_args(struct pgtable_debug_args *args)
 	args->start_ptep = pmd_pgtable(READ_ONCE(*args->pmdp));
 	WARN_ON(!args->start_ptep);
 
-	/*
-	 * PFN for mapping at PTE level is determined from a standard kernel
-	 * text symbol. But pfns for higher page table levels are derived by
-	 * masking lower bits of this real pfn. These derived pfns might not
-	 * exist on the platform but that does not really matter as pfn_pxx()
-	 * helpers will still create appropriate entries for the test. This
-	 * helps avoid large memory block allocations to be used for mapping
-	 * at higher page table levels in some of the tests.
-	 */
-	phys = __pa_symbol(&start_kernel);
-	args->fixed_pgd_pfn = __phys_to_pfn(phys & PGDIR_MASK);
-	args->fixed_p4d_pfn = __phys_to_pfn(phys & P4D_MASK);
-	args->fixed_pud_pfn = __phys_to_pfn(phys & PUD_MASK);
-	args->fixed_pmd_pfn = __phys_to_pfn(phys & PMD_MASK);
-	args->fixed_pte_pfn = __phys_to_pfn(phys & PAGE_MASK);
-	WARN_ON(!pfn_valid(args->fixed_pte_pfn));
+	init_fixed_pfns(args);
 
 	/*
 	 * Allocate (huge) pages because some of the tests need to access