Message ID | 20200916032523.13011-3-gshan@redhat.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | arm64/mm: Enable color zero pages | expand |
On Wed, Sep 16, 2020 at 01:25:23PM +1000, Gavin Shan wrote: > This enables color zero pages by allocating contigous page frames > for it. The number of pages for this is determined by L1 dCache > (or iCache) size, which is probbed from the hardware. > > * Add cache_total_size() to return L1 dCache (or iCache) size > > * Implement setup_zero_pages(), which is called after the page > allocator begins to work, to allocate the contigous pages > needed by color zero page. > > * Reworked ZERO_PAGE() and define __HAVE_COLOR_ZERO_PAGE. > > Signed-off-by: Gavin Shan <gshan@redhat.com> > --- > arch/arm64/include/asm/cache.h | 22 ++++++++++++++++++++ > arch/arm64/include/asm/pgtable.h | 9 ++++++-- > arch/arm64/kernel/cacheinfo.c | 34 +++++++++++++++++++++++++++++++ > arch/arm64/mm/init.c | 35 ++++++++++++++++++++++++++++++++ > arch/arm64/mm/mmu.c | 7 ------- > 5 files changed, 98 insertions(+), 9 deletions(-) > > diff --git a/arch/arm64/include/asm/cache.h b/arch/arm64/include/asm/cache.h > index a4d1b5f771f6..420e9dde2c51 100644 > --- a/arch/arm64/include/asm/cache.h > +++ b/arch/arm64/include/asm/cache.h > @@ -39,6 +39,27 @@ > #define CLIDR_LOC(clidr) (((clidr) >> CLIDR_LOC_SHIFT) & 0x7) > #define CLIDR_LOUIS(clidr) (((clidr) >> CLIDR_LOUIS_SHIFT) & 0x7) > > +#define CSSELR_TND_SHIFT 4 > +#define CSSELR_TND_MASK (UL(1) << CSSELR_TND_SHIFT) > +#define CSSELR_LEVEL_SHIFT 1 > +#define CSSELR_LEVEL_MASK (UL(7) << CSSELR_LEVEL_SHIFT) > +#define CSSELR_IND_SHIFT 0 > +#define CSSERL_IND_MASK (UL(1) << CSSELR_IND_SHIFT) > + > +#define CCSIDR_64_LS_SHIFT 0 > +#define CCSIDR_64_LS_MASK (UL(7) << CCSIDR_64_LS_SHIFT) > +#define CCSIDR_64_ASSOC_SHIFT 3 > +#define CCSIDR_64_ASSOC_MASK (UL(0x1FFFFF) << CCSIDR_64_ASSOC_SHIFT) > +#define CCSIDR_64_SET_SHIFT 32 > +#define CCSIDR_64_SET_MASK (UL(0xFFFFFF) << CCSIDR_64_SET_SHIFT) > + > +#define CCSIDR_32_LS_SHIFT 0 > +#define CCSIDR_32_LS_MASK (UL(7) << CCSIDR_32_LS_SHIFT) > +#define CCSIDR_32_ASSOC_SHIFT 3 > +#define CCSIDR_32_ASSOC_MASK (UL(0x3FF) << CCSIDR_32_ASSOC_SHIFT) > +#define CCSIDR_32_SET_SHIFT 13 > +#define CCSIDR_32_SET_MASK (UL(0x7FFF) << CCSIDR_32_SET_SHIFT) I don't think we should be inferring cache structure from these register values. The Arm ARM helpfully says: | You cannot make any inference about the actual sizes of caches based | on these parameters. so we need to take the topology information from elsewhere. But before we get into that, can you justify why we need to do this at all, please? Do you have data to show the benefit of adding this complexity? Cheers, Will
On 2020-09-16 09:28, Will Deacon wrote: > On Wed, Sep 16, 2020 at 01:25:23PM +1000, Gavin Shan wrote: >> This enables color zero pages by allocating contigous page frames >> for it. The number of pages for this is determined by L1 dCache >> (or iCache) size, which is probbed from the hardware. >> >> * Add cache_total_size() to return L1 dCache (or iCache) size >> >> * Implement setup_zero_pages(), which is called after the page >> allocator begins to work, to allocate the contigous pages >> needed by color zero page. >> >> * Reworked ZERO_PAGE() and define __HAVE_COLOR_ZERO_PAGE. >> >> Signed-off-by: Gavin Shan <gshan@redhat.com> >> --- >> arch/arm64/include/asm/cache.h | 22 ++++++++++++++++++++ >> arch/arm64/include/asm/pgtable.h | 9 ++++++-- >> arch/arm64/kernel/cacheinfo.c | 34 +++++++++++++++++++++++++++++++ >> arch/arm64/mm/init.c | 35 ++++++++++++++++++++++++++++++++ >> arch/arm64/mm/mmu.c | 7 ------- >> 5 files changed, 98 insertions(+), 9 deletions(-) >> >> diff --git a/arch/arm64/include/asm/cache.h b/arch/arm64/include/asm/cache.h >> index a4d1b5f771f6..420e9dde2c51 100644 >> --- a/arch/arm64/include/asm/cache.h >> +++ b/arch/arm64/include/asm/cache.h >> @@ -39,6 +39,27 @@ >> #define CLIDR_LOC(clidr) (((clidr) >> CLIDR_LOC_SHIFT) & 0x7) >> #define CLIDR_LOUIS(clidr) (((clidr) >> CLIDR_LOUIS_SHIFT) & 0x7) >> >> +#define CSSELR_TND_SHIFT 4 >> +#define CSSELR_TND_MASK (UL(1) << CSSELR_TND_SHIFT) >> +#define CSSELR_LEVEL_SHIFT 1 >> +#define CSSELR_LEVEL_MASK (UL(7) << CSSELR_LEVEL_SHIFT) >> +#define CSSELR_IND_SHIFT 0 >> +#define CSSERL_IND_MASK (UL(1) << CSSELR_IND_SHIFT) >> + >> +#define CCSIDR_64_LS_SHIFT 0 >> +#define CCSIDR_64_LS_MASK (UL(7) << CCSIDR_64_LS_SHIFT) >> +#define CCSIDR_64_ASSOC_SHIFT 3 >> +#define CCSIDR_64_ASSOC_MASK (UL(0x1FFFFF) << CCSIDR_64_ASSOC_SHIFT) >> +#define CCSIDR_64_SET_SHIFT 32 >> +#define CCSIDR_64_SET_MASK (UL(0xFFFFFF) << CCSIDR_64_SET_SHIFT) >> + >> +#define CCSIDR_32_LS_SHIFT 0 >> +#define CCSIDR_32_LS_MASK (UL(7) << CCSIDR_32_LS_SHIFT) >> +#define CCSIDR_32_ASSOC_SHIFT 3 >> +#define CCSIDR_32_ASSOC_MASK (UL(0x3FF) << CCSIDR_32_ASSOC_SHIFT) >> +#define CCSIDR_32_SET_SHIFT 13 >> +#define CCSIDR_32_SET_MASK (UL(0x7FFF) << CCSIDR_32_SET_SHIFT) > > I don't think we should be inferring cache structure from these register > values. The Arm ARM helpfully says: > > | You cannot make any inference about the actual sizes of caches based > | on these parameters. > > so we need to take the topology information from elsewhere. Yes, these represent parameters for the low-level cache maintenance by set/way instructions, and nothing more. There are definitely cases where they do not reflect the underlying cache structure (commit 793acf870ea3 is an obvious first call). Robin. > But before we get into that, can you justify why we need to do this at all, > please? Do you have data to show the benefit of adding this complexity? > > Cheers, > > Will > > _______________________________________________ > linux-arm-kernel mailing list > linux-arm-kernel@lists.infradead.org > http://lists.infradead.org/mailman/listinfo/linux-arm-kernel >
Hi Will, On 9/16/20 6:28 PM, Will Deacon wrote: > On Wed, Sep 16, 2020 at 01:25:23PM +1000, Gavin Shan wrote: >> This enables color zero pages by allocating contigous page frames >> for it. The number of pages for this is determined by L1 dCache >> (or iCache) size, which is probbed from the hardware. >> >> * Add cache_total_size() to return L1 dCache (or iCache) size >> >> * Implement setup_zero_pages(), which is called after the page >> allocator begins to work, to allocate the contigous pages >> needed by color zero page. >> >> * Reworked ZERO_PAGE() and define __HAVE_COLOR_ZERO_PAGE. >> >> Signed-off-by: Gavin Shan <gshan@redhat.com> >> --- >> arch/arm64/include/asm/cache.h | 22 ++++++++++++++++++++ >> arch/arm64/include/asm/pgtable.h | 9 ++++++-- >> arch/arm64/kernel/cacheinfo.c | 34 +++++++++++++++++++++++++++++++ >> arch/arm64/mm/init.c | 35 ++++++++++++++++++++++++++++++++ >> arch/arm64/mm/mmu.c | 7 ------- >> 5 files changed, 98 insertions(+), 9 deletions(-) >> >> diff --git a/arch/arm64/include/asm/cache.h b/arch/arm64/include/asm/cache.h >> index a4d1b5f771f6..420e9dde2c51 100644 >> --- a/arch/arm64/include/asm/cache.h >> +++ b/arch/arm64/include/asm/cache.h >> @@ -39,6 +39,27 @@ >> #define CLIDR_LOC(clidr) (((clidr) >> CLIDR_LOC_SHIFT) & 0x7) >> #define CLIDR_LOUIS(clidr) (((clidr) >> CLIDR_LOUIS_SHIFT) & 0x7) >> >> +#define CSSELR_TND_SHIFT 4 >> +#define CSSELR_TND_MASK (UL(1) << CSSELR_TND_SHIFT) >> +#define CSSELR_LEVEL_SHIFT 1 >> +#define CSSELR_LEVEL_MASK (UL(7) << CSSELR_LEVEL_SHIFT) >> +#define CSSELR_IND_SHIFT 0 >> +#define CSSERL_IND_MASK (UL(1) << CSSELR_IND_SHIFT) >> + >> +#define CCSIDR_64_LS_SHIFT 0 >> +#define CCSIDR_64_LS_MASK (UL(7) << CCSIDR_64_LS_SHIFT) >> +#define CCSIDR_64_ASSOC_SHIFT 3 >> +#define CCSIDR_64_ASSOC_MASK (UL(0x1FFFFF) << CCSIDR_64_ASSOC_SHIFT) >> +#define CCSIDR_64_SET_SHIFT 32 >> +#define CCSIDR_64_SET_MASK (UL(0xFFFFFF) << CCSIDR_64_SET_SHIFT) >> + >> +#define CCSIDR_32_LS_SHIFT 0 >> +#define CCSIDR_32_LS_MASK (UL(7) << CCSIDR_32_LS_SHIFT) >> +#define CCSIDR_32_ASSOC_SHIFT 3 >> +#define CCSIDR_32_ASSOC_MASK (UL(0x3FF) << CCSIDR_32_ASSOC_SHIFT) >> +#define CCSIDR_32_SET_SHIFT 13 >> +#define CCSIDR_32_SET_MASK (UL(0x7FFF) << CCSIDR_32_SET_SHIFT) > > I don't think we should be inferring cache structure from these register > values. The Arm ARM helpfully says: > > | You cannot make any inference about the actual sizes of caches based > | on these parameters. > > so we need to take the topology information from elsewhere. > Yeah, I also noticed the statement in the spec. However, the L1 cache size figured out from above registers are matching with "lscpu" on the machine where I did my tests. Note "lscpu" depends on sysfs entries whose information is retrieved from ACPI (PPTT) table. The number of cache levels are partially retrieved from system register (clidr_el1). It's doable to retrieve the L1 cache size from ACPI (PPTT) table. I'll change accordingly in v2 if this enablement is really needed. More clarify is provided below. > But before we get into that, can you justify why we need to do this at all, > please? Do you have data to show the benefit of adding this complexity? > Initially, I found it's the missed feature which has been enabled on mips/s390. Currently, all read-only anonymous VMAs are backed up by same zero page. It means all reads to these VMAs are cached by same set of cache, but still multiple ways if supported. So it would be nice to have multiple zero pages to back up these read-only anonymous VMAs, so that the reads on them can be cached by multiple sets (multiple ways still if supported). It's overall beneficial to the performance. Unfortunately, I didn't find a machine where the size of cache set is larger than page size. So I had one experiment as indication how L1 data cache miss affects the overall performance: L1 data cache size: 32KB L1 data cache line size: 64 Number of L1 data cache set: 64 Number of L1 data cache ways: 8 ---------------------------------------------------------------------- size = (cache_line_size) * (num_of_sets) * (num_of_ways) Kernel configuration: VA_BITS: 48 PAGE_SIZE: 4KB PMD HugeTLB Page Size: 2MB Experiment: I have a program to do the following things and check the consumed time and L1-data-cache-misses by perf. (1) Allocate (mmap) a PMD HugeTLB Page, which is 2MB. (2) Read on the mmap'd region in step of page size (4KB) for 8 or 9 times. Note 8 is the number of data cache ways. (3) Repeat (2) for 1000000 times. Result: (a) when we have 8 for the steps in (2): 37,103 L1-dcache-load-misses 0.217522515 seconds time elapsed 0.217564000 seconds user 0.000000000 seconds sys (b) when we have 9 for the steps in (2): 4,687,932 L1-dcache-load-misses (126 times) 0.248132105 seconds time elapsed (+14.2%) 0.248267000 seconds user 0.000000000 seconds sys Please let me know if it's worthy for a v2, to retrieve the cache size from ACPI (PPTT) table. The cost is to allocate multiple zero pages and the worst case is fail back to one zero page, as before :) Cheers, Gavin
Hi Robin, On 9/16/20 8:46 PM, Robin Murphy wrote: > On 2020-09-16 09:28, Will Deacon wrote: >> On Wed, Sep 16, 2020 at 01:25:23PM +1000, Gavin Shan wrote: >>> This enables color zero pages by allocating contigous page frames >>> for it. The number of pages for this is determined by L1 dCache >>> (or iCache) size, which is probbed from the hardware. >>> >>> * Add cache_total_size() to return L1 dCache (or iCache) size >>> >>> * Implement setup_zero_pages(), which is called after the page >>> allocator begins to work, to allocate the contigous pages >>> needed by color zero page. >>> >>> * Reworked ZERO_PAGE() and define __HAVE_COLOR_ZERO_PAGE. >>> >>> Signed-off-by: Gavin Shan <gshan@redhat.com> >>> --- >>> arch/arm64/include/asm/cache.h | 22 ++++++++++++++++++++ >>> arch/arm64/include/asm/pgtable.h | 9 ++++++-- >>> arch/arm64/kernel/cacheinfo.c | 34 +++++++++++++++++++++++++++++++ >>> arch/arm64/mm/init.c | 35 ++++++++++++++++++++++++++++++++ >>> arch/arm64/mm/mmu.c | 7 ------- >>> 5 files changed, 98 insertions(+), 9 deletions(-) >>> >>> diff --git a/arch/arm64/include/asm/cache.h b/arch/arm64/include/asm/cache.h >>> index a4d1b5f771f6..420e9dde2c51 100644 >>> --- a/arch/arm64/include/asm/cache.h >>> +++ b/arch/arm64/include/asm/cache.h >>> @@ -39,6 +39,27 @@ >>> #define CLIDR_LOC(clidr) (((clidr) >> CLIDR_LOC_SHIFT) & 0x7) >>> #define CLIDR_LOUIS(clidr) (((clidr) >> CLIDR_LOUIS_SHIFT) & 0x7) >>> +#define CSSELR_TND_SHIFT 4 >>> +#define CSSELR_TND_MASK (UL(1) << CSSELR_TND_SHIFT) >>> +#define CSSELR_LEVEL_SHIFT 1 >>> +#define CSSELR_LEVEL_MASK (UL(7) << CSSELR_LEVEL_SHIFT) >>> +#define CSSELR_IND_SHIFT 0 >>> +#define CSSERL_IND_MASK (UL(1) << CSSELR_IND_SHIFT) >>> + >>> +#define CCSIDR_64_LS_SHIFT 0 >>> +#define CCSIDR_64_LS_MASK (UL(7) << CCSIDR_64_LS_SHIFT) >>> +#define CCSIDR_64_ASSOC_SHIFT 3 >>> +#define CCSIDR_64_ASSOC_MASK (UL(0x1FFFFF) << CCSIDR_64_ASSOC_SHIFT) >>> +#define CCSIDR_64_SET_SHIFT 32 >>> +#define CCSIDR_64_SET_MASK (UL(0xFFFFFF) << CCSIDR_64_SET_SHIFT) >>> + >>> +#define CCSIDR_32_LS_SHIFT 0 >>> +#define CCSIDR_32_LS_MASK (UL(7) << CCSIDR_32_LS_SHIFT) >>> +#define CCSIDR_32_ASSOC_SHIFT 3 >>> +#define CCSIDR_32_ASSOC_MASK (UL(0x3FF) << CCSIDR_32_ASSOC_SHIFT) >>> +#define CCSIDR_32_SET_SHIFT 13 >>> +#define CCSIDR_32_SET_MASK (UL(0x7FFF) << CCSIDR_32_SET_SHIFT) >> >> I don't think we should be inferring cache structure from these register >> values. The Arm ARM helpfully says: >> >> | You cannot make any inference about the actual sizes of caches based >> | on these parameters. >> >> so we need to take the topology information from elsewhere. > > Yes, these represent parameters for the low-level cache maintenance by set/way instructions, and nothing more. There are definitely cases where they do not reflect the underlying cache structure (commit 793acf870ea3 is an obvious first call). > Thanks for your confirming. Yeah, the system registers aren't reliable since the cache way/set are hard coded to 1/1. As I suggested in another thread, ACPI (PPTT) table would be correct place to get such kind of information. [...] Cheers, Gavin
On 2020-09-17 04:35, Gavin Shan wrote: > Hi Will, > > On 9/16/20 6:28 PM, Will Deacon wrote: >> On Wed, Sep 16, 2020 at 01:25:23PM +1000, Gavin Shan wrote: >>> This enables color zero pages by allocating contigous page frames >>> for it. The number of pages for this is determined by L1 dCache >>> (or iCache) size, which is probbed from the hardware. >>> >>> * Add cache_total_size() to return L1 dCache (or iCache) size >>> >>> * Implement setup_zero_pages(), which is called after the page >>> allocator begins to work, to allocate the contigous pages >>> needed by color zero page. >>> >>> * Reworked ZERO_PAGE() and define __HAVE_COLOR_ZERO_PAGE. >>> >>> Signed-off-by: Gavin Shan <gshan@redhat.com> >>> --- >>> arch/arm64/include/asm/cache.h | 22 ++++++++++++++++++++ >>> arch/arm64/include/asm/pgtable.h | 9 ++++++-- >>> arch/arm64/kernel/cacheinfo.c | 34 +++++++++++++++++++++++++++++++ >>> arch/arm64/mm/init.c | 35 ++++++++++++++++++++++++++++++++ >>> arch/arm64/mm/mmu.c | 7 ------- >>> 5 files changed, 98 insertions(+), 9 deletions(-) >>> >>> diff --git a/arch/arm64/include/asm/cache.h >>> b/arch/arm64/include/asm/cache.h >>> index a4d1b5f771f6..420e9dde2c51 100644 >>> --- a/arch/arm64/include/asm/cache.h >>> +++ b/arch/arm64/include/asm/cache.h >>> @@ -39,6 +39,27 @@ >>> #define CLIDR_LOC(clidr) (((clidr) >> CLIDR_LOC_SHIFT) & 0x7) >>> #define CLIDR_LOUIS(clidr) (((clidr) >> CLIDR_LOUIS_SHIFT) & 0x7) >>> +#define CSSELR_TND_SHIFT 4 >>> +#define CSSELR_TND_MASK (UL(1) << CSSELR_TND_SHIFT) >>> +#define CSSELR_LEVEL_SHIFT 1 >>> +#define CSSELR_LEVEL_MASK (UL(7) << CSSELR_LEVEL_SHIFT) >>> +#define CSSELR_IND_SHIFT 0 >>> +#define CSSERL_IND_MASK (UL(1) << CSSELR_IND_SHIFT) >>> + >>> +#define CCSIDR_64_LS_SHIFT 0 >>> +#define CCSIDR_64_LS_MASK (UL(7) << CCSIDR_64_LS_SHIFT) >>> +#define CCSIDR_64_ASSOC_SHIFT 3 >>> +#define CCSIDR_64_ASSOC_MASK (UL(0x1FFFFF) << CCSIDR_64_ASSOC_SHIFT) >>> +#define CCSIDR_64_SET_SHIFT 32 >>> +#define CCSIDR_64_SET_MASK (UL(0xFFFFFF) << CCSIDR_64_SET_SHIFT) >>> + >>> +#define CCSIDR_32_LS_SHIFT 0 >>> +#define CCSIDR_32_LS_MASK (UL(7) << CCSIDR_32_LS_SHIFT) >>> +#define CCSIDR_32_ASSOC_SHIFT 3 >>> +#define CCSIDR_32_ASSOC_MASK (UL(0x3FF) << CCSIDR_32_ASSOC_SHIFT) >>> +#define CCSIDR_32_SET_SHIFT 13 >>> +#define CCSIDR_32_SET_MASK (UL(0x7FFF) << CCSIDR_32_SET_SHIFT) >> >> I don't think we should be inferring cache structure from these register >> values. The Arm ARM helpfully says: >> >> | You cannot make any inference about the actual sizes of caches based >> | on these parameters. >> >> so we need to take the topology information from elsewhere. >> > > Yeah, I also noticed the statement in the spec. However, the L1 cache size > figured out from above registers are matching with "lscpu" on the machine > where I did my tests. Note "lscpu" depends on sysfs entries whose > information > is retrieved from ACPI (PPTT) table. The number of cache levels are > partially > retrieved from system register (clidr_el1). > > It's doable to retrieve the L1 cache size from ACPI (PPTT) table. I'll > change accordingly in v2 if this enablement is really needed. More clarify > is provided below. > >> But before we get into that, can you justify why we need to do this at >> all, >> please? Do you have data to show the benefit of adding this complexity? >> > > Initially, I found it's the missed feature which has been enabled on > mips/s390. Currently, all read-only anonymous VMAs are backed up by > same zero page. It means all reads to these VMAs are cached by same > set of cache, but still multiple ways if supported. So it would be > nice to have multiple zero pages to back up these read-only anonymous > VMAs, so that the reads on them can be cached by multiple sets (multiple > ways still if supported). It's overall beneficial to the performance. Is this a concern for true PIPT caches, or is it really just working around a pathological case for alias-detecting VIPT caches? > Unfortunately, I didn't find a machine where the size of cache set is > larger than page size. So I had one experiment as indication how L1 > data cache miss affects the overall performance: > > L1 data cache size: 32KB > L1 data cache line size: 64 > Number of L1 data cache set: 64 > Number of L1 data cache ways: 8 > ---------------------------------------------------------------------- > size = (cache_line_size) * (num_of_sets) * (num_of_ways) > > Kernel configuration: > VA_BITS: 48 > PAGE_SIZE: 4KB > PMD HugeTLB Page Size: 2MB > > Experiment: > I have a program to do the following things and check the > consumed time and L1-data-cache-misses by perf. > > (1) Allocate (mmap) a PMD HugeTLB Page, which is 2MB. > (2) Read on the mmap'd region in step of page size (4KB) > for 8 or 9 times. Note 8 is the number of data cache > ways. > (3) Repeat (2) for 1000000 times. > Result: > (a) when we have 8 for the steps in (2): > 37,103 L1-dcache-load-misses > 0.217522515 seconds time elapsed > 0.217564000 seconds user > 0.000000000 seconds sys > (b) when we have 9 for the steps in (2): > 4,687,932 L1-dcache-load-misses (126 times) > 0.248132105 seconds time elapsed (+14.2%) > 0.248267000 seconds user > 0.000000000 seconds sys I have a vague feeling this may have come up before, but are there real-world applications that have a performance bottleneck on reading from uninitialised memory? As far as synthetic benchmarks go, I'm sure we could equally come up with one that shows a regression due to real data being pushed out of the cache by all those extra zeros ;) Robin. > Please let me know if it's worthy for a v2, to retrieve the cache size > from ACPI (PPTT) table. The cost is to allocate multiple zero pages and > the worst case is fail back to one zero page, as before :) > > Cheers, > Gavin > > > _______________________________________________ > linux-arm-kernel mailing list > linux-arm-kernel@lists.infradead.org > http://lists.infradead.org/mailman/listinfo/linux-arm-kernel
Hi Gavin, Thank you for the patch! Yet something to improve: [auto build test ERROR on arm64/for-next/core] [also build test ERROR on soc/for-next arm/for-next kvmarm/next v5.9-rc5 next-20200917] [cannot apply to xlnx/master] [If your patch is applied to the wrong git tree, kindly drop us a note. And when submitting patch, we suggest to use '--base' as documented in https://git-scm.com/docs/git-format-patch] url: https://github.com/0day-ci/linux/commits/Gavin-Shan/arm64-mm-Enable-color-zero-pages/20200916-112710 base: https://git.kernel.org/pub/scm/linux/kernel/git/arm64/linux.git for-next/core config: arm64-randconfig-r002-20200917 (attached as .config) compiler: clang version 12.0.0 (https://github.com/llvm/llvm-project 0ff28fa6a75617d61b1aeea77463d6a1684c3c89) reproduce (this is a W=1 build): wget https://raw.githubusercontent.com/intel/lkp-tests/master/sbin/make.cross -O ~/bin/make.cross chmod +x ~/bin/make.cross # install arm64 cross compiling tool for clang build # apt-get install binutils-aarch64-linux-gnu # save the attached .config to linux build tree COMPILER_INSTALL_PATH=$HOME/0day COMPILER=clang make.cross ARCH=arm64 If you fix the issue, kindly add following tag as appropriate Reported-by: kernel test robot <lkp@intel.com> All errors (new ones prefixed by >>): >> arch/arm64/kernel/cacheinfo.c:62:8: error: implicit declaration of function 'FIELD_PREP' [-Werror,-Wimplicit-function-declaration] val = FIELD_PREP(CSSELR_LEVEL_MASK, 0) | ^ >> arch/arm64/kernel/cacheinfo.c:68:16: error: implicit declaration of function 'FIELD_GET' [-Werror,-Wimplicit-function-declaration] size = (1 << FIELD_GET(CCSIDR_64_LS_MASK, val)) * ^ arch/arm64/kernel/cacheinfo.c:68:16: note: did you mean 'FIELD_PREP'? arch/arm64/kernel/cacheinfo.c:62:8: note: 'FIELD_PREP' declared here val = FIELD_PREP(CSSELR_LEVEL_MASK, 0) | ^ arch/arm64/kernel/cacheinfo.c:72:16: error: implicit declaration of function 'FIELD_GET' [-Werror,-Wimplicit-function-declaration] size = (1 << FIELD_GET(CCSIDR_32_LS_MASK, val)) * ^ 3 errors generated. # https://github.com/0day-ci/linux/commit/ab02baff46e4889747f134626f5cd3566fd90582 git remote add linux-review https://github.com/0day-ci/linux git fetch --no-tags linux-review Gavin-Shan/arm64-mm-Enable-color-zero-pages/20200916-112710 git checkout ab02baff46e4889747f134626f5cd3566fd90582 vim +/FIELD_PREP +62 arch/arm64/kernel/cacheinfo.c 45 46 int cache_total_size(void) 47 { 48 unsigned int ctype, size; 49 unsigned long val; 50 bool ccidx = false; 51 52 /* Check first level cache is supported */ 53 ctype = get_cache_type(1); 54 if (ctype == CACHE_TYPE_NOCACHE) 55 return 0; 56 57 /* ARMv8.3-CCIDX is supported or not */ 58 val = read_sanitised_ftr_reg(SYS_ID_MMFR4_EL1); 59 ccidx = !!(val & (UL(0xF) << ID_AA64MMFR2_CCIDX_SHIFT)); 60 61 /* Retrieve the information and calculate the total size */ > 62 val = FIELD_PREP(CSSELR_LEVEL_MASK, 0) | 63 FIELD_PREP(CSSERL_IND_MASK, 0); 64 write_sysreg(val, csselr_el1); 65 66 val = read_sysreg(ccsidr_el1); 67 if (ccidx) { > 68 size = (1 << FIELD_GET(CCSIDR_64_LS_MASK, val)) * 69 (FIELD_GET(CCSIDR_64_ASSOC_MASK, val) + 1) * 70 (FIELD_GET(CCSIDR_64_SET_MASK, val) + 1); 71 } else { 72 size = (1 << FIELD_GET(CCSIDR_32_LS_MASK, val)) * 73 (FIELD_GET(CCSIDR_32_ASSOC_MASK, val) + 1) * 74 (FIELD_GET(CCSIDR_32_SET_MASK, val) + 1); 75 } 76 77 return size; 78 } 79 --- 0-DAY CI Kernel Test Service, Intel Corporation https://lists.01.org/hyperkitty/list/kbuild-all@lists.01.org
Hi Robin, On 9/17/20 8:22 PM, Robin Murphy wrote: > On 2020-09-17 04:35, Gavin Shan wrote: >> On 9/16/20 6:28 PM, Will Deacon wrote: >>> On Wed, Sep 16, 2020 at 01:25:23PM +1000, Gavin Shan wrote: >>>> This enables color zero pages by allocating contigous page frames >>>> for it. The number of pages for this is determined by L1 dCache >>>> (or iCache) size, which is probbed from the hardware. >>>> >>>> * Add cache_total_size() to return L1 dCache (or iCache) size >>>> >>>> * Implement setup_zero_pages(), which is called after the page >>>> allocator begins to work, to allocate the contigous pages >>>> needed by color zero page. >>>> >>>> * Reworked ZERO_PAGE() and define __HAVE_COLOR_ZERO_PAGE. >>>> >>>> Signed-off-by: Gavin Shan <gshan@redhat.com> >>>> --- >>>> arch/arm64/include/asm/cache.h | 22 ++++++++++++++++++++ >>>> arch/arm64/include/asm/pgtable.h | 9 ++++++-- >>>> arch/arm64/kernel/cacheinfo.c | 34 +++++++++++++++++++++++++++++++ >>>> arch/arm64/mm/init.c | 35 ++++++++++++++++++++++++++++++++ >>>> arch/arm64/mm/mmu.c | 7 ------- >>>> 5 files changed, 98 insertions(+), 9 deletions(-) >>>> >>>> diff --git a/arch/arm64/include/asm/cache.h b/arch/arm64/include/asm/cache.h >>>> index a4d1b5f771f6..420e9dde2c51 100644 >>>> --- a/arch/arm64/include/asm/cache.h >>>> +++ b/arch/arm64/include/asm/cache.h >>>> @@ -39,6 +39,27 @@ >>>> #define CLIDR_LOC(clidr) (((clidr) >> CLIDR_LOC_SHIFT) & 0x7) >>>> #define CLIDR_LOUIS(clidr) (((clidr) >> CLIDR_LOUIS_SHIFT) & 0x7) >>>> +#define CSSELR_TND_SHIFT 4 >>>> +#define CSSELR_TND_MASK (UL(1) << CSSELR_TND_SHIFT) >>>> +#define CSSELR_LEVEL_SHIFT 1 >>>> +#define CSSELR_LEVEL_MASK (UL(7) << CSSELR_LEVEL_SHIFT) >>>> +#define CSSELR_IND_SHIFT 0 >>>> +#define CSSERL_IND_MASK (UL(1) << CSSELR_IND_SHIFT) >>>> + >>>> +#define CCSIDR_64_LS_SHIFT 0 >>>> +#define CCSIDR_64_LS_MASK (UL(7) << CCSIDR_64_LS_SHIFT) >>>> +#define CCSIDR_64_ASSOC_SHIFT 3 >>>> +#define CCSIDR_64_ASSOC_MASK (UL(0x1FFFFF) << CCSIDR_64_ASSOC_SHIFT) >>>> +#define CCSIDR_64_SET_SHIFT 32 >>>> +#define CCSIDR_64_SET_MASK (UL(0xFFFFFF) << CCSIDR_64_SET_SHIFT) >>>> + >>>> +#define CCSIDR_32_LS_SHIFT 0 >>>> +#define CCSIDR_32_LS_MASK (UL(7) << CCSIDR_32_LS_SHIFT) >>>> +#define CCSIDR_32_ASSOC_SHIFT 3 >>>> +#define CCSIDR_32_ASSOC_MASK (UL(0x3FF) << CCSIDR_32_ASSOC_SHIFT) >>>> +#define CCSIDR_32_SET_SHIFT 13 >>>> +#define CCSIDR_32_SET_MASK (UL(0x7FFF) << CCSIDR_32_SET_SHIFT) >>> >>> I don't think we should be inferring cache structure from these register >>> values. The Arm ARM helpfully says: >>> >>> | You cannot make any inference about the actual sizes of caches based >>> | on these parameters. >>> >>> so we need to take the topology information from elsewhere. >>> >> >> Yeah, I also noticed the statement in the spec. However, the L1 cache size >> figured out from above registers are matching with "lscpu" on the machine >> where I did my tests. Note "lscpu" depends on sysfs entries whose information >> is retrieved from ACPI (PPTT) table. The number of cache levels are partially >> retrieved from system register (clidr_el1). >> >> It's doable to retrieve the L1 cache size from ACPI (PPTT) table. I'll >> change accordingly in v2 if this enablement is really needed. More clarify >> is provided below. >> >>> But before we get into that, can you justify why we need to do this at all, >>> please? Do you have data to show the benefit of adding this complexity? >>> >> >> Initially, I found it's the missed feature which has been enabled on >> mips/s390. Currently, all read-only anonymous VMAs are backed up by >> same zero page. It means all reads to these VMAs are cached by same >> set of cache, but still multiple ways if supported. So it would be >> nice to have multiple zero pages to back up these read-only anonymous >> VMAs, so that the reads on them can be cached by multiple sets (multiple >> ways still if supported). It's overall beneficial to the performance. > > Is this a concern for true PIPT caches, or is it really just working around a pathological case for alias-detecting VIPT caches? > I think it's definitely a concern for PIPT caches. However, I'm not sure about VIPT caches because I failed to understand how it works from ARM8-A spec. If I'm correct, the index of VIPT cache line is still determined by the physical address and the number of sets is another limitation? For example, two virtual addresses (v1) and (v2) are translated to same physical address (p1), there is still one cache line (from particular set) for them. If so, this should helps in terms of performance. However, I'm not sure I understood VIPT caches correctly because there is one statement in the ARM8-A spec as below. It seems (v1) and (v2) can be backed by two different cache line from one particular set. ---- The only architecturally-guaranteed way to invalidate all aliases of a PA from a VIPT instruction cache is to invalidate the entire instruction cache. >> Unfortunately, I didn't find a machine where the size of cache set is >> larger than page size. So I had one experiment as indication how L1 >> data cache miss affects the overall performance: >> >> L1 data cache size: 32KB >> L1 data cache line size: 64 >> Number of L1 data cache set: 64 >> Number of L1 data cache ways: 8 >> ---------------------------------------------------------------------- >> size = (cache_line_size) * (num_of_sets) * (num_of_ways) >> >> Kernel configuration: >> VA_BITS: 48 >> PAGE_SIZE: 4KB >> PMD HugeTLB Page Size: 2MB >> >> Experiment: >> I have a program to do the following things and check the >> consumed time and L1-data-cache-misses by perf. >> >> (1) Allocate (mmap) a PMD HugeTLB Page, which is 2MB. >> (2) Read on the mmap'd region in step of page size (4KB) >> for 8 or 9 times. Note 8 is the number of data cache >> ways. >> (3) Repeat (2) for 1000000 times. >> Result: >> (a) when we have 8 for the steps in (2): >> 37,103 L1-dcache-load-misses >> 0.217522515 seconds time elapsed >> 0.217564000 seconds user >> 0.000000000 seconds sys >> (b) when we have 9 for the steps in (2): >> 4,687,932 L1-dcache-load-misses (126 times) >> 0.248132105 seconds time elapsed (+14.2%) >> 0.248267000 seconds user >> 0.000000000 seconds sys > > I have a vague feeling this may have come up before, but are there real-world applications that have a performance bottleneck on reading from uninitialised memory? As far as synthetic benchmarks go, I'm sure we could equally come up with one that shows a regression due to real data being pushed out of the cache by all those extra zeros ;) [...] Ok. If this was proposed before, I'm not sure if the link to that patchset is still available? :) When I was searching "my_zero_pfn" in upstream kernel, DAX uses the zero pages to fill the holes in one particular file in dax_load_hole(). mmap() on /proc/kcore could use zero page either. Yes, it's correct that the data cache has more chance to be flushed by these extra zeroes. However, it's not why we compromise the performance on reading zero page and depress it intentionally. The cache lines won't be used if there is no reading activities on zero page :) Cheers, Gavin
On 09/21/2020 08:26 AM, Gavin Shan wrote: > Hi Robin, > > On 9/17/20 8:22 PM, Robin Murphy wrote: >> On 2020-09-17 04:35, Gavin Shan wrote: >>> On 9/16/20 6:28 PM, Will Deacon wrote: >>>> On Wed, Sep 16, 2020 at 01:25:23PM +1000, Gavin Shan wrote: >>>>> This enables color zero pages by allocating contigous page frames >>>>> for it. The number of pages for this is determined by L1 dCache >>>>> (or iCache) size, which is probbed from the hardware. >>>>> >>>>> * Add cache_total_size() to return L1 dCache (or iCache) size >>>>> >>>>> * Implement setup_zero_pages(), which is called after the page >>>>> allocator begins to work, to allocate the contigous pages >>>>> needed by color zero page. >>>>> >>>>> * Reworked ZERO_PAGE() and define __HAVE_COLOR_ZERO_PAGE. >>>>> >>>>> Signed-off-by: Gavin Shan <gshan@redhat.com> >>>>> --- >>>>> arch/arm64/include/asm/cache.h | 22 ++++++++++++++++++++ >>>>> arch/arm64/include/asm/pgtable.h | 9 ++++++-- >>>>> arch/arm64/kernel/cacheinfo.c | 34 +++++++++++++++++++++++++++++++ >>>>> arch/arm64/mm/init.c | 35 ++++++++++++++++++++++++++++++++ >>>>> arch/arm64/mm/mmu.c | 7 ------- >>>>> 5 files changed, 98 insertions(+), 9 deletions(-) >>>>> >>>>> diff --git a/arch/arm64/include/asm/cache.h b/arch/arm64/include/asm/cache.h >>>>> index a4d1b5f771f6..420e9dde2c51 100644 >>>>> --- a/arch/arm64/include/asm/cache.h >>>>> +++ b/arch/arm64/include/asm/cache.h >>>>> @@ -39,6 +39,27 @@ >>>>> #define CLIDR_LOC(clidr) (((clidr) >> CLIDR_LOC_SHIFT) & 0x7) >>>>> #define CLIDR_LOUIS(clidr) (((clidr) >> CLIDR_LOUIS_SHIFT) & 0x7) >>>>> +#define CSSELR_TND_SHIFT 4 >>>>> +#define CSSELR_TND_MASK (UL(1) << CSSELR_TND_SHIFT) >>>>> +#define CSSELR_LEVEL_SHIFT 1 >>>>> +#define CSSELR_LEVEL_MASK (UL(7) << CSSELR_LEVEL_SHIFT) >>>>> +#define CSSELR_IND_SHIFT 0 >>>>> +#define CSSERL_IND_MASK (UL(1) << CSSELR_IND_SHIFT) >>>>> + >>>>> +#define CCSIDR_64_LS_SHIFT 0 >>>>> +#define CCSIDR_64_LS_MASK (UL(7) << CCSIDR_64_LS_SHIFT) >>>>> +#define CCSIDR_64_ASSOC_SHIFT 3 >>>>> +#define CCSIDR_64_ASSOC_MASK (UL(0x1FFFFF) << CCSIDR_64_ASSOC_SHIFT) >>>>> +#define CCSIDR_64_SET_SHIFT 32 >>>>> +#define CCSIDR_64_SET_MASK (UL(0xFFFFFF) << CCSIDR_64_SET_SHIFT) >>>>> + >>>>> +#define CCSIDR_32_LS_SHIFT 0 >>>>> +#define CCSIDR_32_LS_MASK (UL(7) << CCSIDR_32_LS_SHIFT) >>>>> +#define CCSIDR_32_ASSOC_SHIFT 3 >>>>> +#define CCSIDR_32_ASSOC_MASK (UL(0x3FF) << CCSIDR_32_ASSOC_SHIFT) >>>>> +#define CCSIDR_32_SET_SHIFT 13 >>>>> +#define CCSIDR_32_SET_MASK (UL(0x7FFF) << CCSIDR_32_SET_SHIFT) >>>> >>>> I don't think we should be inferring cache structure from these register >>>> values. The Arm ARM helpfully says: >>>> >>>> | You cannot make any inference about the actual sizes of caches based >>>> | on these parameters. >>>> >>>> so we need to take the topology information from elsewhere. >>>> >>> >>> Yeah, I also noticed the statement in the spec. However, the L1 cache size >>> figured out from above registers are matching with "lscpu" on the machine >>> where I did my tests. Note "lscpu" depends on sysfs entries whose information >>> is retrieved from ACPI (PPTT) table. The number of cache levels are partially >>> retrieved from system register (clidr_el1). >>> >>> It's doable to retrieve the L1 cache size from ACPI (PPTT) table. I'll >>> change accordingly in v2 if this enablement is really needed. More clarify >>> is provided below. >>> >>>> But before we get into that, can you justify why we need to do this at all, >>>> please? Do you have data to show the benefit of adding this complexity? >>>> >>> >>> Initially, I found it's the missed feature which has been enabled on >>> mips/s390. Currently, all read-only anonymous VMAs are backed up by >>> same zero page. It means all reads to these VMAs are cached by same >>> set of cache, but still multiple ways if supported. So it would be >>> nice to have multiple zero pages to back up these read-only anonymous >>> VMAs, so that the reads on them can be cached by multiple sets (multiple >>> ways still if supported). It's overall beneficial to the performance. >> >> Is this a concern for true PIPT caches, or is it really just working around a pathological case for alias-detecting VIPT caches? >> > > I think it's definitely a concern for PIPT caches. However, I'm not > sure about VIPT caches because I failed to understand how it works > from ARM8-A spec. If I'm correct, the index of VIPT cache line is > still determined by the physical address and the number of sets is > another limitation? For example, two virtual addresses (v1) and (v2) > are translated to same physical address (p1), there is still one > cache line (from particular set) for them. If so, this should helps > in terms of performance. > > However, I'm not sure I understood VIPT caches correctly because there > is one statement in the ARM8-A spec as below. It seems (v1) and (v2) > can be backed by two different cache line from one particular set. > > ---- > > The only architecturally-guaranteed way to invalidate all aliases of > a PA from a VIPT instruction cache is to invalidate the entire instruction > cache. > >>> Unfortunately, I didn't find a machine where the size of cache set is >>> larger than page size. So I had one experiment as indication how L1 >>> data cache miss affects the overall performance: >>> >>> L1 data cache size: 32KB >>> L1 data cache line size: 64 >>> Number of L1 data cache set: 64 >>> Number of L1 data cache ways: 8 >>> ---------------------------------------------------------------------- >>> size = (cache_line_size) * (num_of_sets) * (num_of_ways) >>> >>> Kernel configuration: >>> VA_BITS: 48 >>> PAGE_SIZE: 4KB >>> PMD HugeTLB Page Size: 2MB >>> >>> Experiment: >>> I have a program to do the following things and check the >>> consumed time and L1-data-cache-misses by perf. >>> >>> (1) Allocate (mmap) a PMD HugeTLB Page, which is 2MB. >>> (2) Read on the mmap'd region in step of page size (4KB) >>> for 8 or 9 times. Note 8 is the number of data cache >>> ways. >>> (3) Repeat (2) for 1000000 times. >>> Result: >>> (a) when we have 8 for the steps in (2): >>> 37,103 L1-dcache-load-misses >>> 0.217522515 seconds time elapsed >>> 0.217564000 seconds user >>> 0.000000000 seconds sys >>> (b) when we have 9 for the steps in (2): >>> 4,687,932 L1-dcache-load-misses (126 times) >>> 0.248132105 seconds time elapsed (+14.2%) >>> 0.248267000 seconds user >>> 0.000000000 seconds sys >> >> I have a vague feeling this may have come up before, but are there real-world applications that have a performance bottleneck on reading from uninitialised memory? As far as synthetic benchmarks go, I'm sure we could equally come up with one that shows a regression due to real data being pushed out of the cache by all those extra zeros ;) > > [...] > > Ok. If this was proposed before, I'm not sure if the link to that > patchset is still available? :) > > When I was searching "my_zero_pfn" in upstream kernel, DAX uses the > zero pages to fill the holes in one particular file in dax_load_hole(). > mmap() on /proc/kcore could use zero page either. But how often those mapped areas will be used afterwards either in DAX or /proc/kcore ? It seems like the minimal adaption for this feature so far on platforms (i.e s390 and mips) might have to do with real world workload's frequency of such read accesses on mapped areas using zero pages. > > Yes, it's correct that the data cache has more chance to be flushed by > these extra zeroes. However, it's not why we compromise the performance > on reading zero page and depress it intentionally. The cache lines won't > be used if there is no reading activities on zero page :) > > Cheers, > Gavin > > > >
Hi Anshuman, On 9/21/20 10:40 PM, Anshuman Khandual wrote: > On 09/21/2020 08:26 AM, Gavin Shan wrote: >> On 9/17/20 8:22 PM, Robin Murphy wrote: >>> On 2020-09-17 04:35, Gavin Shan wrote: >>>> On 9/16/20 6:28 PM, Will Deacon wrote: >>>>> On Wed, Sep 16, 2020 at 01:25:23PM +1000, Gavin Shan wrote: >>>>>> This enables color zero pages by allocating contigous page frames >>>>>> for it. The number of pages for this is determined by L1 dCache >>>>>> (or iCache) size, which is probbed from the hardware. >>>>>> >>>>>> * Add cache_total_size() to return L1 dCache (or iCache) size >>>>>> >>>>>> * Implement setup_zero_pages(), which is called after the page >>>>>> allocator begins to work, to allocate the contigous pages >>>>>> needed by color zero page. >>>>>> >>>>>> * Reworked ZERO_PAGE() and define __HAVE_COLOR_ZERO_PAGE. >>>>>> >>>>>> Signed-off-by: Gavin Shan <gshan@redhat.com> >>>>>> --- >>>>>> arch/arm64/include/asm/cache.h | 22 ++++++++++++++++++++ >>>>>> arch/arm64/include/asm/pgtable.h | 9 ++++++-- >>>>>> arch/arm64/kernel/cacheinfo.c | 34 +++++++++++++++++++++++++++++++ >>>>>> arch/arm64/mm/init.c | 35 ++++++++++++++++++++++++++++++++ >>>>>> arch/arm64/mm/mmu.c | 7 ------- >>>>>> 5 files changed, 98 insertions(+), 9 deletions(-) >>>>>> >>>>>> diff --git a/arch/arm64/include/asm/cache.h b/arch/arm64/include/asm/cache.h >>>>>> index a4d1b5f771f6..420e9dde2c51 100644 >>>>>> --- a/arch/arm64/include/asm/cache.h >>>>>> +++ b/arch/arm64/include/asm/cache.h >>>>>> @@ -39,6 +39,27 @@ >>>>>> #define CLIDR_LOC(clidr) (((clidr) >> CLIDR_LOC_SHIFT) & 0x7) >>>>>> #define CLIDR_LOUIS(clidr) (((clidr) >> CLIDR_LOUIS_SHIFT) & 0x7) >>>>>> +#define CSSELR_TND_SHIFT 4 >>>>>> +#define CSSELR_TND_MASK (UL(1) << CSSELR_TND_SHIFT) >>>>>> +#define CSSELR_LEVEL_SHIFT 1 >>>>>> +#define CSSELR_LEVEL_MASK (UL(7) << CSSELR_LEVEL_SHIFT) >>>>>> +#define CSSELR_IND_SHIFT 0 >>>>>> +#define CSSERL_IND_MASK (UL(1) << CSSELR_IND_SHIFT) >>>>>> + >>>>>> +#define CCSIDR_64_LS_SHIFT 0 >>>>>> +#define CCSIDR_64_LS_MASK (UL(7) << CCSIDR_64_LS_SHIFT) >>>>>> +#define CCSIDR_64_ASSOC_SHIFT 3 >>>>>> +#define CCSIDR_64_ASSOC_MASK (UL(0x1FFFFF) << CCSIDR_64_ASSOC_SHIFT) >>>>>> +#define CCSIDR_64_SET_SHIFT 32 >>>>>> +#define CCSIDR_64_SET_MASK (UL(0xFFFFFF) << CCSIDR_64_SET_SHIFT) >>>>>> + >>>>>> +#define CCSIDR_32_LS_SHIFT 0 >>>>>> +#define CCSIDR_32_LS_MASK (UL(7) << CCSIDR_32_LS_SHIFT) >>>>>> +#define CCSIDR_32_ASSOC_SHIFT 3 >>>>>> +#define CCSIDR_32_ASSOC_MASK (UL(0x3FF) << CCSIDR_32_ASSOC_SHIFT) >>>>>> +#define CCSIDR_32_SET_SHIFT 13 >>>>>> +#define CCSIDR_32_SET_MASK (UL(0x7FFF) << CCSIDR_32_SET_SHIFT) >>>>> [...] >> Ok. If this was proposed before, I'm not sure if the link to that >> patchset is still available? :) >> >> When I was searching "my_zero_pfn" in upstream kernel, DAX uses the >> zero pages to fill the holes in one particular file in dax_load_hole(). >> mmap() on /proc/kcore could use zero page either. > > But how often those mapped areas will be used afterwards either in DAX > or /proc/kcore ? It seems like the minimal adaption for this feature so > far on platforms (i.e s390 and mips) might have to do with real world > workload's frequency of such read accesses on mapped areas using zero > pages. > I don't think /proc/kcore is used frequently in real world, still depending on the workload. DAX has been supported by multiple filesystems (ext2/ext4/ xfs). Taking ext4 as an example, all allocated extents (blocks), which isn't be written with the data. The data is retrieved from the zero page. I guess this intends to avoid exposing data, which was written by previous user for safety reason. This would be common case and heavily depend on read performance on zero page. Besides, holes (blocks) in ext4 are also backed by zero pages and it would happen frequently. However, I'm not a filesystem guy. I checked the code and understood the code as above, but I might be wrong completely here. Yeah, I failed to understand why this feature was enabled on s390/mips from the corresponding commit logs. Nothing helpful is provided there. I guess some specific S390/MIPS CPUs have large L1 cache capacity, multiple page sizes in one set. In this case, multiple (color) zero pages can avoid the cache line collisions on reading these pages. However, I'm not sure about arm64. On the CPU where I had my experiment, there is 8-ways/64-sets and 32KB L1 dCache and iCache, meaning 4KB L1 cache in one particular set. This feature has low cost to be enabled as several extra pages are needed and not harmful at least as I can see. [...] Thanks, Gavin
diff --git a/arch/arm64/include/asm/cache.h b/arch/arm64/include/asm/cache.h index a4d1b5f771f6..420e9dde2c51 100644 --- a/arch/arm64/include/asm/cache.h +++ b/arch/arm64/include/asm/cache.h @@ -39,6 +39,27 @@ #define CLIDR_LOC(clidr) (((clidr) >> CLIDR_LOC_SHIFT) & 0x7) #define CLIDR_LOUIS(clidr) (((clidr) >> CLIDR_LOUIS_SHIFT) & 0x7) +#define CSSELR_TND_SHIFT 4 +#define CSSELR_TND_MASK (UL(1) << CSSELR_TND_SHIFT) +#define CSSELR_LEVEL_SHIFT 1 +#define CSSELR_LEVEL_MASK (UL(7) << CSSELR_LEVEL_SHIFT) +#define CSSELR_IND_SHIFT 0 +#define CSSERL_IND_MASK (UL(1) << CSSELR_IND_SHIFT) + +#define CCSIDR_64_LS_SHIFT 0 +#define CCSIDR_64_LS_MASK (UL(7) << CCSIDR_64_LS_SHIFT) +#define CCSIDR_64_ASSOC_SHIFT 3 +#define CCSIDR_64_ASSOC_MASK (UL(0x1FFFFF) << CCSIDR_64_ASSOC_SHIFT) +#define CCSIDR_64_SET_SHIFT 32 +#define CCSIDR_64_SET_MASK (UL(0xFFFFFF) << CCSIDR_64_SET_SHIFT) + +#define CCSIDR_32_LS_SHIFT 0 +#define CCSIDR_32_LS_MASK (UL(7) << CCSIDR_32_LS_SHIFT) +#define CCSIDR_32_ASSOC_SHIFT 3 +#define CCSIDR_32_ASSOC_MASK (UL(0x3FF) << CCSIDR_32_ASSOC_SHIFT) +#define CCSIDR_32_SET_SHIFT 13 +#define CCSIDR_32_SET_MASK (UL(0x7FFF) << CCSIDR_32_SET_SHIFT) + /* * Memory returned by kmalloc() may be used for DMA, so we must make * sure that all such allocations are cache aligned. Otherwise, @@ -89,6 +110,7 @@ static inline int cache_line_size_of_cpu(void) } int cache_line_size(void); +int cache_total_size(void); /* * Read the effective value of CTR_EL0. diff --git a/arch/arm64/include/asm/pgtable.h b/arch/arm64/include/asm/pgtable.h index 6953498f4d40..5cb5f8bb090d 100644 --- a/arch/arm64/include/asm/pgtable.h +++ b/arch/arm64/include/asm/pgtable.h @@ -54,8 +54,13 @@ extern void __pgd_error(const char *file, int line, unsigned long val); * ZERO_PAGE is a global shared page that is always zero: used * for zero-mapped memory areas etc.. */ -extern unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)]; -#define ZERO_PAGE(vaddr) phys_to_page(__pa_symbol(empty_zero_page)) +extern unsigned long empty_zero_page; +extern unsigned long zero_page_mask; + +#define __HAVE_COLOR_ZERO_PAGE +#define ZERO_PAGE(vaddr) \ + (virt_to_page((void *)(empty_zero_page + \ + (((unsigned long)(vaddr)) & zero_page_mask)))) #define pte_ERROR(pte) __pte_error(__FILE__, __LINE__, pte_val(pte)) diff --git a/arch/arm64/kernel/cacheinfo.c b/arch/arm64/kernel/cacheinfo.c index 7fa6828bb488..d3b9ab757014 100644 --- a/arch/arm64/kernel/cacheinfo.c +++ b/arch/arm64/kernel/cacheinfo.c @@ -43,6 +43,40 @@ static void ci_leaf_init(struct cacheinfo *this_leaf, this_leaf->type = type; } +int cache_total_size(void) +{ + unsigned int ctype, size; + unsigned long val; + bool ccidx = false; + + /* Check first level cache is supported */ + ctype = get_cache_type(1); + if (ctype == CACHE_TYPE_NOCACHE) + return 0; + + /* ARMv8.3-CCIDX is supported or not */ + val = read_sanitised_ftr_reg(SYS_ID_MMFR4_EL1); + ccidx = !!(val & (UL(0xF) << ID_AA64MMFR2_CCIDX_SHIFT)); + + /* Retrieve the information and calculate the total size */ + val = FIELD_PREP(CSSELR_LEVEL_MASK, 0) | + FIELD_PREP(CSSERL_IND_MASK, 0); + write_sysreg(val, csselr_el1); + + val = read_sysreg(ccsidr_el1); + if (ccidx) { + size = (1 << FIELD_GET(CCSIDR_64_LS_MASK, val)) * + (FIELD_GET(CCSIDR_64_ASSOC_MASK, val) + 1) * + (FIELD_GET(CCSIDR_64_SET_MASK, val) + 1); + } else { + size = (1 << FIELD_GET(CCSIDR_32_LS_MASK, val)) * + (FIELD_GET(CCSIDR_32_ASSOC_MASK, val) + 1) * + (FIELD_GET(CCSIDR_32_SET_MASK, val) + 1); + } + + return size; +} + static int __init_cache_level(unsigned int cpu) { unsigned int ctype, level, leaves, fw_level; diff --git a/arch/arm64/mm/init.c b/arch/arm64/mm/init.c index 481d22c32a2e..ca6b3cddafb7 100644 --- a/arch/arm64/mm/init.c +++ b/arch/arm64/mm/init.c @@ -69,6 +69,11 @@ EXPORT_SYMBOL(vmemmap); phys_addr_t arm64_dma_phys_limit __ro_after_init; static phys_addr_t arm64_dma32_phys_limit __ro_after_init; +unsigned long empty_zero_page; +EXPORT_SYMBOL(empty_zero_page); +unsigned long zero_page_mask; +EXPORT_SYMBOL(zero_page_mask); + #ifdef CONFIG_KEXEC_CORE /* * reserve_crashkernel() - reserves memory for crash kernel @@ -507,6 +512,35 @@ static void __init free_unused_memmap(void) } #endif /* !CONFIG_SPARSEMEM_VMEMMAP */ +static void __init setup_zero_pages(void) +{ + struct page *page; + int order, size, i; + + size = cache_total_size(); + order = size > 0 ? get_order(PAGE_ALIGN(size)) : 0; + order = min(order, MAX_ORDER - 1); + + do { + empty_zero_page = __get_free_pages(GFP_KERNEL | __GFP_ZERO, + order); + if (empty_zero_page) + break; + } while (--order >= 0); + + if (!empty_zero_page) + panic("%s: out of memory\n", __func__); + + page = virt_to_page((void *) empty_zero_page); + split_page(page, order); + for (i = 1 << order; i > 0; i--) { + mark_page_reserved(page); + page++; + } + + zero_page_mask = ((PAGE_SIZE << order) - 1) & PAGE_MASK; +} + /* * mem_init() marks the free areas in the mem_map and tells us how much memory * is free. This is done after various parts of the system have claimed their @@ -527,6 +561,7 @@ void __init mem_init(void) #endif /* this will put all unused low memory onto the freelists */ memblock_free_all(); + setup_zero_pages(); mem_init_print_info(NULL); diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c index 75df62fea1b6..736939ab3b4f 100644 --- a/arch/arm64/mm/mmu.c +++ b/arch/arm64/mm/mmu.c @@ -49,13 +49,6 @@ EXPORT_SYMBOL(vabits_actual); u64 kimage_voffset __ro_after_init; EXPORT_SYMBOL(kimage_voffset); -/* - * Empty_zero_page is a special page that is used for zero-initialized data - * and COW. - */ -unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)] __page_aligned_bss; -EXPORT_SYMBOL(empty_zero_page); - static pte_t bm_pte[PTRS_PER_PTE] __page_aligned_bss; static pmd_t bm_pmd[PTRS_PER_PMD] __page_aligned_bss __maybe_unused; static pud_t bm_pud[PTRS_PER_PUD] __page_aligned_bss __maybe_unused;
This enables color zero pages by allocating contigous page frames for it. The number of pages for this is determined by L1 dCache (or iCache) size, which is probbed from the hardware. * Add cache_total_size() to return L1 dCache (or iCache) size * Implement setup_zero_pages(), which is called after the page allocator begins to work, to allocate the contigous pages needed by color zero page. * Reworked ZERO_PAGE() and define __HAVE_COLOR_ZERO_PAGE. Signed-off-by: Gavin Shan <gshan@redhat.com> --- arch/arm64/include/asm/cache.h | 22 ++++++++++++++++++++ arch/arm64/include/asm/pgtable.h | 9 ++++++-- arch/arm64/kernel/cacheinfo.c | 34 +++++++++++++++++++++++++++++++ arch/arm64/mm/init.c | 35 ++++++++++++++++++++++++++++++++ arch/arm64/mm/mmu.c | 7 ------- 5 files changed, 98 insertions(+), 9 deletions(-)