Message ID | 20190514235111.2817276-1-guro@fb.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | mm: refactor __vunmap() to avoid duplicated call to find_vm_area() | expand |
On 05/15/2019 05:21 AM, Roman Gushchin wrote: > __vunmap() calls find_vm_area() twice without an obvious reason: > first directly to get the area pointer, second indirectly by calling > vm_remove_mappings()->remove_vm_area(), which is again searching > for the area. > > To remove this redundancy, let's split remove_vm_area() into > __remove_vm_area(struct vmap_area *), which performs the actual area > removal, and remove_vm_area(const void *addr) wrapper, which can > be used everywhere, where it has been used before. Let's pass > a pointer to the vm_area instead of vm_struct to vm_remove_mappings(), > so it can pass it to __remove_vm_area() and avoid the redundant area > lookup. > > On my test setup, I've got 5-10% speed up on vfree()'ing 1000000 > of 4-pages vmalloc blocks. Though results from 1000000 single page vmalloc blocks remain inconclusive, 4-page based vmalloc block's result shows improvement in the range of 5-10%.
On 05/15/2019 05:21 AM, Roman Gushchin wrote: > __vunmap() calls find_vm_area() twice without an obvious reason: > first directly to get the area pointer, second indirectly by calling > vm_remove_mappings()->remove_vm_area(), which is again searching > for the area. > > To remove this redundancy, let's split remove_vm_area() into > __remove_vm_area(struct vmap_area *), which performs the actual area > removal, and remove_vm_area(const void *addr) wrapper, which can > be used everywhere, where it has been used before. Let's pass > a pointer to the vm_area instead of vm_struct to vm_remove_mappings(), > so it can pass it to __remove_vm_area() and avoid the redundant area > lookup. > > On my test setup, I've got 5-10% speed up on vfree()'ing 1000000 > of 4-pages vmalloc blocks. > > Perf report before: > 29.44% cat [kernel.kallsyms] [k] free_unref_page > 11.88% cat [kernel.kallsyms] [k] find_vmap_area > 9.28% cat [kernel.kallsyms] [k] __free_pages > 7.44% cat [kernel.kallsyms] [k] __slab_free > 7.28% cat [kernel.kallsyms] [k] vunmap_page_range > 4.56% cat [kernel.kallsyms] [k] __vunmap > 3.64% cat [kernel.kallsyms] [k] __purge_vmap_area_lazy > 3.04% cat [kernel.kallsyms] [k] __free_vmap_area > > Perf report after: > 32.41% cat [kernel.kallsyms] [k] free_unref_page > 7.79% cat [kernel.kallsyms] [k] find_vmap_area > 7.40% cat [kernel.kallsyms] [k] __slab_free > 7.31% cat [kernel.kallsyms] [k] vunmap_page_range > 6.84% cat [kernel.kallsyms] [k] __free_pages > 6.01% cat [kernel.kallsyms] [k] __vunmap > 3.98% cat [kernel.kallsyms] [k] smp_call_function_single > 3.81% cat [kernel.kallsyms] [k] __purge_vmap_area_lazy > 2.77% cat [kernel.kallsyms] [k] __free_vmap_area > > Signed-off-by: Roman Gushchin <guro@fb.com> > Cc: Johannes Weiner <hannes@cmpxchg.org> > Cc: Matthew Wilcox <willy@infradead.org> > Cc: Vlastimil Babka <vbabka@suse.cz> > --- > mm/vmalloc.c | 52 +++++++++++++++++++++++++++++----------------------- > 1 file changed, 29 insertions(+), 23 deletions(-) > > diff --git a/mm/vmalloc.c b/mm/vmalloc.c > index c42872ed82ac..8d4907865614 100644 > --- a/mm/vmalloc.c > +++ b/mm/vmalloc.c > @@ -2075,6 +2075,22 @@ struct vm_struct *find_vm_area(const void *addr) > return NULL; > } > > +static struct vm_struct *__remove_vm_area(struct vmap_area *va) > +{ > + struct vm_struct *vm = va->vm; > + > + spin_lock(&vmap_area_lock); > + va->vm = NULL; > + va->flags &= ~VM_VM_AREA; > + va->flags |= VM_LAZY_FREE; > + spin_unlock(&vmap_area_lock); > + > + kasan_free_shadow(vm); > + free_unmap_vmap_area(va); > + > + return vm; > +} > + > /** > * remove_vm_area - find and remove a continuous kernel virtual area > * @addr: base address > @@ -2087,26 +2103,14 @@ struct vm_struct *find_vm_area(const void *addr) > */ > struct vm_struct *remove_vm_area(const void *addr) > { > + struct vm_struct *vm = NULL; > struct vmap_area *va; > > - might_sleep(); Is not this necessary any more ? > - > va = find_vmap_area((unsigned long)addr); > - if (va && va->flags & VM_VM_AREA) { > - struct vm_struct *vm = va->vm; > - > - spin_lock(&vmap_area_lock); > - va->vm = NULL; > - va->flags &= ~VM_VM_AREA; > - va->flags |= VM_LAZY_FREE; > - spin_unlock(&vmap_area_lock); > - > - kasan_free_shadow(vm); > - free_unmap_vmap_area(va); > + if (va && va->flags & VM_VM_AREA) > + vm = __remove_vm_area(va); > > - return vm; > - } > - return NULL; > + return vm; > } Other callers of remove_vm_area() cannot use __remove_vm_area() directly as well to save a look up ?
On Wed, May 15, 2019 at 01:11:46PM +0530, Anshuman Khandual wrote: > > > On 05/15/2019 05:21 AM, Roman Gushchin wrote: > > __vunmap() calls find_vm_area() twice without an obvious reason: > > first directly to get the area pointer, second indirectly by calling > > vm_remove_mappings()->remove_vm_area(), which is again searching > > for the area. > > > > To remove this redundancy, let's split remove_vm_area() into > > __remove_vm_area(struct vmap_area *), which performs the actual area > > removal, and remove_vm_area(const void *addr) wrapper, which can > > be used everywhere, where it has been used before. Let's pass > > a pointer to the vm_area instead of vm_struct to vm_remove_mappings(), > > so it can pass it to __remove_vm_area() and avoid the redundant area > > lookup. > > > > On my test setup, I've got 5-10% speed up on vfree()'ing 1000000 > > of 4-pages vmalloc blocks. > > > > Perf report before: > > 29.44% cat [kernel.kallsyms] [k] free_unref_page > > 11.88% cat [kernel.kallsyms] [k] find_vmap_area > > 9.28% cat [kernel.kallsyms] [k] __free_pages > > 7.44% cat [kernel.kallsyms] [k] __slab_free > > 7.28% cat [kernel.kallsyms] [k] vunmap_page_range > > 4.56% cat [kernel.kallsyms] [k] __vunmap > > 3.64% cat [kernel.kallsyms] [k] __purge_vmap_area_lazy > > 3.04% cat [kernel.kallsyms] [k] __free_vmap_area > > > > Perf report after: > > 32.41% cat [kernel.kallsyms] [k] free_unref_page > > 7.79% cat [kernel.kallsyms] [k] find_vmap_area > > 7.40% cat [kernel.kallsyms] [k] __slab_free > > 7.31% cat [kernel.kallsyms] [k] vunmap_page_range > > 6.84% cat [kernel.kallsyms] [k] __free_pages > > 6.01% cat [kernel.kallsyms] [k] __vunmap > > 3.98% cat [kernel.kallsyms] [k] smp_call_function_single > > 3.81% cat [kernel.kallsyms] [k] __purge_vmap_area_lazy > > 2.77% cat [kernel.kallsyms] [k] __free_vmap_area > > > > Signed-off-by: Roman Gushchin <guro@fb.com> > > Cc: Johannes Weiner <hannes@cmpxchg.org> > > Cc: Matthew Wilcox <willy@infradead.org> > > Cc: Vlastimil Babka <vbabka@suse.cz> > > --- > > mm/vmalloc.c | 52 +++++++++++++++++++++++++++++----------------------- > > 1 file changed, 29 insertions(+), 23 deletions(-) > > > > diff --git a/mm/vmalloc.c b/mm/vmalloc.c > > index c42872ed82ac..8d4907865614 100644 > > --- a/mm/vmalloc.c > > +++ b/mm/vmalloc.c > > @@ -2075,6 +2075,22 @@ struct vm_struct *find_vm_area(const void *addr) > > return NULL; > > } > > > > +static struct vm_struct *__remove_vm_area(struct vmap_area *va) > > +{ > > + struct vm_struct *vm = va->vm; > > + > > + spin_lock(&vmap_area_lock); > > + va->vm = NULL; > > + va->flags &= ~VM_VM_AREA; > > + va->flags |= VM_LAZY_FREE; > > + spin_unlock(&vmap_area_lock); > > + > > + kasan_free_shadow(vm); > > + free_unmap_vmap_area(va); > > + > > + return vm; > > +} > > + > > /** > > * remove_vm_area - find and remove a continuous kernel virtual area > > * @addr: base address > > @@ -2087,26 +2103,14 @@ struct vm_struct *find_vm_area(const void *addr) > > */ > > struct vm_struct *remove_vm_area(const void *addr) > > { > > + struct vm_struct *vm = NULL; > > struct vmap_area *va; > > > > - might_sleep(); > > Is not this necessary any more ? We've discussed it here: https://lkml.org/lkml/2019/4/17/1098 . Tl;dr it's not that useful. > > > - > > va = find_vmap_area((unsigned long)addr); > > - if (va && va->flags & VM_VM_AREA) { > > - struct vm_struct *vm = va->vm; > > - > > - spin_lock(&vmap_area_lock); > > - va->vm = NULL; > > - va->flags &= ~VM_VM_AREA; > > - va->flags |= VM_LAZY_FREE; > > - spin_unlock(&vmap_area_lock); > > - > > - kasan_free_shadow(vm); > > - free_unmap_vmap_area(va); > > + if (va && va->flags & VM_VM_AREA) > > + vm = __remove_vm_area(va); > > > > - return vm; > > - } > > - return NULL; > > + return vm; > > } > > Other callers of remove_vm_area() cannot use __remove_vm_area() directly as well > to save a look up ? > I'll take a look. Good idea, thanks! Roman
On Wed, May 15, 2019 at 09:57:11AM +0530, Anshuman Khandual wrote: > > > On 05/15/2019 05:21 AM, Roman Gushchin wrote: > > __vunmap() calls find_vm_area() twice without an obvious reason: > > first directly to get the area pointer, second indirectly by calling > > vm_remove_mappings()->remove_vm_area(), which is again searching > > for the area. > > > > To remove this redundancy, let's split remove_vm_area() into > > __remove_vm_area(struct vmap_area *), which performs the actual area > > removal, and remove_vm_area(const void *addr) wrapper, which can > > be used everywhere, where it has been used before. Let's pass > > a pointer to the vm_area instead of vm_struct to vm_remove_mappings(), > > so it can pass it to __remove_vm_area() and avoid the redundant area > > lookup. > > > > On my test setup, I've got 5-10% speed up on vfree()'ing 1000000 > > of 4-pages vmalloc blocks. > > Though results from 1000000 single page vmalloc blocks remain inconclusive, > 4-page based vmalloc block's result shows improvement in the range of 5-10%. So you can confirm my numbers? Great, thank you!
> > -/* Handle removing and resetting vm mappings related to the vm_struct. */ > -static void vm_remove_mappings(struct vm_struct *area, int deallocate_pages) > +/* Handle removing and resetting vm mappings related to the va->vm vm_struct. */ > +static void vm_remove_mappings(struct vmap_area *va, int deallocate_pages) Does this apply to 5.1? I'm confused because I can't find vm_remove_mappings() in 5.1. Ira > { > + struct vm_struct *area = va->vm; > unsigned long addr = (unsigned long)area->addr; > unsigned long start = ULONG_MAX, end = 0; > int flush_reset = area->flags & VM_FLUSH_RESET_PERMS; > @@ -2138,7 +2143,7 @@ static void vm_remove_mappings(struct vm_struct *area, int deallocate_pages) > set_memory_rw(addr, area->nr_pages); > } > > - remove_vm_area(area->addr); > + __remove_vm_area(va); > > /* If this is not VM_FLUSH_RESET_PERMS memory, no need for the below. */ > if (!flush_reset) > @@ -2178,6 +2183,7 @@ static void vm_remove_mappings(struct vm_struct *area, int deallocate_pages) > static void __vunmap(const void *addr, int deallocate_pages) > { > struct vm_struct *area; > + struct vmap_area *va; > > if (!addr) > return; > @@ -2186,17 +2192,18 @@ static void __vunmap(const void *addr, int deallocate_pages) > addr)) > return; > > - area = find_vm_area(addr); > - if (unlikely(!area)) { > + va = find_vmap_area((unsigned long)addr); > + if (unlikely(!va || !(va->flags & VM_VM_AREA))) { > WARN(1, KERN_ERR "Trying to vfree() nonexistent vm area (%p)\n", > addr); > return; > } > > + area = va->vm; > debug_check_no_locks_freed(area->addr, get_vm_area_size(area)); > debug_check_no_obj_freed(area->addr, get_vm_area_size(area)); > > - vm_remove_mappings(area, deallocate_pages); > + vm_remove_mappings(va, deallocate_pages); > > if (deallocate_pages) { > int i; > @@ -2212,7 +2219,6 @@ static void __vunmap(const void *addr, int deallocate_pages) > } > > kfree(area); > - return; > } > > static inline void __vfree_deferred(const void *addr) > -- > 2.20.1 >
On Wed, May 15, 2019 at 10:35:26AM -0700, Ira Weiny wrote: > > > > -/* Handle removing and resetting vm mappings related to the vm_struct. */ > > -static void vm_remove_mappings(struct vm_struct *area, int deallocate_pages) > > +/* Handle removing and resetting vm mappings related to the va->vm vm_struct. */ > > +static void vm_remove_mappings(struct vmap_area *va, int deallocate_pages) > > Does this apply to 5.1? I'm confused because I can't find vm_remove_mappings() > in 5.1. Not really, it's based on top of the current mm tree. You can find the earlier version which applies on 5.1 here: https://lkml.org/lkml/2019/4/17/954 Thanks!
diff --git a/mm/vmalloc.c b/mm/vmalloc.c index c42872ed82ac..8d4907865614 100644 --- a/mm/vmalloc.c +++ b/mm/vmalloc.c @@ -2075,6 +2075,22 @@ struct vm_struct *find_vm_area(const void *addr) return NULL; } +static struct vm_struct *__remove_vm_area(struct vmap_area *va) +{ + struct vm_struct *vm = va->vm; + + spin_lock(&vmap_area_lock); + va->vm = NULL; + va->flags &= ~VM_VM_AREA; + va->flags |= VM_LAZY_FREE; + spin_unlock(&vmap_area_lock); + + kasan_free_shadow(vm); + free_unmap_vmap_area(va); + + return vm; +} + /** * remove_vm_area - find and remove a continuous kernel virtual area * @addr: base address @@ -2087,26 +2103,14 @@ struct vm_struct *find_vm_area(const void *addr) */ struct vm_struct *remove_vm_area(const void *addr) { + struct vm_struct *vm = NULL; struct vmap_area *va; - might_sleep(); - va = find_vmap_area((unsigned long)addr); - if (va && va->flags & VM_VM_AREA) { - struct vm_struct *vm = va->vm; - - spin_lock(&vmap_area_lock); - va->vm = NULL; - va->flags &= ~VM_VM_AREA; - va->flags |= VM_LAZY_FREE; - spin_unlock(&vmap_area_lock); - - kasan_free_shadow(vm); - free_unmap_vmap_area(va); + if (va && va->flags & VM_VM_AREA) + vm = __remove_vm_area(va); - return vm; - } - return NULL; + return vm; } static inline void set_area_direct_map(const struct vm_struct *area, @@ -2119,9 +2123,10 @@ static inline void set_area_direct_map(const struct vm_struct *area, set_direct_map(area->pages[i]); } -/* Handle removing and resetting vm mappings related to the vm_struct. */ -static void vm_remove_mappings(struct vm_struct *area, int deallocate_pages) +/* Handle removing and resetting vm mappings related to the va->vm vm_struct. */ +static void vm_remove_mappings(struct vmap_area *va, int deallocate_pages) { + struct vm_struct *area = va->vm; unsigned long addr = (unsigned long)area->addr; unsigned long start = ULONG_MAX, end = 0; int flush_reset = area->flags & VM_FLUSH_RESET_PERMS; @@ -2138,7 +2143,7 @@ static void vm_remove_mappings(struct vm_struct *area, int deallocate_pages) set_memory_rw(addr, area->nr_pages); } - remove_vm_area(area->addr); + __remove_vm_area(va); /* If this is not VM_FLUSH_RESET_PERMS memory, no need for the below. */ if (!flush_reset) @@ -2178,6 +2183,7 @@ static void vm_remove_mappings(struct vm_struct *area, int deallocate_pages) static void __vunmap(const void *addr, int deallocate_pages) { struct vm_struct *area; + struct vmap_area *va; if (!addr) return; @@ -2186,17 +2192,18 @@ static void __vunmap(const void *addr, int deallocate_pages) addr)) return; - area = find_vm_area(addr); - if (unlikely(!area)) { + va = find_vmap_area((unsigned long)addr); + if (unlikely(!va || !(va->flags & VM_VM_AREA))) { WARN(1, KERN_ERR "Trying to vfree() nonexistent vm area (%p)\n", addr); return; } + area = va->vm; debug_check_no_locks_freed(area->addr, get_vm_area_size(area)); debug_check_no_obj_freed(area->addr, get_vm_area_size(area)); - vm_remove_mappings(area, deallocate_pages); + vm_remove_mappings(va, deallocate_pages); if (deallocate_pages) { int i; @@ -2212,7 +2219,6 @@ static void __vunmap(const void *addr, int deallocate_pages) } kfree(area); - return; } static inline void __vfree_deferred(const void *addr)
__vunmap() calls find_vm_area() twice without an obvious reason: first directly to get the area pointer, second indirectly by calling vm_remove_mappings()->remove_vm_area(), which is again searching for the area. To remove this redundancy, let's split remove_vm_area() into __remove_vm_area(struct vmap_area *), which performs the actual area removal, and remove_vm_area(const void *addr) wrapper, which can be used everywhere, where it has been used before. Let's pass a pointer to the vm_area instead of vm_struct to vm_remove_mappings(), so it can pass it to __remove_vm_area() and avoid the redundant area lookup. On my test setup, I've got 5-10% speed up on vfree()'ing 1000000 of 4-pages vmalloc blocks. Perf report before: 29.44% cat [kernel.kallsyms] [k] free_unref_page 11.88% cat [kernel.kallsyms] [k] find_vmap_area 9.28% cat [kernel.kallsyms] [k] __free_pages 7.44% cat [kernel.kallsyms] [k] __slab_free 7.28% cat [kernel.kallsyms] [k] vunmap_page_range 4.56% cat [kernel.kallsyms] [k] __vunmap 3.64% cat [kernel.kallsyms] [k] __purge_vmap_area_lazy 3.04% cat [kernel.kallsyms] [k] __free_vmap_area Perf report after: 32.41% cat [kernel.kallsyms] [k] free_unref_page 7.79% cat [kernel.kallsyms] [k] find_vmap_area 7.40% cat [kernel.kallsyms] [k] __slab_free 7.31% cat [kernel.kallsyms] [k] vunmap_page_range 6.84% cat [kernel.kallsyms] [k] __free_pages 6.01% cat [kernel.kallsyms] [k] __vunmap 3.98% cat [kernel.kallsyms] [k] smp_call_function_single 3.81% cat [kernel.kallsyms] [k] __purge_vmap_area_lazy 2.77% cat [kernel.kallsyms] [k] __free_vmap_area Signed-off-by: Roman Gushchin <guro@fb.com> Cc: Johannes Weiner <hannes@cmpxchg.org> Cc: Matthew Wilcox <willy@infradead.org> Cc: Vlastimil Babka <vbabka@suse.cz> --- mm/vmalloc.c | 52 +++++++++++++++++++++++++++++----------------------- 1 file changed, 29 insertions(+), 23 deletions(-)