diff mbox series

[3/4] perf lock contention: Resolve slab object name using BPF

Message ID 20241105172635.2463800-4-namhyung@kernel.org (mailing list archive)
State New
Headers show
Series perf lock contention: Symbolize locks using slab cache names (v1) | expand

Checks

Context Check Description
netdev/tree_selection success Not a local patch

Commit Message

Namhyung Kim Nov. 5, 2024, 5:26 p.m. UTC
The bpf_get_kmem_cache() kfunc can return an address of the slab cache
(kmem_cache).  As it has the name of the slab cache from the iterator,
we can use it to symbolize some dynamic kernel locks in a slab.

Before:
  root@virtme-ng:/home/namhyung/project/linux# tools/perf/perf lock con -abl sleep 1
   contended   total wait     max wait     avg wait            address   symbol

           2      3.34 us      2.87 us      1.67 us   ffff9d7800ad9600    (mutex)
           2      2.16 us      1.93 us      1.08 us   ffff9d7804b992d8    (mutex)
           4      1.37 us       517 ns       343 ns   ffff9d78036e6e00    (mutex)
           1      1.27 us      1.27 us      1.27 us   ffff9d7804b99378    (mutex)
           2       845 ns       599 ns       422 ns   ffffffff9e1c3620   delayed_uprobe_lock (mutex)
           1       845 ns       845 ns       845 ns   ffffffff9da0b280   jiffies_lock (spinlock)
           2       377 ns       259 ns       188 ns   ffffffff9e1cf840   pcpu_alloc_mutex (mutex)
           1       305 ns       305 ns       305 ns   ffffffff9e1b4cf8   tracepoint_srcu_srcu_usage (mutex)
           1       295 ns       295 ns       295 ns   ffffffff9e1c0940   pack_mutex (mutex)
           1       232 ns       232 ns       232 ns   ffff9d7804b7d8d8    (mutex)
           1       180 ns       180 ns       180 ns   ffffffff9e1b4c28   tracepoint_srcu_srcu_usage (mutex)
           1       165 ns       165 ns       165 ns   ffffffff9da8b3a0   text_mutex (mutex)

After:
  root@virtme-ng:/home/namhyung/project/linux# tools/perf/perf lock con -abl sleep 1
   contended   total wait     max wait     avg wait            address   symbol

           2      1.95 us      1.77 us       975 ns   ffff9d5e852d3498   &task_struct (mutex)
           1      1.18 us      1.18 us      1.18 us   ffff9d5e852d3538   &task_struct (mutex)
           4      1.12 us       354 ns       279 ns   ffff9d5e841ca800   &kmalloc-cg-512 (mutex)
           2       859 ns       617 ns       429 ns   ffffffffa41c3620   delayed_uprobe_lock (mutex)
           3       691 ns       388 ns       230 ns   ffffffffa41c0940   pack_mutex (mutex)
           3       421 ns       164 ns       140 ns   ffffffffa3a8b3a0   text_mutex (mutex)
           1       409 ns       409 ns       409 ns   ffffffffa41b4cf8   tracepoint_srcu_srcu_usage (mutex)
           2       362 ns       239 ns       181 ns   ffffffffa41cf840   pcpu_alloc_mutex (mutex)
           1       220 ns       220 ns       220 ns   ffff9d5e82b534d8   &signal_cache (mutex)
           1       215 ns       215 ns       215 ns   ffffffffa41b4c28   tracepoint_srcu_srcu_usage (mutex)

Note that the name starts with '&' sign for slab objects to inform they
are dynamic locks.  It won't give the accurate lock or type names but
it's still useful.  We may add type info to the slab cache later to get
the exact name of the lock in the type later.

Signed-off-by: Namhyung Kim <namhyung@kernel.org>
---
 tools/perf/util/bpf_lock_contention.c         | 52 +++++++++++++++++++
 .../perf/util/bpf_skel/lock_contention.bpf.c  | 21 +++++++-
 2 files changed, 71 insertions(+), 2 deletions(-)

Comments

Ian Rogers Nov. 5, 2024, 5:41 p.m. UTC | #1
On Tue, Nov 5, 2024 at 9:26 AM Namhyung Kim <namhyung@kernel.org> wrote:
>
> The bpf_get_kmem_cache() kfunc can return an address of the slab cache
> (kmem_cache).  As it has the name of the slab cache from the iterator,
> we can use it to symbolize some dynamic kernel locks in a slab.
>
> Before:
>   root@virtme-ng:/home/namhyung/project/linux# tools/perf/perf lock con -abl sleep 1
>    contended   total wait     max wait     avg wait            address   symbol
>
>            2      3.34 us      2.87 us      1.67 us   ffff9d7800ad9600    (mutex)
>            2      2.16 us      1.93 us      1.08 us   ffff9d7804b992d8    (mutex)
>            4      1.37 us       517 ns       343 ns   ffff9d78036e6e00    (mutex)
>            1      1.27 us      1.27 us      1.27 us   ffff9d7804b99378    (mutex)
>            2       845 ns       599 ns       422 ns   ffffffff9e1c3620   delayed_uprobe_lock (mutex)
>            1       845 ns       845 ns       845 ns   ffffffff9da0b280   jiffies_lock (spinlock)
>            2       377 ns       259 ns       188 ns   ffffffff9e1cf840   pcpu_alloc_mutex (mutex)
>            1       305 ns       305 ns       305 ns   ffffffff9e1b4cf8   tracepoint_srcu_srcu_usage (mutex)
>            1       295 ns       295 ns       295 ns   ffffffff9e1c0940   pack_mutex (mutex)
>            1       232 ns       232 ns       232 ns   ffff9d7804b7d8d8    (mutex)
>            1       180 ns       180 ns       180 ns   ffffffff9e1b4c28   tracepoint_srcu_srcu_usage (mutex)
>            1       165 ns       165 ns       165 ns   ffffffff9da8b3a0   text_mutex (mutex)
>
> After:
>   root@virtme-ng:/home/namhyung/project/linux# tools/perf/perf lock con -abl sleep 1
>    contended   total wait     max wait     avg wait            address   symbol
>
>            2      1.95 us      1.77 us       975 ns   ffff9d5e852d3498   &task_struct (mutex)
>            1      1.18 us      1.18 us      1.18 us   ffff9d5e852d3538   &task_struct (mutex)
>            4      1.12 us       354 ns       279 ns   ffff9d5e841ca800   &kmalloc-cg-512 (mutex)
>            2       859 ns       617 ns       429 ns   ffffffffa41c3620   delayed_uprobe_lock (mutex)
>            3       691 ns       388 ns       230 ns   ffffffffa41c0940   pack_mutex (mutex)
>            3       421 ns       164 ns       140 ns   ffffffffa3a8b3a0   text_mutex (mutex)
>            1       409 ns       409 ns       409 ns   ffffffffa41b4cf8   tracepoint_srcu_srcu_usage (mutex)
>            2       362 ns       239 ns       181 ns   ffffffffa41cf840   pcpu_alloc_mutex (mutex)
>            1       220 ns       220 ns       220 ns   ffff9d5e82b534d8   &signal_cache (mutex)
>            1       215 ns       215 ns       215 ns   ffffffffa41b4c28   tracepoint_srcu_srcu_usage (mutex)
>
> Note that the name starts with '&' sign for slab objects to inform they
> are dynamic locks.  It won't give the accurate lock or type names but
> it's still useful.  We may add type info to the slab cache later to get
> the exact name of the lock in the type later.

Many variables may reference a lock through a pointer, should the name
not be associated with the lock or from decoding the task_struct?
The '&' looks redundant as the addresses are clearly different.
How are >1 lock/mutex in the same struct handled?

Thanks,
Ian

> Signed-off-by: Namhyung Kim <namhyung@kernel.org>
> ---
>  tools/perf/util/bpf_lock_contention.c         | 52 +++++++++++++++++++
>  .../perf/util/bpf_skel/lock_contention.bpf.c  | 21 +++++++-
>  2 files changed, 71 insertions(+), 2 deletions(-)
>
> diff --git a/tools/perf/util/bpf_lock_contention.c b/tools/perf/util/bpf_lock_contention.c
> index a2efd40897bad316..50c3039c647d4d77 100644
> --- a/tools/perf/util/bpf_lock_contention.c
> +++ b/tools/perf/util/bpf_lock_contention.c
> @@ -2,6 +2,7 @@
>  #include "util/cgroup.h"
>  #include "util/debug.h"
>  #include "util/evlist.h"
> +#include "util/hashmap.h"
>  #include "util/machine.h"
>  #include "util/map.h"
>  #include "util/symbol.h"
> @@ -20,12 +21,25 @@
>
>  static struct lock_contention_bpf *skel;
>  static bool has_slab_iter;
> +static struct hashmap slab_hash;
> +
> +static size_t slab_cache_hash(long key, void *ctx __maybe_unused)
> +{
> +       return key;
> +}
> +
> +static bool slab_cache_equal(long key1, long key2, void *ctx __maybe_unused)
> +{
> +       return key1 == key2;
> +}
>
>  static void check_slab_cache_iter(struct lock_contention *con)
>  {
>         struct btf *btf = btf__load_vmlinux_btf();
>         s32 ret;
>
> +       hashmap__init(&slab_hash, slab_cache_hash, slab_cache_equal, /*ctx=*/NULL);
> +
>         ret = libbpf_get_error(btf);
>         if (ret) {
>                 pr_debug("BTF loading failed: %d\n", ret);
> @@ -50,6 +64,7 @@ static void run_slab_cache_iter(void)
>  {
>         int fd;
>         char buf[256];
> +       long key, *prev_key;
>
>         if (!has_slab_iter)
>                 return;
> @@ -65,6 +80,34 @@ static void run_slab_cache_iter(void)
>                 continue;
>
>         close(fd);
> +
> +       /* Read the slab cache map and build a hash with IDs */
> +       fd = bpf_map__fd(skel->maps.slab_caches);
> +       prev_key = NULL;
> +       while (!bpf_map_get_next_key(fd, prev_key, &key)) {
> +               struct slab_cache_data *data;
> +
> +               data = malloc(sizeof(*data));
> +               if (data == NULL)
> +                       break;
> +
> +               if (bpf_map_lookup_elem(fd, &key, data) < 0)
> +                       break;
> +
> +               hashmap__add(&slab_hash, data->id, data);
> +               prev_key = &key;
> +       }
> +}
> +
> +static void exit_slab_cache_iter(void)
> +{
> +       struct hashmap_entry *cur;
> +       unsigned bkt;
> +
> +       hashmap__for_each_entry(&slab_hash, cur, bkt)
> +               free(cur->pvalue);
> +
> +       hashmap__clear(&slab_hash);
>  }
>
>  int lock_contention_prepare(struct lock_contention *con)
> @@ -398,6 +441,7 @@ static const char *lock_contention_get_name(struct lock_contention *con,
>
>         if (con->aggr_mode == LOCK_AGGR_ADDR) {
>                 int lock_fd = bpf_map__fd(skel->maps.lock_syms);
> +               struct slab_cache_data *slab_data;
>
>                 /* per-process locks set upper bits of the flags */
>                 if (flags & LCD_F_MMAP_LOCK)
> @@ -416,6 +460,12 @@ static const char *lock_contention_get_name(struct lock_contention *con,
>                                 return "rq_lock";
>                 }
>
> +               /* look slab_hash for dynamic locks in a slab object */
> +               if (hashmap__find(&slab_hash, flags & LCB_F_SLAB_ID_MASK, &slab_data)) {
> +                       snprintf(name_buf, sizeof(name_buf), "&%s", slab_data->name);
> +                       return name_buf;
> +               }
> +
>                 return "";
>         }
>
> @@ -590,5 +640,7 @@ int lock_contention_finish(struct lock_contention *con)
>                 cgroup__put(cgrp);
>         }
>
> +       exit_slab_cache_iter();
> +
>         return 0;
>  }
> diff --git a/tools/perf/util/bpf_skel/lock_contention.bpf.c b/tools/perf/util/bpf_skel/lock_contention.bpf.c
> index fd24ccb00faec0ba..b5bc37955560a58e 100644
> --- a/tools/perf/util/bpf_skel/lock_contention.bpf.c
> +++ b/tools/perf/util/bpf_skel/lock_contention.bpf.c
> @@ -123,6 +123,8 @@ struct mm_struct___new {
>         struct rw_semaphore mmap_lock;
>  } __attribute__((preserve_access_index));
>
> +extern struct kmem_cache *bpf_get_kmem_cache(u64 addr) __ksym __weak;
> +
>  /* control flags */
>  const volatile int has_cpu;
>  const volatile int has_task;
> @@ -496,8 +498,23 @@ int contention_end(u64 *ctx)
>                 };
>                 int err;
>
> -               if (aggr_mode == LOCK_AGGR_ADDR)
> -                       first.flags |= check_lock_type(pelem->lock, pelem->flags);
> +               if (aggr_mode == LOCK_AGGR_ADDR) {
> +                       first.flags |= check_lock_type(pelem->lock,
> +                                                      pelem->flags & LCB_F_TYPE_MASK);
> +
> +                       /* Check if it's from a slab object */
> +                       if (bpf_get_kmem_cache) {
> +                               struct kmem_cache *s;
> +                               struct slab_cache_data *d;
> +
> +                               s = bpf_get_kmem_cache(pelem->lock);
> +                               if (s != NULL) {
> +                                       d = bpf_map_lookup_elem(&slab_caches, &s);
> +                                       if (d != NULL)
> +                                               first.flags |= d->id;
> +                               }
> +                       }
> +               }
>
>                 err = bpf_map_update_elem(&lock_stat, &key, &first, BPF_NOEXIST);
>                 if (err < 0) {
> --
> 2.47.0.199.ga7371fff76-goog
>
Namhyung Kim Nov. 5, 2024, 8:45 p.m. UTC | #2
Hi Ian,

On Tue, Nov 05, 2024 at 09:41:05AM -0800, Ian Rogers wrote:
> On Tue, Nov 5, 2024 at 9:26 AM Namhyung Kim <namhyung@kernel.org> wrote:
> >
> > The bpf_get_kmem_cache() kfunc can return an address of the slab cache
> > (kmem_cache).  As it has the name of the slab cache from the iterator,
> > we can use it to symbolize some dynamic kernel locks in a slab.
> >
> > Before:
> >   root@virtme-ng:/home/namhyung/project/linux# tools/perf/perf lock con -abl sleep 1
> >    contended   total wait     max wait     avg wait            address   symbol
> >
> >            2      3.34 us      2.87 us      1.67 us   ffff9d7800ad9600    (mutex)
> >            2      2.16 us      1.93 us      1.08 us   ffff9d7804b992d8    (mutex)
> >            4      1.37 us       517 ns       343 ns   ffff9d78036e6e00    (mutex)
> >            1      1.27 us      1.27 us      1.27 us   ffff9d7804b99378    (mutex)
> >            2       845 ns       599 ns       422 ns   ffffffff9e1c3620   delayed_uprobe_lock (mutex)
> >            1       845 ns       845 ns       845 ns   ffffffff9da0b280   jiffies_lock (spinlock)
> >            2       377 ns       259 ns       188 ns   ffffffff9e1cf840   pcpu_alloc_mutex (mutex)
> >            1       305 ns       305 ns       305 ns   ffffffff9e1b4cf8   tracepoint_srcu_srcu_usage (mutex)
> >            1       295 ns       295 ns       295 ns   ffffffff9e1c0940   pack_mutex (mutex)
> >            1       232 ns       232 ns       232 ns   ffff9d7804b7d8d8    (mutex)
> >            1       180 ns       180 ns       180 ns   ffffffff9e1b4c28   tracepoint_srcu_srcu_usage (mutex)
> >            1       165 ns       165 ns       165 ns   ffffffff9da8b3a0   text_mutex (mutex)
> >
> > After:
> >   root@virtme-ng:/home/namhyung/project/linux# tools/perf/perf lock con -abl sleep 1
> >    contended   total wait     max wait     avg wait            address   symbol
> >
> >            2      1.95 us      1.77 us       975 ns   ffff9d5e852d3498   &task_struct (mutex)
> >            1      1.18 us      1.18 us      1.18 us   ffff9d5e852d3538   &task_struct (mutex)
> >            4      1.12 us       354 ns       279 ns   ffff9d5e841ca800   &kmalloc-cg-512 (mutex)
> >            2       859 ns       617 ns       429 ns   ffffffffa41c3620   delayed_uprobe_lock (mutex)
> >            3       691 ns       388 ns       230 ns   ffffffffa41c0940   pack_mutex (mutex)
> >            3       421 ns       164 ns       140 ns   ffffffffa3a8b3a0   text_mutex (mutex)
> >            1       409 ns       409 ns       409 ns   ffffffffa41b4cf8   tracepoint_srcu_srcu_usage (mutex)
> >            2       362 ns       239 ns       181 ns   ffffffffa41cf840   pcpu_alloc_mutex (mutex)
> >            1       220 ns       220 ns       220 ns   ffff9d5e82b534d8   &signal_cache (mutex)
> >            1       215 ns       215 ns       215 ns   ffffffffa41b4c28   tracepoint_srcu_srcu_usage (mutex)
> >
> > Note that the name starts with '&' sign for slab objects to inform they
> > are dynamic locks.  It won't give the accurate lock or type names but
> > it's still useful.  We may add type info to the slab cache later to get
> > the exact name of the lock in the type later.
> 
> Many variables may reference a lock through a pointer, should the name
> not be associated with the lock or from decoding the task_struct?

I'm not sure I understood you correctly.  But this only covers when the
lock variable is inside a slab object so that the address falls into the
slab pages.

> The '&' looks redundant as the addresses are clearly different.

Probably.  But sometimes users may want clear separation without looking
at the address values.  Also we want to use it in the filters and that
would need some form of indication for the slab locks.

> How are >1 lock/mutex in the same struct handled?

It cannot distinguish them for now.  It'll be possible once we can have
type info (BTF) for slab objects and a helper to tell the offset inside
the object from the given address.  With that, we could have something
like &task_struct.futex_exit_mutex or &signal_struct.cred_guard_mutex.

Thanks,
Namhyung

> 
> Thanks,
> Ian
> 
> > Signed-off-by: Namhyung Kim <namhyung@kernel.org>
> > ---
> >  tools/perf/util/bpf_lock_contention.c         | 52 +++++++++++++++++++
> >  .../perf/util/bpf_skel/lock_contention.bpf.c  | 21 +++++++-
> >  2 files changed, 71 insertions(+), 2 deletions(-)
> >
> > diff --git a/tools/perf/util/bpf_lock_contention.c b/tools/perf/util/bpf_lock_contention.c
> > index a2efd40897bad316..50c3039c647d4d77 100644
> > --- a/tools/perf/util/bpf_lock_contention.c
> > +++ b/tools/perf/util/bpf_lock_contention.c
> > @@ -2,6 +2,7 @@
> >  #include "util/cgroup.h"
> >  #include "util/debug.h"
> >  #include "util/evlist.h"
> > +#include "util/hashmap.h"
> >  #include "util/machine.h"
> >  #include "util/map.h"
> >  #include "util/symbol.h"
> > @@ -20,12 +21,25 @@
> >
> >  static struct lock_contention_bpf *skel;
> >  static bool has_slab_iter;
> > +static struct hashmap slab_hash;
> > +
> > +static size_t slab_cache_hash(long key, void *ctx __maybe_unused)
> > +{
> > +       return key;
> > +}
> > +
> > +static bool slab_cache_equal(long key1, long key2, void *ctx __maybe_unused)
> > +{
> > +       return key1 == key2;
> > +}
> >
> >  static void check_slab_cache_iter(struct lock_contention *con)
> >  {
> >         struct btf *btf = btf__load_vmlinux_btf();
> >         s32 ret;
> >
> > +       hashmap__init(&slab_hash, slab_cache_hash, slab_cache_equal, /*ctx=*/NULL);
> > +
> >         ret = libbpf_get_error(btf);
> >         if (ret) {
> >                 pr_debug("BTF loading failed: %d\n", ret);
> > @@ -50,6 +64,7 @@ static void run_slab_cache_iter(void)
> >  {
> >         int fd;
> >         char buf[256];
> > +       long key, *prev_key;
> >
> >         if (!has_slab_iter)
> >                 return;
> > @@ -65,6 +80,34 @@ static void run_slab_cache_iter(void)
> >                 continue;
> >
> >         close(fd);
> > +
> > +       /* Read the slab cache map and build a hash with IDs */
> > +       fd = bpf_map__fd(skel->maps.slab_caches);
> > +       prev_key = NULL;
> > +       while (!bpf_map_get_next_key(fd, prev_key, &key)) {
> > +               struct slab_cache_data *data;
> > +
> > +               data = malloc(sizeof(*data));
> > +               if (data == NULL)
> > +                       break;
> > +
> > +               if (bpf_map_lookup_elem(fd, &key, data) < 0)
> > +                       break;
> > +
> > +               hashmap__add(&slab_hash, data->id, data);
> > +               prev_key = &key;
> > +       }
> > +}
> > +
> > +static void exit_slab_cache_iter(void)
> > +{
> > +       struct hashmap_entry *cur;
> > +       unsigned bkt;
> > +
> > +       hashmap__for_each_entry(&slab_hash, cur, bkt)
> > +               free(cur->pvalue);
> > +
> > +       hashmap__clear(&slab_hash);
> >  }
> >
> >  int lock_contention_prepare(struct lock_contention *con)
> > @@ -398,6 +441,7 @@ static const char *lock_contention_get_name(struct lock_contention *con,
> >
> >         if (con->aggr_mode == LOCK_AGGR_ADDR) {
> >                 int lock_fd = bpf_map__fd(skel->maps.lock_syms);
> > +               struct slab_cache_data *slab_data;
> >
> >                 /* per-process locks set upper bits of the flags */
> >                 if (flags & LCD_F_MMAP_LOCK)
> > @@ -416,6 +460,12 @@ static const char *lock_contention_get_name(struct lock_contention *con,
> >                                 return "rq_lock";
> >                 }
> >
> > +               /* look slab_hash for dynamic locks in a slab object */
> > +               if (hashmap__find(&slab_hash, flags & LCB_F_SLAB_ID_MASK, &slab_data)) {
> > +                       snprintf(name_buf, sizeof(name_buf), "&%s", slab_data->name);
> > +                       return name_buf;
> > +               }
> > +
> >                 return "";
> >         }
> >
> > @@ -590,5 +640,7 @@ int lock_contention_finish(struct lock_contention *con)
> >                 cgroup__put(cgrp);
> >         }
> >
> > +       exit_slab_cache_iter();
> > +
> >         return 0;
> >  }
> > diff --git a/tools/perf/util/bpf_skel/lock_contention.bpf.c b/tools/perf/util/bpf_skel/lock_contention.bpf.c
> > index fd24ccb00faec0ba..b5bc37955560a58e 100644
> > --- a/tools/perf/util/bpf_skel/lock_contention.bpf.c
> > +++ b/tools/perf/util/bpf_skel/lock_contention.bpf.c
> > @@ -123,6 +123,8 @@ struct mm_struct___new {
> >         struct rw_semaphore mmap_lock;
> >  } __attribute__((preserve_access_index));
> >
> > +extern struct kmem_cache *bpf_get_kmem_cache(u64 addr) __ksym __weak;
> > +
> >  /* control flags */
> >  const volatile int has_cpu;
> >  const volatile int has_task;
> > @@ -496,8 +498,23 @@ int contention_end(u64 *ctx)
> >                 };
> >                 int err;
> >
> > -               if (aggr_mode == LOCK_AGGR_ADDR)
> > -                       first.flags |= check_lock_type(pelem->lock, pelem->flags);
> > +               if (aggr_mode == LOCK_AGGR_ADDR) {
> > +                       first.flags |= check_lock_type(pelem->lock,
> > +                                                      pelem->flags & LCB_F_TYPE_MASK);
> > +
> > +                       /* Check if it's from a slab object */
> > +                       if (bpf_get_kmem_cache) {
> > +                               struct kmem_cache *s;
> > +                               struct slab_cache_data *d;
> > +
> > +                               s = bpf_get_kmem_cache(pelem->lock);
> > +                               if (s != NULL) {
> > +                                       d = bpf_map_lookup_elem(&slab_caches, &s);
> > +                                       if (d != NULL)
> > +                                               first.flags |= d->id;
> > +                               }
> > +                       }
> > +               }
> >
> >                 err = bpf_map_update_elem(&lock_stat, &key, &first, BPF_NOEXIST);
> >                 if (err < 0) {
> > --
> > 2.47.0.199.ga7371fff76-goog
> >
diff mbox series

Patch

diff --git a/tools/perf/util/bpf_lock_contention.c b/tools/perf/util/bpf_lock_contention.c
index a2efd40897bad316..50c3039c647d4d77 100644
--- a/tools/perf/util/bpf_lock_contention.c
+++ b/tools/perf/util/bpf_lock_contention.c
@@ -2,6 +2,7 @@ 
 #include "util/cgroup.h"
 #include "util/debug.h"
 #include "util/evlist.h"
+#include "util/hashmap.h"
 #include "util/machine.h"
 #include "util/map.h"
 #include "util/symbol.h"
@@ -20,12 +21,25 @@ 
 
 static struct lock_contention_bpf *skel;
 static bool has_slab_iter;
+static struct hashmap slab_hash;
+
+static size_t slab_cache_hash(long key, void *ctx __maybe_unused)
+{
+	return key;
+}
+
+static bool slab_cache_equal(long key1, long key2, void *ctx __maybe_unused)
+{
+	return key1 == key2;
+}
 
 static void check_slab_cache_iter(struct lock_contention *con)
 {
 	struct btf *btf = btf__load_vmlinux_btf();
 	s32 ret;
 
+	hashmap__init(&slab_hash, slab_cache_hash, slab_cache_equal, /*ctx=*/NULL);
+
 	ret = libbpf_get_error(btf);
 	if (ret) {
 		pr_debug("BTF loading failed: %d\n", ret);
@@ -50,6 +64,7 @@  static void run_slab_cache_iter(void)
 {
 	int fd;
 	char buf[256];
+	long key, *prev_key;
 
 	if (!has_slab_iter)
 		return;
@@ -65,6 +80,34 @@  static void run_slab_cache_iter(void)
 		continue;
 
 	close(fd);
+
+	/* Read the slab cache map and build a hash with IDs */
+	fd = bpf_map__fd(skel->maps.slab_caches);
+	prev_key = NULL;
+	while (!bpf_map_get_next_key(fd, prev_key, &key)) {
+		struct slab_cache_data *data;
+
+		data = malloc(sizeof(*data));
+		if (data == NULL)
+			break;
+
+		if (bpf_map_lookup_elem(fd, &key, data) < 0)
+			break;
+
+		hashmap__add(&slab_hash, data->id, data);
+		prev_key = &key;
+	}
+}
+
+static void exit_slab_cache_iter(void)
+{
+	struct hashmap_entry *cur;
+	unsigned bkt;
+
+	hashmap__for_each_entry(&slab_hash, cur, bkt)
+		free(cur->pvalue);
+
+	hashmap__clear(&slab_hash);
 }
 
 int lock_contention_prepare(struct lock_contention *con)
@@ -398,6 +441,7 @@  static const char *lock_contention_get_name(struct lock_contention *con,
 
 	if (con->aggr_mode == LOCK_AGGR_ADDR) {
 		int lock_fd = bpf_map__fd(skel->maps.lock_syms);
+		struct slab_cache_data *slab_data;
 
 		/* per-process locks set upper bits of the flags */
 		if (flags & LCD_F_MMAP_LOCK)
@@ -416,6 +460,12 @@  static const char *lock_contention_get_name(struct lock_contention *con,
 				return "rq_lock";
 		}
 
+		/* look slab_hash for dynamic locks in a slab object */
+		if (hashmap__find(&slab_hash, flags & LCB_F_SLAB_ID_MASK, &slab_data)) {
+			snprintf(name_buf, sizeof(name_buf), "&%s", slab_data->name);
+			return name_buf;
+		}
+
 		return "";
 	}
 
@@ -590,5 +640,7 @@  int lock_contention_finish(struct lock_contention *con)
 		cgroup__put(cgrp);
 	}
 
+	exit_slab_cache_iter();
+
 	return 0;
 }
diff --git a/tools/perf/util/bpf_skel/lock_contention.bpf.c b/tools/perf/util/bpf_skel/lock_contention.bpf.c
index fd24ccb00faec0ba..b5bc37955560a58e 100644
--- a/tools/perf/util/bpf_skel/lock_contention.bpf.c
+++ b/tools/perf/util/bpf_skel/lock_contention.bpf.c
@@ -123,6 +123,8 @@  struct mm_struct___new {
 	struct rw_semaphore mmap_lock;
 } __attribute__((preserve_access_index));
 
+extern struct kmem_cache *bpf_get_kmem_cache(u64 addr) __ksym __weak;
+
 /* control flags */
 const volatile int has_cpu;
 const volatile int has_task;
@@ -496,8 +498,23 @@  int contention_end(u64 *ctx)
 		};
 		int err;
 
-		if (aggr_mode == LOCK_AGGR_ADDR)
-			first.flags |= check_lock_type(pelem->lock, pelem->flags);
+		if (aggr_mode == LOCK_AGGR_ADDR) {
+			first.flags |= check_lock_type(pelem->lock,
+						       pelem->flags & LCB_F_TYPE_MASK);
+
+			/* Check if it's from a slab object */
+			if (bpf_get_kmem_cache) {
+				struct kmem_cache *s;
+				struct slab_cache_data *d;
+
+				s = bpf_get_kmem_cache(pelem->lock);
+				if (s != NULL) {
+					d = bpf_map_lookup_elem(&slab_caches, &s);
+					if (d != NULL)
+						first.flags |= d->id;
+				}
+			}
+		}
 
 		err = bpf_map_update_elem(&lock_stat, &key, &first, BPF_NOEXIST);
 		if (err < 0) {