Message ID | 20231129060211.1890454-10-irogers@google.com (mailing list archive) |
---|---|
State | Not Applicable |
Delegated to: | BPF |
Headers | show |
Series | Clean up libperf cpumap's empty function | expand |
Context | Check | Description |
---|---|---|
netdev/tree_selection | success | Not a local patch |
On 29/11/2023 06:02, Ian Rogers wrote: > Most uses of what was perf_cpu_map__empty but is now > perf_cpu_map__has_any_cpu_or_is_empty want to do something with the > CPU map if it contains CPUs. Replace uses of > perf_cpu_map__has_any_cpu_or_is_empty with other helpers so that CPUs > within the map can be handled. > > Signed-off-by: Ian Rogers <irogers@google.com> Reviewed-by: James Clark <james.clark@arm.com> > --- > tools/perf/builtin-c2c.c | 6 +----- > tools/perf/builtin-stat.c | 9 ++++----- > tools/perf/util/auxtrace.c | 4 ++-- > tools/perf/util/record.c | 2 +- > tools/perf/util/stat.c | 2 +- > 5 files changed, 9 insertions(+), 14 deletions(-) > > diff --git a/tools/perf/builtin-c2c.c b/tools/perf/builtin-c2c.c > index f78eea9e2153..ef7ed53a4b4e 100644 > --- a/tools/perf/builtin-c2c.c > +++ b/tools/perf/builtin-c2c.c > @@ -2319,11 +2319,7 @@ static int setup_nodes(struct perf_session *session) > > nodes[node] = set; > > - /* empty node, skip */ > - if (perf_cpu_map__has_any_cpu_or_is_empty(map)) > - continue; > - > - perf_cpu_map__for_each_cpu(cpu, idx, map) { > + perf_cpu_map__for_each_cpu_skip_any(cpu, idx, map) { > __set_bit(cpu.cpu, set); > > if (WARN_ONCE(cpu2node[cpu.cpu] != -1, "node/cpu topology bug")) > diff --git a/tools/perf/builtin-stat.c b/tools/perf/builtin-stat.c > index 3303aa20f326..f583027a0639 100644 > --- a/tools/perf/builtin-stat.c > +++ b/tools/perf/builtin-stat.c > @@ -1316,10 +1316,9 @@ static int cpu__get_cache_id_from_map(struct perf_cpu cpu, char *map) > * be the first online CPU in the cache domain else use the > * first online CPU of the cache domain as the ID. > */ > - if (perf_cpu_map__has_any_cpu_or_is_empty(cpu_map)) > + id = perf_cpu_map__min(cpu_map).cpu; > + if (id == -1) > id = cpu.cpu; > - else > - id = perf_cpu_map__cpu(cpu_map, 0).cpu; > > /* Free the perf_cpu_map used to find the cache ID */ > perf_cpu_map__put(cpu_map); > @@ -1622,7 +1621,7 @@ static int perf_stat_init_aggr_mode(void) > * taking the highest cpu number to be the size of > * the aggregation translate cpumap. > */ > - if (!perf_cpu_map__has_any_cpu_or_is_empty(evsel_list->core.user_requested_cpus)) > + if (!perf_cpu_map__is_any_cpu_or_is_empty(evsel_list->core.user_requested_cpus)) > nr = perf_cpu_map__max(evsel_list->core.user_requested_cpus).cpu; > else > nr = 0; > @@ -2289,7 +2288,7 @@ int process_stat_config_event(struct perf_session *session, > > perf_event__read_stat_config(&stat_config, &event->stat_config); > > - if (perf_cpu_map__has_any_cpu_or_is_empty(st->cpus)) { > + if (perf_cpu_map__is_empty(st->cpus)) { > if (st->aggr_mode != AGGR_UNSET) > pr_warning("warning: processing task data, aggregation mode not set\n"); > } else if (st->aggr_mode != AGGR_UNSET) { > diff --git a/tools/perf/util/auxtrace.c b/tools/perf/util/auxtrace.c > index 3684e6009b63..6b1d4bafad59 100644 > --- a/tools/perf/util/auxtrace.c > +++ b/tools/perf/util/auxtrace.c > @@ -174,7 +174,7 @@ void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp, > struct evlist *evlist, > struct evsel *evsel, int idx) > { > - bool per_cpu = !perf_cpu_map__has_any_cpu_or_is_empty(evlist->core.user_requested_cpus); > + bool per_cpu = !perf_cpu_map__has_any_cpu(evlist->core.user_requested_cpus); > > mp->mmap_needed = evsel->needs_auxtrace_mmap; > > @@ -648,7 +648,7 @@ int auxtrace_parse_snapshot_options(struct auxtrace_record *itr, > > static int evlist__enable_event_idx(struct evlist *evlist, struct evsel *evsel, int idx) > { > - bool per_cpu_mmaps = !perf_cpu_map__has_any_cpu_or_is_empty(evlist->core.user_requested_cpus); > + bool per_cpu_mmaps = !perf_cpu_map__has_any_cpu(evlist->core.user_requested_cpus); > > if (per_cpu_mmaps) { > struct perf_cpu evlist_cpu = perf_cpu_map__cpu(evlist->core.all_cpus, idx); > diff --git a/tools/perf/util/record.c b/tools/perf/util/record.c > index 87e817b3cf7e..e867de8ddaaa 100644 > --- a/tools/perf/util/record.c > +++ b/tools/perf/util/record.c > @@ -237,7 +237,7 @@ bool evlist__can_select_event(struct evlist *evlist, const char *str) > > evsel = evlist__last(temp_evlist); > > - if (!evlist || perf_cpu_map__has_any_cpu_or_is_empty(evlist->core.user_requested_cpus)) { > + if (!evlist || perf_cpu_map__is_any_cpu_or_is_empty(evlist->core.user_requested_cpus)) { > struct perf_cpu_map *cpus = perf_cpu_map__new_online_cpus(); > > if (cpus) > diff --git a/tools/perf/util/stat.c b/tools/perf/util/stat.c > index 012c4946b9c4..915808a6211a 100644 > --- a/tools/perf/util/stat.c > +++ b/tools/perf/util/stat.c > @@ -315,7 +315,7 @@ static int check_per_pkg(struct evsel *counter, struct perf_counts_values *vals, > if (!counter->per_pkg) > return 0; > > - if (perf_cpu_map__has_any_cpu_or_is_empty(cpus)) > + if (perf_cpu_map__is_any_cpu_or_is_empty(cpus)) > return 0; > > if (!mask) {
diff --git a/tools/perf/builtin-c2c.c b/tools/perf/builtin-c2c.c index f78eea9e2153..ef7ed53a4b4e 100644 --- a/tools/perf/builtin-c2c.c +++ b/tools/perf/builtin-c2c.c @@ -2319,11 +2319,7 @@ static int setup_nodes(struct perf_session *session) nodes[node] = set; - /* empty node, skip */ - if (perf_cpu_map__has_any_cpu_or_is_empty(map)) - continue; - - perf_cpu_map__for_each_cpu(cpu, idx, map) { + perf_cpu_map__for_each_cpu_skip_any(cpu, idx, map) { __set_bit(cpu.cpu, set); if (WARN_ONCE(cpu2node[cpu.cpu] != -1, "node/cpu topology bug")) diff --git a/tools/perf/builtin-stat.c b/tools/perf/builtin-stat.c index 3303aa20f326..f583027a0639 100644 --- a/tools/perf/builtin-stat.c +++ b/tools/perf/builtin-stat.c @@ -1316,10 +1316,9 @@ static int cpu__get_cache_id_from_map(struct perf_cpu cpu, char *map) * be the first online CPU in the cache domain else use the * first online CPU of the cache domain as the ID. */ - if (perf_cpu_map__has_any_cpu_or_is_empty(cpu_map)) + id = perf_cpu_map__min(cpu_map).cpu; + if (id == -1) id = cpu.cpu; - else - id = perf_cpu_map__cpu(cpu_map, 0).cpu; /* Free the perf_cpu_map used to find the cache ID */ perf_cpu_map__put(cpu_map); @@ -1622,7 +1621,7 @@ static int perf_stat_init_aggr_mode(void) * taking the highest cpu number to be the size of * the aggregation translate cpumap. */ - if (!perf_cpu_map__has_any_cpu_or_is_empty(evsel_list->core.user_requested_cpus)) + if (!perf_cpu_map__is_any_cpu_or_is_empty(evsel_list->core.user_requested_cpus)) nr = perf_cpu_map__max(evsel_list->core.user_requested_cpus).cpu; else nr = 0; @@ -2289,7 +2288,7 @@ int process_stat_config_event(struct perf_session *session, perf_event__read_stat_config(&stat_config, &event->stat_config); - if (perf_cpu_map__has_any_cpu_or_is_empty(st->cpus)) { + if (perf_cpu_map__is_empty(st->cpus)) { if (st->aggr_mode != AGGR_UNSET) pr_warning("warning: processing task data, aggregation mode not set\n"); } else if (st->aggr_mode != AGGR_UNSET) { diff --git a/tools/perf/util/auxtrace.c b/tools/perf/util/auxtrace.c index 3684e6009b63..6b1d4bafad59 100644 --- a/tools/perf/util/auxtrace.c +++ b/tools/perf/util/auxtrace.c @@ -174,7 +174,7 @@ void auxtrace_mmap_params__set_idx(struct auxtrace_mmap_params *mp, struct evlist *evlist, struct evsel *evsel, int idx) { - bool per_cpu = !perf_cpu_map__has_any_cpu_or_is_empty(evlist->core.user_requested_cpus); + bool per_cpu = !perf_cpu_map__has_any_cpu(evlist->core.user_requested_cpus); mp->mmap_needed = evsel->needs_auxtrace_mmap; @@ -648,7 +648,7 @@ int auxtrace_parse_snapshot_options(struct auxtrace_record *itr, static int evlist__enable_event_idx(struct evlist *evlist, struct evsel *evsel, int idx) { - bool per_cpu_mmaps = !perf_cpu_map__has_any_cpu_or_is_empty(evlist->core.user_requested_cpus); + bool per_cpu_mmaps = !perf_cpu_map__has_any_cpu(evlist->core.user_requested_cpus); if (per_cpu_mmaps) { struct perf_cpu evlist_cpu = perf_cpu_map__cpu(evlist->core.all_cpus, idx); diff --git a/tools/perf/util/record.c b/tools/perf/util/record.c index 87e817b3cf7e..e867de8ddaaa 100644 --- a/tools/perf/util/record.c +++ b/tools/perf/util/record.c @@ -237,7 +237,7 @@ bool evlist__can_select_event(struct evlist *evlist, const char *str) evsel = evlist__last(temp_evlist); - if (!evlist || perf_cpu_map__has_any_cpu_or_is_empty(evlist->core.user_requested_cpus)) { + if (!evlist || perf_cpu_map__is_any_cpu_or_is_empty(evlist->core.user_requested_cpus)) { struct perf_cpu_map *cpus = perf_cpu_map__new_online_cpus(); if (cpus) diff --git a/tools/perf/util/stat.c b/tools/perf/util/stat.c index 012c4946b9c4..915808a6211a 100644 --- a/tools/perf/util/stat.c +++ b/tools/perf/util/stat.c @@ -315,7 +315,7 @@ static int check_per_pkg(struct evsel *counter, struct perf_counts_values *vals, if (!counter->per_pkg) return 0; - if (perf_cpu_map__has_any_cpu_or_is_empty(cpus)) + if (perf_cpu_map__is_any_cpu_or_is_empty(cpus)) return 0; if (!mask) {
Most uses of what was perf_cpu_map__empty but is now perf_cpu_map__has_any_cpu_or_is_empty want to do something with the CPU map if it contains CPUs. Replace uses of perf_cpu_map__has_any_cpu_or_is_empty with other helpers so that CPUs within the map can be handled. Signed-off-by: Ian Rogers <irogers@google.com> --- tools/perf/builtin-c2c.c | 6 +----- tools/perf/builtin-stat.c | 9 ++++----- tools/perf/util/auxtrace.c | 4 ++-- tools/perf/util/record.c | 2 +- tools/perf/util/stat.c | 2 +- 5 files changed, 9 insertions(+), 14 deletions(-)