Re: [PATCH v1 7/8] perf metrics: Wire up core_wide

From: Arnaldo Carvalho de Melo
Date: Wed Aug 31 2022 - 10:44:21 EST


Em Tue, Aug 30, 2022 at 09:48:45AM -0700, Ian Rogers escreveu:
> Pass state necessary for core_wide into the expression parser. Add
> system_wide and user_requested_cpu_list to perf_stat_config to make it
> available at display time. evlist isn't used as the
> evlist__create_maps, that computes user_requested_cpus, needs the list
> of events which is generated by the metric.
>
> Signed-off-by: Ian Rogers <irogers@xxxxxxxxxx>
> ---
> tools/perf/builtin-stat.c | 9 ++++
> tools/perf/util/expr.c | 10 ++++-
> tools/perf/util/expr.h | 4 +-
> tools/perf/util/expr.l | 6 +--
> tools/perf/util/metricgroup.c | 82 +++++++++++++++++++++++++++--------
> tools/perf/util/metricgroup.h | 2 +
> tools/perf/util/stat-shadow.c | 11 +++--
> tools/perf/util/stat.h | 2 +
> 8 files changed, 101 insertions(+), 25 deletions(-)
>
> diff --git a/tools/perf/builtin-stat.c b/tools/perf/builtin-stat.c
> index c813b1aa7d7c..0554ba6547a5 100644
> --- a/tools/perf/builtin-stat.c
> +++ b/tools/perf/builtin-stat.c
> @@ -1802,6 +1802,8 @@ static int add_default_attributes(void)
> return metricgroup__parse_groups(evsel_list, "transaction",
> stat_config.metric_no_group,
> stat_config.metric_no_merge,
> + stat_config.user_requested_cpu_list,
> + stat_config.system_wide,
> &stat_config.metric_events);
> }
>
> @@ -2435,6 +2437,10 @@ int cmd_stat(int argc, const char **argv)
> if ((stat_config.aggr_mode == AGGR_THREAD) && (target.system_wide))
> target.per_thread = true;
>
> + stat_config.system_wide = target.system_wide;
> + if (target.cpu_list)
> + stat_config.user_requested_cpu_list = strdup(target.cpu_list);

Check strdup() result?

> +
> /*
> * Metric parsing needs to be delayed as metrics may optimize events
> * knowing the target is system-wide.
> @@ -2443,6 +2449,8 @@ int cmd_stat(int argc, const char **argv)
> metricgroup__parse_groups(evsel_list, metrics,
> stat_config.metric_no_group,
> stat_config.metric_no_merge,
> + stat_config.user_requested_cpu_list,
> + stat_config.system_wide,
> &stat_config.metric_events);
> zfree(&metrics);
> }
> @@ -2633,6 +2641,7 @@ int cmd_stat(int argc, const char **argv)
> iostat_release(evsel_list);
>
> zfree(&stat_config.walltime_run);
> + zfree(&stat_config.user_requested_cpu_list);
>
> if (smi_cost && smi_reset)
> sysfs__write_int(FREEZE_ON_SMI_PATH, 0);
> diff --git a/tools/perf/util/expr.c b/tools/perf/util/expr.c
> index 8aa7dafa18b3..ce186bf663c4 100644
> --- a/tools/perf/util/expr.c
> +++ b/tools/perf/util/expr.c
> @@ -310,7 +310,9 @@ struct expr_parse_ctx *expr__ctx_new(void)
> free(ctx);
> return NULL;
> }
> + ctx->sctx.user_requested_cpu_list = NULL;
> ctx->sctx.runtime = 0;
> + ctx->sctx.system_wide = false;
>
> return ctx;
> }
> @@ -332,6 +334,7 @@ void expr__ctx_free(struct expr_parse_ctx *ctx)
> struct hashmap_entry *cur;
> size_t bkt;
>
> + free(ctx->sctx.user_requested_cpu_list);

Isn't better to use zfree(&) here?

> hashmap__for_each_entry(ctx->ids, cur, bkt) {
> free((char *)cur->key);
> free(cur->value);
> @@ -407,7 +410,7 @@ double arch_get_tsc_freq(void)
> }
> #endif
>
> -double expr__get_literal(const char *literal)
> +double expr__get_literal(const char *literal, const struct expr_scanner_ctx *ctx)
> {
> static struct cpu_topology *topology;
> double result = NAN;
> @@ -439,6 +442,11 @@ double expr__get_literal(const char *literal)
> result = smt_on(topology) ? 1.0 : 0.0;
> goto out;
> }
> + if (!strcmp("#core_wide", literal)) {
> + result = core_wide(ctx->system_wide, ctx->user_requested_cpu_list, topology)
> + ? 1.0 : 0.0;
> + goto out;
> + }
> if (!strcmp("#num_packages", literal)) {
> result = topology->package_cpus_lists;
> goto out;
> diff --git a/tools/perf/util/expr.h b/tools/perf/util/expr.h
> index de9b886ec49a..32740e4c81ef 100644
> --- a/tools/perf/util/expr.h
> +++ b/tools/perf/util/expr.h
> @@ -14,7 +14,9 @@
> struct metric_ref;
>
> struct expr_scanner_ctx {
> + char *user_requested_cpu_list;
> int runtime;
> + bool system_wide;
> };
>
> struct expr_parse_ctx {
> @@ -58,6 +60,6 @@ int expr__find_ids(const char *expr, const char *one,
>
> double expr_id_data__value(const struct expr_id_data *data);
> double expr_id_data__source_count(const struct expr_id_data *data);
> -double expr__get_literal(const char *literal);
> +double expr__get_literal(const char *literal, const struct expr_scanner_ctx *ctx);
>
> #endif
> diff --git a/tools/perf/util/expr.l b/tools/perf/util/expr.l
> index 4dc8edbfd9ce..0168a9637330 100644
> --- a/tools/perf/util/expr.l
> +++ b/tools/perf/util/expr.l
> @@ -79,11 +79,11 @@ static int str(yyscan_t scanner, int token, int runtime)
> return token;
> }
>
> -static int literal(yyscan_t scanner)
> +static int literal(yyscan_t scanner, const struct expr_scanner_ctx *sctx)
> {
> YYSTYPE *yylval = expr_get_lval(scanner);
>
> - yylval->num = expr__get_literal(expr_get_text(scanner));
> + yylval->num = expr__get_literal(expr_get_text(scanner), sctx);
> if (isnan(yylval->num))
> return EXPR_ERROR;
>
> @@ -108,7 +108,7 @@ min { return MIN; }
> if { return IF; }
> else { return ELSE; }
> source_count { return SOURCE_COUNT; }
> -{literal} { return literal(yyscanner); }
> +{literal} { return literal(yyscanner, sctx); }
> {number} { return value(yyscanner); }
> {symbol} { return str(yyscanner, ID, sctx->runtime); }
> "|" { return '|'; }
> diff --git a/tools/perf/util/metricgroup.c b/tools/perf/util/metricgroup.c
> index 9151346a16ab..f7d93dc02326 100644
> --- a/tools/perf/util/metricgroup.c
> +++ b/tools/perf/util/metricgroup.c
> @@ -22,6 +22,7 @@
> #include <linux/list_sort.h>
> #include <linux/string.h>
> #include <linux/zalloc.h>
> +#include <perf/cpumap.h>
> #include <subcmd/parse-options.h>
> #include <api/fs/fs.h>
> #include "util.h"
> @@ -192,7 +193,9 @@ static bool metricgroup__has_constraint(const struct pmu_event *pe)
> static struct metric *metric__new(const struct pmu_event *pe,
> const char *modifier,
> bool metric_no_group,
> - int runtime)
> + int runtime,
> + const char *user_requested_cpu_list,
> + bool system_wide)
> {
> struct metric *m;
>
> @@ -215,7 +218,11 @@ static struct metric *metric__new(const struct pmu_event *pe,
> }
> m->metric_expr = pe->metric_expr;
> m->metric_unit = pe->unit;
> + m->pctx->sctx.user_requested_cpu_list = NULL;
> + if (user_requested_cpu_list)
> + m->pctx->sctx.user_requested_cpu_list = strdup(user_requested_cpu_list);

Check?

> m->pctx->sctx.runtime = runtime;
> + m->pctx->sctx.system_wide = system_wide;
> m->has_constraint = metric_no_group || metricgroup__has_constraint(pe);
> m->metric_refs = NULL;
> m->evlist = NULL;
> @@ -874,6 +881,8 @@ struct metricgroup_add_iter_data {
> int *ret;
> bool *has_match;
> bool metric_no_group;
> + const char *user_requested_cpu_list;
> + bool system_wide;
> struct metric *root_metric;
> const struct visited_metric *visited;
> const struct pmu_events_table *table;
> @@ -887,6 +896,8 @@ static int add_metric(struct list_head *metric_list,
> const struct pmu_event *pe,
> const char *modifier,
> bool metric_no_group,
> + const char *user_requested_cpu_list,
> + bool system_wide,
> struct metric *root_metric,
> const struct visited_metric *visited,
> const struct pmu_events_table *table);
> @@ -899,6 +910,8 @@ static int add_metric(struct list_head *metric_list,
> * @metric_no_group: Should events written to events be grouped "{}" or
> * global. Grouping is the default but due to multiplexing the
> * user may override.
> + * @user_requested_cpu_list: Command line specified CPUs to record on.
> + * @system_wide: Are events for all processes recorded.
> * @root_metric: Metrics may reference other metrics to form a tree. In this
> * case the root_metric holds all the IDs and a list of referenced
> * metrics. When adding a root this argument is NULL.
> @@ -910,6 +923,8 @@ static int add_metric(struct list_head *metric_list,
> static int resolve_metric(struct list_head *metric_list,
> const char *modifier,
> bool metric_no_group,
> + const char *user_requested_cpu_list,
> + bool system_wide,
> struct metric *root_metric,
> const struct visited_metric *visited,
> const struct pmu_events_table *table)
> @@ -956,7 +971,8 @@ static int resolve_metric(struct list_head *metric_list,
> */
> for (i = 0; i < pending_cnt; i++) {
> ret = add_metric(metric_list, &pending[i].pe, modifier, metric_no_group,
> - root_metric, visited, table);
> + user_requested_cpu_list, system_wide, root_metric, visited,
> + table);
> if (ret)
> break;
> }
> @@ -974,6 +990,8 @@ static int resolve_metric(struct list_head *metric_list,
> * global. Grouping is the default but due to multiplexing the
> * user may override.
> * @runtime: A special argument for the parser only known at runtime.
> + * @user_requested_cpu_list: Command line specified CPUs to record on.
> + * @system_wide: Are events for all processes recorded.
> * @root_metric: Metrics may reference other metrics to form a tree. In this
> * case the root_metric holds all the IDs and a list of referenced
> * metrics. When adding a root this argument is NULL.
> @@ -987,6 +1005,8 @@ static int __add_metric(struct list_head *metric_list,
> const char *modifier,
> bool metric_no_group,
> int runtime,
> + const char *user_requested_cpu_list,
> + bool system_wide,
> struct metric *root_metric,
> const struct visited_metric *visited,
> const struct pmu_events_table *table)
> @@ -1011,7 +1031,8 @@ static int __add_metric(struct list_head *metric_list,
> * This metric is the root of a tree and may reference other
> * metrics that are added recursively.
> */
> - root_metric = metric__new(pe, modifier, metric_no_group, runtime);
> + root_metric = metric__new(pe, modifier, metric_no_group, runtime,
> + user_requested_cpu_list, system_wide);
> if (!root_metric)
> return -ENOMEM;
>
> @@ -1060,8 +1081,9 @@ static int __add_metric(struct list_head *metric_list,
> ret = -EINVAL;
> } else {
> /* Resolve referenced metrics. */
> - ret = resolve_metric(metric_list, modifier, metric_no_group, root_metric,
> - &visited_node, table);
> + ret = resolve_metric(metric_list, modifier, metric_no_group,
> + user_requested_cpu_list, system_wide,
> + root_metric, &visited_node, table);
> }
>
> if (ret) {
> @@ -1109,6 +1131,8 @@ static int add_metric(struct list_head *metric_list,
> const struct pmu_event *pe,
> const char *modifier,
> bool metric_no_group,
> + const char *user_requested_cpu_list,
> + bool system_wide,
> struct metric *root_metric,
> const struct visited_metric *visited,
> const struct pmu_events_table *table)
> @@ -1119,7 +1143,8 @@ static int add_metric(struct list_head *metric_list,
>
> if (!strstr(pe->metric_expr, "?")) {
> ret = __add_metric(metric_list, pe, modifier, metric_no_group, 0,
> - root_metric, visited, table);
> + user_requested_cpu_list, system_wide, root_metric,
> + visited, table);
> } else {
> int j, count;
>
> @@ -1132,7 +1157,8 @@ static int add_metric(struct list_head *metric_list,
>
> for (j = 0; j < count && !ret; j++)
> ret = __add_metric(metric_list, pe, modifier, metric_no_group, j,
> - root_metric, visited, table);
> + user_requested_cpu_list, system_wide,
> + root_metric, visited, table);
> }
>
> return ret;
> @@ -1149,6 +1175,7 @@ static int metricgroup__add_metric_sys_event_iter(const struct pmu_event *pe,
> return 0;
>
> ret = add_metric(d->metric_list, pe, d->modifier, d->metric_no_group,
> + d->user_requested_cpu_list, d->system_wide,
> d->root_metric, d->visited, d->table);
> if (ret)
> goto out;
> @@ -1191,7 +1218,9 @@ struct metricgroup__add_metric_data {
> struct list_head *list;
> const char *metric_name;
> const char *modifier;
> + const char *user_requested_cpu_list;
> bool metric_no_group;
> + bool system_wide;
> bool has_match;
> };
>
> @@ -1208,8 +1237,8 @@ static int metricgroup__add_metric_callback(const struct pmu_event *pe,
>
> data->has_match = true;
> ret = add_metric(data->list, pe, data->modifier, data->metric_no_group,
> - /*root_metric=*/NULL,
> - /*visited_metrics=*/NULL, table);
> + data->user_requested_cpu_list, data->system_wide,
> + /*root_metric=*/NULL, /*visited_metrics=*/NULL, table);
> }
> return ret;
> }
> @@ -1223,12 +1252,16 @@ static int metricgroup__add_metric_callback(const struct pmu_event *pe,
> * @metric_no_group: Should events written to events be grouped "{}" or
> * global. Grouping is the default but due to multiplexing the
> * user may override.
> + * @user_requested_cpu_list: Command line specified CPUs to record on.
> + * @system_wide: Are events for all processes recorded.
> * @metric_list: The list that the metric or metric group are added to.
> * @table: The table that is searched for metrics, most commonly the table for the
> * architecture perf is running upon.
> */
> static int metricgroup__add_metric(const char *metric_name, const char *modifier,
> bool metric_no_group,
> + const char *user_requested_cpu_list,
> + bool system_wide,
> struct list_head *metric_list,
> const struct pmu_events_table *table)
> {
> @@ -1242,6 +1275,8 @@ static int metricgroup__add_metric(const char *metric_name, const char *modifier
> .metric_name = metric_name,
> .modifier = modifier,
> .metric_no_group = metric_no_group,
> + .user_requested_cpu_list = user_requested_cpu_list,
> + .system_wide = system_wide,
> .has_match = false,
> };
> /*
> @@ -1263,6 +1298,8 @@ static int metricgroup__add_metric(const char *metric_name, const char *modifier
> .metric_name = metric_name,
> .modifier = modifier,
> .metric_no_group = metric_no_group,
> + .user_requested_cpu_list = user_requested_cpu_list,
> + .system_wide = system_wide,
> .has_match = &has_match,
> .ret = &ret,
> .table = table,
> @@ -1293,12 +1330,15 @@ static int metricgroup__add_metric(const char *metric_name, const char *modifier
> * @metric_no_group: Should events written to events be grouped "{}" or
> * global. Grouping is the default but due to multiplexing the
> * user may override.
> + * @user_requested_cpu_list: Command line specified CPUs to record on.
> + * @system_wide: Are events for all processes recorded.
> * @metric_list: The list that metrics are added to.
> * @table: The table that is searched for metrics, most commonly the table for the
> * architecture perf is running upon.
> */
> static int metricgroup__add_metric_list(const char *list, bool metric_no_group,
> - struct list_head *metric_list,
> + const char *user_requested_cpu_list,
> + bool system_wide, struct list_head *metric_list,
> const struct pmu_events_table *table)
> {
> char *list_itr, *list_copy, *metric_name, *modifier;
> @@ -1315,8 +1355,8 @@ static int metricgroup__add_metric_list(const char *list, bool metric_no_group,
> *modifier++ = '\0';
>
> ret = metricgroup__add_metric(metric_name, modifier,
> - metric_no_group, metric_list,
> - table);
> + metric_no_group, user_requested_cpu_list,
> + system_wide, metric_list, table);
> if (ret == -EINVAL)
> pr_err("Cannot find metric or group `%s'\n", metric_name);
>
> @@ -1505,6 +1545,8 @@ static int parse_ids(bool metric_no_merge, struct perf_pmu *fake_pmu,
> static int parse_groups(struct evlist *perf_evlist, const char *str,
> bool metric_no_group,
> bool metric_no_merge,
> + const char *user_requested_cpu_list,
> + bool system_wide,
> struct perf_pmu *fake_pmu,
> struct rblist *metric_events_list,
> const struct pmu_events_table *table)
> @@ -1518,7 +1560,8 @@ static int parse_groups(struct evlist *perf_evlist, const char *str,
> if (metric_events_list->nr_entries == 0)
> metricgroup__rblist_init(metric_events_list);
> ret = metricgroup__add_metric_list(str, metric_no_group,
> - &metric_list, table);
> + user_requested_cpu_list,
> + system_wide, &metric_list, table);
> if (ret)
> goto out;
>
> @@ -1650,6 +1693,8 @@ int metricgroup__parse_groups(struct evlist *perf_evlist,
> const char *str,
> bool metric_no_group,
> bool metric_no_merge,
> + const char *user_requested_cpu_list,
> + bool system_wide,
> struct rblist *metric_events)
> {
> const struct pmu_events_table *table = pmu_events_table__find();
> @@ -1657,8 +1702,9 @@ int metricgroup__parse_groups(struct evlist *perf_evlist,
> if (!table)
> return -EINVAL;
>
> - return parse_groups(perf_evlist, str, metric_no_group,
> - metric_no_merge, NULL, metric_events, table);
> + return parse_groups(perf_evlist, str, metric_no_group, metric_no_merge,
> + user_requested_cpu_list, system_wide,
> + /*fake_pmu=*/NULL, metric_events, table);
> }
>
> int metricgroup__parse_groups_test(struct evlist *evlist,
> @@ -1668,8 +1714,10 @@ int metricgroup__parse_groups_test(struct evlist *evlist,
> bool metric_no_merge,
> struct rblist *metric_events)
> {
> - return parse_groups(evlist, str, metric_no_group,
> - metric_no_merge, &perf_pmu__fake, metric_events, table);
> + return parse_groups(evlist, str, metric_no_group, metric_no_merge,
> + /*user_requested_cpu_list=*/NULL,
> + /*system_wide=*/false,
> + &perf_pmu__fake, metric_events, table);
> }
>
> static int metricgroup__has_metric_callback(const struct pmu_event *pe,
> diff --git a/tools/perf/util/metricgroup.h b/tools/perf/util/metricgroup.h
> index af9ceadaec0f..732d3a0d3334 100644
> --- a/tools/perf/util/metricgroup.h
> +++ b/tools/perf/util/metricgroup.h
> @@ -68,6 +68,8 @@ int metricgroup__parse_groups(struct evlist *perf_evlist,
> const char *str,
> bool metric_no_group,
> bool metric_no_merge,
> + const char *user_requested_cpu_list,
> + bool system_wide,
> struct rblist *metric_events);
> int metricgroup__parse_groups_test(struct evlist *evlist,
> const struct pmu_events_table *table,
> diff --git a/tools/perf/util/stat-shadow.c b/tools/perf/util/stat-shadow.c
> index 1439acd109db..1deb75741df4 100644
> --- a/tools/perf/util/stat-shadow.c
> +++ b/tools/perf/util/stat-shadow.c
> @@ -911,7 +911,10 @@ static void generic_metric(struct perf_stat_config *config,
> if (!pctx)
> return;
>
> + if (config->user_requested_cpu_list)
> + pctx->sctx.user_requested_cpu_list = strdup(config->user_requested_cpu_list);
> pctx->sctx.runtime = runtime;
> + pctx->sctx.system_wide = config->system_wide;
> i = prepare_metric(metric_events, metric_refs, pctx, cpu_map_idx, st);
> if (i < 0) {
> expr__ctx_free(pctx);
> @@ -1304,7 +1307,8 @@ void perf_stat__print_shadow_stats(struct perf_stat_config *config,
> core_bound * 100.);
> } else if (evsel->metric_expr) {
> generic_metric(config, evsel->metric_expr, evsel->metric_events, NULL,
> - evsel->name, evsel->metric_name, NULL, 1, cpu_map_idx, out, st);
> + evsel->name, evsel->metric_name, NULL, 1,
> + cpu_map_idx, out, st);
> } else if (runtime_stat_n(st, STAT_NSECS, cpu_map_idx, &rsd) != 0) {
> char unit = ' ';
> char unit_buf[10] = "/sec";
> @@ -1329,8 +1333,9 @@ void perf_stat__print_shadow_stats(struct perf_stat_config *config,
> if (num++ > 0)
> out->new_line(config, ctxp);
> generic_metric(config, mexp->metric_expr, mexp->metric_events,
> - mexp->metric_refs, evsel->name, mexp->metric_name,
> - mexp->metric_unit, mexp->runtime, cpu_map_idx, out, st);
> + mexp->metric_refs, evsel->name, mexp->metric_name,
> + mexp->metric_unit, mexp->runtime,
> + cpu_map_idx, out, st);
> }
> }
> if (num == 0)
> diff --git a/tools/perf/util/stat.h b/tools/perf/util/stat.h
> index 668250022f8c..72713b344b79 100644
> --- a/tools/perf/util/stat.h
> +++ b/tools/perf/util/stat.h
> @@ -141,6 +141,8 @@ struct perf_stat_config {
> bool stop_read_counter;
> bool quiet;
> bool iostat_run;
> + char *user_requested_cpu_list;
> + bool system_wide;
> FILE *output;
> unsigned int interval;
> unsigned int timeout;
> --
> 2.37.2.672.g94769d06f0-goog

--

- Arnaldo