forked from Minki/linux
71c86cda75
By default, we create two hybrid cache events, one is for cpu_core, and
another is for cpu_atom. But Some hybrid hardware cache events are only
available on one CPU PMU. For example, the 'L1-dcache-load-misses' is only
available on cpu_core, while the 'L1-icache-loads' is only available on
cpu_atom. We need to remove "not supported" hybrid cache events. By
extending is_event_supported() to global API and using it to check if the
hybrid cache events are supported before being created, we can remove the
"not supported" hybrid cache events.
Before:
# ./perf stat -e L1-dcache-load-misses,L1-icache-loads -a sleep 1
Performance counter stats for 'system wide':
52,570 cpu_core/L1-dcache-load-misses/
<not supported> cpu_atom/L1-dcache-load-misses/
<not supported> cpu_core/L1-icache-loads/
1,471,817 cpu_atom/L1-icache-loads/
1.004915229 seconds time elapsed
After:
# ./perf stat -e L1-dcache-load-misses,L1-icache-loads -a sleep 1
Performance counter stats for 'system wide':
54,510 cpu_core/L1-dcache-load-misses/
1,441,286 cpu_atom/L1-icache-loads/
1.005114281 seconds time elapsed
Fixes: 30def61f64
("perf parse-events: Create two hybrid cache events")
Reported-by: Yi Ammy <ammy.yi@intel.com>
Reviewed-by: Kan Liang <kan.liang@linux.intel.com>
Signed-off-by: Xing Zhengjun <zhengjun.xing@linux.intel.com>
Acked-by: Ian Rogers <irogers@google.com>
Cc: Alexander Shishkin <alexander.shishkin@intel.com>
Cc: Andi Kleen <ak@linux.intel.com>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Jin Yao <yao.jin@linux.intel.com>
Cc: Jiri Olsa <jolsa@kernel.org>
Cc: Namhyung Kim <namhyung@kernel.org>
Cc: Peter Zijlstra <peterz@infradead.org>
Link: https://lore.kernel.org/r/20220923030013.3726410-2-zhengjun.xing@linux.intel.com
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
215 lines
5.5 KiB
C
215 lines
5.5 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#include <linux/err.h>
|
|
#include <linux/zalloc.h>
|
|
#include <errno.h>
|
|
#include <sys/types.h>
|
|
#include <sys/stat.h>
|
|
#include <fcntl.h>
|
|
#include <sys/param.h>
|
|
#include "evlist.h"
|
|
#include "evsel.h"
|
|
#include "parse-events.h"
|
|
#include "parse-events-hybrid.h"
|
|
#include "debug.h"
|
|
#include "pmu.h"
|
|
#include "pmu-hybrid.h"
|
|
#include "perf.h"
|
|
|
|
static void config_hybrid_attr(struct perf_event_attr *attr,
|
|
int type, int pmu_type)
|
|
{
|
|
/*
|
|
* attr.config layout for type PERF_TYPE_HARDWARE and
|
|
* PERF_TYPE_HW_CACHE
|
|
*
|
|
* PERF_TYPE_HARDWARE: 0xEEEEEEEE000000AA
|
|
* AA: hardware event ID
|
|
* EEEEEEEE: PMU type ID
|
|
* PERF_TYPE_HW_CACHE: 0xEEEEEEEE00DDCCBB
|
|
* BB: hardware cache ID
|
|
* CC: hardware cache op ID
|
|
* DD: hardware cache op result ID
|
|
* EEEEEEEE: PMU type ID
|
|
* If the PMU type ID is 0, the PERF_TYPE_RAW will be applied.
|
|
*/
|
|
attr->type = type;
|
|
attr->config = (attr->config & PERF_HW_EVENT_MASK) |
|
|
((__u64)pmu_type << PERF_PMU_TYPE_SHIFT);
|
|
}
|
|
|
|
static int create_event_hybrid(__u32 config_type, int *idx,
|
|
struct list_head *list,
|
|
struct perf_event_attr *attr, const char *name,
|
|
const char *metric_id,
|
|
struct list_head *config_terms,
|
|
struct perf_pmu *pmu)
|
|
{
|
|
struct evsel *evsel;
|
|
__u32 type = attr->type;
|
|
__u64 config = attr->config;
|
|
|
|
config_hybrid_attr(attr, config_type, pmu->type);
|
|
|
|
/*
|
|
* Some hybrid hardware cache events are only available on one CPU
|
|
* PMU. For example, the 'L1-dcache-load-misses' is only available
|
|
* on cpu_core, while the 'L1-icache-loads' is only available on
|
|
* cpu_atom. We need to remove "not supported" hybrid cache events.
|
|
*/
|
|
if (attr->type == PERF_TYPE_HW_CACHE
|
|
&& !is_event_supported(attr->type, attr->config))
|
|
return 0;
|
|
|
|
evsel = parse_events__add_event_hybrid(list, idx, attr, name, metric_id,
|
|
pmu, config_terms);
|
|
if (evsel) {
|
|
evsel->pmu_name = strdup(pmu->name);
|
|
if (!evsel->pmu_name)
|
|
return -ENOMEM;
|
|
} else
|
|
return -ENOMEM;
|
|
attr->type = type;
|
|
attr->config = config;
|
|
return 0;
|
|
}
|
|
|
|
static int pmu_cmp(struct parse_events_state *parse_state,
|
|
struct perf_pmu *pmu)
|
|
{
|
|
if (parse_state->evlist && parse_state->evlist->hybrid_pmu_name)
|
|
return strcmp(parse_state->evlist->hybrid_pmu_name, pmu->name);
|
|
|
|
if (parse_state->hybrid_pmu_name)
|
|
return strcmp(parse_state->hybrid_pmu_name, pmu->name);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int add_hw_hybrid(struct parse_events_state *parse_state,
|
|
struct list_head *list, struct perf_event_attr *attr,
|
|
const char *name, const char *metric_id,
|
|
struct list_head *config_terms)
|
|
{
|
|
struct perf_pmu *pmu;
|
|
int ret;
|
|
|
|
perf_pmu__for_each_hybrid_pmu(pmu) {
|
|
LIST_HEAD(terms);
|
|
|
|
if (pmu_cmp(parse_state, pmu))
|
|
continue;
|
|
|
|
copy_config_terms(&terms, config_terms);
|
|
ret = create_event_hybrid(PERF_TYPE_HARDWARE,
|
|
&parse_state->idx, list, attr, name,
|
|
metric_id, &terms, pmu);
|
|
free_config_terms(&terms);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int create_raw_event_hybrid(int *idx, struct list_head *list,
|
|
struct perf_event_attr *attr,
|
|
const char *name,
|
|
const char *metric_id,
|
|
struct list_head *config_terms,
|
|
struct perf_pmu *pmu)
|
|
{
|
|
struct evsel *evsel;
|
|
|
|
attr->type = pmu->type;
|
|
evsel = parse_events__add_event_hybrid(list, idx, attr, name, metric_id,
|
|
pmu, config_terms);
|
|
if (evsel)
|
|
evsel->pmu_name = strdup(pmu->name);
|
|
else
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int add_raw_hybrid(struct parse_events_state *parse_state,
|
|
struct list_head *list, struct perf_event_attr *attr,
|
|
const char *name, const char *metric_id,
|
|
struct list_head *config_terms)
|
|
{
|
|
struct perf_pmu *pmu;
|
|
int ret;
|
|
|
|
perf_pmu__for_each_hybrid_pmu(pmu) {
|
|
LIST_HEAD(terms);
|
|
|
|
if (pmu_cmp(parse_state, pmu))
|
|
continue;
|
|
|
|
copy_config_terms(&terms, config_terms);
|
|
ret = create_raw_event_hybrid(&parse_state->idx, list, attr,
|
|
name, metric_id, &terms, pmu);
|
|
free_config_terms(&terms);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int parse_events__add_numeric_hybrid(struct parse_events_state *parse_state,
|
|
struct list_head *list,
|
|
struct perf_event_attr *attr,
|
|
const char *name, const char *metric_id,
|
|
struct list_head *config_terms,
|
|
bool *hybrid)
|
|
{
|
|
*hybrid = false;
|
|
if (attr->type == PERF_TYPE_SOFTWARE)
|
|
return 0;
|
|
|
|
if (!perf_pmu__has_hybrid())
|
|
return 0;
|
|
|
|
*hybrid = true;
|
|
if (attr->type != PERF_TYPE_RAW) {
|
|
return add_hw_hybrid(parse_state, list, attr, name, metric_id,
|
|
config_terms);
|
|
}
|
|
|
|
return add_raw_hybrid(parse_state, list, attr, name, metric_id,
|
|
config_terms);
|
|
}
|
|
|
|
int parse_events__add_cache_hybrid(struct list_head *list, int *idx,
|
|
struct perf_event_attr *attr,
|
|
const char *name,
|
|
const char *metric_id,
|
|
struct list_head *config_terms,
|
|
bool *hybrid,
|
|
struct parse_events_state *parse_state)
|
|
{
|
|
struct perf_pmu *pmu;
|
|
int ret;
|
|
|
|
*hybrid = false;
|
|
if (!perf_pmu__has_hybrid())
|
|
return 0;
|
|
|
|
*hybrid = true;
|
|
perf_pmu__for_each_hybrid_pmu(pmu) {
|
|
LIST_HEAD(terms);
|
|
|
|
if (pmu_cmp(parse_state, pmu))
|
|
continue;
|
|
|
|
copy_config_terms(&terms, config_terms);
|
|
ret = create_event_hybrid(PERF_TYPE_HW_CACHE, idx, list,
|
|
attr, name, metric_id, &terms, pmu);
|
|
free_config_terms(&terms);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|