mirror of
https://github.com/torvalds/linux.git
synced 2024-12-14 15:13:52 +00:00
2139f74248
PMUs advertise their capabilities via sysfs attribute files but the perf tool currently parses only core(CPU) or hybrid core PMU capabilities. Add support of recording non-core PMU capabilities int perf.data header. Note that a newly proposed HEADER_PMU_CAPS is replacing existing HEADER_HYBRID_CPU_PMU_CAPS. Special care is taken for hybrid core PMUs by writing their capabilities first in the perf.data header to make sure new perf.data file being read by old perf tool does not break. Reviewed-by: Kan Liang <kan.liang@linux.intel.com> Signed-off-by: Ravi Bangoria <ravi.bangoria@amd.com> Acked-by: Namhyung Kim <namhyung@kernel.org> Cc: Ananth Narayan <ananth.narayan@amd.com> Cc: Andi Kleen <ak@linux.intel.com> Cc: Borislav Petkov <bp@alien8.de> Cc: Ian Rogers <irogers@google.com> Cc: Ingo Molnar <mingo@redhat.com> Cc: James Clark <james.clark@arm.com> Cc: Jiri Olsa <jolsa@kernel.org> Cc: Kim Phillips <kim.phillips@amd.com> Cc: Leo Yan <leo.yan@linaro.org> Cc: Mark Rutland <mark.rutland@arm.com> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Robert Richter <rrichter@amd.com> Cc: Sandipan Das <sandipan.das@amd.com> Cc: Santosh Shukla <santosh.shukla@amd.com> Cc: Stephane Eranian <eranian@google.com> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: like.xu.linux@gmail.com Cc: x86@kernel.org Link: https://lore.kernel.org/r/20220604044519.594-6-ravi.bangoria@amd.com Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
178 lines
4.0 KiB
C
178 lines
4.0 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef __PERF_ENV_H
|
|
#define __PERF_ENV_H
|
|
|
|
#include <linux/types.h>
|
|
#include <linux/rbtree.h>
|
|
#include "cpumap.h"
|
|
#include "rwsem.h"
|
|
|
|
struct perf_cpu_map;
|
|
|
|
struct cpu_topology_map {
|
|
int socket_id;
|
|
int die_id;
|
|
int core_id;
|
|
};
|
|
|
|
struct cpu_cache_level {
|
|
u32 level;
|
|
u32 line_size;
|
|
u32 sets;
|
|
u32 ways;
|
|
char *type;
|
|
char *size;
|
|
char *map;
|
|
};
|
|
|
|
struct numa_node {
|
|
u32 node;
|
|
u64 mem_total;
|
|
u64 mem_free;
|
|
struct perf_cpu_map *map;
|
|
};
|
|
|
|
struct memory_node {
|
|
u64 node;
|
|
u64 size;
|
|
unsigned long *set;
|
|
};
|
|
|
|
struct hybrid_node {
|
|
char *pmu_name;
|
|
char *cpus;
|
|
};
|
|
|
|
struct pmu_caps {
|
|
int nr_caps;
|
|
unsigned int max_branches;
|
|
char **caps;
|
|
char *pmu_name;
|
|
};
|
|
|
|
struct perf_env {
|
|
char *hostname;
|
|
char *os_release;
|
|
char *version;
|
|
char *arch;
|
|
int nr_cpus_online;
|
|
int nr_cpus_avail;
|
|
char *cpu_desc;
|
|
char *cpuid;
|
|
unsigned long long total_mem;
|
|
unsigned int msr_pmu_type;
|
|
unsigned int max_branches;
|
|
int kernel_is_64_bit;
|
|
|
|
int nr_cmdline;
|
|
int nr_sibling_cores;
|
|
int nr_sibling_dies;
|
|
int nr_sibling_threads;
|
|
int nr_numa_nodes;
|
|
int nr_memory_nodes;
|
|
int nr_pmu_mappings;
|
|
int nr_groups;
|
|
int nr_cpu_pmu_caps;
|
|
int nr_hybrid_nodes;
|
|
int nr_pmus_with_caps;
|
|
char *cmdline;
|
|
const char **cmdline_argv;
|
|
char *sibling_cores;
|
|
char *sibling_dies;
|
|
char *sibling_threads;
|
|
char *pmu_mappings;
|
|
char **cpu_pmu_caps;
|
|
struct cpu_topology_map *cpu;
|
|
struct cpu_cache_level *caches;
|
|
int caches_cnt;
|
|
u32 comp_ratio;
|
|
u32 comp_ver;
|
|
u32 comp_type;
|
|
u32 comp_level;
|
|
u32 comp_mmap_len;
|
|
struct numa_node *numa_nodes;
|
|
struct memory_node *memory_nodes;
|
|
unsigned long long memory_bsize;
|
|
struct hybrid_node *hybrid_nodes;
|
|
struct pmu_caps *pmu_caps;
|
|
#ifdef HAVE_LIBBPF_SUPPORT
|
|
/*
|
|
* bpf_info_lock protects bpf rbtrees. This is needed because the
|
|
* trees are accessed by different threads in perf-top
|
|
*/
|
|
struct {
|
|
struct rw_semaphore lock;
|
|
struct rb_root infos;
|
|
u32 infos_cnt;
|
|
struct rb_root btfs;
|
|
u32 btfs_cnt;
|
|
} bpf_progs;
|
|
#endif // HAVE_LIBBPF_SUPPORT
|
|
/* same reason as above (for perf-top) */
|
|
struct {
|
|
struct rw_semaphore lock;
|
|
struct rb_root tree;
|
|
} cgroups;
|
|
|
|
/* For fast cpu to numa node lookup via perf_env__numa_node */
|
|
int *numa_map;
|
|
int nr_numa_map;
|
|
|
|
/* For real clock time reference. */
|
|
struct {
|
|
u64 tod_ns;
|
|
u64 clockid_ns;
|
|
u64 clockid_res_ns;
|
|
int clockid;
|
|
/*
|
|
* enabled is valid for report mode, and is true if above
|
|
* values are set, it's set in process_clock_data
|
|
*/
|
|
bool enabled;
|
|
} clock;
|
|
};
|
|
|
|
enum perf_compress_type {
|
|
PERF_COMP_NONE = 0,
|
|
PERF_COMP_ZSTD,
|
|
PERF_COMP_MAX
|
|
};
|
|
|
|
struct bpf_prog_info_node;
|
|
struct btf_node;
|
|
|
|
extern struct perf_env perf_env;
|
|
|
|
void perf_env__exit(struct perf_env *env);
|
|
|
|
int perf_env__kernel_is_64_bit(struct perf_env *env);
|
|
|
|
int perf_env__set_cmdline(struct perf_env *env, int argc, const char *argv[]);
|
|
|
|
int perf_env__read_cpuid(struct perf_env *env);
|
|
int perf_env__read_pmu_mappings(struct perf_env *env);
|
|
int perf_env__nr_pmu_mappings(struct perf_env *env);
|
|
const char *perf_env__pmu_mappings(struct perf_env *env);
|
|
|
|
int perf_env__read_cpu_topology_map(struct perf_env *env);
|
|
|
|
void cpu_cache_level__free(struct cpu_cache_level *cache);
|
|
|
|
const char *perf_env__arch(struct perf_env *env);
|
|
const char *perf_env__cpuid(struct perf_env *env);
|
|
const char *perf_env__raw_arch(struct perf_env *env);
|
|
int perf_env__nr_cpus_avail(struct perf_env *env);
|
|
|
|
void perf_env__init(struct perf_env *env);
|
|
void perf_env__insert_bpf_prog_info(struct perf_env *env,
|
|
struct bpf_prog_info_node *info_node);
|
|
struct bpf_prog_info_node *perf_env__find_bpf_prog_info(struct perf_env *env,
|
|
__u32 prog_id);
|
|
bool perf_env__insert_btf(struct perf_env *env, struct btf_node *btf_node);
|
|
struct btf_node *perf_env__find_btf(struct perf_env *env, __u32 btf_id);
|
|
|
|
int perf_env__numa_node(struct perf_env *env, struct perf_cpu cpu);
|
|
char *perf_env__find_pmu_cap(struct perf_env *env, const char *pmu_name,
|
|
const char *cap);
|
|
#endif /* __PERF_ENV_H */
|