perf stat: Make stats work over the thread dimension
Now that we have space for thread dimension counts, let's store it. Signed-off-by: Jiri Olsa <jolsa@kernel.org> Cc: Adrian Hunter <adrian.hunter@intel.com> Cc: Andi Kleen <ak@linux.intel.com> Cc: David Ahern <dsahern@gmail.com> Cc: Namhyung Kim <namhyung@kernel.org> Cc: Peter Zijlstra <a.p.zijlstra@chello.nl> Cc: Stephane Eranian <eranian@google.com> Link: http://lkml.kernel.org/r/1435310967-14570-7-git-send-email-jolsa@kernel.org Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
This commit is contained in:
parent
a8e02324df
commit
a6fa003855
@ -166,11 +166,12 @@ static void perf_evsel__free_stat_priv(struct perf_evsel *evsel)
|
|||||||
zfree(&evsel->priv);
|
zfree(&evsel->priv);
|
||||||
}
|
}
|
||||||
|
|
||||||
static int perf_evsel__alloc_prev_raw_counts(struct perf_evsel *evsel)
|
static int perf_evsel__alloc_prev_raw_counts(struct perf_evsel *evsel,
|
||||||
|
int ncpus, int nthreads)
|
||||||
{
|
{
|
||||||
struct perf_counts *counts;
|
struct perf_counts *counts;
|
||||||
|
|
||||||
counts = perf_counts__new(perf_evsel__nr_cpus(evsel));
|
counts = perf_counts__new(ncpus, nthreads);
|
||||||
if (counts)
|
if (counts)
|
||||||
evsel->prev_raw_counts = counts;
|
evsel->prev_raw_counts = counts;
|
||||||
|
|
||||||
@ -197,11 +198,14 @@ static void perf_evlist__free_stats(struct perf_evlist *evlist)
|
|||||||
static int perf_evlist__alloc_stats(struct perf_evlist *evlist, bool alloc_raw)
|
static int perf_evlist__alloc_stats(struct perf_evlist *evlist, bool alloc_raw)
|
||||||
{
|
{
|
||||||
struct perf_evsel *evsel;
|
struct perf_evsel *evsel;
|
||||||
|
int nthreads = thread_map__nr(evsel_list->threads);
|
||||||
|
|
||||||
evlist__for_each(evlist, evsel) {
|
evlist__for_each(evlist, evsel) {
|
||||||
|
int ncpus = perf_evsel__nr_cpus(evsel);
|
||||||
|
|
||||||
if (perf_evsel__alloc_stat_priv(evsel) < 0 ||
|
if (perf_evsel__alloc_stat_priv(evsel) < 0 ||
|
||||||
perf_evsel__alloc_counts(evsel, perf_evsel__nr_cpus(evsel)) < 0 ||
|
perf_evsel__alloc_counts(evsel, ncpus, nthreads) < 0 ||
|
||||||
(alloc_raw && perf_evsel__alloc_prev_raw_counts(evsel) < 0))
|
(alloc_raw && perf_evsel__alloc_prev_raw_counts(evsel, ncpus, nthreads) < 0))
|
||||||
goto out_free;
|
goto out_free;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -294,7 +298,7 @@ static int check_per_pkg(struct perf_evsel *counter, int cpu, bool *skip)
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int read_cb(struct perf_evsel *evsel, int cpu, int thread __maybe_unused,
|
static int read_cb(struct perf_evsel *evsel, int cpu, int thread,
|
||||||
struct perf_counts_values *count)
|
struct perf_counts_values *count)
|
||||||
{
|
{
|
||||||
struct perf_counts_values *aggr = &evsel->counts->aggr;
|
struct perf_counts_values *aggr = &evsel->counts->aggr;
|
||||||
@ -314,9 +318,9 @@ static int read_cb(struct perf_evsel *evsel, int cpu, int thread __maybe_unused,
|
|||||||
case AGGR_SOCKET:
|
case AGGR_SOCKET:
|
||||||
case AGGR_NONE:
|
case AGGR_NONE:
|
||||||
if (!evsel->snapshot)
|
if (!evsel->snapshot)
|
||||||
perf_evsel__compute_deltas(evsel, cpu, count);
|
perf_evsel__compute_deltas(evsel, cpu, thread, count);
|
||||||
perf_counts_values__scale(count, scale, NULL);
|
perf_counts_values__scale(count, scale, NULL);
|
||||||
*perf_counts(evsel->counts, cpu) = *count;
|
*perf_counts(evsel->counts, cpu, thread) = *count;
|
||||||
if (aggr_mode == AGGR_NONE)
|
if (aggr_mode == AGGR_NONE)
|
||||||
perf_stat__update_shadow_stats(evsel, count->values, cpu);
|
perf_stat__update_shadow_stats(evsel, count->values, cpu);
|
||||||
break;
|
break;
|
||||||
@ -352,7 +356,7 @@ static int read_counter_aggr(struct perf_evsel *counter)
|
|||||||
return -1;
|
return -1;
|
||||||
|
|
||||||
if (!counter->snapshot)
|
if (!counter->snapshot)
|
||||||
perf_evsel__compute_deltas(counter, -1, aggr);
|
perf_evsel__compute_deltas(counter, -1, -1, aggr);
|
||||||
perf_counts_values__scale(aggr, scale, &counter->counts->scaled);
|
perf_counts_values__scale(aggr, scale, &counter->counts->scaled);
|
||||||
|
|
||||||
for (i = 0; i < 3; i++)
|
for (i = 0; i < 3; i++)
|
||||||
@ -805,9 +809,9 @@ static void print_aggr(char *prefix)
|
|||||||
s2 = aggr_get_id(evsel_list->cpus, cpu2);
|
s2 = aggr_get_id(evsel_list->cpus, cpu2);
|
||||||
if (s2 != id)
|
if (s2 != id)
|
||||||
continue;
|
continue;
|
||||||
val += perf_counts(counter->counts, cpu)->val;
|
val += perf_counts(counter->counts, cpu, 0)->val;
|
||||||
ena += perf_counts(counter->counts, cpu)->ena;
|
ena += perf_counts(counter->counts, cpu, 0)->ena;
|
||||||
run += perf_counts(counter->counts, cpu)->run;
|
run += perf_counts(counter->counts, cpu, 0)->run;
|
||||||
nr++;
|
nr++;
|
||||||
}
|
}
|
||||||
if (prefix)
|
if (prefix)
|
||||||
@ -915,9 +919,9 @@ static void print_counter(struct perf_evsel *counter, char *prefix)
|
|||||||
int cpu;
|
int cpu;
|
||||||
|
|
||||||
for (cpu = 0; cpu < perf_evsel__nr_cpus(counter); cpu++) {
|
for (cpu = 0; cpu < perf_evsel__nr_cpus(counter); cpu++) {
|
||||||
val = perf_counts(counter->counts, cpu)->val;
|
val = perf_counts(counter->counts, cpu, 0)->val;
|
||||||
ena = perf_counts(counter->counts, cpu)->ena;
|
ena = perf_counts(counter->counts, cpu, 0)->ena;
|
||||||
run = perf_counts(counter->counts, cpu)->run;
|
run = perf_counts(counter->counts, cpu, 0)->run;
|
||||||
|
|
||||||
if (prefix)
|
if (prefix)
|
||||||
fprintf(output, "%s", prefix);
|
fprintf(output, "%s", prefix);
|
||||||
|
@ -78,7 +78,7 @@ int test__openat_syscall_event_on_all_cpus(void)
|
|||||||
* we use the auto allocation it will allocate just for 1 cpu,
|
* we use the auto allocation it will allocate just for 1 cpu,
|
||||||
* as we start by cpu 0.
|
* as we start by cpu 0.
|
||||||
*/
|
*/
|
||||||
if (perf_evsel__alloc_counts(evsel, cpus->nr) < 0) {
|
if (perf_evsel__alloc_counts(evsel, cpus->nr, 1) < 0) {
|
||||||
pr_debug("perf_evsel__alloc_counts(ncpus=%d)\n", cpus->nr);
|
pr_debug("perf_evsel__alloc_counts(ncpus=%d)\n", cpus->nr);
|
||||||
goto out_close_fd;
|
goto out_close_fd;
|
||||||
}
|
}
|
||||||
@ -98,9 +98,9 @@ int test__openat_syscall_event_on_all_cpus(void)
|
|||||||
}
|
}
|
||||||
|
|
||||||
expected = nr_openat_calls + cpu;
|
expected = nr_openat_calls + cpu;
|
||||||
if (perf_counts(evsel->counts, cpu)->val != expected) {
|
if (perf_counts(evsel->counts, cpu, 0)->val != expected) {
|
||||||
pr_debug("perf_evsel__read_on_cpu: expected to intercept %d calls on cpu %d, got %" PRIu64 "\n",
|
pr_debug("perf_evsel__read_on_cpu: expected to intercept %d calls on cpu %d, got %" PRIu64 "\n",
|
||||||
expected, cpus->map[cpu], perf_counts(evsel->counts, cpu)->val);
|
expected, cpus->map[cpu], perf_counts(evsel->counts, cpu, 0)->val);
|
||||||
err = -1;
|
err = -1;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -44,9 +44,9 @@ int test__openat_syscall_event(void)
|
|||||||
goto out_close_fd;
|
goto out_close_fd;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (perf_counts(evsel->counts, 0)->val != nr_openat_calls) {
|
if (perf_counts(evsel->counts, 0, 0)->val != nr_openat_calls) {
|
||||||
pr_debug("perf_evsel__read_on_cpu: expected to intercept %d calls, got %" PRIu64 "\n",
|
pr_debug("perf_evsel__read_on_cpu: expected to intercept %d calls, got %" PRIu64 "\n",
|
||||||
nr_openat_calls, perf_counts(evsel->counts, 0)->val);
|
nr_openat_calls, perf_counts(evsel->counts, 0, 0)->val);
|
||||||
goto out_close_fd;
|
goto out_close_fd;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -898,7 +898,7 @@ void perf_evsel__delete(struct perf_evsel *evsel)
|
|||||||
free(evsel);
|
free(evsel);
|
||||||
}
|
}
|
||||||
|
|
||||||
void perf_evsel__compute_deltas(struct perf_evsel *evsel, int cpu,
|
void perf_evsel__compute_deltas(struct perf_evsel *evsel, int cpu, int thread,
|
||||||
struct perf_counts_values *count)
|
struct perf_counts_values *count)
|
||||||
{
|
{
|
||||||
struct perf_counts_values tmp;
|
struct perf_counts_values tmp;
|
||||||
@ -910,8 +910,8 @@ void perf_evsel__compute_deltas(struct perf_evsel *evsel, int cpu,
|
|||||||
tmp = evsel->prev_raw_counts->aggr;
|
tmp = evsel->prev_raw_counts->aggr;
|
||||||
evsel->prev_raw_counts->aggr = *count;
|
evsel->prev_raw_counts->aggr = *count;
|
||||||
} else {
|
} else {
|
||||||
tmp = *perf_counts(evsel->prev_raw_counts, cpu);
|
tmp = *perf_counts(evsel->prev_raw_counts, cpu, thread);
|
||||||
*perf_counts(evsel->prev_raw_counts, cpu) = *count;
|
*perf_counts(evsel->prev_raw_counts, cpu, thread) = *count;
|
||||||
}
|
}
|
||||||
|
|
||||||
count->val = count->val - tmp.val;
|
count->val = count->val - tmp.val;
|
||||||
@ -964,15 +964,15 @@ int __perf_evsel__read_on_cpu(struct perf_evsel *evsel,
|
|||||||
if (FD(evsel, cpu, thread) < 0)
|
if (FD(evsel, cpu, thread) < 0)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
if (evsel->counts == NULL && perf_evsel__alloc_counts(evsel, cpu + 1) < 0)
|
if (evsel->counts == NULL && perf_evsel__alloc_counts(evsel, cpu + 1, thread + 1) < 0)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
|
|
||||||
if (readn(FD(evsel, cpu, thread), &count, nv * sizeof(u64)) < 0)
|
if (readn(FD(evsel, cpu, thread), &count, nv * sizeof(u64)) < 0)
|
||||||
return -errno;
|
return -errno;
|
||||||
|
|
||||||
perf_evsel__compute_deltas(evsel, cpu, &count);
|
perf_evsel__compute_deltas(evsel, cpu, thread, &count);
|
||||||
perf_counts_values__scale(&count, scale, NULL);
|
perf_counts_values__scale(&count, scale, NULL);
|
||||||
*perf_counts(evsel->counts, cpu) = count;
|
*perf_counts(evsel->counts, cpu, thread) = count;
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -112,7 +112,7 @@ static inline int perf_evsel__nr_cpus(struct perf_evsel *evsel)
|
|||||||
void perf_counts_values__scale(struct perf_counts_values *count,
|
void perf_counts_values__scale(struct perf_counts_values *count,
|
||||||
bool scale, s8 *pscaled);
|
bool scale, s8 *pscaled);
|
||||||
|
|
||||||
void perf_evsel__compute_deltas(struct perf_evsel *evsel, int cpu,
|
void perf_evsel__compute_deltas(struct perf_evsel *evsel, int cpu, int thread,
|
||||||
struct perf_counts_values *count);
|
struct perf_counts_values *count);
|
||||||
|
|
||||||
int perf_evsel__object_config(size_t object_size,
|
int perf_evsel__object_config(size_t object_size,
|
||||||
|
@ -95,14 +95,14 @@ void perf_stat_evsel_id_init(struct perf_evsel *evsel)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
struct perf_counts *perf_counts__new(int ncpus)
|
struct perf_counts *perf_counts__new(int ncpus, int nthreads)
|
||||||
{
|
{
|
||||||
struct perf_counts *counts = zalloc(sizeof(*counts));
|
struct perf_counts *counts = zalloc(sizeof(*counts));
|
||||||
|
|
||||||
if (counts) {
|
if (counts) {
|
||||||
struct xyarray *cpu;
|
struct xyarray *cpu;
|
||||||
|
|
||||||
cpu = xyarray__new(ncpus, 1, sizeof(struct perf_counts_values));
|
cpu = xyarray__new(ncpus, nthreads, sizeof(struct perf_counts_values));
|
||||||
if (!cpu) {
|
if (!cpu) {
|
||||||
free(counts);
|
free(counts);
|
||||||
return NULL;
|
return NULL;
|
||||||
@ -132,9 +132,9 @@ void perf_evsel__reset_counts(struct perf_evsel *evsel)
|
|||||||
perf_counts__reset(evsel->counts);
|
perf_counts__reset(evsel->counts);
|
||||||
}
|
}
|
||||||
|
|
||||||
int perf_evsel__alloc_counts(struct perf_evsel *evsel, int ncpus)
|
int perf_evsel__alloc_counts(struct perf_evsel *evsel, int ncpus, int nthreads)
|
||||||
{
|
{
|
||||||
evsel->counts = perf_counts__new(ncpus);
|
evsel->counts = perf_counts__new(ncpus, nthreads);
|
||||||
return evsel->counts != NULL ? 0 : -ENOMEM;
|
return evsel->counts != NULL ? 0 : -ENOMEM;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -50,9 +50,9 @@ struct perf_counts {
|
|||||||
};
|
};
|
||||||
|
|
||||||
static inline struct perf_counts_values*
|
static inline struct perf_counts_values*
|
||||||
perf_counts(struct perf_counts *counts, int cpu)
|
perf_counts(struct perf_counts *counts, int cpu, int thread)
|
||||||
{
|
{
|
||||||
return xyarray__entry(counts->cpu, cpu, 0);
|
return xyarray__entry(counts->cpu, cpu, thread);
|
||||||
}
|
}
|
||||||
|
|
||||||
void update_stats(struct stats *stats, u64 val);
|
void update_stats(struct stats *stats, u64 val);
|
||||||
@ -86,10 +86,10 @@ void perf_stat__update_shadow_stats(struct perf_evsel *counter, u64 *count,
|
|||||||
void perf_stat__print_shadow_stats(FILE *out, struct perf_evsel *evsel,
|
void perf_stat__print_shadow_stats(FILE *out, struct perf_evsel *evsel,
|
||||||
double avg, int cpu, enum aggr_mode aggr);
|
double avg, int cpu, enum aggr_mode aggr);
|
||||||
|
|
||||||
struct perf_counts *perf_counts__new(int ncpus);
|
struct perf_counts *perf_counts__new(int ncpus, int nthreads);
|
||||||
void perf_counts__delete(struct perf_counts *counts);
|
void perf_counts__delete(struct perf_counts *counts);
|
||||||
|
|
||||||
void perf_evsel__reset_counts(struct perf_evsel *evsel);
|
void perf_evsel__reset_counts(struct perf_evsel *evsel);
|
||||||
int perf_evsel__alloc_counts(struct perf_evsel *evsel, int ncpus);
|
int perf_evsel__alloc_counts(struct perf_evsel *evsel, int ncpus, int nthreads);
|
||||||
void perf_evsel__free_counts(struct perf_evsel *evsel);
|
void perf_evsel__free_counts(struct perf_evsel *evsel);
|
||||||
#endif
|
#endif
|
||||||
|
Loading…
Reference in New Issue
Block a user