forked from Minki/linux
perf tools: Pass struct perf_hpp_fmt to its callbacks
Currently ->cmp, ->collapse and ->sort callbacks doesn't pass corresponding fmt. But it'll be needed by upcoming changes in perf diff command. Suggested-by: Jiri Olsa <jolsa@kernel.org> Signed-off-by: Namhyung Kim <namhyung@kernel.org> Acked-by: Jiri Olsa <jolsa@kernel.org> Cc: Ingo Molnar <mingo@kernel.org> Cc: Jiri Olsa <jolsa@redhat.com> Cc: Kan Liang <kan.liang@intel.com> Cc: Peter Zijlstra <a.p.zijlstra@chello.nl> Link: http://lkml.kernel.org/r/1420677949-6719-6-git-send-email-namhyung@kernel.org [ fix build by passing perf_hpp_fmt pointer to hist_entry__cmp_ methods ] Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
This commit is contained in:
parent
ff21cef67e
commit
87bbdf768f
@ -554,14 +554,16 @@ hist_entry__cmp_compute(struct hist_entry *left, struct hist_entry *right,
|
||||
}
|
||||
|
||||
static int64_t
|
||||
hist_entry__cmp_nop(struct hist_entry *left __maybe_unused,
|
||||
hist_entry__cmp_nop(struct perf_hpp_fmt *fmt __maybe_unused,
|
||||
struct hist_entry *left __maybe_unused,
|
||||
struct hist_entry *right __maybe_unused)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int64_t
|
||||
hist_entry__cmp_baseline(struct hist_entry *left, struct hist_entry *right)
|
||||
hist_entry__cmp_baseline(struct perf_hpp_fmt *fmt __maybe_unused,
|
||||
struct hist_entry *left, struct hist_entry *right)
|
||||
{
|
||||
if (sort_compute)
|
||||
return 0;
|
||||
@ -572,19 +574,22 @@ hist_entry__cmp_baseline(struct hist_entry *left, struct hist_entry *right)
|
||||
}
|
||||
|
||||
static int64_t
|
||||
hist_entry__cmp_delta(struct hist_entry *left, struct hist_entry *right)
|
||||
hist_entry__cmp_delta(struct perf_hpp_fmt *fmt __maybe_unused,
|
||||
struct hist_entry *left, struct hist_entry *right)
|
||||
{
|
||||
return hist_entry__cmp_compute(right, left, COMPUTE_DELTA);
|
||||
}
|
||||
|
||||
static int64_t
|
||||
hist_entry__cmp_ratio(struct hist_entry *left, struct hist_entry *right)
|
||||
hist_entry__cmp_ratio(struct perf_hpp_fmt *fmt __maybe_unused,
|
||||
struct hist_entry *left, struct hist_entry *right)
|
||||
{
|
||||
return hist_entry__cmp_compute(right, left, COMPUTE_RATIO);
|
||||
}
|
||||
|
||||
static int64_t
|
||||
hist_entry__cmp_wdiff(struct hist_entry *left, struct hist_entry *right)
|
||||
hist_entry__cmp_wdiff(struct perf_hpp_fmt *fmt __maybe_unused,
|
||||
struct hist_entry *left, struct hist_entry *right)
|
||||
{
|
||||
return hist_entry__cmp_compute(right, left, COMPUTE_WEIGHTED_DIFF);
|
||||
}
|
||||
|
@ -285,7 +285,8 @@ static int hpp__entry_##_type(struct perf_hpp_fmt *fmt, \
|
||||
}
|
||||
|
||||
#define __HPP_SORT_FN(_type, _field) \
|
||||
static int64_t hpp__sort_##_type(struct hist_entry *a, struct hist_entry *b) \
|
||||
static int64_t hpp__sort_##_type(struct perf_hpp_fmt *fmt __maybe_unused, \
|
||||
struct hist_entry *a, struct hist_entry *b) \
|
||||
{ \
|
||||
return __hpp__sort(a, b, he_get_##_field); \
|
||||
}
|
||||
@ -312,7 +313,8 @@ static int hpp__entry_##_type(struct perf_hpp_fmt *fmt, \
|
||||
}
|
||||
|
||||
#define __HPP_SORT_ACC_FN(_type, _field) \
|
||||
static int64_t hpp__sort_##_type(struct hist_entry *a, struct hist_entry *b) \
|
||||
static int64_t hpp__sort_##_type(struct perf_hpp_fmt *fmt __maybe_unused, \
|
||||
struct hist_entry *a, struct hist_entry *b) \
|
||||
{ \
|
||||
return __hpp__sort_acc(a, b, he_get_acc_##_field); \
|
||||
}
|
||||
@ -331,7 +333,8 @@ static int hpp__entry_##_type(struct perf_hpp_fmt *fmt, \
|
||||
}
|
||||
|
||||
#define __HPP_SORT_RAW_FN(_type, _field) \
|
||||
static int64_t hpp__sort_##_type(struct hist_entry *a, struct hist_entry *b) \
|
||||
static int64_t hpp__sort_##_type(struct perf_hpp_fmt *fmt __maybe_unused, \
|
||||
struct hist_entry *a, struct hist_entry *b) \
|
||||
{ \
|
||||
return __hpp__sort(a, b, he_get_raw_##_field); \
|
||||
}
|
||||
@ -361,7 +364,8 @@ HPP_PERCENT_ACC_FNS(overhead_acc, period)
|
||||
HPP_RAW_FNS(samples, nr_events)
|
||||
HPP_RAW_FNS(period, period)
|
||||
|
||||
static int64_t hpp__nop_cmp(struct hist_entry *a __maybe_unused,
|
||||
static int64_t hpp__nop_cmp(struct perf_hpp_fmt *fmt __maybe_unused,
|
||||
struct hist_entry *a __maybe_unused,
|
||||
struct hist_entry *b __maybe_unused)
|
||||
{
|
||||
return 0;
|
||||
|
@ -913,7 +913,7 @@ hist_entry__cmp(struct hist_entry *left, struct hist_entry *right)
|
||||
if (perf_hpp__should_skip(fmt))
|
||||
continue;
|
||||
|
||||
cmp = fmt->cmp(left, right);
|
||||
cmp = fmt->cmp(fmt, left, right);
|
||||
if (cmp)
|
||||
break;
|
||||
}
|
||||
@ -931,7 +931,7 @@ hist_entry__collapse(struct hist_entry *left, struct hist_entry *right)
|
||||
if (perf_hpp__should_skip(fmt))
|
||||
continue;
|
||||
|
||||
cmp = fmt->collapse(left, right);
|
||||
cmp = fmt->collapse(fmt, left, right);
|
||||
if (cmp)
|
||||
break;
|
||||
}
|
||||
@ -1061,7 +1061,7 @@ static int hist_entry__sort(struct hist_entry *a, struct hist_entry *b)
|
||||
if (perf_hpp__should_skip(fmt))
|
||||
continue;
|
||||
|
||||
cmp = fmt->sort(a, b);
|
||||
cmp = fmt->sort(fmt, a, b);
|
||||
if (cmp)
|
||||
break;
|
||||
}
|
||||
|
@ -195,9 +195,12 @@ struct perf_hpp_fmt {
|
||||
struct hist_entry *he);
|
||||
int (*entry)(struct perf_hpp_fmt *fmt, struct perf_hpp *hpp,
|
||||
struct hist_entry *he);
|
||||
int64_t (*cmp)(struct hist_entry *a, struct hist_entry *b);
|
||||
int64_t (*collapse)(struct hist_entry *a, struct hist_entry *b);
|
||||
int64_t (*sort)(struct hist_entry *a, struct hist_entry *b);
|
||||
int64_t (*cmp)(struct perf_hpp_fmt *fmt,
|
||||
struct hist_entry *a, struct hist_entry *b);
|
||||
int64_t (*collapse)(struct perf_hpp_fmt *fmt,
|
||||
struct hist_entry *a, struct hist_entry *b);
|
||||
int64_t (*sort)(struct perf_hpp_fmt *fmt,
|
||||
struct hist_entry *a, struct hist_entry *b);
|
||||
|
||||
struct list_head list;
|
||||
struct list_head sort_list;
|
||||
|
@ -1304,6 +1304,37 @@ static int __sort__hpp_entry(struct perf_hpp_fmt *fmt, struct perf_hpp *hpp,
|
||||
return hse->se->se_snprintf(he, hpp->buf, hpp->size, len);
|
||||
}
|
||||
|
||||
static int64_t __sort__hpp_cmp(struct perf_hpp_fmt *fmt,
|
||||
struct hist_entry *a, struct hist_entry *b)
|
||||
{
|
||||
struct hpp_sort_entry *hse;
|
||||
|
||||
hse = container_of(fmt, struct hpp_sort_entry, hpp);
|
||||
return hse->se->se_cmp(a, b);
|
||||
}
|
||||
|
||||
static int64_t __sort__hpp_collapse(struct perf_hpp_fmt *fmt,
|
||||
struct hist_entry *a, struct hist_entry *b)
|
||||
{
|
||||
struct hpp_sort_entry *hse;
|
||||
int64_t (*collapse_fn)(struct hist_entry *, struct hist_entry *);
|
||||
|
||||
hse = container_of(fmt, struct hpp_sort_entry, hpp);
|
||||
collapse_fn = hse->se->se_collapse ?: hse->se->se_cmp;
|
||||
return collapse_fn(a, b);
|
||||
}
|
||||
|
||||
static int64_t __sort__hpp_sort(struct perf_hpp_fmt *fmt,
|
||||
struct hist_entry *a, struct hist_entry *b)
|
||||
{
|
||||
struct hpp_sort_entry *hse;
|
||||
int64_t (*sort_fn)(struct hist_entry *, struct hist_entry *);
|
||||
|
||||
hse = container_of(fmt, struct hpp_sort_entry, hpp);
|
||||
sort_fn = hse->se->se_sort ?: hse->se->se_cmp;
|
||||
return sort_fn(a, b);
|
||||
}
|
||||
|
||||
static struct hpp_sort_entry *
|
||||
__sort_dimension__alloc_hpp(struct sort_dimension *sd)
|
||||
{
|
||||
@ -1322,9 +1353,9 @@ __sort_dimension__alloc_hpp(struct sort_dimension *sd)
|
||||
hse->hpp.entry = __sort__hpp_entry;
|
||||
hse->hpp.color = NULL;
|
||||
|
||||
hse->hpp.cmp = sd->entry->se_cmp;
|
||||
hse->hpp.collapse = sd->entry->se_collapse ? : sd->entry->se_cmp;
|
||||
hse->hpp.sort = sd->entry->se_sort ? : hse->hpp.collapse;
|
||||
hse->hpp.cmp = __sort__hpp_cmp;
|
||||
hse->hpp.collapse = __sort__hpp_collapse;
|
||||
hse->hpp.sort = __sort__hpp_sort;
|
||||
|
||||
INIT_LIST_HEAD(&hse->hpp.list);
|
||||
INIT_LIST_HEAD(&hse->hpp.sort_list);
|
||||
|
Loading…
Reference in New Issue
Block a user