2009-06-26 14:28:00 +00:00
|
|
|
/*
|
2011-01-14 03:51:58 +00:00
|
|
|
* Copyright (C) 2009-2011, Frederic Weisbecker <fweisbec@gmail.com>
|
2009-06-26 14:28:00 +00:00
|
|
|
*
|
|
|
|
* Handle the callchains from the stream in an ad-hoc radix tree and then
|
|
|
|
* sort them in an rbtree.
|
|
|
|
*
|
2009-07-01 03:35:15 +00:00
|
|
|
* Using a radix for code path provides a fast retrieval and factorizes
|
|
|
|
* memory use. Also that lets us use the paths in a hierarchical graph view.
|
|
|
|
*
|
2009-06-26 14:28:00 +00:00
|
|
|
*/
|
|
|
|
|
|
|
|
#include <stdlib.h>
|
|
|
|
#include <stdio.h>
|
|
|
|
#include <stdbool.h>
|
|
|
|
#include <errno.h>
|
2009-08-09 02:19:15 +00:00
|
|
|
#include <math.h>
|
2009-06-26 14:28:00 +00:00
|
|
|
|
2014-01-14 15:37:15 +00:00
|
|
|
#include "asm/bug.h"
|
|
|
|
|
2013-07-18 22:33:57 +00:00
|
|
|
#include "hist.h"
|
2010-05-20 15:15:33 +00:00
|
|
|
#include "util.h"
|
2014-01-14 05:25:35 +00:00
|
|
|
#include "sort.h"
|
|
|
|
#include "machine.h"
|
2009-06-26 14:28:00 +00:00
|
|
|
#include "callchain.h"
|
|
|
|
|
2012-05-31 05:43:26 +00:00
|
|
|
__thread struct callchain_cursor callchain_cursor;
|
|
|
|
|
2014-09-23 01:01:42 +00:00
|
|
|
#ifdef HAVE_DWARF_UNWIND_SUPPORT
|
|
|
|
static int get_stack_size(const char *str, unsigned long *_size)
|
|
|
|
{
|
|
|
|
char *endptr;
|
|
|
|
unsigned long size;
|
|
|
|
unsigned long max_size = round_down(USHRT_MAX, sizeof(u64));
|
|
|
|
|
|
|
|
size = strtoul(str, &endptr, 0);
|
|
|
|
|
|
|
|
do {
|
|
|
|
if (*endptr)
|
|
|
|
break;
|
|
|
|
|
|
|
|
size = round_up(size, sizeof(u64));
|
|
|
|
if (!size || size > max_size)
|
|
|
|
break;
|
|
|
|
|
|
|
|
*_size = size;
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
} while (0);
|
|
|
|
|
|
|
|
pr_err("callchain: Incorrect stack dump size (max %ld): %s\n",
|
|
|
|
max_size, str);
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
#endif /* HAVE_DWARF_UNWIND_SUPPORT */
|
|
|
|
|
|
|
|
int parse_callchain_record_opt(const char *arg)
|
|
|
|
{
|
|
|
|
char *tok, *name, *saveptr = NULL;
|
|
|
|
char *buf;
|
|
|
|
int ret = -1;
|
|
|
|
|
|
|
|
/* We need buffer that we know we can write to. */
|
|
|
|
buf = malloc(strlen(arg) + 1);
|
|
|
|
if (!buf)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
strcpy(buf, arg);
|
|
|
|
|
|
|
|
tok = strtok_r((char *)buf, ",", &saveptr);
|
|
|
|
name = tok ? : (char *)buf;
|
|
|
|
|
|
|
|
do {
|
|
|
|
/* Framepointer style */
|
|
|
|
if (!strncmp(name, "fp", sizeof("fp"))) {
|
|
|
|
if (!strtok_r(NULL, ",", &saveptr)) {
|
|
|
|
callchain_param.record_mode = CALLCHAIN_FP;
|
|
|
|
ret = 0;
|
|
|
|
} else
|
|
|
|
pr_err("callchain: No more arguments "
|
|
|
|
"needed for -g fp\n");
|
|
|
|
break;
|
|
|
|
|
|
|
|
#ifdef HAVE_DWARF_UNWIND_SUPPORT
|
|
|
|
/* Dwarf style */
|
|
|
|
} else if (!strncmp(name, "dwarf", sizeof("dwarf"))) {
|
|
|
|
const unsigned long default_stack_dump_size = 8192;
|
|
|
|
|
|
|
|
ret = 0;
|
|
|
|
callchain_param.record_mode = CALLCHAIN_DWARF;
|
|
|
|
callchain_param.dump_size = default_stack_dump_size;
|
|
|
|
|
|
|
|
tok = strtok_r(NULL, ",", &saveptr);
|
|
|
|
if (tok) {
|
|
|
|
unsigned long size = 0;
|
|
|
|
|
|
|
|
ret = get_stack_size(tok, &size);
|
|
|
|
callchain_param.dump_size = size;
|
|
|
|
}
|
|
|
|
#endif /* HAVE_DWARF_UNWIND_SUPPORT */
|
|
|
|
} else {
|
|
|
|
pr_err("callchain: Unknown --call-graph option "
|
|
|
|
"value: %s\n", arg);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
} while (0);
|
|
|
|
|
|
|
|
free(buf);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2014-09-23 01:01:43 +00:00
|
|
|
static int parse_callchain_mode(const char *value)
|
|
|
|
{
|
|
|
|
if (!strncmp(value, "graph", strlen(value))) {
|
|
|
|
callchain_param.mode = CHAIN_GRAPH_ABS;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
if (!strncmp(value, "flat", strlen(value))) {
|
|
|
|
callchain_param.mode = CHAIN_FLAT;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
if (!strncmp(value, "fractal", strlen(value))) {
|
|
|
|
callchain_param.mode = CHAIN_GRAPH_REL;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int parse_callchain_order(const char *value)
|
|
|
|
{
|
|
|
|
if (!strncmp(value, "caller", strlen(value))) {
|
|
|
|
callchain_param.order = ORDER_CALLER;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
if (!strncmp(value, "callee", strlen(value))) {
|
|
|
|
callchain_param.order = ORDER_CALLEE;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int parse_callchain_sort_key(const char *value)
|
|
|
|
{
|
|
|
|
if (!strncmp(value, "function", strlen(value))) {
|
|
|
|
callchain_param.key = CCKEY_FUNCTION;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
if (!strncmp(value, "address", strlen(value))) {
|
|
|
|
callchain_param.key = CCKEY_ADDRESS;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
2014-04-07 18:55:24 +00:00
|
|
|
int
|
|
|
|
parse_callchain_report_opt(const char *arg)
|
|
|
|
{
|
2014-08-14 06:01:38 +00:00
|
|
|
char *tok;
|
2014-04-07 18:55:24 +00:00
|
|
|
char *endptr;
|
2014-08-14 06:01:38 +00:00
|
|
|
bool minpcnt_set = false;
|
2014-04-07 18:55:24 +00:00
|
|
|
|
|
|
|
symbol_conf.use_callchain = true;
|
|
|
|
|
|
|
|
if (!arg)
|
|
|
|
return 0;
|
|
|
|
|
2014-08-14 06:01:38 +00:00
|
|
|
while ((tok = strtok((char *)arg, ",")) != NULL) {
|
|
|
|
if (!strncmp(tok, "none", strlen(tok))) {
|
|
|
|
callchain_param.mode = CHAIN_NONE;
|
|
|
|
symbol_conf.use_callchain = false;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2014-09-23 01:01:43 +00:00
|
|
|
if (!parse_callchain_mode(tok) ||
|
|
|
|
!parse_callchain_order(tok) ||
|
|
|
|
!parse_callchain_sort_key(tok)) {
|
|
|
|
/* parsing ok - move on to the next */
|
|
|
|
} else if (!minpcnt_set) {
|
|
|
|
/* try to get the min percent */
|
2014-08-14 06:01:38 +00:00
|
|
|
callchain_param.min_percent = strtod(tok, &endptr);
|
|
|
|
if (tok == endptr)
|
|
|
|
return -1;
|
|
|
|
minpcnt_set = true;
|
|
|
|
} else {
|
|
|
|
/* try print limit at last */
|
|
|
|
callchain_param.print_limit = strtoul(tok, &endptr, 0);
|
|
|
|
if (tok == endptr)
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
arg = NULL;
|
2014-04-07 18:55:24 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (callchain_register_param(&callchain_param) < 0) {
|
|
|
|
pr_err("Can't register callchain params\n");
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2014-09-23 01:01:43 +00:00
|
|
|
int perf_callchain_config(const char *var, const char *value)
|
|
|
|
{
|
|
|
|
char *endptr;
|
|
|
|
|
|
|
|
if (prefixcmp(var, "call-graph."))
|
|
|
|
return 0;
|
|
|
|
var += sizeof("call-graph.") - 1;
|
|
|
|
|
|
|
|
if (!strcmp(var, "record-mode"))
|
|
|
|
return parse_callchain_record_opt(value);
|
|
|
|
#ifdef HAVE_DWARF_UNWIND_SUPPORT
|
|
|
|
if (!strcmp(var, "dump-size")) {
|
|
|
|
unsigned long size = 0;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = get_stack_size(value, &size);
|
|
|
|
callchain_param.dump_size = size;
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
if (!strcmp(var, "print-type"))
|
|
|
|
return parse_callchain_mode(value);
|
|
|
|
if (!strcmp(var, "order"))
|
|
|
|
return parse_callchain_order(value);
|
|
|
|
if (!strcmp(var, "sort-key"))
|
|
|
|
return parse_callchain_sort_key(value);
|
|
|
|
if (!strcmp(var, "threshold")) {
|
|
|
|
callchain_param.min_percent = strtod(value, &endptr);
|
|
|
|
if (value == endptr)
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
if (!strcmp(var, "print-limit")) {
|
|
|
|
callchain_param.print_limit = strtod(value, &endptr);
|
|
|
|
if (value == endptr)
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2009-07-01 03:35:15 +00:00
|
|
|
static void
|
2009-07-02 15:58:21 +00:00
|
|
|
rb_insert_callchain(struct rb_root *root, struct callchain_node *chain,
|
|
|
|
enum chain_mode mode)
|
2009-06-26 14:28:00 +00:00
|
|
|
{
|
|
|
|
struct rb_node **p = &root->rb_node;
|
|
|
|
struct rb_node *parent = NULL;
|
|
|
|
struct callchain_node *rnode;
|
2011-01-14 03:51:59 +00:00
|
|
|
u64 chain_cumul = callchain_cumul_hits(chain);
|
2009-06-26 14:28:00 +00:00
|
|
|
|
|
|
|
while (*p) {
|
2009-08-07 05:11:05 +00:00
|
|
|
u64 rnode_cumul;
|
|
|
|
|
2009-06-26 14:28:00 +00:00
|
|
|
parent = *p;
|
|
|
|
rnode = rb_entry(parent, struct callchain_node, rb_node);
|
2011-01-14 03:51:59 +00:00
|
|
|
rnode_cumul = callchain_cumul_hits(rnode);
|
2009-06-26 14:28:00 +00:00
|
|
|
|
2009-07-02 15:58:21 +00:00
|
|
|
switch (mode) {
|
2009-07-05 05:39:21 +00:00
|
|
|
case CHAIN_FLAT:
|
2009-07-02 15:58:21 +00:00
|
|
|
if (rnode->hit < chain->hit)
|
|
|
|
p = &(*p)->rb_left;
|
|
|
|
else
|
|
|
|
p = &(*p)->rb_right;
|
|
|
|
break;
|
2009-07-05 05:39:21 +00:00
|
|
|
case CHAIN_GRAPH_ABS: /* Falldown */
|
|
|
|
case CHAIN_GRAPH_REL:
|
2009-08-07 05:11:05 +00:00
|
|
|
if (rnode_cumul < chain_cumul)
|
2009-07-02 15:58:21 +00:00
|
|
|
p = &(*p)->rb_left;
|
|
|
|
else
|
|
|
|
p = &(*p)->rb_right;
|
|
|
|
break;
|
2009-08-15 10:26:57 +00:00
|
|
|
case CHAIN_NONE:
|
2009-07-02 15:58:21 +00:00
|
|
|
default:
|
|
|
|
break;
|
|
|
|
}
|
2009-06-26 14:28:00 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
rb_link_node(&chain->rb_node, parent, p);
|
|
|
|
rb_insert_color(&chain->rb_node, root);
|
|
|
|
}
|
|
|
|
|
2009-07-05 05:39:21 +00:00
|
|
|
static void
|
|
|
|
__sort_chain_flat(struct rb_root *rb_root, struct callchain_node *node,
|
|
|
|
u64 min_hit)
|
|
|
|
{
|
2013-10-11 05:15:36 +00:00
|
|
|
struct rb_node *n;
|
2009-07-05 05:39:21 +00:00
|
|
|
struct callchain_node *child;
|
|
|
|
|
2013-10-11 05:15:36 +00:00
|
|
|
n = rb_first(&node->rb_root_in);
|
|
|
|
while (n) {
|
|
|
|
child = rb_entry(n, struct callchain_node, rb_node_in);
|
|
|
|
n = rb_next(n);
|
|
|
|
|
2009-07-05 05:39:21 +00:00
|
|
|
__sort_chain_flat(rb_root, child, min_hit);
|
2013-10-11 05:15:36 +00:00
|
|
|
}
|
2009-07-05 05:39:21 +00:00
|
|
|
|
|
|
|
if (node->hit && node->hit >= min_hit)
|
|
|
|
rb_insert_callchain(rb_root, node, CHAIN_FLAT);
|
|
|
|
}
|
|
|
|
|
2009-06-26 14:28:00 +00:00
|
|
|
/*
|
|
|
|
* Once we get every callchains from the stream, we can now
|
|
|
|
* sort them by hit
|
|
|
|
*/
|
2009-07-05 05:39:21 +00:00
|
|
|
static void
|
2010-08-22 18:05:22 +00:00
|
|
|
sort_chain_flat(struct rb_root *rb_root, struct callchain_root *root,
|
2012-09-10 22:15:03 +00:00
|
|
|
u64 min_hit, struct callchain_param *param __maybe_unused)
|
2009-07-05 05:39:21 +00:00
|
|
|
{
|
2010-08-22 18:05:22 +00:00
|
|
|
__sort_chain_flat(rb_root, &root->node, min_hit);
|
2009-07-05 05:39:21 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void __sort_chain_graph_abs(struct callchain_node *node,
|
|
|
|
u64 min_hit)
|
2009-06-26 14:28:00 +00:00
|
|
|
{
|
2013-10-11 05:15:36 +00:00
|
|
|
struct rb_node *n;
|
2009-06-26 14:28:00 +00:00
|
|
|
struct callchain_node *child;
|
|
|
|
|
2009-07-05 05:39:21 +00:00
|
|
|
node->rb_root = RB_ROOT;
|
2013-10-11 05:15:36 +00:00
|
|
|
n = rb_first(&node->rb_root_in);
|
|
|
|
|
|
|
|
while (n) {
|
|
|
|
child = rb_entry(n, struct callchain_node, rb_node_in);
|
|
|
|
n = rb_next(n);
|
2009-06-26 14:28:00 +00:00
|
|
|
|
2009-07-05 05:39:21 +00:00
|
|
|
__sort_chain_graph_abs(child, min_hit);
|
2011-01-14 03:51:59 +00:00
|
|
|
if (callchain_cumul_hits(child) >= min_hit)
|
2009-07-05 05:39:21 +00:00
|
|
|
rb_insert_callchain(&node->rb_root, child,
|
|
|
|
CHAIN_GRAPH_ABS);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
2010-08-22 18:05:22 +00:00
|
|
|
sort_chain_graph_abs(struct rb_root *rb_root, struct callchain_root *chain_root,
|
2012-09-10 22:15:03 +00:00
|
|
|
u64 min_hit, struct callchain_param *param __maybe_unused)
|
2009-07-05 05:39:21 +00:00
|
|
|
{
|
2010-08-22 18:05:22 +00:00
|
|
|
__sort_chain_graph_abs(&chain_root->node, min_hit);
|
|
|
|
rb_root->rb_node = chain_root->node.rb_root.rb_node;
|
2009-07-02 15:58:21 +00:00
|
|
|
}
|
|
|
|
|
2009-07-05 05:39:21 +00:00
|
|
|
static void __sort_chain_graph_rel(struct callchain_node *node,
|
|
|
|
double min_percent)
|
2009-07-02 15:58:21 +00:00
|
|
|
{
|
2013-10-11 05:15:36 +00:00
|
|
|
struct rb_node *n;
|
2009-07-02 15:58:21 +00:00
|
|
|
struct callchain_node *child;
|
2009-07-05 05:39:21 +00:00
|
|
|
u64 min_hit;
|
2009-07-02 15:58:21 +00:00
|
|
|
|
|
|
|
node->rb_root = RB_ROOT;
|
2009-08-09 02:19:15 +00:00
|
|
|
min_hit = ceil(node->children_hit * min_percent);
|
2009-07-02 15:58:21 +00:00
|
|
|
|
2013-10-11 05:15:36 +00:00
|
|
|
n = rb_first(&node->rb_root_in);
|
|
|
|
while (n) {
|
|
|
|
child = rb_entry(n, struct callchain_node, rb_node_in);
|
|
|
|
n = rb_next(n);
|
|
|
|
|
2009-07-05 05:39:21 +00:00
|
|
|
__sort_chain_graph_rel(child, min_percent);
|
2011-01-14 03:51:59 +00:00
|
|
|
if (callchain_cumul_hits(child) >= min_hit)
|
2009-07-05 05:39:21 +00:00
|
|
|
rb_insert_callchain(&node->rb_root, child,
|
|
|
|
CHAIN_GRAPH_REL);
|
2009-07-02 15:58:21 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2009-07-05 05:39:21 +00:00
|
|
|
static void
|
2010-08-22 18:05:22 +00:00
|
|
|
sort_chain_graph_rel(struct rb_root *rb_root, struct callchain_root *chain_root,
|
2012-09-10 22:15:03 +00:00
|
|
|
u64 min_hit __maybe_unused, struct callchain_param *param)
|
2009-07-02 15:58:21 +00:00
|
|
|
{
|
2010-08-22 18:05:22 +00:00
|
|
|
__sort_chain_graph_rel(&chain_root->node, param->min_percent / 100.0);
|
|
|
|
rb_root->rb_node = chain_root->node.rb_root.rb_node;
|
2009-06-26 14:28:00 +00:00
|
|
|
}
|
|
|
|
|
2011-01-14 03:52:00 +00:00
|
|
|
int callchain_register_param(struct callchain_param *param)
|
2009-07-05 05:39:21 +00:00
|
|
|
{
|
|
|
|
switch (param->mode) {
|
|
|
|
case CHAIN_GRAPH_ABS:
|
|
|
|
param->sort = sort_chain_graph_abs;
|
|
|
|
break;
|
|
|
|
case CHAIN_GRAPH_REL:
|
|
|
|
param->sort = sort_chain_graph_rel;
|
|
|
|
break;
|
|
|
|
case CHAIN_FLAT:
|
|
|
|
param->sort = sort_chain_flat;
|
|
|
|
break;
|
2009-08-15 10:26:57 +00:00
|
|
|
case CHAIN_NONE:
|
2009-07-05 05:39:21 +00:00
|
|
|
default:
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2009-07-01 03:35:15 +00:00
|
|
|
/*
|
|
|
|
* Create a child for a parent. If inherit_children, then the new child
|
|
|
|
* will become the new parent of it's parent children
|
|
|
|
*/
|
|
|
|
static struct callchain_node *
|
|
|
|
create_child(struct callchain_node *parent, bool inherit_children)
|
2009-06-26 14:28:00 +00:00
|
|
|
{
|
|
|
|
struct callchain_node *new;
|
|
|
|
|
2010-05-10 13:56:50 +00:00
|
|
|
new = zalloc(sizeof(*new));
|
2009-06-26 14:28:00 +00:00
|
|
|
if (!new) {
|
|
|
|
perror("not enough memory to create child for code path tree");
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
new->parent = parent;
|
|
|
|
INIT_LIST_HEAD(&new->val);
|
2009-07-01 03:35:15 +00:00
|
|
|
|
|
|
|
if (inherit_children) {
|
2013-10-11 05:15:36 +00:00
|
|
|
struct rb_node *n;
|
|
|
|
struct callchain_node *child;
|
|
|
|
|
|
|
|
new->rb_root_in = parent->rb_root_in;
|
|
|
|
parent->rb_root_in = RB_ROOT;
|
2009-07-01 03:35:15 +00:00
|
|
|
|
2013-10-11 05:15:36 +00:00
|
|
|
n = rb_first(&new->rb_root_in);
|
|
|
|
while (n) {
|
|
|
|
child = rb_entry(n, struct callchain_node, rb_node_in);
|
|
|
|
child->parent = new;
|
|
|
|
n = rb_next(n);
|
|
|
|
}
|
2009-07-01 03:35:15 +00:00
|
|
|
|
2013-10-11 05:15:36 +00:00
|
|
|
/* make it the first child */
|
|
|
|
rb_link_node(&new->rb_node_in, NULL, &parent->rb_root_in.rb_node);
|
|
|
|
rb_insert_color(&new->rb_node_in, &parent->rb_root_in);
|
2009-07-01 03:35:15 +00:00
|
|
|
}
|
2009-06-26 14:28:00 +00:00
|
|
|
|
|
|
|
return new;
|
|
|
|
}
|
|
|
|
|
2010-03-22 16:09:33 +00:00
|
|
|
|
2009-07-01 03:35:15 +00:00
|
|
|
/*
|
|
|
|
* Fill the node with callchain values
|
|
|
|
*/
|
2009-06-26 14:28:00 +00:00
|
|
|
static void
|
2011-01-14 03:51:58 +00:00
|
|
|
fill_node(struct callchain_node *node, struct callchain_cursor *cursor)
|
2009-06-26 14:28:00 +00:00
|
|
|
{
|
2011-01-14 03:51:58 +00:00
|
|
|
struct callchain_cursor_node *cursor_node;
|
|
|
|
|
|
|
|
node->val_nr = cursor->nr - cursor->pos;
|
|
|
|
if (!node->val_nr)
|
|
|
|
pr_warning("Warning: empty node in callchain tree\n");
|
2009-06-26 14:28:00 +00:00
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
cursor_node = callchain_cursor_current(cursor);
|
|
|
|
|
|
|
|
while (cursor_node) {
|
2009-06-26 14:28:00 +00:00
|
|
|
struct callchain_list *call;
|
|
|
|
|
2010-05-10 13:56:50 +00:00
|
|
|
call = zalloc(sizeof(*call));
|
2009-06-26 14:28:00 +00:00
|
|
|
if (!call) {
|
|
|
|
perror("not enough memory for the code path tree");
|
|
|
|
return;
|
|
|
|
}
|
2011-01-14 03:51:58 +00:00
|
|
|
call->ip = cursor_node->ip;
|
|
|
|
call->ms.sym = cursor_node->sym;
|
|
|
|
call->ms.map = cursor_node->map;
|
2009-06-26 14:28:00 +00:00
|
|
|
list_add_tail(&call->list, &node->val);
|
2011-01-14 03:51:58 +00:00
|
|
|
|
|
|
|
callchain_cursor_advance(cursor);
|
|
|
|
cursor_node = callchain_cursor_current(cursor);
|
2009-06-26 14:28:00 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-10-11 05:15:36 +00:00
|
|
|
static struct callchain_node *
|
2011-01-14 03:51:58 +00:00
|
|
|
add_child(struct callchain_node *parent,
|
|
|
|
struct callchain_cursor *cursor,
|
|
|
|
u64 period)
|
2009-06-26 14:28:00 +00:00
|
|
|
{
|
|
|
|
struct callchain_node *new;
|
|
|
|
|
2009-07-01 03:35:15 +00:00
|
|
|
new = create_child(parent, false);
|
2011-01-14 03:51:58 +00:00
|
|
|
fill_node(new, cursor);
|
2009-06-26 14:28:00 +00:00
|
|
|
|
2009-08-07 05:11:05 +00:00
|
|
|
new->children_hit = 0;
|
2010-07-08 01:41:46 +00:00
|
|
|
new->hit = period;
|
2013-10-11 05:15:36 +00:00
|
|
|
return new;
|
|
|
|
}
|
|
|
|
|
|
|
|
static s64 match_chain(struct callchain_cursor_node *node,
|
|
|
|
struct callchain_list *cnode)
|
|
|
|
{
|
|
|
|
struct symbol *sym = node->sym;
|
|
|
|
|
|
|
|
if (cnode->ms.sym && sym &&
|
|
|
|
callchain_param.key == CCKEY_FUNCTION)
|
|
|
|
return cnode->ms.sym->start - sym->start;
|
|
|
|
else
|
|
|
|
return cnode->ip - node->ip;
|
2009-06-26 14:28:00 +00:00
|
|
|
}
|
|
|
|
|
2009-07-01 03:35:15 +00:00
|
|
|
/*
|
|
|
|
* Split the parent in two parts (a new child is created) and
|
|
|
|
* give a part of its callchain to the created child.
|
|
|
|
* Then create another child to host the given callchain of new branch
|
|
|
|
*/
|
2009-06-26 14:28:00 +00:00
|
|
|
static void
|
2011-01-14 03:51:58 +00:00
|
|
|
split_add_child(struct callchain_node *parent,
|
|
|
|
struct callchain_cursor *cursor,
|
|
|
|
struct callchain_list *to_split,
|
|
|
|
u64 idx_parents, u64 idx_local, u64 period)
|
2009-06-26 14:28:00 +00:00
|
|
|
{
|
|
|
|
struct callchain_node *new;
|
2009-07-01 03:35:15 +00:00
|
|
|
struct list_head *old_tail;
|
2009-07-01 10:37:06 +00:00
|
|
|
unsigned int idx_total = idx_parents + idx_local;
|
2009-06-26 14:28:00 +00:00
|
|
|
|
|
|
|
/* split */
|
2009-07-01 03:35:15 +00:00
|
|
|
new = create_child(parent, true);
|
|
|
|
|
|
|
|
/* split the callchain and move a part to the new child */
|
|
|
|
old_tail = parent->val.prev;
|
|
|
|
list_del_range(&to_split->list, old_tail);
|
|
|
|
new->val.next = &to_split->list;
|
|
|
|
new->val.prev = old_tail;
|
|
|
|
to_split->list.prev = &new->val;
|
|
|
|
old_tail->next = &new->val;
|
2009-06-26 14:28:00 +00:00
|
|
|
|
2009-07-01 03:35:15 +00:00
|
|
|
/* split the hits */
|
|
|
|
new->hit = parent->hit;
|
2009-08-07 05:11:05 +00:00
|
|
|
new->children_hit = parent->children_hit;
|
2011-01-14 03:51:59 +00:00
|
|
|
parent->children_hit = callchain_cumul_hits(new);
|
2009-07-01 03:35:15 +00:00
|
|
|
new->val_nr = parent->val_nr - idx_local;
|
|
|
|
parent->val_nr = idx_local;
|
|
|
|
|
|
|
|
/* create a new child for the new branch if any */
|
2011-01-14 03:51:58 +00:00
|
|
|
if (idx_total < cursor->nr) {
|
2013-10-11 05:15:36 +00:00
|
|
|
struct callchain_node *first;
|
|
|
|
struct callchain_list *cnode;
|
|
|
|
struct callchain_cursor_node *node;
|
|
|
|
struct rb_node *p, **pp;
|
|
|
|
|
2009-07-01 03:35:15 +00:00
|
|
|
parent->hit = 0;
|
2010-07-08 01:41:46 +00:00
|
|
|
parent->children_hit += period;
|
2013-10-11 05:15:36 +00:00
|
|
|
|
|
|
|
node = callchain_cursor_current(cursor);
|
|
|
|
new = add_child(parent, cursor, period);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This is second child since we moved parent's children
|
|
|
|
* to new (first) child above.
|
|
|
|
*/
|
|
|
|
p = parent->rb_root_in.rb_node;
|
|
|
|
first = rb_entry(p, struct callchain_node, rb_node_in);
|
|
|
|
cnode = list_first_entry(&first->val, struct callchain_list,
|
|
|
|
list);
|
|
|
|
|
|
|
|
if (match_chain(node, cnode) < 0)
|
|
|
|
pp = &p->rb_left;
|
|
|
|
else
|
|
|
|
pp = &p->rb_right;
|
|
|
|
|
|
|
|
rb_link_node(&new->rb_node_in, p, pp);
|
|
|
|
rb_insert_color(&new->rb_node_in, &parent->rb_root_in);
|
2009-07-01 03:35:15 +00:00
|
|
|
} else {
|
2010-07-08 01:41:46 +00:00
|
|
|
parent->hit = period;
|
2009-07-01 03:35:15 +00:00
|
|
|
}
|
2009-06-26 14:28:00 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2011-01-14 03:51:58 +00:00
|
|
|
append_chain(struct callchain_node *root,
|
|
|
|
struct callchain_cursor *cursor,
|
|
|
|
u64 period);
|
2009-06-26 14:28:00 +00:00
|
|
|
|
2009-07-01 03:35:15 +00:00
|
|
|
static void
|
2011-01-14 03:51:58 +00:00
|
|
|
append_chain_children(struct callchain_node *root,
|
|
|
|
struct callchain_cursor *cursor,
|
|
|
|
u64 period)
|
2009-06-26 14:28:00 +00:00
|
|
|
{
|
|
|
|
struct callchain_node *rnode;
|
2013-10-11 05:15:36 +00:00
|
|
|
struct callchain_cursor_node *node;
|
|
|
|
struct rb_node **p = &root->rb_root_in.rb_node;
|
|
|
|
struct rb_node *parent = NULL;
|
|
|
|
|
|
|
|
node = callchain_cursor_current(cursor);
|
|
|
|
if (!node)
|
|
|
|
return;
|
2009-06-26 14:28:00 +00:00
|
|
|
|
|
|
|
/* lookup in childrens */
|
2013-10-11 05:15:36 +00:00
|
|
|
while (*p) {
|
|
|
|
s64 ret;
|
2009-07-01 10:37:06 +00:00
|
|
|
|
2013-10-11 05:15:36 +00:00
|
|
|
parent = *p;
|
|
|
|
rnode = rb_entry(parent, struct callchain_node, rb_node_in);
|
|
|
|
|
2014-01-14 15:37:15 +00:00
|
|
|
/* If at least first entry matches, rely to children */
|
|
|
|
ret = append_chain(rnode, cursor, period);
|
|
|
|
if (ret == 0)
|
2009-08-07 05:11:05 +00:00
|
|
|
goto inc_children_hit;
|
2013-10-11 05:15:36 +00:00
|
|
|
|
|
|
|
if (ret < 0)
|
|
|
|
p = &parent->rb_left;
|
|
|
|
else
|
|
|
|
p = &parent->rb_right;
|
2009-06-26 14:28:00 +00:00
|
|
|
}
|
2009-07-01 03:35:15 +00:00
|
|
|
/* nothing in children, add to the current node */
|
2013-10-11 05:15:36 +00:00
|
|
|
rnode = add_child(root, cursor, period);
|
|
|
|
rb_link_node(&rnode->rb_node_in, parent, p);
|
|
|
|
rb_insert_color(&rnode->rb_node_in, &root->rb_root_in);
|
2009-07-05 05:39:20 +00:00
|
|
|
|
2009-08-07 05:11:05 +00:00
|
|
|
inc_children_hit:
|
2010-07-08 01:41:46 +00:00
|
|
|
root->children_hit += period;
|
2009-06-26 14:28:00 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2011-01-14 03:51:58 +00:00
|
|
|
append_chain(struct callchain_node *root,
|
|
|
|
struct callchain_cursor *cursor,
|
|
|
|
u64 period)
|
2009-06-26 14:28:00 +00:00
|
|
|
{
|
|
|
|
struct callchain_list *cnode;
|
2011-01-14 03:51:58 +00:00
|
|
|
u64 start = cursor->pos;
|
2009-06-26 14:28:00 +00:00
|
|
|
bool found = false;
|
2011-01-14 03:51:58 +00:00
|
|
|
u64 matches;
|
2014-01-14 15:37:15 +00:00
|
|
|
int cmp = 0;
|
2009-06-26 14:28:00 +00:00
|
|
|
|
2009-07-01 03:35:15 +00:00
|
|
|
/*
|
|
|
|
* Lookup in the current node
|
|
|
|
* If we have a symbol, then compare the start to match
|
2013-07-18 22:33:57 +00:00
|
|
|
* anywhere inside a function, unless function
|
|
|
|
* mode is disabled.
|
2009-07-01 03:35:15 +00:00
|
|
|
*/
|
2009-06-26 14:28:00 +00:00
|
|
|
list_for_each_entry(cnode, &root->val, list) {
|
2011-01-14 03:51:58 +00:00
|
|
|
struct callchain_cursor_node *node;
|
2010-03-22 16:09:33 +00:00
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
node = callchain_cursor_current(cursor);
|
|
|
|
if (!node)
|
2009-07-01 03:35:15 +00:00
|
|
|
break;
|
2010-03-22 16:09:33 +00:00
|
|
|
|
2014-01-14 15:37:15 +00:00
|
|
|
cmp = match_chain(node, cnode);
|
|
|
|
if (cmp)
|
2009-06-26 14:28:00 +00:00
|
|
|
break;
|
2010-03-22 16:09:33 +00:00
|
|
|
|
2013-10-11 05:15:36 +00:00
|
|
|
found = true;
|
2011-01-14 03:51:58 +00:00
|
|
|
|
|
|
|
callchain_cursor_advance(cursor);
|
2009-06-26 14:28:00 +00:00
|
|
|
}
|
|
|
|
|
2013-10-11 05:15:36 +00:00
|
|
|
/* matches not, relay no the parent */
|
2011-01-14 03:51:58 +00:00
|
|
|
if (!found) {
|
2014-01-14 15:37:15 +00:00
|
|
|
WARN_ONCE(!cmp, "Chain comparison error\n");
|
|
|
|
return cmp;
|
2011-01-14 03:51:58 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
matches = cursor->pos - start;
|
2009-06-26 14:28:00 +00:00
|
|
|
|
|
|
|
/* we match only a part of the node. Split it and add the new chain */
|
2011-01-14 03:51:58 +00:00
|
|
|
if (matches < root->val_nr) {
|
|
|
|
split_add_child(root, cursor, cnode, start, matches, period);
|
2009-06-26 14:28:00 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* we match 100% of the path, increment the hit */
|
2011-01-14 03:51:58 +00:00
|
|
|
if (matches == root->val_nr && cursor->pos == cursor->nr) {
|
2010-07-08 01:41:46 +00:00
|
|
|
root->hit += period;
|
2009-06-26 14:28:00 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2009-07-01 03:35:15 +00:00
|
|
|
/* We match the node and still have a part remaining */
|
2011-01-14 03:51:58 +00:00
|
|
|
append_chain_children(root, cursor, period);
|
2009-07-01 03:35:15 +00:00
|
|
|
|
|
|
|
return 0;
|
2009-06-26 14:28:00 +00:00
|
|
|
}
|
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
int callchain_append(struct callchain_root *root,
|
|
|
|
struct callchain_cursor *cursor,
|
|
|
|
u64 period)
|
2009-06-26 14:28:00 +00:00
|
|
|
{
|
2011-01-14 03:51:58 +00:00
|
|
|
if (!cursor->nr)
|
2010-03-22 16:09:33 +00:00
|
|
|
return 0;
|
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
callchain_cursor_commit(cursor);
|
2010-03-22 16:09:33 +00:00
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
append_chain_children(&root->node, cursor, period);
|
2010-08-22 18:05:22 +00:00
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
if (cursor->nr > root->max_depth)
|
|
|
|
root->max_depth = cursor->nr;
|
2010-03-22 16:09:33 +00:00
|
|
|
|
|
|
|
return 0;
|
2009-06-26 14:28:00 +00:00
|
|
|
}
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
|
|
|
|
static int
|
2011-01-14 03:51:58 +00:00
|
|
|
merge_chain_branch(struct callchain_cursor *cursor,
|
|
|
|
struct callchain_node *dst, struct callchain_node *src)
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
{
|
2011-01-14 03:51:58 +00:00
|
|
|
struct callchain_cursor_node **old_last = cursor->last;
|
2013-10-11 05:15:36 +00:00
|
|
|
struct callchain_node *child;
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
struct callchain_list *list, *next_list;
|
2013-10-11 05:15:36 +00:00
|
|
|
struct rb_node *n;
|
2011-01-14 03:51:58 +00:00
|
|
|
int old_pos = cursor->nr;
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
int err = 0;
|
|
|
|
|
|
|
|
list_for_each_entry_safe(list, next_list, &src->val, list) {
|
2011-01-14 03:51:58 +00:00
|
|
|
callchain_cursor_append(cursor, list->ip,
|
|
|
|
list->ms.map, list->ms.sym);
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
list_del(&list->list);
|
|
|
|
free(list);
|
|
|
|
}
|
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
if (src->hit) {
|
|
|
|
callchain_cursor_commit(cursor);
|
|
|
|
append_chain_children(dst, cursor, src->hit);
|
|
|
|
}
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
|
2013-10-11 05:15:36 +00:00
|
|
|
n = rb_first(&src->rb_root_in);
|
|
|
|
while (n) {
|
|
|
|
child = container_of(n, struct callchain_node, rb_node_in);
|
|
|
|
n = rb_next(n);
|
|
|
|
rb_erase(&child->rb_node_in, &src->rb_root_in);
|
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
err = merge_chain_branch(cursor, dst, child);
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
if (err)
|
|
|
|
break;
|
|
|
|
|
|
|
|
free(child);
|
|
|
|
}
|
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
cursor->nr = old_pos;
|
|
|
|
cursor->last = old_last;
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
int callchain_merge(struct callchain_cursor *cursor,
|
|
|
|
struct callchain_root *dst, struct callchain_root *src)
|
|
|
|
{
|
|
|
|
return merge_chain_branch(cursor, &dst->node, &src->node);
|
|
|
|
}
|
|
|
|
|
|
|
|
int callchain_cursor_append(struct callchain_cursor *cursor,
|
|
|
|
u64 ip, struct map *map, struct symbol *sym)
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
{
|
2011-01-14 03:51:58 +00:00
|
|
|
struct callchain_cursor_node *node = *cursor->last;
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
if (!node) {
|
2013-01-31 01:05:49 +00:00
|
|
|
node = calloc(1, sizeof(*node));
|
2011-01-14 03:51:58 +00:00
|
|
|
if (!node)
|
|
|
|
return -ENOMEM;
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
*cursor->last = node;
|
|
|
|
}
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
node->ip = ip;
|
|
|
|
node->map = map;
|
|
|
|
node->sym = sym;
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
cursor->nr++;
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
|
2011-01-14 03:51:58 +00:00
|
|
|
cursor->last = &node->next;
|
|
|
|
|
|
|
|
return 0;
|
perf: Support for callchains merge
If we sort the histograms by comm, which is the default,
we need to merge some of them, typically different thread
histograms of a same process, or just same comm. But during
this merge, we forgot to merge callchains.
So imagine we have three threads (tids: 1000, 1001, 1002) that
belong to comm "foo".
tid 1000 got 100 events
tid 1001 got 10 events
tid 1002 got 3 events
Once we merge these histograms to get a per comm result, we'll
finally get:
"foo" got 113 events
The problem is if we merge 1000 and 1001 histograms into 1002, then
the end merge result, wrt callchains, will be only callchains that
belong to 1002.
This is because we haven't handled callchains in the merge. Only those
from one of the threads inside a common comm survive.
It means during this merge, we can lose a lot of callchains.
Fix this by implementing callchains merge and apply it on histograms
that collapse.
Reported-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Frederic Weisbecker <fweisbec@gmail.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Arnaldo Carvalho de Melo <acme@redhat.com>
Cc: Paul Mackerras <paulus@samba.org>
2010-08-22 19:10:35 +00:00
|
|
|
}
|
2014-01-14 05:25:35 +00:00
|
|
|
|
|
|
|
int sample__resolve_callchain(struct perf_sample *sample, struct symbol **parent,
|
|
|
|
struct perf_evsel *evsel, struct addr_location *al,
|
|
|
|
int max_stack)
|
|
|
|
{
|
|
|
|
if (sample->callchain == NULL)
|
|
|
|
return 0;
|
|
|
|
|
2012-09-11 05:13:04 +00:00
|
|
|
if (symbol_conf.use_callchain || symbol_conf.cumulate_callchain ||
|
|
|
|
sort__has_parent) {
|
2014-01-14 05:25:35 +00:00
|
|
|
return machine__resolve_callchain(al->machine, evsel, al->thread,
|
|
|
|
sample, parent, al, max_stack);
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int hist_entry__append_callchain(struct hist_entry *he, struct perf_sample *sample)
|
|
|
|
{
|
2014-07-14 10:02:35 +00:00
|
|
|
if (!symbol_conf.use_callchain || sample->callchain == NULL)
|
2014-01-14 05:25:35 +00:00
|
|
|
return 0;
|
|
|
|
return callchain_append(he->callchain, &callchain_cursor, sample->period);
|
|
|
|
}
|
2013-10-31 04:58:30 +00:00
|
|
|
|
|
|
|
int fill_callchain_info(struct addr_location *al, struct callchain_cursor_node *node,
|
|
|
|
bool hide_unresolved)
|
|
|
|
{
|
|
|
|
al->map = node->map;
|
|
|
|
al->sym = node->sym;
|
|
|
|
if (node->map)
|
|
|
|
al->addr = node->map->map_ip(node->map, node->ip);
|
|
|
|
else
|
|
|
|
al->addr = node->ip;
|
|
|
|
|
|
|
|
if (al->sym == NULL) {
|
|
|
|
if (hide_unresolved)
|
|
|
|
return 0;
|
|
|
|
if (al->map == NULL)
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (al->map->groups == &al->machine->kmaps) {
|
|
|
|
if (machine__is_host(al->machine)) {
|
|
|
|
al->cpumode = PERF_RECORD_MISC_KERNEL;
|
|
|
|
al->level = 'k';
|
|
|
|
} else {
|
|
|
|
al->cpumode = PERF_RECORD_MISC_GUEST_KERNEL;
|
|
|
|
al->level = 'g';
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
if (machine__is_host(al->machine)) {
|
|
|
|
al->cpumode = PERF_RECORD_MISC_USER;
|
|
|
|
al->level = '.';
|
|
|
|
} else if (perf_guest) {
|
|
|
|
al->cpumode = PERF_RECORD_MISC_GUEST_USER;
|
|
|
|
al->level = 'u';
|
|
|
|
} else {
|
|
|
|
al->cpumode = PERF_RECORD_MISC_HYPERVISOR;
|
|
|
|
al->level = 'H';
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
return 1;
|
|
|
|
}
|