mirror of
https://github.com/torvalds/linux.git
synced 2024-12-12 22:23:55 +00:00
46125590e0
Address/leak sanitizer with reference count checking can identify the location of leaks, so use put rather than delete to avoid free-ing memory when the reference count is >1. Add maps__zput to ensure the variable is cleared. Signed-off-by: Ian Rogers <irogers@google.com> Cc: Adrian Hunter <adrian.hunter@intel.com> Cc: Alexander Shishkin <alexander.shishkin@linux.intel.com> Cc: Ali Saidi <alisaidi@amazon.com> Cc: Andi Kleen <ak@linux.intel.com> Cc: Athira Rajeev <atrajeev@linux.vnet.ibm.com> Cc: Brian Robbins <brianrob@linux.microsoft.com> Cc: Changbin Du <changbin.du@huawei.com> Cc: Dmitrii Dolgov <9erthalion6@gmail.com> Cc: Fangrui Song <maskray@google.com> Cc: German Gomez <german.gomez@arm.com> Cc: Ingo Molnar <mingo@redhat.com> Cc: Ivan Babrou <ivan@cloudflare.com> Cc: James Clark <james.clark@arm.com> Cc: Jing Zhang <renyu.zj@linux.alibaba.com> Cc: Jiri Olsa <jolsa@kernel.org> Cc: John Garry <john.g.garry@oracle.com> Cc: K Prateek Nayak <kprateek.nayak@amd.com> Cc: Kan Liang <kan.liang@linux.intel.com> Cc: Leo Yan <leo.yan@linaro.org> Cc: Liam Howlett <liam.howlett@oracle.com> Cc: Mark Rutland <mark.rutland@arm.com> Cc: Miguel Ojeda <ojeda@kernel.org> Cc: Mike Leach <mike.leach@linaro.org> Cc: Namhyung Kim <namhyung@kernel.org> Cc: Naveen N. Rao <naveen.n.rao@linux.vnet.ibm.com> Cc: Peter Zijlstra <peterz@infradead.org> Cc: Ravi Bangoria <ravi.bangoria@amd.com> Cc: Sean Christopherson <seanjc@google.com> Cc: Steinar H. Gunderson <sesse@google.com> Cc: Suzuki Poulouse <suzuki.poulose@arm.com> Cc: Wenyu Liu <liuwenyu7@huawei.com> Cc: Will Deacon <will@kernel.org> Cc: Yang Jihong <yangjihong1@huawei.com> Cc: Ye Xingchen <ye.xingchen@zte.com.cn> Cc: Yuan Can <yuancan@huawei.com> Cc: coresight@lists.linaro.org Cc: linux-arm-kernel@lists.infradead.org Link: https://lore.kernel.org/r/20230608232823.4027869-5-irogers@google.com Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
477 lines
10 KiB
C
477 lines
10 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#include <errno.h>
|
|
#include <stdlib.h>
|
|
#include <linux/zalloc.h>
|
|
#include "debug.h"
|
|
#include "dso.h"
|
|
#include "map.h"
|
|
#include "maps.h"
|
|
#include "thread.h"
|
|
#include "ui/ui.h"
|
|
#include "unwind.h"
|
|
|
|
static void maps__init(struct maps *maps, struct machine *machine)
|
|
{
|
|
refcount_set(maps__refcnt(maps), 1);
|
|
init_rwsem(maps__lock(maps));
|
|
RC_CHK_ACCESS(maps)->entries = RB_ROOT;
|
|
RC_CHK_ACCESS(maps)->machine = machine;
|
|
RC_CHK_ACCESS(maps)->last_search_by_name = NULL;
|
|
RC_CHK_ACCESS(maps)->nr_maps = 0;
|
|
RC_CHK_ACCESS(maps)->maps_by_name = NULL;
|
|
}
|
|
|
|
static void __maps__free_maps_by_name(struct maps *maps)
|
|
{
|
|
/*
|
|
* Free everything to try to do it from the rbtree in the next search
|
|
*/
|
|
for (unsigned int i = 0; i < maps__nr_maps(maps); i++)
|
|
map__put(maps__maps_by_name(maps)[i]);
|
|
|
|
zfree(&RC_CHK_ACCESS(maps)->maps_by_name);
|
|
RC_CHK_ACCESS(maps)->nr_maps_allocated = 0;
|
|
}
|
|
|
|
static int __maps__insert(struct maps *maps, struct map *map)
|
|
{
|
|
struct rb_node **p = &maps__entries(maps)->rb_node;
|
|
struct rb_node *parent = NULL;
|
|
const u64 ip = map__start(map);
|
|
struct map_rb_node *m, *new_rb_node;
|
|
|
|
new_rb_node = malloc(sizeof(*new_rb_node));
|
|
if (!new_rb_node)
|
|
return -ENOMEM;
|
|
|
|
RB_CLEAR_NODE(&new_rb_node->rb_node);
|
|
new_rb_node->map = map__get(map);
|
|
|
|
while (*p != NULL) {
|
|
parent = *p;
|
|
m = rb_entry(parent, struct map_rb_node, rb_node);
|
|
if (ip < map__start(m->map))
|
|
p = &(*p)->rb_left;
|
|
else
|
|
p = &(*p)->rb_right;
|
|
}
|
|
|
|
rb_link_node(&new_rb_node->rb_node, parent, p);
|
|
rb_insert_color(&new_rb_node->rb_node, maps__entries(maps));
|
|
return 0;
|
|
}
|
|
|
|
int maps__insert(struct maps *maps, struct map *map)
|
|
{
|
|
int err;
|
|
const struct dso *dso = map__dso(map);
|
|
|
|
down_write(maps__lock(maps));
|
|
err = __maps__insert(maps, map);
|
|
if (err)
|
|
goto out;
|
|
|
|
++RC_CHK_ACCESS(maps)->nr_maps;
|
|
|
|
if (dso && dso->kernel) {
|
|
struct kmap *kmap = map__kmap(map);
|
|
|
|
if (kmap)
|
|
kmap->kmaps = maps;
|
|
else
|
|
pr_err("Internal error: kernel dso with non kernel map\n");
|
|
}
|
|
|
|
|
|
/*
|
|
* If we already performed some search by name, then we need to add the just
|
|
* inserted map and resort.
|
|
*/
|
|
if (maps__maps_by_name(maps)) {
|
|
if (maps__nr_maps(maps) > RC_CHK_ACCESS(maps)->nr_maps_allocated) {
|
|
int nr_allocate = maps__nr_maps(maps) * 2;
|
|
struct map **maps_by_name = realloc(maps__maps_by_name(maps),
|
|
nr_allocate * sizeof(map));
|
|
|
|
if (maps_by_name == NULL) {
|
|
__maps__free_maps_by_name(maps);
|
|
err = -ENOMEM;
|
|
goto out;
|
|
}
|
|
|
|
RC_CHK_ACCESS(maps)->maps_by_name = maps_by_name;
|
|
RC_CHK_ACCESS(maps)->nr_maps_allocated = nr_allocate;
|
|
}
|
|
maps__maps_by_name(maps)[maps__nr_maps(maps) - 1] = map__get(map);
|
|
__maps__sort_by_name(maps);
|
|
}
|
|
out:
|
|
up_write(maps__lock(maps));
|
|
return err;
|
|
}
|
|
|
|
static void __maps__remove(struct maps *maps, struct map_rb_node *rb_node)
|
|
{
|
|
rb_erase_init(&rb_node->rb_node, maps__entries(maps));
|
|
map__put(rb_node->map);
|
|
free(rb_node);
|
|
}
|
|
|
|
void maps__remove(struct maps *maps, struct map *map)
|
|
{
|
|
struct map_rb_node *rb_node;
|
|
|
|
down_write(maps__lock(maps));
|
|
if (RC_CHK_ACCESS(maps)->last_search_by_name == map)
|
|
RC_CHK_ACCESS(maps)->last_search_by_name = NULL;
|
|
|
|
rb_node = maps__find_node(maps, map);
|
|
assert(rb_node->RC_CHK_ACCESS(map) == RC_CHK_ACCESS(map));
|
|
__maps__remove(maps, rb_node);
|
|
if (maps__maps_by_name(maps))
|
|
__maps__free_maps_by_name(maps);
|
|
--RC_CHK_ACCESS(maps)->nr_maps;
|
|
up_write(maps__lock(maps));
|
|
}
|
|
|
|
static void __maps__purge(struct maps *maps)
|
|
{
|
|
struct map_rb_node *pos, *next;
|
|
|
|
if (maps__maps_by_name(maps))
|
|
__maps__free_maps_by_name(maps);
|
|
|
|
maps__for_each_entry_safe(maps, pos, next) {
|
|
rb_erase_init(&pos->rb_node, maps__entries(maps));
|
|
map__put(pos->map);
|
|
free(pos);
|
|
}
|
|
}
|
|
|
|
static void maps__exit(struct maps *maps)
|
|
{
|
|
down_write(maps__lock(maps));
|
|
__maps__purge(maps);
|
|
up_write(maps__lock(maps));
|
|
}
|
|
|
|
bool maps__empty(struct maps *maps)
|
|
{
|
|
return !maps__first(maps);
|
|
}
|
|
|
|
struct maps *maps__new(struct machine *machine)
|
|
{
|
|
struct maps *result;
|
|
RC_STRUCT(maps) *maps = zalloc(sizeof(*maps));
|
|
|
|
if (ADD_RC_CHK(result, maps))
|
|
maps__init(result, machine);
|
|
|
|
return result;
|
|
}
|
|
|
|
static void maps__delete(struct maps *maps)
|
|
{
|
|
maps__exit(maps);
|
|
unwind__finish_access(maps);
|
|
RC_CHK_FREE(maps);
|
|
}
|
|
|
|
struct maps *maps__get(struct maps *maps)
|
|
{
|
|
struct maps *result;
|
|
|
|
if (RC_CHK_GET(result, maps))
|
|
refcount_inc(maps__refcnt(maps));
|
|
|
|
return result;
|
|
}
|
|
|
|
void maps__put(struct maps *maps)
|
|
{
|
|
if (maps && refcount_dec_and_test(maps__refcnt(maps)))
|
|
maps__delete(maps);
|
|
else
|
|
RC_CHK_PUT(maps);
|
|
}
|
|
|
|
struct symbol *maps__find_symbol(struct maps *maps, u64 addr, struct map **mapp)
|
|
{
|
|
struct map *map = maps__find(maps, addr);
|
|
|
|
/* Ensure map is loaded before using map->map_ip */
|
|
if (map != NULL && map__load(map) >= 0) {
|
|
if (mapp != NULL)
|
|
*mapp = map;
|
|
return map__find_symbol(map, map__map_ip(map, addr));
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
struct symbol *maps__find_symbol_by_name(struct maps *maps, const char *name, struct map **mapp)
|
|
{
|
|
struct symbol *sym;
|
|
struct map_rb_node *pos;
|
|
|
|
down_read(maps__lock(maps));
|
|
|
|
maps__for_each_entry(maps, pos) {
|
|
sym = map__find_symbol_by_name(pos->map, name);
|
|
|
|
if (sym == NULL)
|
|
continue;
|
|
if (!map__contains_symbol(pos->map, sym)) {
|
|
sym = NULL;
|
|
continue;
|
|
}
|
|
if (mapp != NULL)
|
|
*mapp = pos->map;
|
|
goto out;
|
|
}
|
|
|
|
sym = NULL;
|
|
out:
|
|
up_read(maps__lock(maps));
|
|
return sym;
|
|
}
|
|
|
|
int maps__find_ams(struct maps *maps, struct addr_map_symbol *ams)
|
|
{
|
|
if (ams->addr < map__start(ams->ms.map) || ams->addr >= map__end(ams->ms.map)) {
|
|
if (maps == NULL)
|
|
return -1;
|
|
ams->ms.map = maps__find(maps, ams->addr);
|
|
if (ams->ms.map == NULL)
|
|
return -1;
|
|
}
|
|
|
|
ams->al_addr = map__map_ip(ams->ms.map, ams->addr);
|
|
ams->ms.sym = map__find_symbol(ams->ms.map, ams->al_addr);
|
|
|
|
return ams->ms.sym ? 0 : -1;
|
|
}
|
|
|
|
size_t maps__fprintf(struct maps *maps, FILE *fp)
|
|
{
|
|
size_t printed = 0;
|
|
struct map_rb_node *pos;
|
|
|
|
down_read(maps__lock(maps));
|
|
|
|
maps__for_each_entry(maps, pos) {
|
|
printed += fprintf(fp, "Map:");
|
|
printed += map__fprintf(pos->map, fp);
|
|
if (verbose > 2) {
|
|
printed += dso__fprintf(map__dso(pos->map), fp);
|
|
printed += fprintf(fp, "--\n");
|
|
}
|
|
}
|
|
|
|
up_read(maps__lock(maps));
|
|
|
|
return printed;
|
|
}
|
|
|
|
int maps__fixup_overlappings(struct maps *maps, struct map *map, FILE *fp)
|
|
{
|
|
struct rb_root *root;
|
|
struct rb_node *next, *first;
|
|
int err = 0;
|
|
|
|
down_write(maps__lock(maps));
|
|
|
|
root = maps__entries(maps);
|
|
|
|
/*
|
|
* Find first map where end > map->start.
|
|
* Same as find_vma() in kernel.
|
|
*/
|
|
next = root->rb_node;
|
|
first = NULL;
|
|
while (next) {
|
|
struct map_rb_node *pos = rb_entry(next, struct map_rb_node, rb_node);
|
|
|
|
if (map__end(pos->map) > map__start(map)) {
|
|
first = next;
|
|
if (map__start(pos->map) <= map__start(map))
|
|
break;
|
|
next = next->rb_left;
|
|
} else
|
|
next = next->rb_right;
|
|
}
|
|
|
|
next = first;
|
|
while (next && !err) {
|
|
struct map_rb_node *pos = rb_entry(next, struct map_rb_node, rb_node);
|
|
next = rb_next(&pos->rb_node);
|
|
|
|
/*
|
|
* Stop if current map starts after map->end.
|
|
* Maps are ordered by start: next will not overlap for sure.
|
|
*/
|
|
if (map__start(pos->map) >= map__end(map))
|
|
break;
|
|
|
|
if (verbose >= 2) {
|
|
|
|
if (use_browser) {
|
|
pr_debug("overlapping maps in %s (disable tui for more info)\n",
|
|
map__dso(map)->name);
|
|
} else {
|
|
fputs("overlapping maps:\n", fp);
|
|
map__fprintf(map, fp);
|
|
map__fprintf(pos->map, fp);
|
|
}
|
|
}
|
|
|
|
rb_erase_init(&pos->rb_node, root);
|
|
/*
|
|
* Now check if we need to create new maps for areas not
|
|
* overlapped by the new map:
|
|
*/
|
|
if (map__start(map) > map__start(pos->map)) {
|
|
struct map *before = map__clone(pos->map);
|
|
|
|
if (before == NULL) {
|
|
err = -ENOMEM;
|
|
goto put_map;
|
|
}
|
|
|
|
map__set_end(before, map__start(map));
|
|
err = __maps__insert(maps, before);
|
|
if (err) {
|
|
map__put(before);
|
|
goto put_map;
|
|
}
|
|
|
|
if (verbose >= 2 && !use_browser)
|
|
map__fprintf(before, fp);
|
|
map__put(before);
|
|
}
|
|
|
|
if (map__end(map) < map__end(pos->map)) {
|
|
struct map *after = map__clone(pos->map);
|
|
|
|
if (after == NULL) {
|
|
err = -ENOMEM;
|
|
goto put_map;
|
|
}
|
|
|
|
map__set_start(after, map__end(map));
|
|
map__add_pgoff(after, map__end(map) - map__start(pos->map));
|
|
assert(map__map_ip(pos->map, map__end(map)) ==
|
|
map__map_ip(after, map__end(map)));
|
|
err = __maps__insert(maps, after);
|
|
if (err) {
|
|
map__put(after);
|
|
goto put_map;
|
|
}
|
|
if (verbose >= 2 && !use_browser)
|
|
map__fprintf(after, fp);
|
|
map__put(after);
|
|
}
|
|
put_map:
|
|
map__put(pos->map);
|
|
}
|
|
up_write(maps__lock(maps));
|
|
return err;
|
|
}
|
|
|
|
/*
|
|
* XXX This should not really _copy_ te maps, but refcount them.
|
|
*/
|
|
int maps__clone(struct thread *thread, struct maps *parent)
|
|
{
|
|
struct maps *maps = thread__maps(thread);
|
|
int err;
|
|
struct map_rb_node *rb_node;
|
|
|
|
down_read(maps__lock(parent));
|
|
|
|
maps__for_each_entry(parent, rb_node) {
|
|
struct map *new = map__clone(rb_node->map);
|
|
|
|
if (new == NULL) {
|
|
err = -ENOMEM;
|
|
goto out_unlock;
|
|
}
|
|
|
|
err = unwind__prepare_access(maps, new, NULL);
|
|
if (err)
|
|
goto out_unlock;
|
|
|
|
err = maps__insert(maps, new);
|
|
if (err)
|
|
goto out_unlock;
|
|
|
|
map__put(new);
|
|
}
|
|
|
|
err = 0;
|
|
out_unlock:
|
|
up_read(maps__lock(parent));
|
|
return err;
|
|
}
|
|
|
|
struct map_rb_node *maps__find_node(struct maps *maps, struct map *map)
|
|
{
|
|
struct map_rb_node *rb_node;
|
|
|
|
maps__for_each_entry(maps, rb_node) {
|
|
if (rb_node->RC_CHK_ACCESS(map) == RC_CHK_ACCESS(map))
|
|
return rb_node;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
struct map *maps__find(struct maps *maps, u64 ip)
|
|
{
|
|
struct rb_node *p;
|
|
struct map_rb_node *m;
|
|
|
|
|
|
down_read(maps__lock(maps));
|
|
|
|
p = maps__entries(maps)->rb_node;
|
|
while (p != NULL) {
|
|
m = rb_entry(p, struct map_rb_node, rb_node);
|
|
if (ip < map__start(m->map))
|
|
p = p->rb_left;
|
|
else if (ip >= map__end(m->map))
|
|
p = p->rb_right;
|
|
else
|
|
goto out;
|
|
}
|
|
|
|
m = NULL;
|
|
out:
|
|
up_read(maps__lock(maps));
|
|
return m ? m->map : NULL;
|
|
}
|
|
|
|
struct map_rb_node *maps__first(struct maps *maps)
|
|
{
|
|
struct rb_node *first = rb_first(maps__entries(maps));
|
|
|
|
if (first)
|
|
return rb_entry(first, struct map_rb_node, rb_node);
|
|
return NULL;
|
|
}
|
|
|
|
struct map_rb_node *map_rb_node__next(struct map_rb_node *node)
|
|
{
|
|
struct rb_node *next;
|
|
|
|
if (!node)
|
|
return NULL;
|
|
|
|
next = rb_next(&node->rb_node);
|
|
|
|
if (!next)
|
|
return NULL;
|
|
|
|
return rb_entry(next, struct map_rb_node, rb_node);
|
|
}
|