mirror of
https://github.com/torvalds/linux.git
synced 2024-12-15 23:51:46 +00:00
6f3189f38a
This commit marks kfuncs as such inside the .BTF_ids section. The upshot of these annotations is that we'll be able to automatically generate kfunc prototypes for downstream users. The process is as follows: 1. In source, use BTF_KFUNCS_START/END macro pair to mark kfuncs 2. During build, pahole injects into BTF a "bpf_kfunc" BTF_DECL_TAG for each function inside BTF_KFUNCS sets 3. At runtime, vmlinux or module BTF is made available in sysfs 4. At runtime, bpftool (or similar) can look at provided BTF and generate appropriate prototypes for functions with "bpf_kfunc" tag To ensure future kfunc are similarly tagged, we now also return error inside kfunc registration for untagged kfuncs. For vmlinux kfuncs, we also WARN(), as initcall machinery does not handle errors. Signed-off-by: Daniel Xu <dxu@dxuuu.xyz> Acked-by: Benjamin Tissoires <bentiss@kernel.org> Link: https://lore.kernel.org/r/e55150ceecbf0a5d961e608941165c0bee7bc943.1706491398.git.dxu@dxuuu.xyz Signed-off-by: Alexei Starovoitov <ast@kernel.org>
230 lines
5.3 KiB
C
230 lines
5.3 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/* Copyright (c) 2020 Facebook */
|
|
#include <linux/bpf.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/filter.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/btf_ids.h>
|
|
|
|
struct bpf_iter_seq_map_info {
|
|
u32 map_id;
|
|
};
|
|
|
|
static void *bpf_map_seq_start(struct seq_file *seq, loff_t *pos)
|
|
{
|
|
struct bpf_iter_seq_map_info *info = seq->private;
|
|
struct bpf_map *map;
|
|
|
|
map = bpf_map_get_curr_or_next(&info->map_id);
|
|
if (!map)
|
|
return NULL;
|
|
|
|
if (*pos == 0)
|
|
++*pos;
|
|
return map;
|
|
}
|
|
|
|
static void *bpf_map_seq_next(struct seq_file *seq, void *v, loff_t *pos)
|
|
{
|
|
struct bpf_iter_seq_map_info *info = seq->private;
|
|
|
|
++*pos;
|
|
++info->map_id;
|
|
bpf_map_put((struct bpf_map *)v);
|
|
return bpf_map_get_curr_or_next(&info->map_id);
|
|
}
|
|
|
|
struct bpf_iter__bpf_map {
|
|
__bpf_md_ptr(struct bpf_iter_meta *, meta);
|
|
__bpf_md_ptr(struct bpf_map *, map);
|
|
};
|
|
|
|
DEFINE_BPF_ITER_FUNC(bpf_map, struct bpf_iter_meta *meta, struct bpf_map *map)
|
|
|
|
static int __bpf_map_seq_show(struct seq_file *seq, void *v, bool in_stop)
|
|
{
|
|
struct bpf_iter__bpf_map ctx;
|
|
struct bpf_iter_meta meta;
|
|
struct bpf_prog *prog;
|
|
int ret = 0;
|
|
|
|
ctx.meta = &meta;
|
|
ctx.map = v;
|
|
meta.seq = seq;
|
|
prog = bpf_iter_get_info(&meta, in_stop);
|
|
if (prog)
|
|
ret = bpf_iter_run_prog(prog, &ctx);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int bpf_map_seq_show(struct seq_file *seq, void *v)
|
|
{
|
|
return __bpf_map_seq_show(seq, v, false);
|
|
}
|
|
|
|
static void bpf_map_seq_stop(struct seq_file *seq, void *v)
|
|
{
|
|
if (!v)
|
|
(void)__bpf_map_seq_show(seq, v, true);
|
|
else
|
|
bpf_map_put((struct bpf_map *)v);
|
|
}
|
|
|
|
static const struct seq_operations bpf_map_seq_ops = {
|
|
.start = bpf_map_seq_start,
|
|
.next = bpf_map_seq_next,
|
|
.stop = bpf_map_seq_stop,
|
|
.show = bpf_map_seq_show,
|
|
};
|
|
|
|
BTF_ID_LIST_GLOBAL_SINGLE(btf_bpf_map_id, struct, bpf_map)
|
|
|
|
static const struct bpf_iter_seq_info bpf_map_seq_info = {
|
|
.seq_ops = &bpf_map_seq_ops,
|
|
.init_seq_private = NULL,
|
|
.fini_seq_private = NULL,
|
|
.seq_priv_size = sizeof(struct bpf_iter_seq_map_info),
|
|
};
|
|
|
|
static struct bpf_iter_reg bpf_map_reg_info = {
|
|
.target = "bpf_map",
|
|
.ctx_arg_info_size = 1,
|
|
.ctx_arg_info = {
|
|
{ offsetof(struct bpf_iter__bpf_map, map),
|
|
PTR_TO_BTF_ID_OR_NULL | PTR_TRUSTED },
|
|
},
|
|
.seq_info = &bpf_map_seq_info,
|
|
};
|
|
|
|
static int bpf_iter_attach_map(struct bpf_prog *prog,
|
|
union bpf_iter_link_info *linfo,
|
|
struct bpf_iter_aux_info *aux)
|
|
{
|
|
u32 key_acc_size, value_acc_size, key_size, value_size;
|
|
struct bpf_map *map;
|
|
bool is_percpu = false;
|
|
int err = -EINVAL;
|
|
|
|
if (!linfo->map.map_fd)
|
|
return -EBADF;
|
|
|
|
map = bpf_map_get_with_uref(linfo->map.map_fd);
|
|
if (IS_ERR(map))
|
|
return PTR_ERR(map);
|
|
|
|
if (map->map_type == BPF_MAP_TYPE_PERCPU_HASH ||
|
|
map->map_type == BPF_MAP_TYPE_LRU_PERCPU_HASH ||
|
|
map->map_type == BPF_MAP_TYPE_PERCPU_ARRAY)
|
|
is_percpu = true;
|
|
else if (map->map_type != BPF_MAP_TYPE_HASH &&
|
|
map->map_type != BPF_MAP_TYPE_LRU_HASH &&
|
|
map->map_type != BPF_MAP_TYPE_ARRAY)
|
|
goto put_map;
|
|
|
|
key_acc_size = prog->aux->max_rdonly_access;
|
|
value_acc_size = prog->aux->max_rdwr_access;
|
|
key_size = map->key_size;
|
|
if (!is_percpu)
|
|
value_size = map->value_size;
|
|
else
|
|
value_size = round_up(map->value_size, 8) * num_possible_cpus();
|
|
|
|
if (key_acc_size > key_size || value_acc_size > value_size) {
|
|
err = -EACCES;
|
|
goto put_map;
|
|
}
|
|
|
|
aux->map = map;
|
|
return 0;
|
|
|
|
put_map:
|
|
bpf_map_put_with_uref(map);
|
|
return err;
|
|
}
|
|
|
|
static void bpf_iter_detach_map(struct bpf_iter_aux_info *aux)
|
|
{
|
|
bpf_map_put_with_uref(aux->map);
|
|
}
|
|
|
|
void bpf_iter_map_show_fdinfo(const struct bpf_iter_aux_info *aux,
|
|
struct seq_file *seq)
|
|
{
|
|
seq_printf(seq, "map_id:\t%u\n", aux->map->id);
|
|
}
|
|
|
|
int bpf_iter_map_fill_link_info(const struct bpf_iter_aux_info *aux,
|
|
struct bpf_link_info *info)
|
|
{
|
|
info->iter.map.map_id = aux->map->id;
|
|
return 0;
|
|
}
|
|
|
|
DEFINE_BPF_ITER_FUNC(bpf_map_elem, struct bpf_iter_meta *meta,
|
|
struct bpf_map *map, void *key, void *value)
|
|
|
|
static const struct bpf_iter_reg bpf_map_elem_reg_info = {
|
|
.target = "bpf_map_elem",
|
|
.attach_target = bpf_iter_attach_map,
|
|
.detach_target = bpf_iter_detach_map,
|
|
.show_fdinfo = bpf_iter_map_show_fdinfo,
|
|
.fill_link_info = bpf_iter_map_fill_link_info,
|
|
.ctx_arg_info_size = 2,
|
|
.ctx_arg_info = {
|
|
{ offsetof(struct bpf_iter__bpf_map_elem, key),
|
|
PTR_TO_BUF | PTR_MAYBE_NULL | MEM_RDONLY },
|
|
{ offsetof(struct bpf_iter__bpf_map_elem, value),
|
|
PTR_TO_BUF | PTR_MAYBE_NULL },
|
|
},
|
|
};
|
|
|
|
static int __init bpf_map_iter_init(void)
|
|
{
|
|
int ret;
|
|
|
|
bpf_map_reg_info.ctx_arg_info[0].btf_id = *btf_bpf_map_id;
|
|
ret = bpf_iter_reg_target(&bpf_map_reg_info);
|
|
if (ret)
|
|
return ret;
|
|
|
|
return bpf_iter_reg_target(&bpf_map_elem_reg_info);
|
|
}
|
|
|
|
late_initcall(bpf_map_iter_init);
|
|
|
|
__bpf_kfunc_start_defs();
|
|
|
|
__bpf_kfunc s64 bpf_map_sum_elem_count(const struct bpf_map *map)
|
|
{
|
|
s64 *pcount;
|
|
s64 ret = 0;
|
|
int cpu;
|
|
|
|
if (!map || !map->elem_count)
|
|
return 0;
|
|
|
|
for_each_possible_cpu(cpu) {
|
|
pcount = per_cpu_ptr(map->elem_count, cpu);
|
|
ret += READ_ONCE(*pcount);
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
__bpf_kfunc_end_defs();
|
|
|
|
BTF_KFUNCS_START(bpf_map_iter_kfunc_ids)
|
|
BTF_ID_FLAGS(func, bpf_map_sum_elem_count, KF_TRUSTED_ARGS)
|
|
BTF_KFUNCS_END(bpf_map_iter_kfunc_ids)
|
|
|
|
static const struct btf_kfunc_id_set bpf_map_iter_kfunc_set = {
|
|
.owner = THIS_MODULE,
|
|
.set = &bpf_map_iter_kfunc_ids,
|
|
};
|
|
|
|
static int init_subsystem(void)
|
|
{
|
|
return register_btf_kfunc_id_set(BPF_PROG_TYPE_UNSPEC, &bpf_map_iter_kfunc_set);
|
|
}
|
|
late_initcall(init_subsystem);
|