mirror of
https://github.com/torvalds/linux.git
synced 2024-11-26 06:02:05 +00:00
2d288745eb
When building with clang's -Wincompatible-function-pointer-types-strict (a warning designed to catch potential kCFI failures at build time), there are several warnings along the lines of: fs/bcachefs/bkey_methods.c:118:2: error: incompatible function pointer types initializing 'int (*)(struct btree_trans *, enum btree_id, unsigned int, struct bkey_s_c, struct bkey_s, enum btree_iter_update_trigger_flags)' with an expression of type 'int (struct btree_trans *, enum btree_id, unsigned int, struct bkey_s_c, struct bkey_s, unsigned int)' [-Werror,-Wincompatible-function-pointer-types-strict] 118 | BCH_BKEY_TYPES() | ^~~~~~~~~~~~~~~~ fs/bcachefs/bcachefs_format.h:394:2: note: expanded from macro 'BCH_BKEY_TYPES' 394 | x(inode, 8) \ | ^~~~~~~~~~~~~~~~~~~~~~~~~~ fs/bcachefs/bkey_methods.c:117:41: note: expanded from macro 'x' 117 | #define x(name, nr) [KEY_TYPE_##name] = bch2_bkey_ops_##name, | ^~~~~~~~~~~~~~~~~~~~ <scratch space>:277:1: note: expanded from here 277 | bch2_bkey_ops_inode | ^~~~~~~~~~~~~~~~~~~ fs/bcachefs/inode.h:26:13: note: expanded from macro 'bch2_bkey_ops_inode' 26 | .trigger = bch2_trigger_inode, \ | ^~~~~~~~~~~~~~~~~~ There are several functions that did not have their flags parameter converted to 'enum btree_iter_update_trigger_flags' in the recent unification, which will cause kCFI failures at runtime because the types, while ABI compatible (hence no warning from the non-strict version of this warning), do not match exactly. Fix up these functions (as well as a few other obvious functions that should have it, even if there are no warnings currently) to resolve the warnings and potential kCFI runtime failures. Fixes: 31e4ef3280c8 ("bcachefs: iter/update/trigger/str_hash flag cleanup") Signed-off-by: Nathan Chancellor <nathan@kernel.org> Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
272 lines
6.6 KiB
C
272 lines
6.6 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef _BCACHEFS_SNAPSHOT_H
|
|
#define _BCACHEFS_SNAPSHOT_H
|
|
|
|
enum bkey_invalid_flags;
|
|
|
|
void bch2_snapshot_tree_to_text(struct printbuf *, struct bch_fs *, struct bkey_s_c);
|
|
int bch2_snapshot_tree_invalid(struct bch_fs *, struct bkey_s_c,
|
|
enum bkey_invalid_flags, struct printbuf *);
|
|
|
|
#define bch2_bkey_ops_snapshot_tree ((struct bkey_ops) { \
|
|
.key_invalid = bch2_snapshot_tree_invalid, \
|
|
.val_to_text = bch2_snapshot_tree_to_text, \
|
|
.min_val_size = 8, \
|
|
})
|
|
|
|
struct bkey_i_snapshot_tree *__bch2_snapshot_tree_create(struct btree_trans *);
|
|
|
|
int bch2_snapshot_tree_lookup(struct btree_trans *, u32, struct bch_snapshot_tree *);
|
|
|
|
void bch2_snapshot_to_text(struct printbuf *, struct bch_fs *, struct bkey_s_c);
|
|
int bch2_snapshot_invalid(struct bch_fs *, struct bkey_s_c,
|
|
enum bkey_invalid_flags, struct printbuf *);
|
|
int bch2_mark_snapshot(struct btree_trans *, enum btree_id, unsigned,
|
|
struct bkey_s_c, struct bkey_s,
|
|
enum btree_iter_update_trigger_flags);
|
|
|
|
#define bch2_bkey_ops_snapshot ((struct bkey_ops) { \
|
|
.key_invalid = bch2_snapshot_invalid, \
|
|
.val_to_text = bch2_snapshot_to_text, \
|
|
.trigger = bch2_mark_snapshot, \
|
|
.min_val_size = 24, \
|
|
})
|
|
|
|
static inline struct snapshot_t *__snapshot_t(struct snapshot_table *t, u32 id)
|
|
{
|
|
u32 idx = U32_MAX - id;
|
|
|
|
return likely(t && idx < t->nr)
|
|
? &t->s[idx]
|
|
: NULL;
|
|
}
|
|
|
|
static inline const struct snapshot_t *snapshot_t(struct bch_fs *c, u32 id)
|
|
{
|
|
return __snapshot_t(rcu_dereference(c->snapshots), id);
|
|
}
|
|
|
|
static inline u32 bch2_snapshot_tree(struct bch_fs *c, u32 id)
|
|
{
|
|
rcu_read_lock();
|
|
const struct snapshot_t *s = snapshot_t(c, id);
|
|
id = s ? s->tree : 0;
|
|
rcu_read_unlock();
|
|
|
|
return id;
|
|
}
|
|
|
|
static inline u32 __bch2_snapshot_parent_early(struct bch_fs *c, u32 id)
|
|
{
|
|
const struct snapshot_t *s = snapshot_t(c, id);
|
|
return s ? s->parent : 0;
|
|
}
|
|
|
|
static inline u32 bch2_snapshot_parent_early(struct bch_fs *c, u32 id)
|
|
{
|
|
rcu_read_lock();
|
|
id = __bch2_snapshot_parent_early(c, id);
|
|
rcu_read_unlock();
|
|
|
|
return id;
|
|
}
|
|
|
|
static inline u32 __bch2_snapshot_parent(struct bch_fs *c, u32 id)
|
|
{
|
|
const struct snapshot_t *s = snapshot_t(c, id);
|
|
if (!s)
|
|
return 0;
|
|
|
|
u32 parent = s->parent;
|
|
if (IS_ENABLED(CONFIG_BCACHEFS_DEBUG) &&
|
|
parent &&
|
|
s->depth != snapshot_t(c, parent)->depth + 1)
|
|
panic("id %u depth=%u parent %u depth=%u\n",
|
|
id, snapshot_t(c, id)->depth,
|
|
parent, snapshot_t(c, parent)->depth);
|
|
|
|
return parent;
|
|
}
|
|
|
|
static inline u32 bch2_snapshot_parent(struct bch_fs *c, u32 id)
|
|
{
|
|
rcu_read_lock();
|
|
id = __bch2_snapshot_parent(c, id);
|
|
rcu_read_unlock();
|
|
|
|
return id;
|
|
}
|
|
|
|
static inline u32 bch2_snapshot_nth_parent(struct bch_fs *c, u32 id, u32 n)
|
|
{
|
|
rcu_read_lock();
|
|
while (n--)
|
|
id = __bch2_snapshot_parent(c, id);
|
|
rcu_read_unlock();
|
|
|
|
return id;
|
|
}
|
|
|
|
u32 bch2_snapshot_skiplist_get(struct bch_fs *, u32);
|
|
|
|
static inline u32 bch2_snapshot_root(struct bch_fs *c, u32 id)
|
|
{
|
|
u32 parent;
|
|
|
|
rcu_read_lock();
|
|
while ((parent = __bch2_snapshot_parent(c, id)))
|
|
id = parent;
|
|
rcu_read_unlock();
|
|
|
|
return id;
|
|
}
|
|
|
|
static inline u32 __bch2_snapshot_equiv(struct bch_fs *c, u32 id)
|
|
{
|
|
const struct snapshot_t *s = snapshot_t(c, id);
|
|
return s ? s->equiv : 0;
|
|
}
|
|
|
|
static inline u32 bch2_snapshot_equiv(struct bch_fs *c, u32 id)
|
|
{
|
|
rcu_read_lock();
|
|
id = __bch2_snapshot_equiv(c, id);
|
|
rcu_read_unlock();
|
|
|
|
return id;
|
|
}
|
|
|
|
static inline int bch2_snapshot_is_internal_node(struct bch_fs *c, u32 id)
|
|
{
|
|
rcu_read_lock();
|
|
const struct snapshot_t *s = snapshot_t(c, id);
|
|
int ret = s ? s->children[0] : -BCH_ERR_invalid_snapshot_node;
|
|
rcu_read_unlock();
|
|
|
|
return ret;
|
|
}
|
|
|
|
static inline int bch2_snapshot_is_leaf(struct bch_fs *c, u32 id)
|
|
{
|
|
int ret = bch2_snapshot_is_internal_node(c, id);
|
|
if (ret < 0)
|
|
return ret;
|
|
return !ret;
|
|
}
|
|
|
|
static inline u32 bch2_snapshot_depth(struct bch_fs *c, u32 parent)
|
|
{
|
|
u32 depth;
|
|
|
|
rcu_read_lock();
|
|
depth = parent ? snapshot_t(c, parent)->depth + 1 : 0;
|
|
rcu_read_unlock();
|
|
|
|
return depth;
|
|
}
|
|
|
|
bool __bch2_snapshot_is_ancestor(struct bch_fs *, u32, u32);
|
|
|
|
static inline bool bch2_snapshot_is_ancestor(struct bch_fs *c, u32 id, u32 ancestor)
|
|
{
|
|
return id == ancestor
|
|
? true
|
|
: __bch2_snapshot_is_ancestor(c, id, ancestor);
|
|
}
|
|
|
|
static inline bool bch2_snapshot_has_children(struct bch_fs *c, u32 id)
|
|
{
|
|
const struct snapshot_t *t;
|
|
bool ret;
|
|
|
|
rcu_read_lock();
|
|
t = snapshot_t(c, id);
|
|
ret = (t->children[0]|t->children[1]) != 0;
|
|
rcu_read_unlock();
|
|
|
|
return ret;
|
|
}
|
|
|
|
static inline bool snapshot_list_has_id(snapshot_id_list *s, u32 id)
|
|
{
|
|
darray_for_each(*s, i)
|
|
if (*i == id)
|
|
return true;
|
|
return false;
|
|
}
|
|
|
|
static inline bool snapshot_list_has_ancestor(struct bch_fs *c, snapshot_id_list *s, u32 id)
|
|
{
|
|
darray_for_each(*s, i)
|
|
if (bch2_snapshot_is_ancestor(c, id, *i))
|
|
return true;
|
|
return false;
|
|
}
|
|
|
|
static inline int snapshot_list_add(struct bch_fs *c, snapshot_id_list *s, u32 id)
|
|
{
|
|
BUG_ON(snapshot_list_has_id(s, id));
|
|
int ret = darray_push(s, id);
|
|
if (ret)
|
|
bch_err(c, "error reallocating snapshot_id_list (size %zu)", s->size);
|
|
return ret;
|
|
}
|
|
|
|
static inline int snapshot_list_add_nodup(struct bch_fs *c, snapshot_id_list *s, u32 id)
|
|
{
|
|
int ret = snapshot_list_has_id(s, id)
|
|
? 0
|
|
: darray_push(s, id);
|
|
if (ret)
|
|
bch_err(c, "error reallocating snapshot_id_list (size %zu)", s->size);
|
|
return ret;
|
|
}
|
|
|
|
static inline int snapshot_list_merge(struct bch_fs *c, snapshot_id_list *dst, snapshot_id_list *src)
|
|
{
|
|
darray_for_each(*src, i) {
|
|
int ret = snapshot_list_add_nodup(c, dst, *i);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int bch2_snapshot_lookup(struct btree_trans *trans, u32 id,
|
|
struct bch_snapshot *s);
|
|
int bch2_snapshot_get_subvol(struct btree_trans *, u32,
|
|
struct bch_subvolume *);
|
|
|
|
/* only exported for tests: */
|
|
int bch2_snapshot_node_create(struct btree_trans *, u32,
|
|
u32 *, u32 *, unsigned);
|
|
|
|
int bch2_check_snapshot_trees(struct bch_fs *);
|
|
int bch2_check_snapshots(struct bch_fs *);
|
|
int bch2_reconstruct_snapshots(struct bch_fs *);
|
|
|
|
int bch2_snapshot_node_set_deleted(struct btree_trans *, u32);
|
|
void bch2_delete_dead_snapshots_work(struct work_struct *);
|
|
|
|
int __bch2_key_has_snapshot_overwrites(struct btree_trans *, enum btree_id, struct bpos);
|
|
|
|
static inline int bch2_key_has_snapshot_overwrites(struct btree_trans *trans,
|
|
enum btree_id id,
|
|
struct bpos pos)
|
|
{
|
|
if (!btree_type_has_snapshots(id) ||
|
|
bch2_snapshot_is_leaf(trans->c, pos.snapshot) > 0)
|
|
return 0;
|
|
|
|
return __bch2_key_has_snapshot_overwrites(trans, id, pos);
|
|
}
|
|
|
|
int bch2_propagate_key_to_snapshot_leaves(struct btree_trans *, enum btree_id,
|
|
struct bkey_s_c, struct bpos *);
|
|
|
|
int bch2_snapshots_read(struct bch_fs *);
|
|
void bch2_fs_snapshots_exit(struct bch_fs *);
|
|
|
|
#endif /* _BCACHEFS_SNAPSHOT_H */
|