mirror of
https://github.com/torvalds/linux.git
synced 2024-11-22 04:02:20 +00:00
baefd3f849
When allocating new btree nodes, we were leaving them on the freeable list - unlocked - allowing them to be reclaimed: ouch. Additionally, bch2_btree_node_free_never_used() -> bch2_btree_node_hash_remove was putting it on the freelist, while bch2_btree_node_free_never_used() was putting it back on the btree update reserve list - ouch. Originally, the code was written to always keep btree nodes on a list - live or freeable - and this worked when new nodes were kept locked. But now with the cycle detector, we can't keep nodes locked that aren't tracked by the cycle detector; and this is fine as long as they're not reachable. We also have better and more robust leak detection now, with memory allocation profiling, so the original justification no longer applies. Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
149 lines
4.5 KiB
C
149 lines
4.5 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef _BCACHEFS_BTREE_CACHE_H
|
|
#define _BCACHEFS_BTREE_CACHE_H
|
|
|
|
#include "bcachefs.h"
|
|
#include "btree_types.h"
|
|
#include "bkey_methods.h"
|
|
|
|
extern const char * const bch2_btree_node_flags[];
|
|
|
|
struct btree_iter;
|
|
|
|
void bch2_recalc_btree_reserve(struct bch_fs *);
|
|
|
|
void bch2_btree_node_to_freelist(struct bch_fs *, struct btree *);
|
|
|
|
void __bch2_btree_node_hash_remove(struct btree_cache *, struct btree *);
|
|
void bch2_btree_node_hash_remove(struct btree_cache *, struct btree *);
|
|
|
|
int __bch2_btree_node_hash_insert(struct btree_cache *, struct btree *);
|
|
int bch2_btree_node_hash_insert(struct btree_cache *, struct btree *,
|
|
unsigned, enum btree_id);
|
|
|
|
void bch2_node_pin(struct bch_fs *, struct btree *);
|
|
void bch2_btree_cache_unpin(struct bch_fs *);
|
|
|
|
void bch2_btree_node_update_key_early(struct btree_trans *, enum btree_id, unsigned,
|
|
struct bkey_s_c, struct bkey_i *);
|
|
|
|
void bch2_btree_cache_cannibalize_unlock(struct btree_trans *);
|
|
int bch2_btree_cache_cannibalize_lock(struct btree_trans *, struct closure *);
|
|
|
|
struct btree *__bch2_btree_node_mem_alloc(struct bch_fs *);
|
|
struct btree *bch2_btree_node_mem_alloc(struct btree_trans *, bool);
|
|
|
|
struct btree *bch2_btree_node_get(struct btree_trans *, struct btree_path *,
|
|
const struct bkey_i *, unsigned,
|
|
enum six_lock_type, unsigned long);
|
|
|
|
struct btree *bch2_btree_node_get_noiter(struct btree_trans *, const struct bkey_i *,
|
|
enum btree_id, unsigned, bool);
|
|
|
|
int bch2_btree_node_prefetch(struct btree_trans *, struct btree_path *,
|
|
const struct bkey_i *, enum btree_id, unsigned);
|
|
|
|
void bch2_btree_node_evict(struct btree_trans *, const struct bkey_i *);
|
|
|
|
void bch2_fs_btree_cache_exit(struct bch_fs *);
|
|
int bch2_fs_btree_cache_init(struct bch_fs *);
|
|
void bch2_fs_btree_cache_init_early(struct btree_cache *);
|
|
|
|
static inline u64 btree_ptr_hash_val(const struct bkey_i *k)
|
|
{
|
|
switch (k->k.type) {
|
|
case KEY_TYPE_btree_ptr:
|
|
return *((u64 *) bkey_i_to_btree_ptr_c(k)->v.start);
|
|
case KEY_TYPE_btree_ptr_v2:
|
|
/*
|
|
* The cast/deref is only necessary to avoid sparse endianness
|
|
* warnings:
|
|
*/
|
|
return *((u64 *) &bkey_i_to_btree_ptr_v2_c(k)->v.seq);
|
|
default:
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
static inline struct btree *btree_node_mem_ptr(const struct bkey_i *k)
|
|
{
|
|
return k->k.type == KEY_TYPE_btree_ptr_v2
|
|
? (void *)(unsigned long)bkey_i_to_btree_ptr_v2_c(k)->v.mem_ptr
|
|
: NULL;
|
|
}
|
|
|
|
/* is btree node in hash table? */
|
|
static inline bool btree_node_hashed(struct btree *b)
|
|
{
|
|
return b->hash_val != 0;
|
|
}
|
|
|
|
#define for_each_cached_btree(_b, _c, _tbl, _iter, _pos) \
|
|
for ((_tbl) = rht_dereference_rcu((_c)->btree_cache.table.tbl, \
|
|
&(_c)->btree_cache.table), \
|
|
_iter = 0; _iter < (_tbl)->size; _iter++) \
|
|
rht_for_each_entry_rcu((_b), (_pos), _tbl, _iter, hash)
|
|
|
|
static inline size_t btree_buf_bytes(const struct btree *b)
|
|
{
|
|
return 1UL << b->byte_order;
|
|
}
|
|
|
|
static inline size_t btree_buf_max_u64s(const struct btree *b)
|
|
{
|
|
return (btree_buf_bytes(b) - sizeof(struct btree_node)) / sizeof(u64);
|
|
}
|
|
|
|
static inline size_t btree_max_u64s(const struct bch_fs *c)
|
|
{
|
|
return (c->opts.btree_node_size - sizeof(struct btree_node)) / sizeof(u64);
|
|
}
|
|
|
|
static inline size_t btree_sectors(const struct bch_fs *c)
|
|
{
|
|
return c->opts.btree_node_size >> SECTOR_SHIFT;
|
|
}
|
|
|
|
static inline unsigned btree_blocks(const struct bch_fs *c)
|
|
{
|
|
return btree_sectors(c) >> c->block_bits;
|
|
}
|
|
|
|
#define BTREE_SPLIT_THRESHOLD(c) (btree_max_u64s(c) * 2 / 3)
|
|
|
|
#define BTREE_FOREGROUND_MERGE_THRESHOLD(c) (btree_max_u64s(c) * 1 / 3)
|
|
#define BTREE_FOREGROUND_MERGE_HYSTERESIS(c) \
|
|
(BTREE_FOREGROUND_MERGE_THRESHOLD(c) + \
|
|
(BTREE_FOREGROUND_MERGE_THRESHOLD(c) >> 2))
|
|
|
|
static inline unsigned btree_id_nr_alive(struct bch_fs *c)
|
|
{
|
|
return BTREE_ID_NR + c->btree_roots_extra.nr;
|
|
}
|
|
|
|
static inline struct btree_root *bch2_btree_id_root(struct bch_fs *c, unsigned id)
|
|
{
|
|
if (likely(id < BTREE_ID_NR)) {
|
|
return &c->btree_roots_known[id];
|
|
} else {
|
|
unsigned idx = id - BTREE_ID_NR;
|
|
|
|
EBUG_ON(idx >= c->btree_roots_extra.nr);
|
|
return &c->btree_roots_extra.data[idx];
|
|
}
|
|
}
|
|
|
|
static inline struct btree *btree_node_root(struct bch_fs *c, struct btree *b)
|
|
{
|
|
return bch2_btree_id_root(c, b->c.btree_id)->b;
|
|
}
|
|
|
|
const char *bch2_btree_id_str(enum btree_id);
|
|
void bch2_btree_id_to_text(struct printbuf *, enum btree_id);
|
|
|
|
void bch2_btree_pos_to_text(struct printbuf *, struct bch_fs *, const struct btree *);
|
|
void bch2_btree_node_to_text(struct printbuf *, struct bch_fs *, const struct btree *);
|
|
void bch2_btree_cache_to_text(struct printbuf *, const struct btree_cache *);
|
|
|
|
#endif /* _BCACHEFS_BTREE_CACHE_H */
|