mirror of
https://github.com/torvalds/linux.git
synced 2024-11-25 13:41:51 +00:00
a393f33123
Buckets usually can't be discarded until the transaction that made them empty has been committed in the journal. Tracing has indicated that we're queuing the discard worker excessively, only for it to skip over many buckets that are still waiting on a journal commit, discarding only one or two buckets per iteration. We want to switch to only queuing the discard worker after a journal flush write, but there's an important optimization we need to preserve: if a bucket becomes empty and it was never committed in the journal while it was in use, we want to discard it and reuse it right away - since overwriting it before the previous writes are flushed from the device cache eans those writes only cost bus bandwidth. So, this patch implements a fast path for buckets that can be discarded right away. We need new locking between the two discard workers; the new list of buckets being discarded provides that locking. Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
276 lines
7.6 KiB
C
276 lines
7.6 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef _BCACHEFS_ALLOC_BACKGROUND_H
|
|
#define _BCACHEFS_ALLOC_BACKGROUND_H
|
|
|
|
#include "bcachefs.h"
|
|
#include "alloc_types.h"
|
|
#include "buckets.h"
|
|
#include "debug.h"
|
|
#include "super.h"
|
|
|
|
enum bkey_invalid_flags;
|
|
|
|
/* How out of date a pointer gen is allowed to be: */
|
|
#define BUCKET_GC_GEN_MAX 96U
|
|
|
|
static inline bool bch2_dev_bucket_exists(struct bch_fs *c, struct bpos pos)
|
|
{
|
|
struct bch_dev *ca;
|
|
|
|
if (!bch2_dev_exists2(c, pos.inode))
|
|
return false;
|
|
|
|
ca = bch_dev_bkey_exists(c, pos.inode);
|
|
return pos.offset >= ca->mi.first_bucket &&
|
|
pos.offset < ca->mi.nbuckets;
|
|
}
|
|
|
|
static inline u64 bucket_to_u64(struct bpos bucket)
|
|
{
|
|
return (bucket.inode << 48) | bucket.offset;
|
|
}
|
|
|
|
static inline struct bpos u64_to_bucket(u64 bucket)
|
|
{
|
|
return POS(bucket >> 48, bucket & ~(~0ULL << 48));
|
|
}
|
|
|
|
static inline u8 alloc_gc_gen(struct bch_alloc_v4 a)
|
|
{
|
|
return a.gen - a.oldest_gen;
|
|
}
|
|
|
|
static inline enum bch_data_type __alloc_data_type(u32 dirty_sectors,
|
|
u32 cached_sectors,
|
|
u32 stripe,
|
|
struct bch_alloc_v4 a,
|
|
enum bch_data_type data_type)
|
|
{
|
|
if (stripe)
|
|
return data_type == BCH_DATA_parity ? data_type : BCH_DATA_stripe;
|
|
if (dirty_sectors)
|
|
return data_type;
|
|
if (cached_sectors)
|
|
return BCH_DATA_cached;
|
|
if (BCH_ALLOC_V4_NEED_DISCARD(&a))
|
|
return BCH_DATA_need_discard;
|
|
if (alloc_gc_gen(a) >= BUCKET_GC_GEN_MAX)
|
|
return BCH_DATA_need_gc_gens;
|
|
return BCH_DATA_free;
|
|
}
|
|
|
|
static inline enum bch_data_type alloc_data_type(struct bch_alloc_v4 a,
|
|
enum bch_data_type data_type)
|
|
{
|
|
return __alloc_data_type(a.dirty_sectors, a.cached_sectors,
|
|
a.stripe, a, data_type);
|
|
}
|
|
|
|
static inline enum bch_data_type bucket_data_type(enum bch_data_type data_type)
|
|
{
|
|
return data_type == BCH_DATA_stripe ? BCH_DATA_user : data_type;
|
|
}
|
|
|
|
static inline unsigned bch2_bucket_sectors(struct bch_alloc_v4 a)
|
|
{
|
|
return a.dirty_sectors + a.cached_sectors;
|
|
}
|
|
|
|
static inline unsigned bch2_bucket_sectors_dirty(struct bch_alloc_v4 a)
|
|
{
|
|
return a.dirty_sectors;
|
|
}
|
|
|
|
static inline unsigned bch2_bucket_sectors_fragmented(struct bch_dev *ca,
|
|
struct bch_alloc_v4 a)
|
|
{
|
|
int d = bch2_bucket_sectors_dirty(a);
|
|
|
|
return d ? max(0, ca->mi.bucket_size - d) : 0;
|
|
}
|
|
|
|
static inline u64 alloc_lru_idx_read(struct bch_alloc_v4 a)
|
|
{
|
|
return a.data_type == BCH_DATA_cached ? a.io_time[READ] : 0;
|
|
}
|
|
|
|
#define DATA_TYPES_MOVABLE \
|
|
((1U << BCH_DATA_btree)| \
|
|
(1U << BCH_DATA_user)| \
|
|
(1U << BCH_DATA_stripe))
|
|
|
|
static inline bool data_type_movable(enum bch_data_type type)
|
|
{
|
|
return (1U << type) & DATA_TYPES_MOVABLE;
|
|
}
|
|
|
|
static inline u64 alloc_lru_idx_fragmentation(struct bch_alloc_v4 a,
|
|
struct bch_dev *ca)
|
|
{
|
|
if (!data_type_movable(a.data_type) ||
|
|
!bch2_bucket_sectors_fragmented(ca, a))
|
|
return 0;
|
|
|
|
u64 d = bch2_bucket_sectors_dirty(a);
|
|
return div_u64(d * (1ULL << 31), ca->mi.bucket_size);
|
|
}
|
|
|
|
static inline u64 alloc_freespace_genbits(struct bch_alloc_v4 a)
|
|
{
|
|
return ((u64) alloc_gc_gen(a) >> 4) << 56;
|
|
}
|
|
|
|
static inline struct bpos alloc_freespace_pos(struct bpos pos, struct bch_alloc_v4 a)
|
|
{
|
|
pos.offset |= alloc_freespace_genbits(a);
|
|
return pos;
|
|
}
|
|
|
|
static inline unsigned alloc_v4_u64s(const struct bch_alloc_v4 *a)
|
|
{
|
|
unsigned ret = (BCH_ALLOC_V4_BACKPOINTERS_START(a) ?:
|
|
BCH_ALLOC_V4_U64s_V0) +
|
|
BCH_ALLOC_V4_NR_BACKPOINTERS(a) *
|
|
(sizeof(struct bch_backpointer) / sizeof(u64));
|
|
|
|
BUG_ON(ret > U8_MAX - BKEY_U64s);
|
|
return ret;
|
|
}
|
|
|
|
static inline void set_alloc_v4_u64s(struct bkey_i_alloc_v4 *a)
|
|
{
|
|
set_bkey_val_u64s(&a->k, alloc_v4_u64s(&a->v));
|
|
}
|
|
|
|
struct bkey_i_alloc_v4 *
|
|
bch2_trans_start_alloc_update(struct btree_trans *, struct btree_iter *, struct bpos);
|
|
|
|
void __bch2_alloc_to_v4(struct bkey_s_c, struct bch_alloc_v4 *);
|
|
|
|
static inline const struct bch_alloc_v4 *bch2_alloc_to_v4(struct bkey_s_c k, struct bch_alloc_v4 *convert)
|
|
{
|
|
const struct bch_alloc_v4 *ret;
|
|
|
|
if (unlikely(k.k->type != KEY_TYPE_alloc_v4))
|
|
goto slowpath;
|
|
|
|
ret = bkey_s_c_to_alloc_v4(k).v;
|
|
if (BCH_ALLOC_V4_BACKPOINTERS_START(ret) != BCH_ALLOC_V4_U64s)
|
|
goto slowpath;
|
|
|
|
return ret;
|
|
slowpath:
|
|
__bch2_alloc_to_v4(k, convert);
|
|
return convert;
|
|
}
|
|
|
|
struct bkey_i_alloc_v4 *bch2_alloc_to_v4_mut(struct btree_trans *, struct bkey_s_c);
|
|
|
|
int bch2_bucket_io_time_reset(struct btree_trans *, unsigned, size_t, int);
|
|
|
|
int bch2_alloc_v1_invalid(struct bch_fs *, struct bkey_s_c,
|
|
enum bkey_invalid_flags, struct printbuf *);
|
|
int bch2_alloc_v2_invalid(struct bch_fs *, struct bkey_s_c,
|
|
enum bkey_invalid_flags, struct printbuf *);
|
|
int bch2_alloc_v3_invalid(struct bch_fs *, struct bkey_s_c,
|
|
enum bkey_invalid_flags, struct printbuf *);
|
|
int bch2_alloc_v4_invalid(struct bch_fs *, struct bkey_s_c,
|
|
enum bkey_invalid_flags, struct printbuf *);
|
|
void bch2_alloc_v4_swab(struct bkey_s);
|
|
void bch2_alloc_to_text(struct printbuf *, struct bch_fs *, struct bkey_s_c);
|
|
|
|
#define bch2_bkey_ops_alloc ((struct bkey_ops) { \
|
|
.key_invalid = bch2_alloc_v1_invalid, \
|
|
.val_to_text = bch2_alloc_to_text, \
|
|
.trigger = bch2_trigger_alloc, \
|
|
.min_val_size = 8, \
|
|
})
|
|
|
|
#define bch2_bkey_ops_alloc_v2 ((struct bkey_ops) { \
|
|
.key_invalid = bch2_alloc_v2_invalid, \
|
|
.val_to_text = bch2_alloc_to_text, \
|
|
.trigger = bch2_trigger_alloc, \
|
|
.min_val_size = 8, \
|
|
})
|
|
|
|
#define bch2_bkey_ops_alloc_v3 ((struct bkey_ops) { \
|
|
.key_invalid = bch2_alloc_v3_invalid, \
|
|
.val_to_text = bch2_alloc_to_text, \
|
|
.trigger = bch2_trigger_alloc, \
|
|
.min_val_size = 16, \
|
|
})
|
|
|
|
#define bch2_bkey_ops_alloc_v4 ((struct bkey_ops) { \
|
|
.key_invalid = bch2_alloc_v4_invalid, \
|
|
.val_to_text = bch2_alloc_to_text, \
|
|
.swab = bch2_alloc_v4_swab, \
|
|
.trigger = bch2_trigger_alloc, \
|
|
.min_val_size = 48, \
|
|
})
|
|
|
|
int bch2_bucket_gens_invalid(struct bch_fs *, struct bkey_s_c,
|
|
enum bkey_invalid_flags, struct printbuf *);
|
|
void bch2_bucket_gens_to_text(struct printbuf *, struct bch_fs *, struct bkey_s_c);
|
|
|
|
#define bch2_bkey_ops_bucket_gens ((struct bkey_ops) { \
|
|
.key_invalid = bch2_bucket_gens_invalid, \
|
|
.val_to_text = bch2_bucket_gens_to_text, \
|
|
})
|
|
|
|
int bch2_bucket_gens_init(struct bch_fs *);
|
|
|
|
static inline bool bkey_is_alloc(const struct bkey *k)
|
|
{
|
|
return k->type == KEY_TYPE_alloc ||
|
|
k->type == KEY_TYPE_alloc_v2 ||
|
|
k->type == KEY_TYPE_alloc_v3;
|
|
}
|
|
|
|
int bch2_alloc_read(struct bch_fs *);
|
|
|
|
int bch2_trigger_alloc(struct btree_trans *, enum btree_id, unsigned,
|
|
struct bkey_s_c, struct bkey_s, unsigned);
|
|
int bch2_check_alloc_info(struct bch_fs *);
|
|
int bch2_check_alloc_to_lru_refs(struct bch_fs *);
|
|
void bch2_do_discards(struct bch_fs *);
|
|
|
|
static inline u64 should_invalidate_buckets(struct bch_dev *ca,
|
|
struct bch_dev_usage u)
|
|
{
|
|
u64 want_free = ca->mi.nbuckets >> 7;
|
|
u64 free = max_t(s64, 0,
|
|
u.d[BCH_DATA_free].buckets
|
|
+ u.d[BCH_DATA_need_discard].buckets
|
|
- bch2_dev_buckets_reserved(ca, BCH_WATERMARK_stripe));
|
|
|
|
return clamp_t(s64, want_free - free, 0, u.d[BCH_DATA_cached].buckets);
|
|
}
|
|
|
|
void bch2_do_invalidates(struct bch_fs *);
|
|
|
|
static inline struct bch_backpointer *alloc_v4_backpointers(struct bch_alloc_v4 *a)
|
|
{
|
|
return (void *) ((u64 *) &a->v +
|
|
(BCH_ALLOC_V4_BACKPOINTERS_START(a) ?:
|
|
BCH_ALLOC_V4_U64s_V0));
|
|
}
|
|
|
|
static inline const struct bch_backpointer *alloc_v4_backpointers_c(const struct bch_alloc_v4 *a)
|
|
{
|
|
return (void *) ((u64 *) &a->v + BCH_ALLOC_V4_BACKPOINTERS_START(a));
|
|
}
|
|
|
|
int bch2_dev_freespace_init(struct bch_fs *, struct bch_dev *, u64, u64);
|
|
int bch2_fs_freespace_init(struct bch_fs *);
|
|
|
|
void bch2_recalc_capacity(struct bch_fs *);
|
|
u64 bch2_min_rw_member_capacity(struct bch_fs *);
|
|
|
|
void bch2_dev_allocator_remove(struct bch_fs *, struct bch_dev *);
|
|
void bch2_dev_allocator_add(struct bch_fs *, struct bch_dev *);
|
|
|
|
void bch2_fs_allocator_background_exit(struct bch_fs *);
|
|
void bch2_fs_allocator_background_init(struct bch_fs *);
|
|
|
|
#endif /* _BCACHEFS_ALLOC_BACKGROUND_H */
|