mirror of
https://github.com/torvalds/linux.git
synced 2024-11-25 21:51:40 +00:00
94a0333b92
A concurrent file creation and little writing could unexpectedly return -ENOSPC error since there is a race window that the allocator could get the wrong agf->agf_longest. Write file process steps: 1) Find the entry that best meets the conditions, then calculate the start address and length of the remaining part of the entry after allocation. 2) Delete this entry and update the -current- agf->agf_longest. 3) Insert the remaining unused parts of this entry based on the calculations in 1), and update the agf->agf_longest again if necessary. Create file process steps: 1) Check whether there are free inodes in the inode chunk. 2) If there is no free inode, check whether there has space for creating inode chunks, perform the no-lock judgment first. 3) If the judgment succeeds, the judgment is performed again with agf lock held. Otherwire, an error is returned directly. If the write process is in step 2) but not go to 3) yet, the create file process goes to 2) at this time, it may be mistaken for no space, resulting in the file system still has space but the file creation fails. We have sent two different commits to the community in order to fix this problem[1][2]. Unfortunately, both solutions have flaws. In [2], I discussed with Dave and Darrick, realized that a better solution to this problem requires the "last cnt record tracking" to be ripped out of the generic btree code. And surprisingly, Dave directly provided his fix code. This patch includes appropriate modifications based on his tmp-code to address this issue. The entire fix can be roughly divided into two parts: 1) Delete the code related to lastrec-update in the generic btree code. 2) Place the process of updating longest freespace with cntbt separately to the end of the cntbt modifications. Move the cursor to the rightmost firstly, and update the longest free extent based on the record. Note that we can not update the longest with xfs_alloc_get_rec() after find the longest record, as xfs_verify_agbno() may not pass because pag->block_count is updated on the outside. Therefore, use xfs_btree_get_rec() as a replacement. [1] https://lore.kernel.org/all/20240419061848.1032366-2-yebin10@huawei.com [2] https://lore.kernel.org/all/20240604071121.3981686-1-wozizhi@huawei.com Reported by: Ye Bin <yebin10@huawei.com> Signed-off-by: Zizhi Wo <wozizhi@huawei.com> Reviewed-by: Darrick J. Wong <djwong@kernel.org> Signed-off-by: Chandan Babu R <chandanbabu@kernel.org>
692 lines
21 KiB
C
692 lines
21 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* Copyright (c) 2000-2001,2005 Silicon Graphics, Inc.
|
|
* All Rights Reserved.
|
|
*/
|
|
#ifndef __XFS_BTREE_H__
|
|
#define __XFS_BTREE_H__
|
|
|
|
struct xfs_buf;
|
|
struct xfs_inode;
|
|
struct xfs_mount;
|
|
struct xfs_trans;
|
|
struct xfs_ifork;
|
|
struct xfs_perag;
|
|
|
|
/*
|
|
* Generic key, ptr and record wrapper structures.
|
|
*
|
|
* These are disk format structures, and are converted where necessary
|
|
* by the btree specific code that needs to interpret them.
|
|
*/
|
|
union xfs_btree_ptr {
|
|
__be32 s; /* short form ptr */
|
|
__be64 l; /* long form ptr */
|
|
};
|
|
|
|
/*
|
|
* The in-core btree key. Overlapping btrees actually store two keys
|
|
* per pointer, so we reserve enough memory to hold both. The __*bigkey
|
|
* items should never be accessed directly.
|
|
*/
|
|
union xfs_btree_key {
|
|
struct xfs_bmbt_key bmbt;
|
|
xfs_bmdr_key_t bmbr; /* bmbt root block */
|
|
xfs_alloc_key_t alloc;
|
|
struct xfs_inobt_key inobt;
|
|
struct xfs_rmap_key rmap;
|
|
struct xfs_rmap_key __rmap_bigkey[2];
|
|
struct xfs_refcount_key refc;
|
|
};
|
|
|
|
union xfs_btree_rec {
|
|
struct xfs_bmbt_rec bmbt;
|
|
xfs_bmdr_rec_t bmbr; /* bmbt root block */
|
|
struct xfs_alloc_rec alloc;
|
|
struct xfs_inobt_rec inobt;
|
|
struct xfs_rmap_rec rmap;
|
|
struct xfs_refcount_rec refc;
|
|
};
|
|
|
|
/*
|
|
* This nonsense is to make -wlint happy.
|
|
*/
|
|
#define XFS_LOOKUP_EQ ((xfs_lookup_t)XFS_LOOKUP_EQi)
|
|
#define XFS_LOOKUP_LE ((xfs_lookup_t)XFS_LOOKUP_LEi)
|
|
#define XFS_LOOKUP_GE ((xfs_lookup_t)XFS_LOOKUP_GEi)
|
|
|
|
struct xfs_btree_ops;
|
|
uint32_t xfs_btree_magic(struct xfs_mount *mp, const struct xfs_btree_ops *ops);
|
|
|
|
/*
|
|
* For logging record fields.
|
|
*/
|
|
#define XFS_BB_MAGIC (1u << 0)
|
|
#define XFS_BB_LEVEL (1u << 1)
|
|
#define XFS_BB_NUMRECS (1u << 2)
|
|
#define XFS_BB_LEFTSIB (1u << 3)
|
|
#define XFS_BB_RIGHTSIB (1u << 4)
|
|
#define XFS_BB_BLKNO (1u << 5)
|
|
#define XFS_BB_LSN (1u << 6)
|
|
#define XFS_BB_UUID (1u << 7)
|
|
#define XFS_BB_OWNER (1u << 8)
|
|
#define XFS_BB_NUM_BITS 5
|
|
#define XFS_BB_ALL_BITS ((1u << XFS_BB_NUM_BITS) - 1)
|
|
#define XFS_BB_NUM_BITS_CRC 9
|
|
#define XFS_BB_ALL_BITS_CRC ((1u << XFS_BB_NUM_BITS_CRC) - 1)
|
|
|
|
/*
|
|
* Generic stats interface
|
|
*/
|
|
#define XFS_BTREE_STATS_INC(cur, stat) \
|
|
XFS_STATS_INC_OFF((cur)->bc_mp, \
|
|
(cur)->bc_ops->statoff + __XBTS_ ## stat)
|
|
#define XFS_BTREE_STATS_ADD(cur, stat, val) \
|
|
XFS_STATS_ADD_OFF((cur)->bc_mp, \
|
|
(cur)->bc_ops->statoff + __XBTS_ ## stat, val)
|
|
|
|
enum xbtree_key_contig {
|
|
XBTREE_KEY_GAP = 0,
|
|
XBTREE_KEY_CONTIGUOUS,
|
|
XBTREE_KEY_OVERLAP,
|
|
};
|
|
|
|
/*
|
|
* Decide if these two numeric btree key fields are contiguous, overlapping,
|
|
* or if there's a gap between them. @x should be the field from the high
|
|
* key and @y should be the field from the low key.
|
|
*/
|
|
static inline enum xbtree_key_contig xbtree_key_contig(uint64_t x, uint64_t y)
|
|
{
|
|
x++;
|
|
if (x < y)
|
|
return XBTREE_KEY_GAP;
|
|
if (x == y)
|
|
return XBTREE_KEY_CONTIGUOUS;
|
|
return XBTREE_KEY_OVERLAP;
|
|
}
|
|
|
|
#define XFS_BTREE_LONG_PTR_LEN (sizeof(__be64))
|
|
#define XFS_BTREE_SHORT_PTR_LEN (sizeof(__be32))
|
|
|
|
enum xfs_btree_type {
|
|
XFS_BTREE_TYPE_AG,
|
|
XFS_BTREE_TYPE_INODE,
|
|
XFS_BTREE_TYPE_MEM,
|
|
};
|
|
|
|
struct xfs_btree_ops {
|
|
const char *name;
|
|
|
|
/* Type of btree - AG-rooted or inode-rooted */
|
|
enum xfs_btree_type type;
|
|
|
|
/* XFS_BTGEO_* flags that determine the geometry of the btree */
|
|
unsigned int geom_flags;
|
|
|
|
/* size of the key, pointer, and record structures */
|
|
size_t key_len;
|
|
size_t ptr_len;
|
|
size_t rec_len;
|
|
|
|
/* LRU refcount to set on each btree buffer created */
|
|
unsigned int lru_refs;
|
|
|
|
/* offset of btree stats array */
|
|
unsigned int statoff;
|
|
|
|
/* sick mask for health reporting (only for XFS_BTREE_TYPE_AG) */
|
|
unsigned int sick_mask;
|
|
|
|
/* cursor operations */
|
|
struct xfs_btree_cur *(*dup_cursor)(struct xfs_btree_cur *);
|
|
void (*update_cursor)(struct xfs_btree_cur *src,
|
|
struct xfs_btree_cur *dst);
|
|
|
|
/* update btree root pointer */
|
|
void (*set_root)(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_ptr *nptr, int level_change);
|
|
|
|
/* block allocation / freeing */
|
|
int (*alloc_block)(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_ptr *start_bno,
|
|
union xfs_btree_ptr *new_bno,
|
|
int *stat);
|
|
int (*free_block)(struct xfs_btree_cur *cur, struct xfs_buf *bp);
|
|
|
|
/* records in block/level */
|
|
int (*get_minrecs)(struct xfs_btree_cur *cur, int level);
|
|
int (*get_maxrecs)(struct xfs_btree_cur *cur, int level);
|
|
|
|
/* records on disk. Matter for the root in inode case. */
|
|
int (*get_dmaxrecs)(struct xfs_btree_cur *cur, int level);
|
|
|
|
/* init values of btree structures */
|
|
void (*init_key_from_rec)(union xfs_btree_key *key,
|
|
const union xfs_btree_rec *rec);
|
|
void (*init_rec_from_cur)(struct xfs_btree_cur *cur,
|
|
union xfs_btree_rec *rec);
|
|
void (*init_ptr_from_cur)(struct xfs_btree_cur *cur,
|
|
union xfs_btree_ptr *ptr);
|
|
void (*init_high_key_from_rec)(union xfs_btree_key *key,
|
|
const union xfs_btree_rec *rec);
|
|
|
|
/* difference between key value and cursor value */
|
|
int64_t (*key_diff)(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key);
|
|
|
|
/*
|
|
* Difference between key2 and key1 -- positive if key1 > key2,
|
|
* negative if key1 < key2, and zero if equal. If the @mask parameter
|
|
* is non NULL, each key field to be used in the comparison must
|
|
* contain a nonzero value.
|
|
*/
|
|
int64_t (*diff_two_keys)(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key1,
|
|
const union xfs_btree_key *key2,
|
|
const union xfs_btree_key *mask);
|
|
|
|
const struct xfs_buf_ops *buf_ops;
|
|
|
|
/* check that k1 is lower than k2 */
|
|
int (*keys_inorder)(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *k1,
|
|
const union xfs_btree_key *k2);
|
|
|
|
/* check that r1 is lower than r2 */
|
|
int (*recs_inorder)(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_rec *r1,
|
|
const union xfs_btree_rec *r2);
|
|
|
|
/*
|
|
* Are these two btree keys immediately adjacent?
|
|
*
|
|
* Given two btree keys @key1 and @key2, decide if it is impossible for
|
|
* there to be a third btree key K satisfying the relationship
|
|
* @key1 < K < @key2. To determine if two btree records are
|
|
* immediately adjacent, @key1 should be the high key of the first
|
|
* record and @key2 should be the low key of the second record.
|
|
* If the @mask parameter is non NULL, each key field to be used in the
|
|
* comparison must contain a nonzero value.
|
|
*/
|
|
enum xbtree_key_contig (*keys_contiguous)(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key1,
|
|
const union xfs_btree_key *key2,
|
|
const union xfs_btree_key *mask);
|
|
};
|
|
|
|
/* btree geometry flags */
|
|
#define XFS_BTGEO_OVERLAPPING (1U << 0) /* overlapping intervals */
|
|
|
|
|
|
union xfs_btree_irec {
|
|
struct xfs_alloc_rec_incore a;
|
|
struct xfs_bmbt_irec b;
|
|
struct xfs_inobt_rec_incore i;
|
|
struct xfs_rmap_irec r;
|
|
struct xfs_refcount_irec rc;
|
|
};
|
|
|
|
struct xfs_btree_level {
|
|
/* buffer pointer */
|
|
struct xfs_buf *bp;
|
|
|
|
/* key/record number */
|
|
uint16_t ptr;
|
|
|
|
/* readahead info */
|
|
#define XFS_BTCUR_LEFTRA (1 << 0) /* left sibling has been read-ahead */
|
|
#define XFS_BTCUR_RIGHTRA (1 << 1) /* right sibling has been read-ahead */
|
|
uint16_t ra;
|
|
};
|
|
|
|
/*
|
|
* Btree cursor structure.
|
|
* This collects all information needed by the btree code in one place.
|
|
*/
|
|
struct xfs_btree_cur
|
|
{
|
|
struct xfs_trans *bc_tp; /* transaction we're in, if any */
|
|
struct xfs_mount *bc_mp; /* file system mount struct */
|
|
const struct xfs_btree_ops *bc_ops;
|
|
struct kmem_cache *bc_cache; /* cursor cache */
|
|
unsigned int bc_flags; /* btree features - below */
|
|
union xfs_btree_irec bc_rec; /* current insert/search record value */
|
|
uint8_t bc_nlevels; /* number of levels in the tree */
|
|
uint8_t bc_maxlevels; /* maximum levels for this btree type */
|
|
|
|
/* per-type information */
|
|
union {
|
|
struct {
|
|
struct xfs_inode *ip;
|
|
short forksize;
|
|
char whichfork;
|
|
struct xbtree_ifakeroot *ifake; /* for staging cursor */
|
|
} bc_ino;
|
|
struct {
|
|
struct xfs_perag *pag;
|
|
struct xfs_buf *agbp;
|
|
struct xbtree_afakeroot *afake; /* for staging cursor */
|
|
} bc_ag;
|
|
struct {
|
|
struct xfbtree *xfbtree;
|
|
struct xfs_perag *pag;
|
|
} bc_mem;
|
|
};
|
|
|
|
/* per-format private data */
|
|
union {
|
|
struct {
|
|
int allocated;
|
|
} bc_bmap; /* bmapbt */
|
|
struct {
|
|
unsigned int nr_ops; /* # record updates */
|
|
unsigned int shape_changes; /* # of extent splits */
|
|
} bc_refc; /* refcountbt */
|
|
};
|
|
|
|
/* Must be at the end of the struct! */
|
|
struct xfs_btree_level bc_levels[];
|
|
};
|
|
|
|
/*
|
|
* Compute the size of a btree cursor that can handle a btree of a given
|
|
* height. The bc_levels array handles node and leaf blocks, so its size
|
|
* is exactly nlevels.
|
|
*/
|
|
static inline size_t
|
|
xfs_btree_cur_sizeof(unsigned int nlevels)
|
|
{
|
|
return struct_size_t(struct xfs_btree_cur, bc_levels, nlevels);
|
|
}
|
|
|
|
/* cursor state flags */
|
|
/*
|
|
* The root of this btree is a fakeroot structure so that we can stage a btree
|
|
* rebuild without leaving it accessible via primary metadata. The ops struct
|
|
* is dynamically allocated and must be freed when the cursor is deleted.
|
|
*/
|
|
#define XFS_BTREE_STAGING (1U << 0)
|
|
|
|
/* We are converting a delalloc reservation (only for bmbt btrees) */
|
|
#define XFS_BTREE_BMBT_WASDEL (1U << 1)
|
|
|
|
/* For extent swap, ignore owner check in verifier (only for bmbt btrees) */
|
|
#define XFS_BTREE_BMBT_INVALID_OWNER (1U << 2)
|
|
|
|
/* Cursor is active (only for allocbt btrees) */
|
|
#define XFS_BTREE_ALLOCBT_ACTIVE (1U << 3)
|
|
|
|
#define XFS_BTREE_NOERROR 0
|
|
#define XFS_BTREE_ERROR 1
|
|
|
|
/*
|
|
* Convert from buffer to btree block header.
|
|
*/
|
|
#define XFS_BUF_TO_BLOCK(bp) ((struct xfs_btree_block *)((bp)->b_addr))
|
|
|
|
xfs_failaddr_t __xfs_btree_check_block(struct xfs_btree_cur *cur,
|
|
struct xfs_btree_block *block, int level, struct xfs_buf *bp);
|
|
int __xfs_btree_check_ptr(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_ptr *ptr, int index, int level);
|
|
|
|
/*
|
|
* Check that block header is ok.
|
|
*/
|
|
int
|
|
xfs_btree_check_block(
|
|
struct xfs_btree_cur *cur, /* btree cursor */
|
|
struct xfs_btree_block *block, /* generic btree block pointer */
|
|
int level, /* level of the btree block */
|
|
struct xfs_buf *bp); /* buffer containing block, if any */
|
|
|
|
/*
|
|
* Delete the btree cursor.
|
|
*/
|
|
void
|
|
xfs_btree_del_cursor(
|
|
struct xfs_btree_cur *cur, /* btree cursor */
|
|
int error); /* del because of error */
|
|
|
|
/*
|
|
* Duplicate the btree cursor.
|
|
* Allocate a new one, copy the record, re-get the buffers.
|
|
*/
|
|
int /* error */
|
|
xfs_btree_dup_cursor(
|
|
struct xfs_btree_cur *cur, /* input cursor */
|
|
struct xfs_btree_cur **ncur);/* output cursor */
|
|
|
|
/*
|
|
* Compute first and last byte offsets for the fields given.
|
|
* Interprets the offsets table, which contains struct field offsets.
|
|
*/
|
|
void
|
|
xfs_btree_offsets(
|
|
uint32_t fields, /* bitmask of fields */
|
|
const short *offsets,/* table of field offsets */
|
|
int nbits, /* number of bits to inspect */
|
|
int *first, /* output: first byte offset */
|
|
int *last); /* output: last byte offset */
|
|
|
|
/*
|
|
* Initialise a new btree block header
|
|
*/
|
|
void xfs_btree_init_buf(struct xfs_mount *mp, struct xfs_buf *bp,
|
|
const struct xfs_btree_ops *ops, __u16 level, __u16 numrecs,
|
|
__u64 owner);
|
|
void xfs_btree_init_block(struct xfs_mount *mp,
|
|
struct xfs_btree_block *buf, const struct xfs_btree_ops *ops,
|
|
__u16 level, __u16 numrecs, __u64 owner);
|
|
|
|
/*
|
|
* Common btree core entry points.
|
|
*/
|
|
int xfs_btree_increment(struct xfs_btree_cur *, int, int *);
|
|
int xfs_btree_decrement(struct xfs_btree_cur *, int, int *);
|
|
int xfs_btree_lookup(struct xfs_btree_cur *, xfs_lookup_t, int *);
|
|
int xfs_btree_update(struct xfs_btree_cur *, union xfs_btree_rec *);
|
|
int xfs_btree_new_iroot(struct xfs_btree_cur *, int *, int *);
|
|
int xfs_btree_insert(struct xfs_btree_cur *, int *);
|
|
int xfs_btree_delete(struct xfs_btree_cur *, int *);
|
|
int xfs_btree_get_rec(struct xfs_btree_cur *, union xfs_btree_rec **, int *);
|
|
int xfs_btree_change_owner(struct xfs_btree_cur *cur, uint64_t new_owner,
|
|
struct list_head *buffer_list);
|
|
|
|
/*
|
|
* btree block CRC helpers
|
|
*/
|
|
void xfs_btree_fsblock_calc_crc(struct xfs_buf *);
|
|
bool xfs_btree_fsblock_verify_crc(struct xfs_buf *);
|
|
void xfs_btree_agblock_calc_crc(struct xfs_buf *);
|
|
bool xfs_btree_agblock_verify_crc(struct xfs_buf *);
|
|
|
|
/*
|
|
* Internal btree helpers also used by xfs_bmap.c.
|
|
*/
|
|
void xfs_btree_log_block(struct xfs_btree_cur *, struct xfs_buf *, uint32_t);
|
|
void xfs_btree_log_recs(struct xfs_btree_cur *, struct xfs_buf *, int, int);
|
|
|
|
/*
|
|
* Helpers.
|
|
*/
|
|
static inline int xfs_btree_get_numrecs(const struct xfs_btree_block *block)
|
|
{
|
|
return be16_to_cpu(block->bb_numrecs);
|
|
}
|
|
|
|
static inline void xfs_btree_set_numrecs(struct xfs_btree_block *block,
|
|
uint16_t numrecs)
|
|
{
|
|
block->bb_numrecs = cpu_to_be16(numrecs);
|
|
}
|
|
|
|
static inline int xfs_btree_get_level(const struct xfs_btree_block *block)
|
|
{
|
|
return be16_to_cpu(block->bb_level);
|
|
}
|
|
|
|
|
|
/*
|
|
* Min and max functions for extlen, agblock, fileoff, and filblks types.
|
|
*/
|
|
#define XFS_EXTLEN_MIN(a,b) min_t(xfs_extlen_t, (a), (b))
|
|
#define XFS_EXTLEN_MAX(a,b) max_t(xfs_extlen_t, (a), (b))
|
|
#define XFS_AGBLOCK_MIN(a,b) min_t(xfs_agblock_t, (a), (b))
|
|
#define XFS_AGBLOCK_MAX(a,b) max_t(xfs_agblock_t, (a), (b))
|
|
#define XFS_FILEOFF_MIN(a,b) min_t(xfs_fileoff_t, (a), (b))
|
|
#define XFS_FILEOFF_MAX(a,b) max_t(xfs_fileoff_t, (a), (b))
|
|
#define XFS_FILBLKS_MIN(a,b) min_t(xfs_filblks_t, (a), (b))
|
|
#define XFS_FILBLKS_MAX(a,b) max_t(xfs_filblks_t, (a), (b))
|
|
|
|
xfs_failaddr_t xfs_btree_agblock_v5hdr_verify(struct xfs_buf *bp);
|
|
xfs_failaddr_t xfs_btree_agblock_verify(struct xfs_buf *bp,
|
|
unsigned int max_recs);
|
|
xfs_failaddr_t xfs_btree_fsblock_v5hdr_verify(struct xfs_buf *bp,
|
|
uint64_t owner);
|
|
xfs_failaddr_t xfs_btree_fsblock_verify(struct xfs_buf *bp,
|
|
unsigned int max_recs);
|
|
xfs_failaddr_t xfs_btree_memblock_verify(struct xfs_buf *bp,
|
|
unsigned int max_recs);
|
|
|
|
unsigned int xfs_btree_compute_maxlevels(const unsigned int *limits,
|
|
unsigned long long records);
|
|
unsigned long long xfs_btree_calc_size(const unsigned int *limits,
|
|
unsigned long long records);
|
|
unsigned int xfs_btree_space_to_height(const unsigned int *limits,
|
|
unsigned long long blocks);
|
|
|
|
/*
|
|
* Return codes for the query range iterator function are 0 to continue
|
|
* iterating, and non-zero to stop iterating. Any non-zero value will be
|
|
* passed up to the _query_range caller. The special value -ECANCELED can be
|
|
* used to stop iteration, because _query_range never generates that error
|
|
* code on its own.
|
|
*/
|
|
typedef int (*xfs_btree_query_range_fn)(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_rec *rec, void *priv);
|
|
|
|
int xfs_btree_query_range(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_irec *low_rec,
|
|
const union xfs_btree_irec *high_rec,
|
|
xfs_btree_query_range_fn fn, void *priv);
|
|
int xfs_btree_query_all(struct xfs_btree_cur *cur, xfs_btree_query_range_fn fn,
|
|
void *priv);
|
|
|
|
typedef int (*xfs_btree_visit_blocks_fn)(struct xfs_btree_cur *cur, int level,
|
|
void *data);
|
|
/* Visit record blocks. */
|
|
#define XFS_BTREE_VISIT_RECORDS (1 << 0)
|
|
/* Visit leaf blocks. */
|
|
#define XFS_BTREE_VISIT_LEAVES (1 << 1)
|
|
/* Visit all blocks. */
|
|
#define XFS_BTREE_VISIT_ALL (XFS_BTREE_VISIT_RECORDS | \
|
|
XFS_BTREE_VISIT_LEAVES)
|
|
int xfs_btree_visit_blocks(struct xfs_btree_cur *cur,
|
|
xfs_btree_visit_blocks_fn fn, unsigned int flags, void *data);
|
|
|
|
int xfs_btree_count_blocks(struct xfs_btree_cur *cur, xfs_extlen_t *blocks);
|
|
|
|
union xfs_btree_rec *xfs_btree_rec_addr(struct xfs_btree_cur *cur, int n,
|
|
struct xfs_btree_block *block);
|
|
union xfs_btree_key *xfs_btree_key_addr(struct xfs_btree_cur *cur, int n,
|
|
struct xfs_btree_block *block);
|
|
union xfs_btree_key *xfs_btree_high_key_addr(struct xfs_btree_cur *cur, int n,
|
|
struct xfs_btree_block *block);
|
|
union xfs_btree_ptr *xfs_btree_ptr_addr(struct xfs_btree_cur *cur, int n,
|
|
struct xfs_btree_block *block);
|
|
int xfs_btree_lookup_get_block(struct xfs_btree_cur *cur, int level,
|
|
const union xfs_btree_ptr *pp, struct xfs_btree_block **blkp);
|
|
struct xfs_btree_block *xfs_btree_get_block(struct xfs_btree_cur *cur,
|
|
int level, struct xfs_buf **bpp);
|
|
bool xfs_btree_ptr_is_null(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_ptr *ptr);
|
|
int64_t xfs_btree_diff_two_ptrs(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_ptr *a,
|
|
const union xfs_btree_ptr *b);
|
|
void xfs_btree_get_sibling(struct xfs_btree_cur *cur,
|
|
struct xfs_btree_block *block,
|
|
union xfs_btree_ptr *ptr, int lr);
|
|
void xfs_btree_get_keys(struct xfs_btree_cur *cur,
|
|
struct xfs_btree_block *block, union xfs_btree_key *key);
|
|
union xfs_btree_key *xfs_btree_high_key_from_key(struct xfs_btree_cur *cur,
|
|
union xfs_btree_key *key);
|
|
typedef bool (*xfs_btree_key_gap_fn)(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key1,
|
|
const union xfs_btree_key *key2);
|
|
|
|
int xfs_btree_has_records(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_irec *low,
|
|
const union xfs_btree_irec *high,
|
|
const union xfs_btree_key *mask,
|
|
enum xbtree_recpacking *outcome);
|
|
|
|
bool xfs_btree_has_more_records(struct xfs_btree_cur *cur);
|
|
struct xfs_ifork *xfs_btree_ifork_ptr(struct xfs_btree_cur *cur);
|
|
|
|
/* Key comparison helpers */
|
|
static inline bool
|
|
xfs_btree_keycmp_lt(
|
|
struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key1,
|
|
const union xfs_btree_key *key2)
|
|
{
|
|
return cur->bc_ops->diff_two_keys(cur, key1, key2, NULL) < 0;
|
|
}
|
|
|
|
static inline bool
|
|
xfs_btree_keycmp_gt(
|
|
struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key1,
|
|
const union xfs_btree_key *key2)
|
|
{
|
|
return cur->bc_ops->diff_two_keys(cur, key1, key2, NULL) > 0;
|
|
}
|
|
|
|
static inline bool
|
|
xfs_btree_keycmp_eq(
|
|
struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key1,
|
|
const union xfs_btree_key *key2)
|
|
{
|
|
return cur->bc_ops->diff_two_keys(cur, key1, key2, NULL) == 0;
|
|
}
|
|
|
|
static inline bool
|
|
xfs_btree_keycmp_le(
|
|
struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key1,
|
|
const union xfs_btree_key *key2)
|
|
{
|
|
return !xfs_btree_keycmp_gt(cur, key1, key2);
|
|
}
|
|
|
|
static inline bool
|
|
xfs_btree_keycmp_ge(
|
|
struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key1,
|
|
const union xfs_btree_key *key2)
|
|
{
|
|
return !xfs_btree_keycmp_lt(cur, key1, key2);
|
|
}
|
|
|
|
static inline bool
|
|
xfs_btree_keycmp_ne(
|
|
struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key1,
|
|
const union xfs_btree_key *key2)
|
|
{
|
|
return !xfs_btree_keycmp_eq(cur, key1, key2);
|
|
}
|
|
|
|
/* Masked key comparison helpers */
|
|
static inline bool
|
|
xfs_btree_masked_keycmp_lt(
|
|
struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key1,
|
|
const union xfs_btree_key *key2,
|
|
const union xfs_btree_key *mask)
|
|
{
|
|
return cur->bc_ops->diff_two_keys(cur, key1, key2, mask) < 0;
|
|
}
|
|
|
|
static inline bool
|
|
xfs_btree_masked_keycmp_gt(
|
|
struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key1,
|
|
const union xfs_btree_key *key2,
|
|
const union xfs_btree_key *mask)
|
|
{
|
|
return cur->bc_ops->diff_two_keys(cur, key1, key2, mask) > 0;
|
|
}
|
|
|
|
static inline bool
|
|
xfs_btree_masked_keycmp_ge(
|
|
struct xfs_btree_cur *cur,
|
|
const union xfs_btree_key *key1,
|
|
const union xfs_btree_key *key2,
|
|
const union xfs_btree_key *mask)
|
|
{
|
|
return !xfs_btree_masked_keycmp_lt(cur, key1, key2, mask);
|
|
}
|
|
|
|
/* Does this cursor point to the last block in the given level? */
|
|
static inline bool
|
|
xfs_btree_islastblock(
|
|
struct xfs_btree_cur *cur,
|
|
int level)
|
|
{
|
|
struct xfs_btree_block *block;
|
|
struct xfs_buf *bp;
|
|
|
|
block = xfs_btree_get_block(cur, level, &bp);
|
|
|
|
if (cur->bc_ops->ptr_len == XFS_BTREE_LONG_PTR_LEN)
|
|
return block->bb_u.l.bb_rightsib == cpu_to_be64(NULLFSBLOCK);
|
|
return block->bb_u.s.bb_rightsib == cpu_to_be32(NULLAGBLOCK);
|
|
}
|
|
|
|
void xfs_btree_set_ptr_null(struct xfs_btree_cur *cur,
|
|
union xfs_btree_ptr *ptr);
|
|
int xfs_btree_get_buf_block(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_ptr *ptr, struct xfs_btree_block **block,
|
|
struct xfs_buf **bpp);
|
|
int xfs_btree_read_buf_block(struct xfs_btree_cur *cur,
|
|
const union xfs_btree_ptr *ptr, int flags,
|
|
struct xfs_btree_block **block, struct xfs_buf **bpp);
|
|
void xfs_btree_set_sibling(struct xfs_btree_cur *cur,
|
|
struct xfs_btree_block *block, const union xfs_btree_ptr *ptr,
|
|
int lr);
|
|
void xfs_btree_init_block_cur(struct xfs_btree_cur *cur,
|
|
struct xfs_buf *bp, int level, int numrecs);
|
|
void xfs_btree_copy_ptrs(struct xfs_btree_cur *cur,
|
|
union xfs_btree_ptr *dst_ptr,
|
|
const union xfs_btree_ptr *src_ptr, int numptrs);
|
|
void xfs_btree_copy_keys(struct xfs_btree_cur *cur,
|
|
union xfs_btree_key *dst_key,
|
|
const union xfs_btree_key *src_key, int numkeys);
|
|
void xfs_btree_init_ptr_from_cur(struct xfs_btree_cur *cur,
|
|
union xfs_btree_ptr *ptr);
|
|
|
|
static inline struct xfs_btree_cur *
|
|
xfs_btree_alloc_cursor(
|
|
struct xfs_mount *mp,
|
|
struct xfs_trans *tp,
|
|
const struct xfs_btree_ops *ops,
|
|
uint8_t maxlevels,
|
|
struct kmem_cache *cache)
|
|
{
|
|
struct xfs_btree_cur *cur;
|
|
|
|
ASSERT(ops->ptr_len == XFS_BTREE_LONG_PTR_LEN ||
|
|
ops->ptr_len == XFS_BTREE_SHORT_PTR_LEN);
|
|
|
|
/* BMBT allocations can come through from non-transactional context. */
|
|
cur = kmem_cache_zalloc(cache,
|
|
GFP_KERNEL | __GFP_NOLOCKDEP | __GFP_NOFAIL);
|
|
cur->bc_ops = ops;
|
|
cur->bc_tp = tp;
|
|
cur->bc_mp = mp;
|
|
cur->bc_maxlevels = maxlevels;
|
|
cur->bc_cache = cache;
|
|
|
|
return cur;
|
|
}
|
|
|
|
int __init xfs_btree_init_cur_caches(void);
|
|
void xfs_btree_destroy_cur_caches(void);
|
|
|
|
int xfs_btree_goto_left_edge(struct xfs_btree_cur *cur);
|
|
|
|
/* Does this level of the cursor point to the inode root (and not a block)? */
|
|
static inline bool
|
|
xfs_btree_at_iroot(
|
|
const struct xfs_btree_cur *cur,
|
|
int level)
|
|
{
|
|
return cur->bc_ops->type == XFS_BTREE_TYPE_INODE &&
|
|
level == cur->bc_nlevels - 1;
|
|
}
|
|
|
|
#endif /* __XFS_BTREE_H__ */
|