2018-06-06 02:42:14 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
2013-08-12 10:49:33 +00:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2000-2003,2005 Silicon Graphics, Inc.
|
|
|
|
* All Rights Reserved.
|
|
|
|
*/
|
|
|
|
#ifndef __XFS_INODE_FORK_H__
|
|
|
|
#define __XFS_INODE_FORK_H__
|
|
|
|
|
|
|
|
struct xfs_inode_log_item;
|
2013-10-22 23:51:50 +00:00
|
|
|
struct xfs_dinode;
|
2013-08-12 10:49:33 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* File incore extent information, present for each of data & attr forks.
|
|
|
|
*/
|
2018-07-17 23:51:50 +00:00
|
|
|
struct xfs_ifork {
|
xfs: fix inode fork extent count overflow
[commit message is verbose for discussion purposes - will trim it
down later. Some questions about implementation details at the end.]
Zorro Lang recently ran a new test to stress single inode extent
counts now that they are no longer limited by memory allocation.
The test was simply:
# xfs_io -f -c "falloc 0 40t" /mnt/scratch/big-file
# ~/src/xfstests-dev/punch-alternating /mnt/scratch/big-file
This test uncovered a problem where the hole punching operation
appeared to finish with no error, but apparently only created 268M
extents instead of the 10 billion it was supposed to.
Further, trying to punch out extents that should have been present
resulted in success, but no change in the extent count. It looked
like a silent failure.
While running the test and observing the behaviour in real time,
I observed the extent coutn growing at ~2M extents/minute, and saw
this after about an hour:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next ; \
> sleep 60 ; \
> xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 127657993
fsxattr.nextents = 129683339
#
And a few minutes later this:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 4177861124
#
Ah, what? Where did that 4 billion extra extents suddenly come from?
Stop the workload, unmount, mount:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 166044375
#
And it's back at the expected number. i.e. the extent count is
correct on disk, but it's screwed up in memory. I loaded up the
extent list, and immediately:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 4192576215
#
It's bad again. So, where does that number come from?
xfs_fill_fsxattr():
if (ip->i_df.if_flags & XFS_IFEXTENTS)
fa->fsx_nextents = xfs_iext_count(&ip->i_df);
else
fa->fsx_nextents = ip->i_d.di_nextents;
And that's the behaviour I just saw in a nutshell. The on disk count
is correct, but once the tree is loaded into memory, it goes whacky.
Clearly there's something wrong with xfs_iext_count():
inline xfs_extnum_t xfs_iext_count(struct xfs_ifork *ifp)
{
return ifp->if_bytes / sizeof(struct xfs_iext_rec);
}
Simple enough, but 134M extents is 2**27, and that's right about
where things went wrong. A struct xfs_iext_rec is 16 bytes in size,
which means 2**27 * 2**4 = 2**31 and we're right on target for an
integer overflow. And, sure enough:
struct xfs_ifork {
int if_bytes; /* bytes in if_u1 */
....
Once we get 2**27 extents in a file, we overflow if_bytes and the
in-core extent count goes wrong. And when we reach 2**28 extents,
if_bytes wraps back to zero and things really start to go wrong
there. This is where the silent failure comes from - only the first
2**28 extents can be looked up directly due to the overflow, all the
extents above this index wrap back to somewhere in the first 2**28
extents. Hence with a regular pattern, trying to punch a hole in the
range that didn't have holes mapped to a hole in the first 2**28
extents and so "succeeded" without changing anything. Hence "silent
failure"...
Fix this by converting if_bytes to a int64_t and converting all the
index variables and size calculations to use int64_t types to avoid
overflows in future. Signed integers are still used to enable easy
detection of extent count underflows. This enables scalability of
extent counts to the limits of the on-disk format - MAXEXTNUM
(2**31) extents.
Current testing is at over 500M extents and still going:
fsxattr.nextents = 517310478
Reported-by: Zorro Lang <zlang@redhat.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com>
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
2019-10-17 20:40:33 +00:00
|
|
|
int64_t if_bytes; /* bytes in if_u1 */
|
2013-08-12 10:49:33 +00:00
|
|
|
struct xfs_btree_block *if_broot; /* file's incore btree root */
|
xfs: fix inode fork extent count overflow
[commit message is verbose for discussion purposes - will trim it
down later. Some questions about implementation details at the end.]
Zorro Lang recently ran a new test to stress single inode extent
counts now that they are no longer limited by memory allocation.
The test was simply:
# xfs_io -f -c "falloc 0 40t" /mnt/scratch/big-file
# ~/src/xfstests-dev/punch-alternating /mnt/scratch/big-file
This test uncovered a problem where the hole punching operation
appeared to finish with no error, but apparently only created 268M
extents instead of the 10 billion it was supposed to.
Further, trying to punch out extents that should have been present
resulted in success, but no change in the extent count. It looked
like a silent failure.
While running the test and observing the behaviour in real time,
I observed the extent coutn growing at ~2M extents/minute, and saw
this after about an hour:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next ; \
> sleep 60 ; \
> xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 127657993
fsxattr.nextents = 129683339
#
And a few minutes later this:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 4177861124
#
Ah, what? Where did that 4 billion extra extents suddenly come from?
Stop the workload, unmount, mount:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 166044375
#
And it's back at the expected number. i.e. the extent count is
correct on disk, but it's screwed up in memory. I loaded up the
extent list, and immediately:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 4192576215
#
It's bad again. So, where does that number come from?
xfs_fill_fsxattr():
if (ip->i_df.if_flags & XFS_IFEXTENTS)
fa->fsx_nextents = xfs_iext_count(&ip->i_df);
else
fa->fsx_nextents = ip->i_d.di_nextents;
And that's the behaviour I just saw in a nutshell. The on disk count
is correct, but once the tree is loaded into memory, it goes whacky.
Clearly there's something wrong with xfs_iext_count():
inline xfs_extnum_t xfs_iext_count(struct xfs_ifork *ifp)
{
return ifp->if_bytes / sizeof(struct xfs_iext_rec);
}
Simple enough, but 134M extents is 2**27, and that's right about
where things went wrong. A struct xfs_iext_rec is 16 bytes in size,
which means 2**27 * 2**4 = 2**31 and we're right on target for an
integer overflow. And, sure enough:
struct xfs_ifork {
int if_bytes; /* bytes in if_u1 */
....
Once we get 2**27 extents in a file, we overflow if_bytes and the
in-core extent count goes wrong. And when we reach 2**28 extents,
if_bytes wraps back to zero and things really start to go wrong
there. This is where the silent failure comes from - only the first
2**28 extents can be looked up directly due to the overflow, all the
extents above this index wrap back to somewhere in the first 2**28
extents. Hence with a regular pattern, trying to punch a hole in the
range that didn't have holes mapped to a hole in the first 2**28
extents and so "succeeded" without changing anything. Hence "silent
failure"...
Fix this by converting if_bytes to a int64_t and converting all the
index variables and size calculations to use int64_t types to avoid
overflows in future. Signed integers are still used to enable easy
detection of extent count underflows. This enables scalability of
extent counts to the limits of the on-disk format - MAXEXTNUM
(2**31) extents.
Current testing is at over 500M extents and still going:
fsxattr.nextents = 517310478
Reported-by: Zorro Lang <zlang@redhat.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com>
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
2019-10-17 20:40:33 +00:00
|
|
|
unsigned int if_seq; /* fork mod counter */
|
2017-11-03 17:34:46 +00:00
|
|
|
int if_height; /* height of the extent tree */
|
2013-08-12 10:49:33 +00:00
|
|
|
union {
|
2017-11-03 17:34:46 +00:00
|
|
|
void *if_root; /* extent tree root */
|
2013-08-12 10:49:33 +00:00
|
|
|
char *if_data; /* inline file data */
|
|
|
|
} if_u1;
|
2021-11-16 07:28:40 +00:00
|
|
|
xfs_extnum_t if_nextents; /* # of extents in this fork */
|
xfs: fix inode fork extent count overflow
[commit message is verbose for discussion purposes - will trim it
down later. Some questions about implementation details at the end.]
Zorro Lang recently ran a new test to stress single inode extent
counts now that they are no longer limited by memory allocation.
The test was simply:
# xfs_io -f -c "falloc 0 40t" /mnt/scratch/big-file
# ~/src/xfstests-dev/punch-alternating /mnt/scratch/big-file
This test uncovered a problem where the hole punching operation
appeared to finish with no error, but apparently only created 268M
extents instead of the 10 billion it was supposed to.
Further, trying to punch out extents that should have been present
resulted in success, but no change in the extent count. It looked
like a silent failure.
While running the test and observing the behaviour in real time,
I observed the extent coutn growing at ~2M extents/minute, and saw
this after about an hour:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next ; \
> sleep 60 ; \
> xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 127657993
fsxattr.nextents = 129683339
#
And a few minutes later this:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 4177861124
#
Ah, what? Where did that 4 billion extra extents suddenly come from?
Stop the workload, unmount, mount:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 166044375
#
And it's back at the expected number. i.e. the extent count is
correct on disk, but it's screwed up in memory. I loaded up the
extent list, and immediately:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 4192576215
#
It's bad again. So, where does that number come from?
xfs_fill_fsxattr():
if (ip->i_df.if_flags & XFS_IFEXTENTS)
fa->fsx_nextents = xfs_iext_count(&ip->i_df);
else
fa->fsx_nextents = ip->i_d.di_nextents;
And that's the behaviour I just saw in a nutshell. The on disk count
is correct, but once the tree is loaded into memory, it goes whacky.
Clearly there's something wrong with xfs_iext_count():
inline xfs_extnum_t xfs_iext_count(struct xfs_ifork *ifp)
{
return ifp->if_bytes / sizeof(struct xfs_iext_rec);
}
Simple enough, but 134M extents is 2**27, and that's right about
where things went wrong. A struct xfs_iext_rec is 16 bytes in size,
which means 2**27 * 2**4 = 2**31 and we're right on target for an
integer overflow. And, sure enough:
struct xfs_ifork {
int if_bytes; /* bytes in if_u1 */
....
Once we get 2**27 extents in a file, we overflow if_bytes and the
in-core extent count goes wrong. And when we reach 2**28 extents,
if_bytes wraps back to zero and things really start to go wrong
there. This is where the silent failure comes from - only the first
2**28 extents can be looked up directly due to the overflow, all the
extents above this index wrap back to somewhere in the first 2**28
extents. Hence with a regular pattern, trying to punch a hole in the
range that didn't have holes mapped to a hole in the first 2**28
extents and so "succeeded" without changing anything. Hence "silent
failure"...
Fix this by converting if_bytes to a int64_t and converting all the
index variables and size calculations to use int64_t types to avoid
overflows in future. Signed integers are still used to enable easy
detection of extent count underflows. This enables scalability of
extent counts to the limits of the on-disk format - MAXEXTNUM
(2**31) extents.
Current testing is at over 500M extents and still going:
fsxattr.nextents = 517310478
Reported-by: Zorro Lang <zlang@redhat.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com>
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
2019-10-17 20:40:33 +00:00
|
|
|
short if_broot_bytes; /* bytes allocated for root */
|
2020-05-18 17:28:05 +00:00
|
|
|
int8_t if_format; /* format of this fork */
|
2018-07-17 23:51:50 +00:00
|
|
|
};
|
2013-08-12 10:49:33 +00:00
|
|
|
|
2021-01-23 00:48:11 +00:00
|
|
|
/*
|
|
|
|
* Worst-case increase in the fork extent count when we're adding a single
|
|
|
|
* extent to a fork and there's no possibility of splitting an existing mapping.
|
|
|
|
*/
|
|
|
|
#define XFS_IEXT_ADD_NOSPLIT_CNT (1)
|
|
|
|
|
2021-01-23 00:48:11 +00:00
|
|
|
/*
|
|
|
|
* Punching out an extent from the middle of an existing extent can cause the
|
|
|
|
* extent count to increase by 1.
|
|
|
|
* i.e. | Old extent | Hole | Old extent |
|
|
|
|
*/
|
|
|
|
#define XFS_IEXT_PUNCH_HOLE_CNT (1)
|
|
|
|
|
2021-01-23 00:48:13 +00:00
|
|
|
/*
|
|
|
|
* Adding/removing an xattr can cause XFS_DA_NODE_MAXDEPTH extents to
|
|
|
|
* be added. One extra extent for dabtree in case a local attr is
|
|
|
|
* large enough to cause a double split. It can also cause extent
|
|
|
|
* count to increase proportional to the size of a remote xattr's
|
|
|
|
* value.
|
|
|
|
*/
|
|
|
|
#define XFS_IEXT_ATTR_MANIP_CNT(rmt_blks) \
|
|
|
|
(XFS_DA_NODE_MAXDEPTH + max(1, rmt_blks))
|
|
|
|
|
2021-01-23 00:48:14 +00:00
|
|
|
/*
|
|
|
|
* A write to a sub-interval of an existing unwritten extent causes the original
|
|
|
|
* extent to be split into 3 extents
|
|
|
|
* i.e. | Unwritten | Real | Unwritten |
|
|
|
|
* Hence extent count can increase by 2.
|
|
|
|
*/
|
|
|
|
#define XFS_IEXT_WRITE_UNWRITTEN_CNT (2)
|
|
|
|
|
|
|
|
|
2021-01-23 00:48:14 +00:00
|
|
|
/*
|
|
|
|
* Moving an extent to data fork can cause a sub-interval of an existing extent
|
|
|
|
* to be unmapped. This will increase extent count by 1. Mapping in the new
|
|
|
|
* extent can increase the extent count by 1 again i.e.
|
|
|
|
* | Old extent | New extent | Old extent |
|
|
|
|
* Hence number of extents increases by 2.
|
|
|
|
*/
|
|
|
|
#define XFS_IEXT_REFLINK_END_COW_CNT (2)
|
|
|
|
|
2021-01-23 00:48:15 +00:00
|
|
|
/*
|
|
|
|
* Removing an initial range of source/donor file's extent and adding a new
|
|
|
|
* extent (from donor/source file) in its place will cause extent count to
|
|
|
|
* increase by 1.
|
|
|
|
*/
|
|
|
|
#define XFS_IEXT_SWAP_RMAP_CNT (1)
|
|
|
|
|
2013-08-12 10:49:33 +00:00
|
|
|
/*
|
|
|
|
* Fork handling.
|
|
|
|
*/
|
|
|
|
#define XFS_IFORK_MAXEXT(ip, w) \
|
2022-07-09 17:56:07 +00:00
|
|
|
(xfs_inode_fork_size(ip, w) / sizeof(xfs_bmbt_rec_t))
|
2013-08-12 10:49:33 +00:00
|
|
|
|
2020-05-18 17:28:05 +00:00
|
|
|
static inline bool xfs_ifork_has_extents(struct xfs_ifork *ifp)
|
|
|
|
{
|
|
|
|
return ifp->if_format == XFS_DINODE_FMT_EXTENTS ||
|
|
|
|
ifp->if_format == XFS_DINODE_FMT_BTREE;
|
|
|
|
}
|
xfs: refactor "does this fork map blocks" predicate
Replace the open-coded checks for whether or not an inode fork maps
blocks with a macro that will implant the code for us. This helps us
declutter the bmap code a bit.
Note that I had to use a macro instead of a static inline function
because of C header dependency problems between xfs_inode.h and
xfs_inode_fork.h.
Conversion was performed with the following Coccinelle script:
@@
expression ip, w;
@@
- XFS_IFORK_FORMAT(ip, w) == XFS_DINODE_FMT_EXTENTS || XFS_IFORK_FORMAT(ip, w) == XFS_DINODE_FMT_BTREE
+ xfs_ifork_has_extents(ip, w)
@@
expression ip, w;
@@
- XFS_IFORK_FORMAT(ip, w) != XFS_DINODE_FMT_EXTENTS && XFS_IFORK_FORMAT(ip, w) != XFS_DINODE_FMT_BTREE
+ !xfs_ifork_has_extents(ip, w)
@@
expression ip, w;
@@
- XFS_IFORK_FORMAT(ip, w) == XFS_DINODE_FMT_BTREE || XFS_IFORK_FORMAT(ip, w) == XFS_DINODE_FMT_EXTENTS
+ xfs_ifork_has_extents(ip, w)
@@
expression ip, w;
@@
- XFS_IFORK_FORMAT(ip, w) != XFS_DINODE_FMT_BTREE && XFS_IFORK_FORMAT(ip, w) != XFS_DINODE_FMT_EXTENTS
+ !xfs_ifork_has_extents(ip, w)
@@
expression ip, w;
@@
- (xfs_ifork_has_extents(ip, w))
+ xfs_ifork_has_extents(ip, w)
@@
expression ip, w;
@@
- (!xfs_ifork_has_extents(ip, w))
+ !xfs_ifork_has_extents(ip, w)
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2019-11-07 23:05:21 +00:00
|
|
|
|
2020-05-18 17:27:22 +00:00
|
|
|
static inline xfs_extnum_t xfs_ifork_nextents(struct xfs_ifork *ifp)
|
|
|
|
{
|
|
|
|
if (!ifp)
|
|
|
|
return 0;
|
|
|
|
return ifp->if_nextents;
|
|
|
|
}
|
|
|
|
|
2020-05-18 17:28:05 +00:00
|
|
|
static inline int8_t xfs_ifork_format(struct xfs_ifork *ifp)
|
|
|
|
{
|
|
|
|
if (!ifp)
|
|
|
|
return XFS_DINODE_FMT_EXTENTS;
|
|
|
|
return ifp->if_format;
|
|
|
|
}
|
|
|
|
|
2021-11-16 09:54:37 +00:00
|
|
|
static inline xfs_extnum_t xfs_iext_max_nextents(bool has_large_extent_counts,
|
|
|
|
int whichfork)
|
2020-08-27 09:39:10 +00:00
|
|
|
{
|
2021-11-16 09:54:37 +00:00
|
|
|
switch (whichfork) {
|
|
|
|
case XFS_DATA_FORK:
|
|
|
|
case XFS_COW_FORK:
|
|
|
|
if (has_large_extent_counts)
|
|
|
|
return XFS_MAX_EXTCNT_DATA_FORK_LARGE;
|
|
|
|
return XFS_MAX_EXTCNT_DATA_FORK_SMALL;
|
|
|
|
|
|
|
|
case XFS_ATTR_FORK:
|
|
|
|
if (has_large_extent_counts)
|
|
|
|
return XFS_MAX_EXTCNT_ATTR_FORK_LARGE;
|
|
|
|
return XFS_MAX_EXTCNT_ATTR_FORK_SMALL;
|
|
|
|
|
|
|
|
default:
|
|
|
|
ASSERT(0);
|
|
|
|
return 0;
|
|
|
|
}
|
2020-08-27 09:39:10 +00:00
|
|
|
}
|
|
|
|
|
2020-08-27 10:04:34 +00:00
|
|
|
static inline xfs_extnum_t
|
|
|
|
xfs_dfork_data_extents(
|
|
|
|
struct xfs_dinode *dip)
|
|
|
|
{
|
2022-03-08 09:34:28 +00:00
|
|
|
if (xfs_dinode_has_large_extent_counts(dip))
|
|
|
|
return be64_to_cpu(dip->di_big_nextents);
|
|
|
|
|
2020-08-27 10:04:34 +00:00
|
|
|
return be32_to_cpu(dip->di_nextents);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline xfs_extnum_t
|
|
|
|
xfs_dfork_attr_extents(
|
|
|
|
struct xfs_dinode *dip)
|
|
|
|
{
|
2022-03-08 09:34:28 +00:00
|
|
|
if (xfs_dinode_has_large_extent_counts(dip))
|
|
|
|
return be32_to_cpu(dip->di_big_anextents);
|
|
|
|
|
2020-08-27 10:04:34 +00:00
|
|
|
return be16_to_cpu(dip->di_anextents);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline xfs_extnum_t
|
|
|
|
xfs_dfork_nextents(
|
|
|
|
struct xfs_dinode *dip,
|
|
|
|
int whichfork)
|
|
|
|
{
|
|
|
|
switch (whichfork) {
|
|
|
|
case XFS_DATA_FORK:
|
|
|
|
return xfs_dfork_data_extents(dip);
|
|
|
|
case XFS_ATTR_FORK:
|
|
|
|
return xfs_dfork_attr_extents(dip);
|
|
|
|
default:
|
|
|
|
ASSERT(0);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
xfs: make inode attribute forks a permanent part of struct xfs_inode
Syzkaller reported a UAF bug a while back:
==================================================================
BUG: KASAN: use-after-free in xfs_ilock_attr_map_shared+0xe3/0xf6 fs/xfs/xfs_inode.c:127
Read of size 4 at addr ffff88802cec919c by task syz-executor262/2958
CPU: 2 PID: 2958 Comm: syz-executor262 Not tainted
5.15.0-0.30.3-20220406_1406 #3
Hardware name: Red Hat KVM, BIOS 1.13.0-2.module+el8.3.0+7860+a7792d29
04/01/2014
Call Trace:
<TASK>
__dump_stack lib/dump_stack.c:88 [inline]
dump_stack_lvl+0x82/0xa9 lib/dump_stack.c:106
print_address_description.constprop.9+0x21/0x2d5 mm/kasan/report.c:256
__kasan_report mm/kasan/report.c:442 [inline]
kasan_report.cold.14+0x7f/0x11b mm/kasan/report.c:459
xfs_ilock_attr_map_shared+0xe3/0xf6 fs/xfs/xfs_inode.c:127
xfs_attr_get+0x378/0x4c2 fs/xfs/libxfs/xfs_attr.c:159
xfs_xattr_get+0xe3/0x150 fs/xfs/xfs_xattr.c:36
__vfs_getxattr+0xdf/0x13d fs/xattr.c:399
cap_inode_need_killpriv+0x41/0x5d security/commoncap.c:300
security_inode_need_killpriv+0x4c/0x97 security/security.c:1408
dentry_needs_remove_privs.part.28+0x21/0x63 fs/inode.c:1912
dentry_needs_remove_privs+0x80/0x9e fs/inode.c:1908
do_truncate+0xc3/0x1e0 fs/open.c:56
handle_truncate fs/namei.c:3084 [inline]
do_open fs/namei.c:3432 [inline]
path_openat+0x30ab/0x396d fs/namei.c:3561
do_filp_open+0x1c4/0x290 fs/namei.c:3588
do_sys_openat2+0x60d/0x98c fs/open.c:1212
do_sys_open+0xcf/0x13c fs/open.c:1228
do_syscall_x64 arch/x86/entry/common.c:50 [inline]
do_syscall_64+0x3a/0x7e arch/x86/entry/common.c:80
entry_SYSCALL_64_after_hwframe+0x44/0x0
RIP: 0033:0x7f7ef4bb753d
Code: 00 c3 66 2e 0f 1f 84 00 00 00 00 00 90 f3 0f 1e fa 48 89 f8 48 89 f7 48
89 d6 48 89 ca 4d 89 c2 4d 89 c8 4c 8b 4c 24 08 0f 05 <48> 3d 01 f0 ff ff 73
01 c3 48 8b 0d 1b 79 2c 00 f7 d8 64 89 01 48
RSP: 002b:00007f7ef52c2ed8 EFLAGS: 00000246 ORIG_RAX: 0000000000000055
RAX: ffffffffffffffda RBX: 0000000000404148 RCX: 00007f7ef4bb753d
RDX: 00007f7ef4bb753d RSI: 0000000000000000 RDI: 0000000020004fc0
RBP: 0000000000404140 R08: 0000000000000000 R09: 0000000000000000
R10: 0000000000000000 R11: 0000000000000246 R12: 0030656c69662f2e
R13: 00007ffd794db37f R14: 00007ffd794db470 R15: 00007f7ef52c2fc0
</TASK>
Allocated by task 2953:
kasan_save_stack+0x19/0x38 mm/kasan/common.c:38
kasan_set_track mm/kasan/common.c:46 [inline]
set_alloc_info mm/kasan/common.c:434 [inline]
__kasan_slab_alloc+0x68/0x7c mm/kasan/common.c:467
kasan_slab_alloc include/linux/kasan.h:254 [inline]
slab_post_alloc_hook mm/slab.h:519 [inline]
slab_alloc_node mm/slub.c:3213 [inline]
slab_alloc mm/slub.c:3221 [inline]
kmem_cache_alloc+0x11b/0x3eb mm/slub.c:3226
kmem_cache_zalloc include/linux/slab.h:711 [inline]
xfs_ifork_alloc+0x25/0xa2 fs/xfs/libxfs/xfs_inode_fork.c:287
xfs_bmap_add_attrfork+0x3f2/0x9b1 fs/xfs/libxfs/xfs_bmap.c:1098
xfs_attr_set+0xe38/0x12a7 fs/xfs/libxfs/xfs_attr.c:746
xfs_xattr_set+0xeb/0x1a9 fs/xfs/xfs_xattr.c:59
__vfs_setxattr+0x11b/0x177 fs/xattr.c:180
__vfs_setxattr_noperm+0x128/0x5e0 fs/xattr.c:214
__vfs_setxattr_locked+0x1d4/0x258 fs/xattr.c:275
vfs_setxattr+0x154/0x33d fs/xattr.c:301
setxattr+0x216/0x29f fs/xattr.c:575
__do_sys_fsetxattr fs/xattr.c:632 [inline]
__se_sys_fsetxattr fs/xattr.c:621 [inline]
__x64_sys_fsetxattr+0x243/0x2fe fs/xattr.c:621
do_syscall_x64 arch/x86/entry/common.c:50 [inline]
do_syscall_64+0x3a/0x7e arch/x86/entry/common.c:80
entry_SYSCALL_64_after_hwframe+0x44/0x0
Freed by task 2949:
kasan_save_stack+0x19/0x38 mm/kasan/common.c:38
kasan_set_track+0x1c/0x21 mm/kasan/common.c:46
kasan_set_free_info+0x20/0x30 mm/kasan/generic.c:360
____kasan_slab_free mm/kasan/common.c:366 [inline]
____kasan_slab_free mm/kasan/common.c:328 [inline]
__kasan_slab_free+0xe2/0x10e mm/kasan/common.c:374
kasan_slab_free include/linux/kasan.h:230 [inline]
slab_free_hook mm/slub.c:1700 [inline]
slab_free_freelist_hook mm/slub.c:1726 [inline]
slab_free mm/slub.c:3492 [inline]
kmem_cache_free+0xdc/0x3ce mm/slub.c:3508
xfs_attr_fork_remove+0x8d/0x132 fs/xfs/libxfs/xfs_attr_leaf.c:773
xfs_attr_sf_removename+0x5dd/0x6cb fs/xfs/libxfs/xfs_attr_leaf.c:822
xfs_attr_remove_iter+0x68c/0x805 fs/xfs/libxfs/xfs_attr.c:1413
xfs_attr_remove_args+0xb1/0x10d fs/xfs/libxfs/xfs_attr.c:684
xfs_attr_set+0xf1e/0x12a7 fs/xfs/libxfs/xfs_attr.c:802
xfs_xattr_set+0xeb/0x1a9 fs/xfs/xfs_xattr.c:59
__vfs_removexattr+0x106/0x16a fs/xattr.c:468
cap_inode_killpriv+0x24/0x47 security/commoncap.c:324
security_inode_killpriv+0x54/0xa1 security/security.c:1414
setattr_prepare+0x1a6/0x897 fs/attr.c:146
xfs_vn_change_ok+0x111/0x15e fs/xfs/xfs_iops.c:682
xfs_vn_setattr_size+0x5f/0x15a fs/xfs/xfs_iops.c:1065
xfs_vn_setattr+0x125/0x2ad fs/xfs/xfs_iops.c:1093
notify_change+0xae5/0x10a1 fs/attr.c:410
do_truncate+0x134/0x1e0 fs/open.c:64
handle_truncate fs/namei.c:3084 [inline]
do_open fs/namei.c:3432 [inline]
path_openat+0x30ab/0x396d fs/namei.c:3561
do_filp_open+0x1c4/0x290 fs/namei.c:3588
do_sys_openat2+0x60d/0x98c fs/open.c:1212
do_sys_open+0xcf/0x13c fs/open.c:1228
do_syscall_x64 arch/x86/entry/common.c:50 [inline]
do_syscall_64+0x3a/0x7e arch/x86/entry/common.c:80
entry_SYSCALL_64_after_hwframe+0x44/0x0
The buggy address belongs to the object at ffff88802cec9188
which belongs to the cache xfs_ifork of size 40
The buggy address is located 20 bytes inside of
40-byte region [ffff88802cec9188, ffff88802cec91b0)
The buggy address belongs to the page:
page:00000000c3af36a1 refcount:1 mapcount:0 mapping:0000000000000000
index:0x0 pfn:0x2cec9
flags: 0xfffffc0000200(slab|node=0|zone=1|lastcpupid=0x1fffff)
raw: 000fffffc0000200 ffffea00009d2580 0000000600000006 ffff88801a9ffc80
raw: 0000000000000000 0000000080490049 00000001ffffffff 0000000000000000
page dumped because: kasan: bad access detected
Memory state around the buggy address:
ffff88802cec9080: fb fb fb fc fc fa fb fb fb fb fc fc fb fb fb fb
ffff88802cec9100: fb fc fc fb fb fb fb fb fc fc fb fb fb fb fb fc
>ffff88802cec9180: fc fa fb fb fb fb fc fc fa fb fb fb fb fc fc fb
^
ffff88802cec9200: fb fb fb fb fc fc fb fb fb fb fb fc fc fb fb fb
ffff88802cec9280: fb fb fc fc fa fb fb fb fb fc fc fa fb fb fb fb
==================================================================
The root cause of this bug is the unlocked access to xfs_inode.i_afp
from the getxattr code paths while trying to determine which ILOCK mode
to use to stabilize the xattr data. Unfortunately, the VFS does not
acquire i_rwsem when vfs_getxattr (or listxattr) call into the
filesystem, which means that getxattr can race with a removexattr that's
tearing down the attr fork and crash:
xfs_attr_set: xfs_attr_get:
xfs_attr_fork_remove: xfs_ilock_attr_map_shared:
xfs_idestroy_fork(ip->i_afp);
kmem_cache_free(xfs_ifork_cache, ip->i_afp);
if (ip->i_afp &&
ip->i_afp = NULL;
xfs_need_iread_extents(ip->i_afp))
<KABOOM>
ip->i_forkoff = 0;
Regrettably, the VFS is much more lax about i_rwsem and getxattr than
is immediately obvious -- not only does it not guarantee that we hold
i_rwsem, it actually doesn't guarantee that we *don't* hold it either.
The getxattr system call won't acquire the lock before calling XFS, but
the file capabilities code calls getxattr with and without i_rwsem held
to determine if the "security.capabilities" xattr is set on the file.
Fixing the VFS locking requires a treewide investigation into every code
path that could touch an xattr and what i_rwsem state it expects or sets
up. That could take years or even prove impossible; fortunately, we
can fix this UAF problem inside XFS.
An earlier version of this patch used smp_wmb in xfs_attr_fork_remove to
ensure that i_forkoff is always zeroed before i_afp is set to null and
changed the read paths to use smp_rmb before accessing i_forkoff and
i_afp, which avoided these UAF problems. However, the patch author was
too busy dealing with other problems in the meantime, and by the time he
came back to this issue, the situation had changed a bit.
On a modern system with selinux, each inode will always have at least
one xattr for the selinux label, so it doesn't make much sense to keep
incurring the extra pointer dereference. Furthermore, Allison's
upcoming parent pointer patchset will also cause nearly every inode in
the filesystem to have extended attributes. Therefore, make the inode
attribute fork structure part of struct xfs_inode, at a cost of 40 more
bytes.
This patch adds a clunky if_present field where necessary to maintain
the existing logic of xattr fork null pointer testing in the existing
codebase. The next patch switches the logic over to XFS_IFORK_Q and it
all goes away.
Signed-off-by: Darrick J. Wong <djwong@kernel.org>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
2022-07-09 17:56:06 +00:00
|
|
|
void xfs_ifork_zap_attr(struct xfs_inode *ip);
|
|
|
|
void xfs_ifork_init_attr(struct xfs_inode *ip, enum xfs_dinode_fmt format,
|
|
|
|
xfs_extnum_t nextents);
|
2016-10-03 16:11:32 +00:00
|
|
|
struct xfs_ifork *xfs_iext_state_to_fork(struct xfs_inode *ip, int state);
|
|
|
|
|
2020-05-14 21:01:17 +00:00
|
|
|
int xfs_iformat_data_fork(struct xfs_inode *, struct xfs_dinode *);
|
|
|
|
int xfs_iformat_attr_fork(struct xfs_inode *, struct xfs_dinode *);
|
2017-04-03 19:22:20 +00:00
|
|
|
void xfs_iflush_fork(struct xfs_inode *, struct xfs_dinode *,
|
2014-04-14 09:04:46 +00:00
|
|
|
struct xfs_inode_log_item *, int);
|
2020-05-18 17:29:27 +00:00
|
|
|
void xfs_idestroy_fork(struct xfs_ifork *ifp);
|
xfs: fix inode fork extent count overflow
[commit message is verbose for discussion purposes - will trim it
down later. Some questions about implementation details at the end.]
Zorro Lang recently ran a new test to stress single inode extent
counts now that they are no longer limited by memory allocation.
The test was simply:
# xfs_io -f -c "falloc 0 40t" /mnt/scratch/big-file
# ~/src/xfstests-dev/punch-alternating /mnt/scratch/big-file
This test uncovered a problem where the hole punching operation
appeared to finish with no error, but apparently only created 268M
extents instead of the 10 billion it was supposed to.
Further, trying to punch out extents that should have been present
resulted in success, but no change in the extent count. It looked
like a silent failure.
While running the test and observing the behaviour in real time,
I observed the extent coutn growing at ~2M extents/minute, and saw
this after about an hour:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next ; \
> sleep 60 ; \
> xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 127657993
fsxattr.nextents = 129683339
#
And a few minutes later this:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 4177861124
#
Ah, what? Where did that 4 billion extra extents suddenly come from?
Stop the workload, unmount, mount:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 166044375
#
And it's back at the expected number. i.e. the extent count is
correct on disk, but it's screwed up in memory. I loaded up the
extent list, and immediately:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 4192576215
#
It's bad again. So, where does that number come from?
xfs_fill_fsxattr():
if (ip->i_df.if_flags & XFS_IFEXTENTS)
fa->fsx_nextents = xfs_iext_count(&ip->i_df);
else
fa->fsx_nextents = ip->i_d.di_nextents;
And that's the behaviour I just saw in a nutshell. The on disk count
is correct, but once the tree is loaded into memory, it goes whacky.
Clearly there's something wrong with xfs_iext_count():
inline xfs_extnum_t xfs_iext_count(struct xfs_ifork *ifp)
{
return ifp->if_bytes / sizeof(struct xfs_iext_rec);
}
Simple enough, but 134M extents is 2**27, and that's right about
where things went wrong. A struct xfs_iext_rec is 16 bytes in size,
which means 2**27 * 2**4 = 2**31 and we're right on target for an
integer overflow. And, sure enough:
struct xfs_ifork {
int if_bytes; /* bytes in if_u1 */
....
Once we get 2**27 extents in a file, we overflow if_bytes and the
in-core extent count goes wrong. And when we reach 2**28 extents,
if_bytes wraps back to zero and things really start to go wrong
there. This is where the silent failure comes from - only the first
2**28 extents can be looked up directly due to the overflow, all the
extents above this index wrap back to somewhere in the first 2**28
extents. Hence with a regular pattern, trying to punch a hole in the
range that didn't have holes mapped to a hole in the first 2**28
extents and so "succeeded" without changing anything. Hence "silent
failure"...
Fix this by converting if_bytes to a int64_t and converting all the
index variables and size calculations to use int64_t types to avoid
overflows in future. Signed integers are still used to enable easy
detection of extent count underflows. This enables scalability of
extent counts to the limits of the on-disk format - MAXEXTNUM
(2**31) extents.
Current testing is at over 500M extents and still going:
fsxattr.nextents = 517310478
Reported-by: Zorro Lang <zlang@redhat.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com>
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
2019-10-17 20:40:33 +00:00
|
|
|
void xfs_idata_realloc(struct xfs_inode *ip, int64_t byte_diff,
|
|
|
|
int whichfork);
|
2013-08-12 10:49:33 +00:00
|
|
|
void xfs_iroot_realloc(struct xfs_inode *, int, int);
|
|
|
|
int xfs_iread_extents(struct xfs_trans *, struct xfs_inode *, int);
|
|
|
|
int xfs_iextents_copy(struct xfs_inode *, struct xfs_bmbt_rec *,
|
|
|
|
int);
|
xfs: fix inode fork extent count overflow
[commit message is verbose for discussion purposes - will trim it
down later. Some questions about implementation details at the end.]
Zorro Lang recently ran a new test to stress single inode extent
counts now that they are no longer limited by memory allocation.
The test was simply:
# xfs_io -f -c "falloc 0 40t" /mnt/scratch/big-file
# ~/src/xfstests-dev/punch-alternating /mnt/scratch/big-file
This test uncovered a problem where the hole punching operation
appeared to finish with no error, but apparently only created 268M
extents instead of the 10 billion it was supposed to.
Further, trying to punch out extents that should have been present
resulted in success, but no change in the extent count. It looked
like a silent failure.
While running the test and observing the behaviour in real time,
I observed the extent coutn growing at ~2M extents/minute, and saw
this after about an hour:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next ; \
> sleep 60 ; \
> xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 127657993
fsxattr.nextents = 129683339
#
And a few minutes later this:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 4177861124
#
Ah, what? Where did that 4 billion extra extents suddenly come from?
Stop the workload, unmount, mount:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 166044375
#
And it's back at the expected number. i.e. the extent count is
correct on disk, but it's screwed up in memory. I loaded up the
extent list, and immediately:
# xfs_io -f -c "stat" /mnt/scratch/big-file |grep next
fsxattr.nextents = 4192576215
#
It's bad again. So, where does that number come from?
xfs_fill_fsxattr():
if (ip->i_df.if_flags & XFS_IFEXTENTS)
fa->fsx_nextents = xfs_iext_count(&ip->i_df);
else
fa->fsx_nextents = ip->i_d.di_nextents;
And that's the behaviour I just saw in a nutshell. The on disk count
is correct, but once the tree is loaded into memory, it goes whacky.
Clearly there's something wrong with xfs_iext_count():
inline xfs_extnum_t xfs_iext_count(struct xfs_ifork *ifp)
{
return ifp->if_bytes / sizeof(struct xfs_iext_rec);
}
Simple enough, but 134M extents is 2**27, and that's right about
where things went wrong. A struct xfs_iext_rec is 16 bytes in size,
which means 2**27 * 2**4 = 2**31 and we're right on target for an
integer overflow. And, sure enough:
struct xfs_ifork {
int if_bytes; /* bytes in if_u1 */
....
Once we get 2**27 extents in a file, we overflow if_bytes and the
in-core extent count goes wrong. And when we reach 2**28 extents,
if_bytes wraps back to zero and things really start to go wrong
there. This is where the silent failure comes from - only the first
2**28 extents can be looked up directly due to the overflow, all the
extents above this index wrap back to somewhere in the first 2**28
extents. Hence with a regular pattern, trying to punch a hole in the
range that didn't have holes mapped to a hole in the first 2**28
extents and so "succeeded" without changing anything. Hence "silent
failure"...
Fix this by converting if_bytes to a int64_t and converting all the
index variables and size calculations to use int64_t types to avoid
overflows in future. Signed integers are still used to enable easy
detection of extent count underflows. This enables scalability of
extent counts to the limits of the on-disk format - MAXEXTNUM
(2**31) extents.
Current testing is at over 500M extents and still going:
fsxattr.nextents = 517310478
Reported-by: Zorro Lang <zlang@redhat.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com>
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
2019-10-17 20:40:33 +00:00
|
|
|
void xfs_init_local_fork(struct xfs_inode *ip, int whichfork,
|
|
|
|
const void *data, int64_t size);
|
2013-08-12 10:49:33 +00:00
|
|
|
|
2017-11-03 17:34:46 +00:00
|
|
|
xfs_extnum_t xfs_iext_count(struct xfs_ifork *ifp);
|
2017-11-03 17:34:43 +00:00
|
|
|
void xfs_iext_insert(struct xfs_inode *, struct xfs_iext_cursor *cur,
|
2017-11-03 17:34:46 +00:00
|
|
|
struct xfs_bmbt_irec *, int);
|
2017-11-03 17:34:43 +00:00
|
|
|
void xfs_iext_remove(struct xfs_inode *, struct xfs_iext_cursor *,
|
2017-11-03 17:34:47 +00:00
|
|
|
int);
|
2013-08-12 10:49:33 +00:00
|
|
|
void xfs_iext_destroy(struct xfs_ifork *);
|
|
|
|
|
2016-11-24 00:39:32 +00:00
|
|
|
bool xfs_iext_lookup_extent(struct xfs_inode *ip,
|
|
|
|
struct xfs_ifork *ifp, xfs_fileoff_t bno,
|
2017-11-03 17:34:43 +00:00
|
|
|
struct xfs_iext_cursor *cur,
|
|
|
|
struct xfs_bmbt_irec *gotp);
|
2017-10-23 23:32:39 +00:00
|
|
|
bool xfs_iext_lookup_extent_before(struct xfs_inode *ip,
|
|
|
|
struct xfs_ifork *ifp, xfs_fileoff_t *end,
|
2017-11-03 17:34:43 +00:00
|
|
|
struct xfs_iext_cursor *cur,
|
|
|
|
struct xfs_bmbt_irec *gotp);
|
|
|
|
bool xfs_iext_get_extent(struct xfs_ifork *ifp,
|
|
|
|
struct xfs_iext_cursor *cur,
|
2016-11-24 00:39:32 +00:00
|
|
|
struct xfs_bmbt_irec *gotp);
|
2017-10-19 18:04:44 +00:00
|
|
|
void xfs_iext_update_extent(struct xfs_inode *ip, int state,
|
2017-11-03 17:34:43 +00:00
|
|
|
struct xfs_iext_cursor *cur,
|
|
|
|
struct xfs_bmbt_irec *gotp);
|
|
|
|
|
2017-11-03 17:34:46 +00:00
|
|
|
void xfs_iext_first(struct xfs_ifork *, struct xfs_iext_cursor *);
|
|
|
|
void xfs_iext_last(struct xfs_ifork *, struct xfs_iext_cursor *);
|
|
|
|
void xfs_iext_next(struct xfs_ifork *, struct xfs_iext_cursor *);
|
|
|
|
void xfs_iext_prev(struct xfs_ifork *, struct xfs_iext_cursor *);
|
2017-11-03 17:34:43 +00:00
|
|
|
|
|
|
|
static inline bool xfs_iext_next_extent(struct xfs_ifork *ifp,
|
|
|
|
struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp)
|
|
|
|
{
|
|
|
|
xfs_iext_next(ifp, cur);
|
|
|
|
return xfs_iext_get_extent(ifp, cur, gotp);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline bool xfs_iext_prev_extent(struct xfs_ifork *ifp,
|
|
|
|
struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp)
|
|
|
|
{
|
|
|
|
xfs_iext_prev(ifp, cur);
|
|
|
|
return xfs_iext_get_extent(ifp, cur, gotp);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Return the extent after cur in gotp without updating the cursor.
|
|
|
|
*/
|
|
|
|
static inline bool xfs_iext_peek_next_extent(struct xfs_ifork *ifp,
|
|
|
|
struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp)
|
|
|
|
{
|
|
|
|
struct xfs_iext_cursor ncur = *cur;
|
|
|
|
|
|
|
|
xfs_iext_next(ifp, &ncur);
|
|
|
|
return xfs_iext_get_extent(ifp, &ncur, gotp);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Return the extent before cur in gotp without updating the cursor.
|
|
|
|
*/
|
|
|
|
static inline bool xfs_iext_peek_prev_extent(struct xfs_ifork *ifp,
|
|
|
|
struct xfs_iext_cursor *cur, struct xfs_bmbt_irec *gotp)
|
|
|
|
{
|
|
|
|
struct xfs_iext_cursor ncur = *cur;
|
|
|
|
|
|
|
|
xfs_iext_prev(ifp, &ncur);
|
|
|
|
return xfs_iext_get_extent(ifp, &ncur, gotp);
|
|
|
|
}
|
|
|
|
|
|
|
|
#define for_each_xfs_iext(ifp, ext, got) \
|
|
|
|
for (xfs_iext_first((ifp), (ext)); \
|
|
|
|
xfs_iext_get_extent((ifp), (ext), (got)); \
|
|
|
|
xfs_iext_next((ifp), (ext)))
|
2016-11-24 00:39:32 +00:00
|
|
|
|
2021-10-12 18:09:23 +00:00
|
|
|
extern struct kmem_cache *xfs_ifork_cache;
|
2013-08-12 10:49:33 +00:00
|
|
|
|
2016-10-03 16:11:32 +00:00
|
|
|
extern void xfs_ifork_init_cow(struct xfs_inode *ip);
|
|
|
|
|
2020-05-14 21:01:19 +00:00
|
|
|
int xfs_ifork_verify_local_data(struct xfs_inode *ip);
|
|
|
|
int xfs_ifork_verify_local_attr(struct xfs_inode *ip);
|
2021-01-23 00:48:10 +00:00
|
|
|
int xfs_iext_count_may_overflow(struct xfs_inode *ip, int whichfork,
|
|
|
|
int nr_to_add);
|
2022-03-09 07:49:36 +00:00
|
|
|
int xfs_iext_count_upgrade(struct xfs_trans *tp, struct xfs_inode *ip,
|
|
|
|
uint nr_to_add);
|
2018-01-08 18:51:06 +00:00
|
|
|
|
2021-04-13 18:15:12 +00:00
|
|
|
/* returns true if the fork has extents but they are not read in yet. */
|
|
|
|
static inline bool xfs_need_iread_extents(struct xfs_ifork *ifp)
|
|
|
|
{
|
|
|
|
return ifp->if_format == XFS_DINODE_FMT_BTREE && ifp->if_height == 0;
|
|
|
|
}
|
|
|
|
|
2013-08-12 10:49:33 +00:00
|
|
|
#endif /* __XFS_INODE_FORK_H__ */
|