2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2006-06-09 04:48:12 +00:00
|
|
|
* Copyright (c) 2000-2006 Silicon Graphics, Inc.
|
2005-11-02 03:58:39 +00:00
|
|
|
* All Rights Reserved.
|
2005-04-16 22:20:36 +00:00
|
|
|
*
|
2005-11-02 03:58:39 +00:00
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public License as
|
2005-04-16 22:20:36 +00:00
|
|
|
* published by the Free Software Foundation.
|
|
|
|
*
|
2005-11-02 03:58:39 +00:00
|
|
|
* This program is distributed in the hope that it would be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
* GNU General Public License for more details.
|
2005-04-16 22:20:36 +00:00
|
|
|
*
|
2005-11-02 03:58:39 +00:00
|
|
|
* You should have received a copy of the GNU General Public License
|
|
|
|
* along with this program; if not, write the Free Software Foundation,
|
|
|
|
* Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
#include "xfs.h"
|
2005-11-02 03:38:42 +00:00
|
|
|
#include "xfs_fs.h"
|
2013-10-22 23:36:05 +00:00
|
|
|
#include "xfs_shared.h"
|
2013-10-22 23:50:10 +00:00
|
|
|
#include "xfs_format.h"
|
|
|
|
#include "xfs_log_format.h"
|
|
|
|
#include "xfs_trans_resv.h"
|
2005-11-02 03:38:42 +00:00
|
|
|
#include "xfs_bit.h"
|
2005-04-16 22:20:36 +00:00
|
|
|
#include "xfs_sb.h"
|
2013-04-24 08:58:02 +00:00
|
|
|
#include "xfs_mount.h"
|
2016-08-03 01:15:38 +00:00
|
|
|
#include "xfs_defer.h"
|
2013-10-14 22:17:51 +00:00
|
|
|
#include "xfs_da_format.h"
|
2005-11-02 03:38:42 +00:00
|
|
|
#include "xfs_da_btree.h"
|
2013-08-12 10:49:37 +00:00
|
|
|
#include "xfs_dir2.h"
|
2005-04-16 22:20:36 +00:00
|
|
|
#include "xfs_inode.h"
|
2005-11-02 03:38:42 +00:00
|
|
|
#include "xfs_btree.h"
|
2013-10-22 23:50:10 +00:00
|
|
|
#include "xfs_trans.h"
|
2005-11-02 03:38:42 +00:00
|
|
|
#include "xfs_inode_item.h"
|
2005-04-16 22:20:36 +00:00
|
|
|
#include "xfs_extfree_item.h"
|
|
|
|
#include "xfs_alloc.h"
|
|
|
|
#include "xfs_bmap.h"
|
2013-08-12 10:49:42 +00:00
|
|
|
#include "xfs_bmap_util.h"
|
2013-10-22 23:51:50 +00:00
|
|
|
#include "xfs_bmap_btree.h"
|
2005-04-16 22:20:36 +00:00
|
|
|
#include "xfs_rtalloc.h"
|
|
|
|
#include "xfs_error.h"
|
|
|
|
#include "xfs_quota.h"
|
|
|
|
#include "xfs_trans_space.h"
|
|
|
|
#include "xfs_buf_item.h"
|
2009-12-14 23:14:59 +00:00
|
|
|
#include "xfs_trace.h"
|
2013-04-03 05:11:18 +00:00
|
|
|
#include "xfs_symlink.h"
|
2013-10-22 23:51:50 +00:00
|
|
|
#include "xfs_attr_leaf.h"
|
|
|
|
#include "xfs_filestream.h"
|
2016-08-03 01:33:42 +00:00
|
|
|
#include "xfs_rmap.h"
|
xfs: set up per-AG free space reservations
One unfortunate quirk of the reference count and reverse mapping
btrees -- they can expand in size when blocks are written to *other*
allocation groups if, say, one large extent becomes a lot of tiny
extents. Since we don't want to start throwing errors in the middle
of CoWing, we need to reserve some blocks to handle future expansion.
The transaction block reservation counters aren't sufficient here
because we have to have a reserve of blocks in every AG, not just
somewhere in the filesystem.
Therefore, create two per-AG block reservation pools. One feeds the
AGFL so that rmapbt expansion always succeeds, and the other feeds all
other metadata so that refcountbt expansion never fails.
Use the count of how many reserved blocks we need to have on hand to
create a virtual reservation in the AG. Through selective clamping of
the maximum length of allocation requests and of the length of the
longest free extent, we can make it look like there's less free space
in the AG unless the reservation owner is asking for blocks.
In other words, play some accounting tricks in-core to make sure that
we always have blocks available. On the plus side, there's nothing to
clean up if we crash, which is contrast to the strategy that the rough
draft used (actually removing extents from the freespace btrees).
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-19 00:30:52 +00:00
|
|
|
#include "xfs_ag_resv.h"
|
2016-10-03 16:11:23 +00:00
|
|
|
#include "xfs_refcount.h"
|
2016-11-28 03:57:42 +00:00
|
|
|
#include "xfs_rmap_btree.h"
|
2016-11-28 03:57:42 +00:00
|
|
|
#include "xfs_icache.h"
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
|
|
|
|
kmem_zone_t *xfs_bmap_free_item_zone;
|
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Miscellaneous helper functions
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Compute and fill in the value of the maximum depth of a bmap btree
|
|
|
|
* in this filesystem. Done once, during mount.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
void
|
|
|
|
xfs_bmap_compute_maxlevels(
|
|
|
|
xfs_mount_t *mp, /* file system mount structure */
|
|
|
|
int whichfork) /* data or attr fork */
|
|
|
|
{
|
|
|
|
int level; /* btree level */
|
|
|
|
uint maxblocks; /* max blocks at this level */
|
|
|
|
uint maxleafents; /* max leaf entries possible */
|
|
|
|
int maxrootrecs; /* max records in root block */
|
|
|
|
int minleafrecs; /* min records in leaf block */
|
|
|
|
int minnoderecs; /* min records in node block */
|
|
|
|
int sz; /* root block size */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* The maximum number of extents in a file, hence the maximum
|
|
|
|
* number of leaf entries, is controlled by the type of di_nextents
|
|
|
|
* (a signed 32-bit number, xfs_extnum_t), or by di_anextents
|
|
|
|
* (a signed 16-bit number, xfs_aextnum_t).
|
|
|
|
*
|
|
|
|
* Note that we can no longer assume that if we are in ATTR1 that
|
|
|
|
* the fork offset of all the inodes will be
|
|
|
|
* (xfs_default_attroffset(ip) >> 3) because we could have mounted
|
|
|
|
* with ATTR2 and then mounted back with ATTR1, keeping the
|
|
|
|
* di_forkoff's fixed but probably at various positions. Therefore,
|
|
|
|
* for both ATTR1 and ATTR2 we have to assume the worst case scenario
|
|
|
|
* of a minimum size available.
|
|
|
|
*/
|
|
|
|
if (whichfork == XFS_DATA_FORK) {
|
|
|
|
maxleafents = MAXEXTNUM;
|
|
|
|
sz = XFS_BMDR_SPACE_CALC(MINDBTPTRS);
|
|
|
|
} else {
|
|
|
|
maxleafents = MAXAEXTNUM;
|
|
|
|
sz = XFS_BMDR_SPACE_CALC(MINABTPTRS);
|
|
|
|
}
|
2014-04-14 08:58:51 +00:00
|
|
|
maxrootrecs = xfs_bmdr_maxrecs(sz, 0);
|
2013-02-25 01:31:26 +00:00
|
|
|
minleafrecs = mp->m_bmap_dmnr[0];
|
|
|
|
minnoderecs = mp->m_bmap_dmnr[1];
|
|
|
|
maxblocks = (maxleafents + minleafrecs - 1) / minleafrecs;
|
|
|
|
for (level = 1; maxblocks > 1; level++) {
|
|
|
|
if (maxblocks <= maxrootrecs)
|
|
|
|
maxblocks = 1;
|
|
|
|
else
|
|
|
|
maxblocks = (maxblocks + minnoderecs - 1) / minnoderecs;
|
|
|
|
}
|
|
|
|
mp->m_bm_maxlevels[whichfork] = level;
|
|
|
|
}
|
2005-11-02 04:10:24 +00:00
|
|
|
|
2008-10-30 05:56:09 +00:00
|
|
|
STATIC int /* error */
|
|
|
|
xfs_bmbt_lookup_eq(
|
|
|
|
struct xfs_btree_cur *cur,
|
|
|
|
xfs_fileoff_t off,
|
|
|
|
xfs_fsblock_t bno,
|
|
|
|
xfs_filblks_t len,
|
|
|
|
int *stat) /* success/failure */
|
|
|
|
{
|
|
|
|
cur->bc_rec.b.br_startoff = off;
|
|
|
|
cur->bc_rec.b.br_startblock = bno;
|
|
|
|
cur->bc_rec.b.br_blockcount = len;
|
|
|
|
return xfs_btree_lookup(cur, XFS_LOOKUP_EQ, stat);
|
|
|
|
}
|
|
|
|
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_bmbt_lookup_ge(
|
|
|
|
struct xfs_btree_cur *cur,
|
|
|
|
xfs_fileoff_t off,
|
|
|
|
xfs_fsblock_t bno,
|
|
|
|
xfs_filblks_t len,
|
|
|
|
int *stat) /* success/failure */
|
|
|
|
{
|
|
|
|
cur->bc_rec.b.br_startoff = off;
|
|
|
|
cur->bc_rec.b.br_startblock = bno;
|
|
|
|
cur->bc_rec.b.br_blockcount = len;
|
|
|
|
return xfs_btree_lookup(cur, XFS_LOOKUP_GE, stat);
|
|
|
|
}
|
|
|
|
|
2008-10-30 05:56:32 +00:00
|
|
|
/*
|
2011-12-18 20:00:07 +00:00
|
|
|
* Check if the inode needs to be converted to btree format.
|
|
|
|
*/
|
|
|
|
static inline bool xfs_bmap_needs_btree(struct xfs_inode *ip, int whichfork)
|
|
|
|
{
|
2016-10-03 16:11:34 +00:00
|
|
|
return whichfork != XFS_COW_FORK &&
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_EXTENTS &&
|
2011-12-18 20:00:07 +00:00
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) >
|
|
|
|
XFS_IFORK_MAXEXT(ip, whichfork);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check if the inode should be converted to extent format.
|
|
|
|
*/
|
|
|
|
static inline bool xfs_bmap_wants_extents(struct xfs_inode *ip, int whichfork)
|
|
|
|
{
|
2016-10-03 16:11:34 +00:00
|
|
|
return whichfork != XFS_COW_FORK &&
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_BTREE &&
|
2011-12-18 20:00:07 +00:00
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) <=
|
|
|
|
XFS_IFORK_MAXEXT(ip, whichfork);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Update the record referred to by cur to the value given
|
2008-10-30 05:56:32 +00:00
|
|
|
* by [off, bno, len, state].
|
|
|
|
* This either works (return 0) or gets an EFSCORRUPTED error.
|
|
|
|
*/
|
|
|
|
STATIC int
|
|
|
|
xfs_bmbt_update(
|
|
|
|
struct xfs_btree_cur *cur,
|
|
|
|
xfs_fileoff_t off,
|
|
|
|
xfs_fsblock_t bno,
|
|
|
|
xfs_filblks_t len,
|
|
|
|
xfs_exntst_t state)
|
|
|
|
{
|
|
|
|
union xfs_btree_rec rec;
|
|
|
|
|
|
|
|
xfs_bmbt_disk_set_allf(&rec.bmbt, off, bno, len, state);
|
|
|
|
return xfs_btree_update(cur, &rec);
|
|
|
|
}
|
2008-10-30 05:56:09 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Compute the worst-case number of indirect blocks that will be used
|
|
|
|
* for ip's delayed extent of length "len".
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
STATIC xfs_filblks_t
|
|
|
|
xfs_bmap_worst_indlen(
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
|
|
|
xfs_filblks_t len) /* delayed extent length */
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
int level; /* btree level number */
|
|
|
|
int maxrecs; /* maximum record count at this level */
|
|
|
|
xfs_mount_t *mp; /* mount structure */
|
|
|
|
xfs_filblks_t rval; /* return value */
|
2016-11-28 03:57:42 +00:00
|
|
|
xfs_filblks_t orig_len;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
mp = ip->i_mount;
|
2016-11-28 03:57:42 +00:00
|
|
|
|
|
|
|
/* Calculate the worst-case size of the bmbt. */
|
|
|
|
orig_len = len;
|
2013-02-25 01:31:26 +00:00
|
|
|
maxrecs = mp->m_bmap_dmxr[0];
|
|
|
|
for (level = 0, rval = 0;
|
|
|
|
level < XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK);
|
|
|
|
level++) {
|
|
|
|
len += maxrecs - 1;
|
|
|
|
do_div(len, maxrecs);
|
|
|
|
rval += len;
|
2016-11-28 03:57:42 +00:00
|
|
|
if (len == 1) {
|
|
|
|
rval += XFS_BM_MAXLEVELS(mp, XFS_DATA_FORK) -
|
2013-02-25 01:31:26 +00:00
|
|
|
level - 1;
|
2016-11-28 03:57:42 +00:00
|
|
|
break;
|
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
if (level == 0)
|
|
|
|
maxrecs = mp->m_bmap_dmxr[1];
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2016-11-28 03:57:42 +00:00
|
|
|
|
|
|
|
/* Calculate the worst-case size of the rmapbt. */
|
|
|
|
if (xfs_sb_version_hasrmapbt(&mp->m_sb))
|
|
|
|
rval += 1 + xfs_rmapbt_calc_size(mp, orig_len) +
|
|
|
|
mp->m_rmap_maxlevels;
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
return rval;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Calculate the default attribute fork offset for newly created inodes.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
uint
|
|
|
|
xfs_default_attroffset(
|
|
|
|
struct xfs_inode *ip)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
|
|
|
uint offset;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (mp->m_sb.sb_inodesize == 256) {
|
2013-03-12 12:30:36 +00:00
|
|
|
offset = XFS_LITINO(mp, ip->i_d.di_version) -
|
2013-02-25 01:31:26 +00:00
|
|
|
XFS_BMDR_SPACE_CALC(MINABTPTRS);
|
|
|
|
} else {
|
|
|
|
offset = XFS_BMDR_SPACE_CALC(6 * MINABTPTRS);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
|
2013-03-12 12:30:36 +00:00
|
|
|
ASSERT(offset < XFS_LITINO(mp, ip->i_d.di_version));
|
2013-02-25 01:31:26 +00:00
|
|
|
return offset;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Helper routine to reset inode di_forkoff field when switching
|
|
|
|
* attribute fork from local to extent format - we reset it where
|
|
|
|
* possible to make space available for inline data fork extents.
|
2013-02-11 04:58:13 +00:00
|
|
|
*/
|
|
|
|
STATIC void
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmap_forkoff_reset(
|
|
|
|
xfs_inode_t *ip,
|
|
|
|
int whichfork)
|
2013-02-11 04:58:13 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
if (whichfork == XFS_ATTR_FORK &&
|
|
|
|
ip->i_d.di_format != XFS_DINODE_FMT_DEV &&
|
|
|
|
ip->i_d.di_format != XFS_DINODE_FMT_UUID &&
|
|
|
|
ip->i_d.di_format != XFS_DINODE_FMT_BTREE) {
|
|
|
|
uint dfl_forkoff = xfs_default_attroffset(ip) >> 3;
|
|
|
|
|
|
|
|
if (dfl_forkoff > ip->i_d.di_forkoff)
|
|
|
|
ip->i_d.di_forkoff = dfl_forkoff;
|
|
|
|
}
|
2013-02-11 04:58:13 +00:00
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
#ifdef DEBUG
|
|
|
|
STATIC struct xfs_buf *
|
|
|
|
xfs_bmap_get_bp(
|
|
|
|
struct xfs_btree_cur *cur,
|
|
|
|
xfs_fsblock_t bno)
|
|
|
|
{
|
|
|
|
struct xfs_log_item_desc *lidp;
|
|
|
|
int i;
|
2009-11-25 00:00:19 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (!cur)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
for (i = 0; i < XFS_BTREE_MAXLEVELS; i++) {
|
|
|
|
if (!cur->bc_bufs[i])
|
|
|
|
break;
|
|
|
|
if (XFS_BUF_ADDR(cur->bc_bufs[i]) == bno)
|
|
|
|
return cur->bc_bufs[i];
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2009-11-25 00:00:19 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/* Chase down all the log items to see if the bp is there */
|
|
|
|
list_for_each_entry(lidp, &cur->bc_tp->t_items, lid_trans) {
|
|
|
|
struct xfs_buf_log_item *bip;
|
|
|
|
bip = (struct xfs_buf_log_item *)lidp->lid_item;
|
|
|
|
if (bip->bli_item.li_type == XFS_LI_BUF &&
|
|
|
|
XFS_BUF_ADDR(bip->bli_buf) == bno)
|
|
|
|
return bip->bli_buf;
|
|
|
|
}
|
2009-11-25 00:00:19 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
return NULL;
|
|
|
|
}
|
2009-12-14 23:14:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
STATIC void
|
|
|
|
xfs_check_block(
|
|
|
|
struct xfs_btree_block *block,
|
|
|
|
xfs_mount_t *mp,
|
|
|
|
int root,
|
|
|
|
short sz)
|
|
|
|
{
|
|
|
|
int i, j, dmxr;
|
|
|
|
__be64 *pp, *thispa; /* pointer to block address */
|
|
|
|
xfs_bmbt_key_t *prevp, *keyp;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(be16_to_cpu(block->bb_level) > 0);
|
2011-05-23 08:52:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
prevp = NULL;
|
|
|
|
for( i = 1; i <= xfs_btree_get_numrecs(block); i++) {
|
|
|
|
dmxr = mp->m_bmap_dmxr[0];
|
|
|
|
keyp = XFS_BMBT_KEY_ADDR(mp, block, i);
|
2009-12-14 23:14:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (prevp) {
|
|
|
|
ASSERT(be64_to_cpu(prevp->br_startoff) <
|
|
|
|
be64_to_cpu(keyp->br_startoff));
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
prevp = keyp;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Compare the block numbers to see if there are dups.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (root)
|
|
|
|
pp = XFS_BMAP_BROOT_PTR_ADDR(mp, block, i, sz);
|
|
|
|
else
|
|
|
|
pp = XFS_BMBT_PTR_ADDR(mp, block, i, dmxr);
|
2009-12-14 23:14:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
for (j = i+1; j <= be16_to_cpu(block->bb_numrecs); j++) {
|
|
|
|
if (root)
|
|
|
|
thispa = XFS_BMAP_BROOT_PTR_ADDR(mp, block, j, sz);
|
|
|
|
else
|
|
|
|
thispa = XFS_BMBT_PTR_ADDR(mp, block, j, dmxr);
|
|
|
|
if (*thispa == *pp) {
|
|
|
|
xfs_warn(mp, "%s: thispa(%d) == pp(%d) %Ld",
|
|
|
|
__func__, j, i,
|
|
|
|
(unsigned long long)be64_to_cpu(*thispa));
|
|
|
|
panic("%s: ptrs are equal in node\n",
|
|
|
|
__func__);
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Check that the extents for the inode ip are in the right order in all
|
2016-01-08 00:28:49 +00:00
|
|
|
* btree leaves. THis becomes prohibitively expensive for large extent count
|
|
|
|
* files, so don't bother with inodes that have more than 10,000 extents in
|
|
|
|
* them. The btree record ordering checks will still be done, so for such large
|
|
|
|
* bmapbt constructs that is going to catch most corruptions.
|
2013-02-25 01:31:26 +00:00
|
|
|
*/
|
|
|
|
STATIC void
|
|
|
|
xfs_bmap_check_leaf_extents(
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor or null */
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
|
|
|
int whichfork) /* data or attr fork */
|
|
|
|
{
|
|
|
|
struct xfs_btree_block *block; /* current btree block */
|
|
|
|
xfs_fsblock_t bno; /* block # of "block" */
|
|
|
|
xfs_buf_t *bp; /* buffer for "block" */
|
|
|
|
int error; /* error return value */
|
|
|
|
xfs_extnum_t i=0, j; /* index into the extents list */
|
|
|
|
xfs_ifork_t *ifp; /* fork structure */
|
|
|
|
int level; /* btree level, for checking */
|
|
|
|
xfs_mount_t *mp; /* file system mount structure */
|
|
|
|
__be64 *pp; /* pointer to block address */
|
|
|
|
xfs_bmbt_rec_t *ep; /* pointer to current extent */
|
|
|
|
xfs_bmbt_rec_t last = {0, 0}; /* last extent in prev block */
|
|
|
|
xfs_bmbt_rec_t *nextp; /* pointer to next extent */
|
|
|
|
int bp_release = 0;
|
|
|
|
|
|
|
|
if (XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE) {
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2016-01-08 00:28:49 +00:00
|
|
|
/* skip large extent count inodes */
|
|
|
|
if (ip->i_d.di_nextents > 10000)
|
|
|
|
return;
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
bno = NULLFSBLOCK;
|
|
|
|
mp = ip->i_mount;
|
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
block = ifp->if_broot;
|
|
|
|
/*
|
|
|
|
* Root level must use BMAP_BROOT_PTR_ADDR macro to get ptr out.
|
|
|
|
*/
|
|
|
|
level = be16_to_cpu(block->bb_level);
|
|
|
|
ASSERT(level > 0);
|
|
|
|
xfs_check_block(block, mp, 1, ifp->if_broot_bytes);
|
|
|
|
pp = XFS_BMAP_BROOT_PTR_ADDR(mp, block, 1, ifp->if_broot_bytes);
|
|
|
|
bno = be64_to_cpu(*pp);
|
|
|
|
|
2014-07-29 23:12:05 +00:00
|
|
|
ASSERT(bno != NULLFSBLOCK);
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(XFS_FSB_TO_AGNO(mp, bno) < mp->m_sb.sb_agcount);
|
|
|
|
ASSERT(XFS_FSB_TO_AGBNO(mp, bno) < mp->m_sb.sb_agblocks);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Go down the tree until leaf level is reached, following the first
|
|
|
|
* pointer (leftmost) at each level.
|
|
|
|
*/
|
|
|
|
while (level-- > 0) {
|
|
|
|
/* See if buf is in cur first */
|
|
|
|
bp_release = 0;
|
|
|
|
bp = xfs_bmap_get_bp(cur, XFS_FSB_TO_DADDR(mp, bno));
|
|
|
|
if (!bp) {
|
|
|
|
bp_release = 1;
|
|
|
|
error = xfs_btree_read_bufl(mp, NULL, bno, 0, &bp,
|
|
|
|
XFS_BMAP_BTREE_REF,
|
|
|
|
&xfs_bmbt_buf_ops);
|
2011-09-18 20:41:04 +00:00
|
|
|
if (error)
|
2013-02-25 01:31:26 +00:00
|
|
|
goto error_norelse;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
block = XFS_BUF_TO_BLOCK(bp);
|
|
|
|
if (level == 0)
|
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Check this block for basic sanity (increasing keys and
|
|
|
|
* no duplicate blocks).
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2009-12-14 23:14:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_check_block(block, mp, 0, 0);
|
|
|
|
pp = XFS_BMBT_PTR_ADDR(mp, block, 1, mp->m_bmap_dmxr[1]);
|
|
|
|
bno = be64_to_cpu(*pp);
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp,
|
|
|
|
XFS_FSB_SANITY_CHECK(mp, bno), error0);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (bp_release) {
|
|
|
|
bp_release = 0;
|
|
|
|
xfs_trans_brelse(NULL, bp);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
2011-05-23 08:52:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Here with bp and block set to the leftmost leaf node in the tree.
|
|
|
|
*/
|
|
|
|
i = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Loop over all leaf nodes checking that all extents are in the right order.
|
|
|
|
*/
|
|
|
|
for (;;) {
|
|
|
|
xfs_fsblock_t nextbno;
|
|
|
|
xfs_extnum_t num_recs;
|
|
|
|
|
|
|
|
|
|
|
|
num_recs = xfs_btree_get_numrecs(block);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Read-ahead the next leaf block, if any.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2011-12-18 20:00:07 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
nextbno = be64_to_cpu(block->bb_u.l.bb_rightsib);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Check all the extents to make sure they are OK.
|
|
|
|
* If we had a previous block, the last entry should
|
|
|
|
* conform with the first entry in this one.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2011-05-23 08:52:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
ep = XFS_BMBT_REC_ADDR(mp, block, 1);
|
|
|
|
if (i) {
|
|
|
|
ASSERT(xfs_bmbt_disk_get_startoff(&last) +
|
|
|
|
xfs_bmbt_disk_get_blockcount(&last) <=
|
|
|
|
xfs_bmbt_disk_get_startoff(ep));
|
|
|
|
}
|
|
|
|
for (j = 1; j < num_recs; j++) {
|
|
|
|
nextp = XFS_BMBT_REC_ADDR(mp, block, j + 1);
|
|
|
|
ASSERT(xfs_bmbt_disk_get_startoff(ep) +
|
|
|
|
xfs_bmbt_disk_get_blockcount(ep) <=
|
|
|
|
xfs_bmbt_disk_get_startoff(nextp));
|
|
|
|
ep = nextp;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
last = *ep;
|
|
|
|
i += num_recs;
|
|
|
|
if (bp_release) {
|
|
|
|
bp_release = 0;
|
|
|
|
xfs_trans_brelse(NULL, bp);
|
|
|
|
}
|
|
|
|
bno = nextbno;
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* If we've reached the end, stop.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (bno == NULLFSBLOCK)
|
|
|
|
break;
|
2011-12-18 20:00:07 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
bp_release = 0;
|
|
|
|
bp = xfs_bmap_get_bp(cur, XFS_FSB_TO_DADDR(mp, bno));
|
|
|
|
if (!bp) {
|
|
|
|
bp_release = 1;
|
|
|
|
error = xfs_btree_read_bufl(mp, NULL, bno, 0, &bp,
|
|
|
|
XFS_BMAP_BTREE_REF,
|
|
|
|
&xfs_bmbt_buf_ops);
|
2011-09-18 20:40:42 +00:00
|
|
|
if (error)
|
2013-02-25 01:31:26 +00:00
|
|
|
goto error_norelse;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
block = XFS_BUF_TO_BLOCK(bp);
|
2011-09-18 20:40:54 +00:00
|
|
|
}
|
2016-03-08 21:17:56 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
return;
|
2011-09-18 20:40:54 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
error0:
|
|
|
|
xfs_warn(mp, "%s: at error0", __func__);
|
|
|
|
if (bp_release)
|
|
|
|
xfs_trans_brelse(NULL, bp);
|
|
|
|
error_norelse:
|
|
|
|
xfs_warn(mp, "%s: BAD after btree leaves for %d extents",
|
|
|
|
__func__, i);
|
|
|
|
panic("%s: CORRUPTED BTREE OR SOMETHING", __func__);
|
|
|
|
return;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Add bmap trace insert entries for all the contents of the extent records.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
void
|
|
|
|
xfs_bmap_trace_exlist(
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
|
|
|
xfs_extnum_t cnt, /* count of entries in the list */
|
2016-12-05 01:32:00 +00:00
|
|
|
int whichfork, /* data or attr or cow fork */
|
2013-02-25 01:31:26 +00:00
|
|
|
unsigned long caller_ip)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_extnum_t idx; /* extent record index */
|
|
|
|
xfs_ifork_t *ifp; /* inode fork pointer */
|
|
|
|
int state = 0;
|
2011-09-18 20:40:54 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (whichfork == XFS_ATTR_FORK)
|
|
|
|
state |= BMAP_ATTRFORK;
|
2016-12-05 01:32:00 +00:00
|
|
|
else if (whichfork == XFS_COW_FORK)
|
|
|
|
state |= BMAP_COWFORK;
|
2011-09-18 20:40:54 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
2016-11-08 01:59:42 +00:00
|
|
|
ASSERT(cnt == xfs_iext_count(ifp));
|
2013-02-25 01:31:26 +00:00
|
|
|
for (idx = 0; idx < cnt; idx++)
|
2016-12-05 01:31:50 +00:00
|
|
|
trace_xfs_extlist(ip, idx, state, caller_ip);
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
2011-09-18 20:40:54 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Validate that the bmbt_irecs being returned from bmapi are valid
|
2013-08-12 03:14:53 +00:00
|
|
|
* given the caller's original parameters. Specifically check the
|
|
|
|
* ranges of the returned irecs to ensure that they only extend beyond
|
2013-02-25 01:31:26 +00:00
|
|
|
* the given parameters if the XFS_BMAPI_ENTIRE flag was set.
|
|
|
|
*/
|
|
|
|
STATIC void
|
|
|
|
xfs_bmap_validate_ret(
|
|
|
|
xfs_fileoff_t bno,
|
|
|
|
xfs_filblks_t len,
|
|
|
|
int flags,
|
|
|
|
xfs_bmbt_irec_t *mval,
|
|
|
|
int nmap,
|
|
|
|
int ret_nmap)
|
|
|
|
{
|
|
|
|
int i; /* index to map values */
|
2011-09-18 20:40:54 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(ret_nmap <= nmap);
|
2011-09-18 20:40:54 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
for (i = 0; i < ret_nmap; i++) {
|
|
|
|
ASSERT(mval[i].br_blockcount > 0);
|
|
|
|
if (!(flags & XFS_BMAPI_ENTIRE)) {
|
|
|
|
ASSERT(mval[i].br_startoff >= bno);
|
|
|
|
ASSERT(mval[i].br_blockcount <= len);
|
|
|
|
ASSERT(mval[i].br_startoff + mval[i].br_blockcount <=
|
|
|
|
bno + len);
|
|
|
|
} else {
|
|
|
|
ASSERT(mval[i].br_startoff < bno + len);
|
|
|
|
ASSERT(mval[i].br_startoff + mval[i].br_blockcount >
|
|
|
|
bno);
|
|
|
|
}
|
|
|
|
ASSERT(i == 0 ||
|
|
|
|
mval[i - 1].br_startoff + mval[i - 1].br_blockcount ==
|
|
|
|
mval[i].br_startoff);
|
|
|
|
ASSERT(mval[i].br_startblock != DELAYSTARTBLOCK &&
|
|
|
|
mval[i].br_startblock != HOLESTARTBLOCK);
|
|
|
|
ASSERT(mval[i].br_state == XFS_EXT_NORM ||
|
|
|
|
mval[i].br_state == XFS_EXT_UNWRITTEN);
|
|
|
|
}
|
|
|
|
}
|
2009-11-25 00:00:19 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
#else
|
|
|
|
#define xfs_bmap_check_leaf_extents(cur, ip, whichfork) do { } while (0)
|
|
|
|
#define xfs_bmap_validate_ret(bno,len,flags,mval,onmap,nmap)
|
|
|
|
#endif /* DEBUG */
|
2009-11-25 00:00:19 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* bmap free list manipulation functions
|
|
|
|
*/
|
2009-11-25 00:00:19 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Add the extent to the list of extents to be free at transaction end.
|
|
|
|
* The list is maintained sorted (by block number).
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
xfs_bmap_add_free(
|
2016-08-03 01:33:42 +00:00
|
|
|
struct xfs_mount *mp,
|
|
|
|
struct xfs_defer_ops *dfops,
|
|
|
|
xfs_fsblock_t bno,
|
|
|
|
xfs_filblks_t len,
|
|
|
|
struct xfs_owner_info *oinfo)
|
2013-02-25 01:31:26 +00:00
|
|
|
{
|
2016-08-03 01:18:10 +00:00
|
|
|
struct xfs_extent_free_item *new; /* new element */
|
2013-02-25 01:31:26 +00:00
|
|
|
#ifdef DEBUG
|
|
|
|
xfs_agnumber_t agno;
|
|
|
|
xfs_agblock_t agbno;
|
|
|
|
|
|
|
|
ASSERT(bno != NULLFSBLOCK);
|
|
|
|
ASSERT(len > 0);
|
|
|
|
ASSERT(len <= MAXEXTLEN);
|
|
|
|
ASSERT(!isnullstartblock(bno));
|
|
|
|
agno = XFS_FSB_TO_AGNO(mp, bno);
|
|
|
|
agbno = XFS_FSB_TO_AGBNO(mp, bno);
|
|
|
|
ASSERT(agno < mp->m_sb.sb_agcount);
|
|
|
|
ASSERT(agbno < mp->m_sb.sb_agblocks);
|
|
|
|
ASSERT(len < mp->m_sb.sb_agblocks);
|
|
|
|
ASSERT(agbno + len <= mp->m_sb.sb_agblocks);
|
|
|
|
#endif
|
|
|
|
ASSERT(xfs_bmap_free_item_zone != NULL);
|
2016-08-03 01:33:42 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
new = kmem_zone_alloc(xfs_bmap_free_item_zone, KM_SLEEP);
|
2016-08-03 01:18:10 +00:00
|
|
|
new->xefi_startblock = bno;
|
|
|
|
new->xefi_blockcount = (xfs_extlen_t)len;
|
2016-08-03 01:33:42 +00:00
|
|
|
if (oinfo)
|
|
|
|
new->xefi_oinfo = *oinfo;
|
|
|
|
else
|
|
|
|
xfs_rmap_skip_owner_update(&new->xefi_oinfo);
|
2016-08-03 01:26:33 +00:00
|
|
|
trace_xfs_bmap_free_defer(mp, XFS_FSB_TO_AGNO(mp, bno), 0,
|
|
|
|
XFS_FSB_TO_AGBNO(mp, bno), len);
|
2016-08-03 01:19:29 +00:00
|
|
|
xfs_defer_add(dfops, XFS_DEFER_OPS_TYPE_FREE, &new->xefi_list);
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
2009-12-14 23:14:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Inode fork format manipulation functions
|
|
|
|
*/
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Transform a btree format file with only one leaf node, where the
|
|
|
|
* extents list will fit in the inode, into an extents format file.
|
|
|
|
* Since the file extents are already in-core, all we have to do is
|
|
|
|
* give up the space for the btree root and pitch the leaf block.
|
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_bmap_btree_to_extents(
|
|
|
|
xfs_trans_t *tp, /* transaction pointer */
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
|
|
|
xfs_btree_cur_t *cur, /* btree cursor */
|
|
|
|
int *logflagsp, /* inode logging flags */
|
|
|
|
int whichfork) /* data or attr fork */
|
|
|
|
{
|
|
|
|
/* REFERENCED */
|
|
|
|
struct xfs_btree_block *cblock;/* child btree block */
|
|
|
|
xfs_fsblock_t cbno; /* child block number */
|
|
|
|
xfs_buf_t *cbp; /* child block's buffer */
|
|
|
|
int error; /* error return value */
|
|
|
|
xfs_ifork_t *ifp; /* inode fork data */
|
|
|
|
xfs_mount_t *mp; /* mount point structure */
|
|
|
|
__be64 *pp; /* ptr to block address */
|
|
|
|
struct xfs_btree_block *rblock;/* root btree block */
|
2016-08-03 01:33:42 +00:00
|
|
|
struct xfs_owner_info oinfo;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
mp = ip->i_mount;
|
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
2016-10-03 16:11:34 +00:00
|
|
|
ASSERT(whichfork != XFS_COW_FORK);
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(ifp->if_flags & XFS_IFEXTENTS);
|
|
|
|
ASSERT(XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_BTREE);
|
|
|
|
rblock = ifp->if_broot;
|
|
|
|
ASSERT(be16_to_cpu(rblock->bb_level) == 1);
|
|
|
|
ASSERT(be16_to_cpu(rblock->bb_numrecs) == 1);
|
|
|
|
ASSERT(xfs_bmbt_maxrecs(mp, ifp->if_broot_bytes, 0) == 1);
|
|
|
|
pp = XFS_BMAP_BROOT_PTR_ADDR(mp, rblock, 1, ifp->if_broot_bytes);
|
|
|
|
cbno = be64_to_cpu(*pp);
|
|
|
|
*logflagsp = 0;
|
|
|
|
#ifdef DEBUG
|
|
|
|
if ((error = xfs_btree_check_lptr(cur, cbno, 1)))
|
|
|
|
return error;
|
|
|
|
#endif
|
|
|
|
error = xfs_btree_read_bufl(mp, tp, cbno, 0, &cbp, XFS_BMAP_BTREE_REF,
|
|
|
|
&xfs_bmbt_buf_ops);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
cblock = XFS_BUF_TO_BLOCK(cbp);
|
|
|
|
if ((error = xfs_btree_check_block(cur, cblock, 0, cbp)))
|
|
|
|
return error;
|
2016-08-03 01:33:42 +00:00
|
|
|
xfs_rmap_ino_bmbt_owner(&oinfo, ip->i_ino, whichfork);
|
|
|
|
xfs_bmap_add_free(mp, cur->bc_private.b.dfops, cbno, 1, &oinfo);
|
2013-02-25 01:31:26 +00:00
|
|
|
ip->i_d.di_nblocks--;
|
|
|
|
xfs_trans_mod_dquot_byino(tp, ip, XFS_TRANS_DQ_BCOUNT, -1L);
|
|
|
|
xfs_trans_binval(tp, cbp);
|
|
|
|
if (cur->bc_bufs[0] == cbp)
|
|
|
|
cur->bc_bufs[0] = NULL;
|
|
|
|
xfs_iroot_realloc(ip, -1, whichfork);
|
|
|
|
ASSERT(ifp->if_broot == NULL);
|
|
|
|
ASSERT((ifp->if_flags & XFS_IFBROOT) == 0);
|
|
|
|
XFS_IFORK_FMT_SET(ip, whichfork, XFS_DINODE_FMT_EXTENTS);
|
|
|
|
*logflagsp = XFS_ILOG_CORE | xfs_ilog_fext(whichfork);
|
|
|
|
return 0;
|
|
|
|
}
|
2009-12-14 23:14:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Convert an extents-format file into a btree-format file.
|
|
|
|
* The new file will have a root block (in the inode) and a single child block.
|
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_bmap_extents_to_btree(
|
|
|
|
xfs_trans_t *tp, /* transaction pointer */
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
|
|
|
xfs_fsblock_t *firstblock, /* first-block-allocated */
|
2016-08-03 01:19:29 +00:00
|
|
|
struct xfs_defer_ops *dfops, /* blocks freed in xaction */
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_btree_cur_t **curp, /* cursor returned to caller */
|
|
|
|
int wasdel, /* converting a delayed alloc */
|
|
|
|
int *logflagsp, /* inode logging flags */
|
|
|
|
int whichfork) /* data or attr fork */
|
|
|
|
{
|
|
|
|
struct xfs_btree_block *ablock; /* allocated (child) bt block */
|
|
|
|
xfs_buf_t *abp; /* buffer for ablock */
|
|
|
|
xfs_alloc_arg_t args; /* allocation arguments */
|
|
|
|
xfs_bmbt_rec_t *arp; /* child record pointer */
|
|
|
|
struct xfs_btree_block *block; /* btree root block */
|
|
|
|
xfs_btree_cur_t *cur; /* bmap btree cursor */
|
|
|
|
xfs_bmbt_rec_host_t *ep; /* extent record pointer */
|
|
|
|
int error; /* error return value */
|
|
|
|
xfs_extnum_t i, cnt; /* extent record index */
|
|
|
|
xfs_ifork_t *ifp; /* inode fork pointer */
|
|
|
|
xfs_bmbt_key_t *kp; /* root block key pointer */
|
|
|
|
xfs_mount_t *mp; /* mount structure */
|
|
|
|
xfs_extnum_t nextents; /* number of file extents */
|
|
|
|
xfs_bmbt_ptr_t *pp; /* root block address pointer */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-04-21 19:53:46 +00:00
|
|
|
mp = ip->i_mount;
|
2016-10-03 16:11:34 +00:00
|
|
|
ASSERT(whichfork != XFS_COW_FORK);
|
2013-02-25 01:31:26 +00:00
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
ASSERT(XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_EXTENTS);
|
2009-12-14 23:14:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Make space in the inode incore.
|
|
|
|
*/
|
|
|
|
xfs_iroot_realloc(ip, 1, whichfork);
|
|
|
|
ifp->if_flags |= XFS_IFBROOT;
|
2011-05-23 08:52:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Fill in the root.
|
|
|
|
*/
|
|
|
|
block = ifp->if_broot;
|
2017-01-28 07:16:39 +00:00
|
|
|
xfs_btree_init_block_int(mp, block, XFS_BUF_DADDR_NULL,
|
|
|
|
XFS_BTNUM_BMAP, 1, 1, ip->i_ino,
|
2017-01-28 07:16:37 +00:00
|
|
|
XFS_BTREE_LONG_PTRS);
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Need a cursor. Can't allocate until bb_level is filled in.
|
|
|
|
*/
|
|
|
|
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
|
|
|
|
cur->bc_private.b.firstblock = *firstblock;
|
2016-08-03 01:19:29 +00:00
|
|
|
cur->bc_private.b.dfops = dfops;
|
2013-02-25 01:31:26 +00:00
|
|
|
cur->bc_private.b.flags = wasdel ? XFS_BTCUR_BPRV_WASDEL : 0;
|
|
|
|
/*
|
|
|
|
* Convert to a btree with two levels, one record in root.
|
|
|
|
*/
|
|
|
|
XFS_IFORK_FMT_SET(ip, whichfork, XFS_DINODE_FMT_BTREE);
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
|
|
args.tp = tp;
|
|
|
|
args.mp = mp;
|
2016-08-03 01:33:42 +00:00
|
|
|
xfs_rmap_ino_bmbt_owner(&args.oinfo, ip->i_ino, whichfork);
|
2013-02-25 01:31:26 +00:00
|
|
|
args.firstblock = *firstblock;
|
|
|
|
if (*firstblock == NULLFSBLOCK) {
|
|
|
|
args.type = XFS_ALLOCTYPE_START_BNO;
|
|
|
|
args.fsbno = XFS_INO_TO_FSB(mp, ip->i_ino);
|
2016-08-03 01:19:29 +00:00
|
|
|
} else if (dfops->dop_low) {
|
2013-02-25 01:31:26 +00:00
|
|
|
args.type = XFS_ALLOCTYPE_START_BNO;
|
|
|
|
args.fsbno = *firstblock;
|
|
|
|
} else {
|
|
|
|
args.type = XFS_ALLOCTYPE_NEAR_BNO;
|
|
|
|
args.fsbno = *firstblock;
|
|
|
|
}
|
|
|
|
args.minlen = args.maxlen = args.prod = 1;
|
|
|
|
args.wasdel = wasdel;
|
|
|
|
*logflagsp = 0;
|
|
|
|
if ((error = xfs_alloc_vextent(&args))) {
|
|
|
|
xfs_iroot_realloc(ip, -1, whichfork);
|
|
|
|
xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
|
|
|
|
return error;
|
|
|
|
}
|
2016-10-03 16:11:45 +00:00
|
|
|
|
2017-03-08 18:38:53 +00:00
|
|
|
if (WARN_ON_ONCE(args.fsbno == NULLFSBLOCK)) {
|
|
|
|
xfs_iroot_realloc(ip, -1, whichfork);
|
|
|
|
xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
|
|
|
|
return -ENOSPC;
|
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Allocation can't fail, the space was reserved.
|
|
|
|
*/
|
|
|
|
ASSERT(*firstblock == NULLFSBLOCK ||
|
2017-02-17 01:12:51 +00:00
|
|
|
args.agno >= XFS_FSB_TO_AGNO(mp, *firstblock));
|
2013-02-25 01:31:26 +00:00
|
|
|
*firstblock = cur->bc_private.b.firstblock = args.fsbno;
|
|
|
|
cur->bc_private.b.allocated++;
|
|
|
|
ip->i_d.di_nblocks++;
|
|
|
|
xfs_trans_mod_dquot_byino(tp, ip, XFS_TRANS_DQ_BCOUNT, 1L);
|
|
|
|
abp = xfs_btree_get_bufl(mp, tp, args.fsbno, 0);
|
|
|
|
/*
|
|
|
|
* Fill in the child block.
|
|
|
|
*/
|
|
|
|
abp->b_ops = &xfs_bmbt_buf_ops;
|
|
|
|
ablock = XFS_BUF_TO_BLOCK(abp);
|
2017-01-28 07:16:39 +00:00
|
|
|
xfs_btree_init_block_int(mp, ablock, abp->b_bn,
|
|
|
|
XFS_BTNUM_BMAP, 0, 0, ip->i_ino,
|
2013-04-21 19:53:46 +00:00
|
|
|
XFS_BTREE_LONG_PTRS);
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
arp = XFS_BMBT_REC_ADDR(mp, ablock, 1);
|
2016-11-08 01:59:42 +00:00
|
|
|
nextents = xfs_iext_count(ifp);
|
2013-02-25 01:31:26 +00:00
|
|
|
for (cnt = i = 0; i < nextents; i++) {
|
|
|
|
ep = xfs_iext_get_ext(ifp, i);
|
|
|
|
if (!isnullstartblock(xfs_bmbt_get_startblock(ep))) {
|
|
|
|
arp->l0 = cpu_to_be64(ep->l0);
|
|
|
|
arp->l1 = cpu_to_be64(ep->l1);
|
|
|
|
arp++; cnt++;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
|
|
|
ASSERT(cnt == XFS_IFORK_NEXTENTS(ip, whichfork));
|
|
|
|
xfs_btree_set_numrecs(ablock, cnt);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Fill in the root key and pointer.
|
|
|
|
*/
|
|
|
|
kp = XFS_BMBT_KEY_ADDR(mp, block, 1);
|
|
|
|
arp = XFS_BMBT_REC_ADDR(mp, ablock, 1);
|
|
|
|
kp->br_startoff = cpu_to_be64(xfs_bmbt_disk_get_startoff(arp));
|
|
|
|
pp = XFS_BMBT_PTR_ADDR(mp, block, 1, xfs_bmbt_get_maxrecs(cur,
|
|
|
|
be16_to_cpu(block->bb_level)));
|
|
|
|
*pp = cpu_to_be64(args.fsbno);
|
2011-05-23 08:52:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Do all this logging at the end so that
|
|
|
|
* the root is at the right level.
|
|
|
|
*/
|
|
|
|
xfs_btree_log_block(cur, abp, XFS_BB_ALL_BITS);
|
|
|
|
xfs_btree_log_recs(cur, abp, 1, be16_to_cpu(ablock->bb_numrecs));
|
|
|
|
ASSERT(*curp == NULL);
|
|
|
|
*curp = cur;
|
|
|
|
*logflagsp = XFS_ILOG_CORE | xfs_ilog_fbroot(whichfork);
|
|
|
|
return 0;
|
|
|
|
}
|
2011-05-23 08:52:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Convert a local file to an extents file.
|
|
|
|
* This code is out of bounds for data forks of regular files,
|
|
|
|
* since the file data needs to get logged so things will stay consistent.
|
|
|
|
* (The bmap-level manipulations are ok, though).
|
|
|
|
*/
|
2013-07-09 21:04:00 +00:00
|
|
|
void
|
|
|
|
xfs_bmap_local_to_extents_empty(
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
int whichfork)
|
|
|
|
{
|
|
|
|
struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
|
2016-10-03 16:11:34 +00:00
|
|
|
ASSERT(whichfork != XFS_COW_FORK);
|
2013-07-09 21:04:00 +00:00
|
|
|
ASSERT(XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_LOCAL);
|
|
|
|
ASSERT(ifp->if_bytes == 0);
|
|
|
|
ASSERT(XFS_IFORK_NEXTENTS(ip, whichfork) == 0);
|
|
|
|
|
2014-04-14 08:59:26 +00:00
|
|
|
xfs_bmap_forkoff_reset(ip, whichfork);
|
2013-07-09 21:04:00 +00:00
|
|
|
ifp->if_flags &= ~XFS_IFINLINE;
|
|
|
|
ifp->if_flags |= XFS_IFEXTENTS;
|
|
|
|
XFS_IFORK_FMT_SET(ip, whichfork, XFS_DINODE_FMT_EXTENTS);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
STATIC int /* error */
|
|
|
|
xfs_bmap_local_to_extents(
|
|
|
|
xfs_trans_t *tp, /* transaction pointer */
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
|
|
|
xfs_fsblock_t *firstblock, /* first block allocated in xaction */
|
|
|
|
xfs_extlen_t total, /* total blocks needed by transaction */
|
|
|
|
int *logflagsp, /* inode logging flags */
|
|
|
|
int whichfork,
|
2013-04-21 19:53:46 +00:00
|
|
|
void (*init_fn)(struct xfs_trans *tp,
|
|
|
|
struct xfs_buf *bp,
|
2013-02-25 01:31:26 +00:00
|
|
|
struct xfs_inode *ip,
|
|
|
|
struct xfs_ifork *ifp))
|
|
|
|
{
|
2013-07-09 21:04:00 +00:00
|
|
|
int error = 0;
|
2013-02-25 01:31:26 +00:00
|
|
|
int flags; /* logging flags returned */
|
|
|
|
xfs_ifork_t *ifp; /* inode fork pointer */
|
2013-07-09 21:04:00 +00:00
|
|
|
xfs_alloc_arg_t args; /* allocation arguments */
|
|
|
|
xfs_buf_t *bp; /* buffer for extent block */
|
2017-08-29 22:44:11 +00:00
|
|
|
struct xfs_bmbt_irec rec;
|
2009-12-14 23:14:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* We don't want to deal with the case of keeping inode data inline yet.
|
|
|
|
* So sending the data fork of a regular inode is invalid.
|
|
|
|
*/
|
2016-02-09 05:54:58 +00:00
|
|
|
ASSERT(!(S_ISREG(VFS_I(ip)->i_mode) && whichfork == XFS_DATA_FORK));
|
2013-02-25 01:31:26 +00:00
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
ASSERT(XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_LOCAL);
|
2013-07-09 21:04:00 +00:00
|
|
|
|
|
|
|
if (!ifp->if_bytes) {
|
|
|
|
xfs_bmap_local_to_extents_empty(ip, whichfork);
|
|
|
|
flags = XFS_ILOG_CORE;
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
flags = 0;
|
|
|
|
error = 0;
|
2013-07-09 21:04:00 +00:00
|
|
|
ASSERT((ifp->if_flags & (XFS_IFINLINE|XFS_IFEXTENTS|XFS_IFEXTIREC)) ==
|
|
|
|
XFS_IFINLINE);
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
|
|
args.tp = tp;
|
|
|
|
args.mp = ip->i_mount;
|
2016-08-03 01:33:42 +00:00
|
|
|
xfs_rmap_ino_owner(&args.oinfo, ip->i_ino, whichfork, 0);
|
2013-07-09 21:04:00 +00:00
|
|
|
args.firstblock = *firstblock;
|
|
|
|
/*
|
|
|
|
* Allocate a block. We know we need only one, since the
|
|
|
|
* file currently fits in an inode.
|
|
|
|
*/
|
|
|
|
if (*firstblock == NULLFSBLOCK) {
|
|
|
|
args.fsbno = XFS_INO_TO_FSB(args.mp, ip->i_ino);
|
|
|
|
args.type = XFS_ALLOCTYPE_START_BNO;
|
2013-02-25 01:31:26 +00:00
|
|
|
} else {
|
2013-07-09 21:04:00 +00:00
|
|
|
args.fsbno = *firstblock;
|
|
|
|
args.type = XFS_ALLOCTYPE_NEAR_BNO;
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
2013-07-09 21:04:00 +00:00
|
|
|
args.total = total;
|
|
|
|
args.minlen = args.maxlen = args.prod = 1;
|
|
|
|
error = xfs_alloc_vextent(&args);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
|
|
|
|
/* Can't fail, the space was reserved. */
|
|
|
|
ASSERT(args.fsbno != NULLFSBLOCK);
|
|
|
|
ASSERT(args.len == 1);
|
|
|
|
*firstblock = args.fsbno;
|
|
|
|
bp = xfs_btree_get_bufl(args.mp, tp, args.fsbno, 0);
|
|
|
|
|
2015-01-21 22:30:06 +00:00
|
|
|
/*
|
2015-10-12 04:40:24 +00:00
|
|
|
* Initialize the block, copy the data and log the remote buffer.
|
2015-01-21 22:30:06 +00:00
|
|
|
*
|
2015-10-12 04:40:24 +00:00
|
|
|
* The callout is responsible for logging because the remote format
|
|
|
|
* might differ from the local format and thus we don't know how much to
|
|
|
|
* log here. Note that init_fn must also set the buffer log item type
|
|
|
|
* correctly.
|
2015-01-21 22:30:06 +00:00
|
|
|
*/
|
2013-07-09 21:04:00 +00:00
|
|
|
init_fn(tp, bp, ip, ifp);
|
|
|
|
|
2015-10-12 04:40:24 +00:00
|
|
|
/* account for the change in fork size */
|
2013-07-09 21:04:00 +00:00
|
|
|
xfs_idata_realloc(ip, -ifp->if_bytes, whichfork);
|
|
|
|
xfs_bmap_local_to_extents_empty(ip, whichfork);
|
2013-02-25 01:31:26 +00:00
|
|
|
flags |= XFS_ILOG_CORE;
|
2013-07-09 21:04:00 +00:00
|
|
|
|
2017-08-29 22:44:11 +00:00
|
|
|
rec.br_startoff = 0;
|
|
|
|
rec.br_startblock = args.fsbno;
|
|
|
|
rec.br_blockcount = 1;
|
|
|
|
rec.br_state = XFS_EXT_NORM;
|
|
|
|
xfs_iext_insert(ip, 0, 1, &rec, 0);
|
|
|
|
|
2013-07-09 21:04:00 +00:00
|
|
|
trace_xfs_bmap_post_update(ip, 0,
|
|
|
|
whichfork == XFS_ATTR_FORK ? BMAP_ATTRFORK : 0,
|
|
|
|
_THIS_IP_);
|
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork, 1);
|
|
|
|
ip->i_d.di_nblocks = 1;
|
|
|
|
xfs_trans_mod_dquot_byino(tp, ip,
|
|
|
|
XFS_TRANS_DQ_BCOUNT, 1L);
|
|
|
|
flags |= xfs_ilog_fext(whichfork);
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
done:
|
|
|
|
*logflagsp = flags;
|
|
|
|
return error;
|
|
|
|
}
|
2011-05-23 08:52:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Called from xfs_bmap_add_attrfork to handle btree format files.
|
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_bmap_add_attrfork_btree(
|
|
|
|
xfs_trans_t *tp, /* transaction pointer */
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
|
|
|
xfs_fsblock_t *firstblock, /* first block allocated */
|
2016-08-03 01:19:29 +00:00
|
|
|
struct xfs_defer_ops *dfops, /* blocks to free at commit */
|
2013-02-25 01:31:26 +00:00
|
|
|
int *flags) /* inode logging flags */
|
|
|
|
{
|
|
|
|
xfs_btree_cur_t *cur; /* btree cursor */
|
|
|
|
int error; /* error return value */
|
|
|
|
xfs_mount_t *mp; /* file system mount struct */
|
|
|
|
int stat; /* newroot status */
|
2011-05-23 08:52:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
mp = ip->i_mount;
|
|
|
|
if (ip->i_df.if_broot_bytes <= XFS_IFORK_DSIZE(ip))
|
|
|
|
*flags |= XFS_ILOG_DBROOT;
|
|
|
|
else {
|
|
|
|
cur = xfs_bmbt_init_cursor(mp, tp, ip, XFS_DATA_FORK);
|
2016-08-03 01:19:29 +00:00
|
|
|
cur->bc_private.b.dfops = dfops;
|
2013-02-25 01:31:26 +00:00
|
|
|
cur->bc_private.b.firstblock = *firstblock;
|
|
|
|
if ((error = xfs_bmbt_lookup_ge(cur, 0, 0, 0, &stat)))
|
|
|
|
goto error0;
|
|
|
|
/* must be at least one entry */
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, stat == 1, error0);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_btree_new_iroot(cur, flags, &stat)))
|
|
|
|
goto error0;
|
|
|
|
if (stat == 0) {
|
|
|
|
xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
|
2014-06-25 04:58:08 +00:00
|
|
|
return -ENOSPC;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
*firstblock = cur->bc_private.b.firstblock;
|
|
|
|
cur->bc_private.b.allocated = 0;
|
|
|
|
xfs_btree_del_cursor(cur, XFS_BTREE_NOERROR);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
return 0;
|
|
|
|
error0:
|
|
|
|
xfs_btree_del_cursor(cur, XFS_BTREE_ERROR);
|
|
|
|
return error;
|
|
|
|
}
|
2011-09-18 20:40:54 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Called from xfs_bmap_add_attrfork to handle extents format files.
|
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_bmap_add_attrfork_extents(
|
|
|
|
xfs_trans_t *tp, /* transaction pointer */
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
|
|
|
xfs_fsblock_t *firstblock, /* first block allocated */
|
2016-08-03 01:19:29 +00:00
|
|
|
struct xfs_defer_ops *dfops, /* blocks to free at commit */
|
2013-02-25 01:31:26 +00:00
|
|
|
int *flags) /* inode logging flags */
|
|
|
|
{
|
|
|
|
xfs_btree_cur_t *cur; /* bmap btree cursor */
|
|
|
|
int error; /* error return value */
|
2011-09-18 20:40:54 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (ip->i_d.di_nextents * sizeof(xfs_bmbt_rec_t) <= XFS_IFORK_DSIZE(ip))
|
|
|
|
return 0;
|
|
|
|
cur = NULL;
|
2016-08-03 01:19:29 +00:00
|
|
|
error = xfs_bmap_extents_to_btree(tp, ip, firstblock, dfops, &cur, 0,
|
2013-02-25 01:31:26 +00:00
|
|
|
flags, XFS_DATA_FORK);
|
2011-09-18 20:40:54 +00:00
|
|
|
if (cur) {
|
|
|
|
cur->bc_private.b.allocated = 0;
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_btree_del_cursor(cur,
|
|
|
|
error ? XFS_BTREE_ERROR : XFS_BTREE_NOERROR);
|
2011-09-18 20:40:54 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Called from xfs_bmap_add_attrfork to handle local format files. Each
|
|
|
|
* different data fork content type needs a different callout to do the
|
|
|
|
* conversion. Some are basic and only require special block initialisation
|
|
|
|
* callouts for the data formating, others (directories) are so specialised they
|
|
|
|
* handle everything themselves.
|
|
|
|
*
|
|
|
|
* XXX (dgc): investigate whether directory conversion can use the generic
|
|
|
|
* formatting callout. It should be possible - it's just a very complex
|
2013-04-21 19:53:46 +00:00
|
|
|
* formatter.
|
2013-02-25 01:31:26 +00:00
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_bmap_add_attrfork_local(
|
|
|
|
xfs_trans_t *tp, /* transaction pointer */
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
|
|
|
xfs_fsblock_t *firstblock, /* first block allocated */
|
2016-08-03 01:19:29 +00:00
|
|
|
struct xfs_defer_ops *dfops, /* blocks to free at commit */
|
2013-02-25 01:31:26 +00:00
|
|
|
int *flags) /* inode logging flags */
|
|
|
|
{
|
|
|
|
xfs_da_args_t dargs; /* args for dir/attr code */
|
2009-11-25 00:00:19 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (ip->i_df.if_bytes <= XFS_IFORK_DSIZE(ip))
|
|
|
|
return 0;
|
2009-11-25 00:00:19 +00:00
|
|
|
|
2016-02-09 05:54:58 +00:00
|
|
|
if (S_ISDIR(VFS_I(ip)->i_mode)) {
|
2013-02-25 01:31:26 +00:00
|
|
|
memset(&dargs, 0, sizeof(dargs));
|
2014-06-06 05:14:11 +00:00
|
|
|
dargs.geo = ip->i_mount->m_dir_geo;
|
2013-02-25 01:31:26 +00:00
|
|
|
dargs.dp = ip;
|
|
|
|
dargs.firstblock = firstblock;
|
2016-08-03 01:19:29 +00:00
|
|
|
dargs.dfops = dfops;
|
2014-06-06 05:14:11 +00:00
|
|
|
dargs.total = dargs.geo->fsbcount;
|
2013-02-25 01:31:26 +00:00
|
|
|
dargs.whichfork = XFS_DATA_FORK;
|
|
|
|
dargs.trans = tp;
|
|
|
|
return xfs_dir2_sf_to_block(&dargs);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2009-11-25 00:00:19 +00:00
|
|
|
|
2016-02-09 05:54:58 +00:00
|
|
|
if (S_ISLNK(VFS_I(ip)->i_mode))
|
2013-02-25 01:31:26 +00:00
|
|
|
return xfs_bmap_local_to_extents(tp, ip, firstblock, 1,
|
|
|
|
flags, XFS_DATA_FORK,
|
|
|
|
xfs_symlink_local_to_remote);
|
2009-11-25 00:00:19 +00:00
|
|
|
|
2013-07-09 21:04:00 +00:00
|
|
|
/* should only be called for types that support local format data */
|
|
|
|
ASSERT(0);
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EFSCORRUPTED;
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
2009-12-14 23:14:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Convert inode from non-attributed to attributed.
|
|
|
|
* Must not be in a transaction, ip must not be locked.
|
|
|
|
*/
|
|
|
|
int /* error code */
|
|
|
|
xfs_bmap_add_attrfork(
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
|
|
|
int size, /* space new attribute needs */
|
|
|
|
int rsvd) /* xact may use reserved blks */
|
|
|
|
{
|
|
|
|
xfs_fsblock_t firstblock; /* 1st block/ag allocated */
|
2016-08-03 01:19:29 +00:00
|
|
|
struct xfs_defer_ops dfops; /* freed extent records */
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_mount_t *mp; /* mount structure */
|
|
|
|
xfs_trans_t *tp; /* transaction pointer */
|
|
|
|
int blks; /* space reservation */
|
|
|
|
int version = 1; /* superblock attr version */
|
|
|
|
int logflags; /* logging flags */
|
|
|
|
int error; /* error return value */
|
2009-12-14 23:14:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(XFS_IFORK_Q(ip) == 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
mp = ip->i_mount;
|
|
|
|
ASSERT(!XFS_NOT_DQATTACHED(mp, ip));
|
2016-04-05 23:19:55 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
blks = XFS_ADDAFORK_SPACE_RES(mp);
|
2016-04-05 23:19:55 +00:00
|
|
|
|
|
|
|
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_addafork, blks, 0,
|
|
|
|
rsvd ? XFS_TRANS_RESERVE : 0, &tp);
|
|
|
|
if (error)
|
2013-11-07 21:43:28 +00:00
|
|
|
return error;
|
2016-04-05 23:19:55 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_ilock(ip, XFS_ILOCK_EXCL);
|
|
|
|
error = xfs_trans_reserve_quota_nblks(tp, ip, blks, 0, rsvd ?
|
|
|
|
XFS_QMOPT_RES_REGBLKS | XFS_QMOPT_FORCE_RES :
|
|
|
|
XFS_QMOPT_RES_REGBLKS);
|
2013-11-07 21:43:28 +00:00
|
|
|
if (error)
|
|
|
|
goto trans_cancel;
|
2013-02-25 01:31:26 +00:00
|
|
|
if (XFS_IFORK_Q(ip))
|
2013-11-07 21:43:28 +00:00
|
|
|
goto trans_cancel;
|
2016-12-05 01:38:11 +00:00
|
|
|
if (ip->i_d.di_anextents != 0) {
|
|
|
|
error = -EFSCORRUPTED;
|
|
|
|
goto trans_cancel;
|
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
if (ip->i_d.di_aformat != XFS_DINODE_FMT_EXTENTS) {
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* For inodes coming from pre-6.2 filesystems.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(ip->i_d.di_aformat == 0);
|
|
|
|
ip->i_d.di_aformat = XFS_DINODE_FMT_EXTENTS;
|
|
|
|
}
|
2011-05-23 08:52:53 +00:00
|
|
|
|
2013-11-07 21:43:28 +00:00
|
|
|
xfs_trans_ijoin(tp, ip, 0);
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
switch (ip->i_d.di_format) {
|
|
|
|
case XFS_DINODE_FMT_DEV:
|
|
|
|
ip->i_d.di_forkoff = roundup(sizeof(xfs_dev_t), 8) >> 3;
|
|
|
|
break;
|
|
|
|
case XFS_DINODE_FMT_UUID:
|
|
|
|
ip->i_d.di_forkoff = roundup(sizeof(uuid_t), 8) >> 3;
|
|
|
|
break;
|
|
|
|
case XFS_DINODE_FMT_LOCAL:
|
|
|
|
case XFS_DINODE_FMT_EXTENTS:
|
|
|
|
case XFS_DINODE_FMT_BTREE:
|
|
|
|
ip->i_d.di_forkoff = xfs_attr_shortform_bytesfit(ip, size);
|
|
|
|
if (!ip->i_d.di_forkoff)
|
|
|
|
ip->i_d.di_forkoff = xfs_default_attroffset(ip) >> 3;
|
|
|
|
else if (mp->m_flags & XFS_MOUNT_ATTR2)
|
|
|
|
version = 2;
|
2005-04-16 22:20:36 +00:00
|
|
|
break;
|
2013-02-25 01:31:26 +00:00
|
|
|
default:
|
|
|
|
ASSERT(0);
|
2014-06-25 04:58:08 +00:00
|
|
|
error = -EINVAL;
|
2013-11-07 21:43:28 +00:00
|
|
|
goto trans_cancel;
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(ip->i_afp == NULL);
|
|
|
|
ip->i_afp = kmem_zone_zalloc(xfs_ifork_zone, KM_SLEEP);
|
|
|
|
ip->i_afp->if_flags = XFS_IFEXTENTS;
|
|
|
|
logflags = 0;
|
2016-08-03 01:19:29 +00:00
|
|
|
xfs_defer_init(&dfops, &firstblock);
|
2013-02-25 01:31:26 +00:00
|
|
|
switch (ip->i_d.di_format) {
|
|
|
|
case XFS_DINODE_FMT_LOCAL:
|
2016-08-03 01:19:29 +00:00
|
|
|
error = xfs_bmap_add_attrfork_local(tp, ip, &firstblock, &dfops,
|
2013-02-25 01:31:26 +00:00
|
|
|
&logflags);
|
|
|
|
break;
|
|
|
|
case XFS_DINODE_FMT_EXTENTS:
|
|
|
|
error = xfs_bmap_add_attrfork_extents(tp, ip, &firstblock,
|
2016-08-03 01:19:29 +00:00
|
|
|
&dfops, &logflags);
|
2013-02-25 01:31:26 +00:00
|
|
|
break;
|
|
|
|
case XFS_DINODE_FMT_BTREE:
|
2016-08-03 01:19:29 +00:00
|
|
|
error = xfs_bmap_add_attrfork_btree(tp, ip, &firstblock, &dfops,
|
2013-02-25 01:31:26 +00:00
|
|
|
&logflags);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
error = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
break;
|
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
if (logflags)
|
|
|
|
xfs_trans_log_inode(tp, ip, logflags);
|
|
|
|
if (error)
|
2013-11-07 21:43:28 +00:00
|
|
|
goto bmap_cancel;
|
2013-02-25 01:31:26 +00:00
|
|
|
if (!xfs_sb_version_hasattr(&mp->m_sb) ||
|
|
|
|
(!xfs_sb_version_hasattr2(&mp->m_sb) && version == 2)) {
|
2015-01-21 22:10:31 +00:00
|
|
|
bool log_sb = false;
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
spin_lock(&mp->m_sb_lock);
|
|
|
|
if (!xfs_sb_version_hasattr(&mp->m_sb)) {
|
|
|
|
xfs_sb_version_addattr(&mp->m_sb);
|
2015-01-21 22:10:31 +00:00
|
|
|
log_sb = true;
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
|
|
|
if (!xfs_sb_version_hasattr2(&mp->m_sb) && version == 2) {
|
|
|
|
xfs_sb_version_addattr2(&mp->m_sb);
|
2015-01-21 22:10:31 +00:00
|
|
|
log_sb = true;
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
xfs: remove bitfield based superblock updates
When we log changes to the superblock, we first have to write them
to the on-disk buffer, and then log that. Right now we have a
complex bitfield based arrangement to only write the modified field
to the buffer before we log it.
This used to be necessary as a performance optimisation because we
logged the superblock buffer in every extent or inode allocation or
freeing, and so performance was extremely important. We haven't done
this for years, however, ever since the lazy superblock counters
pulled the superblock logging out of the transaction commit
fast path.
Hence we have a bunch of complexity that is not necessary that makes
writing the in-core superblock to disk much more complex than it
needs to be. We only need to log the superblock now during
management operations (e.g. during mount, unmount or quota control
operations) so it is not a performance critical path anymore.
As such, remove the complex field based logging mechanism and
replace it with a simple conversion function similar to what we use
for all other on-disk structures.
This means we always log the entirity of the superblock, but again
because we rarely modify the superblock this is not an issue for log
bandwidth or CPU time. Indeed, if we do log the superblock
frequently, delayed logging will minimise the impact of this
overhead.
[Fixed gquota/pquota inode sharing regression noticed by bfoster.]
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2015-01-21 22:10:26 +00:00
|
|
|
spin_unlock(&mp->m_sb_lock);
|
2015-01-21 22:10:31 +00:00
|
|
|
if (log_sb)
|
|
|
|
xfs_log_sb(tp);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
|
2017-08-28 17:21:04 +00:00
|
|
|
error = xfs_defer_finish(&tp, &dfops);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (error)
|
2013-11-07 21:43:28 +00:00
|
|
|
goto bmap_cancel;
|
2015-06-04 03:48:08 +00:00
|
|
|
error = xfs_trans_commit(tp);
|
2013-11-07 21:43:28 +00:00
|
|
|
xfs_iunlock(ip, XFS_ILOCK_EXCL);
|
|
|
|
return error;
|
|
|
|
|
|
|
|
bmap_cancel:
|
2016-08-03 01:19:29 +00:00
|
|
|
xfs_defer_cancel(&dfops);
|
2013-11-07 21:43:28 +00:00
|
|
|
trans_cancel:
|
2015-06-04 03:47:56 +00:00
|
|
|
xfs_trans_cancel(tp);
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_iunlock(ip, XFS_ILOCK_EXCL);
|
|
|
|
return error;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Internal and external extent tree search functions.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2011-09-18 20:40:54 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Read in the extents to if_extents.
|
|
|
|
* All inode fields are set up by caller, we just traverse the btree
|
|
|
|
* and copy the records in. If the file system cannot contain unwritten
|
|
|
|
* extents, the records are checked for no "state" flags.
|
|
|
|
*/
|
|
|
|
int /* error */
|
|
|
|
xfs_bmap_read_extents(
|
|
|
|
xfs_trans_t *tp, /* transaction pointer */
|
|
|
|
xfs_inode_t *ip, /* incore inode */
|
|
|
|
int whichfork) /* data or attr fork */
|
|
|
|
{
|
|
|
|
struct xfs_btree_block *block; /* current btree block */
|
|
|
|
xfs_fsblock_t bno; /* block # of "block" */
|
|
|
|
xfs_buf_t *bp; /* buffer for "block" */
|
|
|
|
int error; /* error return value */
|
|
|
|
xfs_extnum_t i, j; /* index into the extents list */
|
|
|
|
xfs_ifork_t *ifp; /* fork structure */
|
|
|
|
int level; /* btree level, for checking */
|
|
|
|
xfs_mount_t *mp; /* file system mount structure */
|
|
|
|
__be64 *pp; /* pointer to block address */
|
|
|
|
/* REFERENCED */
|
|
|
|
xfs_extnum_t room; /* number of entries there's room for */
|
2009-11-25 00:00:21 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
mp = ip->i_mount;
|
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
block = ifp->if_broot;
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Root level must use BMAP_BROOT_PTR_ADDR macro to get ptr out.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
level = be16_to_cpu(block->bb_level);
|
|
|
|
ASSERT(level > 0);
|
|
|
|
pp = XFS_BMAP_BROOT_PTR_ADDR(mp, block, 1, ifp->if_broot_bytes);
|
|
|
|
bno = be64_to_cpu(*pp);
|
2017-02-02 23:13:58 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Go down the tree until leaf level is reached, following the first
|
|
|
|
* pointer (leftmost) at each level.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
while (level-- > 0) {
|
|
|
|
error = xfs_btree_read_bufl(mp, tp, bno, 0, &bp,
|
|
|
|
XFS_BMAP_BTREE_REF, &xfs_bmbt_buf_ops);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
block = XFS_BUF_TO_BLOCK(bp);
|
|
|
|
if (level == 0)
|
|
|
|
break;
|
|
|
|
pp = XFS_BMBT_PTR_ADDR(mp, block, 1, mp->m_bmap_dmxr[1]);
|
|
|
|
bno = be64_to_cpu(*pp);
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp,
|
|
|
|
XFS_FSB_SANITY_CHECK(mp, bno), error0);
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_trans_brelse(tp, bp);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Here with bp and block set to the leftmost leaf node in the tree.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2016-11-08 01:59:42 +00:00
|
|
|
room = xfs_iext_count(ifp);
|
2013-02-25 01:31:26 +00:00
|
|
|
i = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Loop over all leaf nodes. Copy information to the extent records.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
for (;;) {
|
|
|
|
xfs_bmbt_rec_t *frp;
|
|
|
|
xfs_fsblock_t nextbno;
|
|
|
|
xfs_extnum_t num_recs;
|
2009-12-14 23:14:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
num_recs = xfs_btree_get_numrecs(block);
|
|
|
|
if (unlikely(i + num_recs > room)) {
|
|
|
|
ASSERT(i + num_recs <= room);
|
|
|
|
xfs_warn(ip->i_mount,
|
|
|
|
"corrupt dinode %Lu, (btree extents).",
|
|
|
|
(unsigned long long) ip->i_ino);
|
|
|
|
XFS_CORRUPTION_ERROR("xfs_bmap_read_extents(1)",
|
|
|
|
XFS_ERRLEVEL_LOW, ip->i_mount, block);
|
|
|
|
goto error0;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Read-ahead the next leaf block, if any.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
nextbno = be64_to_cpu(block->bb_u.l.bb_rightsib);
|
|
|
|
if (nextbno != NULLFSBLOCK)
|
|
|
|
xfs_btree_reada_bufl(mp, nextbno, 1,
|
|
|
|
&xfs_bmbt_buf_ops);
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Copy records into the extent records.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
frp = XFS_BMBT_REC_ADDR(mp, block, 1);
|
|
|
|
for (j = 0; j < num_recs; j++, i++, frp++) {
|
|
|
|
xfs_bmbt_rec_host_t *trp = xfs_iext_get_ext(ifp, i);
|
|
|
|
trp->l0 = be64_to_cpu(frp->l0);
|
|
|
|
trp->l1 = be64_to_cpu(frp->l1);
|
2017-04-20 16:42:48 +00:00
|
|
|
if (!xfs_bmbt_validate_extent(mp, whichfork, trp)) {
|
2013-02-25 01:31:26 +00:00
|
|
|
XFS_ERROR_REPORT("xfs_bmap_read_extents(2)",
|
2017-04-20 16:42:48 +00:00
|
|
|
XFS_ERRLEVEL_LOW, mp);
|
2013-02-25 01:31:26 +00:00
|
|
|
goto error0;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
xfs_trans_brelse(tp, bp);
|
|
|
|
bno = nextbno;
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* If we've reached the end, stop.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (bno == NULLFSBLOCK)
|
|
|
|
break;
|
|
|
|
error = xfs_btree_read_bufl(mp, tp, bno, 0, &bp,
|
|
|
|
XFS_BMAP_BTREE_REF, &xfs_bmbt_buf_ops);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
block = XFS_BUF_TO_BLOCK(bp);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2016-12-05 01:36:56 +00:00
|
|
|
if (i != XFS_IFORK_NEXTENTS(ip, whichfork))
|
|
|
|
return -EFSCORRUPTED;
|
2016-11-08 01:59:42 +00:00
|
|
|
ASSERT(i == xfs_iext_count(ifp));
|
2013-02-25 01:31:26 +00:00
|
|
|
XFS_BMAP_TRACE_EXLIST(ip, i, whichfork);
|
|
|
|
return 0;
|
|
|
|
error0:
|
|
|
|
xfs_trans_brelse(tp, bp);
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EFSCORRUPTED;
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
2011-09-18 20:40:54 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Returns the file-relative block number of the first unused block(s)
|
|
|
|
* in the file with at least "len" logically contiguous blocks free.
|
|
|
|
* This is the lowest-address hole if the file has holes, else the first block
|
|
|
|
* past the end of file.
|
|
|
|
* Return 0 if the file is currently local (in-inode).
|
|
|
|
*/
|
|
|
|
int /* error */
|
|
|
|
xfs_bmap_first_unused(
|
|
|
|
xfs_trans_t *tp, /* transaction pointer */
|
|
|
|
xfs_inode_t *ip, /* incore inode */
|
|
|
|
xfs_extlen_t len, /* size of hole to find */
|
|
|
|
xfs_fileoff_t *first_unused, /* unused block */
|
|
|
|
int whichfork) /* data or attr fork */
|
|
|
|
{
|
|
|
|
int error; /* error return value */
|
|
|
|
int idx; /* extent record index */
|
|
|
|
xfs_ifork_t *ifp; /* inode fork pointer */
|
|
|
|
xfs_fileoff_t lastaddr; /* last block number seen */
|
|
|
|
xfs_fileoff_t lowest; /* lowest useful block */
|
|
|
|
xfs_fileoff_t max; /* starting useful block */
|
|
|
|
xfs_fileoff_t off; /* offset for this block */
|
|
|
|
xfs_extnum_t nextents; /* number of extent entries */
|
|
|
|
|
|
|
|
ASSERT(XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_BTREE ||
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_EXTENTS ||
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_LOCAL);
|
|
|
|
if (XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_LOCAL) {
|
|
|
|
*first_unused = 0;
|
|
|
|
return 0;
|
2006-01-11 04:28:28 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
if (!(ifp->if_flags & XFS_IFEXTENTS) &&
|
|
|
|
(error = xfs_iread_extents(tp, ip, whichfork)))
|
|
|
|
return error;
|
|
|
|
lowest = *first_unused;
|
2016-11-08 01:59:42 +00:00
|
|
|
nextents = xfs_iext_count(ifp);
|
2013-02-25 01:31:26 +00:00
|
|
|
for (idx = 0, lastaddr = 0, max = lowest; idx < nextents; idx++) {
|
|
|
|
xfs_bmbt_rec_host_t *ep = xfs_iext_get_ext(ifp, idx);
|
|
|
|
off = xfs_bmbt_get_startoff(ep);
|
|
|
|
/*
|
|
|
|
* See if the hole before this extent will work.
|
|
|
|
*/
|
|
|
|
if (off >= lowest + len && off - max >= len) {
|
|
|
|
*first_unused = max;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
lastaddr = off + xfs_bmbt_get_blockcount(ep);
|
|
|
|
max = XFS_FILEOFF_MAX(lastaddr, lowest);
|
2006-01-11 04:28:28 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
*first_unused = max;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2013-08-12 03:14:54 +00:00
|
|
|
* Returns the file-relative block number of the last block - 1 before
|
2013-02-25 01:31:26 +00:00
|
|
|
* last_block (input value) in the file.
|
|
|
|
* This is not based on i_size, it is based on the extent records.
|
|
|
|
* Returns 0 for local files, as they do not have extent records.
|
|
|
|
*/
|
|
|
|
int /* error */
|
|
|
|
xfs_bmap_last_before(
|
2016-11-24 00:39:38 +00:00
|
|
|
struct xfs_trans *tp, /* transaction pointer */
|
|
|
|
struct xfs_inode *ip, /* incore inode */
|
|
|
|
xfs_fileoff_t *last_block, /* last block */
|
|
|
|
int whichfork) /* data or attr fork */
|
2013-02-25 01:31:26 +00:00
|
|
|
{
|
2016-11-24 00:39:38 +00:00
|
|
|
struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
struct xfs_bmbt_irec got;
|
|
|
|
xfs_extnum_t idx;
|
|
|
|
int error;
|
2013-02-25 01:31:26 +00:00
|
|
|
|
2016-11-24 00:39:38 +00:00
|
|
|
switch (XFS_IFORK_FORMAT(ip, whichfork)) {
|
|
|
|
case XFS_DINODE_FMT_LOCAL:
|
2013-02-25 01:31:26 +00:00
|
|
|
*last_block = 0;
|
|
|
|
return 0;
|
2016-11-24 00:39:38 +00:00
|
|
|
case XFS_DINODE_FMT_BTREE:
|
|
|
|
case XFS_DINODE_FMT_EXTENTS:
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
return -EIO;
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
2016-11-24 00:39:38 +00:00
|
|
|
|
|
|
|
if (!(ifp->if_flags & XFS_IFEXTENTS)) {
|
|
|
|
error = xfs_iread_extents(tp, ip, whichfork);
|
|
|
|
if (error)
|
|
|
|
return error;
|
2006-01-11 04:28:28 +00:00
|
|
|
}
|
2016-11-24 00:39:38 +00:00
|
|
|
|
|
|
|
if (xfs_iext_lookup_extent(ip, ifp, *last_block - 1, &idx, &got)) {
|
|
|
|
if (got.br_startoff <= *last_block - 1)
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (xfs_iext_get_extent(ifp, idx - 1, &got)) {
|
|
|
|
*last_block = got.br_startoff + got.br_blockcount;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
*last_block = 0;
|
2013-02-25 01:31:26 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-08-12 10:49:42 +00:00
|
|
|
int
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmap_last_extent(
|
|
|
|
struct xfs_trans *tp,
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
int whichfork,
|
|
|
|
struct xfs_bmbt_irec *rec,
|
|
|
|
int *is_empty)
|
|
|
|
{
|
|
|
|
struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
int error;
|
|
|
|
int nextents;
|
|
|
|
|
|
|
|
if (!(ifp->if_flags & XFS_IFEXTENTS)) {
|
|
|
|
error = xfs_iread_extents(tp, ip, whichfork);
|
|
|
|
if (error)
|
|
|
|
return error;
|
2006-01-11 04:28:28 +00:00
|
|
|
}
|
|
|
|
|
2016-11-08 01:59:42 +00:00
|
|
|
nextents = xfs_iext_count(ifp);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (nextents == 0) {
|
|
|
|
*is_empty = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
2006-01-11 04:28:28 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(ifp, nextents - 1), rec);
|
|
|
|
*is_empty = 0;
|
2006-01-11 04:28:28 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Check the last inode extent to determine whether this allocation will result
|
|
|
|
* in blocks being allocated at the end of the file. When we allocate new data
|
|
|
|
* blocks at the end of the file which do not start at the previous data block,
|
|
|
|
* we will try to align the new blocks at stripe unit boundaries.
|
|
|
|
*
|
2013-11-21 23:41:16 +00:00
|
|
|
* Returns 1 in bma->aeof if the file (fork) is empty as any new write will be
|
2013-02-25 01:31:26 +00:00
|
|
|
* at, or past the EOF.
|
|
|
|
*/
|
|
|
|
STATIC int
|
|
|
|
xfs_bmap_isaeof(
|
|
|
|
struct xfs_bmalloca *bma,
|
|
|
|
int whichfork)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
struct xfs_bmbt_irec rec;
|
|
|
|
int is_empty;
|
|
|
|
int error;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
bma->aeof = 0;
|
|
|
|
error = xfs_bmap_last_extent(NULL, bma->ip, whichfork, &rec,
|
|
|
|
&is_empty);
|
2013-11-21 23:41:16 +00:00
|
|
|
if (error)
|
2013-02-25 01:31:26 +00:00
|
|
|
return error;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-11-21 23:41:16 +00:00
|
|
|
if (is_empty) {
|
|
|
|
bma->aeof = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Check if we are allocation or past the last extent, or at least into
|
|
|
|
* the last delayed allocated extent.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
bma->aeof = bma->offset >= rec.br_startoff + rec.br_blockcount ||
|
|
|
|
(bma->offset >= rec.br_startoff &&
|
|
|
|
isnullstartblock(rec.br_startblock));
|
|
|
|
return 0;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Returns the file-relative block number of the first block past eof in
|
|
|
|
* the file. This is not based on i_size, it is based on the extent records.
|
|
|
|
* Returns 0 for local files, as they do not have extent records.
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
xfs_bmap_last_offset(
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
xfs_fileoff_t *last_block,
|
|
|
|
int whichfork)
|
|
|
|
{
|
|
|
|
struct xfs_bmbt_irec rec;
|
|
|
|
int is_empty;
|
|
|
|
int error;
|
2011-01-25 09:06:19 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
*last_block = 0;
|
2007-06-28 06:46:56 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_LOCAL)
|
|
|
|
return 0;
|
2006-03-14 02:34:16 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE &&
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS)
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EIO;
|
2006-03-14 02:34:16 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_bmap_last_extent(NULL, ip, whichfork, &rec, &is_empty);
|
|
|
|
if (error || is_empty)
|
2006-03-14 02:34:16 +00:00
|
|
|
return error;
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
*last_block = rec.br_startoff + rec.br_blockcount;
|
2006-03-14 02:34:16 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Returns whether the selected fork of the inode has exactly one
|
|
|
|
* block or not. For the data fork we check this matches di_size,
|
|
|
|
* implying the file's range is 0..bsize-1.
|
|
|
|
*/
|
|
|
|
int /* 1=>1 block, 0=>otherwise */
|
|
|
|
xfs_bmap_one_block(
|
|
|
|
xfs_inode_t *ip, /* incore inode */
|
|
|
|
int whichfork) /* data or attr fork */
|
2010-02-15 23:34:42 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmbt_rec_host_t *ep; /* ptr to fork's extent */
|
|
|
|
xfs_ifork_t *ifp; /* inode fork pointer */
|
|
|
|
int rval; /* return value */
|
|
|
|
xfs_bmbt_irec_t s; /* internal version of extent */
|
2010-02-15 23:34:42 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
#ifndef DEBUG
|
|
|
|
if (whichfork == XFS_DATA_FORK)
|
|
|
|
return XFS_ISIZE(ip) == ip->i_mount->m_sb.sb_blocksize;
|
|
|
|
#endif /* !DEBUG */
|
|
|
|
if (XFS_IFORK_NEXTENTS(ip, whichfork) != 1)
|
|
|
|
return 0;
|
|
|
|
if (XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS)
|
|
|
|
return 0;
|
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
ASSERT(ifp->if_flags & XFS_IFEXTENTS);
|
|
|
|
ep = xfs_iext_get_ext(ifp, 0);
|
|
|
|
xfs_bmbt_get_all(ep, &s);
|
|
|
|
rval = s.br_startoff == 0 && s.br_blockcount == 1;
|
|
|
|
if (rval && whichfork == XFS_DATA_FORK)
|
|
|
|
ASSERT(XFS_ISIZE(ip) == ip->i_mount->m_sb.sb_blocksize);
|
|
|
|
return rval;
|
|
|
|
}
|
2010-02-15 23:34:42 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Extent tree manipulation functions used during allocation.
|
|
|
|
*/
|
2010-02-15 23:34:42 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Convert a delayed allocation to a real allocation.
|
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_bmap_add_extent_delay_real(
|
2016-10-03 16:11:34 +00:00
|
|
|
struct xfs_bmalloca *bma,
|
|
|
|
int whichfork)
|
2013-02-25 01:31:26 +00:00
|
|
|
{
|
|
|
|
struct xfs_bmbt_irec *new = &bma->got;
|
|
|
|
int diff; /* temp value */
|
|
|
|
xfs_bmbt_rec_host_t *ep; /* extent entry for idx */
|
|
|
|
int error; /* error return value */
|
|
|
|
int i; /* temp state */
|
|
|
|
xfs_ifork_t *ifp; /* inode fork pointer */
|
|
|
|
xfs_fileoff_t new_endoff; /* end offset of new entry */
|
|
|
|
xfs_bmbt_irec_t r[3]; /* neighbor extent entries */
|
|
|
|
/* left is 0, right is 1, prev is 2 */
|
|
|
|
int rval=0; /* return value (logging flags) */
|
|
|
|
int state = 0;/* state bits, accessed thru macros */
|
|
|
|
xfs_filblks_t da_new; /* new count del alloc blocks used */
|
|
|
|
xfs_filblks_t da_old; /* old count del alloc blocks used */
|
|
|
|
xfs_filblks_t temp=0; /* value for da_new calculations */
|
|
|
|
xfs_filblks_t temp2=0;/* value for da_new calculations */
|
|
|
|
int tmp_rval; /* partial logging flags */
|
2015-02-23 11:39:08 +00:00
|
|
|
struct xfs_mount *mp;
|
2016-10-03 16:11:34 +00:00
|
|
|
xfs_extnum_t *nextents;
|
2010-02-15 23:34:42 +00:00
|
|
|
|
2016-01-04 05:10:42 +00:00
|
|
|
mp = bma->ip->i_mount;
|
2016-01-04 05:12:42 +00:00
|
|
|
ifp = XFS_IFORK_PTR(bma->ip, whichfork);
|
2016-10-03 16:11:34 +00:00
|
|
|
ASSERT(whichfork != XFS_ATTR_FORK);
|
|
|
|
nextents = (whichfork == XFS_COW_FORK ? &bma->ip->i_cnextents :
|
|
|
|
&bma->ip->i_d.di_nextents);
|
2010-02-15 23:34:42 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(bma->idx >= 0);
|
2016-11-08 01:59:42 +00:00
|
|
|
ASSERT(bma->idx <= xfs_iext_count(ifp));
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(!isnullstartblock(new->br_startblock));
|
|
|
|
ASSERT(!bma->cur ||
|
|
|
|
(bma->cur->bc_private.b.flags & XFS_BTCUR_BPRV_WASDEL));
|
2010-02-15 23:34:42 +00:00
|
|
|
|
2015-10-12 07:21:22 +00:00
|
|
|
XFS_STATS_INC(mp, xs_add_exlist);
|
2010-02-15 23:34:42 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
#define LEFT r[0]
|
|
|
|
#define RIGHT r[1]
|
|
|
|
#define PREV r[2]
|
2010-02-15 23:34:42 +00:00
|
|
|
|
2016-10-03 16:11:34 +00:00
|
|
|
if (whichfork == XFS_COW_FORK)
|
|
|
|
state |= BMAP_COWFORK;
|
|
|
|
|
2010-02-15 23:34:42 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Set up a bunch of variables to make the tests simpler.
|
2010-02-15 23:34:42 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
ep = xfs_iext_get_ext(ifp, bma->idx);
|
|
|
|
xfs_bmbt_get_all(ep, &PREV);
|
|
|
|
new_endoff = new->br_startoff + new->br_blockcount;
|
|
|
|
ASSERT(PREV.br_startoff <= new->br_startoff);
|
|
|
|
ASSERT(PREV.br_startoff + PREV.br_blockcount >= new_endoff);
|
|
|
|
|
|
|
|
da_old = startblockval(PREV.br_startblock);
|
|
|
|
da_new = 0;
|
|
|
|
|
2010-02-15 23:34:42 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Set flags determining what part of the previous delayed allocation
|
|
|
|
* extent is being replaced by a real allocation.
|
2010-02-15 23:34:42 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (PREV.br_startoff == new->br_startoff)
|
|
|
|
state |= BMAP_LEFT_FILLING;
|
|
|
|
if (PREV.br_startoff + PREV.br_blockcount == new_endoff)
|
|
|
|
state |= BMAP_RIGHT_FILLING;
|
2010-02-15 23:34:42 +00:00
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Check and set flags if this segment has a left neighbor.
|
|
|
|
* Don't set contiguous if the combined extent would be too large.
|
2010-02-15 23:34:42 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (bma->idx > 0) {
|
|
|
|
state |= BMAP_LEFT_VALID;
|
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(ifp, bma->idx - 1), &LEFT);
|
2011-12-01 11:24:20 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (isnullstartblock(LEFT.br_startblock))
|
|
|
|
state |= BMAP_LEFT_DELAY;
|
2006-03-14 02:34:16 +00:00
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((state & BMAP_LEFT_VALID) && !(state & BMAP_LEFT_DELAY) &&
|
|
|
|
LEFT.br_startoff + LEFT.br_blockcount == new->br_startoff &&
|
|
|
|
LEFT.br_startblock + LEFT.br_blockcount == new->br_startblock &&
|
|
|
|
LEFT.br_state == new->br_state &&
|
|
|
|
LEFT.br_blockcount + new->br_blockcount <= MAXEXTLEN)
|
|
|
|
state |= BMAP_LEFT_CONTIG;
|
2006-03-14 02:34:16 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Check and set flags if this segment has a right neighbor.
|
|
|
|
* Don't set contiguous if the combined extent would be too large.
|
|
|
|
* Also check for all-three-contiguous being too large.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2016-11-08 01:59:42 +00:00
|
|
|
if (bma->idx < xfs_iext_count(ifp) - 1) {
|
2013-02-25 01:31:26 +00:00
|
|
|
state |= BMAP_RIGHT_VALID;
|
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(ifp, bma->idx + 1), &RIGHT);
|
2011-01-27 01:16:28 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (isnullstartblock(RIGHT.br_startblock))
|
|
|
|
state |= BMAP_RIGHT_DELAY;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
if ((state & BMAP_RIGHT_VALID) && !(state & BMAP_RIGHT_DELAY) &&
|
|
|
|
new_endoff == RIGHT.br_startoff &&
|
|
|
|
new->br_startblock + new->br_blockcount == RIGHT.br_startblock &&
|
|
|
|
new->br_state == RIGHT.br_state &&
|
|
|
|
new->br_blockcount + RIGHT.br_blockcount <= MAXEXTLEN &&
|
|
|
|
((state & (BMAP_LEFT_CONTIG | BMAP_LEFT_FILLING |
|
|
|
|
BMAP_RIGHT_FILLING)) !=
|
|
|
|
(BMAP_LEFT_CONTIG | BMAP_LEFT_FILLING |
|
|
|
|
BMAP_RIGHT_FILLING) ||
|
|
|
|
LEFT.br_blockcount + new->br_blockcount + RIGHT.br_blockcount
|
|
|
|
<= MAXEXTLEN))
|
|
|
|
state |= BMAP_RIGHT_CONTIG;
|
|
|
|
|
|
|
|
error = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Switch out based on the FILLING and CONTIG state bits.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
switch (state & (BMAP_LEFT_FILLING | BMAP_LEFT_CONTIG |
|
|
|
|
BMAP_RIGHT_FILLING | BMAP_RIGHT_CONTIG)) {
|
|
|
|
case BMAP_LEFT_FILLING | BMAP_LEFT_CONTIG |
|
|
|
|
BMAP_RIGHT_FILLING | BMAP_RIGHT_CONTIG:
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Filling in all of a previously delayed allocation extent.
|
|
|
|
* The left and right neighbors are both contiguous with new.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
bma->idx--;
|
|
|
|
trace_xfs_bmap_pre_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, bma->idx),
|
|
|
|
LEFT.br_blockcount + PREV.br_blockcount +
|
|
|
|
RIGHT.br_blockcount);
|
|
|
|
trace_xfs_bmap_post_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
|
|
|
|
xfs_iext_remove(bma->ip, bma->idx + 1, 2, state);
|
2016-10-03 16:11:34 +00:00
|
|
|
(*nextents)--;
|
2013-02-25 01:31:26 +00:00
|
|
|
if (bma->cur == NULL)
|
|
|
|
rval = XFS_ILOG_CORE | XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = XFS_ILOG_CORE;
|
|
|
|
error = xfs_bmbt_lookup_eq(bma->cur, RIGHT.br_startoff,
|
|
|
|
RIGHT.br_startblock,
|
|
|
|
RIGHT.br_blockcount, &i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_btree_delete(bma->cur, &i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_btree_decrement(bma->cur, 0, &i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_bmbt_update(bma->cur, LEFT.br_startoff,
|
|
|
|
LEFT.br_startblock,
|
|
|
|
LEFT.br_blockcount +
|
|
|
|
PREV.br_blockcount +
|
|
|
|
RIGHT.br_blockcount, LEFT.br_state);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case BMAP_LEFT_FILLING | BMAP_RIGHT_FILLING | BMAP_LEFT_CONTIG:
|
2006-03-14 02:34:16 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Filling in all of a previously delayed allocation extent.
|
|
|
|
* The left neighbor is contiguous, the right is not.
|
2006-03-14 02:34:16 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
bma->idx--;
|
|
|
|
|
|
|
|
trace_xfs_bmap_pre_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, bma->idx),
|
|
|
|
LEFT.br_blockcount + PREV.br_blockcount);
|
|
|
|
trace_xfs_bmap_post_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
|
|
|
|
xfs_iext_remove(bma->ip, bma->idx + 1, 1, state);
|
|
|
|
if (bma->cur == NULL)
|
|
|
|
rval = XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = 0;
|
|
|
|
error = xfs_bmbt_lookup_eq(bma->cur, LEFT.br_startoff,
|
|
|
|
LEFT.br_startblock, LEFT.br_blockcount,
|
|
|
|
&i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_bmbt_update(bma->cur, LEFT.br_startoff,
|
|
|
|
LEFT.br_startblock,
|
|
|
|
LEFT.br_blockcount +
|
|
|
|
PREV.br_blockcount, LEFT.br_state);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case BMAP_LEFT_FILLING | BMAP_RIGHT_FILLING | BMAP_RIGHT_CONTIG:
|
2011-09-18 20:40:57 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Filling in all of a previously delayed allocation extent.
|
|
|
|
* The right neighbor is contiguous, the left is not.
|
2011-09-18 20:40:57 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_pre_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_startblock(ep, new->br_startblock);
|
|
|
|
xfs_bmbt_set_blockcount(ep,
|
|
|
|
PREV.br_blockcount + RIGHT.br_blockcount);
|
|
|
|
trace_xfs_bmap_post_update(bma->ip, bma->idx, state, _THIS_IP_);
|
2011-09-18 20:40:57 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_iext_remove(bma->ip, bma->idx + 1, 1, state);
|
|
|
|
if (bma->cur == NULL)
|
|
|
|
rval = XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = 0;
|
|
|
|
error = xfs_bmbt_lookup_eq(bma->cur, RIGHT.br_startoff,
|
|
|
|
RIGHT.br_startblock,
|
|
|
|
RIGHT.br_blockcount, &i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_bmbt_update(bma->cur, PREV.br_startoff,
|
|
|
|
new->br_startblock,
|
|
|
|
PREV.br_blockcount +
|
|
|
|
RIGHT.br_blockcount, PREV.br_state);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case BMAP_LEFT_FILLING | BMAP_RIGHT_FILLING:
|
2006-03-14 02:34:16 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Filling in all of a previously delayed allocation extent.
|
|
|
|
* Neither the left nor right neighbors are contiguous with
|
|
|
|
* the new one.
|
2006-03-14 02:34:16 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_pre_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_startblock(ep, new->br_startblock);
|
2017-02-02 23:14:01 +00:00
|
|
|
xfs_bmbt_set_state(ep, new->br_state);
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_post_update(bma->ip, bma->idx, state, _THIS_IP_);
|
2006-03-14 02:34:16 +00:00
|
|
|
|
2016-10-03 16:11:34 +00:00
|
|
|
(*nextents)++;
|
2013-02-25 01:31:26 +00:00
|
|
|
if (bma->cur == NULL)
|
|
|
|
rval = XFS_ILOG_CORE | XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = XFS_ILOG_CORE;
|
|
|
|
error = xfs_bmbt_lookup_eq(bma->cur, new->br_startoff,
|
|
|
|
new->br_startblock, new->br_blockcount,
|
|
|
|
&i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 0, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
bma->cur->bc_rec.b.br_state = XFS_EXT_NORM;
|
|
|
|
error = xfs_btree_insert(bma->cur, &i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_LEFT_FILLING | BMAP_LEFT_CONTIG:
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Filling in the first part of a previous delayed allocation.
|
|
|
|
* The left neighbor is contiguous.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_pre_update(bma->ip, bma->idx - 1, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, bma->idx - 1),
|
|
|
|
LEFT.br_blockcount + new->br_blockcount);
|
|
|
|
xfs_bmbt_set_startoff(ep,
|
|
|
|
PREV.br_startoff + new->br_blockcount);
|
|
|
|
trace_xfs_bmap_post_update(bma->ip, bma->idx - 1, state, _THIS_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
temp = PREV.br_blockcount - new->br_blockcount;
|
|
|
|
trace_xfs_bmap_pre_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(ep, temp);
|
|
|
|
if (bma->cur == NULL)
|
|
|
|
rval = XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = 0;
|
|
|
|
error = xfs_bmbt_lookup_eq(bma->cur, LEFT.br_startoff,
|
|
|
|
LEFT.br_startblock, LEFT.br_blockcount,
|
|
|
|
&i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_bmbt_update(bma->cur, LEFT.br_startoff,
|
|
|
|
LEFT.br_startblock,
|
|
|
|
LEFT.br_blockcount +
|
|
|
|
new->br_blockcount,
|
|
|
|
LEFT.br_state);
|
2011-07-08 12:34:47 +00:00
|
|
|
if (error)
|
2005-04-16 22:20:36 +00:00
|
|
|
goto done;
|
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
da_new = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(bma->ip, temp),
|
|
|
|
startblockval(PREV.br_startblock));
|
|
|
|
xfs_bmbt_set_startblock(ep, nullstartblock(da_new));
|
|
|
|
trace_xfs_bmap_post_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
|
|
|
|
bma->idx--;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case BMAP_LEFT_FILLING:
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Filling in the first part of a previous delayed allocation.
|
|
|
|
* The left neighbor is not contiguous.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_pre_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_startoff(ep, new_endoff);
|
|
|
|
temp = PREV.br_blockcount - new->br_blockcount;
|
|
|
|
xfs_bmbt_set_blockcount(ep, temp);
|
|
|
|
xfs_iext_insert(bma->ip, bma->idx, 1, new, state);
|
2016-10-03 16:11:34 +00:00
|
|
|
(*nextents)++;
|
2013-02-25 01:31:26 +00:00
|
|
|
if (bma->cur == NULL)
|
|
|
|
rval = XFS_ILOG_CORE | XFS_ILOG_DEXT;
|
2005-04-16 22:20:36 +00:00
|
|
|
else {
|
2013-02-25 01:31:26 +00:00
|
|
|
rval = XFS_ILOG_CORE;
|
|
|
|
error = xfs_bmbt_lookup_eq(bma->cur, new->br_startoff,
|
|
|
|
new->br_startblock, new->br_blockcount,
|
|
|
|
&i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 0, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
bma->cur->bc_rec.b.br_state = XFS_EXT_NORM;
|
|
|
|
error = xfs_btree_insert(bma->cur, &i);
|
|
|
|
if (error)
|
2005-04-16 22:20:36 +00:00
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2011-05-11 15:04:05 +00:00
|
|
|
|
2016-01-04 05:12:42 +00:00
|
|
|
if (xfs_bmap_needs_btree(bma->ip, whichfork)) {
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_bmap_extents_to_btree(bma->tp, bma->ip,
|
2016-08-03 01:19:29 +00:00
|
|
|
bma->firstblock, bma->dfops,
|
2016-01-04 05:12:42 +00:00
|
|
|
&bma->cur, 1, &tmp_rval, whichfork);
|
2013-02-25 01:31:26 +00:00
|
|
|
rval |= tmp_rval;
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
da_new = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(bma->ip, temp),
|
|
|
|
startblockval(PREV.br_startblock) -
|
|
|
|
(bma->cur ? bma->cur->bc_private.b.allocated : 0));
|
|
|
|
ep = xfs_iext_get_ext(ifp, bma->idx + 1);
|
|
|
|
xfs_bmbt_set_startblock(ep, nullstartblock(da_new));
|
|
|
|
trace_xfs_bmap_post_update(bma->ip, bma->idx + 1, state, _THIS_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
break;
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_RIGHT_FILLING | BMAP_RIGHT_CONTIG:
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Filling in the last part of a previous delayed allocation.
|
|
|
|
* The right neighbor is contiguous with the new allocation.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
temp = PREV.br_blockcount - new->br_blockcount;
|
|
|
|
trace_xfs_bmap_pre_update(bma->ip, bma->idx + 1, state, _THIS_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
xfs_bmbt_set_blockcount(ep, temp);
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmbt_set_allf(xfs_iext_get_ext(ifp, bma->idx + 1),
|
|
|
|
new->br_startoff, new->br_startblock,
|
|
|
|
new->br_blockcount + RIGHT.br_blockcount,
|
|
|
|
RIGHT.br_state);
|
|
|
|
trace_xfs_bmap_post_update(bma->ip, bma->idx + 1, state, _THIS_IP_);
|
|
|
|
if (bma->cur == NULL)
|
|
|
|
rval = XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = 0;
|
|
|
|
error = xfs_bmbt_lookup_eq(bma->cur, RIGHT.br_startoff,
|
|
|
|
RIGHT.br_startblock,
|
|
|
|
RIGHT.br_blockcount, &i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_bmbt_update(bma->cur, new->br_startoff,
|
|
|
|
new->br_startblock,
|
|
|
|
new->br_blockcount +
|
|
|
|
RIGHT.br_blockcount,
|
|
|
|
RIGHT.br_state);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
da_new = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(bma->ip, temp),
|
|
|
|
startblockval(PREV.br_startblock));
|
|
|
|
trace_xfs_bmap_pre_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_startblock(ep, nullstartblock(da_new));
|
|
|
|
trace_xfs_bmap_post_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
|
|
|
|
bma->idx++;
|
2005-04-16 22:20:36 +00:00
|
|
|
break;
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_RIGHT_FILLING:
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Filling in the last part of a previous delayed allocation.
|
|
|
|
* The right neighbor is not contiguous.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
temp = PREV.br_blockcount - new->br_blockcount;
|
|
|
|
trace_xfs_bmap_pre_update(bma->ip, bma->idx, state, _THIS_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
xfs_bmbt_set_blockcount(ep, temp);
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_iext_insert(bma->ip, bma->idx + 1, 1, new, state);
|
2016-10-03 16:11:34 +00:00
|
|
|
(*nextents)++;
|
2013-02-25 01:31:26 +00:00
|
|
|
if (bma->cur == NULL)
|
|
|
|
rval = XFS_ILOG_CORE | XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = XFS_ILOG_CORE;
|
|
|
|
error = xfs_bmbt_lookup_eq(bma->cur, new->br_startoff,
|
|
|
|
new->br_startblock, new->br_blockcount,
|
|
|
|
&i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 0, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
bma->cur->bc_rec.b.br_state = XFS_EXT_NORM;
|
|
|
|
error = xfs_btree_insert(bma->cur, &i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
|
2016-01-04 05:12:42 +00:00
|
|
|
if (xfs_bmap_needs_btree(bma->ip, whichfork)) {
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_bmap_extents_to_btree(bma->tp, bma->ip,
|
2016-08-03 01:19:29 +00:00
|
|
|
bma->firstblock, bma->dfops, &bma->cur, 1,
|
2016-01-04 05:12:42 +00:00
|
|
|
&tmp_rval, whichfork);
|
2013-02-25 01:31:26 +00:00
|
|
|
rval |= tmp_rval;
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
da_new = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(bma->ip, temp),
|
|
|
|
startblockval(PREV.br_startblock) -
|
|
|
|
(bma->cur ? bma->cur->bc_private.b.allocated : 0));
|
|
|
|
ep = xfs_iext_get_ext(ifp, bma->idx);
|
|
|
|
xfs_bmbt_set_startblock(ep, nullstartblock(da_new));
|
|
|
|
trace_xfs_bmap_post_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
|
|
|
|
bma->idx++;
|
2005-04-16 22:20:36 +00:00
|
|
|
break;
|
|
|
|
|
|
|
|
case 0:
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Filling in the middle part of a previous delayed allocation.
|
|
|
|
* Contiguity is impossible here.
|
|
|
|
* This case is avoided almost all the time.
|
|
|
|
*
|
|
|
|
* We start with a delayed allocation:
|
|
|
|
*
|
|
|
|
* +ddddddddddddddddddddddddddddddddddddddddddddddddddddddd+
|
|
|
|
* PREV @ idx
|
|
|
|
*
|
|
|
|
* and we are allocating:
|
|
|
|
* +rrrrrrrrrrrrrrrrr+
|
|
|
|
* new
|
|
|
|
*
|
|
|
|
* and we set it up for insertion as:
|
|
|
|
* +ddddddddddddddddddd+rrrrrrrrrrrrrrrrr+ddddddddddddddddd+
|
|
|
|
* new
|
|
|
|
* PREV @ idx LEFT RIGHT
|
|
|
|
* inserted at idx + 1
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
temp = new->br_startoff - PREV.br_startoff;
|
|
|
|
temp2 = PREV.br_startoff + PREV.br_blockcount - new_endoff;
|
|
|
|
trace_xfs_bmap_pre_update(bma->ip, bma->idx, 0, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(ep, temp); /* truncate PREV */
|
|
|
|
LEFT = *new;
|
|
|
|
RIGHT.br_state = PREV.br_state;
|
|
|
|
RIGHT.br_startblock = nullstartblock(
|
|
|
|
(int)xfs_bmap_worst_indlen(bma->ip, temp2));
|
|
|
|
RIGHT.br_startoff = new_endoff;
|
|
|
|
RIGHT.br_blockcount = temp2;
|
|
|
|
/* insert LEFT (r[0]) and RIGHT (r[1]) at the same time */
|
|
|
|
xfs_iext_insert(bma->ip, bma->idx + 1, 2, &LEFT, state);
|
2016-10-03 16:11:34 +00:00
|
|
|
(*nextents)++;
|
2013-02-25 01:31:26 +00:00
|
|
|
if (bma->cur == NULL)
|
|
|
|
rval = XFS_ILOG_CORE | XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = XFS_ILOG_CORE;
|
|
|
|
error = xfs_bmbt_lookup_eq(bma->cur, new->br_startoff,
|
|
|
|
new->br_startblock, new->br_blockcount,
|
|
|
|
&i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 0, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
bma->cur->bc_rec.b.br_state = XFS_EXT_NORM;
|
|
|
|
error = xfs_btree_insert(bma->cur, &i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2016-01-04 05:12:42 +00:00
|
|
|
if (xfs_bmap_needs_btree(bma->ip, whichfork)) {
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_bmap_extents_to_btree(bma->tp, bma->ip,
|
2016-08-03 01:19:29 +00:00
|
|
|
bma->firstblock, bma->dfops, &bma->cur,
|
2016-01-04 05:12:42 +00:00
|
|
|
1, &tmp_rval, whichfork);
|
2013-02-25 01:31:26 +00:00
|
|
|
rval |= tmp_rval;
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
temp = xfs_bmap_worst_indlen(bma->ip, temp);
|
|
|
|
temp2 = xfs_bmap_worst_indlen(bma->ip, temp2);
|
2017-05-12 17:44:08 +00:00
|
|
|
diff = (int)(temp + temp2 -
|
|
|
|
(startblockval(PREV.br_startblock) -
|
|
|
|
(bma->cur ?
|
|
|
|
bma->cur->bc_private.b.allocated : 0)));
|
2013-02-25 01:31:26 +00:00
|
|
|
if (diff > 0) {
|
2015-02-23 10:22:03 +00:00
|
|
|
error = xfs_mod_fdblocks(bma->ip->i_mount,
|
|
|
|
-((int64_t)diff), false);
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(!error);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
|
|
|
|
ep = xfs_iext_get_ext(ifp, bma->idx);
|
|
|
|
xfs_bmbt_set_startblock(ep, nullstartblock((int)temp));
|
|
|
|
trace_xfs_bmap_post_update(bma->ip, bma->idx, state, _THIS_IP_);
|
|
|
|
trace_xfs_bmap_pre_update(bma->ip, bma->idx + 2, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_startblock(xfs_iext_get_ext(ifp, bma->idx + 2),
|
|
|
|
nullstartblock((int)temp2));
|
|
|
|
trace_xfs_bmap_post_update(bma->ip, bma->idx + 2, state, _THIS_IP_);
|
|
|
|
|
|
|
|
bma->idx++;
|
|
|
|
da_new = temp + temp2;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case BMAP_LEFT_FILLING | BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG:
|
|
|
|
case BMAP_RIGHT_FILLING | BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG:
|
|
|
|
case BMAP_LEFT_FILLING | BMAP_RIGHT_CONTIG:
|
|
|
|
case BMAP_RIGHT_FILLING | BMAP_LEFT_CONTIG:
|
|
|
|
case BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG:
|
|
|
|
case BMAP_LEFT_CONTIG:
|
|
|
|
case BMAP_RIGHT_CONTIG:
|
|
|
|
/*
|
|
|
|
* These cases are all impossible.
|
|
|
|
*/
|
|
|
|
ASSERT(0);
|
|
|
|
}
|
|
|
|
|
xfs: propagate bmap updates to rmapbt
When we map, unmap, or convert an extent in a file's data or attr
fork, schedule a respective update in the rmapbt. Previous versions
of this patch required a 1:1 correspondence between bmap and rmap,
but this is no longer true as we now have ability to make interval
queries against the rmapbt.
We use the deferred operations code to handle redo operations
atomically and deadlock free. This plumbs in all five rmap actions
(map, unmap, convert extent, alloc, free); we'll use the first three
now for file data, and reflink will want the last two. We also add
an error injection site to test log recovery.
Finally, we need to fix the bmap shift extent code to adjust the
rmaps correctly.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-08-03 02:16:05 +00:00
|
|
|
/* add reverse mapping */
|
|
|
|
error = xfs_rmap_map_extent(mp, bma->dfops, bma->ip, whichfork, new);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/* convert to a btree if necessary */
|
2016-01-04 05:12:42 +00:00
|
|
|
if (xfs_bmap_needs_btree(bma->ip, whichfork)) {
|
2013-02-25 01:31:26 +00:00
|
|
|
int tmp_logflags; /* partial log flag return val */
|
|
|
|
|
|
|
|
ASSERT(bma->cur == NULL);
|
|
|
|
error = xfs_bmap_extents_to_btree(bma->tp, bma->ip,
|
2016-08-03 01:19:29 +00:00
|
|
|
bma->firstblock, bma->dfops, &bma->cur,
|
2016-01-04 05:12:42 +00:00
|
|
|
da_old > 0, &tmp_logflags, whichfork);
|
2013-02-25 01:31:26 +00:00
|
|
|
bma->logflags |= tmp_logflags;
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* adjust for changes in reserved delayed indirect blocks */
|
|
|
|
if (da_old || da_new) {
|
|
|
|
temp = da_new;
|
|
|
|
if (bma->cur)
|
|
|
|
temp += bma->cur->bc_private.b.allocated;
|
|
|
|
if (temp < da_old)
|
2015-02-23 10:22:03 +00:00
|
|
|
xfs_mod_fdblocks(bma->ip->i_mount,
|
|
|
|
(int64_t)(da_old - temp), false);
|
2010-09-30 02:25:55 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
/* clear out the allocated field, done with it now in any case. */
|
|
|
|
if (bma->cur)
|
|
|
|
bma->cur->bc_private.b.allocated = 0;
|
|
|
|
|
2016-01-04 05:12:42 +00:00
|
|
|
xfs_bmap_check_leaf_extents(bma->cur, bma->ip, whichfork);
|
2005-04-16 22:20:36 +00:00
|
|
|
done:
|
2016-10-03 16:11:34 +00:00
|
|
|
if (whichfork != XFS_COW_FORK)
|
|
|
|
bma->logflags |= rval;
|
2005-04-16 22:20:36 +00:00
|
|
|
return error;
|
2013-02-25 01:31:26 +00:00
|
|
|
#undef LEFT
|
|
|
|
#undef RIGHT
|
|
|
|
#undef PREV
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Convert an unwritten allocation to a real allocation or vice versa.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
STATIC int /* error */
|
|
|
|
xfs_bmap_add_extent_unwritten_real(
|
|
|
|
struct xfs_trans *tp,
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
2017-02-02 23:14:01 +00:00
|
|
|
int whichfork,
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_extnum_t *idx, /* extent number to update/insert */
|
|
|
|
xfs_btree_cur_t **curp, /* if *curp is null, not a btree */
|
|
|
|
xfs_bmbt_irec_t *new, /* new data to add to file extents */
|
|
|
|
xfs_fsblock_t *first, /* pointer to firstblock variable */
|
2016-08-03 01:19:29 +00:00
|
|
|
struct xfs_defer_ops *dfops, /* list of extents to be freed */
|
2013-02-25 01:31:26 +00:00
|
|
|
int *logflagsp) /* inode logging flags */
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_btree_cur_t *cur; /* btree cursor */
|
|
|
|
xfs_bmbt_rec_host_t *ep; /* extent entry for idx */
|
|
|
|
int error; /* error return value */
|
|
|
|
int i; /* temp state */
|
|
|
|
xfs_ifork_t *ifp; /* inode fork pointer */
|
|
|
|
xfs_fileoff_t new_endoff; /* end offset of new entry */
|
|
|
|
xfs_exntst_t newext; /* new extent state */
|
|
|
|
xfs_exntst_t oldext; /* old extent state */
|
|
|
|
xfs_bmbt_irec_t r[3]; /* neighbor extent entries */
|
|
|
|
/* left is 0, right is 1, prev is 2 */
|
|
|
|
int rval=0; /* return value (logging flags) */
|
|
|
|
int state = 0;/* state bits, accessed thru macros */
|
2017-02-02 23:14:01 +00:00
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
*logflagsp = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
cur = *curp;
|
2017-02-02 23:14:01 +00:00
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
if (whichfork == XFS_COW_FORK)
|
|
|
|
state |= BMAP_COWFORK;
|
2011-12-18 20:00:07 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(*idx >= 0);
|
2016-11-08 01:59:42 +00:00
|
|
|
ASSERT(*idx <= xfs_iext_count(ifp));
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(!isnullstartblock(new->br_startblock));
|
|
|
|
|
2015-10-12 07:21:22 +00:00
|
|
|
XFS_STATS_INC(mp, xs_add_exlist);
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
#define LEFT r[0]
|
|
|
|
#define RIGHT r[1]
|
|
|
|
#define PREV r[2]
|
2008-10-30 06:14:34 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Set up a bunch of variables to make the tests simpler.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
error = 0;
|
|
|
|
ep = xfs_iext_get_ext(ifp, *idx);
|
|
|
|
xfs_bmbt_get_all(ep, &PREV);
|
|
|
|
newext = new->br_state;
|
|
|
|
oldext = (newext == XFS_EXT_UNWRITTEN) ?
|
|
|
|
XFS_EXT_NORM : XFS_EXT_UNWRITTEN;
|
|
|
|
ASSERT(PREV.br_state == oldext);
|
|
|
|
new_endoff = new->br_startoff + new->br_blockcount;
|
|
|
|
ASSERT(PREV.br_startoff <= new->br_startoff);
|
|
|
|
ASSERT(PREV.br_startoff + PREV.br_blockcount >= new_endoff);
|
2008-10-30 06:14:34 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Set flags determining what part of the previous oldext allocation
|
|
|
|
* extent is being replaced by a newext allocation.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (PREV.br_startoff == new->br_startoff)
|
|
|
|
state |= BMAP_LEFT_FILLING;
|
|
|
|
if (PREV.br_startoff + PREV.br_blockcount == new_endoff)
|
|
|
|
state |= BMAP_RIGHT_FILLING;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Check and set flags if this segment has a left neighbor.
|
|
|
|
* Don't set contiguous if the combined extent would be too large.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (*idx > 0) {
|
|
|
|
state |= BMAP_LEFT_VALID;
|
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(ifp, *idx - 1), &LEFT);
|
|
|
|
|
|
|
|
if (isnullstartblock(LEFT.br_startblock))
|
|
|
|
state |= BMAP_LEFT_DELAY;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
if ((state & BMAP_LEFT_VALID) && !(state & BMAP_LEFT_DELAY) &&
|
|
|
|
LEFT.br_startoff + LEFT.br_blockcount == new->br_startoff &&
|
|
|
|
LEFT.br_startblock + LEFT.br_blockcount == new->br_startblock &&
|
|
|
|
LEFT.br_state == newext &&
|
|
|
|
LEFT.br_blockcount + new->br_blockcount <= MAXEXTLEN)
|
|
|
|
state |= BMAP_LEFT_CONTIG;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Check and set flags if this segment has a right neighbor.
|
|
|
|
* Don't set contiguous if the combined extent would be too large.
|
|
|
|
* Also check for all-three-contiguous being too large.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2017-02-02 23:14:01 +00:00
|
|
|
if (*idx < xfs_iext_count(ifp) - 1) {
|
2013-02-25 01:31:26 +00:00
|
|
|
state |= BMAP_RIGHT_VALID;
|
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(ifp, *idx + 1), &RIGHT);
|
|
|
|
if (isnullstartblock(RIGHT.br_startblock))
|
|
|
|
state |= BMAP_RIGHT_DELAY;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2008-10-30 06:14:34 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((state & BMAP_RIGHT_VALID) && !(state & BMAP_RIGHT_DELAY) &&
|
|
|
|
new_endoff == RIGHT.br_startoff &&
|
|
|
|
new->br_startblock + new->br_blockcount == RIGHT.br_startblock &&
|
|
|
|
newext == RIGHT.br_state &&
|
|
|
|
new->br_blockcount + RIGHT.br_blockcount <= MAXEXTLEN &&
|
|
|
|
((state & (BMAP_LEFT_CONTIG | BMAP_LEFT_FILLING |
|
|
|
|
BMAP_RIGHT_FILLING)) !=
|
|
|
|
(BMAP_LEFT_CONTIG | BMAP_LEFT_FILLING |
|
|
|
|
BMAP_RIGHT_FILLING) ||
|
|
|
|
LEFT.br_blockcount + new->br_blockcount + RIGHT.br_blockcount
|
|
|
|
<= MAXEXTLEN))
|
|
|
|
state |= BMAP_RIGHT_CONTIG;
|
2008-10-30 06:11:40 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Switch out based on the FILLING and CONTIG state bits.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
switch (state & (BMAP_LEFT_FILLING | BMAP_LEFT_CONTIG |
|
|
|
|
BMAP_RIGHT_FILLING | BMAP_RIGHT_CONTIG)) {
|
|
|
|
case BMAP_LEFT_FILLING | BMAP_LEFT_CONTIG |
|
|
|
|
BMAP_RIGHT_FILLING | BMAP_RIGHT_CONTIG:
|
|
|
|
/*
|
|
|
|
* Setting all of a previous oldext extent to newext.
|
|
|
|
* The left and right neighbors are both contiguous with new.
|
|
|
|
*/
|
|
|
|
--*idx;
|
2009-03-29 17:26:46 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, *idx),
|
|
|
|
LEFT.br_blockcount + PREV.br_blockcount +
|
|
|
|
RIGHT.br_blockcount);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2009-03-29 17:26:46 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_iext_remove(ip, *idx + 1, 2, state);
|
2017-02-02 23:14:01 +00:00
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork,
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) - 2);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (cur == NULL)
|
|
|
|
rval = XFS_ILOG_CORE | XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = XFS_ILOG_CORE;
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur, RIGHT.br_startoff,
|
|
|
|
RIGHT.br_startblock,
|
|
|
|
RIGHT.br_blockcount, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_btree_delete(cur, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_btree_decrement(cur, 0, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_btree_delete(cur, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_btree_decrement(cur, 0, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_bmbt_update(cur, LEFT.br_startoff,
|
|
|
|
LEFT.br_startblock,
|
|
|
|
LEFT.br_blockcount + PREV.br_blockcount +
|
|
|
|
RIGHT.br_blockcount, LEFT.br_state)))
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
break;
|
2009-03-29 17:26:46 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_LEFT_FILLING | BMAP_RIGHT_FILLING | BMAP_LEFT_CONTIG:
|
|
|
|
/*
|
|
|
|
* Setting all of a previous oldext extent to newext.
|
|
|
|
* The left neighbor is contiguous, the right is not.
|
|
|
|
*/
|
|
|
|
--*idx;
|
2005-11-01 23:34:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, *idx),
|
|
|
|
LEFT.br_blockcount + PREV.br_blockcount);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_iext_remove(ip, *idx + 1, 1, state);
|
2017-02-02 23:14:01 +00:00
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork,
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) - 1);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (cur == NULL)
|
|
|
|
rval = XFS_ILOG_CORE | XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = XFS_ILOG_CORE;
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur, PREV.br_startoff,
|
|
|
|
PREV.br_startblock, PREV.br_blockcount,
|
|
|
|
&i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_btree_delete(cur, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_btree_decrement(cur, 0, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_bmbt_update(cur, LEFT.br_startoff,
|
|
|
|
LEFT.br_startblock,
|
|
|
|
LEFT.br_blockcount + PREV.br_blockcount,
|
|
|
|
LEFT.br_state)))
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_LEFT_FILLING | BMAP_RIGHT_FILLING | BMAP_RIGHT_CONTIG:
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Setting all of a previous oldext extent to newext.
|
|
|
|
* The right neighbor is contiguous, the left is not.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(ep,
|
|
|
|
PREV.br_blockcount + RIGHT.br_blockcount);
|
|
|
|
xfs_bmbt_set_state(ep, newext);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_iext_remove(ip, *idx + 1, 1, state);
|
2017-02-02 23:14:01 +00:00
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork,
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) - 1);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (cur == NULL)
|
|
|
|
rval = XFS_ILOG_CORE | XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = XFS_ILOG_CORE;
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur, RIGHT.br_startoff,
|
|
|
|
RIGHT.br_startblock,
|
|
|
|
RIGHT.br_blockcount, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_btree_delete(cur, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_btree_decrement(cur, 0, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_bmbt_update(cur, new->br_startoff,
|
|
|
|
new->br_startblock,
|
|
|
|
new->br_blockcount + RIGHT.br_blockcount,
|
|
|
|
newext)))
|
|
|
|
goto done;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
break;
|
2013-02-11 04:58:13 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_LEFT_FILLING | BMAP_RIGHT_FILLING:
|
|
|
|
/*
|
|
|
|
* Setting all of a previous oldext extent to newext.
|
|
|
|
* Neither the left nor right neighbors are contiguous with
|
|
|
|
* the new one.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_state(ep, newext);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2013-02-11 04:58:13 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (cur == NULL)
|
|
|
|
rval = XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = 0;
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur, new->br_startoff,
|
|
|
|
new->br_startblock, new->br_blockcount,
|
|
|
|
&i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_bmbt_update(cur, new->br_startoff,
|
|
|
|
new->br_startblock, new->br_blockcount,
|
|
|
|
newext)))
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
break;
|
2013-02-11 04:58:13 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_LEFT_FILLING | BMAP_LEFT_CONTIG:
|
|
|
|
/*
|
|
|
|
* Setting the first part of a previous oldext extent to newext.
|
|
|
|
* The left neighbor is contiguous.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx - 1, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, *idx - 1),
|
|
|
|
LEFT.br_blockcount + new->br_blockcount);
|
|
|
|
xfs_bmbt_set_startoff(ep,
|
|
|
|
PREV.br_startoff + new->br_blockcount);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx - 1, state, _THIS_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_startblock(ep,
|
|
|
|
new->br_startblock + new->br_blockcount);
|
|
|
|
xfs_bmbt_set_blockcount(ep,
|
|
|
|
PREV.br_blockcount - new->br_blockcount);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2006-03-14 02:30:23 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
--*idx;
|
[XFS] There are a few problems with the new
xfs_bmap_search_multi_extents() wrapper function that I introduced in mod
xfs-linux:xfs-kern:207393a. The function was added as a wrapper around
xfs_bmap_do_search_extents() to avoid breaking the top-of-tree CXFS
interface. The idea of the function was basically to extract the target
extent buffer (if muli- level extent allocation mode), then call
xfs_bmap_do_search_extents() with either a pointer to the first extent in
the target buffer or a pointer to the first extent in the file, depending
on which extent mode was being used. However, in addition to locating the
target extent record for block bno, xfs_bmap_do_search_extents() also sets
four parameters needed by the caller: *lastx, *eofp, *gotp, *prevp.
Passing only the target extent buffer to xfs_bmap_do_search_extents()
causes *eofp to be set incorrectly if the extent is at the end of the
target list but there are actually more extents in the next er_extbuf.
Likewise, if the extent is the first one in the buffer but NOT the first
in the file, *prevp is incorrectly set to NULL. Adding the needed
functionality to xfs_bmap_search_multi_extents() to re-set any incorrectly
set fields is redundant and makes the call to xfs_bmap_do_search_extents()
not make much sense when multi-level extent allocation mode is being used.
This mod basically extracts the two functional components from
xfs_bmap_do_search_extents(), with the intent of obsoleting/removing
xfs_bmap_do_search_extents() after the CXFS mult-level in-core extent
changes are checked in. The two components are: 1) The binary search to
locate the target extent record, and 2) Setting the four parameters needed
by the caller (*lastx, *eofp, *gotp, *prevp). Component 1: I created a
new function in xfs_inode.c called xfs_iext_bno_to_ext(), which executes
the binary search to find the target extent record.
xfs_bmap_search_multi_extents() has been modified to call
xfs_iext_bno_to_ext() rather than xfs_bmap_do_search_extents(). Component
2: The parameter setting functionality has been added to
xfs_bmap_search_multi_extents(), eliminating the need for
xfs_bmap_do_search_extents(). These changes make the removal of
xfs_bmap_do_search_extents() trival once the CXFS changes are in place.
They also allow us to maintain the current XFS interface, using the new
search function introduced in mod xfs-linux:xfs-kern:207393a.
SGI-PV: 928864
SGI-Modid: xfs-linux-melb:xfs-kern:207866a
Signed-off-by: Mandy Kirkconnell <alkirkco@sgi.com>
Signed-off-by: Nathan Scott <nathans@sgi.com>
2006-03-17 06:25:04 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (cur == NULL)
|
|
|
|
rval = XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = 0;
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur, PREV.br_startoff,
|
|
|
|
PREV.br_startblock, PREV.br_blockcount,
|
|
|
|
&i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_bmbt_update(cur,
|
|
|
|
PREV.br_startoff + new->br_blockcount,
|
|
|
|
PREV.br_startblock + new->br_blockcount,
|
|
|
|
PREV.br_blockcount - new->br_blockcount,
|
|
|
|
oldext)))
|
|
|
|
goto done;
|
|
|
|
if ((error = xfs_btree_decrement(cur, 0, &i)))
|
|
|
|
goto done;
|
|
|
|
error = xfs_bmbt_update(cur, LEFT.br_startoff,
|
|
|
|
LEFT.br_startblock,
|
|
|
|
LEFT.br_blockcount + new->br_blockcount,
|
|
|
|
LEFT.br_state);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
[XFS] There are a few problems with the new
xfs_bmap_search_multi_extents() wrapper function that I introduced in mod
xfs-linux:xfs-kern:207393a. The function was added as a wrapper around
xfs_bmap_do_search_extents() to avoid breaking the top-of-tree CXFS
interface. The idea of the function was basically to extract the target
extent buffer (if muli- level extent allocation mode), then call
xfs_bmap_do_search_extents() with either a pointer to the first extent in
the target buffer or a pointer to the first extent in the file, depending
on which extent mode was being used. However, in addition to locating the
target extent record for block bno, xfs_bmap_do_search_extents() also sets
four parameters needed by the caller: *lastx, *eofp, *gotp, *prevp.
Passing only the target extent buffer to xfs_bmap_do_search_extents()
causes *eofp to be set incorrectly if the extent is at the end of the
target list but there are actually more extents in the next er_extbuf.
Likewise, if the extent is the first one in the buffer but NOT the first
in the file, *prevp is incorrectly set to NULL. Adding the needed
functionality to xfs_bmap_search_multi_extents() to re-set any incorrectly
set fields is redundant and makes the call to xfs_bmap_do_search_extents()
not make much sense when multi-level extent allocation mode is being used.
This mod basically extracts the two functional components from
xfs_bmap_do_search_extents(), with the intent of obsoleting/removing
xfs_bmap_do_search_extents() after the CXFS mult-level in-core extent
changes are checked in. The two components are: 1) The binary search to
locate the target extent record, and 2) Setting the four parameters needed
by the caller (*lastx, *eofp, *gotp, *prevp). Component 1: I created a
new function in xfs_inode.c called xfs_iext_bno_to_ext(), which executes
the binary search to find the target extent record.
xfs_bmap_search_multi_extents() has been modified to call
xfs_iext_bno_to_ext() rather than xfs_bmap_do_search_extents(). Component
2: The parameter setting functionality has been added to
xfs_bmap_search_multi_extents(), eliminating the need for
xfs_bmap_do_search_extents(). These changes make the removal of
xfs_bmap_do_search_extents() trival once the CXFS changes are in place.
They also allow us to maintain the current XFS interface, using the new
search function introduced in mod xfs-linux:xfs-kern:207393a.
SGI-PV: 928864
SGI-Modid: xfs-linux-melb:xfs-kern:207866a
Signed-off-by: Mandy Kirkconnell <alkirkco@sgi.com>
Signed-off-by: Nathan Scott <nathans@sgi.com>
2006-03-17 06:25:04 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
break;
|
2006-03-14 02:30:23 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_LEFT_FILLING:
|
|
|
|
/*
|
|
|
|
* Setting the first part of a previous oldext extent to newext.
|
|
|
|
* The left neighbor is not contiguous.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
ASSERT(ep && xfs_bmbt_get_state(ep) == oldext);
|
|
|
|
xfs_bmbt_set_startoff(ep, new_endoff);
|
|
|
|
xfs_bmbt_set_blockcount(ep,
|
|
|
|
PREV.br_blockcount - new->br_blockcount);
|
|
|
|
xfs_bmbt_set_startblock(ep,
|
|
|
|
new->br_startblock + new->br_blockcount);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_iext_insert(ip, *idx, 1, new, state);
|
2017-02-02 23:14:01 +00:00
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork,
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) + 1);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (cur == NULL)
|
|
|
|
rval = XFS_ILOG_CORE | XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = XFS_ILOG_CORE;
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur, PREV.br_startoff,
|
|
|
|
PREV.br_startblock, PREV.br_blockcount,
|
|
|
|
&i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_bmbt_update(cur,
|
|
|
|
PREV.br_startoff + new->br_blockcount,
|
|
|
|
PREV.br_startblock + new->br_blockcount,
|
|
|
|
PREV.br_blockcount - new->br_blockcount,
|
|
|
|
oldext)))
|
|
|
|
goto done;
|
|
|
|
cur->bc_rec.b = *new;
|
|
|
|
if ((error = xfs_btree_insert(cur, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_RIGHT_FILLING | BMAP_RIGHT_CONTIG:
|
|
|
|
/*
|
|
|
|
* Setting the last part of a previous oldext extent to newext.
|
|
|
|
* The right neighbor is contiguous with the new allocation.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(ep,
|
|
|
|
PREV.br_blockcount - new->br_blockcount);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2006-03-14 02:30:23 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
++*idx;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_allf(xfs_iext_get_ext(ifp, *idx),
|
|
|
|
new->br_startoff, new->br_startblock,
|
|
|
|
new->br_blockcount + RIGHT.br_blockcount, newext);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (cur == NULL)
|
|
|
|
rval = XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = 0;
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur, PREV.br_startoff,
|
|
|
|
PREV.br_startblock,
|
|
|
|
PREV.br_blockcount, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_bmbt_update(cur, PREV.br_startoff,
|
|
|
|
PREV.br_startblock,
|
|
|
|
PREV.br_blockcount - new->br_blockcount,
|
|
|
|
oldext)))
|
|
|
|
goto done;
|
|
|
|
if ((error = xfs_btree_increment(cur, 0, &i)))
|
|
|
|
goto done;
|
|
|
|
if ((error = xfs_bmbt_update(cur, new->br_startoff,
|
|
|
|
new->br_startblock,
|
|
|
|
new->br_blockcount + RIGHT.br_blockcount,
|
|
|
|
newext)))
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_RIGHT_FILLING:
|
|
|
|
/*
|
|
|
|
* Setting the last part of a previous oldext extent to newext.
|
|
|
|
* The right neighbor is not contiguous.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(ep,
|
|
|
|
PREV.br_blockcount - new->br_blockcount);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
++*idx;
|
|
|
|
xfs_iext_insert(ip, *idx, 1, new, state);
|
2005-11-01 23:34:53 +00:00
|
|
|
|
2017-02-02 23:14:01 +00:00
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork,
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) + 1);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (cur == NULL)
|
|
|
|
rval = XFS_ILOG_CORE | XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = XFS_ILOG_CORE;
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur, PREV.br_startoff,
|
|
|
|
PREV.br_startblock, PREV.br_blockcount,
|
|
|
|
&i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((error = xfs_bmbt_update(cur, PREV.br_startoff,
|
|
|
|
PREV.br_startblock,
|
|
|
|
PREV.br_blockcount - new->br_blockcount,
|
|
|
|
oldext)))
|
|
|
|
goto done;
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur, new->br_startoff,
|
|
|
|
new->br_startblock, new->br_blockcount,
|
|
|
|
&i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 0, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
cur->bc_rec.b.br_state = XFS_EXT_NORM;
|
|
|
|
if ((error = xfs_btree_insert(cur, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case 0:
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Setting the middle part of a previous oldext extent to
|
|
|
|
* newext. Contiguity is impossible here.
|
|
|
|
* One extent becomes three extents.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(ep,
|
|
|
|
new->br_startoff - PREV.br_startoff);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2010-06-24 01:36:58 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
r[0] = *new;
|
|
|
|
r[1].br_startoff = new_endoff;
|
|
|
|
r[1].br_blockcount =
|
|
|
|
PREV.br_startoff + PREV.br_blockcount - new_endoff;
|
|
|
|
r[1].br_startblock = new->br_startblock + new->br_blockcount;
|
|
|
|
r[1].br_state = oldext;
|
2010-06-24 01:36:58 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
++*idx;
|
|
|
|
xfs_iext_insert(ip, *idx, 2, &r[0], state);
|
|
|
|
|
2017-02-02 23:14:01 +00:00
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork,
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) + 2);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (cur == NULL)
|
|
|
|
rval = XFS_ILOG_CORE | XFS_ILOG_DEXT;
|
|
|
|
else {
|
|
|
|
rval = XFS_ILOG_CORE;
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur, PREV.br_startoff,
|
|
|
|
PREV.br_startblock, PREV.br_blockcount,
|
|
|
|
&i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
/* new right extent - oldext */
|
|
|
|
if ((error = xfs_bmbt_update(cur, r[1].br_startoff,
|
|
|
|
r[1].br_startblock, r[1].br_blockcount,
|
|
|
|
r[1].br_state)))
|
|
|
|
goto done;
|
|
|
|
/* new left extent - oldext */
|
|
|
|
cur->bc_rec.b = PREV;
|
|
|
|
cur->bc_rec.b.br_blockcount =
|
|
|
|
new->br_startoff - PREV.br_startoff;
|
|
|
|
if ((error = xfs_btree_insert(cur, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Reset the cursor to the position of the new extent
|
|
|
|
* we are about to insert as we can't trust it after
|
|
|
|
* the previous insert.
|
|
|
|
*/
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur, new->br_startoff,
|
|
|
|
new->br_startblock, new->br_blockcount,
|
|
|
|
&i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 0, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
/* new middle extent - newext */
|
|
|
|
cur->bc_rec.b.br_state = new->br_state;
|
|
|
|
if ((error = xfs_btree_insert(cur, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
break;
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
case BMAP_LEFT_FILLING | BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG:
|
|
|
|
case BMAP_RIGHT_FILLING | BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG:
|
|
|
|
case BMAP_LEFT_FILLING | BMAP_RIGHT_CONTIG:
|
|
|
|
case BMAP_RIGHT_FILLING | BMAP_LEFT_CONTIG:
|
|
|
|
case BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG:
|
|
|
|
case BMAP_LEFT_CONTIG:
|
|
|
|
case BMAP_RIGHT_CONTIG:
|
|
|
|
/*
|
|
|
|
* These cases are all impossible.
|
|
|
|
*/
|
2005-04-16 22:20:36 +00:00
|
|
|
ASSERT(0);
|
|
|
|
}
|
2011-12-18 20:00:07 +00:00
|
|
|
|
xfs: propagate bmap updates to rmapbt
When we map, unmap, or convert an extent in a file's data or attr
fork, schedule a respective update in the rmapbt. Previous versions
of this patch required a 1:1 correspondence between bmap and rmap,
but this is no longer true as we now have ability to make interval
queries against the rmapbt.
We use the deferred operations code to handle redo operations
atomically and deadlock free. This plumbs in all five rmap actions
(map, unmap, convert extent, alloc, free); we'll use the first three
now for file data, and reflink will want the last two. We also add
an error injection site to test log recovery.
Finally, we need to fix the bmap shift extent code to adjust the
rmaps correctly.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-08-03 02:16:05 +00:00
|
|
|
/* update reverse mappings */
|
2017-02-02 23:14:01 +00:00
|
|
|
error = xfs_rmap_convert_extent(mp, dfops, ip, whichfork, new);
|
xfs: propagate bmap updates to rmapbt
When we map, unmap, or convert an extent in a file's data or attr
fork, schedule a respective update in the rmapbt. Previous versions
of this patch required a 1:1 correspondence between bmap and rmap,
but this is no longer true as we now have ability to make interval
queries against the rmapbt.
We use the deferred operations code to handle redo operations
atomically and deadlock free. This plumbs in all five rmap actions
(map, unmap, convert extent, alloc, free); we'll use the first three
now for file data, and reflink will want the last two. We also add
an error injection site to test log recovery.
Finally, we need to fix the bmap shift extent code to adjust the
rmaps correctly.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-08-03 02:16:05 +00:00
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/* convert to a btree if necessary */
|
2017-02-02 23:14:01 +00:00
|
|
|
if (xfs_bmap_needs_btree(ip, whichfork)) {
|
2013-02-25 01:31:26 +00:00
|
|
|
int tmp_logflags; /* partial log flag return val */
|
|
|
|
|
|
|
|
ASSERT(cur == NULL);
|
2016-08-03 01:19:29 +00:00
|
|
|
error = xfs_bmap_extents_to_btree(tp, ip, first, dfops, &cur,
|
2017-02-02 23:14:01 +00:00
|
|
|
0, &tmp_logflags, whichfork);
|
2013-02-25 01:31:26 +00:00
|
|
|
*logflagsp |= tmp_logflags;
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2005-11-02 04:00:20 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/* clear out the allocated field, done with it now in any case. */
|
|
|
|
if (cur) {
|
|
|
|
cur->bc_private.b.allocated = 0;
|
|
|
|
*curp = cur;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2011-12-18 20:00:07 +00:00
|
|
|
|
2017-02-02 23:14:01 +00:00
|
|
|
xfs_bmap_check_leaf_extents(*curp, ip, whichfork);
|
2013-02-25 01:31:26 +00:00
|
|
|
done:
|
|
|
|
*logflagsp |= rval;
|
2005-04-16 22:20:36 +00:00
|
|
|
return error;
|
2013-02-25 01:31:26 +00:00
|
|
|
#undef LEFT
|
|
|
|
#undef RIGHT
|
|
|
|
#undef PREV
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Convert a hole to a delayed allocation.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
STATIC void
|
|
|
|
xfs_bmap_add_extent_hole_delay(
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
2016-10-03 16:11:32 +00:00
|
|
|
int whichfork,
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_extnum_t *idx, /* extent number to update/insert */
|
|
|
|
xfs_bmbt_irec_t *new) /* new data to add to file extents */
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_ifork_t *ifp; /* inode fork pointer */
|
|
|
|
xfs_bmbt_irec_t left; /* left neighbor extent entry */
|
|
|
|
xfs_filblks_t newlen=0; /* new indirect size */
|
|
|
|
xfs_filblks_t oldlen=0; /* old indirect size */
|
|
|
|
xfs_bmbt_irec_t right; /* right neighbor extent entry */
|
|
|
|
int state; /* state bits, accessed thru macros */
|
|
|
|
xfs_filblks_t temp=0; /* temp for indirect calculations */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-10-03 16:11:32 +00:00
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
2013-02-25 01:31:26 +00:00
|
|
|
state = 0;
|
2016-10-03 16:11:32 +00:00
|
|
|
if (whichfork == XFS_COW_FORK)
|
|
|
|
state |= BMAP_COWFORK;
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(isnullstartblock(new->br_startblock));
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Check and set flags if this segment has a left neighbor
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (*idx > 0) {
|
|
|
|
state |= BMAP_LEFT_VALID;
|
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(ifp, *idx - 1), &left);
|
|
|
|
|
|
|
|
if (isnullstartblock(left.br_startblock))
|
|
|
|
state |= BMAP_LEFT_DELAY;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Check and set flags if the current (right) segment exists.
|
|
|
|
* If it doesn't exist, we're converting the hole at end-of-file.
|
|
|
|
*/
|
2016-11-08 01:59:42 +00:00
|
|
|
if (*idx < xfs_iext_count(ifp)) {
|
2013-02-25 01:31:26 +00:00
|
|
|
state |= BMAP_RIGHT_VALID;
|
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(ifp, *idx), &right);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (isnullstartblock(right.br_startblock))
|
|
|
|
state |= BMAP_RIGHT_DELAY;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Set contiguity flags on the left and right neighbors.
|
|
|
|
* Don't let extents get too large, even if the pieces are contiguous.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((state & BMAP_LEFT_VALID) && (state & BMAP_LEFT_DELAY) &&
|
|
|
|
left.br_startoff + left.br_blockcount == new->br_startoff &&
|
|
|
|
left.br_blockcount + new->br_blockcount <= MAXEXTLEN)
|
|
|
|
state |= BMAP_LEFT_CONTIG;
|
|
|
|
|
|
|
|
if ((state & BMAP_RIGHT_VALID) && (state & BMAP_RIGHT_DELAY) &&
|
|
|
|
new->br_startoff + new->br_blockcount == right.br_startoff &&
|
|
|
|
new->br_blockcount + right.br_blockcount <= MAXEXTLEN &&
|
|
|
|
(!(state & BMAP_LEFT_CONTIG) ||
|
|
|
|
(left.br_blockcount + new->br_blockcount +
|
|
|
|
right.br_blockcount <= MAXEXTLEN)))
|
|
|
|
state |= BMAP_RIGHT_CONTIG;
|
2008-11-17 06:37:10 +00:00
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Switch out based on the contiguity flags.
|
2008-11-17 06:37:10 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
switch (state & (BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG)) {
|
|
|
|
case BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG:
|
|
|
|
/*
|
|
|
|
* New allocation is contiguous with delayed allocations
|
|
|
|
* on the left and on the right.
|
|
|
|
* Merge all three into a single extent record.
|
|
|
|
*/
|
|
|
|
--*idx;
|
|
|
|
temp = left.br_blockcount + new->br_blockcount +
|
|
|
|
right.br_blockcount;
|
2008-11-17 06:37:10 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, *idx), temp);
|
|
|
|
oldlen = startblockval(left.br_startblock) +
|
|
|
|
startblockval(new->br_startblock) +
|
|
|
|
startblockval(right.br_startblock);
|
xfs: handle indlen shortage on delalloc extent merge
When a delalloc extent is created, it can be merged with pre-existing,
contiguous, delalloc extents. When this occurs,
xfs_bmap_add_extent_hole_delay() merges the extents along with the
associated indirect block reservations. The expectation here is that the
combined worst case indlen reservation is always less than or equal to
the indlen reservation for the individual extents.
This is not always the case, however, as existing extents can less than
the expected indlen reservation if the extent was previously split due
to a hole punch. If a new extent merges with such an extent, the total
indlen requirement may be larger than the sum of the indlen reservations
held by both extents.
xfs_bmap_add_extent_hole_delay() assumes that the worst case indlen
reservation is always available and assigns it to the merged extent
without consideration for the indlen held by the pre-existing extent. As
a result, the subsequent xfs_mod_fdblocks() call can attempt an
unintentional allocation rather than a free (indicated by an ASSERT()
failure). Further, if the allocation happens to fail in this context,
the failure goes unhandled and creates a filesystem wide block
accounting inconsistency.
Fix xfs_bmap_add_extent_hole_delay() to function as designed. Cap the
indlen reservation assigned to the merged extent to the sum of the
indlen reservations held by each of the individual extents.
Signed-off-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com>
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
2017-02-14 06:48:18 +00:00
|
|
|
newlen = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp),
|
|
|
|
oldlen);
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmbt_set_startblock(xfs_iext_get_ext(ifp, *idx),
|
|
|
|
nullstartblock((int)newlen));
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_iext_remove(ip, *idx + 1, 1, state);
|
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_LEFT_CONTIG:
|
|
|
|
/*
|
|
|
|
* New allocation is contiguous with a delayed allocation
|
|
|
|
* on the left.
|
|
|
|
* Merge the new allocation with the left neighbor.
|
|
|
|
*/
|
|
|
|
--*idx;
|
|
|
|
temp = left.br_blockcount + new->br_blockcount;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, *idx), temp);
|
|
|
|
oldlen = startblockval(left.br_startblock) +
|
|
|
|
startblockval(new->br_startblock);
|
xfs: handle indlen shortage on delalloc extent merge
When a delalloc extent is created, it can be merged with pre-existing,
contiguous, delalloc extents. When this occurs,
xfs_bmap_add_extent_hole_delay() merges the extents along with the
associated indirect block reservations. The expectation here is that the
combined worst case indlen reservation is always less than or equal to
the indlen reservation for the individual extents.
This is not always the case, however, as existing extents can less than
the expected indlen reservation if the extent was previously split due
to a hole punch. If a new extent merges with such an extent, the total
indlen requirement may be larger than the sum of the indlen reservations
held by both extents.
xfs_bmap_add_extent_hole_delay() assumes that the worst case indlen
reservation is always available and assigns it to the merged extent
without consideration for the indlen held by the pre-existing extent. As
a result, the subsequent xfs_mod_fdblocks() call can attempt an
unintentional allocation rather than a free (indicated by an ASSERT()
failure). Further, if the allocation happens to fail in this context,
the failure goes unhandled and creates a filesystem wide block
accounting inconsistency.
Fix xfs_bmap_add_extent_hole_delay() to function as designed. Cap the
indlen reservation assigned to the merged extent to the sum of the
indlen reservations held by each of the individual extents.
Signed-off-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com>
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
2017-02-14 06:48:18 +00:00
|
|
|
newlen = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp),
|
|
|
|
oldlen);
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmbt_set_startblock(xfs_iext_get_ext(ifp, *idx),
|
|
|
|
nullstartblock((int)newlen));
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_RIGHT_CONTIG:
|
|
|
|
/*
|
|
|
|
* New allocation is contiguous with a delayed allocation
|
|
|
|
* on the right.
|
|
|
|
* Merge the new allocation with the right neighbor.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
temp = new->br_blockcount + right.br_blockcount;
|
|
|
|
oldlen = startblockval(new->br_startblock) +
|
|
|
|
startblockval(right.br_startblock);
|
xfs: handle indlen shortage on delalloc extent merge
When a delalloc extent is created, it can be merged with pre-existing,
contiguous, delalloc extents. When this occurs,
xfs_bmap_add_extent_hole_delay() merges the extents along with the
associated indirect block reservations. The expectation here is that the
combined worst case indlen reservation is always less than or equal to
the indlen reservation for the individual extents.
This is not always the case, however, as existing extents can less than
the expected indlen reservation if the extent was previously split due
to a hole punch. If a new extent merges with such an extent, the total
indlen requirement may be larger than the sum of the indlen reservations
held by both extents.
xfs_bmap_add_extent_hole_delay() assumes that the worst case indlen
reservation is always available and assigns it to the merged extent
without consideration for the indlen held by the pre-existing extent. As
a result, the subsequent xfs_mod_fdblocks() call can attempt an
unintentional allocation rather than a free (indicated by an ASSERT()
failure). Further, if the allocation happens to fail in this context,
the failure goes unhandled and creates a filesystem wide block
accounting inconsistency.
Fix xfs_bmap_add_extent_hole_delay() to function as designed. Cap the
indlen reservation assigned to the merged extent to the sum of the
indlen reservations held by each of the individual extents.
Signed-off-by: Brian Foster <bfoster@redhat.com>
Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com>
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
2017-02-14 06:48:18 +00:00
|
|
|
newlen = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp),
|
|
|
|
oldlen);
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmbt_set_allf(xfs_iext_get_ext(ifp, *idx),
|
|
|
|
new->br_startoff,
|
|
|
|
nullstartblock((int)newlen), temp, right.br_state);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case 0:
|
|
|
|
/*
|
|
|
|
* New allocation is not contiguous with another
|
|
|
|
* delayed allocation.
|
|
|
|
* Insert a new entry.
|
|
|
|
*/
|
|
|
|
oldlen = newlen = 0;
|
|
|
|
xfs_iext_insert(ip, *idx, 1, new, state);
|
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
if (oldlen != newlen) {
|
|
|
|
ASSERT(oldlen > newlen);
|
2015-02-23 10:22:03 +00:00
|
|
|
xfs_mod_fdblocks(ip->i_mount, (int64_t)(oldlen - newlen),
|
|
|
|
false);
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Nothing to do for disk quota accounting here.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Convert a hole to a real allocation.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
STATIC int /* error */
|
|
|
|
xfs_bmap_add_extent_hole_real(
|
2017-04-11 23:45:54 +00:00
|
|
|
struct xfs_trans *tp,
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
int whichfork,
|
|
|
|
xfs_extnum_t *idx,
|
|
|
|
struct xfs_btree_cur **curp,
|
|
|
|
struct xfs_bmbt_irec *new,
|
|
|
|
xfs_fsblock_t *first,
|
|
|
|
struct xfs_defer_ops *dfops,
|
|
|
|
int *logflagsp)
|
2011-09-18 20:40:53 +00:00
|
|
|
{
|
2017-04-11 23:45:54 +00:00
|
|
|
struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
|
|
|
struct xfs_btree_cur *cur = *curp;
|
2013-02-25 01:31:26 +00:00
|
|
|
int error; /* error return value */
|
|
|
|
int i; /* temp state */
|
|
|
|
xfs_bmbt_irec_t left; /* left neighbor extent entry */
|
|
|
|
xfs_bmbt_irec_t right; /* right neighbor extent entry */
|
|
|
|
int rval=0; /* return value (logging flags) */
|
|
|
|
int state; /* state bits, accessed thru macros */
|
2011-09-18 20:40:53 +00:00
|
|
|
|
2017-04-11 23:45:54 +00:00
|
|
|
ASSERT(*idx >= 0);
|
|
|
|
ASSERT(*idx <= xfs_iext_count(ifp));
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(!isnullstartblock(new->br_startblock));
|
2017-04-11 23:45:54 +00:00
|
|
|
ASSERT(!cur || !(cur->bc_private.b.flags & XFS_BTCUR_BPRV_WASDEL));
|
2011-09-18 20:40:53 +00:00
|
|
|
|
2015-10-12 07:21:22 +00:00
|
|
|
XFS_STATS_INC(mp, xs_add_exlist);
|
2011-09-18 20:40:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
state = 0;
|
|
|
|
if (whichfork == XFS_ATTR_FORK)
|
|
|
|
state |= BMAP_ATTRFORK;
|
2017-02-06 18:50:49 +00:00
|
|
|
if (whichfork == XFS_COW_FORK)
|
|
|
|
state |= BMAP_COWFORK;
|
2011-09-18 20:40:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Check and set flags if this segment has a left neighbor.
|
|
|
|
*/
|
2017-04-11 23:45:54 +00:00
|
|
|
if (*idx > 0) {
|
2013-02-25 01:31:26 +00:00
|
|
|
state |= BMAP_LEFT_VALID;
|
2017-04-11 23:45:54 +00:00
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(ifp, *idx - 1), &left);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (isnullstartblock(left.br_startblock))
|
|
|
|
state |= BMAP_LEFT_DELAY;
|
|
|
|
}
|
2011-09-18 20:40:53 +00:00
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Check and set flags if this segment has a current value.
|
|
|
|
* Not true if we're inserting into the "hole" at eof.
|
2011-09-18 20:40:53 +00:00
|
|
|
*/
|
2017-04-11 23:45:54 +00:00
|
|
|
if (*idx < xfs_iext_count(ifp)) {
|
2013-02-25 01:31:26 +00:00
|
|
|
state |= BMAP_RIGHT_VALID;
|
2017-04-11 23:45:54 +00:00
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(ifp, *idx), &right);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (isnullstartblock(right.br_startblock))
|
|
|
|
state |= BMAP_RIGHT_DELAY;
|
|
|
|
}
|
2011-09-18 20:40:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* We're inserting a real allocation between "left" and "right".
|
|
|
|
* Set the contiguity flags. Don't let extents get too large.
|
|
|
|
*/
|
|
|
|
if ((state & BMAP_LEFT_VALID) && !(state & BMAP_LEFT_DELAY) &&
|
|
|
|
left.br_startoff + left.br_blockcount == new->br_startoff &&
|
|
|
|
left.br_startblock + left.br_blockcount == new->br_startblock &&
|
|
|
|
left.br_state == new->br_state &&
|
|
|
|
left.br_blockcount + new->br_blockcount <= MAXEXTLEN)
|
|
|
|
state |= BMAP_LEFT_CONTIG;
|
2011-09-18 20:40:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((state & BMAP_RIGHT_VALID) && !(state & BMAP_RIGHT_DELAY) &&
|
|
|
|
new->br_startoff + new->br_blockcount == right.br_startoff &&
|
|
|
|
new->br_startblock + new->br_blockcount == right.br_startblock &&
|
|
|
|
new->br_state == right.br_state &&
|
|
|
|
new->br_blockcount + right.br_blockcount <= MAXEXTLEN &&
|
|
|
|
(!(state & BMAP_LEFT_CONTIG) ||
|
|
|
|
left.br_blockcount + new->br_blockcount +
|
|
|
|
right.br_blockcount <= MAXEXTLEN))
|
|
|
|
state |= BMAP_RIGHT_CONTIG;
|
2011-09-18 20:40:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
error = 0;
|
|
|
|
/*
|
|
|
|
* Select which case we're in here, and implement it.
|
|
|
|
*/
|
|
|
|
switch (state & (BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG)) {
|
|
|
|
case BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG:
|
|
|
|
/*
|
|
|
|
* New allocation is contiguous with real allocations on the
|
|
|
|
* left and on the right.
|
|
|
|
* Merge all three into a single extent record.
|
|
|
|
*/
|
2017-04-11 23:45:54 +00:00
|
|
|
--*idx;
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, *idx),
|
2013-02-25 01:31:26 +00:00
|
|
|
left.br_blockcount + new->br_blockcount +
|
|
|
|
right.br_blockcount);
|
2017-04-11 23:45:54 +00:00
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2011-09-18 20:40:53 +00:00
|
|
|
|
2017-04-11 23:45:54 +00:00
|
|
|
xfs_iext_remove(ip, *idx + 1, 1, state);
|
2011-09-18 20:40:53 +00:00
|
|
|
|
2017-04-11 23:45:54 +00:00
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork,
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) - 1);
|
|
|
|
if (cur == NULL) {
|
2013-02-25 01:31:26 +00:00
|
|
|
rval = XFS_ILOG_CORE | xfs_ilog_fext(whichfork);
|
|
|
|
} else {
|
|
|
|
rval = XFS_ILOG_CORE;
|
2017-04-11 23:45:54 +00:00
|
|
|
error = xfs_bmbt_lookup_eq(cur, right.br_startoff,
|
2013-02-25 01:31:26 +00:00
|
|
|
right.br_startblock, right.br_blockcount,
|
|
|
|
&i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2017-04-11 23:45:54 +00:00
|
|
|
error = xfs_btree_delete(cur, &i);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2017-04-11 23:45:54 +00:00
|
|
|
error = xfs_btree_decrement(cur, 0, &i);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2017-04-11 23:45:54 +00:00
|
|
|
error = xfs_bmbt_update(cur, left.br_startoff,
|
2013-02-25 01:31:26 +00:00
|
|
|
left.br_startblock,
|
|
|
|
left.br_blockcount +
|
|
|
|
new->br_blockcount +
|
|
|
|
right.br_blockcount,
|
|
|
|
left.br_state);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
break;
|
2011-09-18 20:40:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_LEFT_CONTIG:
|
|
|
|
/*
|
|
|
|
* New allocation is contiguous with a real allocation
|
|
|
|
* on the left.
|
|
|
|
* Merge the new allocation with the left neighbor.
|
|
|
|
*/
|
2017-04-11 23:45:54 +00:00
|
|
|
--*idx;
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(xfs_iext_get_ext(ifp, *idx),
|
2013-02-25 01:31:26 +00:00
|
|
|
left.br_blockcount + new->br_blockcount);
|
2017-04-11 23:45:54 +00:00
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2017-04-11 23:45:54 +00:00
|
|
|
if (cur == NULL) {
|
2013-02-25 01:31:26 +00:00
|
|
|
rval = xfs_ilog_fext(whichfork);
|
|
|
|
} else {
|
|
|
|
rval = 0;
|
2017-04-11 23:45:54 +00:00
|
|
|
error = xfs_bmbt_lookup_eq(cur, left.br_startoff,
|
2013-02-25 01:31:26 +00:00
|
|
|
left.br_startblock, left.br_blockcount,
|
|
|
|
&i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2017-04-11 23:45:54 +00:00
|
|
|
error = xfs_bmbt_update(cur, left.br_startoff,
|
2013-02-25 01:31:26 +00:00
|
|
|
left.br_startblock,
|
|
|
|
left.br_blockcount +
|
|
|
|
new->br_blockcount,
|
|
|
|
left.br_state);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
break;
|
2011-09-18 20:40:53 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
case BMAP_RIGHT_CONTIG:
|
|
|
|
/*
|
|
|
|
* New allocation is contiguous with a real allocation
|
|
|
|
* on the right.
|
|
|
|
* Merge the new allocation with the right neighbor.
|
|
|
|
*/
|
2017-04-11 23:45:54 +00:00
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_allf(xfs_iext_get_ext(ifp, *idx),
|
2013-02-25 01:31:26 +00:00
|
|
|
new->br_startoff, new->br_startblock,
|
|
|
|
new->br_blockcount + right.br_blockcount,
|
|
|
|
right.br_state);
|
2017-04-11 23:45:54 +00:00
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
2011-09-18 20:40:53 +00:00
|
|
|
|
2017-04-11 23:45:54 +00:00
|
|
|
if (cur == NULL) {
|
2013-02-25 01:31:26 +00:00
|
|
|
rval = xfs_ilog_fext(whichfork);
|
|
|
|
} else {
|
|
|
|
rval = 0;
|
2017-04-11 23:45:54 +00:00
|
|
|
error = xfs_bmbt_lookup_eq(cur,
|
2013-02-25 01:31:26 +00:00
|
|
|
right.br_startoff,
|
|
|
|
right.br_startblock,
|
|
|
|
right.br_blockcount, &i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2017-04-11 23:45:54 +00:00
|
|
|
error = xfs_bmbt_update(cur, new->br_startoff,
|
2013-02-25 01:31:26 +00:00
|
|
|
new->br_startblock,
|
|
|
|
new->br_blockcount +
|
|
|
|
right.br_blockcount,
|
|
|
|
right.br_state);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
case 0:
|
|
|
|
/*
|
|
|
|
* New allocation is not contiguous with another
|
|
|
|
* real allocation.
|
|
|
|
* Insert a new entry.
|
|
|
|
*/
|
2017-04-11 23:45:54 +00:00
|
|
|
xfs_iext_insert(ip, *idx, 1, new, state);
|
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork,
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) + 1);
|
|
|
|
if (cur == NULL) {
|
2013-02-25 01:31:26 +00:00
|
|
|
rval = XFS_ILOG_CORE | xfs_ilog_fext(whichfork);
|
|
|
|
} else {
|
|
|
|
rval = XFS_ILOG_CORE;
|
2017-04-11 23:45:54 +00:00
|
|
|
error = xfs_bmbt_lookup_eq(cur,
|
2013-02-25 01:31:26 +00:00
|
|
|
new->br_startoff,
|
|
|
|
new->br_startblock,
|
|
|
|
new->br_blockcount, &i);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 0, done);
|
2017-04-11 23:45:54 +00:00
|
|
|
cur->bc_rec.b.br_state = new->br_state;
|
|
|
|
error = xfs_btree_insert(cur, &i);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (error)
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
xfs: propagate bmap updates to rmapbt
When we map, unmap, or convert an extent in a file's data or attr
fork, schedule a respective update in the rmapbt. Previous versions
of this patch required a 1:1 correspondence between bmap and rmap,
but this is no longer true as we now have ability to make interval
queries against the rmapbt.
We use the deferred operations code to handle redo operations
atomically and deadlock free. This plumbs in all five rmap actions
(map, unmap, convert extent, alloc, free); we'll use the first three
now for file data, and reflink will want the last two. We also add
an error injection site to test log recovery.
Finally, we need to fix the bmap shift extent code to adjust the
rmaps correctly.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-08-03 02:16:05 +00:00
|
|
|
/* add reverse mapping */
|
2017-04-11 23:45:54 +00:00
|
|
|
error = xfs_rmap_map_extent(mp, dfops, ip, whichfork, new);
|
xfs: propagate bmap updates to rmapbt
When we map, unmap, or convert an extent in a file's data or attr
fork, schedule a respective update in the rmapbt. Previous versions
of this patch required a 1:1 correspondence between bmap and rmap,
but this is no longer true as we now have ability to make interval
queries against the rmapbt.
We use the deferred operations code to handle redo operations
atomically and deadlock free. This plumbs in all five rmap actions
(map, unmap, convert extent, alloc, free); we'll use the first three
now for file data, and reflink will want the last two. We also add
an error injection site to test log recovery.
Finally, we need to fix the bmap shift extent code to adjust the
rmaps correctly.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-08-03 02:16:05 +00:00
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/* convert to a btree if necessary */
|
2017-04-11 23:45:54 +00:00
|
|
|
if (xfs_bmap_needs_btree(ip, whichfork)) {
|
2013-02-25 01:31:26 +00:00
|
|
|
int tmp_logflags; /* partial log flag return val */
|
|
|
|
|
2017-04-11 23:45:54 +00:00
|
|
|
ASSERT(cur == NULL);
|
|
|
|
error = xfs_bmap_extents_to_btree(tp, ip, first, dfops, curp,
|
2013-02-25 01:31:26 +00:00
|
|
|
0, &tmp_logflags, whichfork);
|
2017-04-11 23:45:54 +00:00
|
|
|
*logflagsp |= tmp_logflags;
|
|
|
|
cur = *curp;
|
2013-02-25 01:31:26 +00:00
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* clear out the allocated field, done with it now in any case. */
|
2017-04-11 23:45:54 +00:00
|
|
|
if (cur)
|
|
|
|
cur->bc_private.b.allocated = 0;
|
2013-02-25 01:31:26 +00:00
|
|
|
|
2017-04-11 23:45:54 +00:00
|
|
|
xfs_bmap_check_leaf_extents(cur, ip, whichfork);
|
2013-02-25 01:31:26 +00:00
|
|
|
done:
|
2017-04-11 23:45:54 +00:00
|
|
|
*logflagsp |= rval;
|
2013-02-25 01:31:26 +00:00
|
|
|
return error;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Functions used in the extent read, allocate and remove paths
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Adjust the size of the new extent based on di_extsize and rt extsize.
|
|
|
|
*/
|
2013-08-12 10:49:42 +00:00
|
|
|
int
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmap_extsize_align(
|
|
|
|
xfs_mount_t *mp,
|
|
|
|
xfs_bmbt_irec_t *gotp, /* next extent pointer */
|
|
|
|
xfs_bmbt_irec_t *prevp, /* previous extent pointer */
|
|
|
|
xfs_extlen_t extsz, /* align to this extent size */
|
|
|
|
int rt, /* is this a realtime inode? */
|
|
|
|
int eof, /* is extent at end-of-file? */
|
|
|
|
int delay, /* creating delalloc extent? */
|
|
|
|
int convert, /* overwriting unwritten extent? */
|
|
|
|
xfs_fileoff_t *offp, /* in/out: aligned offset */
|
|
|
|
xfs_extlen_t *lenp) /* in/out: aligned length */
|
2008-10-30 06:14:43 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_fileoff_t orig_off; /* original offset */
|
|
|
|
xfs_extlen_t orig_alen; /* original length */
|
|
|
|
xfs_fileoff_t orig_end; /* original off+len */
|
|
|
|
xfs_fileoff_t nexto; /* next file offset */
|
|
|
|
xfs_fileoff_t prevo; /* previous file offset */
|
|
|
|
xfs_fileoff_t align_off; /* temp for offset */
|
|
|
|
xfs_extlen_t align_alen; /* temp for length */
|
|
|
|
xfs_extlen_t temp; /* temp for calculations */
|
2008-10-30 06:14:43 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (convert)
|
2008-10-30 06:14:43 +00:00
|
|
|
return 0;
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
orig_off = align_off = *offp;
|
|
|
|
orig_alen = align_alen = *lenp;
|
|
|
|
orig_end = orig_off + orig_alen;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* If this request overlaps an existing extent, then don't
|
|
|
|
* attempt to perform any additional alignment.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (!delay && !eof &&
|
|
|
|
(orig_off >= gotp->br_startoff) &&
|
|
|
|
(orig_end <= gotp->br_startoff + gotp->br_blockcount)) {
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* If the file offset is unaligned vs. the extent size
|
|
|
|
* we need to align it. This will be possible unless
|
|
|
|
* the file was previously written with a kernel that didn't
|
|
|
|
* perform this alignment, or if a truncate shot us in the
|
|
|
|
* foot.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
temp = do_mod(orig_off, extsz);
|
|
|
|
if (temp) {
|
|
|
|
align_alen += temp;
|
|
|
|
align_off -= temp;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
xfs: extent size hints can round up extents past MAXEXTLEN
This results in BMBT corruption, as seen by this test:
# mkfs.xfs -f -d size=40051712b,agcount=4 /dev/vdc
....
# mount /dev/vdc /mnt/scratch
# xfs_io -ft -c "extsize 16m" -c "falloc 0 30g" -c "bmap -vp" /mnt/scratch/foo
which results in this failure on a debug kernel:
XFS: Assertion failed: (blockcount & xfs_mask64hi(64-BMBT_BLOCKCOUNT_BITLEN)) == 0, file: fs/xfs/libxfs/xfs_bmap_btree.c, line: 211
....
Call Trace:
[<ffffffff814cf0ff>] xfs_bmbt_set_allf+0x8f/0x100
[<ffffffff814cf18d>] xfs_bmbt_set_all+0x1d/0x20
[<ffffffff814f2efe>] xfs_iext_insert+0x9e/0x120
[<ffffffff814c7956>] ? xfs_bmap_add_extent_hole_real+0x1c6/0xc70
[<ffffffff814c7956>] xfs_bmap_add_extent_hole_real+0x1c6/0xc70
[<ffffffff814caaab>] xfs_bmapi_write+0x72b/0xed0
[<ffffffff811c72ac>] ? kmem_cache_alloc+0x15c/0x170
[<ffffffff814fe070>] xfs_alloc_file_space+0x160/0x400
[<ffffffff81ddcc29>] ? down_write+0x29/0x60
[<ffffffff815063eb>] xfs_file_fallocate+0x29b/0x310
[<ffffffff811d2bc8>] ? __sb_start_write+0x58/0x120
[<ffffffff811e3e18>] ? do_vfs_ioctl+0x318/0x570
[<ffffffff811cd680>] vfs_fallocate+0x140/0x260
[<ffffffff811ce6f8>] SyS_fallocate+0x48/0x80
[<ffffffff81ddec09>] system_call_fastpath+0x12/0x17
The tracepoint that indicates the extent that triggered the assert
failure is:
xfs_iext_insert: idx 0 offset 0 block 16777224 count 2097152 flag 1
Clearly indicating that the extent length is greater than MAXEXTLEN,
which is 2097151. A prior trace point shows the allocation was an
exact size match and that a length greater than MAXEXTLEN was asked
for:
xfs_alloc_size_done: agno 1 agbno 8 minlen 2097152 maxlen 2097152
^^^^^^^ ^^^^^^^
We don't see this problem with extent size hints through the IO path
because we can't do single IOs large enough to trigger MAXEXTLEN
allocation. fallocate(), OTOH, is not limited in it's allocation
sizes and so needs help here.
The issue is that the extent size hint alignment is rounding up the
extent size past MAXEXTLEN, because xfs_bmapi_write() is not taking
into account extent size hints when calculating the maximum extent
length to allocate. xfs_bmapi_reserve_delalloc() is already doing
this, but direct extent allocation is not.
Unfortunately, the calculation in xfs_bmapi_reserve_delalloc() is
wrong, and it works only because delayed allocation extents are not
limited in size to MAXEXTLEN in the in-core extent tree. hence this
calculation does not work for direct allocation, and the delalloc
code needs fixing. This may, in fact be the underlying bug that
occassionally causes transaction overruns in delayed allocation
extent conversion, so now we know it's wrong we should fix it, too.
Many thanks to Brian Foster for finding this problem during review
of this patch.
Hence the fix, after much code reading, is to allow
xfs_bmap_extsize_align() to align partial extents when full
alignment would extend the alignment past MAXEXTLEN. We can safely
do this because all callers have higher layer allocation loops that
already handle short allocations, and so will simply run another
allocation to cover the remainder of the requested allocation range
that we ignored during alignment. The advantage of this approach is
that it also removes the need for callers to do anything other than
limit their requests to MAXEXTLEN - they don't really need to be
aware of extent size hints at all.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2015-05-28 21:40:06 +00:00
|
|
|
|
|
|
|
/* Same adjustment for the end of the requested area. */
|
|
|
|
temp = (align_alen % extsz);
|
|
|
|
if (temp)
|
|
|
|
align_alen += extsz - temp;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
xfs: extent size hints can round up extents past MAXEXTLEN
This results in BMBT corruption, as seen by this test:
# mkfs.xfs -f -d size=40051712b,agcount=4 /dev/vdc
....
# mount /dev/vdc /mnt/scratch
# xfs_io -ft -c "extsize 16m" -c "falloc 0 30g" -c "bmap -vp" /mnt/scratch/foo
which results in this failure on a debug kernel:
XFS: Assertion failed: (blockcount & xfs_mask64hi(64-BMBT_BLOCKCOUNT_BITLEN)) == 0, file: fs/xfs/libxfs/xfs_bmap_btree.c, line: 211
....
Call Trace:
[<ffffffff814cf0ff>] xfs_bmbt_set_allf+0x8f/0x100
[<ffffffff814cf18d>] xfs_bmbt_set_all+0x1d/0x20
[<ffffffff814f2efe>] xfs_iext_insert+0x9e/0x120
[<ffffffff814c7956>] ? xfs_bmap_add_extent_hole_real+0x1c6/0xc70
[<ffffffff814c7956>] xfs_bmap_add_extent_hole_real+0x1c6/0xc70
[<ffffffff814caaab>] xfs_bmapi_write+0x72b/0xed0
[<ffffffff811c72ac>] ? kmem_cache_alloc+0x15c/0x170
[<ffffffff814fe070>] xfs_alloc_file_space+0x160/0x400
[<ffffffff81ddcc29>] ? down_write+0x29/0x60
[<ffffffff815063eb>] xfs_file_fallocate+0x29b/0x310
[<ffffffff811d2bc8>] ? __sb_start_write+0x58/0x120
[<ffffffff811e3e18>] ? do_vfs_ioctl+0x318/0x570
[<ffffffff811cd680>] vfs_fallocate+0x140/0x260
[<ffffffff811ce6f8>] SyS_fallocate+0x48/0x80
[<ffffffff81ddec09>] system_call_fastpath+0x12/0x17
The tracepoint that indicates the extent that triggered the assert
failure is:
xfs_iext_insert: idx 0 offset 0 block 16777224 count 2097152 flag 1
Clearly indicating that the extent length is greater than MAXEXTLEN,
which is 2097151. A prior trace point shows the allocation was an
exact size match and that a length greater than MAXEXTLEN was asked
for:
xfs_alloc_size_done: agno 1 agbno 8 minlen 2097152 maxlen 2097152
^^^^^^^ ^^^^^^^
We don't see this problem with extent size hints through the IO path
because we can't do single IOs large enough to trigger MAXEXTLEN
allocation. fallocate(), OTOH, is not limited in it's allocation
sizes and so needs help here.
The issue is that the extent size hint alignment is rounding up the
extent size past MAXEXTLEN, because xfs_bmapi_write() is not taking
into account extent size hints when calculating the maximum extent
length to allocate. xfs_bmapi_reserve_delalloc() is already doing
this, but direct extent allocation is not.
Unfortunately, the calculation in xfs_bmapi_reserve_delalloc() is
wrong, and it works only because delayed allocation extents are not
limited in size to MAXEXTLEN in the in-core extent tree. hence this
calculation does not work for direct allocation, and the delalloc
code needs fixing. This may, in fact be the underlying bug that
occassionally causes transaction overruns in delayed allocation
extent conversion, so now we know it's wrong we should fix it, too.
Many thanks to Brian Foster for finding this problem during review
of this patch.
Hence the fix, after much code reading, is to allow
xfs_bmap_extsize_align() to align partial extents when full
alignment would extend the alignment past MAXEXTLEN. We can safely
do this because all callers have higher layer allocation loops that
already handle short allocations, and so will simply run another
allocation to cover the remainder of the requested allocation range
that we ignored during alignment. The advantage of this approach is
that it also removes the need for callers to do anything other than
limit their requests to MAXEXTLEN - they don't really need to be
aware of extent size hints at all.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2015-05-28 21:40:06 +00:00
|
|
|
* For large extent hint sizes, the aligned extent might be larger than
|
|
|
|
* MAXEXTLEN. In that case, reduce the size by an extsz so that it pulls
|
|
|
|
* the length back under MAXEXTLEN. The outer allocation loops handle
|
|
|
|
* short allocation just fine, so it is safe to do this. We only want to
|
|
|
|
* do it when we are forced to, though, because it means more allocation
|
|
|
|
* operations are required.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
xfs: extent size hints can round up extents past MAXEXTLEN
This results in BMBT corruption, as seen by this test:
# mkfs.xfs -f -d size=40051712b,agcount=4 /dev/vdc
....
# mount /dev/vdc /mnt/scratch
# xfs_io -ft -c "extsize 16m" -c "falloc 0 30g" -c "bmap -vp" /mnt/scratch/foo
which results in this failure on a debug kernel:
XFS: Assertion failed: (blockcount & xfs_mask64hi(64-BMBT_BLOCKCOUNT_BITLEN)) == 0, file: fs/xfs/libxfs/xfs_bmap_btree.c, line: 211
....
Call Trace:
[<ffffffff814cf0ff>] xfs_bmbt_set_allf+0x8f/0x100
[<ffffffff814cf18d>] xfs_bmbt_set_all+0x1d/0x20
[<ffffffff814f2efe>] xfs_iext_insert+0x9e/0x120
[<ffffffff814c7956>] ? xfs_bmap_add_extent_hole_real+0x1c6/0xc70
[<ffffffff814c7956>] xfs_bmap_add_extent_hole_real+0x1c6/0xc70
[<ffffffff814caaab>] xfs_bmapi_write+0x72b/0xed0
[<ffffffff811c72ac>] ? kmem_cache_alloc+0x15c/0x170
[<ffffffff814fe070>] xfs_alloc_file_space+0x160/0x400
[<ffffffff81ddcc29>] ? down_write+0x29/0x60
[<ffffffff815063eb>] xfs_file_fallocate+0x29b/0x310
[<ffffffff811d2bc8>] ? __sb_start_write+0x58/0x120
[<ffffffff811e3e18>] ? do_vfs_ioctl+0x318/0x570
[<ffffffff811cd680>] vfs_fallocate+0x140/0x260
[<ffffffff811ce6f8>] SyS_fallocate+0x48/0x80
[<ffffffff81ddec09>] system_call_fastpath+0x12/0x17
The tracepoint that indicates the extent that triggered the assert
failure is:
xfs_iext_insert: idx 0 offset 0 block 16777224 count 2097152 flag 1
Clearly indicating that the extent length is greater than MAXEXTLEN,
which is 2097151. A prior trace point shows the allocation was an
exact size match and that a length greater than MAXEXTLEN was asked
for:
xfs_alloc_size_done: agno 1 agbno 8 minlen 2097152 maxlen 2097152
^^^^^^^ ^^^^^^^
We don't see this problem with extent size hints through the IO path
because we can't do single IOs large enough to trigger MAXEXTLEN
allocation. fallocate(), OTOH, is not limited in it's allocation
sizes and so needs help here.
The issue is that the extent size hint alignment is rounding up the
extent size past MAXEXTLEN, because xfs_bmapi_write() is not taking
into account extent size hints when calculating the maximum extent
length to allocate. xfs_bmapi_reserve_delalloc() is already doing
this, but direct extent allocation is not.
Unfortunately, the calculation in xfs_bmapi_reserve_delalloc() is
wrong, and it works only because delayed allocation extents are not
limited in size to MAXEXTLEN in the in-core extent tree. hence this
calculation does not work for direct allocation, and the delalloc
code needs fixing. This may, in fact be the underlying bug that
occassionally causes transaction overruns in delayed allocation
extent conversion, so now we know it's wrong we should fix it, too.
Many thanks to Brian Foster for finding this problem during review
of this patch.
Hence the fix, after much code reading, is to allow
xfs_bmap_extsize_align() to align partial extents when full
alignment would extend the alignment past MAXEXTLEN. We can safely
do this because all callers have higher layer allocation loops that
already handle short allocations, and so will simply run another
allocation to cover the remainder of the requested allocation range
that we ignored during alignment. The advantage of this approach is
that it also removes the need for callers to do anything other than
limit their requests to MAXEXTLEN - they don't really need to be
aware of extent size hints at all.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2015-05-28 21:40:06 +00:00
|
|
|
while (align_alen > MAXEXTLEN)
|
|
|
|
align_alen -= extsz;
|
|
|
|
ASSERT(align_alen <= MAXEXTLEN);
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* If the previous block overlaps with this proposed allocation
|
|
|
|
* then move the start forward without adjusting the length.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (prevp->br_startoff != NULLFILEOFF) {
|
|
|
|
if (prevp->br_startblock == HOLESTARTBLOCK)
|
|
|
|
prevo = prevp->br_startoff;
|
|
|
|
else
|
|
|
|
prevo = prevp->br_startoff + prevp->br_blockcount;
|
|
|
|
} else
|
|
|
|
prevo = 0;
|
|
|
|
if (align_off != orig_off && align_off < prevo)
|
|
|
|
align_off = prevo;
|
|
|
|
/*
|
|
|
|
* If the next block overlaps with this proposed allocation
|
|
|
|
* then move the start back without adjusting the length,
|
|
|
|
* but not before offset 0.
|
|
|
|
* This may of course make the start overlap previous block,
|
|
|
|
* and if we hit the offset 0 limit then the next block
|
|
|
|
* can still overlap too.
|
|
|
|
*/
|
|
|
|
if (!eof && gotp->br_startoff != NULLFILEOFF) {
|
|
|
|
if ((delay && gotp->br_startblock == HOLESTARTBLOCK) ||
|
|
|
|
(!delay && gotp->br_startblock == DELAYSTARTBLOCK))
|
|
|
|
nexto = gotp->br_startoff + gotp->br_blockcount;
|
|
|
|
else
|
|
|
|
nexto = gotp->br_startoff;
|
|
|
|
} else
|
|
|
|
nexto = NULLFILEOFF;
|
|
|
|
if (!eof &&
|
|
|
|
align_off + align_alen != orig_end &&
|
|
|
|
align_off + align_alen > nexto)
|
|
|
|
align_off = nexto > align_alen ? nexto - align_alen : 0;
|
|
|
|
/*
|
|
|
|
* If we're now overlapping the next or previous extent that
|
|
|
|
* means we can't fit an extsz piece in this hole. Just move
|
|
|
|
* the start forward to the first valid spot and set
|
|
|
|
* the length so we hit the end.
|
|
|
|
*/
|
|
|
|
if (align_off != orig_off && align_off < prevo)
|
|
|
|
align_off = prevo;
|
|
|
|
if (align_off + align_alen != orig_end &&
|
|
|
|
align_off + align_alen > nexto &&
|
|
|
|
nexto != NULLFILEOFF) {
|
|
|
|
ASSERT(nexto > prevo);
|
|
|
|
align_alen = nexto - align_off;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* If realtime, and the result isn't a multiple of the realtime
|
|
|
|
* extent size we need to remove blocks until it is.
|
|
|
|
*/
|
|
|
|
if (rt && (temp = (align_alen % mp->m_sb.sb_rextsize))) {
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* We're not covering the original request, or
|
|
|
|
* we won't be able to once we fix the length.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (orig_off < align_off ||
|
|
|
|
orig_end > align_off + align_alen ||
|
|
|
|
align_alen - temp < orig_alen)
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EINVAL;
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Try to fix it by moving the start up.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (align_off + temp <= orig_off) {
|
|
|
|
align_alen -= temp;
|
|
|
|
align_off += temp;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Try to fix it by moving the end in.
|
|
|
|
*/
|
|
|
|
else if (align_off + align_alen - temp >= orig_end)
|
|
|
|
align_alen -= temp;
|
|
|
|
/*
|
|
|
|
* Set the start to the minimum then trim the length.
|
|
|
|
*/
|
|
|
|
else {
|
|
|
|
align_alen -= orig_off - align_off;
|
|
|
|
align_off = orig_off;
|
|
|
|
align_alen -= align_alen % mp->m_sb.sb_rextsize;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Result doesn't cover the request, fail it.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (orig_off < align_off || orig_end > align_off + align_alen)
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EINVAL;
|
2013-02-25 01:31:26 +00:00
|
|
|
} else {
|
|
|
|
ASSERT(orig_off >= align_off);
|
xfs: extent size hints can round up extents past MAXEXTLEN
This results in BMBT corruption, as seen by this test:
# mkfs.xfs -f -d size=40051712b,agcount=4 /dev/vdc
....
# mount /dev/vdc /mnt/scratch
# xfs_io -ft -c "extsize 16m" -c "falloc 0 30g" -c "bmap -vp" /mnt/scratch/foo
which results in this failure on a debug kernel:
XFS: Assertion failed: (blockcount & xfs_mask64hi(64-BMBT_BLOCKCOUNT_BITLEN)) == 0, file: fs/xfs/libxfs/xfs_bmap_btree.c, line: 211
....
Call Trace:
[<ffffffff814cf0ff>] xfs_bmbt_set_allf+0x8f/0x100
[<ffffffff814cf18d>] xfs_bmbt_set_all+0x1d/0x20
[<ffffffff814f2efe>] xfs_iext_insert+0x9e/0x120
[<ffffffff814c7956>] ? xfs_bmap_add_extent_hole_real+0x1c6/0xc70
[<ffffffff814c7956>] xfs_bmap_add_extent_hole_real+0x1c6/0xc70
[<ffffffff814caaab>] xfs_bmapi_write+0x72b/0xed0
[<ffffffff811c72ac>] ? kmem_cache_alloc+0x15c/0x170
[<ffffffff814fe070>] xfs_alloc_file_space+0x160/0x400
[<ffffffff81ddcc29>] ? down_write+0x29/0x60
[<ffffffff815063eb>] xfs_file_fallocate+0x29b/0x310
[<ffffffff811d2bc8>] ? __sb_start_write+0x58/0x120
[<ffffffff811e3e18>] ? do_vfs_ioctl+0x318/0x570
[<ffffffff811cd680>] vfs_fallocate+0x140/0x260
[<ffffffff811ce6f8>] SyS_fallocate+0x48/0x80
[<ffffffff81ddec09>] system_call_fastpath+0x12/0x17
The tracepoint that indicates the extent that triggered the assert
failure is:
xfs_iext_insert: idx 0 offset 0 block 16777224 count 2097152 flag 1
Clearly indicating that the extent length is greater than MAXEXTLEN,
which is 2097151. A prior trace point shows the allocation was an
exact size match and that a length greater than MAXEXTLEN was asked
for:
xfs_alloc_size_done: agno 1 agbno 8 minlen 2097152 maxlen 2097152
^^^^^^^ ^^^^^^^
We don't see this problem with extent size hints through the IO path
because we can't do single IOs large enough to trigger MAXEXTLEN
allocation. fallocate(), OTOH, is not limited in it's allocation
sizes and so needs help here.
The issue is that the extent size hint alignment is rounding up the
extent size past MAXEXTLEN, because xfs_bmapi_write() is not taking
into account extent size hints when calculating the maximum extent
length to allocate. xfs_bmapi_reserve_delalloc() is already doing
this, but direct extent allocation is not.
Unfortunately, the calculation in xfs_bmapi_reserve_delalloc() is
wrong, and it works only because delayed allocation extents are not
limited in size to MAXEXTLEN in the in-core extent tree. hence this
calculation does not work for direct allocation, and the delalloc
code needs fixing. This may, in fact be the underlying bug that
occassionally causes transaction overruns in delayed allocation
extent conversion, so now we know it's wrong we should fix it, too.
Many thanks to Brian Foster for finding this problem during review
of this patch.
Hence the fix, after much code reading, is to allow
xfs_bmap_extsize_align() to align partial extents when full
alignment would extend the alignment past MAXEXTLEN. We can safely
do this because all callers have higher layer allocation loops that
already handle short allocations, and so will simply run another
allocation to cover the remainder of the requested allocation range
that we ignored during alignment. The advantage of this approach is
that it also removes the need for callers to do anything other than
limit their requests to MAXEXTLEN - they don't really need to be
aware of extent size hints at all.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2015-05-28 21:40:06 +00:00
|
|
|
/* see MAXEXTLEN handling above */
|
|
|
|
ASSERT(orig_end <= align_off + align_alen ||
|
|
|
|
align_alen + extsz > MAXEXTLEN);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2009-12-14 23:14:59 +00:00
|
|
|
#ifdef DEBUG
|
2013-02-25 01:31:26 +00:00
|
|
|
if (!eof && gotp->br_startoff != NULLFILEOFF)
|
|
|
|
ASSERT(align_off + align_alen <= gotp->br_startoff);
|
|
|
|
if (prevp->br_startoff != NULLFILEOFF)
|
|
|
|
ASSERT(align_off >= prevp->br_startoff + prevp->br_blockcount);
|
|
|
|
#endif
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
*lenp = align_alen;
|
|
|
|
*offp = align_off;
|
|
|
|
return 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
#define XFS_ALLOC_GAP_UNITS 4
|
|
|
|
|
2013-08-12 10:49:42 +00:00
|
|
|
void
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmap_adjacent(
|
2013-08-12 10:49:42 +00:00
|
|
|
struct xfs_bmalloca *ap) /* bmap alloc argument struct */
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_fsblock_t adjust; /* adjustment to block numbers */
|
|
|
|
xfs_agnumber_t fb_agno; /* ag number of ap->firstblock */
|
|
|
|
xfs_mount_t *mp; /* mount point structure */
|
|
|
|
int nullfb; /* true if ap->firstblock isn't set */
|
|
|
|
int rt; /* true if inode is realtime */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
#define ISVALID(x,y) \
|
|
|
|
(rt ? \
|
|
|
|
(x) < mp->m_sb.sb_rblocks : \
|
|
|
|
XFS_FSB_TO_AGNO(mp, x) == XFS_FSB_TO_AGNO(mp, y) && \
|
|
|
|
XFS_FSB_TO_AGNO(mp, x) < mp->m_sb.sb_agcount && \
|
|
|
|
XFS_FSB_TO_AGBNO(mp, x) < mp->m_sb.sb_agblocks)
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
mp = ap->ip->i_mount;
|
|
|
|
nullfb = *ap->firstblock == NULLFSBLOCK;
|
xfs: remote attribute blocks aren't really userdata
When adding a new remote attribute, we write the attribute to the
new extent before the allocation transaction is committed. This
means we cannot reuse busy extents as that violates crash
consistency semantics. Hence we currently treat remote attribute
extent allocation like userdata because it has the same overwrite
ordering constraints as userdata.
Unfortunately, this also allows the allocator to incorrectly apply
extent size hints to the remote attribute extent allocation. This
results in interesting failures, such as transaction block
reservation overruns and in-memory inode attribute fork corruption.
To fix this, we need to separate the busy extent reuse configuration
from the userdata configuration. This changes the definition of
XFS_BMAPI_METADATA slightly - it now means that allocation is
metadata and reuse of busy extents is acceptible due to the metadata
ordering semantics of the journal. If this flag is not set, it
means the allocation is that has unordered data writeback, and hence
busy extent reuse is not allowed. It no longer implies the
allocation is for user data, just that the data write will not be
strictly ordered. This matches the semantics for both user data
and remote attribute block allocation.
As such, This patch changes the "userdata" field to a "datatype"
field, and adds a "no busy reuse" flag to the field.
When we detect an unordered data extent allocation, we immediately set
the no reuse flag. We then set the "user data" flags based on the
inode fork we are allocating the extent to. Hence we only set
userdata flags on data fork allocations now and consider attribute
fork remote extents to be an unordered metadata extent.
The result is that remote attribute extents now have the expected
allocation semantics, and the data fork allocation behaviour is
completely unchanged.
It should be noted that there may be other ways to fix this (e.g.
use ordered metadata buffers for the remote attribute extent data
write) but they are more invasive and difficult to validate both
from a design and implementation POV. Hence this patch takes the
simple, obvious route to fixing the problem...
Reported-and-tested-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-25 22:21:28 +00:00
|
|
|
rt = XFS_IS_REALTIME_INODE(ap->ip) &&
|
|
|
|
xfs_alloc_is_userdata(ap->datatype);
|
2013-02-25 01:31:26 +00:00
|
|
|
fb_agno = nullfb ? NULLAGNUMBER : XFS_FSB_TO_AGNO(mp, *ap->firstblock);
|
|
|
|
/*
|
|
|
|
* If allocating at eof, and there's a previous real block,
|
|
|
|
* try to use its last block as our starting point.
|
|
|
|
*/
|
|
|
|
if (ap->eof && ap->prev.br_startoff != NULLFILEOFF &&
|
|
|
|
!isnullstartblock(ap->prev.br_startblock) &&
|
|
|
|
ISVALID(ap->prev.br_startblock + ap->prev.br_blockcount,
|
|
|
|
ap->prev.br_startblock)) {
|
|
|
|
ap->blkno = ap->prev.br_startblock + ap->prev.br_blockcount;
|
|
|
|
/*
|
|
|
|
* Adjust for the gap between prevp and us.
|
|
|
|
*/
|
|
|
|
adjust = ap->offset -
|
|
|
|
(ap->prev.br_startoff + ap->prev.br_blockcount);
|
|
|
|
if (adjust &&
|
|
|
|
ISVALID(ap->blkno + adjust, ap->prev.br_startblock))
|
|
|
|
ap->blkno += adjust;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* If not at eof, then compare the two neighbor blocks.
|
|
|
|
* Figure out whether either one gives us a good starting point,
|
|
|
|
* and pick the better one.
|
|
|
|
*/
|
|
|
|
else if (!ap->eof) {
|
|
|
|
xfs_fsblock_t gotbno; /* right side block number */
|
|
|
|
xfs_fsblock_t gotdiff=0; /* right side difference */
|
|
|
|
xfs_fsblock_t prevbno; /* left side block number */
|
|
|
|
xfs_fsblock_t prevdiff=0; /* left side difference */
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If there's a previous (left) block, select a requested
|
|
|
|
* start block based on it.
|
|
|
|
*/
|
|
|
|
if (ap->prev.br_startoff != NULLFILEOFF &&
|
|
|
|
!isnullstartblock(ap->prev.br_startblock) &&
|
|
|
|
(prevbno = ap->prev.br_startblock +
|
|
|
|
ap->prev.br_blockcount) &&
|
|
|
|
ISVALID(prevbno, ap->prev.br_startblock)) {
|
|
|
|
/*
|
|
|
|
* Calculate gap to end of previous block.
|
|
|
|
*/
|
|
|
|
adjust = prevdiff = ap->offset -
|
|
|
|
(ap->prev.br_startoff +
|
|
|
|
ap->prev.br_blockcount);
|
|
|
|
/*
|
|
|
|
* Figure the startblock based on the previous block's
|
|
|
|
* end and the gap size.
|
|
|
|
* Heuristic!
|
|
|
|
* If the gap is large relative to the piece we're
|
|
|
|
* allocating, or using it gives us an invalid block
|
|
|
|
* number, then just use the end of the previous block.
|
|
|
|
*/
|
|
|
|
if (prevdiff <= XFS_ALLOC_GAP_UNITS * ap->length &&
|
|
|
|
ISVALID(prevbno + prevdiff,
|
|
|
|
ap->prev.br_startblock))
|
|
|
|
prevbno += adjust;
|
|
|
|
else
|
|
|
|
prevdiff += adjust;
|
|
|
|
/*
|
|
|
|
* If the firstblock forbids it, can't use it,
|
|
|
|
* must use default.
|
|
|
|
*/
|
|
|
|
if (!rt && !nullfb &&
|
|
|
|
XFS_FSB_TO_AGNO(mp, prevbno) != fb_agno)
|
|
|
|
prevbno = NULLFSBLOCK;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* No previous block or can't follow it, just default.
|
|
|
|
*/
|
|
|
|
else
|
|
|
|
prevbno = NULLFSBLOCK;
|
|
|
|
/*
|
|
|
|
* If there's a following (right) block, select a requested
|
|
|
|
* start block based on it.
|
|
|
|
*/
|
|
|
|
if (!isnullstartblock(ap->got.br_startblock)) {
|
|
|
|
/*
|
|
|
|
* Calculate gap to start of next block.
|
|
|
|
*/
|
|
|
|
adjust = gotdiff = ap->got.br_startoff - ap->offset;
|
|
|
|
/*
|
|
|
|
* Figure the startblock based on the next block's
|
|
|
|
* start and the gap size.
|
|
|
|
*/
|
|
|
|
gotbno = ap->got.br_startblock;
|
|
|
|
/*
|
|
|
|
* Heuristic!
|
|
|
|
* If the gap is large relative to the piece we're
|
|
|
|
* allocating, or using it gives us an invalid block
|
|
|
|
* number, then just use the start of the next block
|
|
|
|
* offset by our length.
|
|
|
|
*/
|
|
|
|
if (gotdiff <= XFS_ALLOC_GAP_UNITS * ap->length &&
|
|
|
|
ISVALID(gotbno - gotdiff, gotbno))
|
|
|
|
gotbno -= adjust;
|
|
|
|
else if (ISVALID(gotbno - ap->length, gotbno)) {
|
|
|
|
gotbno -= ap->length;
|
|
|
|
gotdiff += adjust - ap->length;
|
|
|
|
} else
|
|
|
|
gotdiff += adjust;
|
|
|
|
/*
|
|
|
|
* If the firstblock forbids it, can't use it,
|
|
|
|
* must use default.
|
|
|
|
*/
|
|
|
|
if (!rt && !nullfb &&
|
|
|
|
XFS_FSB_TO_AGNO(mp, gotbno) != fb_agno)
|
|
|
|
gotbno = NULLFSBLOCK;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* No next block, just default.
|
|
|
|
*/
|
|
|
|
else
|
|
|
|
gotbno = NULLFSBLOCK;
|
|
|
|
/*
|
|
|
|
* If both valid, pick the better one, else the only good
|
|
|
|
* one, else ap->blkno is already set (to 0 or the inode block).
|
|
|
|
*/
|
|
|
|
if (prevbno != NULLFSBLOCK && gotbno != NULLFSBLOCK)
|
|
|
|
ap->blkno = prevdiff <= gotdiff ? prevbno : gotbno;
|
|
|
|
else if (prevbno != NULLFSBLOCK)
|
|
|
|
ap->blkno = prevbno;
|
|
|
|
else if (gotbno != NULLFSBLOCK)
|
|
|
|
ap->blkno = gotbno;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
#undef ISVALID
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2014-04-22 21:11:41 +00:00
|
|
|
static int
|
|
|
|
xfs_bmap_longest_free_extent(
|
|
|
|
struct xfs_trans *tp,
|
|
|
|
xfs_agnumber_t ag,
|
|
|
|
xfs_extlen_t *blen,
|
|
|
|
int *notinit)
|
|
|
|
{
|
|
|
|
struct xfs_mount *mp = tp->t_mountp;
|
|
|
|
struct xfs_perag *pag;
|
|
|
|
xfs_extlen_t longest;
|
|
|
|
int error = 0;
|
|
|
|
|
|
|
|
pag = xfs_perag_get(mp, ag);
|
|
|
|
if (!pag->pagf_init) {
|
|
|
|
error = xfs_alloc_pagf_init(mp, tp, ag, XFS_ALLOC_FLAG_TRYLOCK);
|
|
|
|
if (error)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
if (!pag->pagf_init) {
|
|
|
|
*notinit = 1;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2015-06-22 00:04:31 +00:00
|
|
|
longest = xfs_alloc_longest_free_extent(mp, pag,
|
xfs: set up per-AG free space reservations
One unfortunate quirk of the reference count and reverse mapping
btrees -- they can expand in size when blocks are written to *other*
allocation groups if, say, one large extent becomes a lot of tiny
extents. Since we don't want to start throwing errors in the middle
of CoWing, we need to reserve some blocks to handle future expansion.
The transaction block reservation counters aren't sufficient here
because we have to have a reserve of blocks in every AG, not just
somewhere in the filesystem.
Therefore, create two per-AG block reservation pools. One feeds the
AGFL so that rmapbt expansion always succeeds, and the other feeds all
other metadata so that refcountbt expansion never fails.
Use the count of how many reserved blocks we need to have on hand to
create a virtual reservation in the AG. Through selective clamping of
the maximum length of allocation requests and of the length of the
longest free extent, we can make it look like there's less free space
in the AG unless the reservation owner is asking for blocks.
In other words, play some accounting tricks in-core to make sure that
we always have blocks available. On the plus side, there's nothing to
clean up if we crash, which is contrast to the strategy that the rough
draft used (actually removing extents from the freespace btrees).
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-19 00:30:52 +00:00
|
|
|
xfs_alloc_min_freelist(mp, pag),
|
|
|
|
xfs_ag_resv_needed(pag, XFS_AG_RESV_NONE));
|
2014-04-22 21:11:41 +00:00
|
|
|
if (*blen < longest)
|
|
|
|
*blen = longest;
|
|
|
|
|
|
|
|
out:
|
|
|
|
xfs_perag_put(pag);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
xfs_bmap_select_minlen(
|
|
|
|
struct xfs_bmalloca *ap,
|
|
|
|
struct xfs_alloc_arg *args,
|
|
|
|
xfs_extlen_t *blen,
|
|
|
|
int notinit)
|
|
|
|
{
|
|
|
|
if (notinit || *blen < ap->minlen) {
|
|
|
|
/*
|
|
|
|
* Since we did a BUF_TRYLOCK above, it is possible that
|
|
|
|
* there is space for this request.
|
|
|
|
*/
|
|
|
|
args->minlen = ap->minlen;
|
|
|
|
} else if (*blen < args->maxlen) {
|
|
|
|
/*
|
|
|
|
* If the best seen length is less than the request length,
|
|
|
|
* use the best as the minimum.
|
|
|
|
*/
|
|
|
|
args->minlen = *blen;
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* Otherwise we've seen an extent as big as maxlen, use that
|
|
|
|
* as the minimum.
|
|
|
|
*/
|
|
|
|
args->minlen = args->maxlen;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-09-18 20:40:47 +00:00
|
|
|
STATIC int
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmap_btalloc_nullfb(
|
|
|
|
struct xfs_bmalloca *ap,
|
|
|
|
struct xfs_alloc_arg *args,
|
|
|
|
xfs_extlen_t *blen)
|
2011-09-18 20:40:47 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
struct xfs_mount *mp = ap->ip->i_mount;
|
|
|
|
xfs_agnumber_t ag, startag;
|
|
|
|
int notinit = 0;
|
2011-09-18 20:40:47 +00:00
|
|
|
int error;
|
|
|
|
|
2014-04-22 21:11:41 +00:00
|
|
|
args->type = XFS_ALLOCTYPE_START_BNO;
|
2013-02-25 01:31:26 +00:00
|
|
|
args->total = ap->total;
|
2011-09-18 20:40:47 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
startag = ag = XFS_FSB_TO_AGNO(mp, args->fsbno);
|
|
|
|
if (startag == NULLAGNUMBER)
|
|
|
|
startag = ag = 0;
|
2011-09-18 20:40:47 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
while (*blen < args->maxlen) {
|
2014-04-22 21:11:41 +00:00
|
|
|
error = xfs_bmap_longest_free_extent(args->tp, ag, blen,
|
|
|
|
¬init);
|
|
|
|
if (error)
|
|
|
|
return error;
|
2011-09-18 20:40:47 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (++ag == mp->m_sb.sb_agcount)
|
|
|
|
ag = 0;
|
|
|
|
if (ag == startag)
|
|
|
|
break;
|
|
|
|
}
|
2011-09-18 20:40:47 +00:00
|
|
|
|
2014-04-22 21:11:41 +00:00
|
|
|
xfs_bmap_select_minlen(ap, args, blen, notinit);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
STATIC int
|
|
|
|
xfs_bmap_btalloc_filestreams(
|
|
|
|
struct xfs_bmalloca *ap,
|
|
|
|
struct xfs_alloc_arg *args,
|
|
|
|
xfs_extlen_t *blen)
|
|
|
|
{
|
|
|
|
struct xfs_mount *mp = ap->ip->i_mount;
|
|
|
|
xfs_agnumber_t ag;
|
|
|
|
int notinit = 0;
|
|
|
|
int error;
|
|
|
|
|
|
|
|
args->type = XFS_ALLOCTYPE_NEAR_BNO;
|
|
|
|
args->total = ap->total;
|
|
|
|
|
|
|
|
ag = XFS_FSB_TO_AGNO(mp, args->fsbno);
|
|
|
|
if (ag == NULLAGNUMBER)
|
|
|
|
ag = 0;
|
|
|
|
|
|
|
|
error = xfs_bmap_longest_free_extent(args->tp, ag, blen, ¬init);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
|
|
|
|
if (*blen < args->maxlen) {
|
|
|
|
error = xfs_filestream_new_ag(ap, &ag);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
|
|
|
|
error = xfs_bmap_longest_free_extent(args->tp, ag, blen,
|
|
|
|
¬init);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
xfs_bmap_select_minlen(ap, args, blen, notinit);
|
2011-09-18 20:40:47 +00:00
|
|
|
|
|
|
|
/*
|
2014-04-22 21:11:41 +00:00
|
|
|
* Set the failure fallback case to look in the selected AG as stream
|
|
|
|
* may have moved.
|
2011-09-18 20:40:47 +00:00
|
|
|
*/
|
2014-04-22 21:11:41 +00:00
|
|
|
ap->blkno = args->fsbno = XFS_AGB_TO_FSB(mp, ag, 0);
|
2011-09-18 20:40:47 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
STATIC int
|
|
|
|
xfs_bmap_btalloc(
|
2013-08-12 10:49:42 +00:00
|
|
|
struct xfs_bmalloca *ap) /* bmap alloc argument struct */
|
2011-09-18 20:40:48 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_mount_t *mp; /* mount point structure */
|
|
|
|
xfs_alloctype_t atype = 0; /* type for allocation routines */
|
xfs: remote attribute blocks aren't really userdata
When adding a new remote attribute, we write the attribute to the
new extent before the allocation transaction is committed. This
means we cannot reuse busy extents as that violates crash
consistency semantics. Hence we currently treat remote attribute
extent allocation like userdata because it has the same overwrite
ordering constraints as userdata.
Unfortunately, this also allows the allocator to incorrectly apply
extent size hints to the remote attribute extent allocation. This
results in interesting failures, such as transaction block
reservation overruns and in-memory inode attribute fork corruption.
To fix this, we need to separate the busy extent reuse configuration
from the userdata configuration. This changes the definition of
XFS_BMAPI_METADATA slightly - it now means that allocation is
metadata and reuse of busy extents is acceptible due to the metadata
ordering semantics of the journal. If this flag is not set, it
means the allocation is that has unordered data writeback, and hence
busy extent reuse is not allowed. It no longer implies the
allocation is for user data, just that the data write will not be
strictly ordered. This matches the semantics for both user data
and remote attribute block allocation.
As such, This patch changes the "userdata" field to a "datatype"
field, and adds a "no busy reuse" flag to the field.
When we detect an unordered data extent allocation, we immediately set
the no reuse flag. We then set the "user data" flags based on the
inode fork we are allocating the extent to. Hence we only set
userdata flags on data fork allocations now and consider attribute
fork remote extents to be an unordered metadata extent.
The result is that remote attribute extents now have the expected
allocation semantics, and the data fork allocation behaviour is
completely unchanged.
It should be noted that there may be other ways to fix this (e.g.
use ordered metadata buffers for the remote attribute extent data
write) but they are more invasive and difficult to validate both
from a design and implementation POV. Hence this patch takes the
simple, obvious route to fixing the problem...
Reported-and-tested-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-25 22:21:28 +00:00
|
|
|
xfs_extlen_t align = 0; /* minimum allocation alignment */
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_agnumber_t fb_agno; /* ag number of ap->firstblock */
|
|
|
|
xfs_agnumber_t ag;
|
|
|
|
xfs_alloc_arg_t args;
|
|
|
|
xfs_extlen_t blen;
|
|
|
|
xfs_extlen_t nextminlen = 0;
|
|
|
|
int nullfb; /* true if ap->firstblock isn't set */
|
|
|
|
int isaligned;
|
|
|
|
int tryagain;
|
|
|
|
int error;
|
2013-12-12 05:34:36 +00:00
|
|
|
int stripe_align;
|
2011-09-18 20:40:48 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(ap->length);
|
2011-09-18 20:40:48 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
mp = ap->ip->i_mount;
|
2013-12-12 05:34:36 +00:00
|
|
|
|
|
|
|
/* stripe alignment for allocation is determined by mount parameters */
|
|
|
|
stripe_align = 0;
|
|
|
|
if (mp->m_swidth && (mp->m_flags & XFS_MOUNT_SWALLOC))
|
|
|
|
stripe_align = mp->m_swidth;
|
|
|
|
else if (mp->m_dalign)
|
|
|
|
stripe_align = mp->m_dalign;
|
|
|
|
|
2016-10-03 16:11:43 +00:00
|
|
|
if (ap->flags & XFS_BMAPI_COWFORK)
|
|
|
|
align = xfs_get_cowextsz_hint(ap->ip);
|
|
|
|
else if (xfs_alloc_is_userdata(ap->datatype))
|
xfs: remote attribute blocks aren't really userdata
When adding a new remote attribute, we write the attribute to the
new extent before the allocation transaction is committed. This
means we cannot reuse busy extents as that violates crash
consistency semantics. Hence we currently treat remote attribute
extent allocation like userdata because it has the same overwrite
ordering constraints as userdata.
Unfortunately, this also allows the allocator to incorrectly apply
extent size hints to the remote attribute extent allocation. This
results in interesting failures, such as transaction block
reservation overruns and in-memory inode attribute fork corruption.
To fix this, we need to separate the busy extent reuse configuration
from the userdata configuration. This changes the definition of
XFS_BMAPI_METADATA slightly - it now means that allocation is
metadata and reuse of busy extents is acceptible due to the metadata
ordering semantics of the journal. If this flag is not set, it
means the allocation is that has unordered data writeback, and hence
busy extent reuse is not allowed. It no longer implies the
allocation is for user data, just that the data write will not be
strictly ordered. This matches the semantics for both user data
and remote attribute block allocation.
As such, This patch changes the "userdata" field to a "datatype"
field, and adds a "no busy reuse" flag to the field.
When we detect an unordered data extent allocation, we immediately set
the no reuse flag. We then set the "user data" flags based on the
inode fork we are allocating the extent to. Hence we only set
userdata flags on data fork allocations now and consider attribute
fork remote extents to be an unordered metadata extent.
The result is that remote attribute extents now have the expected
allocation semantics, and the data fork allocation behaviour is
completely unchanged.
It should be noted that there may be other ways to fix this (e.g.
use ordered metadata buffers for the remote attribute extent data
write) but they are more invasive and difficult to validate both
from a design and implementation POV. Hence this patch takes the
simple, obvious route to fixing the problem...
Reported-and-tested-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-25 22:21:28 +00:00
|
|
|
align = xfs_get_extsz_hint(ap->ip);
|
2017-01-25 16:59:43 +00:00
|
|
|
if (align) {
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_bmap_extsize_align(mp, &ap->got, &ap->prev,
|
|
|
|
align, 0, ap->eof, 0, ap->conv,
|
|
|
|
&ap->offset, &ap->length);
|
|
|
|
ASSERT(!error);
|
|
|
|
ASSERT(ap->length);
|
2011-09-18 20:40:48 +00:00
|
|
|
}
|
2013-12-12 05:34:36 +00:00
|
|
|
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
nullfb = *ap->firstblock == NULLFSBLOCK;
|
|
|
|
fb_agno = nullfb ? NULLAGNUMBER : XFS_FSB_TO_AGNO(mp, *ap->firstblock);
|
|
|
|
if (nullfb) {
|
xfs: remote attribute blocks aren't really userdata
When adding a new remote attribute, we write the attribute to the
new extent before the allocation transaction is committed. This
means we cannot reuse busy extents as that violates crash
consistency semantics. Hence we currently treat remote attribute
extent allocation like userdata because it has the same overwrite
ordering constraints as userdata.
Unfortunately, this also allows the allocator to incorrectly apply
extent size hints to the remote attribute extent allocation. This
results in interesting failures, such as transaction block
reservation overruns and in-memory inode attribute fork corruption.
To fix this, we need to separate the busy extent reuse configuration
from the userdata configuration. This changes the definition of
XFS_BMAPI_METADATA slightly - it now means that allocation is
metadata and reuse of busy extents is acceptible due to the metadata
ordering semantics of the journal. If this flag is not set, it
means the allocation is that has unordered data writeback, and hence
busy extent reuse is not allowed. It no longer implies the
allocation is for user data, just that the data write will not be
strictly ordered. This matches the semantics for both user data
and remote attribute block allocation.
As such, This patch changes the "userdata" field to a "datatype"
field, and adds a "no busy reuse" flag to the field.
When we detect an unordered data extent allocation, we immediately set
the no reuse flag. We then set the "user data" flags based on the
inode fork we are allocating the extent to. Hence we only set
userdata flags on data fork allocations now and consider attribute
fork remote extents to be an unordered metadata extent.
The result is that remote attribute extents now have the expected
allocation semantics, and the data fork allocation behaviour is
completely unchanged.
It should be noted that there may be other ways to fix this (e.g.
use ordered metadata buffers for the remote attribute extent data
write) but they are more invasive and difficult to validate both
from a design and implementation POV. Hence this patch takes the
simple, obvious route to fixing the problem...
Reported-and-tested-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-25 22:21:28 +00:00
|
|
|
if (xfs_alloc_is_userdata(ap->datatype) &&
|
|
|
|
xfs_inode_is_filestream(ap->ip)) {
|
2013-02-25 01:31:26 +00:00
|
|
|
ag = xfs_filestream_lookup_ag(ap->ip);
|
|
|
|
ag = (ag != NULLAGNUMBER) ? ag : 0;
|
|
|
|
ap->blkno = XFS_AGB_TO_FSB(mp, ag, 0);
|
|
|
|
} else {
|
|
|
|
ap->blkno = XFS_INO_TO_FSB(mp, ap->ip->i_ino);
|
|
|
|
}
|
|
|
|
} else
|
|
|
|
ap->blkno = *ap->firstblock;
|
2011-09-18 20:40:48 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmap_adjacent(ap);
|
2011-09-18 20:40:48 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* If allowed, use ap->blkno; otherwise must use firstblock since
|
|
|
|
* it's in the right allocation group.
|
|
|
|
*/
|
|
|
|
if (nullfb || XFS_FSB_TO_AGNO(mp, ap->blkno) == fb_agno)
|
|
|
|
;
|
|
|
|
else
|
|
|
|
ap->blkno = *ap->firstblock;
|
|
|
|
/*
|
|
|
|
* Normal allocation, done through xfs_alloc_vextent.
|
|
|
|
*/
|
|
|
|
tryagain = isaligned = 0;
|
|
|
|
memset(&args, 0, sizeof(args));
|
|
|
|
args.tp = ap->tp;
|
|
|
|
args.mp = mp;
|
|
|
|
args.fsbno = ap->blkno;
|
2016-08-03 01:33:42 +00:00
|
|
|
xfs_rmap_skip_owner_update(&args.oinfo);
|
2011-09-18 20:40:48 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/* Trim the allocation back to the maximum an AG can fit. */
|
2016-08-03 01:38:24 +00:00
|
|
|
args.maxlen = MIN(ap->length, mp->m_ag_max_usable);
|
2013-02-25 01:31:26 +00:00
|
|
|
args.firstblock = *ap->firstblock;
|
|
|
|
blen = 0;
|
|
|
|
if (nullfb) {
|
2014-04-22 21:11:41 +00:00
|
|
|
/*
|
|
|
|
* Search for an allocation group with a single extent large
|
|
|
|
* enough for the request. If one isn't found, then adjust
|
|
|
|
* the minimum allocation size to the largest space found.
|
|
|
|
*/
|
xfs: remote attribute blocks aren't really userdata
When adding a new remote attribute, we write the attribute to the
new extent before the allocation transaction is committed. This
means we cannot reuse busy extents as that violates crash
consistency semantics. Hence we currently treat remote attribute
extent allocation like userdata because it has the same overwrite
ordering constraints as userdata.
Unfortunately, this also allows the allocator to incorrectly apply
extent size hints to the remote attribute extent allocation. This
results in interesting failures, such as transaction block
reservation overruns and in-memory inode attribute fork corruption.
To fix this, we need to separate the busy extent reuse configuration
from the userdata configuration. This changes the definition of
XFS_BMAPI_METADATA slightly - it now means that allocation is
metadata and reuse of busy extents is acceptible due to the metadata
ordering semantics of the journal. If this flag is not set, it
means the allocation is that has unordered data writeback, and hence
busy extent reuse is not allowed. It no longer implies the
allocation is for user data, just that the data write will not be
strictly ordered. This matches the semantics for both user data
and remote attribute block allocation.
As such, This patch changes the "userdata" field to a "datatype"
field, and adds a "no busy reuse" flag to the field.
When we detect an unordered data extent allocation, we immediately set
the no reuse flag. We then set the "user data" flags based on the
inode fork we are allocating the extent to. Hence we only set
userdata flags on data fork allocations now and consider attribute
fork remote extents to be an unordered metadata extent.
The result is that remote attribute extents now have the expected
allocation semantics, and the data fork allocation behaviour is
completely unchanged.
It should be noted that there may be other ways to fix this (e.g.
use ordered metadata buffers for the remote attribute extent data
write) but they are more invasive and difficult to validate both
from a design and implementation POV. Hence this patch takes the
simple, obvious route to fixing the problem...
Reported-and-tested-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-25 22:21:28 +00:00
|
|
|
if (xfs_alloc_is_userdata(ap->datatype) &&
|
|
|
|
xfs_inode_is_filestream(ap->ip))
|
2014-04-22 21:11:41 +00:00
|
|
|
error = xfs_bmap_btalloc_filestreams(ap, &args, &blen);
|
|
|
|
else
|
|
|
|
error = xfs_bmap_btalloc_nullfb(ap, &args, &blen);
|
2011-09-18 20:40:48 +00:00
|
|
|
if (error)
|
|
|
|
return error;
|
2016-08-03 01:19:29 +00:00
|
|
|
} else if (ap->dfops->dop_low) {
|
2013-02-25 01:31:26 +00:00
|
|
|
if (xfs_inode_is_filestream(ap->ip))
|
|
|
|
args.type = XFS_ALLOCTYPE_FIRST_AG;
|
|
|
|
else
|
|
|
|
args.type = XFS_ALLOCTYPE_START_BNO;
|
|
|
|
args.total = args.minlen = ap->minlen;
|
|
|
|
} else {
|
|
|
|
args.type = XFS_ALLOCTYPE_NEAR_BNO;
|
|
|
|
args.total = ap->total;
|
|
|
|
args.minlen = ap->minlen;
|
2011-09-18 20:40:48 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
/* apply extent size hints if obtained earlier */
|
2017-01-25 16:59:43 +00:00
|
|
|
if (align) {
|
2013-02-25 01:31:26 +00:00
|
|
|
args.prod = align;
|
|
|
|
if ((args.mod = (xfs_extlen_t)do_mod(ap->offset, args.prod)))
|
|
|
|
args.mod = (xfs_extlen_t)(args.prod - args.mod);
|
mm, fs: get rid of PAGE_CACHE_* and page_cache_{get,release} macros
PAGE_CACHE_{SIZE,SHIFT,MASK,ALIGN} macros were introduced *long* time
ago with promise that one day it will be possible to implement page
cache with bigger chunks than PAGE_SIZE.
This promise never materialized. And unlikely will.
We have many places where PAGE_CACHE_SIZE assumed to be equal to
PAGE_SIZE. And it's constant source of confusion on whether
PAGE_CACHE_* or PAGE_* constant should be used in a particular case,
especially on the border between fs and mm.
Global switching to PAGE_CACHE_SIZE != PAGE_SIZE would cause to much
breakage to be doable.
Let's stop pretending that pages in page cache are special. They are
not.
The changes are pretty straight-forward:
- <foo> << (PAGE_CACHE_SHIFT - PAGE_SHIFT) -> <foo>;
- <foo> >> (PAGE_CACHE_SHIFT - PAGE_SHIFT) -> <foo>;
- PAGE_CACHE_{SIZE,SHIFT,MASK,ALIGN} -> PAGE_{SIZE,SHIFT,MASK,ALIGN};
- page_cache_get() -> get_page();
- page_cache_release() -> put_page();
This patch contains automated changes generated with coccinelle using
script below. For some reason, coccinelle doesn't patch header files.
I've called spatch for them manually.
The only adjustment after coccinelle is revert of changes to
PAGE_CAHCE_ALIGN definition: we are going to drop it later.
There are few places in the code where coccinelle didn't reach. I'll
fix them manually in a separate patch. Comments and documentation also
will be addressed with the separate patch.
virtual patch
@@
expression E;
@@
- E << (PAGE_CACHE_SHIFT - PAGE_SHIFT)
+ E
@@
expression E;
@@
- E >> (PAGE_CACHE_SHIFT - PAGE_SHIFT)
+ E
@@
@@
- PAGE_CACHE_SHIFT
+ PAGE_SHIFT
@@
@@
- PAGE_CACHE_SIZE
+ PAGE_SIZE
@@
@@
- PAGE_CACHE_MASK
+ PAGE_MASK
@@
expression E;
@@
- PAGE_CACHE_ALIGN(E)
+ PAGE_ALIGN(E)
@@
expression E;
@@
- page_cache_get(E)
+ get_page(E)
@@
expression E;
@@
- page_cache_release(E)
+ put_page(E)
Signed-off-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2016-04-01 12:29:47 +00:00
|
|
|
} else if (mp->m_sb.sb_blocksize >= PAGE_SIZE) {
|
2013-02-25 01:31:26 +00:00
|
|
|
args.prod = 1;
|
|
|
|
args.mod = 0;
|
|
|
|
} else {
|
mm, fs: get rid of PAGE_CACHE_* and page_cache_{get,release} macros
PAGE_CACHE_{SIZE,SHIFT,MASK,ALIGN} macros were introduced *long* time
ago with promise that one day it will be possible to implement page
cache with bigger chunks than PAGE_SIZE.
This promise never materialized. And unlikely will.
We have many places where PAGE_CACHE_SIZE assumed to be equal to
PAGE_SIZE. And it's constant source of confusion on whether
PAGE_CACHE_* or PAGE_* constant should be used in a particular case,
especially on the border between fs and mm.
Global switching to PAGE_CACHE_SIZE != PAGE_SIZE would cause to much
breakage to be doable.
Let's stop pretending that pages in page cache are special. They are
not.
The changes are pretty straight-forward:
- <foo> << (PAGE_CACHE_SHIFT - PAGE_SHIFT) -> <foo>;
- <foo> >> (PAGE_CACHE_SHIFT - PAGE_SHIFT) -> <foo>;
- PAGE_CACHE_{SIZE,SHIFT,MASK,ALIGN} -> PAGE_{SIZE,SHIFT,MASK,ALIGN};
- page_cache_get() -> get_page();
- page_cache_release() -> put_page();
This patch contains automated changes generated with coccinelle using
script below. For some reason, coccinelle doesn't patch header files.
I've called spatch for them manually.
The only adjustment after coccinelle is revert of changes to
PAGE_CAHCE_ALIGN definition: we are going to drop it later.
There are few places in the code where coccinelle didn't reach. I'll
fix them manually in a separate patch. Comments and documentation also
will be addressed with the separate patch.
virtual patch
@@
expression E;
@@
- E << (PAGE_CACHE_SHIFT - PAGE_SHIFT)
+ E
@@
expression E;
@@
- E >> (PAGE_CACHE_SHIFT - PAGE_SHIFT)
+ E
@@
@@
- PAGE_CACHE_SHIFT
+ PAGE_SHIFT
@@
@@
- PAGE_CACHE_SIZE
+ PAGE_SIZE
@@
@@
- PAGE_CACHE_MASK
+ PAGE_MASK
@@
expression E;
@@
- PAGE_CACHE_ALIGN(E)
+ PAGE_ALIGN(E)
@@
expression E;
@@
- page_cache_get(E)
+ get_page(E)
@@
expression E;
@@
- page_cache_release(E)
+ put_page(E)
Signed-off-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2016-04-01 12:29:47 +00:00
|
|
|
args.prod = PAGE_SIZE >> mp->m_sb.sb_blocklog;
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((args.mod = (xfs_extlen_t)(do_mod(ap->offset, args.prod))))
|
|
|
|
args.mod = (xfs_extlen_t)(args.prod - args.mod);
|
2011-09-18 20:40:48 +00:00
|
|
|
}
|
2011-09-18 20:40:50 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* If we are not low on available data blocks, and the
|
|
|
|
* underlying logical volume manager is a stripe, and
|
|
|
|
* the file offset is zero then try to allocate data
|
|
|
|
* blocks on stripe unit boundary.
|
|
|
|
* NOTE: ap->aeof is only set if the allocation length
|
|
|
|
* is >= the stripe unit and the allocation offset is
|
|
|
|
* at the end of file.
|
2011-09-18 20:40:50 +00:00
|
|
|
*/
|
2016-08-03 01:19:29 +00:00
|
|
|
if (!ap->dfops->dop_low && ap->aeof) {
|
2013-02-25 01:31:26 +00:00
|
|
|
if (!ap->offset) {
|
2013-12-12 05:34:36 +00:00
|
|
|
args.alignment = stripe_align;
|
2013-02-25 01:31:26 +00:00
|
|
|
atype = args.type;
|
|
|
|
isaligned = 1;
|
|
|
|
/*
|
|
|
|
* Adjust for alignment
|
|
|
|
*/
|
|
|
|
if (blen > args.alignment && blen <= args.maxlen)
|
|
|
|
args.minlen = blen - args.alignment;
|
|
|
|
args.minalignslop = 0;
|
|
|
|
} else {
|
|
|
|
/*
|
|
|
|
* First try an exact bno allocation.
|
|
|
|
* If it fails then do a near or start bno
|
|
|
|
* allocation with alignment turned on.
|
|
|
|
*/
|
|
|
|
atype = args.type;
|
|
|
|
tryagain = 1;
|
|
|
|
args.type = XFS_ALLOCTYPE_THIS_BNO;
|
|
|
|
args.alignment = 1;
|
|
|
|
/*
|
|
|
|
* Compute the minlen+alignment for the
|
|
|
|
* next case. Set slop so that the value
|
|
|
|
* of minlen+alignment+slop doesn't go up
|
|
|
|
* between the calls.
|
|
|
|
*/
|
2013-12-12 05:34:36 +00:00
|
|
|
if (blen > stripe_align && blen <= args.maxlen)
|
|
|
|
nextminlen = blen - stripe_align;
|
2013-02-25 01:31:26 +00:00
|
|
|
else
|
|
|
|
nextminlen = args.minlen;
|
2013-12-12 05:34:36 +00:00
|
|
|
if (nextminlen + stripe_align > args.minlen + 1)
|
2013-02-25 01:31:26 +00:00
|
|
|
args.minalignslop =
|
2013-12-12 05:34:36 +00:00
|
|
|
nextminlen + stripe_align -
|
2013-02-25 01:31:26 +00:00
|
|
|
args.minlen - 1;
|
|
|
|
else
|
|
|
|
args.minalignslop = 0;
|
2011-09-18 20:40:50 +00:00
|
|
|
}
|
|
|
|
} else {
|
2013-02-25 01:31:26 +00:00
|
|
|
args.alignment = 1;
|
|
|
|
args.minalignslop = 0;
|
2011-09-18 20:40:50 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
args.minleft = ap->minleft;
|
|
|
|
args.wasdel = ap->wasdel;
|
xfs: set up per-AG free space reservations
One unfortunate quirk of the reference count and reverse mapping
btrees -- they can expand in size when blocks are written to *other*
allocation groups if, say, one large extent becomes a lot of tiny
extents. Since we don't want to start throwing errors in the middle
of CoWing, we need to reserve some blocks to handle future expansion.
The transaction block reservation counters aren't sufficient here
because we have to have a reserve of blocks in every AG, not just
somewhere in the filesystem.
Therefore, create two per-AG block reservation pools. One feeds the
AGFL so that rmapbt expansion always succeeds, and the other feeds all
other metadata so that refcountbt expansion never fails.
Use the count of how many reserved blocks we need to have on hand to
create a virtual reservation in the AG. Through selective clamping of
the maximum length of allocation requests and of the length of the
longest free extent, we can make it look like there's less free space
in the AG unless the reservation owner is asking for blocks.
In other words, play some accounting tricks in-core to make sure that
we always have blocks available. On the plus side, there's nothing to
clean up if we crash, which is contrast to the strategy that the rough
draft used (actually removing extents from the freespace btrees).
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-19 00:30:52 +00:00
|
|
|
args.resv = XFS_AG_RESV_NONE;
|
xfs: remote attribute blocks aren't really userdata
When adding a new remote attribute, we write the attribute to the
new extent before the allocation transaction is committed. This
means we cannot reuse busy extents as that violates crash
consistency semantics. Hence we currently treat remote attribute
extent allocation like userdata because it has the same overwrite
ordering constraints as userdata.
Unfortunately, this also allows the allocator to incorrectly apply
extent size hints to the remote attribute extent allocation. This
results in interesting failures, such as transaction block
reservation overruns and in-memory inode attribute fork corruption.
To fix this, we need to separate the busy extent reuse configuration
from the userdata configuration. This changes the definition of
XFS_BMAPI_METADATA slightly - it now means that allocation is
metadata and reuse of busy extents is acceptible due to the metadata
ordering semantics of the journal. If this flag is not set, it
means the allocation is that has unordered data writeback, and hence
busy extent reuse is not allowed. It no longer implies the
allocation is for user data, just that the data write will not be
strictly ordered. This matches the semantics for both user data
and remote attribute block allocation.
As such, This patch changes the "userdata" field to a "datatype"
field, and adds a "no busy reuse" flag to the field.
When we detect an unordered data extent allocation, we immediately set
the no reuse flag. We then set the "user data" flags based on the
inode fork we are allocating the extent to. Hence we only set
userdata flags on data fork allocations now and consider attribute
fork remote extents to be an unordered metadata extent.
The result is that remote attribute extents now have the expected
allocation semantics, and the data fork allocation behaviour is
completely unchanged.
It should be noted that there may be other ways to fix this (e.g.
use ordered metadata buffers for the remote attribute extent data
write) but they are more invasive and difficult to validate both
from a design and implementation POV. Hence this patch takes the
simple, obvious route to fixing the problem...
Reported-and-tested-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-25 22:21:28 +00:00
|
|
|
args.datatype = ap->datatype;
|
|
|
|
if (ap->datatype & XFS_ALLOC_USERDATA_ZERO)
|
2015-11-03 01:27:22 +00:00
|
|
|
args.ip = ap->ip;
|
|
|
|
|
|
|
|
error = xfs_alloc_vextent(&args);
|
|
|
|
if (error)
|
2013-02-25 01:31:26 +00:00
|
|
|
return error;
|
2015-11-03 01:27:22 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (tryagain && args.fsbno == NULLFSBLOCK) {
|
|
|
|
/*
|
|
|
|
* Exact allocation failed. Now try with alignment
|
|
|
|
* turned on.
|
|
|
|
*/
|
|
|
|
args.type = atype;
|
|
|
|
args.fsbno = ap->blkno;
|
2013-12-12 05:34:36 +00:00
|
|
|
args.alignment = stripe_align;
|
2013-02-25 01:31:26 +00:00
|
|
|
args.minlen = nextminlen;
|
|
|
|
args.minalignslop = 0;
|
|
|
|
isaligned = 1;
|
|
|
|
if ((error = xfs_alloc_vextent(&args)))
|
|
|
|
return error;
|
2011-09-18 20:40:50 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
if (isaligned && args.fsbno == NULLFSBLOCK) {
|
|
|
|
/*
|
|
|
|
* allocation failed, so turn off alignment and
|
|
|
|
* try again.
|
|
|
|
*/
|
|
|
|
args.type = atype;
|
|
|
|
args.fsbno = ap->blkno;
|
|
|
|
args.alignment = 0;
|
|
|
|
if ((error = xfs_alloc_vextent(&args)))
|
2011-09-18 20:40:50 +00:00
|
|
|
return error;
|
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
if (args.fsbno == NULLFSBLOCK && nullfb &&
|
|
|
|
args.minlen > ap->minlen) {
|
|
|
|
args.minlen = ap->minlen;
|
|
|
|
args.type = XFS_ALLOCTYPE_START_BNO;
|
|
|
|
args.fsbno = ap->blkno;
|
|
|
|
if ((error = xfs_alloc_vextent(&args)))
|
|
|
|
return error;
|
2011-09-18 20:40:50 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
if (args.fsbno == NULLFSBLOCK && nullfb) {
|
|
|
|
args.fsbno = 0;
|
|
|
|
args.type = XFS_ALLOCTYPE_FIRST_AG;
|
|
|
|
args.total = ap->minlen;
|
|
|
|
if ((error = xfs_alloc_vextent(&args)))
|
|
|
|
return error;
|
2016-08-03 01:19:29 +00:00
|
|
|
ap->dfops->dop_low = true;
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
|
|
|
if (args.fsbno != NULLFSBLOCK) {
|
|
|
|
/*
|
|
|
|
* check the allocation happened at the same or higher AG than
|
|
|
|
* the first block that was allocated.
|
|
|
|
*/
|
|
|
|
ASSERT(*ap->firstblock == NULLFSBLOCK ||
|
2017-02-17 01:12:51 +00:00
|
|
|
XFS_FSB_TO_AGNO(mp, *ap->firstblock) <=
|
|
|
|
XFS_FSB_TO_AGNO(mp, args.fsbno));
|
2011-09-18 20:40:50 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
ap->blkno = args.fsbno;
|
|
|
|
if (*ap->firstblock == NULLFSBLOCK)
|
|
|
|
*ap->firstblock = args.fsbno;
|
2017-02-17 01:12:51 +00:00
|
|
|
ASSERT(nullfb || fb_agno <= args.agno);
|
2013-02-25 01:31:26 +00:00
|
|
|
ap->length = args.len;
|
2016-10-03 16:11:34 +00:00
|
|
|
if (!(ap->flags & XFS_BMAPI_COWFORK))
|
|
|
|
ap->ip->i_d.di_nblocks += args.len;
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_trans_log_inode(ap->tp, ap->ip, XFS_ILOG_CORE);
|
|
|
|
if (ap->wasdel)
|
|
|
|
ap->ip->i_delayed_blks -= args.len;
|
|
|
|
/*
|
|
|
|
* Adjust the disk quota also. This was reserved
|
|
|
|
* earlier.
|
|
|
|
*/
|
|
|
|
xfs_trans_mod_dquot_byino(ap->tp, ap->ip,
|
|
|
|
ap->wasdel ? XFS_TRANS_DQ_DELBCOUNT :
|
|
|
|
XFS_TRANS_DQ_BCOUNT,
|
|
|
|
(long) args.len);
|
|
|
|
} else {
|
|
|
|
ap->blkno = NULLFSBLOCK;
|
|
|
|
ap->length = 0;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
2011-09-18 20:40:50 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* xfs_bmap_alloc is called by xfs_bmapi to allocate an extent for a file.
|
|
|
|
* It figures out where to ask the underlying allocator to put the new extent.
|
|
|
|
*/
|
|
|
|
STATIC int
|
|
|
|
xfs_bmap_alloc(
|
2013-08-12 10:49:42 +00:00
|
|
|
struct xfs_bmalloca *ap) /* bmap alloc argument struct */
|
2013-02-25 01:31:26 +00:00
|
|
|
{
|
xfs: remote attribute blocks aren't really userdata
When adding a new remote attribute, we write the attribute to the
new extent before the allocation transaction is committed. This
means we cannot reuse busy extents as that violates crash
consistency semantics. Hence we currently treat remote attribute
extent allocation like userdata because it has the same overwrite
ordering constraints as userdata.
Unfortunately, this also allows the allocator to incorrectly apply
extent size hints to the remote attribute extent allocation. This
results in interesting failures, such as transaction block
reservation overruns and in-memory inode attribute fork corruption.
To fix this, we need to separate the busy extent reuse configuration
from the userdata configuration. This changes the definition of
XFS_BMAPI_METADATA slightly - it now means that allocation is
metadata and reuse of busy extents is acceptible due to the metadata
ordering semantics of the journal. If this flag is not set, it
means the allocation is that has unordered data writeback, and hence
busy extent reuse is not allowed. It no longer implies the
allocation is for user data, just that the data write will not be
strictly ordered. This matches the semantics for both user data
and remote attribute block allocation.
As such, This patch changes the "userdata" field to a "datatype"
field, and adds a "no busy reuse" flag to the field.
When we detect an unordered data extent allocation, we immediately set
the no reuse flag. We then set the "user data" flags based on the
inode fork we are allocating the extent to. Hence we only set
userdata flags on data fork allocations now and consider attribute
fork remote extents to be an unordered metadata extent.
The result is that remote attribute extents now have the expected
allocation semantics, and the data fork allocation behaviour is
completely unchanged.
It should be noted that there may be other ways to fix this (e.g.
use ordered metadata buffers for the remote attribute extent data
write) but they are more invasive and difficult to validate both
from a design and implementation POV. Hence this patch takes the
simple, obvious route to fixing the problem...
Reported-and-tested-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-25 22:21:28 +00:00
|
|
|
if (XFS_IS_REALTIME_INODE(ap->ip) &&
|
|
|
|
xfs_alloc_is_userdata(ap->datatype))
|
2013-02-25 01:31:26 +00:00
|
|
|
return xfs_bmap_rtalloc(ap);
|
|
|
|
return xfs_bmap_btalloc(ap);
|
|
|
|
}
|
2011-09-18 20:40:54 +00:00
|
|
|
|
2016-10-20 04:51:50 +00:00
|
|
|
/* Trim extent to fit a logical block range. */
|
|
|
|
void
|
|
|
|
xfs_trim_extent(
|
|
|
|
struct xfs_bmbt_irec *irec,
|
|
|
|
xfs_fileoff_t bno,
|
|
|
|
xfs_filblks_t len)
|
|
|
|
{
|
|
|
|
xfs_fileoff_t distance;
|
|
|
|
xfs_fileoff_t end = bno + len;
|
|
|
|
|
|
|
|
if (irec->br_startoff + irec->br_blockcount <= bno ||
|
|
|
|
irec->br_startoff >= end) {
|
|
|
|
irec->br_blockcount = 0;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (irec->br_startoff < bno) {
|
|
|
|
distance = bno - irec->br_startoff;
|
|
|
|
if (isnullstartblock(irec->br_startblock))
|
|
|
|
irec->br_startblock = DELAYSTARTBLOCK;
|
|
|
|
if (irec->br_startblock != DELAYSTARTBLOCK &&
|
|
|
|
irec->br_startblock != HOLESTARTBLOCK)
|
|
|
|
irec->br_startblock += distance;
|
|
|
|
irec->br_startoff += distance;
|
|
|
|
irec->br_blockcount -= distance;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (end < irec->br_startoff + irec->br_blockcount) {
|
|
|
|
distance = irec->br_startoff + irec->br_blockcount - end;
|
|
|
|
irec->br_blockcount -= distance;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Trim the returned map to the required bounds
|
|
|
|
*/
|
|
|
|
STATIC void
|
|
|
|
xfs_bmapi_trim_map(
|
|
|
|
struct xfs_bmbt_irec *mval,
|
|
|
|
struct xfs_bmbt_irec *got,
|
|
|
|
xfs_fileoff_t *bno,
|
|
|
|
xfs_filblks_t len,
|
|
|
|
xfs_fileoff_t obno,
|
|
|
|
xfs_fileoff_t end,
|
|
|
|
int n,
|
|
|
|
int flags)
|
|
|
|
{
|
|
|
|
if ((flags & XFS_BMAPI_ENTIRE) ||
|
|
|
|
got->br_startoff + got->br_blockcount <= obno) {
|
|
|
|
*mval = *got;
|
|
|
|
if (isnullstartblock(got->br_startblock))
|
|
|
|
mval->br_startblock = DELAYSTARTBLOCK;
|
|
|
|
return;
|
|
|
|
}
|
2011-09-18 20:40:50 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (obno > *bno)
|
|
|
|
*bno = obno;
|
|
|
|
ASSERT((*bno >= obno) || (n == 0));
|
|
|
|
ASSERT(*bno < end);
|
|
|
|
mval->br_startoff = *bno;
|
|
|
|
if (isnullstartblock(got->br_startblock))
|
|
|
|
mval->br_startblock = DELAYSTARTBLOCK;
|
|
|
|
else
|
|
|
|
mval->br_startblock = got->br_startblock +
|
|
|
|
(*bno - got->br_startoff);
|
2011-09-18 20:40:50 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Return the minimum of what we got and what we asked for for
|
|
|
|
* the length. We can use the len variable here because it is
|
|
|
|
* modified below and we could have been there before coming
|
|
|
|
* here if the first part of the allocation didn't overlap what
|
|
|
|
* was asked for.
|
2011-09-18 20:40:50 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
mval->br_blockcount = XFS_FILBLKS_MIN(end - *bno,
|
|
|
|
got->br_blockcount - (*bno - got->br_startoff));
|
|
|
|
mval->br_state = got->br_state;
|
|
|
|
ASSERT(mval->br_blockcount <= len);
|
|
|
|
return;
|
2011-09-18 20:40:50 +00:00
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Update and validate the extent map to return
|
|
|
|
*/
|
|
|
|
STATIC void
|
|
|
|
xfs_bmapi_update_map(
|
|
|
|
struct xfs_bmbt_irec **map,
|
|
|
|
xfs_fileoff_t *bno,
|
|
|
|
xfs_filblks_t *len,
|
|
|
|
xfs_fileoff_t obno,
|
|
|
|
xfs_fileoff_t end,
|
|
|
|
int *n,
|
|
|
|
int flags)
|
2012-10-05 01:06:59 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmbt_irec_t *mval = *map;
|
2012-10-05 01:06:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT((flags & XFS_BMAPI_ENTIRE) ||
|
|
|
|
((mval->br_startoff + mval->br_blockcount) <= end));
|
|
|
|
ASSERT((flags & XFS_BMAPI_ENTIRE) || (mval->br_blockcount <= *len) ||
|
|
|
|
(mval->br_startoff < obno));
|
2012-10-05 01:06:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
*bno = mval->br_startoff + mval->br_blockcount;
|
|
|
|
*len = end - *bno;
|
|
|
|
if (*n > 0 && mval->br_startoff == mval[-1].br_startoff) {
|
|
|
|
/* update previous map with new information */
|
|
|
|
ASSERT(mval->br_startblock == mval[-1].br_startblock);
|
|
|
|
ASSERT(mval->br_blockcount > mval[-1].br_blockcount);
|
|
|
|
ASSERT(mval->br_state == mval[-1].br_state);
|
|
|
|
mval[-1].br_blockcount = mval->br_blockcount;
|
|
|
|
mval[-1].br_state = mval->br_state;
|
|
|
|
} else if (*n > 0 && mval->br_startblock != DELAYSTARTBLOCK &&
|
|
|
|
mval[-1].br_startblock != DELAYSTARTBLOCK &&
|
|
|
|
mval[-1].br_startblock != HOLESTARTBLOCK &&
|
|
|
|
mval->br_startblock == mval[-1].br_startblock +
|
|
|
|
mval[-1].br_blockcount &&
|
|
|
|
((flags & XFS_BMAPI_IGSTATE) ||
|
|
|
|
mval[-1].br_state == mval->br_state)) {
|
|
|
|
ASSERT(mval->br_startoff ==
|
|
|
|
mval[-1].br_startoff + mval[-1].br_blockcount);
|
|
|
|
mval[-1].br_blockcount += mval->br_blockcount;
|
|
|
|
} else if (*n > 0 &&
|
|
|
|
mval->br_startblock == DELAYSTARTBLOCK &&
|
|
|
|
mval[-1].br_startblock == DELAYSTARTBLOCK &&
|
|
|
|
mval->br_startoff ==
|
|
|
|
mval[-1].br_startoff + mval[-1].br_blockcount) {
|
|
|
|
mval[-1].br_blockcount += mval->br_blockcount;
|
|
|
|
mval[-1].br_state = mval->br_state;
|
|
|
|
} else if (!((*n == 0) &&
|
|
|
|
((mval->br_startoff + mval->br_blockcount) <=
|
|
|
|
obno))) {
|
|
|
|
mval++;
|
|
|
|
(*n)++;
|
|
|
|
}
|
|
|
|
*map = mval;
|
2012-10-05 01:06:59 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Map file blocks to filesystem blocks without allocation.
|
2012-10-05 01:06:59 +00:00
|
|
|
*/
|
|
|
|
int
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmapi_read(
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
xfs_fileoff_t bno,
|
2011-09-18 20:40:51 +00:00
|
|
|
xfs_filblks_t len,
|
2013-02-25 01:31:26 +00:00
|
|
|
struct xfs_bmbt_irec *mval,
|
|
|
|
int *nmap,
|
2011-09-18 20:41:02 +00:00
|
|
|
int flags)
|
2011-09-18 20:40:51 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
|
|
|
struct xfs_ifork *ifp;
|
|
|
|
struct xfs_bmbt_irec got;
|
|
|
|
xfs_fileoff_t obno;
|
|
|
|
xfs_fileoff_t end;
|
2016-11-24 00:39:43 +00:00
|
|
|
xfs_extnum_t idx;
|
2013-02-25 01:31:26 +00:00
|
|
|
int error;
|
2016-11-24 00:39:43 +00:00
|
|
|
bool eof = false;
|
2013-02-25 01:31:26 +00:00
|
|
|
int n = 0;
|
2016-10-03 16:11:32 +00:00
|
|
|
int whichfork = xfs_bmapi_whichfork(flags);
|
2011-09-18 20:40:51 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(*nmap >= 1);
|
|
|
|
ASSERT(!(flags & ~(XFS_BMAPI_ATTRFORK|XFS_BMAPI_ENTIRE|
|
2016-10-03 16:11:32 +00:00
|
|
|
XFS_BMAPI_IGSTATE|XFS_BMAPI_COWFORK)));
|
2013-12-06 20:30:17 +00:00
|
|
|
ASSERT(xfs_isilocked(ip, XFS_ILOCK_SHARED|XFS_ILOCK_EXCL));
|
2011-09-18 20:40:51 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (unlikely(XFS_TEST_ERROR(
|
|
|
|
(XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS &&
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE),
|
2017-06-21 00:54:47 +00:00
|
|
|
mp, XFS_ERRTAG_BMAPIFORMAT))) {
|
2013-02-25 01:31:26 +00:00
|
|
|
XFS_ERROR_REPORT("xfs_bmapi_read", XFS_ERRLEVEL_LOW, mp);
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EFSCORRUPTED;
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
2011-09-18 20:40:51 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (XFS_FORCED_SHUTDOWN(mp))
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EIO;
|
2013-02-25 01:31:26 +00:00
|
|
|
|
2015-10-12 07:21:22 +00:00
|
|
|
XFS_STATS_INC(mp, xs_blk_mapr);
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
|
2016-10-03 16:11:32 +00:00
|
|
|
/* No CoW fork? Return a hole. */
|
|
|
|
if (whichfork == XFS_COW_FORK && !ifp) {
|
|
|
|
mval->br_startoff = bno;
|
|
|
|
mval->br_startblock = HOLESTARTBLOCK;
|
|
|
|
mval->br_blockcount = len;
|
|
|
|
mval->br_state = XFS_EXT_NORM;
|
|
|
|
*nmap = 1;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (!(ifp->if_flags & XFS_IFEXTENTS)) {
|
|
|
|
error = xfs_iread_extents(NULL, ip, whichfork);
|
|
|
|
if (error)
|
|
|
|
return error;
|
2011-09-18 20:40:51 +00:00
|
|
|
}
|
|
|
|
|
2016-11-24 00:39:43 +00:00
|
|
|
if (!xfs_iext_lookup_extent(ip, ifp, bno, &idx, &got))
|
|
|
|
eof = true;
|
2013-02-25 01:31:26 +00:00
|
|
|
end = bno + len;
|
|
|
|
obno = bno;
|
2011-09-18 20:40:51 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
while (bno < end && n < *nmap) {
|
|
|
|
/* Reading past eof, act as though there's a hole up to end. */
|
|
|
|
if (eof)
|
|
|
|
got.br_startoff = end;
|
|
|
|
if (got.br_startoff > bno) {
|
|
|
|
/* Reading in a hole. */
|
|
|
|
mval->br_startoff = bno;
|
|
|
|
mval->br_startblock = HOLESTARTBLOCK;
|
|
|
|
mval->br_blockcount =
|
|
|
|
XFS_FILBLKS_MIN(len, got.br_startoff - bno);
|
|
|
|
mval->br_state = XFS_EXT_NORM;
|
|
|
|
bno += mval->br_blockcount;
|
|
|
|
len -= mval->br_blockcount;
|
|
|
|
mval++;
|
|
|
|
n++;
|
|
|
|
continue;
|
|
|
|
}
|
2011-09-18 20:40:51 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/* set up the extent map to return. */
|
|
|
|
xfs_bmapi_trim_map(mval, &got, &bno, len, obno, end, n, flags);
|
|
|
|
xfs_bmapi_update_map(&mval, &bno, &len, obno, end, &n, flags);
|
|
|
|
|
|
|
|
/* If we're done, stop now. */
|
|
|
|
if (bno >= end || n >= *nmap)
|
|
|
|
break;
|
|
|
|
|
|
|
|
/* Else go on to the next record. */
|
2016-11-24 00:39:43 +00:00
|
|
|
if (!xfs_iext_get_extent(ifp, ++idx, &got))
|
|
|
|
eof = true;
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
|
|
|
*nmap = n;
|
2011-09-18 20:40:51 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-03-08 17:58:08 +00:00
|
|
|
/*
|
|
|
|
* Add a delayed allocation extent to an inode. Blocks are reserved from the
|
|
|
|
* global pool and the extent inserted into the inode in-core extent tree.
|
|
|
|
*
|
|
|
|
* On entry, got refers to the first extent beyond the offset of the extent to
|
|
|
|
* allocate or eof is specified if no such extent exists. On return, got refers
|
|
|
|
* to the extent record that was inserted to the inode fork.
|
|
|
|
*
|
|
|
|
* Note that the allocated extent may have been merged with contiguous extents
|
|
|
|
* during insertion into the inode fork. Thus, got does not reflect the current
|
|
|
|
* state of the inode fork on return. If necessary, the caller can use lastx to
|
|
|
|
* look up the updated record in the inode fork.
|
|
|
|
*/
|
2016-09-19 01:10:21 +00:00
|
|
|
int
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmapi_reserve_delalloc(
|
|
|
|
struct xfs_inode *ip,
|
2016-10-03 16:11:32 +00:00
|
|
|
int whichfork,
|
2016-11-28 03:57:42 +00:00
|
|
|
xfs_fileoff_t off,
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_filblks_t len,
|
2016-11-28 03:57:42 +00:00
|
|
|
xfs_filblks_t prealloc,
|
2013-02-25 01:31:26 +00:00
|
|
|
struct xfs_bmbt_irec *got,
|
|
|
|
xfs_extnum_t *lastx,
|
|
|
|
int eof)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2011-09-18 20:40:52 +00:00
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
2016-10-03 16:11:32 +00:00
|
|
|
struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, whichfork);
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_extlen_t alen;
|
|
|
|
xfs_extlen_t indlen;
|
|
|
|
char rt = XFS_IS_REALTIME_INODE(ip);
|
|
|
|
xfs_extlen_t extsz;
|
|
|
|
int error;
|
2016-11-28 03:57:42 +00:00
|
|
|
xfs_fileoff_t aoff = off;
|
2011-09-18 20:40:52 +00:00
|
|
|
|
2016-11-28 03:57:42 +00:00
|
|
|
/*
|
|
|
|
* Cap the alloc length. Keep track of prealloc so we know whether to
|
|
|
|
* tag the inode before we return.
|
|
|
|
*/
|
|
|
|
alen = XFS_FILBLKS_MIN(len + prealloc, MAXEXTLEN);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (!eof)
|
|
|
|
alen = XFS_FILBLKS_MIN(alen, got->br_startoff - aoff);
|
2016-11-28 03:57:42 +00:00
|
|
|
if (prealloc && alen >= len)
|
|
|
|
prealloc = alen - len;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/* Figure out the extent size, adjust alen */
|
2016-10-03 16:11:43 +00:00
|
|
|
if (whichfork == XFS_COW_FORK)
|
|
|
|
extsz = xfs_get_cowextsz_hint(ip);
|
|
|
|
else
|
|
|
|
extsz = xfs_get_extsz_hint(ip);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (extsz) {
|
2016-11-24 00:39:44 +00:00
|
|
|
struct xfs_bmbt_irec prev;
|
|
|
|
|
|
|
|
if (!xfs_iext_get_extent(ifp, *lastx - 1, &prev))
|
|
|
|
prev.br_startoff = NULLFILEOFF;
|
|
|
|
|
|
|
|
error = xfs_bmap_extsize_align(mp, got, &prev, extsz, rt, eof,
|
2013-02-25 01:31:26 +00:00
|
|
|
1, 0, &aoff, &alen);
|
|
|
|
ASSERT(!error);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (rt)
|
|
|
|
extsz = alen / mp->m_sb.sb_rextsize;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Make a transaction-less quota reservation for delayed allocation
|
|
|
|
* blocks. This number gets adjusted later. We return if we haven't
|
|
|
|
* allocated blocks already inside this loop.
|
|
|
|
*/
|
|
|
|
error = xfs_trans_reserve_quota_nblks(NULL, ip, (long)alen, 0,
|
|
|
|
rt ? XFS_QMOPT_RES_RTBLKS : XFS_QMOPT_RES_REGBLKS);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Split changing sb for alen and indlen since they could be coming
|
|
|
|
* from different places.
|
|
|
|
*/
|
|
|
|
indlen = (xfs_extlen_t)xfs_bmap_worst_indlen(ip, alen);
|
|
|
|
ASSERT(indlen > 0);
|
|
|
|
|
|
|
|
if (rt) {
|
2015-02-23 10:22:54 +00:00
|
|
|
error = xfs_mod_frextents(mp, -((int64_t)extsz));
|
2013-02-25 01:31:26 +00:00
|
|
|
} else {
|
2015-02-23 10:22:03 +00:00
|
|
|
error = xfs_mod_fdblocks(mp, -((int64_t)alen), false);
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (error)
|
|
|
|
goto out_unreserve_quota;
|
|
|
|
|
2015-02-23 10:22:03 +00:00
|
|
|
error = xfs_mod_fdblocks(mp, -((int64_t)indlen), false);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (error)
|
|
|
|
goto out_unreserve_blocks;
|
|
|
|
|
|
|
|
|
|
|
|
ip->i_delayed_blks += alen;
|
|
|
|
|
|
|
|
got->br_startoff = aoff;
|
|
|
|
got->br_startblock = nullstartblock(indlen);
|
|
|
|
got->br_blockcount = alen;
|
|
|
|
got->br_state = XFS_EXT_NORM;
|
|
|
|
|
2017-03-08 17:58:08 +00:00
|
|
|
xfs_bmap_add_extent_hole_delay(ip, whichfork, lastx, got);
|
2013-02-25 01:31:26 +00:00
|
|
|
|
2016-11-28 03:57:42 +00:00
|
|
|
/*
|
|
|
|
* Tag the inode if blocks were preallocated. Note that COW fork
|
|
|
|
* preallocation can occur at the start or end of the extent, even when
|
|
|
|
* prealloc == 0, so we must also check the aligned offset and length.
|
|
|
|
*/
|
|
|
|
if (whichfork == XFS_DATA_FORK && prealloc)
|
|
|
|
xfs_inode_set_eofblocks_tag(ip);
|
|
|
|
if (whichfork == XFS_COW_FORK && (prealloc || aoff < off || alen > len))
|
|
|
|
xfs_inode_set_cowblocks_tag(ip);
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
out_unreserve_blocks:
|
|
|
|
if (rt)
|
2015-02-23 10:22:54 +00:00
|
|
|
xfs_mod_frextents(mp, extsz);
|
2013-02-25 01:31:26 +00:00
|
|
|
else
|
2015-02-23 10:22:03 +00:00
|
|
|
xfs_mod_fdblocks(mp, alen, false);
|
2013-02-25 01:31:26 +00:00
|
|
|
out_unreserve_quota:
|
|
|
|
if (XFS_IS_QUOTA_ON(mp))
|
|
|
|
xfs_trans_unreserve_quota_nblks(NULL, ip, (long)alen, 0, rt ?
|
|
|
|
XFS_QMOPT_RES_RTBLKS : XFS_QMOPT_RES_REGBLKS);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
xfs: refine the allocation stack switch
The allocation stack switch at xfs_bmapi_allocate() has served it's
purpose, but is no longer a sufficient solution to the stack usage
problem we have in the XFS allocation path.
Whilst the kernel stack size is now 16k, that is not a valid reason
for undoing all our "keep stack usage down" modifications. What it
does allow us to do is have the freedom to refine and perfect the
modifications knowing that if we get it wrong it won't blow up in
our faces - we have a safety net now.
This is important because we still have the issue of older kernels
having smaller stacks and that they are still supported and are
demonstrating a wide range of different stack overflows. Red Hat
has several open bugs for allocation based stack overflows from
directory modifications and direct IO block allocation and these
problems still need to be solved. If we can solve them upstream,
then distro's won't need to bake their own unique solutions.
To that end, I've observed that every allocation based stack
overflow report has had a specific characteristic - it has happened
during or directly after a bmap btree block split. That event
requires a new block to be allocated to the tree, and so we
effectively stack one allocation stack on top of another, and that's
when we get into trouble.
A further observation is that bmap btree block splits are much rarer
than writeback allocation - over a range of different workloads I've
observed the ratio of bmap btree inserts to splits ranges from 100:1
(xfstests run) to 10000:1 (local VM image server with sparse files
that range in the hundreds of thousands to millions of extents).
Either way, bmap btree split events are much, much rarer than
allocation events.
Finally, we have to move the kswapd state to the allocation workqueue
work when allocation is done on behalf of kswapd. This is proving to
cause significant perturbation in performance under memory pressure
and appears to be generating allocation deadlock warnings under some
workloads, so avoiding the use of a workqueue for the majority of
kswapd writeback allocation will minimise the impact of such
behaviour.
Hence it makes sense to move the stack switch to xfs_btree_split()
and only do it for bmap btree splits. Stack switches during
allocation will be much rarer, so there won't be significant
performacne overhead caused by switching stacks. The worse case
stack from all allocation paths will be split, not just writeback.
And the majority of memory allocations will be done in the correct
context (e.g. kswapd) without causing additional latency, and so we
simplify the memory reclaim interactions between processes,
workqueues and kswapd.
The worst stack I've been able to generate with this patch in place
is 5600 bytes deep. It's very revealing because we exit XFS at:
37) 1768 64 kmem_cache_alloc+0x13b/0x170
about 1800 bytes of stack consumed, and the remaining 3800 bytes
(and 36 functions) is memory reclaim, swap and the IO stack. And
this occurs in the inode allocation from an open(O_CREAT) syscall,
not writeback.
The amount of stack being used is much less than I've previously be
able to generate - fs_mark testing has been able to generate stack
usage of around 7k without too much trouble; with this patch it's
only just getting to 5.5k. This is primarily because the metadata
allocation paths (e.g. directory blocks) are no longer causing
double splits on the same stack, and hence now stack tracing is
showing swapping being the worst stack consumer rather than XFS.
Performance of fs_mark inode create workloads is unchanged.
Performance of fs_mark async fsync workloads is consistently good
with context switches reduced by around 150,000/s (30%).
Performance of dbench, streaming IO and postmark is unchanged.
Allocation deadlock warnings have not been seen on the workloads
that generated them since adding this patch.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-07-14 21:08:24 +00:00
|
|
|
static int
|
|
|
|
xfs_bmapi_allocate(
|
2013-02-25 01:31:26 +00:00
|
|
|
struct xfs_bmalloca *bma)
|
|
|
|
{
|
|
|
|
struct xfs_mount *mp = bma->ip->i_mount;
|
2016-10-03 16:11:34 +00:00
|
|
|
int whichfork = xfs_bmapi_whichfork(bma->flags);
|
2013-02-25 01:31:26 +00:00
|
|
|
struct xfs_ifork *ifp = XFS_IFORK_PTR(bma->ip, whichfork);
|
|
|
|
int tmp_logflags = 0;
|
|
|
|
int error;
|
|
|
|
|
|
|
|
ASSERT(bma->length > 0);
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* For the wasdelay case, we could also just allocate the stuff asked
|
|
|
|
* for in this bmap call but that wouldn't be as good.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (bma->wasdel) {
|
|
|
|
bma->length = (xfs_extlen_t)bma->got.br_blockcount;
|
|
|
|
bma->offset = bma->got.br_startoff;
|
2016-11-24 00:40:32 +00:00
|
|
|
if (bma->idx) {
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(ifp, bma->idx - 1),
|
|
|
|
&bma->prev);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
} else {
|
|
|
|
bma->length = XFS_FILBLKS_MIN(bma->length, MAXEXTLEN);
|
|
|
|
if (!bma->eof)
|
|
|
|
bma->length = XFS_FILBLKS_MIN(bma->length,
|
|
|
|
bma->got.br_startoff - bma->offset);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
xfs: remote attribute blocks aren't really userdata
When adding a new remote attribute, we write the attribute to the
new extent before the allocation transaction is committed. This
means we cannot reuse busy extents as that violates crash
consistency semantics. Hence we currently treat remote attribute
extent allocation like userdata because it has the same overwrite
ordering constraints as userdata.
Unfortunately, this also allows the allocator to incorrectly apply
extent size hints to the remote attribute extent allocation. This
results in interesting failures, such as transaction block
reservation overruns and in-memory inode attribute fork corruption.
To fix this, we need to separate the busy extent reuse configuration
from the userdata configuration. This changes the definition of
XFS_BMAPI_METADATA slightly - it now means that allocation is
metadata and reuse of busy extents is acceptible due to the metadata
ordering semantics of the journal. If this flag is not set, it
means the allocation is that has unordered data writeback, and hence
busy extent reuse is not allowed. It no longer implies the
allocation is for user data, just that the data write will not be
strictly ordered. This matches the semantics for both user data
and remote attribute block allocation.
As such, This patch changes the "userdata" field to a "datatype"
field, and adds a "no busy reuse" flag to the field.
When we detect an unordered data extent allocation, we immediately set
the no reuse flag. We then set the "user data" flags based on the
inode fork we are allocating the extent to. Hence we only set
userdata flags on data fork allocations now and consider attribute
fork remote extents to be an unordered metadata extent.
The result is that remote attribute extents now have the expected
allocation semantics, and the data fork allocation behaviour is
completely unchanged.
It should be noted that there may be other ways to fix this (e.g.
use ordered metadata buffers for the remote attribute extent data
write) but they are more invasive and difficult to validate both
from a design and implementation POV. Hence this patch takes the
simple, obvious route to fixing the problem...
Reported-and-tested-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-25 22:21:28 +00:00
|
|
|
* Set the data type being allocated. For the data fork, the first data
|
|
|
|
* in the file is treated differently to all other allocations. For the
|
|
|
|
* attribute fork, we only need to ensure the allocated range is not on
|
|
|
|
* the busy list.
|
2013-02-25 01:31:26 +00:00
|
|
|
*/
|
|
|
|
if (!(bma->flags & XFS_BMAPI_METADATA)) {
|
xfs: remote attribute blocks aren't really userdata
When adding a new remote attribute, we write the attribute to the
new extent before the allocation transaction is committed. This
means we cannot reuse busy extents as that violates crash
consistency semantics. Hence we currently treat remote attribute
extent allocation like userdata because it has the same overwrite
ordering constraints as userdata.
Unfortunately, this also allows the allocator to incorrectly apply
extent size hints to the remote attribute extent allocation. This
results in interesting failures, such as transaction block
reservation overruns and in-memory inode attribute fork corruption.
To fix this, we need to separate the busy extent reuse configuration
from the userdata configuration. This changes the definition of
XFS_BMAPI_METADATA slightly - it now means that allocation is
metadata and reuse of busy extents is acceptible due to the metadata
ordering semantics of the journal. If this flag is not set, it
means the allocation is that has unordered data writeback, and hence
busy extent reuse is not allowed. It no longer implies the
allocation is for user data, just that the data write will not be
strictly ordered. This matches the semantics for both user data
and remote attribute block allocation.
As such, This patch changes the "userdata" field to a "datatype"
field, and adds a "no busy reuse" flag to the field.
When we detect an unordered data extent allocation, we immediately set
the no reuse flag. We then set the "user data" flags based on the
inode fork we are allocating the extent to. Hence we only set
userdata flags on data fork allocations now and consider attribute
fork remote extents to be an unordered metadata extent.
The result is that remote attribute extents now have the expected
allocation semantics, and the data fork allocation behaviour is
completely unchanged.
It should be noted that there may be other ways to fix this (e.g.
use ordered metadata buffers for the remote attribute extent data
write) but they are more invasive and difficult to validate both
from a design and implementation POV. Hence this patch takes the
simple, obvious route to fixing the problem...
Reported-and-tested-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-25 22:21:28 +00:00
|
|
|
bma->datatype = XFS_ALLOC_NOBUSY;
|
|
|
|
if (whichfork == XFS_DATA_FORK) {
|
|
|
|
if (bma->offset == 0)
|
|
|
|
bma->datatype |= XFS_ALLOC_INITIAL_USER_DATA;
|
|
|
|
else
|
|
|
|
bma->datatype |= XFS_ALLOC_USERDATA;
|
|
|
|
}
|
2015-11-03 01:27:22 +00:00
|
|
|
if (bma->flags & XFS_BMAPI_ZERO)
|
xfs: remote attribute blocks aren't really userdata
When adding a new remote attribute, we write the attribute to the
new extent before the allocation transaction is committed. This
means we cannot reuse busy extents as that violates crash
consistency semantics. Hence we currently treat remote attribute
extent allocation like userdata because it has the same overwrite
ordering constraints as userdata.
Unfortunately, this also allows the allocator to incorrectly apply
extent size hints to the remote attribute extent allocation. This
results in interesting failures, such as transaction block
reservation overruns and in-memory inode attribute fork corruption.
To fix this, we need to separate the busy extent reuse configuration
from the userdata configuration. This changes the definition of
XFS_BMAPI_METADATA slightly - it now means that allocation is
metadata and reuse of busy extents is acceptible due to the metadata
ordering semantics of the journal. If this flag is not set, it
means the allocation is that has unordered data writeback, and hence
busy extent reuse is not allowed. It no longer implies the
allocation is for user data, just that the data write will not be
strictly ordered. This matches the semantics for both user data
and remote attribute block allocation.
As such, This patch changes the "userdata" field to a "datatype"
field, and adds a "no busy reuse" flag to the field.
When we detect an unordered data extent allocation, we immediately set
the no reuse flag. We then set the "user data" flags based on the
inode fork we are allocating the extent to. Hence we only set
userdata flags on data fork allocations now and consider attribute
fork remote extents to be an unordered metadata extent.
The result is that remote attribute extents now have the expected
allocation semantics, and the data fork allocation behaviour is
completely unchanged.
It should be noted that there may be other ways to fix this (e.g.
use ordered metadata buffers for the remote attribute extent data
write) but they are more invasive and difficult to validate both
from a design and implementation POV. Hence this patch takes the
simple, obvious route to fixing the problem...
Reported-and-tested-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-25 22:21:28 +00:00
|
|
|
bma->datatype |= XFS_ALLOC_USERDATA_ZERO;
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
bma->minlen = (bma->flags & XFS_BMAPI_CONTIG) ? bma->length : 1;
|
2009-12-14 23:14:59 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Only want to do the alignment at the eof if it is userdata and
|
|
|
|
* allocation length is larger than a stripe unit.
|
|
|
|
*/
|
|
|
|
if (mp->m_dalign && bma->length >= mp->m_dalign &&
|
|
|
|
!(bma->flags & XFS_BMAPI_METADATA) && whichfork == XFS_DATA_FORK) {
|
|
|
|
error = xfs_bmap_isaeof(bma, whichfork);
|
|
|
|
if (error)
|
|
|
|
return error;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2011-12-18 20:00:07 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
error = xfs_bmap_alloc(bma);
|
|
|
|
if (error)
|
2005-04-16 22:20:36 +00:00
|
|
|
return error;
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
if (bma->cur)
|
|
|
|
bma->cur->bc_private.b.firstblock = *bma->firstblock;
|
|
|
|
if (bma->blkno == NULLFSBLOCK)
|
2005-04-16 22:20:36 +00:00
|
|
|
return 0;
|
2013-02-25 01:31:26 +00:00
|
|
|
if ((ifp->if_flags & XFS_IFBROOT) && !bma->cur) {
|
|
|
|
bma->cur = xfs_bmbt_init_cursor(mp, bma->tp, bma->ip, whichfork);
|
|
|
|
bma->cur->bc_private.b.firstblock = *bma->firstblock;
|
2016-08-03 01:19:29 +00:00
|
|
|
bma->cur->bc_private.b.dfops = bma->dfops;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Bump the number of extents we've allocated
|
|
|
|
* in this call.
|
|
|
|
*/
|
|
|
|
bma->nallocs++;
|
|
|
|
|
|
|
|
if (bma->cur)
|
|
|
|
bma->cur->bc_private.b.flags =
|
|
|
|
bma->wasdel ? XFS_BTCUR_BPRV_WASDEL : 0;
|
|
|
|
|
|
|
|
bma->got.br_startoff = bma->offset;
|
|
|
|
bma->got.br_startblock = bma->blkno;
|
|
|
|
bma->got.br_blockcount = bma->length;
|
|
|
|
bma->got.br_state = XFS_EXT_NORM;
|
2010-06-23 08:11:15 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2017-02-02 23:14:01 +00:00
|
|
|
* In the data fork, a wasdelay extent has been initialized, so
|
|
|
|
* shouldn't be flagged as unwritten.
|
|
|
|
*
|
|
|
|
* For the cow fork, however, we convert delalloc reservations
|
|
|
|
* (extents allocated for speculative preallocation) to
|
|
|
|
* allocated unwritten extents, and only convert the unwritten
|
|
|
|
* extents to real extents when we're about to write the data.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2017-02-02 23:14:01 +00:00
|
|
|
if ((!bma->wasdel || (bma->flags & XFS_BMAPI_COWFORK)) &&
|
|
|
|
(bma->flags & XFS_BMAPI_PREALLOC) &&
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_sb_version_hasextflgbit(&mp->m_sb))
|
|
|
|
bma->got.br_state = XFS_EXT_UNWRITTEN;
|
|
|
|
|
|
|
|
if (bma->wasdel)
|
2016-10-03 16:11:34 +00:00
|
|
|
error = xfs_bmap_add_extent_delay_real(bma, whichfork);
|
2013-02-25 01:31:26 +00:00
|
|
|
else
|
2017-04-11 23:45:54 +00:00
|
|
|
error = xfs_bmap_add_extent_hole_real(bma->tp, bma->ip,
|
|
|
|
whichfork, &bma->idx, &bma->cur, &bma->got,
|
|
|
|
bma->firstblock, bma->dfops, &bma->logflags);
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
bma->logflags |= tmp_logflags;
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Update our extent pointer, given that xfs_bmap_add_extent_delay_real
|
|
|
|
* or xfs_bmap_add_extent_hole_real might have merged it into one of
|
|
|
|
* the neighbouring ones.
|
|
|
|
*/
|
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(ifp, bma->idx), &bma->got);
|
|
|
|
|
|
|
|
ASSERT(bma->got.br_startoff <= bma->offset);
|
|
|
|
ASSERT(bma->got.br_startoff + bma->got.br_blockcount >=
|
|
|
|
bma->offset + bma->length);
|
|
|
|
ASSERT(bma->got.br_state == XFS_EXT_NORM ||
|
|
|
|
bma->got.br_state == XFS_EXT_UNWRITTEN);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
STATIC int
|
|
|
|
xfs_bmapi_convert_unwritten(
|
|
|
|
struct xfs_bmalloca *bma,
|
|
|
|
struct xfs_bmbt_irec *mval,
|
|
|
|
xfs_filblks_t len,
|
|
|
|
int flags)
|
|
|
|
{
|
2016-10-03 16:11:32 +00:00
|
|
|
int whichfork = xfs_bmapi_whichfork(flags);
|
2013-02-25 01:31:26 +00:00
|
|
|
struct xfs_ifork *ifp = XFS_IFORK_PTR(bma->ip, whichfork);
|
|
|
|
int tmp_logflags = 0;
|
|
|
|
int error;
|
|
|
|
|
|
|
|
/* check if we need to do unwritten->real conversion */
|
|
|
|
if (mval->br_state == XFS_EXT_UNWRITTEN &&
|
|
|
|
(flags & XFS_BMAPI_PREALLOC))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* check if we need to do real->unwritten conversion */
|
|
|
|
if (mval->br_state == XFS_EXT_NORM &&
|
|
|
|
(flags & (XFS_BMAPI_PREALLOC | XFS_BMAPI_CONVERT)) !=
|
|
|
|
(XFS_BMAPI_PREALLOC | XFS_BMAPI_CONVERT))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Modify (by adding) the state flag, if writing.
|
|
|
|
*/
|
|
|
|
ASSERT(mval->br_blockcount <= len);
|
|
|
|
if ((ifp->if_flags & XFS_IFBROOT) && !bma->cur) {
|
|
|
|
bma->cur = xfs_bmbt_init_cursor(bma->ip->i_mount, bma->tp,
|
|
|
|
bma->ip, whichfork);
|
|
|
|
bma->cur->bc_private.b.firstblock = *bma->firstblock;
|
2016-08-03 01:19:29 +00:00
|
|
|
bma->cur->bc_private.b.dfops = bma->dfops;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
mval->br_state = (mval->br_state == XFS_EXT_UNWRITTEN)
|
|
|
|
? XFS_EXT_NORM : XFS_EXT_UNWRITTEN;
|
2012-02-23 00:41:39 +00:00
|
|
|
|
2015-11-03 01:27:22 +00:00
|
|
|
/*
|
|
|
|
* Before insertion into the bmbt, zero the range being converted
|
|
|
|
* if required.
|
|
|
|
*/
|
|
|
|
if (flags & XFS_BMAPI_ZERO) {
|
|
|
|
error = xfs_zero_extent(bma->ip, mval->br_startblock,
|
|
|
|
mval->br_blockcount);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2017-02-02 23:14:01 +00:00
|
|
|
error = xfs_bmap_add_extent_unwritten_real(bma->tp, bma->ip, whichfork,
|
|
|
|
&bma->idx, &bma->cur, mval, bma->firstblock, bma->dfops,
|
2013-02-25 01:31:26 +00:00
|
|
|
&tmp_logflags);
|
2015-05-31 21:15:23 +00:00
|
|
|
/*
|
|
|
|
* Log the inode core unconditionally in the unwritten extent conversion
|
|
|
|
* path because the conversion might not have done so (e.g., if the
|
|
|
|
* extent count hasn't changed). We need to make sure the inode is dirty
|
|
|
|
* in the transaction for the sake of fsync(), even if nothing has
|
|
|
|
* changed, because fsync() will not force the log for this transaction
|
|
|
|
* unless it sees the inode pinned.
|
2017-02-02 23:14:01 +00:00
|
|
|
*
|
|
|
|
* Note: If we're only converting cow fork extents, there aren't
|
|
|
|
* any on-disk updates to make, so we don't need to log anything.
|
2015-05-31 21:15:23 +00:00
|
|
|
*/
|
2017-02-02 23:14:01 +00:00
|
|
|
if (whichfork != XFS_COW_FORK)
|
|
|
|
bma->logflags |= tmp_logflags | XFS_ILOG_CORE;
|
2013-02-25 01:31:26 +00:00
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Update our extent pointer, given that
|
|
|
|
* xfs_bmap_add_extent_unwritten_real might have merged it into one
|
|
|
|
* of the neighbouring ones.
|
|
|
|
*/
|
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(ifp, bma->idx), &bma->got);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We may have combined previously unwritten space with written space,
|
|
|
|
* so generate another request.
|
|
|
|
*/
|
|
|
|
if (mval->br_blockcount < len)
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EAGAIN;
|
2013-02-25 01:31:26 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Map file blocks to filesystem blocks, and allocate blocks or convert the
|
|
|
|
* extent state if necessary. Details behaviour is controlled by the flags
|
|
|
|
* parameter. Only allocates blocks from a single allocation group, to avoid
|
|
|
|
* locking problems.
|
|
|
|
*
|
|
|
|
* The returned value in "firstblock" from the first call in a transaction
|
|
|
|
* must be remembered and presented to subsequent calls in "firstblock".
|
|
|
|
* An upper bound for the number of blocks to be allocated is supplied to
|
|
|
|
* the first call in "total"; if no allocation group has that many free
|
|
|
|
* blocks then the call will fail (return NULLFSBLOCK in "firstblock").
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
xfs_bmapi_write(
|
|
|
|
struct xfs_trans *tp, /* transaction pointer */
|
|
|
|
struct xfs_inode *ip, /* incore inode */
|
|
|
|
xfs_fileoff_t bno, /* starting file offs. mapped */
|
|
|
|
xfs_filblks_t len, /* length to map in file */
|
|
|
|
int flags, /* XFS_BMAPI_... */
|
|
|
|
xfs_fsblock_t *firstblock, /* first allocated block
|
|
|
|
controls a.g. for allocs */
|
|
|
|
xfs_extlen_t total, /* total blocks needed */
|
|
|
|
struct xfs_bmbt_irec *mval, /* output: map values */
|
|
|
|
int *nmap, /* i/o: mval size/count */
|
2016-08-03 01:19:29 +00:00
|
|
|
struct xfs_defer_ops *dfops) /* i/o: list extents to free */
|
2013-02-25 01:31:26 +00:00
|
|
|
{
|
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
|
|
|
struct xfs_ifork *ifp;
|
2013-09-02 10:49:36 +00:00
|
|
|
struct xfs_bmalloca bma = { NULL }; /* args for xfs_bmap_alloc */
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_fileoff_t end; /* end of mapped file region */
|
2016-11-24 00:39:43 +00:00
|
|
|
bool eof = false; /* after the end of extents */
|
2013-02-25 01:31:26 +00:00
|
|
|
int error; /* error return */
|
|
|
|
int n; /* current extent index */
|
|
|
|
xfs_fileoff_t obno; /* old block number (offset) */
|
|
|
|
int whichfork; /* data or attr fork */
|
|
|
|
|
|
|
|
#ifdef DEBUG
|
|
|
|
xfs_fileoff_t orig_bno; /* original block number value */
|
|
|
|
int orig_flags; /* original flags arg value */
|
|
|
|
xfs_filblks_t orig_len; /* original value of len arg */
|
|
|
|
struct xfs_bmbt_irec *orig_mval; /* original value of mval */
|
|
|
|
int orig_nmap; /* original value of *nmap */
|
|
|
|
|
|
|
|
orig_bno = bno;
|
|
|
|
orig_len = len;
|
|
|
|
orig_flags = flags;
|
|
|
|
orig_mval = mval;
|
|
|
|
orig_nmap = *nmap;
|
|
|
|
#endif
|
2016-10-03 16:11:34 +00:00
|
|
|
whichfork = xfs_bmapi_whichfork(flags);
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
ASSERT(*nmap >= 1);
|
|
|
|
ASSERT(*nmap <= XFS_BMAP_MAX_NMAP);
|
|
|
|
ASSERT(!(flags & XFS_BMAPI_IGSTATE));
|
2017-02-02 23:14:01 +00:00
|
|
|
ASSERT(tp != NULL ||
|
|
|
|
(flags & (XFS_BMAPI_CONVERT | XFS_BMAPI_COWFORK)) ==
|
|
|
|
(XFS_BMAPI_CONVERT | XFS_BMAPI_COWFORK));
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(len > 0);
|
2013-07-09 21:04:00 +00:00
|
|
|
ASSERT(XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_LOCAL);
|
2013-12-06 20:30:17 +00:00
|
|
|
ASSERT(xfs_isilocked(ip, XFS_ILOCK_EXCL));
|
2017-04-11 23:45:55 +00:00
|
|
|
ASSERT(!(flags & XFS_BMAPI_REMAP));
|
2013-02-25 01:31:26 +00:00
|
|
|
|
2015-11-03 01:27:22 +00:00
|
|
|
/* zeroing is for currently only for data extents, not metadata */
|
|
|
|
ASSERT((flags & (XFS_BMAPI_METADATA | XFS_BMAPI_ZERO)) !=
|
|
|
|
(XFS_BMAPI_METADATA | XFS_BMAPI_ZERO));
|
|
|
|
/*
|
|
|
|
* we can allocate unwritten extents or pre-zero allocated blocks,
|
|
|
|
* but it makes no sense to do both at once. This would result in
|
|
|
|
* zeroing the unwritten extent twice, but it still being an
|
|
|
|
* unwritten extent....
|
|
|
|
*/
|
|
|
|
ASSERT((flags & (XFS_BMAPI_PREALLOC | XFS_BMAPI_ZERO)) !=
|
|
|
|
(XFS_BMAPI_PREALLOC | XFS_BMAPI_ZERO));
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (unlikely(XFS_TEST_ERROR(
|
|
|
|
(XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS &&
|
2013-07-09 21:04:00 +00:00
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE),
|
2017-06-21 00:54:47 +00:00
|
|
|
mp, XFS_ERRTAG_BMAPIFORMAT))) {
|
2013-02-25 01:31:26 +00:00
|
|
|
XFS_ERROR_REPORT("xfs_bmapi_write", XFS_ERRLEVEL_LOW, mp);
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EFSCORRUPTED;
|
2012-02-23 00:41:39 +00:00
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (XFS_FORCED_SHUTDOWN(mp))
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EIO;
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
|
2015-10-12 07:21:22 +00:00
|
|
|
XFS_STATS_INC(mp, xs_blk_mapw);
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
if (*firstblock == NULLFSBLOCK) {
|
|
|
|
if (XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_BTREE)
|
|
|
|
bma.minleft = be16_to_cpu(ifp->if_broot->bb_level) + 1;
|
|
|
|
else
|
|
|
|
bma.minleft = 1;
|
|
|
|
} else {
|
|
|
|
bma.minleft = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!(ifp->if_flags & XFS_IFEXTENTS)) {
|
|
|
|
error = xfs_iread_extents(tp, ip, whichfork);
|
|
|
|
if (error)
|
|
|
|
goto error0;
|
|
|
|
}
|
|
|
|
|
|
|
|
n = 0;
|
|
|
|
end = bno + len;
|
|
|
|
obno = bno;
|
|
|
|
|
2016-11-24 00:39:43 +00:00
|
|
|
if (!xfs_iext_lookup_extent(ip, ifp, bno, &bma.idx, &bma.got))
|
|
|
|
eof = true;
|
|
|
|
if (!xfs_iext_get_extent(ifp, bma.idx - 1, &bma.prev))
|
|
|
|
bma.prev.br_startoff = NULLFILEOFF;
|
2013-02-25 01:31:26 +00:00
|
|
|
bma.tp = tp;
|
|
|
|
bma.ip = ip;
|
|
|
|
bma.total = total;
|
xfs: remote attribute blocks aren't really userdata
When adding a new remote attribute, we write the attribute to the
new extent before the allocation transaction is committed. This
means we cannot reuse busy extents as that violates crash
consistency semantics. Hence we currently treat remote attribute
extent allocation like userdata because it has the same overwrite
ordering constraints as userdata.
Unfortunately, this also allows the allocator to incorrectly apply
extent size hints to the remote attribute extent allocation. This
results in interesting failures, such as transaction block
reservation overruns and in-memory inode attribute fork corruption.
To fix this, we need to separate the busy extent reuse configuration
from the userdata configuration. This changes the definition of
XFS_BMAPI_METADATA slightly - it now means that allocation is
metadata and reuse of busy extents is acceptible due to the metadata
ordering semantics of the journal. If this flag is not set, it
means the allocation is that has unordered data writeback, and hence
busy extent reuse is not allowed. It no longer implies the
allocation is for user data, just that the data write will not be
strictly ordered. This matches the semantics for both user data
and remote attribute block allocation.
As such, This patch changes the "userdata" field to a "datatype"
field, and adds a "no busy reuse" flag to the field.
When we detect an unordered data extent allocation, we immediately set
the no reuse flag. We then set the "user data" flags based on the
inode fork we are allocating the extent to. Hence we only set
userdata flags on data fork allocations now and consider attribute
fork remote extents to be an unordered metadata extent.
The result is that remote attribute extents now have the expected
allocation semantics, and the data fork allocation behaviour is
completely unchanged.
It should be noted that there may be other ways to fix this (e.g.
use ordered metadata buffers for the remote attribute extent data
write) but they are more invasive and difficult to validate both
from a design and implementation POV. Hence this patch takes the
simple, obvious route to fixing the problem...
Reported-and-tested-by: Ross Zwisler <ross.zwisler@linux.intel.com>
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-25 22:21:28 +00:00
|
|
|
bma.datatype = 0;
|
2016-08-03 01:19:29 +00:00
|
|
|
bma.dfops = dfops;
|
2013-02-25 01:31:26 +00:00
|
|
|
bma.firstblock = firstblock;
|
|
|
|
|
|
|
|
while (bno < end && n < *nmap) {
|
2017-01-20 17:31:54 +00:00
|
|
|
bool need_alloc = false, wasdelay = false;
|
|
|
|
|
|
|
|
/* in hole or beyoned EOF? */
|
|
|
|
if (eof || bma.got.br_startoff > bno) {
|
|
|
|
if (flags & XFS_BMAPI_DELALLOC) {
|
|
|
|
/*
|
|
|
|
* For the COW fork we can reasonably get a
|
|
|
|
* request for converting an extent that races
|
|
|
|
* with other threads already having converted
|
|
|
|
* part of it, as there converting COW to
|
|
|
|
* regular blocks is not protected using the
|
|
|
|
* IOLOCK.
|
|
|
|
*/
|
|
|
|
ASSERT(flags & XFS_BMAPI_COWFORK);
|
|
|
|
if (!(flags & XFS_BMAPI_COWFORK)) {
|
|
|
|
error = -EIO;
|
|
|
|
goto error0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (eof || bno >= end)
|
|
|
|
break;
|
|
|
|
} else {
|
|
|
|
need_alloc = true;
|
|
|
|
}
|
2017-04-11 23:45:55 +00:00
|
|
|
} else if (isnullstartblock(bma.got.br_startblock)) {
|
|
|
|
wasdelay = true;
|
2017-01-20 17:31:54 +00:00
|
|
|
}
|
2016-10-03 16:11:27 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* First, deal with the hole before the allocated space
|
|
|
|
* that we found, if any.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2017-01-20 17:31:54 +00:00
|
|
|
if (need_alloc || wasdelay) {
|
2013-02-25 01:31:26 +00:00
|
|
|
bma.eof = eof;
|
|
|
|
bma.conv = !!(flags & XFS_BMAPI_CONVERT);
|
|
|
|
bma.wasdel = wasdelay;
|
|
|
|
bma.offset = bno;
|
|
|
|
bma.flags = flags;
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* There's a 32/64 bit type mismatch between the
|
|
|
|
* allocation length request (which can be 64 bits in
|
|
|
|
* length) and the bma length request, which is
|
|
|
|
* xfs_extlen_t and therefore 32 bits. Hence we have to
|
|
|
|
* check for 32-bit overflows and handle them here.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (len > (xfs_filblks_t)MAXEXTLEN)
|
|
|
|
bma.length = MAXEXTLEN;
|
|
|
|
else
|
|
|
|
bma.length = len;
|
|
|
|
|
|
|
|
ASSERT(len > 0);
|
|
|
|
ASSERT(bma.length > 0);
|
|
|
|
error = xfs_bmapi_allocate(&bma);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (error)
|
|
|
|
goto error0;
|
2013-02-25 01:31:26 +00:00
|
|
|
if (bma.blkno == NULLFSBLOCK)
|
|
|
|
break;
|
2016-10-03 16:11:39 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If this is a CoW allocation, record the data in
|
|
|
|
* the refcount btree for orphan recovery.
|
|
|
|
*/
|
|
|
|
if (whichfork == XFS_COW_FORK) {
|
|
|
|
error = xfs_refcount_alloc_cow_extent(mp, dfops,
|
|
|
|
bma.blkno, bma.length);
|
|
|
|
if (error)
|
|
|
|
goto error0;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
/* Deal with the allocated space we found. */
|
|
|
|
xfs_bmapi_trim_map(mval, &bma.got, &bno, len, obno,
|
|
|
|
end, n, flags);
|
|
|
|
|
|
|
|
/* Execute unwritten extent conversion if necessary */
|
|
|
|
error = xfs_bmapi_convert_unwritten(&bma, mval, len, flags);
|
2014-06-25 04:58:08 +00:00
|
|
|
if (error == -EAGAIN)
|
2013-02-25 01:31:26 +00:00
|
|
|
continue;
|
|
|
|
if (error)
|
|
|
|
goto error0;
|
|
|
|
|
|
|
|
/* update the extent map to return */
|
|
|
|
xfs_bmapi_update_map(&mval, &bno, &len, obno, end, &n, flags);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we're done, stop now. Stop when we've allocated
|
|
|
|
* XFS_BMAP_MAX_NMAP extents no matter what. Otherwise
|
|
|
|
* the transaction may get too big.
|
|
|
|
*/
|
|
|
|
if (bno >= end || n >= *nmap || bma.nallocs >= *nmap)
|
|
|
|
break;
|
|
|
|
|
|
|
|
/* Else go on to the next record. */
|
|
|
|
bma.prev = bma.got;
|
2016-11-24 00:39:43 +00:00
|
|
|
if (!xfs_iext_get_extent(ifp, ++bma.idx, &bma.got))
|
|
|
|
eof = true;
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
|
|
|
*nmap = n;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Transform from btree to extents, give it cur.
|
|
|
|
*/
|
|
|
|
if (xfs_bmap_wants_extents(ip, whichfork)) {
|
|
|
|
int tmp_logflags = 0;
|
|
|
|
|
|
|
|
ASSERT(bma.cur);
|
|
|
|
error = xfs_bmap_btree_to_extents(tp, ip, bma.cur,
|
|
|
|
&tmp_logflags, whichfork);
|
|
|
|
bma.logflags |= tmp_logflags;
|
|
|
|
if (error)
|
|
|
|
goto error0;
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT(XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE ||
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) >
|
|
|
|
XFS_IFORK_MAXEXT(ip, whichfork));
|
|
|
|
error = 0;
|
|
|
|
error0:
|
|
|
|
/*
|
|
|
|
* Log everything. Do this after conversion, there's no point in
|
|
|
|
* logging the extent records if we've converted to btree format.
|
|
|
|
*/
|
|
|
|
if ((bma.logflags & xfs_ilog_fext(whichfork)) &&
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS)
|
|
|
|
bma.logflags &= ~xfs_ilog_fext(whichfork);
|
|
|
|
else if ((bma.logflags & xfs_ilog_fbroot(whichfork)) &&
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE)
|
|
|
|
bma.logflags &= ~xfs_ilog_fbroot(whichfork);
|
|
|
|
/*
|
|
|
|
* Log whatever the flags say, even if error. Otherwise we might miss
|
|
|
|
* detecting a case where the data is changed, there's an error,
|
|
|
|
* and it's not logged so we don't shutdown when we should.
|
|
|
|
*/
|
|
|
|
if (bma.logflags)
|
|
|
|
xfs_trans_log_inode(tp, ip, bma.logflags);
|
|
|
|
|
|
|
|
if (bma.cur) {
|
|
|
|
if (!error) {
|
|
|
|
ASSERT(*firstblock == NULLFSBLOCK ||
|
2017-02-17 01:12:51 +00:00
|
|
|
XFS_FSB_TO_AGNO(mp, *firstblock) <=
|
2013-02-25 01:31:26 +00:00
|
|
|
XFS_FSB_TO_AGNO(mp,
|
2017-02-17 01:12:51 +00:00
|
|
|
bma.cur->bc_private.b.firstblock));
|
2013-02-25 01:31:26 +00:00
|
|
|
*firstblock = bma.cur->bc_private.b.firstblock;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_btree_del_cursor(bma.cur,
|
|
|
|
error ? XFS_BTREE_ERROR : XFS_BTREE_NOERROR);
|
|
|
|
}
|
|
|
|
if (!error)
|
|
|
|
xfs_bmap_validate_ret(orig_bno, orig_len, orig_flags, orig_mval,
|
|
|
|
orig_nmap, *nmap);
|
|
|
|
return error;
|
|
|
|
}
|
2005-06-21 05:48:47 +00:00
|
|
|
|
2017-04-11 23:45:55 +00:00
|
|
|
static int
|
|
|
|
xfs_bmapi_remap(
|
|
|
|
struct xfs_trans *tp,
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
xfs_fileoff_t bno,
|
|
|
|
xfs_filblks_t len,
|
|
|
|
xfs_fsblock_t startblock,
|
|
|
|
struct xfs_defer_ops *dfops)
|
|
|
|
{
|
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
|
|
|
struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, XFS_DATA_FORK);
|
|
|
|
struct xfs_btree_cur *cur = NULL;
|
|
|
|
xfs_fsblock_t firstblock = NULLFSBLOCK;
|
|
|
|
struct xfs_bmbt_irec got;
|
|
|
|
xfs_extnum_t idx;
|
|
|
|
int logflags = 0, error;
|
|
|
|
|
|
|
|
ASSERT(len > 0);
|
|
|
|
ASSERT(len <= (xfs_filblks_t)MAXEXTLEN);
|
|
|
|
ASSERT(xfs_isilocked(ip, XFS_ILOCK_EXCL));
|
|
|
|
|
|
|
|
if (unlikely(XFS_TEST_ERROR(
|
|
|
|
(XFS_IFORK_FORMAT(ip, XFS_DATA_FORK) != XFS_DINODE_FMT_EXTENTS &&
|
|
|
|
XFS_IFORK_FORMAT(ip, XFS_DATA_FORK) != XFS_DINODE_FMT_BTREE),
|
2017-06-21 00:54:47 +00:00
|
|
|
mp, XFS_ERRTAG_BMAPIFORMAT))) {
|
2017-04-11 23:45:55 +00:00
|
|
|
XFS_ERROR_REPORT("xfs_bmapi_remap", XFS_ERRLEVEL_LOW, mp);
|
|
|
|
return -EFSCORRUPTED;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (XFS_FORCED_SHUTDOWN(mp))
|
|
|
|
return -EIO;
|
|
|
|
|
|
|
|
if (!(ifp->if_flags & XFS_IFEXTENTS)) {
|
|
|
|
error = xfs_iread_extents(NULL, ip, XFS_DATA_FORK);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (xfs_iext_lookup_extent(ip, ifp, bno, &idx, &got)) {
|
|
|
|
/* make sure we only reflink into a hole. */
|
|
|
|
ASSERT(got.br_startoff > bno);
|
|
|
|
ASSERT(got.br_startoff - bno >= len);
|
|
|
|
}
|
|
|
|
|
2017-04-11 23:45:56 +00:00
|
|
|
ip->i_d.di_nblocks += len;
|
|
|
|
xfs_trans_log_inode(tp, ip, XFS_ILOG_CORE);
|
2017-04-11 23:45:55 +00:00
|
|
|
|
|
|
|
if (ifp->if_flags & XFS_IFBROOT) {
|
|
|
|
cur = xfs_bmbt_init_cursor(mp, tp, ip, XFS_DATA_FORK);
|
|
|
|
cur->bc_private.b.firstblock = firstblock;
|
|
|
|
cur->bc_private.b.dfops = dfops;
|
|
|
|
cur->bc_private.b.flags = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
got.br_startoff = bno;
|
|
|
|
got.br_startblock = startblock;
|
|
|
|
got.br_blockcount = len;
|
|
|
|
got.br_state = XFS_EXT_NORM;
|
|
|
|
|
|
|
|
error = xfs_bmap_add_extent_hole_real(tp, ip, XFS_DATA_FORK, &idx, &cur,
|
|
|
|
&got, &firstblock, dfops, &logflags);
|
|
|
|
if (error)
|
|
|
|
goto error0;
|
|
|
|
|
|
|
|
if (xfs_bmap_wants_extents(ip, XFS_DATA_FORK)) {
|
|
|
|
int tmp_logflags = 0;
|
|
|
|
|
|
|
|
error = xfs_bmap_btree_to_extents(tp, ip, cur,
|
|
|
|
&tmp_logflags, XFS_DATA_FORK);
|
|
|
|
logflags |= tmp_logflags;
|
|
|
|
}
|
|
|
|
|
|
|
|
error0:
|
|
|
|
if (ip->i_d.di_format != XFS_DINODE_FMT_EXTENTS)
|
|
|
|
logflags &= ~XFS_ILOG_DEXT;
|
|
|
|
else if (ip->i_d.di_format != XFS_DINODE_FMT_BTREE)
|
|
|
|
logflags &= ~XFS_ILOG_DBROOT;
|
|
|
|
|
|
|
|
if (logflags)
|
|
|
|
xfs_trans_log_inode(tp, ip, logflags);
|
|
|
|
if (cur) {
|
|
|
|
xfs_btree_del_cursor(cur,
|
|
|
|
error ? XFS_BTREE_ERROR : XFS_BTREE_NOERROR);
|
|
|
|
}
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2016-03-15 00:42:46 +00:00
|
|
|
/*
|
|
|
|
* When a delalloc extent is split (e.g., due to a hole punch), the original
|
|
|
|
* indlen reservation must be shared across the two new extents that are left
|
|
|
|
* behind.
|
|
|
|
*
|
|
|
|
* Given the original reservation and the worst case indlen for the two new
|
|
|
|
* extents (as calculated by xfs_bmap_worst_indlen()), split the original
|
2016-03-15 00:42:47 +00:00
|
|
|
* reservation fairly across the two new extents. If necessary, steal available
|
|
|
|
* blocks from a deleted extent to make up a reservation deficiency (e.g., if
|
|
|
|
* ores == 1). The number of stolen blocks is returned. The availability and
|
|
|
|
* subsequent accounting of stolen blocks is the responsibility of the caller.
|
2016-03-15 00:42:46 +00:00
|
|
|
*/
|
2016-03-15 00:42:47 +00:00
|
|
|
static xfs_filblks_t
|
2016-03-15 00:42:46 +00:00
|
|
|
xfs_bmap_split_indlen(
|
|
|
|
xfs_filblks_t ores, /* original res. */
|
|
|
|
xfs_filblks_t *indlen1, /* ext1 worst indlen */
|
2016-03-15 00:42:47 +00:00
|
|
|
xfs_filblks_t *indlen2, /* ext2 worst indlen */
|
|
|
|
xfs_filblks_t avail) /* stealable blocks */
|
2016-03-15 00:42:46 +00:00
|
|
|
{
|
|
|
|
xfs_filblks_t len1 = *indlen1;
|
|
|
|
xfs_filblks_t len2 = *indlen2;
|
|
|
|
xfs_filblks_t nres = len1 + len2; /* new total res. */
|
2016-03-15 00:42:47 +00:00
|
|
|
xfs_filblks_t stolen = 0;
|
2017-02-14 06:48:30 +00:00
|
|
|
xfs_filblks_t resfactor;
|
2016-03-15 00:42:47 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Steal as many blocks as we can to try and satisfy the worst case
|
|
|
|
* indlen for both new extents.
|
|
|
|
*/
|
2017-02-14 06:48:30 +00:00
|
|
|
if (ores < nres && avail)
|
|
|
|
stolen = XFS_FILBLKS_MIN(nres - ores, avail);
|
|
|
|
ores += stolen;
|
|
|
|
|
|
|
|
/* nothing else to do if we've satisfied the new reservation */
|
|
|
|
if (ores >= nres)
|
|
|
|
return stolen;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We can't meet the total required reservation for the two extents.
|
|
|
|
* Calculate the percent of the overall shortage between both extents
|
|
|
|
* and apply this percentage to each of the requested indlen values.
|
|
|
|
* This distributes the shortage fairly and reduces the chances that one
|
|
|
|
* of the two extents is left with nothing when extents are repeatedly
|
|
|
|
* split.
|
|
|
|
*/
|
|
|
|
resfactor = (ores * 100);
|
|
|
|
do_div(resfactor, nres);
|
|
|
|
len1 *= resfactor;
|
|
|
|
do_div(len1, 100);
|
|
|
|
len2 *= resfactor;
|
|
|
|
do_div(len2, 100);
|
|
|
|
ASSERT(len1 + len2 <= ores);
|
|
|
|
ASSERT(len1 < *indlen1 && len2 < *indlen2);
|
2016-03-15 00:42:46 +00:00
|
|
|
|
|
|
|
/*
|
2017-02-14 06:48:30 +00:00
|
|
|
* Hand out the remainder to each extent. If one of the two reservations
|
|
|
|
* is zero, we want to make sure that one gets a block first. The loop
|
|
|
|
* below starts with len1, so hand len2 a block right off the bat if it
|
|
|
|
* is zero.
|
2016-03-15 00:42:46 +00:00
|
|
|
*/
|
2017-02-14 06:48:30 +00:00
|
|
|
ores -= (len1 + len2);
|
|
|
|
ASSERT((*indlen1 - len1) + (*indlen2 - len2) >= ores);
|
|
|
|
if (ores && !len2 && *indlen2) {
|
|
|
|
len2++;
|
|
|
|
ores--;
|
|
|
|
}
|
|
|
|
while (ores) {
|
|
|
|
if (len1 < *indlen1) {
|
|
|
|
len1++;
|
|
|
|
ores--;
|
2016-03-15 00:42:46 +00:00
|
|
|
}
|
2017-02-14 06:48:30 +00:00
|
|
|
if (!ores)
|
2016-03-15 00:42:46 +00:00
|
|
|
break;
|
2017-02-14 06:48:30 +00:00
|
|
|
if (len2 < *indlen2) {
|
|
|
|
len2++;
|
|
|
|
ores--;
|
2016-03-15 00:42:46 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
*indlen1 = len1;
|
|
|
|
*indlen2 = len2;
|
2016-03-15 00:42:47 +00:00
|
|
|
|
|
|
|
return stolen;
|
2016-03-15 00:42:46 +00:00
|
|
|
}
|
|
|
|
|
2016-10-20 04:54:14 +00:00
|
|
|
int
|
|
|
|
xfs_bmap_del_extent_delay(
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
int whichfork,
|
|
|
|
xfs_extnum_t *idx,
|
|
|
|
struct xfs_bmbt_irec *got,
|
|
|
|
struct xfs_bmbt_irec *del)
|
|
|
|
{
|
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
|
|
|
struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
struct xfs_bmbt_irec new;
|
|
|
|
int64_t da_old, da_new, da_diff = 0;
|
|
|
|
xfs_fileoff_t del_endoff, got_endoff;
|
|
|
|
xfs_filblks_t got_indlen, new_indlen, stolen;
|
|
|
|
int error = 0, state = 0;
|
|
|
|
bool isrt;
|
|
|
|
|
|
|
|
XFS_STATS_INC(mp, xs_del_exlist);
|
|
|
|
|
|
|
|
isrt = (whichfork == XFS_DATA_FORK) && XFS_IS_REALTIME_INODE(ip);
|
|
|
|
del_endoff = del->br_startoff + del->br_blockcount;
|
|
|
|
got_endoff = got->br_startoff + got->br_blockcount;
|
|
|
|
da_old = startblockval(got->br_startblock);
|
|
|
|
da_new = 0;
|
|
|
|
|
|
|
|
ASSERT(*idx >= 0);
|
2016-11-08 01:59:42 +00:00
|
|
|
ASSERT(*idx <= xfs_iext_count(ifp));
|
2016-10-20 04:54:14 +00:00
|
|
|
ASSERT(del->br_blockcount > 0);
|
|
|
|
ASSERT(got->br_startoff <= del->br_startoff);
|
|
|
|
ASSERT(got_endoff >= del_endoff);
|
|
|
|
|
|
|
|
if (isrt) {
|
2017-04-19 22:19:32 +00:00
|
|
|
uint64_t rtexts = XFS_FSB_TO_B(mp, del->br_blockcount);
|
2016-10-20 04:54:14 +00:00
|
|
|
|
|
|
|
do_div(rtexts, mp->m_sb.sb_rextsize);
|
|
|
|
xfs_mod_frextents(mp, rtexts);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Update the inode delalloc counter now and wait to update the
|
|
|
|
* sb counters as we might have to borrow some blocks for the
|
|
|
|
* indirect block accounting.
|
|
|
|
*/
|
2016-11-08 00:59:26 +00:00
|
|
|
error = xfs_trans_reserve_quota_nblks(NULL, ip,
|
|
|
|
-((long)del->br_blockcount), 0,
|
2016-10-20 04:54:14 +00:00
|
|
|
isrt ? XFS_QMOPT_RES_RTBLKS : XFS_QMOPT_RES_REGBLKS);
|
2016-11-08 00:59:26 +00:00
|
|
|
if (error)
|
|
|
|
return error;
|
2016-10-20 04:54:14 +00:00
|
|
|
ip->i_delayed_blks -= del->br_blockcount;
|
|
|
|
|
|
|
|
if (whichfork == XFS_COW_FORK)
|
|
|
|
state |= BMAP_COWFORK;
|
|
|
|
|
|
|
|
if (got->br_startoff == del->br_startoff)
|
|
|
|
state |= BMAP_LEFT_CONTIG;
|
|
|
|
if (got_endoff == del_endoff)
|
|
|
|
state |= BMAP_RIGHT_CONTIG;
|
|
|
|
|
|
|
|
switch (state & (BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG)) {
|
|
|
|
case BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG:
|
|
|
|
/*
|
|
|
|
* Matches the whole extent. Delete the entry.
|
|
|
|
*/
|
|
|
|
xfs_iext_remove(ip, *idx, 1, state);
|
|
|
|
--*idx;
|
|
|
|
break;
|
|
|
|
case BMAP_LEFT_CONTIG:
|
|
|
|
/*
|
|
|
|
* Deleting the first part of the extent.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
got->br_startoff = del_endoff;
|
|
|
|
got->br_blockcount -= del->br_blockcount;
|
|
|
|
da_new = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip,
|
|
|
|
got->br_blockcount), da_old);
|
|
|
|
got->br_startblock = nullstartblock((int)da_new);
|
2017-08-29 22:44:11 +00:00
|
|
|
xfs_iext_update_extent(ifp, *idx, got);
|
2016-10-20 04:54:14 +00:00
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
break;
|
|
|
|
case BMAP_RIGHT_CONTIG:
|
|
|
|
/*
|
|
|
|
* Deleting the last part of the extent.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
got->br_blockcount = got->br_blockcount - del->br_blockcount;
|
|
|
|
da_new = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip,
|
|
|
|
got->br_blockcount), da_old);
|
|
|
|
got->br_startblock = nullstartblock((int)da_new);
|
2017-08-29 22:44:11 +00:00
|
|
|
xfs_iext_update_extent(ifp, *idx, got);
|
2016-10-20 04:54:14 +00:00
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
break;
|
|
|
|
case 0:
|
|
|
|
/*
|
|
|
|
* Deleting the middle of the extent.
|
|
|
|
*
|
|
|
|
* Distribute the original indlen reservation across the two new
|
|
|
|
* extents. Steal blocks from the deleted extent if necessary.
|
|
|
|
* Stealing blocks simply fudges the fdblocks accounting below.
|
|
|
|
* Warn if either of the new indlen reservations is zero as this
|
|
|
|
* can lead to delalloc problems.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
|
|
|
|
got->br_blockcount = del->br_startoff - got->br_startoff;
|
|
|
|
got_indlen = xfs_bmap_worst_indlen(ip, got->br_blockcount);
|
|
|
|
|
|
|
|
new.br_blockcount = got_endoff - del_endoff;
|
|
|
|
new_indlen = xfs_bmap_worst_indlen(ip, new.br_blockcount);
|
|
|
|
|
|
|
|
WARN_ON_ONCE(!got_indlen || !new_indlen);
|
|
|
|
stolen = xfs_bmap_split_indlen(da_old, &got_indlen, &new_indlen,
|
|
|
|
del->br_blockcount);
|
|
|
|
|
|
|
|
got->br_startblock = nullstartblock((int)got_indlen);
|
2017-08-29 22:44:11 +00:00
|
|
|
xfs_iext_update_extent(ifp, *idx, got);
|
2016-10-20 04:54:14 +00:00
|
|
|
trace_xfs_bmap_post_update(ip, *idx, 0, _THIS_IP_);
|
|
|
|
|
|
|
|
new.br_startoff = del_endoff;
|
|
|
|
new.br_state = got->br_state;
|
|
|
|
new.br_startblock = nullstartblock((int)new_indlen);
|
|
|
|
|
|
|
|
++*idx;
|
|
|
|
xfs_iext_insert(ip, *idx, 1, &new, state);
|
|
|
|
|
|
|
|
da_new = got_indlen + new_indlen - stolen;
|
|
|
|
del->br_blockcount -= stolen;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
ASSERT(da_old >= da_new);
|
|
|
|
da_diff = da_old - da_new;
|
|
|
|
if (!isrt)
|
|
|
|
da_diff += del->br_blockcount;
|
|
|
|
if (da_diff)
|
|
|
|
xfs_mod_fdblocks(mp, da_diff, false);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
xfs_bmap_del_extent_cow(
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
xfs_extnum_t *idx,
|
|
|
|
struct xfs_bmbt_irec *got,
|
|
|
|
struct xfs_bmbt_irec *del)
|
|
|
|
{
|
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
|
|
|
struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, XFS_COW_FORK);
|
|
|
|
struct xfs_bmbt_irec new;
|
|
|
|
xfs_fileoff_t del_endoff, got_endoff;
|
|
|
|
int state = BMAP_COWFORK;
|
|
|
|
|
|
|
|
XFS_STATS_INC(mp, xs_del_exlist);
|
|
|
|
|
|
|
|
del_endoff = del->br_startoff + del->br_blockcount;
|
|
|
|
got_endoff = got->br_startoff + got->br_blockcount;
|
|
|
|
|
|
|
|
ASSERT(*idx >= 0);
|
2016-11-08 01:59:42 +00:00
|
|
|
ASSERT(*idx <= xfs_iext_count(ifp));
|
2016-10-20 04:54:14 +00:00
|
|
|
ASSERT(del->br_blockcount > 0);
|
|
|
|
ASSERT(got->br_startoff <= del->br_startoff);
|
|
|
|
ASSERT(got_endoff >= del_endoff);
|
|
|
|
ASSERT(!isnullstartblock(got->br_startblock));
|
|
|
|
|
|
|
|
if (got->br_startoff == del->br_startoff)
|
|
|
|
state |= BMAP_LEFT_CONTIG;
|
|
|
|
if (got_endoff == del_endoff)
|
|
|
|
state |= BMAP_RIGHT_CONTIG;
|
|
|
|
|
|
|
|
switch (state & (BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG)) {
|
|
|
|
case BMAP_LEFT_CONTIG | BMAP_RIGHT_CONTIG:
|
|
|
|
/*
|
|
|
|
* Matches the whole extent. Delete the entry.
|
|
|
|
*/
|
|
|
|
xfs_iext_remove(ip, *idx, 1, state);
|
|
|
|
--*idx;
|
|
|
|
break;
|
|
|
|
case BMAP_LEFT_CONTIG:
|
|
|
|
/*
|
|
|
|
* Deleting the first part of the extent.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
got->br_startoff = del_endoff;
|
|
|
|
got->br_blockcount -= del->br_blockcount;
|
|
|
|
got->br_startblock = del->br_startblock + del->br_blockcount;
|
2017-08-29 22:44:11 +00:00
|
|
|
xfs_iext_update_extent(ifp, *idx, got);
|
2016-10-20 04:54:14 +00:00
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
break;
|
|
|
|
case BMAP_RIGHT_CONTIG:
|
|
|
|
/*
|
|
|
|
* Deleting the last part of the extent.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
got->br_blockcount -= del->br_blockcount;
|
2017-08-29 22:44:11 +00:00
|
|
|
xfs_iext_update_extent(ifp, *idx, got);
|
2016-10-20 04:54:14 +00:00
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
break;
|
|
|
|
case 0:
|
|
|
|
/*
|
|
|
|
* Deleting the middle of the extent.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
got->br_blockcount = del->br_startoff - got->br_startoff;
|
2017-08-29 22:44:11 +00:00
|
|
|
xfs_iext_update_extent(ifp, *idx, got);
|
2016-10-20 04:54:14 +00:00
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
|
|
|
|
new.br_startoff = del_endoff;
|
|
|
|
new.br_blockcount = got_endoff - del_endoff;
|
|
|
|
new.br_state = got->br_state;
|
|
|
|
new.br_startblock = del->br_startblock + del->br_blockcount;
|
|
|
|
|
|
|
|
++*idx;
|
|
|
|
xfs_iext_insert(ip, *idx, 1, &new, state);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Called by xfs_bmapi to update file extent records and the btree
|
|
|
|
* after removing space (or undoing a delayed allocation).
|
|
|
|
*/
|
|
|
|
STATIC int /* error */
|
|
|
|
xfs_bmap_del_extent(
|
|
|
|
xfs_inode_t *ip, /* incore inode pointer */
|
|
|
|
xfs_trans_t *tp, /* current transaction pointer */
|
|
|
|
xfs_extnum_t *idx, /* extent number to update/delete */
|
2016-08-03 01:19:29 +00:00
|
|
|
struct xfs_defer_ops *dfops, /* list of extents to be freed */
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_btree_cur_t *cur, /* if null, not a btree */
|
|
|
|
xfs_bmbt_irec_t *del, /* data to remove from extents */
|
|
|
|
int *logflagsp, /* inode logging flags */
|
2016-10-03 16:11:27 +00:00
|
|
|
int whichfork, /* data or attr fork */
|
|
|
|
int bflags) /* bmapi flags */
|
2013-02-25 01:31:26 +00:00
|
|
|
{
|
|
|
|
xfs_filblks_t da_new; /* new delay-alloc indirect blocks */
|
|
|
|
xfs_filblks_t da_old; /* old delay-alloc indirect blocks */
|
|
|
|
xfs_fsblock_t del_endblock=0; /* first block past del */
|
|
|
|
xfs_fileoff_t del_endoff; /* first offset past del */
|
|
|
|
int delay; /* current block is delayed allocated */
|
|
|
|
int do_fx; /* free extent at end of routine */
|
|
|
|
xfs_bmbt_rec_host_t *ep; /* current extent entry pointer */
|
|
|
|
int error; /* error return value */
|
|
|
|
int flags; /* inode logging flags */
|
|
|
|
xfs_bmbt_irec_t got; /* current extent entry */
|
|
|
|
xfs_fileoff_t got_endoff; /* first offset past got */
|
|
|
|
int i; /* temp state */
|
|
|
|
xfs_ifork_t *ifp; /* inode fork pointer */
|
|
|
|
xfs_mount_t *mp; /* mount structure */
|
|
|
|
xfs_filblks_t nblks; /* quota/sb block count */
|
|
|
|
xfs_bmbt_irec_t new; /* new record to be inserted */
|
|
|
|
/* REFERENCED */
|
|
|
|
uint qfield; /* quota field to update */
|
|
|
|
xfs_filblks_t temp; /* for indirect length calculations */
|
|
|
|
xfs_filblks_t temp2; /* for indirect length calculations */
|
|
|
|
int state = 0;
|
|
|
|
|
2015-10-12 07:21:22 +00:00
|
|
|
mp = ip->i_mount;
|
|
|
|
XFS_STATS_INC(mp, xs_del_exlist);
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
if (whichfork == XFS_ATTR_FORK)
|
|
|
|
state |= BMAP_ATTRFORK;
|
2016-10-03 16:11:32 +00:00
|
|
|
else if (whichfork == XFS_COW_FORK)
|
|
|
|
state |= BMAP_COWFORK;
|
2013-02-25 01:31:26 +00:00
|
|
|
|
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
2016-11-08 01:59:42 +00:00
|
|
|
ASSERT((*idx >= 0) && (*idx < xfs_iext_count(ifp)));
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(del->br_blockcount > 0);
|
|
|
|
ep = xfs_iext_get_ext(ifp, *idx);
|
|
|
|
xfs_bmbt_get_all(ep, &got);
|
|
|
|
ASSERT(got.br_startoff <= del->br_startoff);
|
|
|
|
del_endoff = del->br_startoff + del->br_blockcount;
|
|
|
|
got_endoff = got.br_startoff + got.br_blockcount;
|
|
|
|
ASSERT(got_endoff >= del_endoff);
|
|
|
|
delay = isnullstartblock(got.br_startblock);
|
|
|
|
ASSERT(isnullstartblock(del->br_startblock) == delay);
|
|
|
|
flags = 0;
|
|
|
|
qfield = 0;
|
|
|
|
error = 0;
|
|
|
|
/*
|
|
|
|
* If deleting a real allocation, must free up the disk space.
|
|
|
|
*/
|
|
|
|
if (!delay) {
|
|
|
|
flags = XFS_ILOG_CORE;
|
|
|
|
/*
|
|
|
|
* Realtime allocation. Free it and record di_nblocks update.
|
|
|
|
*/
|
|
|
|
if (whichfork == XFS_DATA_FORK && XFS_IS_REALTIME_INODE(ip)) {
|
|
|
|
xfs_fsblock_t bno;
|
|
|
|
xfs_filblks_t len;
|
|
|
|
|
|
|
|
ASSERT(do_mod(del->br_blockcount,
|
|
|
|
mp->m_sb.sb_rextsize) == 0);
|
|
|
|
ASSERT(do_mod(del->br_startblock,
|
|
|
|
mp->m_sb.sb_rextsize) == 0);
|
|
|
|
bno = del->br_startblock;
|
|
|
|
len = del->br_blockcount;
|
|
|
|
do_div(bno, mp->m_sb.sb_rextsize);
|
|
|
|
do_div(len, mp->m_sb.sb_rextsize);
|
|
|
|
error = xfs_rtfree_extent(tp, bno, (xfs_extlen_t)len);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
do_fx = 0;
|
|
|
|
nblks = len * mp->m_sb.sb_rextsize;
|
|
|
|
qfield = XFS_TRANS_DQ_RTBCOUNT;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Ordinary allocation.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
else {
|
|
|
|
do_fx = 1;
|
|
|
|
nblks = del->br_blockcount;
|
|
|
|
qfield = XFS_TRANS_DQ_BCOUNT;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Set up del_endblock and cur for later.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
del_endblock = del->br_startblock + del->br_blockcount;
|
|
|
|
if (cur) {
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur, got.br_startoff,
|
|
|
|
got.br_startblock, got.br_blockcount,
|
|
|
|
&i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
da_old = da_new = 0;
|
|
|
|
} else {
|
|
|
|
da_old = startblockval(got.br_startblock);
|
|
|
|
da_new = 0;
|
|
|
|
nblks = 0;
|
|
|
|
do_fx = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2016-08-03 01:33:42 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Set flag value to use in switch statement.
|
|
|
|
* Left-contig is 2, right-contig is 1.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
switch (((got.br_startoff == del->br_startoff) << 1) |
|
|
|
|
(got_endoff == del_endoff)) {
|
|
|
|
case 3:
|
|
|
|
/*
|
|
|
|
* Matches the whole extent. Delete the entry.
|
|
|
|
*/
|
2016-10-03 16:11:35 +00:00
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_iext_remove(ip, *idx, 1,
|
|
|
|
whichfork == XFS_ATTR_FORK ? BMAP_ATTRFORK : 0);
|
|
|
|
--*idx;
|
|
|
|
if (delay)
|
|
|
|
break;
|
|
|
|
|
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork,
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) - 1);
|
|
|
|
flags |= XFS_ILOG_CORE;
|
|
|
|
if (!cur) {
|
|
|
|
flags |= xfs_ilog_fext(whichfork);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if ((error = xfs_btree_delete(cur, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
break;
|
|
|
|
|
|
|
|
case 2:
|
|
|
|
/*
|
|
|
|
* Deleting the first part of the extent.
|
|
|
|
*/
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_startoff(ep, del_endoff);
|
|
|
|
temp = got.br_blockcount - del->br_blockcount;
|
|
|
|
xfs_bmbt_set_blockcount(ep, temp);
|
|
|
|
if (delay) {
|
|
|
|
temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp),
|
|
|
|
da_old);
|
|
|
|
xfs_bmbt_set_startblock(ep, nullstartblock((int)temp));
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
da_new = temp;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
xfs_bmbt_set_startblock(ep, del_endblock);
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
if (!cur) {
|
|
|
|
flags |= xfs_ilog_fext(whichfork);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if ((error = xfs_bmbt_update(cur, del_endoff, del_endblock,
|
|
|
|
got.br_blockcount - del->br_blockcount,
|
|
|
|
got.br_state)))
|
|
|
|
goto done;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case 1:
|
|
|
|
/*
|
|
|
|
* Deleting the last part of the extent.
|
|
|
|
*/
|
|
|
|
temp = got.br_blockcount - del->br_blockcount;
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(ep, temp);
|
|
|
|
if (delay) {
|
|
|
|
temp = XFS_FILBLKS_MIN(xfs_bmap_worst_indlen(ip, temp),
|
|
|
|
da_old);
|
|
|
|
xfs_bmbt_set_startblock(ep, nullstartblock((int)temp));
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
da_new = temp;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
if (!cur) {
|
|
|
|
flags |= xfs_ilog_fext(whichfork);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if ((error = xfs_bmbt_update(cur, got.br_startoff,
|
|
|
|
got.br_startblock,
|
|
|
|
got.br_blockcount - del->br_blockcount,
|
|
|
|
got.br_state)))
|
|
|
|
goto done;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case 0:
|
|
|
|
/*
|
|
|
|
* Deleting the middle of the extent.
|
|
|
|
*/
|
|
|
|
temp = del->br_startoff - got.br_startoff;
|
|
|
|
trace_xfs_bmap_pre_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_bmbt_set_blockcount(ep, temp);
|
|
|
|
new.br_startoff = del_endoff;
|
|
|
|
temp2 = got_endoff - del_endoff;
|
|
|
|
new.br_blockcount = temp2;
|
|
|
|
new.br_state = got.br_state;
|
|
|
|
if (!delay) {
|
|
|
|
new.br_startblock = del_endblock;
|
|
|
|
flags |= XFS_ILOG_CORE;
|
|
|
|
if (cur) {
|
|
|
|
if ((error = xfs_bmbt_update(cur,
|
|
|
|
got.br_startoff,
|
|
|
|
got.br_startblock, temp,
|
|
|
|
got.br_state)))
|
|
|
|
goto done;
|
|
|
|
if ((error = xfs_btree_increment(cur, 0, &i)))
|
|
|
|
goto done;
|
|
|
|
cur->bc_rec.b = new;
|
|
|
|
error = xfs_btree_insert(cur, &i);
|
2014-06-25 04:58:08 +00:00
|
|
|
if (error && error != -ENOSPC)
|
2013-02-25 01:31:26 +00:00
|
|
|
goto done;
|
|
|
|
/*
|
|
|
|
* If get no-space back from btree insert,
|
|
|
|
* it tried a split, and we have a zero
|
|
|
|
* block reservation.
|
|
|
|
* Fix up our state and return the error.
|
|
|
|
*/
|
2014-06-25 04:58:08 +00:00
|
|
|
if (error == -ENOSPC) {
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Reset the cursor, don't trust
|
|
|
|
* it after any insert operation.
|
|
|
|
*/
|
|
|
|
if ((error = xfs_bmbt_lookup_eq(cur,
|
|
|
|
got.br_startoff,
|
|
|
|
got.br_startblock,
|
|
|
|
temp, &i)))
|
|
|
|
goto done;
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp,
|
|
|
|
i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Update the btree record back
|
|
|
|
* to the original value.
|
|
|
|
*/
|
|
|
|
if ((error = xfs_bmbt_update(cur,
|
|
|
|
got.br_startoff,
|
|
|
|
got.br_startblock,
|
|
|
|
got.br_blockcount,
|
|
|
|
got.br_state)))
|
|
|
|
goto done;
|
|
|
|
/*
|
|
|
|
* Reset the extent record back
|
|
|
|
* to the original value.
|
|
|
|
*/
|
|
|
|
xfs_bmbt_set_blockcount(ep,
|
|
|
|
got.br_blockcount);
|
|
|
|
flags = 0;
|
2014-06-25 04:58:08 +00:00
|
|
|
error = -ENOSPC;
|
2013-02-25 01:31:26 +00:00
|
|
|
goto done;
|
|
|
|
}
|
2015-02-23 11:39:08 +00:00
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, done);
|
2013-02-25 01:31:26 +00:00
|
|
|
} else
|
|
|
|
flags |= xfs_ilog_fext(whichfork);
|
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork,
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) + 1);
|
|
|
|
} else {
|
2016-03-15 00:42:47 +00:00
|
|
|
xfs_filblks_t stolen;
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(whichfork == XFS_DATA_FORK);
|
2016-03-15 00:42:46 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Distribute the original indlen reservation across the
|
2016-03-15 00:42:47 +00:00
|
|
|
* two new extents. Steal blocks from the deleted extent
|
|
|
|
* if necessary. Stealing blocks simply fudges the
|
|
|
|
* fdblocks accounting in xfs_bunmapi().
|
2016-03-15 00:42:46 +00:00
|
|
|
*/
|
|
|
|
temp = xfs_bmap_worst_indlen(ip, got.br_blockcount);
|
|
|
|
temp2 = xfs_bmap_worst_indlen(ip, new.br_blockcount);
|
2016-03-15 00:42:47 +00:00
|
|
|
stolen = xfs_bmap_split_indlen(da_old, &temp, &temp2,
|
|
|
|
del->br_blockcount);
|
|
|
|
da_new = temp + temp2 - stolen;
|
|
|
|
del->br_blockcount -= stolen;
|
2016-03-15 00:42:46 +00:00
|
|
|
|
|
|
|
/*
|
2016-03-15 00:42:47 +00:00
|
|
|
* Set the reservation for each extent. Warn if either
|
|
|
|
* is zero as this can lead to delalloc problems.
|
2016-03-15 00:42:46 +00:00
|
|
|
*/
|
2016-03-15 00:42:47 +00:00
|
|
|
WARN_ON_ONCE(!temp || !temp2);
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_bmbt_set_startblock(ep, nullstartblock((int)temp));
|
|
|
|
new.br_startblock = nullstartblock((int)temp2);
|
|
|
|
}
|
|
|
|
trace_xfs_bmap_post_update(ip, *idx, state, _THIS_IP_);
|
|
|
|
xfs_iext_insert(ip, *idx + 1, 1, &new, state);
|
|
|
|
++*idx;
|
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
xfs: propagate bmap updates to rmapbt
When we map, unmap, or convert an extent in a file's data or attr
fork, schedule a respective update in the rmapbt. Previous versions
of this patch required a 1:1 correspondence between bmap and rmap,
but this is no longer true as we now have ability to make interval
queries against the rmapbt.
We use the deferred operations code to handle redo operations
atomically and deadlock free. This plumbs in all five rmap actions
(map, unmap, convert extent, alloc, free); we'll use the first three
now for file data, and reflink will want the last two. We also add
an error injection site to test log recovery.
Finally, we need to fix the bmap shift extent code to adjust the
rmaps correctly.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-08-03 02:16:05 +00:00
|
|
|
|
|
|
|
/* remove reverse mapping */
|
|
|
|
if (!delay) {
|
|
|
|
error = xfs_rmap_unmap_extent(mp, dfops, ip, whichfork, del);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* If we need to, add to list of extents to delete.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2016-10-03 16:11:27 +00:00
|
|
|
if (do_fx && !(bflags & XFS_BMAPI_REMAP)) {
|
2016-10-03 16:11:23 +00:00
|
|
|
if (xfs_is_reflink_inode(ip) && whichfork == XFS_DATA_FORK) {
|
|
|
|
error = xfs_refcount_decrease_extent(mp, dfops, del);
|
|
|
|
if (error)
|
|
|
|
goto done;
|
|
|
|
} else
|
|
|
|
xfs_bmap_add_free(mp, dfops, del->br_startblock,
|
|
|
|
del->br_blockcount, NULL);
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Adjust inode # blocks in the file.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (nblks)
|
|
|
|
ip->i_d.di_nblocks -= nblks;
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Adjust quota data.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2016-10-03 16:11:27 +00:00
|
|
|
if (qfield && !(bflags & XFS_BMAPI_REMAP))
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_trans_mod_dquot_byino(tp, ip, qfield, (long)-nblks);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Account for change in delayed indirect blocks.
|
|
|
|
* Nothing to do for disk quota accounting here.
|
|
|
|
*/
|
|
|
|
ASSERT(da_old >= da_new);
|
2015-02-23 10:22:03 +00:00
|
|
|
if (da_old > da_new)
|
|
|
|
xfs_mod_fdblocks(mp, (int64_t)(da_old - da_new), false);
|
2013-02-25 01:31:26 +00:00
|
|
|
done:
|
|
|
|
*logflagsp = flags;
|
2005-04-16 22:20:36 +00:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2007-08-16 05:20:25 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Unmap (remove) blocks from a file.
|
|
|
|
* If nexts is nonzero then the number of extents to remove is limited to
|
|
|
|
* that value. If not all extents in the block range can be removed then
|
|
|
|
* *done is set.
|
2007-08-16 05:20:25 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
int /* error */
|
2016-10-03 16:11:29 +00:00
|
|
|
__xfs_bunmapi(
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_trans_t *tp, /* transaction pointer */
|
|
|
|
struct xfs_inode *ip, /* incore inode */
|
|
|
|
xfs_fileoff_t bno, /* starting offset to unmap */
|
2016-10-03 16:11:29 +00:00
|
|
|
xfs_filblks_t *rlen, /* i/o: amount remaining */
|
2013-02-25 01:31:26 +00:00
|
|
|
int flags, /* misc flags */
|
|
|
|
xfs_extnum_t nexts, /* number of extents max */
|
|
|
|
xfs_fsblock_t *firstblock, /* first allocated block
|
|
|
|
controls a.g. for allocs */
|
2016-10-03 16:11:29 +00:00
|
|
|
struct xfs_defer_ops *dfops) /* i/o: deferred updates */
|
2007-08-16 05:20:25 +00:00
|
|
|
{
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_btree_cur_t *cur; /* bmap btree cursor */
|
|
|
|
xfs_bmbt_irec_t del; /* extent being deleted */
|
|
|
|
int error; /* error return value */
|
|
|
|
xfs_extnum_t extno; /* extent number in list */
|
|
|
|
xfs_bmbt_irec_t got; /* current extent record */
|
2008-11-28 03:23:35 +00:00
|
|
|
xfs_ifork_t *ifp; /* inode fork pointer */
|
2013-02-25 01:31:26 +00:00
|
|
|
int isrt; /* freeing in rt area */
|
|
|
|
xfs_extnum_t lastx; /* last extent index used */
|
|
|
|
int logflags; /* transaction logging flags */
|
|
|
|
xfs_extlen_t mod; /* rt extent offset */
|
|
|
|
xfs_mount_t *mp; /* mount structure */
|
|
|
|
xfs_fileoff_t start; /* first file offset deleted */
|
|
|
|
int tmp_logflags; /* partial logging flags */
|
|
|
|
int wasdel; /* was a delayed alloc extent */
|
|
|
|
int whichfork; /* data or attribute fork */
|
|
|
|
xfs_fsblock_t sum;
|
2016-10-03 16:11:29 +00:00
|
|
|
xfs_filblks_t len = *rlen; /* length to unmap in file */
|
2017-06-15 04:25:57 +00:00
|
|
|
xfs_fileoff_t max_len;
|
2017-07-18 18:16:51 +00:00
|
|
|
xfs_agnumber_t prev_agno = NULLAGNUMBER, agno;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
trace_xfs_bunmap(ip, bno, len, flags, _RET_IP_);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2016-10-03 16:11:32 +00:00
|
|
|
whichfork = xfs_bmapi_whichfork(flags);
|
|
|
|
ASSERT(whichfork != XFS_COW_FORK);
|
2013-02-25 01:31:26 +00:00
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
if (unlikely(
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS &&
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE)) {
|
|
|
|
XFS_ERROR_REPORT("xfs_bunmapi", XFS_ERRLEVEL_LOW,
|
|
|
|
ip->i_mount);
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EFSCORRUPTED;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
mp = ip->i_mount;
|
|
|
|
if (XFS_FORCED_SHUTDOWN(mp))
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EIO;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-12-06 20:30:17 +00:00
|
|
|
ASSERT(xfs_isilocked(ip, XFS_ILOCK_EXCL));
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(len > 0);
|
|
|
|
ASSERT(nexts >= 0);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2017-06-15 04:25:57 +00:00
|
|
|
/*
|
|
|
|
* Guesstimate how many blocks we can unmap without running the risk of
|
|
|
|
* blowing out the transaction with a mix of EFIs and reflink
|
|
|
|
* adjustments.
|
|
|
|
*/
|
|
|
|
if (xfs_is_reflink_inode(ip) && whichfork == XFS_DATA_FORK)
|
|
|
|
max_len = min(len, xfs_refcount_max_unmap(tp->t_log_res));
|
|
|
|
else
|
|
|
|
max_len = len;
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
if (!(ifp->if_flags & XFS_IFEXTENTS) &&
|
|
|
|
(error = xfs_iread_extents(tp, ip, whichfork)))
|
|
|
|
return error;
|
2016-11-08 01:59:42 +00:00
|
|
|
if (xfs_iext_count(ifp) == 0) {
|
2016-10-03 16:11:29 +00:00
|
|
|
*rlen = 0;
|
2013-02-25 01:31:26 +00:00
|
|
|
return 0;
|
|
|
|
}
|
2015-10-12 07:21:22 +00:00
|
|
|
XFS_STATS_INC(mp, xs_blk_unmap);
|
2013-02-25 01:31:26 +00:00
|
|
|
isrt = (whichfork == XFS_DATA_FORK) && XFS_IS_REALTIME_INODE(ip);
|
|
|
|
start = bno;
|
|
|
|
bno = start + len - 1;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Check to see if the given block number is past the end of the
|
|
|
|
* file, back up to the last block if so...
|
|
|
|
*/
|
2016-11-24 00:39:44 +00:00
|
|
|
if (!xfs_iext_lookup_extent(ip, ifp, bno, &lastx, &got)) {
|
|
|
|
ASSERT(lastx > 0);
|
|
|
|
xfs_iext_get_extent(ifp, --lastx, &got);
|
2013-02-25 01:31:26 +00:00
|
|
|
bno = got.br_startoff + got.br_blockcount - 1;
|
|
|
|
}
|
2016-11-24 00:39:44 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
logflags = 0;
|
|
|
|
if (ifp->if_flags & XFS_IFBROOT) {
|
|
|
|
ASSERT(XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_BTREE);
|
|
|
|
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
|
|
|
|
cur->bc_private.b.firstblock = *firstblock;
|
2016-08-03 01:19:29 +00:00
|
|
|
cur->bc_private.b.dfops = dfops;
|
2013-02-25 01:31:26 +00:00
|
|
|
cur->bc_private.b.flags = 0;
|
|
|
|
} else
|
|
|
|
cur = NULL;
|
|
|
|
|
|
|
|
if (isrt) {
|
|
|
|
/*
|
|
|
|
* Synchronize by locking the bitmap inode.
|
|
|
|
*/
|
2016-08-03 01:00:42 +00:00
|
|
|
xfs_ilock(mp->m_rbmip, XFS_ILOCK_EXCL|XFS_ILOCK_RTBITMAP);
|
2013-02-25 01:31:26 +00:00
|
|
|
xfs_trans_ijoin(tp, mp->m_rbmip, XFS_ILOCK_EXCL);
|
2016-08-03 01:00:42 +00:00
|
|
|
xfs_ilock(mp->m_rsumip, XFS_ILOCK_EXCL|XFS_ILOCK_RTSUM);
|
|
|
|
xfs_trans_ijoin(tp, mp->m_rsumip, XFS_ILOCK_EXCL);
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
2012-04-29 11:16:17 +00:00
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
extno = 0;
|
|
|
|
while (bno != (xfs_fileoff_t)-1 && bno >= start && lastx >= 0 &&
|
2017-06-15 04:25:57 +00:00
|
|
|
(nexts == 0 || extno < nexts) && max_len > 0) {
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* Is the found extent after a hole in which bno lives?
|
|
|
|
* Just back up to the previous extent, if so.
|
|
|
|
*/
|
|
|
|
if (got.br_startoff > bno) {
|
|
|
|
if (--lastx < 0)
|
|
|
|
break;
|
2016-11-24 00:39:44 +00:00
|
|
|
xfs_iext_get_extent(ifp, lastx, &got);
|
2013-02-25 01:31:26 +00:00
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Is the last block of this extent before the range
|
|
|
|
* we're supposed to delete? If so, we're done.
|
|
|
|
*/
|
|
|
|
bno = XFS_FILEOFF_MIN(bno,
|
|
|
|
got.br_startoff + got.br_blockcount - 1);
|
|
|
|
if (bno < start)
|
|
|
|
break;
|
|
|
|
/*
|
|
|
|
* Then deal with the (possibly delayed) allocated space
|
|
|
|
* we found.
|
|
|
|
*/
|
|
|
|
del = got;
|
|
|
|
wasdel = isnullstartblock(del.br_startblock);
|
2017-07-18 18:16:51 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Make sure we don't touch multiple AGF headers out of order
|
|
|
|
* in a single transaction, as that could cause AB-BA deadlocks.
|
|
|
|
*/
|
|
|
|
if (!wasdel) {
|
|
|
|
agno = XFS_FSB_TO_AGNO(mp, del.br_startblock);
|
|
|
|
if (prev_agno != NULLAGNUMBER && prev_agno > agno)
|
|
|
|
break;
|
|
|
|
prev_agno = agno;
|
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
if (got.br_startoff < start) {
|
|
|
|
del.br_startoff = start;
|
|
|
|
del.br_blockcount -= start - got.br_startoff;
|
|
|
|
if (!wasdel)
|
|
|
|
del.br_startblock += start - got.br_startoff;
|
|
|
|
}
|
|
|
|
if (del.br_startoff + del.br_blockcount > bno + 1)
|
|
|
|
del.br_blockcount = bno + 1 - del.br_startoff;
|
2017-06-15 04:25:57 +00:00
|
|
|
|
|
|
|
/* How much can we safely unmap? */
|
|
|
|
if (max_len < del.br_blockcount) {
|
|
|
|
del.br_startoff += del.br_blockcount - max_len;
|
|
|
|
if (!wasdel)
|
|
|
|
del.br_startblock += del.br_blockcount - max_len;
|
|
|
|
del.br_blockcount = max_len;
|
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
sum = del.br_startblock + del.br_blockcount;
|
|
|
|
if (isrt &&
|
|
|
|
(mod = do_mod(sum, mp->m_sb.sb_rextsize))) {
|
2012-04-29 11:16:17 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Realtime extent not lined up at the end.
|
|
|
|
* The extent could have been split into written
|
|
|
|
* and unwritten pieces, or we could just be
|
|
|
|
* unmapping part of it. But we can't really
|
|
|
|
* get rid of part of a realtime extent.
|
2012-04-29 11:16:17 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (del.br_state == XFS_EXT_UNWRITTEN ||
|
|
|
|
!xfs_sb_version_hasextflgbit(&mp->m_sb)) {
|
|
|
|
/*
|
|
|
|
* This piece is unwritten, or we're not
|
|
|
|
* using unwritten extents. Skip over it.
|
|
|
|
*/
|
|
|
|
ASSERT(bno >= mod);
|
|
|
|
bno -= mod > del.br_blockcount ?
|
|
|
|
del.br_blockcount : mod;
|
|
|
|
if (bno < got.br_startoff) {
|
|
|
|
if (--lastx >= 0)
|
|
|
|
xfs_bmbt_get_all(xfs_iext_get_ext(
|
|
|
|
ifp, lastx), &got);
|
|
|
|
}
|
|
|
|
continue;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
xfs: Make fiemap work with sparse files
In xfs_vn_fiemap, we set bvm_count to fi_extent_max + 1 and want
to return fi_extent_max extents, but actually it won't work for
a sparse file. The reason is that in xfs_getbmap we will
calculate holes and set it in 'out', while out is malloced by
bmv_count(fi_extent_max+1) which didn't consider holes. So in the
worst case, if 'out' vector looks like
[hole, extent, hole, extent, hole, ... hole, extent, hole],
we will only return half of fi_extent_max extents.
This patch add a new parameter BMV_IF_NO_HOLES for bvm_iflags.
So with this flags, we don't use our 'out' in xfs_getbmap for
a hole. The solution is a bit ugly by just don't increasing
index of 'out' vector. I felt that it is not easy to skip it
at the very beginning since we have the complicated check and
some function like xfs_getbmapx_fix_eof_hole to adjust 'out'.
Cc: Dave Chinner <david@fromorbit.com>
Signed-off-by: Tao Ma <tao.ma@oracle.com>
Signed-off-by: Alex Elder <aelder@sgi.com>
2010-08-30 02:44:03 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* It's written, turn it unwritten.
|
|
|
|
* This is better than zeroing it.
|
xfs: Make fiemap work with sparse files
In xfs_vn_fiemap, we set bvm_count to fi_extent_max + 1 and want
to return fi_extent_max extents, but actually it won't work for
a sparse file. The reason is that in xfs_getbmap we will
calculate holes and set it in 'out', while out is malloced by
bmv_count(fi_extent_max+1) which didn't consider holes. So in the
worst case, if 'out' vector looks like
[hole, extent, hole, extent, hole, ... hole, extent, hole],
we will only return half of fi_extent_max extents.
This patch add a new parameter BMV_IF_NO_HOLES for bvm_iflags.
So with this flags, we don't use our 'out' in xfs_getbmap for
a hole. The solution is a bit ugly by just don't increasing
index of 'out' vector. I felt that it is not easy to skip it
at the very beginning since we have the complicated check and
some function like xfs_getbmapx_fix_eof_hole to adjust 'out'.
Cc: Dave Chinner <david@fromorbit.com>
Signed-off-by: Tao Ma <tao.ma@oracle.com>
Signed-off-by: Alex Elder <aelder@sgi.com>
2010-08-30 02:44:03 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(del.br_state == XFS_EXT_NORM);
|
2016-03-01 22:58:21 +00:00
|
|
|
ASSERT(tp->t_blk_res > 0);
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* If this spans a realtime extent boundary,
|
|
|
|
* chop it back to the start of the one we end at.
|
|
|
|
*/
|
|
|
|
if (del.br_blockcount > mod) {
|
|
|
|
del.br_startoff += del.br_blockcount - mod;
|
|
|
|
del.br_startblock += del.br_blockcount - mod;
|
|
|
|
del.br_blockcount = mod;
|
|
|
|
}
|
|
|
|
del.br_state = XFS_EXT_UNWRITTEN;
|
|
|
|
error = xfs_bmap_add_extent_unwritten_real(tp, ip,
|
2017-02-02 23:14:01 +00:00
|
|
|
whichfork, &lastx, &cur, &del,
|
|
|
|
firstblock, dfops, &logflags);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (error)
|
|
|
|
goto error0;
|
|
|
|
goto nodelete;
|
|
|
|
}
|
|
|
|
if (isrt && (mod = do_mod(del.br_startblock, mp->m_sb.sb_rextsize))) {
|
|
|
|
/*
|
|
|
|
* Realtime extent is lined up at the end but not
|
|
|
|
* at the front. We'll get rid of full extents if
|
|
|
|
* we can.
|
|
|
|
*/
|
|
|
|
mod = mp->m_sb.sb_rextsize - mod;
|
|
|
|
if (del.br_blockcount > mod) {
|
|
|
|
del.br_blockcount -= mod;
|
|
|
|
del.br_startoff += mod;
|
|
|
|
del.br_startblock += mod;
|
|
|
|
} else if ((del.br_startoff == start &&
|
|
|
|
(del.br_state == XFS_EXT_UNWRITTEN ||
|
2016-03-01 22:58:21 +00:00
|
|
|
tp->t_blk_res == 0)) ||
|
2013-02-25 01:31:26 +00:00
|
|
|
!xfs_sb_version_hasextflgbit(&mp->m_sb)) {
|
|
|
|
/*
|
|
|
|
* Can't make it unwritten. There isn't
|
|
|
|
* a full extent here so just skip it.
|
|
|
|
*/
|
|
|
|
ASSERT(bno >= del.br_blockcount);
|
|
|
|
bno -= del.br_blockcount;
|
2016-11-24 00:39:44 +00:00
|
|
|
if (got.br_startoff > bno && --lastx >= 0)
|
|
|
|
xfs_iext_get_extent(ifp, lastx, &got);
|
xfs: Make fiemap work with sparse files
In xfs_vn_fiemap, we set bvm_count to fi_extent_max + 1 and want
to return fi_extent_max extents, but actually it won't work for
a sparse file. The reason is that in xfs_getbmap we will
calculate holes and set it in 'out', while out is malloced by
bmv_count(fi_extent_max+1) which didn't consider holes. So in the
worst case, if 'out' vector looks like
[hole, extent, hole, extent, hole, ... hole, extent, hole],
we will only return half of fi_extent_max extents.
This patch add a new parameter BMV_IF_NO_HOLES for bvm_iflags.
So with this flags, we don't use our 'out' in xfs_getbmap for
a hole. The solution is a bit ugly by just don't increasing
index of 'out' vector. I felt that it is not easy to skip it
at the very beginning since we have the complicated check and
some function like xfs_getbmapx_fix_eof_hole to adjust 'out'.
Cc: Dave Chinner <david@fromorbit.com>
Signed-off-by: Tao Ma <tao.ma@oracle.com>
Signed-off-by: Alex Elder <aelder@sgi.com>
2010-08-30 02:44:03 +00:00
|
|
|
continue;
|
2013-02-25 01:31:26 +00:00
|
|
|
} else if (del.br_state == XFS_EXT_UNWRITTEN) {
|
2016-11-24 00:39:44 +00:00
|
|
|
struct xfs_bmbt_irec prev;
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* This one is already unwritten.
|
|
|
|
* It must have a written left neighbor.
|
|
|
|
* Unwrite the killed part of that one and
|
|
|
|
* try again.
|
|
|
|
*/
|
|
|
|
ASSERT(lastx > 0);
|
2016-11-24 00:39:44 +00:00
|
|
|
xfs_iext_get_extent(ifp, lastx - 1, &prev);
|
2013-02-25 01:31:26 +00:00
|
|
|
ASSERT(prev.br_state == XFS_EXT_NORM);
|
|
|
|
ASSERT(!isnullstartblock(prev.br_startblock));
|
|
|
|
ASSERT(del.br_startblock ==
|
|
|
|
prev.br_startblock + prev.br_blockcount);
|
|
|
|
if (prev.br_startoff < start) {
|
|
|
|
mod = start - prev.br_startoff;
|
|
|
|
prev.br_blockcount -= mod;
|
|
|
|
prev.br_startblock += mod;
|
|
|
|
prev.br_startoff = start;
|
|
|
|
}
|
|
|
|
prev.br_state = XFS_EXT_UNWRITTEN;
|
|
|
|
lastx--;
|
|
|
|
error = xfs_bmap_add_extent_unwritten_real(tp,
|
2017-02-02 23:14:01 +00:00
|
|
|
ip, whichfork, &lastx, &cur,
|
|
|
|
&prev, firstblock, dfops,
|
|
|
|
&logflags);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (error)
|
|
|
|
goto error0;
|
|
|
|
goto nodelete;
|
|
|
|
} else {
|
|
|
|
ASSERT(del.br_state == XFS_EXT_NORM);
|
|
|
|
del.br_state = XFS_EXT_UNWRITTEN;
|
|
|
|
error = xfs_bmap_add_extent_unwritten_real(tp,
|
2017-02-02 23:14:01 +00:00
|
|
|
ip, whichfork, &lastx, &cur,
|
|
|
|
&del, firstblock, dfops,
|
|
|
|
&logflags);
|
2013-02-25 01:31:26 +00:00
|
|
|
if (error)
|
|
|
|
goto error0;
|
|
|
|
goto nodelete;
|
xfs: Make fiemap work with sparse files
In xfs_vn_fiemap, we set bvm_count to fi_extent_max + 1 and want
to return fi_extent_max extents, but actually it won't work for
a sparse file. The reason is that in xfs_getbmap we will
calculate holes and set it in 'out', while out is malloced by
bmv_count(fi_extent_max+1) which didn't consider holes. So in the
worst case, if 'out' vector looks like
[hole, extent, hole, extent, hole, ... hole, extent, hole],
we will only return half of fi_extent_max extents.
This patch add a new parameter BMV_IF_NO_HOLES for bvm_iflags.
So with this flags, we don't use our 'out' in xfs_getbmap for
a hole. The solution is a bit ugly by just don't increasing
index of 'out' vector. I felt that it is not easy to skip it
at the very beginning since we have the complicated check and
some function like xfs_getbmapx_fix_eof_hole to adjust 'out'.
Cc: Dave Chinner <david@fromorbit.com>
Signed-off-by: Tao Ma <tao.ma@oracle.com>
Signed-off-by: Alex Elder <aelder@sgi.com>
2010-08-30 02:44:03 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2013-02-25 01:31:26 +00:00
|
|
|
/*
|
|
|
|
* If it's the case where the directory code is running
|
|
|
|
* with no block reservation, and the deleted block is in
|
|
|
|
* the middle of its extent, and the resulting insert
|
|
|
|
* of an extent would cause transformation to btree format,
|
|
|
|
* then reject it. The calling code will then swap
|
|
|
|
* blocks around instead.
|
|
|
|
* We have to do this now, rather than waiting for the
|
|
|
|
* conversion to btree format, since the transaction
|
|
|
|
* will be dirty.
|
|
|
|
*/
|
2016-03-01 22:58:21 +00:00
|
|
|
if (!wasdel && tp->t_blk_res == 0 &&
|
2013-02-25 01:31:26 +00:00
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) == XFS_DINODE_FMT_EXTENTS &&
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) >= /* Note the >= */
|
|
|
|
XFS_IFORK_MAXEXT(ip, whichfork) &&
|
|
|
|
del.br_startoff > got.br_startoff &&
|
|
|
|
del.br_startoff + del.br_blockcount <
|
|
|
|
got.br_startoff + got.br_blockcount) {
|
2014-06-25 04:58:08 +00:00
|
|
|
error = -ENOSPC;
|
2013-02-25 01:31:26 +00:00
|
|
|
goto error0;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2016-03-15 00:42:46 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Unreserve quota and update realtime free space, if
|
|
|
|
* appropriate. If delayed allocation, update the inode delalloc
|
|
|
|
* counter now and wait to update the sb counters as
|
|
|
|
* xfs_bmap_del_extent() might need to borrow some blocks.
|
|
|
|
*/
|
|
|
|
if (wasdel) {
|
|
|
|
ASSERT(startblockval(del.br_startblock) > 0);
|
|
|
|
if (isrt) {
|
|
|
|
xfs_filblks_t rtexts;
|
|
|
|
|
|
|
|
rtexts = XFS_FSB_TO_B(mp, del.br_blockcount);
|
|
|
|
do_div(rtexts, mp->m_sb.sb_rextsize);
|
|
|
|
xfs_mod_frextents(mp, (int64_t)rtexts);
|
|
|
|
(void)xfs_trans_reserve_quota_nblks(NULL,
|
|
|
|
ip, -((long)del.br_blockcount), 0,
|
|
|
|
XFS_QMOPT_RES_RTBLKS);
|
|
|
|
} else {
|
|
|
|
(void)xfs_trans_reserve_quota_nblks(NULL,
|
|
|
|
ip, -((long)del.br_blockcount), 0,
|
|
|
|
XFS_QMOPT_RES_REGBLKS);
|
|
|
|
}
|
|
|
|
ip->i_delayed_blks -= del.br_blockcount;
|
|
|
|
if (cur)
|
|
|
|
cur->bc_private.b.flags |=
|
|
|
|
XFS_BTCUR_BPRV_WASDEL;
|
|
|
|
} else if (cur)
|
|
|
|
cur->bc_private.b.flags &= ~XFS_BTCUR_BPRV_WASDEL;
|
|
|
|
|
2016-08-03 01:19:29 +00:00
|
|
|
error = xfs_bmap_del_extent(ip, tp, &lastx, dfops, cur, &del,
|
2016-10-03 16:11:27 +00:00
|
|
|
&tmp_logflags, whichfork, flags);
|
2013-02-25 01:31:26 +00:00
|
|
|
logflags |= tmp_logflags;
|
|
|
|
if (error)
|
|
|
|
goto error0;
|
2016-03-15 00:42:46 +00:00
|
|
|
|
|
|
|
if (!isrt && wasdel)
|
|
|
|
xfs_mod_fdblocks(mp, (int64_t)del.br_blockcount, false);
|
|
|
|
|
2017-06-15 04:25:57 +00:00
|
|
|
max_len -= del.br_blockcount;
|
2013-02-25 01:31:26 +00:00
|
|
|
bno = del.br_startoff - 1;
|
|
|
|
nodelete:
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* If not done go on to the next (previous) record.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (bno != (xfs_fileoff_t)-1 && bno >= start) {
|
|
|
|
if (lastx >= 0) {
|
2016-11-24 00:39:44 +00:00
|
|
|
xfs_iext_get_extent(ifp, lastx, &got);
|
|
|
|
if (got.br_startoff > bno && --lastx >= 0)
|
|
|
|
xfs_iext_get_extent(ifp, lastx, &got);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2013-02-25 01:31:26 +00:00
|
|
|
extno++;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
}
|
2016-10-03 16:11:29 +00:00
|
|
|
if (bno == (xfs_fileoff_t)-1 || bno < start || lastx < 0)
|
|
|
|
*rlen = 0;
|
|
|
|
else
|
|
|
|
*rlen = bno - start + 1;
|
2006-09-28 00:58:06 +00:00
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* Convert to a btree if necessary.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
if (xfs_bmap_needs_btree(ip, whichfork)) {
|
|
|
|
ASSERT(cur == NULL);
|
2016-08-03 01:19:29 +00:00
|
|
|
error = xfs_bmap_extents_to_btree(tp, ip, firstblock, dfops,
|
2013-02-25 01:31:26 +00:00
|
|
|
&cur, 0, &tmp_logflags, whichfork);
|
|
|
|
logflags |= tmp_logflags;
|
|
|
|
if (error)
|
|
|
|
goto error0;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* transform from btree to extents, give it cur
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
else if (xfs_bmap_wants_extents(ip, whichfork)) {
|
|
|
|
ASSERT(cur != NULL);
|
|
|
|
error = xfs_bmap_btree_to_extents(tp, ip, cur, &tmp_logflags,
|
|
|
|
whichfork);
|
|
|
|
logflags |= tmp_logflags;
|
|
|
|
if (error)
|
|
|
|
goto error0;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2013-02-25 01:31:26 +00:00
|
|
|
* transform from extents to local?
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2013-02-25 01:31:26 +00:00
|
|
|
error = 0;
|
|
|
|
error0:
|
|
|
|
/*
|
|
|
|
* Log everything. Do this after conversion, there's no point in
|
|
|
|
* logging the extent records if we've converted to btree format.
|
|
|
|
*/
|
|
|
|
if ((logflags & xfs_ilog_fext(whichfork)) &&
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS)
|
|
|
|
logflags &= ~xfs_ilog_fext(whichfork);
|
|
|
|
else if ((logflags & xfs_ilog_fbroot(whichfork)) &&
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE)
|
|
|
|
logflags &= ~xfs_ilog_fbroot(whichfork);
|
|
|
|
/*
|
|
|
|
* Log inode even in the error case, if the transaction
|
|
|
|
* is dirty we'll need to shut down the filesystem.
|
|
|
|
*/
|
|
|
|
if (logflags)
|
|
|
|
xfs_trans_log_inode(tp, ip, logflags);
|
|
|
|
if (cur) {
|
|
|
|
if (!error) {
|
|
|
|
*firstblock = cur->bc_private.b.firstblock;
|
|
|
|
cur->bc_private.b.allocated = 0;
|
|
|
|
}
|
|
|
|
xfs_btree_del_cursor(cur,
|
|
|
|
error ? XFS_BTREE_ERROR : XFS_BTREE_NOERROR);
|
|
|
|
}
|
|
|
|
return error;
|
|
|
|
}
|
2014-02-23 23:58:19 +00:00
|
|
|
|
2016-10-03 16:11:29 +00:00
|
|
|
/* Unmap a range of a file. */
|
|
|
|
int
|
|
|
|
xfs_bunmapi(
|
|
|
|
xfs_trans_t *tp,
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
xfs_fileoff_t bno,
|
|
|
|
xfs_filblks_t len,
|
|
|
|
int flags,
|
|
|
|
xfs_extnum_t nexts,
|
|
|
|
xfs_fsblock_t *firstblock,
|
|
|
|
struct xfs_defer_ops *dfops,
|
|
|
|
int *done)
|
|
|
|
{
|
|
|
|
int error;
|
|
|
|
|
|
|
|
error = __xfs_bunmapi(tp, ip, bno, &len, flags, nexts, firstblock,
|
|
|
|
dfops);
|
|
|
|
*done = (len == 0);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2014-09-23 05:38:09 +00:00
|
|
|
/*
|
|
|
|
* Determine whether an extent shift can be accomplished by a merge with the
|
|
|
|
* extent that precedes the target hole of the shift.
|
|
|
|
*/
|
|
|
|
STATIC bool
|
|
|
|
xfs_bmse_can_merge(
|
|
|
|
struct xfs_bmbt_irec *left, /* preceding extent */
|
|
|
|
struct xfs_bmbt_irec *got, /* current extent to shift */
|
|
|
|
xfs_fileoff_t shift) /* shift fsb */
|
|
|
|
{
|
|
|
|
xfs_fileoff_t startoff;
|
|
|
|
|
|
|
|
startoff = got->br_startoff - shift;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The extent, once shifted, must be adjacent in-file and on-disk with
|
|
|
|
* the preceding extent.
|
|
|
|
*/
|
|
|
|
if ((left->br_startoff + left->br_blockcount != startoff) ||
|
|
|
|
(left->br_startblock + left->br_blockcount != got->br_startblock) ||
|
|
|
|
(left->br_state != got->br_state) ||
|
|
|
|
(left->br_blockcount + got->br_blockcount > MAXEXTLEN))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* A bmap extent shift adjusts the file offset of an extent to fill a preceding
|
|
|
|
* hole in the file. If an extent shift would result in the extent being fully
|
|
|
|
* adjacent to the extent that currently precedes the hole, we can merge with
|
|
|
|
* the preceding extent rather than do the shift.
|
|
|
|
*
|
|
|
|
* This function assumes the caller has verified a shift-by-merge is possible
|
|
|
|
* with the provided extents via xfs_bmse_can_merge().
|
|
|
|
*/
|
|
|
|
STATIC int
|
|
|
|
xfs_bmse_merge(
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
int whichfork,
|
|
|
|
xfs_fileoff_t shift, /* shift fsb */
|
|
|
|
int current_ext, /* idx of gotp */
|
|
|
|
struct xfs_bmbt_rec_host *gotp, /* extent to shift */
|
|
|
|
struct xfs_bmbt_rec_host *leftp, /* preceding extent */
|
|
|
|
struct xfs_btree_cur *cur,
|
|
|
|
int *logflags) /* output */
|
|
|
|
{
|
|
|
|
struct xfs_bmbt_irec got;
|
|
|
|
struct xfs_bmbt_irec left;
|
|
|
|
xfs_filblks_t blockcount;
|
|
|
|
int error, i;
|
2015-02-23 11:39:13 +00:00
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
2014-09-23 05:38:09 +00:00
|
|
|
|
|
|
|
xfs_bmbt_get_all(gotp, &got);
|
|
|
|
xfs_bmbt_get_all(leftp, &left);
|
|
|
|
blockcount = left.br_blockcount + got.br_blockcount;
|
|
|
|
|
|
|
|
ASSERT(xfs_isilocked(ip, XFS_IOLOCK_EXCL));
|
|
|
|
ASSERT(xfs_isilocked(ip, XFS_ILOCK_EXCL));
|
|
|
|
ASSERT(xfs_bmse_can_merge(&left, &got, shift));
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Merge the in-core extents. Note that the host record pointers and
|
|
|
|
* current_ext index are invalid once the extent has been removed via
|
|
|
|
* xfs_iext_remove().
|
|
|
|
*/
|
|
|
|
xfs_bmbt_set_blockcount(leftp, blockcount);
|
|
|
|
xfs_iext_remove(ip, current_ext, 1, 0);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Update the on-disk extent count, the btree if necessary and log the
|
|
|
|
* inode.
|
|
|
|
*/
|
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork,
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) - 1);
|
|
|
|
*logflags |= XFS_ILOG_CORE;
|
|
|
|
if (!cur) {
|
|
|
|
*logflags |= XFS_ILOG_DEXT;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* lookup and remove the extent to merge */
|
|
|
|
error = xfs_bmbt_lookup_eq(cur, got.br_startoff, got.br_startblock,
|
|
|
|
got.br_blockcount, &i);
|
|
|
|
if (error)
|
2014-12-03 22:42:40 +00:00
|
|
|
return error;
|
2015-02-23 11:39:13 +00:00
|
|
|
XFS_WANT_CORRUPTED_RETURN(mp, i == 1);
|
2014-09-23 05:38:09 +00:00
|
|
|
|
|
|
|
error = xfs_btree_delete(cur, &i);
|
|
|
|
if (error)
|
2014-12-03 22:42:40 +00:00
|
|
|
return error;
|
2015-02-23 11:39:13 +00:00
|
|
|
XFS_WANT_CORRUPTED_RETURN(mp, i == 1);
|
2014-09-23 05:38:09 +00:00
|
|
|
|
|
|
|
/* lookup and update size of the previous extent */
|
|
|
|
error = xfs_bmbt_lookup_eq(cur, left.br_startoff, left.br_startblock,
|
|
|
|
left.br_blockcount, &i);
|
|
|
|
if (error)
|
2014-12-03 22:42:40 +00:00
|
|
|
return error;
|
2015-02-23 11:39:13 +00:00
|
|
|
XFS_WANT_CORRUPTED_RETURN(mp, i == 1);
|
2014-09-23 05:38:09 +00:00
|
|
|
|
|
|
|
left.br_blockcount = blockcount;
|
|
|
|
|
2014-12-03 22:42:40 +00:00
|
|
|
return xfs_bmbt_update(cur, left.br_startoff, left.br_startblock,
|
|
|
|
left.br_blockcount, left.br_state);
|
2014-09-23 05:38:09 +00:00
|
|
|
}
|
|
|
|
|
2014-09-23 05:39:04 +00:00
|
|
|
/*
|
|
|
|
* Shift a single extent.
|
|
|
|
*/
|
|
|
|
STATIC int
|
|
|
|
xfs_bmse_shift_one(
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
int whichfork,
|
|
|
|
xfs_fileoff_t offset_shift_fsb,
|
|
|
|
int *current_ext,
|
|
|
|
struct xfs_bmbt_rec_host *gotp,
|
|
|
|
struct xfs_btree_cur *cur,
|
2015-03-25 04:08:56 +00:00
|
|
|
int *logflags,
|
xfs: propagate bmap updates to rmapbt
When we map, unmap, or convert an extent in a file's data or attr
fork, schedule a respective update in the rmapbt. Previous versions
of this patch required a 1:1 correspondence between bmap and rmap,
but this is no longer true as we now have ability to make interval
queries against the rmapbt.
We use the deferred operations code to handle redo operations
atomically and deadlock free. This plumbs in all five rmap actions
(map, unmap, convert extent, alloc, free); we'll use the first three
now for file data, and reflink will want the last two. We also add
an error injection site to test log recovery.
Finally, we need to fix the bmap shift extent code to adjust the
rmaps correctly.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-08-03 02:16:05 +00:00
|
|
|
enum shift_direction direction,
|
|
|
|
struct xfs_defer_ops *dfops)
|
2014-09-23 05:39:04 +00:00
|
|
|
{
|
|
|
|
struct xfs_ifork *ifp;
|
2015-02-23 11:39:13 +00:00
|
|
|
struct xfs_mount *mp;
|
2014-09-23 05:39:04 +00:00
|
|
|
xfs_fileoff_t startoff;
|
2015-03-25 04:08:56 +00:00
|
|
|
struct xfs_bmbt_rec_host *adj_irecp;
|
2014-09-23 05:39:04 +00:00
|
|
|
struct xfs_bmbt_irec got;
|
2015-03-25 04:08:56 +00:00
|
|
|
struct xfs_bmbt_irec adj_irec;
|
2014-09-23 05:39:04 +00:00
|
|
|
int error;
|
|
|
|
int i;
|
2015-03-25 04:08:56 +00:00
|
|
|
int total_extents;
|
2014-09-23 05:39:04 +00:00
|
|
|
|
2015-02-23 11:39:13 +00:00
|
|
|
mp = ip->i_mount;
|
2014-09-23 05:39:04 +00:00
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
2016-11-08 01:59:42 +00:00
|
|
|
total_extents = xfs_iext_count(ifp);
|
2014-09-23 05:39:04 +00:00
|
|
|
|
|
|
|
xfs_bmbt_get_all(gotp, &got);
|
|
|
|
|
2014-09-23 05:39:05 +00:00
|
|
|
/* delalloc extents should be prevented by caller */
|
2015-02-23 11:39:13 +00:00
|
|
|
XFS_WANT_CORRUPTED_RETURN(mp, !isnullstartblock(got.br_startblock));
|
2014-09-23 05:39:05 +00:00
|
|
|
|
2015-03-25 04:08:56 +00:00
|
|
|
if (direction == SHIFT_LEFT) {
|
|
|
|
startoff = got.br_startoff - offset_shift_fsb;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check for merge if we've got an extent to the left,
|
|
|
|
* otherwise make sure there's enough room at the start
|
|
|
|
* of the file for the shift.
|
|
|
|
*/
|
|
|
|
if (!*current_ext) {
|
|
|
|
if (got.br_startoff < offset_shift_fsb)
|
|
|
|
return -EINVAL;
|
|
|
|
goto update_current_ext;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* grab the left extent and check for a large
|
|
|
|
* enough hole.
|
|
|
|
*/
|
|
|
|
adj_irecp = xfs_iext_get_ext(ifp, *current_ext - 1);
|
|
|
|
xfs_bmbt_get_all(adj_irecp, &adj_irec);
|
2014-09-23 05:39:04 +00:00
|
|
|
|
2015-03-25 04:08:56 +00:00
|
|
|
if (startoff <
|
|
|
|
adj_irec.br_startoff + adj_irec.br_blockcount)
|
2014-12-03 22:42:24 +00:00
|
|
|
return -EINVAL;
|
2014-09-23 05:39:04 +00:00
|
|
|
|
2014-12-03 22:42:24 +00:00
|
|
|
/* check whether to merge the extent or shift it down */
|
2015-03-25 04:08:56 +00:00
|
|
|
if (xfs_bmse_can_merge(&adj_irec, &got,
|
|
|
|
offset_shift_fsb)) {
|
xfs: propagate bmap updates to rmapbt
When we map, unmap, or convert an extent in a file's data or attr
fork, schedule a respective update in the rmapbt. Previous versions
of this patch required a 1:1 correspondence between bmap and rmap,
but this is no longer true as we now have ability to make interval
queries against the rmapbt.
We use the deferred operations code to handle redo operations
atomically and deadlock free. This plumbs in all five rmap actions
(map, unmap, convert extent, alloc, free); we'll use the first three
now for file data, and reflink will want the last two. We also add
an error injection site to test log recovery.
Finally, we need to fix the bmap shift extent code to adjust the
rmaps correctly.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-08-03 02:16:05 +00:00
|
|
|
error = xfs_bmse_merge(ip, whichfork, offset_shift_fsb,
|
|
|
|
*current_ext, gotp, adj_irecp,
|
|
|
|
cur, logflags);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
adj_irec = got;
|
|
|
|
goto update_rmap;
|
2014-12-03 22:42:24 +00:00
|
|
|
}
|
2015-03-25 04:08:56 +00:00
|
|
|
} else {
|
|
|
|
startoff = got.br_startoff + offset_shift_fsb;
|
|
|
|
/* nothing to move if this is the last extent */
|
|
|
|
if (*current_ext >= (total_extents - 1))
|
|
|
|
goto update_current_ext;
|
|
|
|
/*
|
|
|
|
* If this is not the last extent in the file, make sure there
|
|
|
|
* is enough room between current extent and next extent for
|
|
|
|
* accommodating the shift.
|
|
|
|
*/
|
|
|
|
adj_irecp = xfs_iext_get_ext(ifp, *current_ext + 1);
|
|
|
|
xfs_bmbt_get_all(adj_irecp, &adj_irec);
|
|
|
|
if (startoff + got.br_blockcount > adj_irec.br_startoff)
|
|
|
|
return -EINVAL;
|
|
|
|
/*
|
|
|
|
* Unlike a left shift (which involves a hole punch),
|
|
|
|
* a right shift does not modify extent neighbors
|
|
|
|
* in any way. We should never find mergeable extents
|
|
|
|
* in this scenario. Check anyways and warn if we
|
|
|
|
* encounter two extents that could be one.
|
|
|
|
*/
|
|
|
|
if (xfs_bmse_can_merge(&got, &adj_irec, offset_shift_fsb))
|
|
|
|
WARN_ON_ONCE(1);
|
|
|
|
}
|
2014-09-23 05:39:04 +00:00
|
|
|
/*
|
|
|
|
* Increment the extent index for the next iteration, update the start
|
|
|
|
* offset of the in-core extent and update the btree if applicable.
|
|
|
|
*/
|
2015-03-25 04:08:56 +00:00
|
|
|
update_current_ext:
|
|
|
|
if (direction == SHIFT_LEFT)
|
|
|
|
(*current_ext)++;
|
|
|
|
else
|
|
|
|
(*current_ext)--;
|
2014-09-23 05:39:04 +00:00
|
|
|
xfs_bmbt_set_startoff(gotp, startoff);
|
|
|
|
*logflags |= XFS_ILOG_CORE;
|
xfs: propagate bmap updates to rmapbt
When we map, unmap, or convert an extent in a file's data or attr
fork, schedule a respective update in the rmapbt. Previous versions
of this patch required a 1:1 correspondence between bmap and rmap,
but this is no longer true as we now have ability to make interval
queries against the rmapbt.
We use the deferred operations code to handle redo operations
atomically and deadlock free. This plumbs in all five rmap actions
(map, unmap, convert extent, alloc, free); we'll use the first three
now for file data, and reflink will want the last two. We also add
an error injection site to test log recovery.
Finally, we need to fix the bmap shift extent code to adjust the
rmaps correctly.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-08-03 02:16:05 +00:00
|
|
|
adj_irec = got;
|
2014-09-23 05:39:04 +00:00
|
|
|
if (!cur) {
|
|
|
|
*logflags |= XFS_ILOG_DEXT;
|
xfs: propagate bmap updates to rmapbt
When we map, unmap, or convert an extent in a file's data or attr
fork, schedule a respective update in the rmapbt. Previous versions
of this patch required a 1:1 correspondence between bmap and rmap,
but this is no longer true as we now have ability to make interval
queries against the rmapbt.
We use the deferred operations code to handle redo operations
atomically and deadlock free. This plumbs in all five rmap actions
(map, unmap, convert extent, alloc, free); we'll use the first three
now for file data, and reflink will want the last two. We also add
an error injection site to test log recovery.
Finally, we need to fix the bmap shift extent code to adjust the
rmaps correctly.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-08-03 02:16:05 +00:00
|
|
|
goto update_rmap;
|
2014-09-23 05:39:04 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
error = xfs_bmbt_lookup_eq(cur, got.br_startoff, got.br_startblock,
|
|
|
|
got.br_blockcount, &i);
|
|
|
|
if (error)
|
|
|
|
return error;
|
2015-02-23 11:39:13 +00:00
|
|
|
XFS_WANT_CORRUPTED_RETURN(mp, i == 1);
|
2014-09-23 05:39:04 +00:00
|
|
|
|
|
|
|
got.br_startoff = startoff;
|
xfs: propagate bmap updates to rmapbt
When we map, unmap, or convert an extent in a file's data or attr
fork, schedule a respective update in the rmapbt. Previous versions
of this patch required a 1:1 correspondence between bmap and rmap,
but this is no longer true as we now have ability to make interval
queries against the rmapbt.
We use the deferred operations code to handle redo operations
atomically and deadlock free. This plumbs in all five rmap actions
(map, unmap, convert extent, alloc, free); we'll use the first three
now for file data, and reflink will want the last two. We also add
an error injection site to test log recovery.
Finally, we need to fix the bmap shift extent code to adjust the
rmaps correctly.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-08-03 02:16:05 +00:00
|
|
|
error = xfs_bmbt_update(cur, got.br_startoff, got.br_startblock,
|
|
|
|
got.br_blockcount, got.br_state);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
|
|
|
|
update_rmap:
|
|
|
|
/* update reverse mapping */
|
|
|
|
error = xfs_rmap_unmap_extent(mp, dfops, ip, whichfork, &adj_irec);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
adj_irec.br_startoff = startoff;
|
|
|
|
return xfs_rmap_map_extent(mp, dfops, ip, whichfork, &adj_irec);
|
2014-09-23 05:39:04 +00:00
|
|
|
}
|
|
|
|
|
2014-02-23 23:58:19 +00:00
|
|
|
/*
|
2015-03-25 04:08:56 +00:00
|
|
|
* Shift extent records to the left/right to cover/create a hole.
|
2014-02-23 23:58:19 +00:00
|
|
|
*
|
2014-09-23 05:37:09 +00:00
|
|
|
* The maximum number of extents to be shifted in a single operation is
|
2015-03-25 04:08:56 +00:00
|
|
|
* @num_exts. @stop_fsb specifies the file offset at which to stop shift and the
|
2014-09-23 05:37:09 +00:00
|
|
|
* file offset where we've left off is returned in @next_fsb. @offset_shift_fsb
|
|
|
|
* is the length by which each extent is shifted. If there is no hole to shift
|
|
|
|
* the extents into, this will be considered invalid operation and we abort
|
|
|
|
* immediately.
|
2014-02-23 23:58:19 +00:00
|
|
|
*/
|
|
|
|
int
|
|
|
|
xfs_bmap_shift_extents(
|
|
|
|
struct xfs_trans *tp,
|
|
|
|
struct xfs_inode *ip,
|
2015-03-25 04:08:56 +00:00
|
|
|
xfs_fileoff_t *next_fsb,
|
2014-02-23 23:58:19 +00:00
|
|
|
xfs_fileoff_t offset_shift_fsb,
|
2014-09-23 05:37:09 +00:00
|
|
|
int *done,
|
2015-03-25 04:08:56 +00:00
|
|
|
xfs_fileoff_t stop_fsb,
|
2014-02-23 23:58:19 +00:00
|
|
|
xfs_fsblock_t *firstblock,
|
2016-08-03 01:19:29 +00:00
|
|
|
struct xfs_defer_ops *dfops,
|
2015-03-25 04:08:56 +00:00
|
|
|
enum shift_direction direction,
|
2014-02-23 23:58:19 +00:00
|
|
|
int num_exts)
|
|
|
|
{
|
2014-09-02 02:12:53 +00:00
|
|
|
struct xfs_btree_cur *cur = NULL;
|
2014-02-23 23:58:19 +00:00
|
|
|
struct xfs_bmbt_rec_host *gotp;
|
|
|
|
struct xfs_bmbt_irec got;
|
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
|
|
|
struct xfs_ifork *ifp;
|
|
|
|
xfs_extnum_t nexts = 0;
|
2014-09-23 05:37:09 +00:00
|
|
|
xfs_extnum_t current_ext;
|
2015-03-25 04:08:56 +00:00
|
|
|
xfs_extnum_t total_extents;
|
|
|
|
xfs_extnum_t stop_extent;
|
2014-02-23 23:58:19 +00:00
|
|
|
int error = 0;
|
|
|
|
int whichfork = XFS_DATA_FORK;
|
2014-09-02 02:12:53 +00:00
|
|
|
int logflags = 0;
|
2014-02-23 23:58:19 +00:00
|
|
|
|
|
|
|
if (unlikely(XFS_TEST_ERROR(
|
|
|
|
(XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS &&
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE),
|
2017-06-21 00:54:47 +00:00
|
|
|
mp, XFS_ERRTAG_BMAPIFORMAT))) {
|
2014-02-23 23:58:19 +00:00
|
|
|
XFS_ERROR_REPORT("xfs_bmap_shift_extents",
|
|
|
|
XFS_ERRLEVEL_LOW, mp);
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EFSCORRUPTED;
|
2014-02-23 23:58:19 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
if (XFS_FORCED_SHUTDOWN(mp))
|
2014-06-25 04:58:08 +00:00
|
|
|
return -EIO;
|
2014-02-23 23:58:19 +00:00
|
|
|
|
2014-09-23 05:37:09 +00:00
|
|
|
ASSERT(xfs_isilocked(ip, XFS_IOLOCK_EXCL));
|
|
|
|
ASSERT(xfs_isilocked(ip, XFS_ILOCK_EXCL));
|
2015-03-25 04:08:56 +00:00
|
|
|
ASSERT(direction == SHIFT_LEFT || direction == SHIFT_RIGHT);
|
|
|
|
ASSERT(*next_fsb != NULLFSBLOCK || direction == SHIFT_RIGHT);
|
2014-02-23 23:58:19 +00:00
|
|
|
|
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
if (!(ifp->if_flags & XFS_IFEXTENTS)) {
|
|
|
|
/* Read in all the extents */
|
|
|
|
error = xfs_iread_extents(tp, ip, whichfork);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2014-09-23 05:38:09 +00:00
|
|
|
if (ifp->if_flags & XFS_IFBROOT) {
|
|
|
|
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
|
|
|
|
cur->bc_private.b.firstblock = *firstblock;
|
2016-08-03 01:19:29 +00:00
|
|
|
cur->bc_private.b.dfops = dfops;
|
2014-09-23 05:38:09 +00:00
|
|
|
cur->bc_private.b.flags = 0;
|
|
|
|
}
|
|
|
|
|
2015-03-25 04:08:56 +00:00
|
|
|
/*
|
|
|
|
* There may be delalloc extents in the data fork before the range we
|
|
|
|
* are collapsing out, so we cannot use the count of real extents here.
|
|
|
|
* Instead we have to calculate it from the incore fork.
|
|
|
|
*/
|
2016-11-08 01:59:42 +00:00
|
|
|
total_extents = xfs_iext_count(ifp);
|
2015-03-25 04:08:56 +00:00
|
|
|
if (total_extents == 0) {
|
|
|
|
*done = 1;
|
|
|
|
goto del_cursor;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* In case of first right shift, we need to initialize next_fsb
|
|
|
|
*/
|
|
|
|
if (*next_fsb == NULLFSBLOCK) {
|
|
|
|
gotp = xfs_iext_get_ext(ifp, total_extents - 1);
|
|
|
|
xfs_bmbt_get_all(gotp, &got);
|
|
|
|
*next_fsb = got.br_startoff;
|
|
|
|
if (stop_fsb > *next_fsb) {
|
|
|
|
*done = 1;
|
|
|
|
goto del_cursor;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Lookup the extent index at which we have to stop */
|
|
|
|
if (direction == SHIFT_RIGHT) {
|
|
|
|
gotp = xfs_iext_bno_to_ext(ifp, stop_fsb, &stop_extent);
|
|
|
|
/* Make stop_extent exclusive of shift range */
|
|
|
|
stop_extent--;
|
|
|
|
} else
|
|
|
|
stop_extent = total_extents;
|
|
|
|
|
2014-02-23 23:58:19 +00:00
|
|
|
/*
|
2014-09-23 05:37:09 +00:00
|
|
|
* Look up the extent index for the fsb where we start shifting. We can
|
|
|
|
* henceforth iterate with current_ext as extent list changes are locked
|
|
|
|
* out via ilock.
|
|
|
|
*
|
|
|
|
* gotp can be null in 2 cases: 1) if there are no extents or 2)
|
2015-03-25 04:08:56 +00:00
|
|
|
* *next_fsb lies in a hole beyond which there are no extents. Either
|
2014-09-23 05:37:09 +00:00
|
|
|
* way, we are done.
|
2014-02-23 23:58:19 +00:00
|
|
|
*/
|
2015-03-25 04:08:56 +00:00
|
|
|
gotp = xfs_iext_bno_to_ext(ifp, *next_fsb, ¤t_ext);
|
2014-09-23 05:37:09 +00:00
|
|
|
if (!gotp) {
|
|
|
|
*done = 1;
|
2014-09-23 05:38:09 +00:00
|
|
|
goto del_cursor;
|
2014-02-23 23:58:19 +00:00
|
|
|
}
|
|
|
|
|
2015-03-25 04:08:56 +00:00
|
|
|
/* some sanity checking before we finally start shifting extents */
|
|
|
|
if ((direction == SHIFT_LEFT && current_ext >= stop_extent) ||
|
|
|
|
(direction == SHIFT_RIGHT && current_ext <= stop_extent)) {
|
|
|
|
error = -EIO;
|
|
|
|
goto del_cursor;
|
|
|
|
}
|
|
|
|
|
|
|
|
while (nexts++ < num_exts) {
|
2014-09-23 05:39:04 +00:00
|
|
|
error = xfs_bmse_shift_one(ip, whichfork, offset_shift_fsb,
|
2015-03-25 04:08:56 +00:00
|
|
|
¤t_ext, gotp, cur, &logflags,
|
xfs: propagate bmap updates to rmapbt
When we map, unmap, or convert an extent in a file's data or attr
fork, schedule a respective update in the rmapbt. Previous versions
of this patch required a 1:1 correspondence between bmap and rmap,
but this is no longer true as we now have ability to make interval
queries against the rmapbt.
We use the deferred operations code to handle redo operations
atomically and deadlock free. This plumbs in all five rmap actions
(map, unmap, convert extent, alloc, free); we'll use the first three
now for file data, and reflink will want the last two. We also add
an error injection site to test log recovery.
Finally, we need to fix the bmap shift extent code to adjust the
rmaps correctly.
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-08-03 02:16:05 +00:00
|
|
|
direction, dfops);
|
2014-09-23 05:39:04 +00:00
|
|
|
if (error)
|
|
|
|
goto del_cursor;
|
2015-03-25 04:08:56 +00:00
|
|
|
/*
|
|
|
|
* If there was an extent merge during the shift, the extent
|
|
|
|
* count can change. Update the total and grade the next record.
|
|
|
|
*/
|
|
|
|
if (direction == SHIFT_LEFT) {
|
2016-11-08 01:59:42 +00:00
|
|
|
total_extents = xfs_iext_count(ifp);
|
2015-03-25 04:08:56 +00:00
|
|
|
stop_extent = total_extents;
|
|
|
|
}
|
2014-02-23 23:58:19 +00:00
|
|
|
|
2015-03-25 04:08:56 +00:00
|
|
|
if (current_ext == stop_extent) {
|
|
|
|
*done = 1;
|
|
|
|
*next_fsb = NULLFSBLOCK;
|
2014-09-23 05:38:09 +00:00
|
|
|
break;
|
2015-03-25 04:08:56 +00:00
|
|
|
}
|
2014-09-23 05:38:09 +00:00
|
|
|
gotp = xfs_iext_get_ext(ifp, current_ext);
|
2014-02-23 23:58:19 +00:00
|
|
|
}
|
|
|
|
|
2015-03-25 04:08:56 +00:00
|
|
|
if (!*done) {
|
2014-09-23 05:39:04 +00:00
|
|
|
xfs_bmbt_get_all(gotp, &got);
|
2014-09-23 05:37:09 +00:00
|
|
|
*next_fsb = got.br_startoff;
|
2014-09-23 05:39:04 +00:00
|
|
|
}
|
2014-02-23 23:58:19 +00:00
|
|
|
|
|
|
|
del_cursor:
|
|
|
|
if (cur)
|
|
|
|
xfs_btree_del_cursor(cur,
|
|
|
|
error ? XFS_BTREE_ERROR : XFS_BTREE_NOERROR);
|
|
|
|
|
2014-09-02 02:12:53 +00:00
|
|
|
if (logflags)
|
|
|
|
xfs_trans_log_inode(tp, ip, logflags);
|
2014-09-23 05:37:09 +00:00
|
|
|
|
2014-02-23 23:58:19 +00:00
|
|
|
return error;
|
|
|
|
}
|
2015-03-25 04:08:56 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Splits an extent into two extents at split_fsb block such that it is
|
|
|
|
* the first block of the current_ext. @current_ext is a target extent
|
|
|
|
* to be split. @split_fsb is a block where the extents is split.
|
|
|
|
* If split_fsb lies in a hole or the first block of extents, just return 0.
|
|
|
|
*/
|
|
|
|
STATIC int
|
|
|
|
xfs_bmap_split_extent_at(
|
|
|
|
struct xfs_trans *tp,
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
xfs_fileoff_t split_fsb,
|
|
|
|
xfs_fsblock_t *firstfsb,
|
2016-08-03 01:19:29 +00:00
|
|
|
struct xfs_defer_ops *dfops)
|
2015-03-25 04:08:56 +00:00
|
|
|
{
|
|
|
|
int whichfork = XFS_DATA_FORK;
|
|
|
|
struct xfs_btree_cur *cur = NULL;
|
|
|
|
struct xfs_bmbt_rec_host *gotp;
|
|
|
|
struct xfs_bmbt_irec got;
|
|
|
|
struct xfs_bmbt_irec new; /* split extent */
|
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
|
|
|
struct xfs_ifork *ifp;
|
|
|
|
xfs_fsblock_t gotblkcnt; /* new block count for got */
|
|
|
|
xfs_extnum_t current_ext;
|
|
|
|
int error = 0;
|
|
|
|
int logflags = 0;
|
|
|
|
int i = 0;
|
|
|
|
|
|
|
|
if (unlikely(XFS_TEST_ERROR(
|
|
|
|
(XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_EXTENTS &&
|
|
|
|
XFS_IFORK_FORMAT(ip, whichfork) != XFS_DINODE_FMT_BTREE),
|
2017-06-21 00:54:47 +00:00
|
|
|
mp, XFS_ERRTAG_BMAPIFORMAT))) {
|
2015-03-25 04:08:56 +00:00
|
|
|
XFS_ERROR_REPORT("xfs_bmap_split_extent_at",
|
|
|
|
XFS_ERRLEVEL_LOW, mp);
|
|
|
|
return -EFSCORRUPTED;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (XFS_FORCED_SHUTDOWN(mp))
|
|
|
|
return -EIO;
|
|
|
|
|
|
|
|
ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
if (!(ifp->if_flags & XFS_IFEXTENTS)) {
|
|
|
|
/* Read in all the extents */
|
|
|
|
error = xfs_iread_extents(tp, ip, whichfork);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* gotp can be null in 2 cases: 1) if there are no extents
|
|
|
|
* or 2) split_fsb lies in a hole beyond which there are
|
|
|
|
* no extents. Either way, we are done.
|
|
|
|
*/
|
|
|
|
gotp = xfs_iext_bno_to_ext(ifp, split_fsb, ¤t_ext);
|
|
|
|
if (!gotp)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
xfs_bmbt_get_all(gotp, &got);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Check split_fsb lies in a hole or the start boundary offset
|
|
|
|
* of the extent.
|
|
|
|
*/
|
|
|
|
if (got.br_startoff >= split_fsb)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
gotblkcnt = split_fsb - got.br_startoff;
|
|
|
|
new.br_startoff = split_fsb;
|
|
|
|
new.br_startblock = got.br_startblock + gotblkcnt;
|
|
|
|
new.br_blockcount = got.br_blockcount - gotblkcnt;
|
|
|
|
new.br_state = got.br_state;
|
|
|
|
|
|
|
|
if (ifp->if_flags & XFS_IFBROOT) {
|
|
|
|
cur = xfs_bmbt_init_cursor(mp, tp, ip, whichfork);
|
|
|
|
cur->bc_private.b.firstblock = *firstfsb;
|
2016-08-03 01:19:29 +00:00
|
|
|
cur->bc_private.b.dfops = dfops;
|
2015-03-25 04:08:56 +00:00
|
|
|
cur->bc_private.b.flags = 0;
|
|
|
|
error = xfs_bmbt_lookup_eq(cur, got.br_startoff,
|
|
|
|
got.br_startblock,
|
|
|
|
got.br_blockcount,
|
|
|
|
&i);
|
|
|
|
if (error)
|
|
|
|
goto del_cursor;
|
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, del_cursor);
|
|
|
|
}
|
|
|
|
|
|
|
|
xfs_bmbt_set_blockcount(gotp, gotblkcnt);
|
|
|
|
got.br_blockcount = gotblkcnt;
|
|
|
|
|
|
|
|
logflags = XFS_ILOG_CORE;
|
|
|
|
if (cur) {
|
|
|
|
error = xfs_bmbt_update(cur, got.br_startoff,
|
|
|
|
got.br_startblock,
|
|
|
|
got.br_blockcount,
|
|
|
|
got.br_state);
|
|
|
|
if (error)
|
|
|
|
goto del_cursor;
|
|
|
|
} else
|
|
|
|
logflags |= XFS_ILOG_DEXT;
|
|
|
|
|
|
|
|
/* Add new extent */
|
|
|
|
current_ext++;
|
|
|
|
xfs_iext_insert(ip, current_ext, 1, &new, 0);
|
|
|
|
XFS_IFORK_NEXT_SET(ip, whichfork,
|
|
|
|
XFS_IFORK_NEXTENTS(ip, whichfork) + 1);
|
|
|
|
|
|
|
|
if (cur) {
|
|
|
|
error = xfs_bmbt_lookup_eq(cur, new.br_startoff,
|
|
|
|
new.br_startblock, new.br_blockcount,
|
|
|
|
&i);
|
|
|
|
if (error)
|
|
|
|
goto del_cursor;
|
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 0, del_cursor);
|
|
|
|
cur->bc_rec.b.br_state = new.br_state;
|
|
|
|
|
|
|
|
error = xfs_btree_insert(cur, &i);
|
|
|
|
if (error)
|
|
|
|
goto del_cursor;
|
|
|
|
XFS_WANT_CORRUPTED_GOTO(mp, i == 1, del_cursor);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Convert to a btree if necessary.
|
|
|
|
*/
|
|
|
|
if (xfs_bmap_needs_btree(ip, whichfork)) {
|
|
|
|
int tmp_logflags; /* partial log flag return val */
|
|
|
|
|
|
|
|
ASSERT(cur == NULL);
|
2016-08-03 01:19:29 +00:00
|
|
|
error = xfs_bmap_extents_to_btree(tp, ip, firstfsb, dfops,
|
2015-03-25 04:08:56 +00:00
|
|
|
&cur, 0, &tmp_logflags, whichfork);
|
|
|
|
logflags |= tmp_logflags;
|
|
|
|
}
|
|
|
|
|
|
|
|
del_cursor:
|
|
|
|
if (cur) {
|
|
|
|
cur->bc_private.b.allocated = 0;
|
|
|
|
xfs_btree_del_cursor(cur,
|
|
|
|
error ? XFS_BTREE_ERROR : XFS_BTREE_NOERROR);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (logflags)
|
|
|
|
xfs_trans_log_inode(tp, ip, logflags);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
xfs_bmap_split_extent(
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
xfs_fileoff_t split_fsb)
|
|
|
|
{
|
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
|
|
|
struct xfs_trans *tp;
|
2016-08-03 01:19:29 +00:00
|
|
|
struct xfs_defer_ops dfops;
|
2015-03-25 04:08:56 +00:00
|
|
|
xfs_fsblock_t firstfsb;
|
|
|
|
int error;
|
|
|
|
|
2016-04-05 23:19:55 +00:00
|
|
|
error = xfs_trans_alloc(mp, &M_RES(mp)->tr_write,
|
|
|
|
XFS_DIOSTRAT_SPACE_RES(mp, 0), 0, 0, &tp);
|
|
|
|
if (error)
|
2015-03-25 04:08:56 +00:00
|
|
|
return error;
|
|
|
|
|
|
|
|
xfs_ilock(ip, XFS_ILOCK_EXCL);
|
|
|
|
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
|
|
|
|
|
2016-08-03 01:19:29 +00:00
|
|
|
xfs_defer_init(&dfops, &firstfsb);
|
2015-03-25 04:08:56 +00:00
|
|
|
|
|
|
|
error = xfs_bmap_split_extent_at(tp, ip, split_fsb,
|
2016-08-03 01:19:29 +00:00
|
|
|
&firstfsb, &dfops);
|
2015-03-25 04:08:56 +00:00
|
|
|
if (error)
|
|
|
|
goto out;
|
|
|
|
|
2017-08-28 17:21:04 +00:00
|
|
|
error = xfs_defer_finish(&tp, &dfops);
|
2015-03-25 04:08:56 +00:00
|
|
|
if (error)
|
|
|
|
goto out;
|
|
|
|
|
2015-06-04 03:48:08 +00:00
|
|
|
return xfs_trans_commit(tp);
|
2015-03-25 04:08:56 +00:00
|
|
|
|
|
|
|
out:
|
2016-08-03 01:19:29 +00:00
|
|
|
xfs_defer_cancel(&dfops);
|
2015-06-04 03:47:56 +00:00
|
|
|
xfs_trans_cancel(tp);
|
2015-03-25 04:08:56 +00:00
|
|
|
return error;
|
|
|
|
}
|
2016-10-03 16:11:28 +00:00
|
|
|
|
|
|
|
/* Deferred mapping is only for real extents in the data fork. */
|
|
|
|
static bool
|
|
|
|
xfs_bmap_is_update_needed(
|
|
|
|
struct xfs_bmbt_irec *bmap)
|
|
|
|
{
|
|
|
|
return bmap->br_startblock != HOLESTARTBLOCK &&
|
|
|
|
bmap->br_startblock != DELAYSTARTBLOCK;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Record a bmap intent. */
|
|
|
|
static int
|
|
|
|
__xfs_bmap_add(
|
|
|
|
struct xfs_mount *mp,
|
|
|
|
struct xfs_defer_ops *dfops,
|
|
|
|
enum xfs_bmap_intent_type type,
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
int whichfork,
|
|
|
|
struct xfs_bmbt_irec *bmap)
|
|
|
|
{
|
|
|
|
int error;
|
|
|
|
struct xfs_bmap_intent *bi;
|
|
|
|
|
|
|
|
trace_xfs_bmap_defer(mp,
|
|
|
|
XFS_FSB_TO_AGNO(mp, bmap->br_startblock),
|
|
|
|
type,
|
|
|
|
XFS_FSB_TO_AGBNO(mp, bmap->br_startblock),
|
|
|
|
ip->i_ino, whichfork,
|
|
|
|
bmap->br_startoff,
|
|
|
|
bmap->br_blockcount,
|
|
|
|
bmap->br_state);
|
|
|
|
|
|
|
|
bi = kmem_alloc(sizeof(struct xfs_bmap_intent), KM_SLEEP | KM_NOFS);
|
|
|
|
INIT_LIST_HEAD(&bi->bi_list);
|
|
|
|
bi->bi_type = type;
|
|
|
|
bi->bi_owner = ip;
|
|
|
|
bi->bi_whichfork = whichfork;
|
|
|
|
bi->bi_bmap = *bmap;
|
|
|
|
|
2017-08-28 17:21:03 +00:00
|
|
|
error = xfs_defer_ijoin(dfops, bi->bi_owner);
|
2016-10-03 16:11:28 +00:00
|
|
|
if (error) {
|
|
|
|
kmem_free(bi);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
xfs_defer_add(dfops, XFS_DEFER_OPS_TYPE_BMAP, &bi->bi_list);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Map an extent into a file. */
|
|
|
|
int
|
|
|
|
xfs_bmap_map_extent(
|
|
|
|
struct xfs_mount *mp,
|
|
|
|
struct xfs_defer_ops *dfops,
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
struct xfs_bmbt_irec *PREV)
|
|
|
|
{
|
|
|
|
if (!xfs_bmap_is_update_needed(PREV))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
return __xfs_bmap_add(mp, dfops, XFS_BMAP_MAP, ip,
|
|
|
|
XFS_DATA_FORK, PREV);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Unmap an extent out of a file. */
|
|
|
|
int
|
|
|
|
xfs_bmap_unmap_extent(
|
|
|
|
struct xfs_mount *mp,
|
|
|
|
struct xfs_defer_ops *dfops,
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
struct xfs_bmbt_irec *PREV)
|
|
|
|
{
|
|
|
|
if (!xfs_bmap_is_update_needed(PREV))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
return __xfs_bmap_add(mp, dfops, XFS_BMAP_UNMAP, ip,
|
|
|
|
XFS_DATA_FORK, PREV);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Process one of the deferred bmap operations. We pass back the
|
|
|
|
* btree cursor to maintain our lock on the bmapbt between calls.
|
|
|
|
*/
|
|
|
|
int
|
|
|
|
xfs_bmap_finish_one(
|
|
|
|
struct xfs_trans *tp,
|
|
|
|
struct xfs_defer_ops *dfops,
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
enum xfs_bmap_intent_type type,
|
|
|
|
int whichfork,
|
|
|
|
xfs_fileoff_t startoff,
|
|
|
|
xfs_fsblock_t startblock,
|
2017-06-15 04:25:57 +00:00
|
|
|
xfs_filblks_t *blockcount,
|
2016-10-03 16:11:28 +00:00
|
|
|
xfs_exntst_t state)
|
|
|
|
{
|
2017-06-15 04:25:57 +00:00
|
|
|
xfs_fsblock_t firstfsb;
|
|
|
|
int error = 0;
|
2016-10-03 16:11:28 +00:00
|
|
|
|
2017-07-17 21:30:51 +00:00
|
|
|
/*
|
|
|
|
* firstfsb is tied to the transaction lifetime and is used to
|
|
|
|
* ensure correct AG locking order and schedule work item
|
|
|
|
* continuations. XFS_BUI_MAX_FAST_EXTENTS (== 1) restricts us
|
|
|
|
* to only making one bmap call per transaction, so it should
|
|
|
|
* be safe to have it as a local variable here.
|
|
|
|
*/
|
|
|
|
firstfsb = NULLFSBLOCK;
|
|
|
|
|
2016-10-03 16:11:28 +00:00
|
|
|
trace_xfs_bmap_deferred(tp->t_mountp,
|
|
|
|
XFS_FSB_TO_AGNO(tp->t_mountp, startblock), type,
|
|
|
|
XFS_FSB_TO_AGBNO(tp->t_mountp, startblock),
|
2017-06-15 04:25:57 +00:00
|
|
|
ip->i_ino, whichfork, startoff, *blockcount, state);
|
2016-10-03 16:11:28 +00:00
|
|
|
|
2017-04-11 23:45:53 +00:00
|
|
|
if (WARN_ON_ONCE(whichfork != XFS_DATA_FORK))
|
2016-10-03 16:11:28 +00:00
|
|
|
return -EFSCORRUPTED;
|
|
|
|
|
|
|
|
if (XFS_TEST_ERROR(false, tp->t_mountp,
|
2017-06-21 00:54:47 +00:00
|
|
|
XFS_ERRTAG_BMAP_FINISH_ONE))
|
2016-10-03 16:11:28 +00:00
|
|
|
return -EIO;
|
|
|
|
|
|
|
|
switch (type) {
|
|
|
|
case XFS_BMAP_MAP:
|
2017-06-15 04:25:57 +00:00
|
|
|
error = xfs_bmapi_remap(tp, ip, startoff, *blockcount,
|
2017-04-11 23:45:55 +00:00
|
|
|
startblock, dfops);
|
2017-06-15 04:25:57 +00:00
|
|
|
*blockcount = 0;
|
2016-10-03 16:11:28 +00:00
|
|
|
break;
|
|
|
|
case XFS_BMAP_UNMAP:
|
2017-06-15 04:25:57 +00:00
|
|
|
error = __xfs_bunmapi(tp, ip, startoff, blockcount,
|
|
|
|
XFS_BMAPI_REMAP, 1, &firstfsb, dfops);
|
2016-10-03 16:11:28 +00:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
ASSERT(0);
|
|
|
|
error = -EFSCORRUPTED;
|
|
|
|
}
|
|
|
|
|
|
|
|
return error;
|
|
|
|
}
|