License cleanup: add SPDX GPL-2.0 license identifier to files with no license
Many source files in the tree are missing licensing information, which
makes it harder for compliance tools to determine the correct license.
By default all files without license information are under the default
license of the kernel, which is GPL version 2.
Update the files which contain no license information with the 'GPL-2.0'
SPDX license identifier. The SPDX identifier is a legally binding
shorthand, which can be used instead of the full boiler plate text.
This patch is based on work done by Thomas Gleixner and Kate Stewart and
Philippe Ombredanne.
How this work was done:
Patches were generated and checked against linux-4.14-rc6 for a subset of
the use cases:
- file had no licensing information it it.
- file was a */uapi/* one with no licensing information in it,
- file was a */uapi/* one with existing licensing information,
Further patches will be generated in subsequent months to fix up cases
where non-standard license headers were used, and references to license
had to be inferred by heuristics based on keywords.
The analysis to determine which SPDX License Identifier to be applied to
a file was done in a spreadsheet of side by side results from of the
output of two independent scanners (ScanCode & Windriver) producing SPDX
tag:value files created by Philippe Ombredanne. Philippe prepared the
base worksheet, and did an initial spot review of a few 1000 files.
The 4.13 kernel was the starting point of the analysis with 60,537 files
assessed. Kate Stewart did a file by file comparison of the scanner
results in the spreadsheet to determine which SPDX license identifier(s)
to be applied to the file. She confirmed any determination that was not
immediately clear with lawyers working with the Linux Foundation.
Criteria used to select files for SPDX license identifier tagging was:
- Files considered eligible had to be source code files.
- Make and config files were included as candidates if they contained >5
lines of source
- File already had some variant of a license header in it (even if <5
lines).
All documentation files were explicitly excluded.
The following heuristics were used to determine which SPDX license
identifiers to apply.
- when both scanners couldn't find any license traces, file was
considered to have no license information in it, and the top level
COPYING file license applied.
For non */uapi/* files that summary was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 11139
and resulted in the first patch in this series.
If that file was a */uapi/* path one, it was "GPL-2.0 WITH
Linux-syscall-note" otherwise it was "GPL-2.0". Results of that was:
SPDX license identifier # files
---------------------------------------------------|-------
GPL-2.0 WITH Linux-syscall-note 930
and resulted in the second patch in this series.
- if a file had some form of licensing information in it, and was one
of the */uapi/* ones, it was denoted with the Linux-syscall-note if
any GPL family license was found in the file or had no licensing in
it (per prior point). Results summary:
SPDX license identifier # files
---------------------------------------------------|------
GPL-2.0 WITH Linux-syscall-note 270
GPL-2.0+ WITH Linux-syscall-note 169
((GPL-2.0 WITH Linux-syscall-note) OR BSD-2-Clause) 21
((GPL-2.0 WITH Linux-syscall-note) OR BSD-3-Clause) 17
LGPL-2.1+ WITH Linux-syscall-note 15
GPL-1.0+ WITH Linux-syscall-note 14
((GPL-2.0+ WITH Linux-syscall-note) OR BSD-3-Clause) 5
LGPL-2.0+ WITH Linux-syscall-note 4
LGPL-2.1 WITH Linux-syscall-note 3
((GPL-2.0 WITH Linux-syscall-note) OR MIT) 3
((GPL-2.0 WITH Linux-syscall-note) AND MIT) 1
and that resulted in the third patch in this series.
- when the two scanners agreed on the detected license(s), that became
the concluded license(s).
- when there was disagreement between the two scanners (one detected a
license but the other didn't, or they both detected different
licenses) a manual inspection of the file occurred.
- In most cases a manual inspection of the information in the file
resulted in a clear resolution of the license that should apply (and
which scanner probably needed to revisit its heuristics).
- When it was not immediately clear, the license identifier was
confirmed with lawyers working with the Linux Foundation.
- If there was any question as to the appropriate license identifier,
the file was flagged for further research and to be revisited later
in time.
In total, over 70 hours of logged manual review was done on the
spreadsheet to determine the SPDX license identifiers to apply to the
source files by Kate, Philippe, Thomas and, in some cases, confirmation
by lawyers working with the Linux Foundation.
Kate also obtained a third independent scan of the 4.13 code base from
FOSSology, and compared selected files where the other two scanners
disagreed against that SPDX file, to see if there was new insights. The
Windriver scanner is based on an older version of FOSSology in part, so
they are related.
Thomas did random spot checks in about 500 files from the spreadsheets
for the uapi headers and agreed with SPDX license identifier in the
files he inspected. For the non-uapi files Thomas did random spot checks
in about 15000 files.
In initial set of patches against 4.14-rc6, 3 files were found to have
copy/paste license identifier errors, and have been fixed to reflect the
correct identifier.
Additionally Philippe spent 10 hours this week doing a detailed manual
inspection and review of the 12,461 patched files from the initial patch
version early this week with:
- a full scancode scan run, collecting the matched texts, detected
license ids and scores
- reviewing anything where there was a license detected (about 500+
files) to ensure that the applied SPDX license was correct
- reviewing anything where there was no detection but the patch license
was not GPL-2.0 WITH Linux-syscall-note to ensure that the applied
SPDX license was correct
This produced a worksheet with 20 files needing minor correction. This
worksheet was then exported into 3 different .csv files for the
different types of files to be modified.
These .csv files were then reviewed by Greg. Thomas wrote a script to
parse the csv files and add the proper SPDX tag to the file, in the
format that the file expected. This script was further refined by Greg
based on the output to detect more types of files automatically and to
distinguish between header and source .c files (which need different
comment types.) Finally Greg ran the script using the .csv files to
generate the patches.
Reviewed-by: Kate Stewart <kstewart@linuxfoundation.org>
Reviewed-by: Philippe Ombredanne <pombredanne@nexb.com>
Reviewed-by: Thomas Gleixner <tglx@linutronix.de>
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
2017-11-01 14:07:57 +00:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* linux/fs/ext2/inode.c
|
|
|
|
*
|
|
|
|
* Copyright (C) 1992, 1993, 1994, 1995
|
|
|
|
* Remy Card (card@masi.ibp.fr)
|
|
|
|
* Laboratoire MASI - Institut Blaise Pascal
|
|
|
|
* Universite Pierre et Marie Curie (Paris VI)
|
|
|
|
*
|
|
|
|
* from
|
|
|
|
*
|
|
|
|
* linux/fs/minix/inode.c
|
|
|
|
*
|
|
|
|
* Copyright (C) 1991, 1992 Linus Torvalds
|
|
|
|
*
|
|
|
|
* Goal-directed block allocation by Stephen Tweedie
|
|
|
|
* (sct@dcs.ed.ac.uk), 1993, 1998
|
|
|
|
* Big-endian to little-endian byte-swapping/bitmaps by
|
|
|
|
* David S. Miller (davem@caip.rutgers.edu), 1995
|
|
|
|
* 64-bit file support on 64-bit platforms by Jakub Jelinek
|
|
|
|
* (jj@sunsite.ms.mff.cuni.cz)
|
|
|
|
*
|
|
|
|
* Assorted race fixes, rewrite of ext2_get_block() by Al Viro, 2000
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/time.h>
|
|
|
|
#include <linux/highuid.h>
|
|
|
|
#include <linux/pagemap.h>
|
2015-09-08 21:58:40 +00:00
|
|
|
#include <linux/dax.h>
|
2016-03-15 17:20:41 +00:00
|
|
|
#include <linux/blkdev.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include <linux/quotaops.h>
|
|
|
|
#include <linux/writeback.h>
|
|
|
|
#include <linux/buffer_head.h>
|
|
|
|
#include <linux/mpage.h>
|
2008-10-03 21:32:43 +00:00
|
|
|
#include <linux/fiemap.h>
|
2016-09-19 01:30:29 +00:00
|
|
|
#include <linux/iomap.h>
|
2008-12-19 20:47:13 +00:00
|
|
|
#include <linux/namei.h>
|
2015-02-22 16:58:50 +00:00
|
|
|
#include <linux/uio.h>
|
2005-04-16 22:20:36 +00:00
|
|
|
#include "ext2.h"
|
|
|
|
#include "acl.h"
|
2013-03-13 11:57:08 +00:00
|
|
|
#include "xattr.h"
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2010-03-05 08:21:37 +00:00
|
|
|
static int __ext2_write_inode(struct inode *inode, int do_sync);
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* Test whether an inode is a fast symlink.
|
|
|
|
*/
|
|
|
|
static inline int ext2_inode_is_fast_symlink(struct inode *inode)
|
|
|
|
{
|
|
|
|
int ea_blocks = EXT2_I(inode)->i_file_acl ?
|
|
|
|
(inode->i_sb->s_blocksize >> 9) : 0;
|
|
|
|
|
|
|
|
return (S_ISLNK(inode->i_mode) &&
|
|
|
|
inode->i_blocks - ea_blocks == 0);
|
|
|
|
}
|
|
|
|
|
2010-05-26 15:05:37 +00:00
|
|
|
static void ext2_truncate_blocks(struct inode *inode, loff_t offset);
|
|
|
|
|
|
|
|
static void ext2_write_failed(struct address_space *mapping, loff_t to)
|
|
|
|
{
|
|
|
|
struct inode *inode = mapping->host;
|
|
|
|
|
|
|
|
if (to > inode->i_size) {
|
2013-09-12 22:13:56 +00:00
|
|
|
truncate_pagecache(inode, inode->i_size);
|
2010-05-26 15:05:37 +00:00
|
|
|
ext2_truncate_blocks(inode, inode->i_size);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* Called at the last iput() if i_nlink is zero.
|
|
|
|
*/
|
2010-06-05 03:32:28 +00:00
|
|
|
void ext2_evict_inode(struct inode * inode)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2010-06-05 03:32:28 +00:00
|
|
|
struct ext2_block_alloc_info *rsv;
|
|
|
|
int want_delete = 0;
|
|
|
|
|
|
|
|
if (!inode->i_nlink && !is_bad_inode(inode)) {
|
|
|
|
want_delete = 1;
|
2010-03-03 14:05:07 +00:00
|
|
|
dquot_initialize(inode);
|
2010-06-05 03:32:28 +00:00
|
|
|
} else {
|
|
|
|
dquot_drop(inode);
|
|
|
|
}
|
|
|
|
|
2014-04-03 21:47:49 +00:00
|
|
|
truncate_inode_pages_final(&inode->i_data);
|
2005-09-09 20:01:31 +00:00
|
|
|
|
2010-06-05 03:32:28 +00:00
|
|
|
if (want_delete) {
|
2012-06-12 14:20:46 +00:00
|
|
|
sb_start_intwrite(inode->i_sb);
|
2010-06-05 03:32:28 +00:00
|
|
|
/* set dtime */
|
2018-06-20 14:31:40 +00:00
|
|
|
EXT2_I(inode)->i_dtime = ktime_get_real_seconds();
|
2010-06-05 03:32:28 +00:00
|
|
|
mark_inode_dirty(inode);
|
|
|
|
__ext2_write_inode(inode, inode_needs_sync(inode));
|
|
|
|
/* truncate to 0 */
|
|
|
|
inode->i_size = 0;
|
|
|
|
if (inode->i_blocks)
|
|
|
|
ext2_truncate_blocks(inode, 0);
|
2013-03-13 11:57:08 +00:00
|
|
|
ext2_xattr_delete_inode(inode);
|
2010-06-05 03:32:28 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
invalidate_inode_buffers(inode);
|
2012-05-03 12:48:02 +00:00
|
|
|
clear_inode(inode);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2010-06-05 03:32:28 +00:00
|
|
|
ext2_discard_reservation(inode);
|
|
|
|
rsv = EXT2_I(inode)->i_block_alloc_info;
|
|
|
|
EXT2_I(inode)->i_block_alloc_info = NULL;
|
|
|
|
if (unlikely(rsv))
|
|
|
|
kfree(rsv);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2012-06-12 14:20:46 +00:00
|
|
|
if (want_delete) {
|
2010-06-05 03:32:28 +00:00
|
|
|
ext2_free_inode(inode);
|
2012-06-12 14:20:46 +00:00
|
|
|
sb_end_intwrite(inode->i_sb);
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
typedef struct {
|
|
|
|
__le32 *p;
|
|
|
|
__le32 key;
|
|
|
|
struct buffer_head *bh;
|
|
|
|
} Indirect;
|
|
|
|
|
|
|
|
static inline void add_chain(Indirect *p, struct buffer_head *bh, __le32 *v)
|
|
|
|
{
|
|
|
|
p->key = *(p->p = v);
|
|
|
|
p->bh = bh;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int verify_chain(Indirect *from, Indirect *to)
|
|
|
|
{
|
|
|
|
while (from <= to && from->key == *from->p)
|
|
|
|
from++;
|
|
|
|
return (from > to);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ext2_block_to_path - parse the block number into array of offsets
|
|
|
|
* @inode: inode in question (we are only interested in its superblock)
|
|
|
|
* @i_block: block number to be parsed
|
|
|
|
* @offsets: array to store the offsets in
|
|
|
|
* @boundary: set this non-zero if the referred-to block is likely to be
|
|
|
|
* followed (on disk) by an indirect block.
|
|
|
|
* To store the locations of file's data ext2 uses a data structure common
|
|
|
|
* for UNIX filesystems - tree of pointers anchored in the inode, with
|
|
|
|
* data blocks at leaves and indirect blocks in intermediate nodes.
|
|
|
|
* This function translates the block number into path in that tree -
|
|
|
|
* return value is the path length and @offsets[n] is the offset of
|
|
|
|
* pointer to (n+1)th node in the nth one. If @block is out of range
|
|
|
|
* (negative or too large) warning is printed and zero returned.
|
|
|
|
*
|
|
|
|
* Note: function doesn't find node addresses, so no IO is needed. All
|
|
|
|
* we need to know is the capacity of indirect blocks (taken from the
|
|
|
|
* inode->i_sb).
|
|
|
|
*/
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Portability note: the last comparison (check that we fit into triple
|
|
|
|
* indirect block) is spelled differently, because otherwise on an
|
|
|
|
* architecture with 32-bit longs and 8Kb pages we might get into trouble
|
|
|
|
* if our filesystem had 8Kb blocks. We might use long long, but that would
|
|
|
|
* kill us on x86. Oh, well, at least the sign propagation does not matter -
|
|
|
|
* i_block would have to be negative in the very beginning, so we would not
|
|
|
|
* get there at all.
|
|
|
|
*/
|
|
|
|
|
|
|
|
static int ext2_block_to_path(struct inode *inode,
|
|
|
|
long i_block, int offsets[4], int *boundary)
|
|
|
|
{
|
|
|
|
int ptrs = EXT2_ADDR_PER_BLOCK(inode->i_sb);
|
|
|
|
int ptrs_bits = EXT2_ADDR_PER_BLOCK_BITS(inode->i_sb);
|
|
|
|
const long direct_blocks = EXT2_NDIR_BLOCKS,
|
|
|
|
indirect_blocks = ptrs,
|
|
|
|
double_blocks = (1 << (ptrs_bits * 2));
|
|
|
|
int n = 0;
|
|
|
|
int final = 0;
|
|
|
|
|
|
|
|
if (i_block < 0) {
|
ext2: Unify log messages in ext2
make messages produced by ext2 more unified. It should be
easy to parse.
dmesg before patch:
[ 4893.684892] reservations ON
[ 4893.684896] xip option not supported
[ 4893.684961] EXT2-fs warning: mounting ext3 filesystem as ext2
[ 4893.684964] EXT2-fs warning: maximal mount count reached, running
e2fsck is recommended
[ 4893.684990] EXT II FS: 0.5b, 95/08/09, bs=1024, fs=1024, gc=2,
bpg=8192, ipg=1280, mo=80010]
dmesg after patch:
[ 4893.684892] EXT2-fs (loop0): reservations ON
[ 4893.684896] EXT2-fs (loop0): xip option not supported
[ 4893.684961] EXT2-fs (loop0): warning: mounting ext3 filesystem as
ext2
[ 4893.684964] EXT2-fs (loop0): warning: maximal mount count reached,
running e2fsck is recommended
[ 4893.684990] EXT2-fs (loop0): 0.5b, 95/08/09, bs=1024, fs=1024, gc=2,
bpg=8192, ipg=1280, mo=80010]
Signed-off-by: Alexey Fisher <bug-track@fisher-privat.net>
Reviewed-by: Andreas Dilger <adilger@sun.com>
Signed-off-by: Jan Kara <jack@suse.cz>
2009-11-19 18:12:51 +00:00
|
|
|
ext2_msg(inode->i_sb, KERN_WARNING,
|
|
|
|
"warning: %s: block < 0", __func__);
|
2005-04-16 22:20:36 +00:00
|
|
|
} else if (i_block < direct_blocks) {
|
|
|
|
offsets[n++] = i_block;
|
|
|
|
final = direct_blocks;
|
|
|
|
} else if ( (i_block -= direct_blocks) < indirect_blocks) {
|
|
|
|
offsets[n++] = EXT2_IND_BLOCK;
|
|
|
|
offsets[n++] = i_block;
|
|
|
|
final = ptrs;
|
|
|
|
} else if ((i_block -= indirect_blocks) < double_blocks) {
|
|
|
|
offsets[n++] = EXT2_DIND_BLOCK;
|
|
|
|
offsets[n++] = i_block >> ptrs_bits;
|
|
|
|
offsets[n++] = i_block & (ptrs - 1);
|
|
|
|
final = ptrs;
|
|
|
|
} else if (((i_block -= double_blocks) >> (ptrs_bits * 2)) < ptrs) {
|
|
|
|
offsets[n++] = EXT2_TIND_BLOCK;
|
|
|
|
offsets[n++] = i_block >> (ptrs_bits * 2);
|
|
|
|
offsets[n++] = (i_block >> ptrs_bits) & (ptrs - 1);
|
|
|
|
offsets[n++] = i_block & (ptrs - 1);
|
|
|
|
final = ptrs;
|
|
|
|
} else {
|
ext2: Unify log messages in ext2
make messages produced by ext2 more unified. It should be
easy to parse.
dmesg before patch:
[ 4893.684892] reservations ON
[ 4893.684896] xip option not supported
[ 4893.684961] EXT2-fs warning: mounting ext3 filesystem as ext2
[ 4893.684964] EXT2-fs warning: maximal mount count reached, running
e2fsck is recommended
[ 4893.684990] EXT II FS: 0.5b, 95/08/09, bs=1024, fs=1024, gc=2,
bpg=8192, ipg=1280, mo=80010]
dmesg after patch:
[ 4893.684892] EXT2-fs (loop0): reservations ON
[ 4893.684896] EXT2-fs (loop0): xip option not supported
[ 4893.684961] EXT2-fs (loop0): warning: mounting ext3 filesystem as
ext2
[ 4893.684964] EXT2-fs (loop0): warning: maximal mount count reached,
running e2fsck is recommended
[ 4893.684990] EXT2-fs (loop0): 0.5b, 95/08/09, bs=1024, fs=1024, gc=2,
bpg=8192, ipg=1280, mo=80010]
Signed-off-by: Alexey Fisher <bug-track@fisher-privat.net>
Reviewed-by: Andreas Dilger <adilger@sun.com>
Signed-off-by: Jan Kara <jack@suse.cz>
2009-11-19 18:12:51 +00:00
|
|
|
ext2_msg(inode->i_sb, KERN_WARNING,
|
|
|
|
"warning: %s: block is too big", __func__);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
if (boundary)
|
2007-10-17 06:30:46 +00:00
|
|
|
*boundary = final - 1 - (i_block & (ptrs - 1));
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
return n;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ext2_get_branch - read the chain of indirect blocks leading to data
|
|
|
|
* @inode: inode in question
|
|
|
|
* @depth: depth of the chain (1 - direct pointer, etc.)
|
|
|
|
* @offsets: offsets of pointers in inode/indirect blocks
|
|
|
|
* @chain: place to store the result
|
|
|
|
* @err: here we store the error value
|
|
|
|
*
|
|
|
|
* Function fills the array of triples <key, p, bh> and returns %NULL
|
|
|
|
* if everything went OK or the pointer to the last filled triple
|
|
|
|
* (incomplete one) otherwise. Upon the return chain[i].key contains
|
|
|
|
* the number of (i+1)-th block in the chain (as it is stored in memory,
|
|
|
|
* i.e. little-endian 32-bit), chain[i].p contains the address of that
|
|
|
|
* number (it points into struct inode for i==0 and into the bh->b_data
|
|
|
|
* for i>0) and chain[i].bh points to the buffer_head of i-th indirect
|
|
|
|
* block for i>0 and NULL for i==0. In other words, it holds the block
|
|
|
|
* numbers of the chain, addresses they were taken from (and where we can
|
|
|
|
* verify that chain did not change) and buffer_heads hosting these
|
|
|
|
* numbers.
|
|
|
|
*
|
|
|
|
* Function stops when it stumbles upon zero pointer (absent block)
|
|
|
|
* (pointer to last triple returned, *@err == 0)
|
|
|
|
* or when it gets an IO error reading an indirect block
|
|
|
|
* (ditto, *@err == -EIO)
|
|
|
|
* or when it notices that chain had been changed while it was reading
|
|
|
|
* (ditto, *@err == -EAGAIN)
|
|
|
|
* or when it reads all @depth-1 indirect blocks successfully and finds
|
|
|
|
* the whole chain, all way to the data (returns %NULL, *err == 0).
|
|
|
|
*/
|
|
|
|
static Indirect *ext2_get_branch(struct inode *inode,
|
|
|
|
int depth,
|
|
|
|
int *offsets,
|
|
|
|
Indirect chain[4],
|
|
|
|
int *err)
|
|
|
|
{
|
|
|
|
struct super_block *sb = inode->i_sb;
|
|
|
|
Indirect *p = chain;
|
|
|
|
struct buffer_head *bh;
|
|
|
|
|
|
|
|
*err = 0;
|
|
|
|
/* i_data is not going away, no lock needed */
|
|
|
|
add_chain (chain, NULL, EXT2_I(inode)->i_data + *offsets);
|
|
|
|
if (!p->key)
|
|
|
|
goto no_block;
|
|
|
|
while (--depth) {
|
|
|
|
bh = sb_bread(sb, le32_to_cpu(p->key));
|
|
|
|
if (!bh)
|
|
|
|
goto failure;
|
|
|
|
read_lock(&EXT2_I(inode)->i_meta_lock);
|
|
|
|
if (!verify_chain(chain, p))
|
|
|
|
goto changed;
|
|
|
|
add_chain(++p, bh, (__le32*)bh->b_data + *++offsets);
|
|
|
|
read_unlock(&EXT2_I(inode)->i_meta_lock);
|
|
|
|
if (!p->key)
|
|
|
|
goto no_block;
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
changed:
|
|
|
|
read_unlock(&EXT2_I(inode)->i_meta_lock);
|
|
|
|
brelse(bh);
|
|
|
|
*err = -EAGAIN;
|
|
|
|
goto no_block;
|
|
|
|
failure:
|
|
|
|
*err = -EIO;
|
|
|
|
no_block:
|
|
|
|
return p;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ext2_find_near - find a place for allocation with sufficient locality
|
|
|
|
* @inode: owner
|
|
|
|
* @ind: descriptor of indirect block.
|
|
|
|
*
|
2008-04-21 22:45:55 +00:00
|
|
|
* This function returns the preferred place for block allocation.
|
2005-04-16 22:20:36 +00:00
|
|
|
* It is used when heuristic for sequential allocation fails.
|
|
|
|
* Rules are:
|
|
|
|
* + if there is a block to the left of our position - allocate near it.
|
|
|
|
* + if pointer will live in indirect block - allocate near that block.
|
|
|
|
* + if pointer will live in inode - allocate in the same cylinder group.
|
|
|
|
*
|
|
|
|
* In the latter case we colour the starting block by the callers PID to
|
|
|
|
* prevent it from clashing with concurrent allocations for a different inode
|
|
|
|
* in the same block group. The PID is used here so that functionally related
|
|
|
|
* files will be close-by on-disk.
|
|
|
|
*
|
|
|
|
* Caller must make sure that @ind is valid and will stay that way.
|
|
|
|
*/
|
|
|
|
|
2008-04-28 09:16:02 +00:00
|
|
|
static ext2_fsblk_t ext2_find_near(struct inode *inode, Indirect *ind)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct ext2_inode_info *ei = EXT2_I(inode);
|
|
|
|
__le32 *start = ind->bh ? (__le32 *) ind->bh->b_data : ei->i_data;
|
|
|
|
__le32 *p;
|
2008-04-28 09:16:02 +00:00
|
|
|
ext2_fsblk_t bg_start;
|
|
|
|
ext2_fsblk_t colour;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* Try to find previous block */
|
|
|
|
for (p = ind->p - 1; p >= start; p--)
|
|
|
|
if (*p)
|
|
|
|
return le32_to_cpu(*p);
|
|
|
|
|
|
|
|
/* No such thing, so let's try location of indirect block */
|
|
|
|
if (ind->bh)
|
|
|
|
return ind->bh->b_blocknr;
|
|
|
|
|
|
|
|
/*
|
2011-03-31 01:57:33 +00:00
|
|
|
* It is going to be referred from inode itself? OK, just put it into
|
2005-04-16 22:20:36 +00:00
|
|
|
* the same cylinder group then.
|
|
|
|
*/
|
2008-04-28 09:16:01 +00:00
|
|
|
bg_start = ext2_group_first_block_no(inode->i_sb, ei->i_block_group);
|
2005-04-16 22:20:36 +00:00
|
|
|
colour = (current->pid % 16) *
|
|
|
|
(EXT2_BLOCKS_PER_GROUP(inode->i_sb) / 16);
|
|
|
|
return bg_start + colour;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2008-04-21 22:45:55 +00:00
|
|
|
* ext2_find_goal - find a preferred place for allocation.
|
2005-04-16 22:20:36 +00:00
|
|
|
* @inode: owner
|
|
|
|
* @block: block we want
|
|
|
|
* @partial: pointer to the last triple within a chain
|
|
|
|
*
|
2007-10-17 06:30:46 +00:00
|
|
|
* Returns preferred place for a block (the goal).
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
|
|
|
|
2008-04-28 09:16:02 +00:00
|
|
|
static inline ext2_fsblk_t ext2_find_goal(struct inode *inode, long block,
|
|
|
|
Indirect *partial)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2007-10-17 06:30:46 +00:00
|
|
|
struct ext2_block_alloc_info *block_i;
|
|
|
|
|
|
|
|
block_i = EXT2_I(inode)->i_block_alloc_info;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* try the heuristic for sequential allocation,
|
|
|
|
* failing that at least try to get decent locality.
|
|
|
|
*/
|
|
|
|
if (block_i && (block == block_i->last_alloc_logical_block + 1)
|
|
|
|
&& (block_i->last_alloc_physical_block != 0)) {
|
|
|
|
return block_i->last_alloc_physical_block + 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
return ext2_find_near(inode, partial);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ext2_blks_to_allocate: Look up the block map and count the number
|
|
|
|
* of direct blocks need to be allocated for the given branch.
|
|
|
|
*
|
|
|
|
* @branch: chain of indirect blocks
|
|
|
|
* @k: number of blocks need for indirect blocks
|
|
|
|
* @blks: number of data blocks to be mapped.
|
|
|
|
* @blocks_to_boundary: the offset in the indirect block
|
|
|
|
*
|
2020-07-03 12:44:11 +00:00
|
|
|
* return the number of direct blocks to allocate.
|
2007-10-17 06:30:46 +00:00
|
|
|
*/
|
|
|
|
static int
|
|
|
|
ext2_blks_to_allocate(Indirect * branch, int k, unsigned long blks,
|
|
|
|
int blocks_to_boundary)
|
|
|
|
{
|
|
|
|
unsigned long count = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Simple case, [t,d]Indirect block(s) has not allocated yet
|
|
|
|
* then it's clear blocks on that path have not allocated
|
|
|
|
*/
|
|
|
|
if (k > 0) {
|
|
|
|
/* right now don't hanel cross boundary allocation */
|
|
|
|
if (blks < blocks_to_boundary + 1)
|
|
|
|
count += blks;
|
|
|
|
else
|
|
|
|
count += blocks_to_boundary + 1;
|
|
|
|
return count;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2007-10-17 06:30:46 +00:00
|
|
|
|
|
|
|
count++;
|
|
|
|
while (count < blks && count <= blocks_to_boundary
|
|
|
|
&& le32_to_cpu(*(branch[0].p + count)) == 0) {
|
|
|
|
count++;
|
|
|
|
}
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ext2_alloc_blocks: multiple allocate blocks needed for a branch
|
|
|
|
* @indirect_blks: the number of blocks need to allocate for indirect
|
|
|
|
* blocks
|
2020-07-03 12:44:11 +00:00
|
|
|
* @blks: the number of blocks need to allocate for direct blocks
|
2007-10-17 06:30:46 +00:00
|
|
|
* @new_blocks: on return it will store the new block numbers for
|
|
|
|
* the indirect blocks(if needed) and the first direct block,
|
|
|
|
*/
|
|
|
|
static int ext2_alloc_blocks(struct inode *inode,
|
|
|
|
ext2_fsblk_t goal, int indirect_blks, int blks,
|
|
|
|
ext2_fsblk_t new_blocks[4], int *err)
|
|
|
|
{
|
|
|
|
int target, i;
|
|
|
|
unsigned long count = 0;
|
|
|
|
int index = 0;
|
|
|
|
ext2_fsblk_t current_block = 0;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Here we try to allocate the requested multiple blocks at once,
|
|
|
|
* on a best-effort basis.
|
|
|
|
* To build a branch, we should allocate blocks for
|
|
|
|
* the indirect blocks(if not allocated yet), and at least
|
|
|
|
* the first direct block of this branch. That's the
|
|
|
|
* minimum number of blocks need to allocate(required)
|
|
|
|
*/
|
|
|
|
target = blks + indirect_blks;
|
|
|
|
|
|
|
|
while (1) {
|
|
|
|
count = target;
|
|
|
|
/* allocating blocks for indirect blocks and direct blocks */
|
|
|
|
current_block = ext2_new_blocks(inode,goal,&count,err);
|
|
|
|
if (*err)
|
|
|
|
goto failed_out;
|
|
|
|
|
|
|
|
target -= count;
|
|
|
|
/* allocate blocks for indirect blocks */
|
|
|
|
while (index < indirect_blks && count) {
|
|
|
|
new_blocks[index++] = current_block++;
|
|
|
|
count--;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (count > 0)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* save the new block number for the first direct block */
|
|
|
|
new_blocks[index] = current_block;
|
|
|
|
|
|
|
|
/* total number of blocks allocated for direct blocks */
|
|
|
|
ret = count;
|
|
|
|
*err = 0;
|
|
|
|
return ret;
|
|
|
|
failed_out:
|
|
|
|
for (i = 0; i <index; i++)
|
|
|
|
ext2_free_blocks(inode, new_blocks[i], 1);
|
2010-07-21 21:19:42 +00:00
|
|
|
if (index)
|
|
|
|
mark_inode_dirty(inode);
|
2007-10-17 06:30:46 +00:00
|
|
|
return ret;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ext2_alloc_branch - allocate and set up a chain of blocks.
|
|
|
|
* @inode: owner
|
2019-03-06 01:47:35 +00:00
|
|
|
* @indirect_blks: depth of the chain (number of blocks to allocate)
|
|
|
|
* @blks: number of allocated direct blocks
|
|
|
|
* @goal: preferred place for allocation
|
2005-04-16 22:20:36 +00:00
|
|
|
* @offsets: offsets (in the blocks) to store the pointers to next.
|
|
|
|
* @branch: place to store the chain in.
|
|
|
|
*
|
|
|
|
* This function allocates @num blocks, zeroes out all but the last one,
|
|
|
|
* links them into chain and (if we are synchronous) writes them to disk.
|
|
|
|
* In other words, it prepares a branch that can be spliced onto the
|
|
|
|
* inode. It stores the information about that chain in the branch[], in
|
|
|
|
* the same format as ext2_get_branch() would do. We are calling it after
|
|
|
|
* we had read the existing part of chain and partial points to the last
|
|
|
|
* triple of that (one with zero ->key). Upon the exit we have the same
|
2010-09-15 12:46:02 +00:00
|
|
|
* picture as after the successful ext2_get_block(), except that in one
|
2005-04-16 22:20:36 +00:00
|
|
|
* place chain is disconnected - *branch->p is still zero (we did not
|
|
|
|
* set the last link), but branch->key contains the number that should
|
|
|
|
* be placed into *branch->p to fill that gap.
|
|
|
|
*
|
|
|
|
* If allocation fails we free all blocks we've allocated (and forget
|
|
|
|
* their buffer_heads) and return the error value the from failed
|
|
|
|
* ext2_alloc_block() (normally -ENOSPC). Otherwise we set the chain
|
|
|
|
* as described above and return 0.
|
|
|
|
*/
|
|
|
|
|
|
|
|
static int ext2_alloc_branch(struct inode *inode,
|
2007-10-17 06:30:46 +00:00
|
|
|
int indirect_blks, int *blks, ext2_fsblk_t goal,
|
|
|
|
int *offsets, Indirect *branch)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
int blocksize = inode->i_sb->s_blocksize;
|
2007-10-17 06:30:46 +00:00
|
|
|
int i, n = 0;
|
|
|
|
int err = 0;
|
|
|
|
struct buffer_head *bh;
|
|
|
|
int num;
|
|
|
|
ext2_fsblk_t new_blocks[4];
|
|
|
|
ext2_fsblk_t current_block;
|
|
|
|
|
|
|
|
num = ext2_alloc_blocks(inode, goal, indirect_blks,
|
|
|
|
*blks, new_blocks, &err);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
|
|
|
|
branch[0].key = cpu_to_le32(new_blocks[0]);
|
|
|
|
/*
|
|
|
|
* metadata blocks and data blocks are allocated.
|
|
|
|
*/
|
|
|
|
for (n = 1; n <= indirect_blks; n++) {
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
2007-10-17 06:30:46 +00:00
|
|
|
* Get buffer_head for parent block, zero it out
|
|
|
|
* and set the pointer to new one, then send
|
|
|
|
* parent to disk.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-10-17 06:30:46 +00:00
|
|
|
bh = sb_getblk(inode->i_sb, new_blocks[n-1]);
|
2013-01-12 09:19:32 +00:00
|
|
|
if (unlikely(!bh)) {
|
|
|
|
err = -ENOMEM;
|
|
|
|
goto failed;
|
|
|
|
}
|
2007-10-17 06:30:46 +00:00
|
|
|
branch[n].bh = bh;
|
2005-04-16 22:20:36 +00:00
|
|
|
lock_buffer(bh);
|
|
|
|
memset(bh->b_data, 0, blocksize);
|
|
|
|
branch[n].p = (__le32 *) bh->b_data + offsets[n];
|
2007-10-17 06:30:46 +00:00
|
|
|
branch[n].key = cpu_to_le32(new_blocks[n]);
|
2005-04-16 22:20:36 +00:00
|
|
|
*branch[n].p = branch[n].key;
|
2007-10-17 06:30:46 +00:00
|
|
|
if ( n == indirect_blks) {
|
|
|
|
current_block = new_blocks[n];
|
|
|
|
/*
|
|
|
|
* End of chain, update the last new metablock of
|
|
|
|
* the chain to point to the new allocated
|
|
|
|
* data blocks numbers
|
|
|
|
*/
|
|
|
|
for (i=1; i < num; i++)
|
|
|
|
*(branch[n].p + i) = cpu_to_le32(++current_block);
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
set_buffer_uptodate(bh);
|
|
|
|
unlock_buffer(bh);
|
|
|
|
mark_buffer_dirty_inode(bh, inode);
|
|
|
|
/* We used to sync bh here if IS_SYNC(inode).
|
2009-08-18 15:54:11 +00:00
|
|
|
* But we now rely upon generic_write_sync()
|
2005-04-16 22:20:36 +00:00
|
|
|
* and b_inode_buffers. But not for directories.
|
|
|
|
*/
|
|
|
|
if (S_ISDIR(inode->i_mode) && IS_DIRSYNC(inode))
|
|
|
|
sync_dirty_buffer(bh);
|
|
|
|
}
|
2007-10-17 06:30:46 +00:00
|
|
|
*blks = num;
|
2005-04-16 22:20:36 +00:00
|
|
|
return err;
|
2013-01-12 09:19:32 +00:00
|
|
|
|
|
|
|
failed:
|
|
|
|
for (i = 1; i < n; i++)
|
|
|
|
bforget(branch[i].bh);
|
|
|
|
for (i = 0; i < indirect_blks; i++)
|
|
|
|
ext2_free_blocks(inode, new_blocks[i], 1);
|
|
|
|
ext2_free_blocks(inode, new_blocks[i], num);
|
|
|
|
return err;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
2007-10-17 06:30:46 +00:00
|
|
|
* ext2_splice_branch - splice the allocated branch onto inode.
|
|
|
|
* @inode: owner
|
|
|
|
* @block: (logical) number of block we are adding
|
|
|
|
* @where: location of missing link
|
|
|
|
* @num: number of indirect blocks we are adding
|
|
|
|
* @blks: number of direct blocks we are adding
|
2005-04-16 22:20:36 +00:00
|
|
|
*
|
2007-10-17 06:30:46 +00:00
|
|
|
* This function fills the missing link and does all housekeeping needed in
|
|
|
|
* inode (->i_blocks, etc.). In case of success we end up with the full
|
|
|
|
* chain to new block and return 0.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-10-17 06:30:46 +00:00
|
|
|
static void ext2_splice_branch(struct inode *inode,
|
|
|
|
long block, Indirect *where, int num, int blks)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
int i;
|
2007-10-17 06:30:46 +00:00
|
|
|
struct ext2_block_alloc_info *block_i;
|
|
|
|
ext2_fsblk_t current_block;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-17 06:30:46 +00:00
|
|
|
block_i = EXT2_I(inode)->i_block_alloc_info;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-17 06:30:46 +00:00
|
|
|
/* XXX LOCKING probably should have i_meta_lock ?*/
|
2005-04-16 22:20:36 +00:00
|
|
|
/* That's it */
|
|
|
|
|
|
|
|
*where->p = where->key;
|
|
|
|
|
2007-10-17 06:30:46 +00:00
|
|
|
/*
|
|
|
|
* Update the host buffer_head or inode to point to more just allocated
|
|
|
|
* direct blocks blocks
|
|
|
|
*/
|
|
|
|
if (num == 0 && blks > 1) {
|
|
|
|
current_block = le32_to_cpu(where->key) + 1;
|
|
|
|
for (i = 1; i < blks; i++)
|
|
|
|
*(where->p + i ) = cpu_to_le32(current_block++);
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-17 06:30:46 +00:00
|
|
|
/*
|
|
|
|
* update the most recently allocated logical & physical block
|
|
|
|
* in i_block_alloc_info, to assist find the proper goal block for next
|
|
|
|
* allocation
|
|
|
|
*/
|
|
|
|
if (block_i) {
|
|
|
|
block_i->last_alloc_logical_block = block + blks - 1;
|
|
|
|
block_i->last_alloc_physical_block =
|
|
|
|
le32_to_cpu(where[num].key) + blks - 1;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-17 06:30:46 +00:00
|
|
|
/* We are done with atomic stuff, now do the rest of housekeeping */
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/* had we spliced it onto indirect block? */
|
|
|
|
if (where->bh)
|
|
|
|
mark_buffer_dirty_inode(where->bh, inode);
|
|
|
|
|
2016-09-14 14:48:05 +00:00
|
|
|
inode->i_ctime = current_time(inode);
|
2005-04-16 22:20:36 +00:00
|
|
|
mark_inode_dirty(inode);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocation strategy is simple: if we have to allocate something, we will
|
|
|
|
* have to go the whole way to leaf. So let's do it before attaching anything
|
|
|
|
* to tree, set linkage between the newborn blocks, write them if sync is
|
|
|
|
* required, recheck the path, free and repeat if check fails, otherwise
|
|
|
|
* set the last missing link (that will protect us from any truncate-generated
|
|
|
|
* removals - all blocks on the path are immune now) and possibly force the
|
|
|
|
* write on the parent block.
|
|
|
|
* That has a nice additional property: no special recovery from the failed
|
|
|
|
* allocations is needed - we simply release blocks and do not touch anything
|
|
|
|
* reachable from inode.
|
2007-10-17 06:30:46 +00:00
|
|
|
*
|
|
|
|
* `handle' can be NULL if create == 0.
|
|
|
|
*
|
|
|
|
* return > 0, # of blocks mapped or allocated.
|
|
|
|
* return = 0, if plain lookup failed.
|
|
|
|
* return < 0, error case.
|
2005-04-16 22:20:36 +00:00
|
|
|
*/
|
2007-10-17 06:30:46 +00:00
|
|
|
static int ext2_get_blocks(struct inode *inode,
|
|
|
|
sector_t iblock, unsigned long maxblocks,
|
2016-09-19 01:28:39 +00:00
|
|
|
u32 *bno, bool *new, bool *boundary,
|
2007-10-17 06:30:46 +00:00
|
|
|
int create)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2016-10-17 22:05:34 +00:00
|
|
|
int err;
|
2005-04-16 22:20:36 +00:00
|
|
|
int offsets[4];
|
|
|
|
Indirect chain[4];
|
|
|
|
Indirect *partial;
|
2007-10-17 06:30:46 +00:00
|
|
|
ext2_fsblk_t goal;
|
|
|
|
int indirect_blks;
|
|
|
|
int blocks_to_boundary = 0;
|
|
|
|
int depth;
|
|
|
|
struct ext2_inode_info *ei = EXT2_I(inode);
|
|
|
|
int count = 0;
|
|
|
|
ext2_fsblk_t first_block = 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2013-11-05 00:15:38 +00:00
|
|
|
BUG_ON(maxblocks == 0);
|
|
|
|
|
2007-10-17 06:30:46 +00:00
|
|
|
depth = ext2_block_to_path(inode,iblock,offsets,&blocks_to_boundary);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-17 06:30:46 +00:00
|
|
|
if (depth == 0)
|
2016-10-17 22:05:34 +00:00
|
|
|
return -EIO;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
ext2: fix data corruption for racing writes
If two writers allocating blocks to file race with each other (e.g.
because writepages races with ordinary write or two writepages race with
each other), ext2_getblock() can be called on the same inode in parallel.
Before we are going to allocate new blocks, we have to recheck the block
chain we have obtained so far without holding truncate_mutex. Otherwise
we could overwrite the indirect block pointer set by the other writer
leading to data loss.
The below test program by Ying is able to reproduce the data loss with ext2
on in BRD in a few minutes if the machine is under memory pressure:
long kMemSize = 50 << 20;
int kPageSize = 4096;
int main(int argc, char **argv) {
int status;
int count = 0;
int i;
char *fname = "/mnt/test.mmap";
char *mem;
unlink(fname);
int fd = open(fname, O_CREAT | O_EXCL | O_RDWR, 0600);
status = ftruncate(fd, kMemSize);
mem = mmap(0, kMemSize, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
// Fill the memory with 1s.
memset(mem, 1, kMemSize);
sleep(2);
for (i = 0; i < kMemSize; i++) {
int byte_good = mem[i] != 0;
if (!byte_good && ((i % kPageSize) == 0)) {
//printf("%d ", i / kPageSize);
count++;
}
}
munmap(mem, kMemSize);
close(fd);
unlink(fname);
if (count > 0) {
printf("Running %d bad page\n", count);
return 1;
}
return 0;
}
Cc: Ying Han <yinghan@google.com>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Jan Kara <jack@suse.cz>
Cc: Mingming Cao <cmm@us.ibm.com>
Cc: <linux-ext4@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-13 21:40:14 +00:00
|
|
|
partial = ext2_get_branch(inode, depth, offsets, chain, &err);
|
2005-04-16 22:20:36 +00:00
|
|
|
/* Simplest case - block found, no allocation needed */
|
|
|
|
if (!partial) {
|
2007-10-17 06:30:46 +00:00
|
|
|
first_block = le32_to_cpu(chain[depth - 1].key);
|
|
|
|
count++;
|
|
|
|
/*map more blocks*/
|
|
|
|
while (count < maxblocks && count <= blocks_to_boundary) {
|
|
|
|
ext2_fsblk_t blk;
|
|
|
|
|
ext2: fix data corruption for racing writes
If two writers allocating blocks to file race with each other (e.g.
because writepages races with ordinary write or two writepages race with
each other), ext2_getblock() can be called on the same inode in parallel.
Before we are going to allocate new blocks, we have to recheck the block
chain we have obtained so far without holding truncate_mutex. Otherwise
we could overwrite the indirect block pointer set by the other writer
leading to data loss.
The below test program by Ying is able to reproduce the data loss with ext2
on in BRD in a few minutes if the machine is under memory pressure:
long kMemSize = 50 << 20;
int kPageSize = 4096;
int main(int argc, char **argv) {
int status;
int count = 0;
int i;
char *fname = "/mnt/test.mmap";
char *mem;
unlink(fname);
int fd = open(fname, O_CREAT | O_EXCL | O_RDWR, 0600);
status = ftruncate(fd, kMemSize);
mem = mmap(0, kMemSize, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
// Fill the memory with 1s.
memset(mem, 1, kMemSize);
sleep(2);
for (i = 0; i < kMemSize; i++) {
int byte_good = mem[i] != 0;
if (!byte_good && ((i % kPageSize) == 0)) {
//printf("%d ", i / kPageSize);
count++;
}
}
munmap(mem, kMemSize);
close(fd);
unlink(fname);
if (count > 0) {
printf("Running %d bad page\n", count);
return 1;
}
return 0;
}
Cc: Ying Han <yinghan@google.com>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Jan Kara <jack@suse.cz>
Cc: Mingming Cao <cmm@us.ibm.com>
Cc: <linux-ext4@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-13 21:40:14 +00:00
|
|
|
if (!verify_chain(chain, chain + depth - 1)) {
|
2007-10-17 06:30:46 +00:00
|
|
|
/*
|
|
|
|
* Indirect block might be removed by
|
|
|
|
* truncate while we were reading it.
|
|
|
|
* Handling of that case: forget what we've
|
|
|
|
* got now, go to reread.
|
|
|
|
*/
|
ext2: fix data corruption for racing writes
If two writers allocating blocks to file race with each other (e.g.
because writepages races with ordinary write or two writepages race with
each other), ext2_getblock() can be called on the same inode in parallel.
Before we are going to allocate new blocks, we have to recheck the block
chain we have obtained so far without holding truncate_mutex. Otherwise
we could overwrite the indirect block pointer set by the other writer
leading to data loss.
The below test program by Ying is able to reproduce the data loss with ext2
on in BRD in a few minutes if the machine is under memory pressure:
long kMemSize = 50 << 20;
int kPageSize = 4096;
int main(int argc, char **argv) {
int status;
int count = 0;
int i;
char *fname = "/mnt/test.mmap";
char *mem;
unlink(fname);
int fd = open(fname, O_CREAT | O_EXCL | O_RDWR, 0600);
status = ftruncate(fd, kMemSize);
mem = mmap(0, kMemSize, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
// Fill the memory with 1s.
memset(mem, 1, kMemSize);
sleep(2);
for (i = 0; i < kMemSize; i++) {
int byte_good = mem[i] != 0;
if (!byte_good && ((i % kPageSize) == 0)) {
//printf("%d ", i / kPageSize);
count++;
}
}
munmap(mem, kMemSize);
close(fd);
unlink(fname);
if (count > 0) {
printf("Running %d bad page\n", count);
return 1;
}
return 0;
}
Cc: Ying Han <yinghan@google.com>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Jan Kara <jack@suse.cz>
Cc: Mingming Cao <cmm@us.ibm.com>
Cc: <linux-ext4@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-13 21:40:14 +00:00
|
|
|
err = -EAGAIN;
|
2007-10-17 06:30:46 +00:00
|
|
|
count = 0;
|
2017-06-24 00:37:21 +00:00
|
|
|
partial = chain + depth - 1;
|
ext2: fix data corruption for racing writes
If two writers allocating blocks to file race with each other (e.g.
because writepages races with ordinary write or two writepages race with
each other), ext2_getblock() can be called on the same inode in parallel.
Before we are going to allocate new blocks, we have to recheck the block
chain we have obtained so far without holding truncate_mutex. Otherwise
we could overwrite the indirect block pointer set by the other writer
leading to data loss.
The below test program by Ying is able to reproduce the data loss with ext2
on in BRD in a few minutes if the machine is under memory pressure:
long kMemSize = 50 << 20;
int kPageSize = 4096;
int main(int argc, char **argv) {
int status;
int count = 0;
int i;
char *fname = "/mnt/test.mmap";
char *mem;
unlink(fname);
int fd = open(fname, O_CREAT | O_EXCL | O_RDWR, 0600);
status = ftruncate(fd, kMemSize);
mem = mmap(0, kMemSize, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
// Fill the memory with 1s.
memset(mem, 1, kMemSize);
sleep(2);
for (i = 0; i < kMemSize; i++) {
int byte_good = mem[i] != 0;
if (!byte_good && ((i % kPageSize) == 0)) {
//printf("%d ", i / kPageSize);
count++;
}
}
munmap(mem, kMemSize);
close(fd);
unlink(fname);
if (count > 0) {
printf("Running %d bad page\n", count);
return 1;
}
return 0;
}
Cc: Ying Han <yinghan@google.com>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Jan Kara <jack@suse.cz>
Cc: Mingming Cao <cmm@us.ibm.com>
Cc: <linux-ext4@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-13 21:40:14 +00:00
|
|
|
break;
|
2007-10-17 06:30:46 +00:00
|
|
|
}
|
|
|
|
blk = le32_to_cpu(*(chain[depth-1].p + count));
|
|
|
|
if (blk == first_block + count)
|
|
|
|
count++;
|
|
|
|
else
|
|
|
|
break;
|
|
|
|
}
|
ext2: fix data corruption for racing writes
If two writers allocating blocks to file race with each other (e.g.
because writepages races with ordinary write or two writepages race with
each other), ext2_getblock() can be called on the same inode in parallel.
Before we are going to allocate new blocks, we have to recheck the block
chain we have obtained so far without holding truncate_mutex. Otherwise
we could overwrite the indirect block pointer set by the other writer
leading to data loss.
The below test program by Ying is able to reproduce the data loss with ext2
on in BRD in a few minutes if the machine is under memory pressure:
long kMemSize = 50 << 20;
int kPageSize = 4096;
int main(int argc, char **argv) {
int status;
int count = 0;
int i;
char *fname = "/mnt/test.mmap";
char *mem;
unlink(fname);
int fd = open(fname, O_CREAT | O_EXCL | O_RDWR, 0600);
status = ftruncate(fd, kMemSize);
mem = mmap(0, kMemSize, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
// Fill the memory with 1s.
memset(mem, 1, kMemSize);
sleep(2);
for (i = 0; i < kMemSize; i++) {
int byte_good = mem[i] != 0;
if (!byte_good && ((i % kPageSize) == 0)) {
//printf("%d ", i / kPageSize);
count++;
}
}
munmap(mem, kMemSize);
close(fd);
unlink(fname);
if (count > 0) {
printf("Running %d bad page\n", count);
return 1;
}
return 0;
}
Cc: Ying Han <yinghan@google.com>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Jan Kara <jack@suse.cz>
Cc: Mingming Cao <cmm@us.ibm.com>
Cc: <linux-ext4@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-13 21:40:14 +00:00
|
|
|
if (err != -EAGAIN)
|
|
|
|
goto got_it;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/* Next simple case - plain lookup or failed read of indirect block */
|
2007-10-17 06:30:46 +00:00
|
|
|
if (!create || err == -EIO)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
mutex_lock(&ei->truncate_mutex);
|
ext2: fix data corruption for racing writes
If two writers allocating blocks to file race with each other (e.g.
because writepages races with ordinary write or two writepages race with
each other), ext2_getblock() can be called on the same inode in parallel.
Before we are going to allocate new blocks, we have to recheck the block
chain we have obtained so far without holding truncate_mutex. Otherwise
we could overwrite the indirect block pointer set by the other writer
leading to data loss.
The below test program by Ying is able to reproduce the data loss with ext2
on in BRD in a few minutes if the machine is under memory pressure:
long kMemSize = 50 << 20;
int kPageSize = 4096;
int main(int argc, char **argv) {
int status;
int count = 0;
int i;
char *fname = "/mnt/test.mmap";
char *mem;
unlink(fname);
int fd = open(fname, O_CREAT | O_EXCL | O_RDWR, 0600);
status = ftruncate(fd, kMemSize);
mem = mmap(0, kMemSize, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
// Fill the memory with 1s.
memset(mem, 1, kMemSize);
sleep(2);
for (i = 0; i < kMemSize; i++) {
int byte_good = mem[i] != 0;
if (!byte_good && ((i % kPageSize) == 0)) {
//printf("%d ", i / kPageSize);
count++;
}
}
munmap(mem, kMemSize);
close(fd);
unlink(fname);
if (count > 0) {
printf("Running %d bad page\n", count);
return 1;
}
return 0;
}
Cc: Ying Han <yinghan@google.com>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Jan Kara <jack@suse.cz>
Cc: Mingming Cao <cmm@us.ibm.com>
Cc: <linux-ext4@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-13 21:40:14 +00:00
|
|
|
/*
|
|
|
|
* If the indirect block is missing while we are reading
|
2010-09-15 12:46:02 +00:00
|
|
|
* the chain(ext2_get_branch() returns -EAGAIN err), or
|
ext2: fix data corruption for racing writes
If two writers allocating blocks to file race with each other (e.g.
because writepages races with ordinary write or two writepages race with
each other), ext2_getblock() can be called on the same inode in parallel.
Before we are going to allocate new blocks, we have to recheck the block
chain we have obtained so far without holding truncate_mutex. Otherwise
we could overwrite the indirect block pointer set by the other writer
leading to data loss.
The below test program by Ying is able to reproduce the data loss with ext2
on in BRD in a few minutes if the machine is under memory pressure:
long kMemSize = 50 << 20;
int kPageSize = 4096;
int main(int argc, char **argv) {
int status;
int count = 0;
int i;
char *fname = "/mnt/test.mmap";
char *mem;
unlink(fname);
int fd = open(fname, O_CREAT | O_EXCL | O_RDWR, 0600);
status = ftruncate(fd, kMemSize);
mem = mmap(0, kMemSize, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
// Fill the memory with 1s.
memset(mem, 1, kMemSize);
sleep(2);
for (i = 0; i < kMemSize; i++) {
int byte_good = mem[i] != 0;
if (!byte_good && ((i % kPageSize) == 0)) {
//printf("%d ", i / kPageSize);
count++;
}
}
munmap(mem, kMemSize);
close(fd);
unlink(fname);
if (count > 0) {
printf("Running %d bad page\n", count);
return 1;
}
return 0;
}
Cc: Ying Han <yinghan@google.com>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Jan Kara <jack@suse.cz>
Cc: Mingming Cao <cmm@us.ibm.com>
Cc: <linux-ext4@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-13 21:40:14 +00:00
|
|
|
* if the chain has been changed after we grab the semaphore,
|
|
|
|
* (either because another process truncated this branch, or
|
|
|
|
* another get_block allocated this branch) re-grab the chain to see if
|
|
|
|
* the request block has been allocated or not.
|
|
|
|
*
|
|
|
|
* Since we already block the truncate/other get_block
|
|
|
|
* at this point, we will have the current copy of the chain when we
|
|
|
|
* splice the branch into the tree.
|
|
|
|
*/
|
|
|
|
if (err == -EAGAIN || !verify_chain(chain, partial)) {
|
|
|
|
while (partial > chain) {
|
|
|
|
brelse(partial->bh);
|
|
|
|
partial--;
|
|
|
|
}
|
|
|
|
partial = ext2_get_branch(inode, depth, offsets, chain, &err);
|
|
|
|
if (!partial) {
|
|
|
|
count++;
|
|
|
|
mutex_unlock(&ei->truncate_mutex);
|
|
|
|
goto got_it;
|
|
|
|
}
|
2019-11-05 04:51:00 +00:00
|
|
|
|
|
|
|
if (err) {
|
|
|
|
mutex_unlock(&ei->truncate_mutex);
|
|
|
|
goto cleanup;
|
|
|
|
}
|
ext2: fix data corruption for racing writes
If two writers allocating blocks to file race with each other (e.g.
because writepages races with ordinary write or two writepages race with
each other), ext2_getblock() can be called on the same inode in parallel.
Before we are going to allocate new blocks, we have to recheck the block
chain we have obtained so far without holding truncate_mutex. Otherwise
we could overwrite the indirect block pointer set by the other writer
leading to data loss.
The below test program by Ying is able to reproduce the data loss with ext2
on in BRD in a few minutes if the machine is under memory pressure:
long kMemSize = 50 << 20;
int kPageSize = 4096;
int main(int argc, char **argv) {
int status;
int count = 0;
int i;
char *fname = "/mnt/test.mmap";
char *mem;
unlink(fname);
int fd = open(fname, O_CREAT | O_EXCL | O_RDWR, 0600);
status = ftruncate(fd, kMemSize);
mem = mmap(0, kMemSize, PROT_READ | PROT_WRITE, MAP_SHARED, fd, 0);
// Fill the memory with 1s.
memset(mem, 1, kMemSize);
sleep(2);
for (i = 0; i < kMemSize; i++) {
int byte_good = mem[i] != 0;
if (!byte_good && ((i % kPageSize) == 0)) {
//printf("%d ", i / kPageSize);
count++;
}
}
munmap(mem, kMemSize);
close(fd);
unlink(fname);
if (count > 0) {
printf("Running %d bad page\n", count);
return 1;
}
return 0;
}
Cc: Ying Han <yinghan@google.com>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Signed-off-by: Jan Kara <jack@suse.cz>
Cc: Mingming Cao <cmm@us.ibm.com>
Cc: <linux-ext4@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2009-04-13 21:40:14 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
/*
|
2007-10-17 06:30:46 +00:00
|
|
|
* Okay, we need to do block allocation. Lazily initialize the block
|
|
|
|
* allocation info here if necessary
|
|
|
|
*/
|
|
|
|
if (S_ISREG(inode->i_mode) && (!ei->i_block_alloc_info))
|
|
|
|
ext2_init_block_alloc_info(inode);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-02-06 09:40:16 +00:00
|
|
|
goal = ext2_find_goal(inode, iblock, partial);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2007-10-17 06:30:46 +00:00
|
|
|
/* the number of blocks need to allocate for [d,t]indirect blocks */
|
|
|
|
indirect_blks = (chain + depth) - partial - 1;
|
|
|
|
/*
|
2019-01-29 12:39:49 +00:00
|
|
|
* Next look up the indirect map to count the total number of
|
2007-10-17 06:30:46 +00:00
|
|
|
* direct blocks to allocate for this branch.
|
|
|
|
*/
|
|
|
|
count = ext2_blks_to_allocate(partial, indirect_blks,
|
|
|
|
maxblocks, blocks_to_boundary);
|
|
|
|
/*
|
|
|
|
* XXX ???? Block out ext2_truncate while we alter the tree
|
|
|
|
*/
|
|
|
|
err = ext2_alloc_branch(inode, indirect_blks, &count, goal,
|
|
|
|
offsets + (partial - chain), partial);
|
|
|
|
|
|
|
|
if (err) {
|
|
|
|
mutex_unlock(&ei->truncate_mutex);
|
2005-04-16 22:20:36 +00:00
|
|
|
goto cleanup;
|
2007-10-17 06:30:46 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2015-02-16 23:58:53 +00:00
|
|
|
if (IS_DAX(inode)) {
|
2016-09-27 14:35:45 +00:00
|
|
|
/*
|
|
|
|
* We must unmap blocks before zeroing so that writeback cannot
|
|
|
|
* overwrite zeros with stale data from block device page cache.
|
|
|
|
*/
|
2016-11-04 17:08:14 +00:00
|
|
|
clean_bdev_aliases(inode->i_sb->s_bdev,
|
|
|
|
le32_to_cpu(chain[depth-1].key),
|
|
|
|
count);
|
2005-06-24 05:05:26 +00:00
|
|
|
/*
|
2015-02-16 23:58:59 +00:00
|
|
|
* block must be initialised before we put it in the tree
|
|
|
|
* so that it's not found by another thread before it's
|
|
|
|
* initialised
|
2005-06-24 05:05:26 +00:00
|
|
|
*/
|
2016-03-15 17:20:41 +00:00
|
|
|
err = sb_issue_zeroout(inode->i_sb,
|
|
|
|
le32_to_cpu(chain[depth-1].key), count,
|
|
|
|
GFP_NOFS);
|
2007-10-17 06:30:46 +00:00
|
|
|
if (err) {
|
|
|
|
mutex_unlock(&ei->truncate_mutex);
|
2005-06-24 05:05:26 +00:00
|
|
|
goto cleanup;
|
2007-10-17 06:30:46 +00:00
|
|
|
}
|
2016-09-19 01:28:39 +00:00
|
|
|
}
|
2016-08-10 14:42:53 +00:00
|
|
|
*new = true;
|
2005-06-24 05:05:26 +00:00
|
|
|
|
2007-10-17 06:30:46 +00:00
|
|
|
ext2_splice_branch(inode, iblock, partial, indirect_blks, count);
|
|
|
|
mutex_unlock(&ei->truncate_mutex);
|
|
|
|
got_it:
|
|
|
|
if (count > blocks_to_boundary)
|
2016-09-19 01:28:39 +00:00
|
|
|
*boundary = true;
|
2007-10-17 06:30:46 +00:00
|
|
|
err = count;
|
|
|
|
/* Clean up and exit */
|
|
|
|
partial = chain + depth - 1; /* the whole chain */
|
|
|
|
cleanup:
|
|
|
|
while (partial > chain) {
|
|
|
|
brelse(partial->bh);
|
|
|
|
partial--;
|
|
|
|
}
|
2016-10-17 22:05:34 +00:00
|
|
|
if (err > 0)
|
|
|
|
*bno = le32_to_cpu(chain[depth-1].key);
|
2007-10-17 06:30:46 +00:00
|
|
|
return err;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2016-09-19 01:28:39 +00:00
|
|
|
int ext2_get_block(struct inode *inode, sector_t iblock,
|
|
|
|
struct buffer_head *bh_result, int create)
|
2007-10-17 06:30:46 +00:00
|
|
|
{
|
|
|
|
unsigned max_blocks = bh_result->b_size >> inode->i_blkbits;
|
2016-09-19 01:28:39 +00:00
|
|
|
bool new = false, boundary = false;
|
|
|
|
u32 bno;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = ext2_get_blocks(inode, iblock, max_blocks, &bno, &new, &boundary,
|
|
|
|
create);
|
|
|
|
if (ret <= 0)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
map_bh(bh_result, inode->i_sb, bno);
|
|
|
|
bh_result->b_size = (ret << inode->i_blkbits);
|
|
|
|
if (new)
|
|
|
|
set_buffer_new(bh_result);
|
|
|
|
if (boundary)
|
|
|
|
set_buffer_boundary(bh_result);
|
|
|
|
return 0;
|
2007-10-17 06:30:46 +00:00
|
|
|
|
|
|
|
}
|
|
|
|
|
2016-09-19 01:30:29 +00:00
|
|
|
static int ext2_iomap_begin(struct inode *inode, loff_t offset, loff_t length,
|
2019-10-18 23:44:10 +00:00
|
|
|
unsigned flags, struct iomap *iomap, struct iomap *srcmap)
|
2016-09-19 01:30:29 +00:00
|
|
|
{
|
|
|
|
unsigned int blkbits = inode->i_blkbits;
|
|
|
|
unsigned long first_block = offset >> blkbits;
|
|
|
|
unsigned long max_blocks = (length + (1 << blkbits) - 1) >> blkbits;
|
2017-08-30 16:43:34 +00:00
|
|
|
struct ext2_sb_info *sbi = EXT2_SB(inode->i_sb);
|
2016-09-19 01:30:29 +00:00
|
|
|
bool new = false, boundary = false;
|
|
|
|
u32 bno;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = ext2_get_blocks(inode, first_block, max_blocks,
|
|
|
|
&bno, &new, &boundary, flags & IOMAP_WRITE);
|
|
|
|
if (ret < 0)
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
iomap->flags = 0;
|
2017-08-30 16:43:34 +00:00
|
|
|
iomap->bdev = inode->i_sb->s_bdev;
|
2016-10-02 22:46:04 +00:00
|
|
|
iomap->offset = (u64)first_block << blkbits;
|
2017-08-30 16:43:34 +00:00
|
|
|
iomap->dax_dev = sbi->s_daxdev;
|
2016-09-19 01:30:29 +00:00
|
|
|
|
|
|
|
if (ret == 0) {
|
|
|
|
iomap->type = IOMAP_HOLE;
|
2017-10-01 21:55:54 +00:00
|
|
|
iomap->addr = IOMAP_NULL_ADDR;
|
2016-09-19 01:30:29 +00:00
|
|
|
iomap->length = 1 << blkbits;
|
|
|
|
} else {
|
|
|
|
iomap->type = IOMAP_MAPPED;
|
2017-10-01 21:55:54 +00:00
|
|
|
iomap->addr = (u64)bno << blkbits;
|
2016-09-19 01:30:29 +00:00
|
|
|
iomap->length = (u64)ret << blkbits;
|
|
|
|
iomap->flags |= IOMAP_F_MERGED;
|
2007-10-17 06:30:46 +00:00
|
|
|
}
|
|
|
|
|
2016-09-19 01:30:29 +00:00
|
|
|
if (new)
|
|
|
|
iomap->flags |= IOMAP_F_NEW;
|
|
|
|
return 0;
|
2007-10-17 06:30:46 +00:00
|
|
|
}
|
|
|
|
|
2016-09-19 01:30:29 +00:00
|
|
|
static int
|
|
|
|
ext2_iomap_end(struct inode *inode, loff_t offset, loff_t length,
|
|
|
|
ssize_t written, unsigned flags, struct iomap *iomap)
|
|
|
|
{
|
|
|
|
if (iomap->type == IOMAP_MAPPED &&
|
|
|
|
written < length &&
|
|
|
|
(flags & IOMAP_WRITE))
|
|
|
|
ext2_write_failed(inode->i_mapping, offset + length);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-01-28 07:20:26 +00:00
|
|
|
const struct iomap_ops ext2_iomap_ops = {
|
2016-09-19 01:30:29 +00:00
|
|
|
.iomap_begin = ext2_iomap_begin,
|
|
|
|
.iomap_end = ext2_iomap_end,
|
|
|
|
};
|
|
|
|
|
2008-10-03 21:32:43 +00:00
|
|
|
int ext2_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo,
|
|
|
|
u64 start, u64 len)
|
|
|
|
{
|
|
|
|
return generic_block_fiemap(inode, fieinfo, start, len,
|
|
|
|
ext2_get_block);
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
static int ext2_writepage(struct page *page, struct writeback_control *wbc)
|
|
|
|
{
|
|
|
|
return block_write_full_page(page, ext2_get_block, wbc);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int ext2_readpage(struct file *file, struct page *page)
|
|
|
|
{
|
|
|
|
return mpage_readpage(page, ext2_get_block);
|
|
|
|
}
|
|
|
|
|
fs: convert mpage_readpages to mpage_readahead
Implement the new readahead aop and convert all callers (block_dev,
exfat, ext2, fat, gfs2, hpfs, isofs, jfs, nilfs2, ocfs2, omfs, qnx6,
reiserfs & udf).
The callers are all trivial except for GFS2 & OCFS2.
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Junxiao Bi <junxiao.bi@oracle.com> # ocfs2
Reviewed-by: Joseph Qi <joseph.qi@linux.alibaba.com> # ocfs2
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: John Hubbard <jhubbard@nvidia.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: William Kucharski <william.kucharski@oracle.com>
Cc: Chao Yu <yuchao0@huawei.com>
Cc: Cong Wang <xiyou.wangcong@gmail.com>
Cc: Darrick J. Wong <darrick.wong@oracle.com>
Cc: Eric Biggers <ebiggers@google.com>
Cc: Gao Xiang <gaoxiang25@huawei.com>
Cc: Jaegeuk Kim <jaegeuk@kernel.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Zi Yan <ziy@nvidia.com>
Cc: Johannes Thumshirn <johannes.thumshirn@wdc.com>
Cc: Miklos Szeredi <mszeredi@redhat.com>
Link: http://lkml.kernel.org/r/20200414150233.24495-17-willy@infradead.org
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2020-06-02 04:47:02 +00:00
|
|
|
static void ext2_readahead(struct readahead_control *rac)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
fs: convert mpage_readpages to mpage_readahead
Implement the new readahead aop and convert all callers (block_dev,
exfat, ext2, fat, gfs2, hpfs, isofs, jfs, nilfs2, ocfs2, omfs, qnx6,
reiserfs & udf).
The callers are all trivial except for GFS2 & OCFS2.
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Junxiao Bi <junxiao.bi@oracle.com> # ocfs2
Reviewed-by: Joseph Qi <joseph.qi@linux.alibaba.com> # ocfs2
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: John Hubbard <jhubbard@nvidia.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: William Kucharski <william.kucharski@oracle.com>
Cc: Chao Yu <yuchao0@huawei.com>
Cc: Cong Wang <xiyou.wangcong@gmail.com>
Cc: Darrick J. Wong <darrick.wong@oracle.com>
Cc: Eric Biggers <ebiggers@google.com>
Cc: Gao Xiang <gaoxiang25@huawei.com>
Cc: Jaegeuk Kim <jaegeuk@kernel.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Zi Yan <ziy@nvidia.com>
Cc: Johannes Thumshirn <johannes.thumshirn@wdc.com>
Cc: Miklos Szeredi <mszeredi@redhat.com>
Link: http://lkml.kernel.org/r/20200414150233.24495-17-willy@infradead.org
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2020-06-02 04:47:02 +00:00
|
|
|
mpage_readahead(rac, ext2_get_block);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2007-10-16 08:25:04 +00:00
|
|
|
ext2_write_begin(struct file *file, struct address_space *mapping,
|
|
|
|
loff_t pos, unsigned len, unsigned flags,
|
|
|
|
struct page **pagep, void **fsdata)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2010-05-26 15:05:37 +00:00
|
|
|
int ret;
|
|
|
|
|
2010-06-04 09:29:58 +00:00
|
|
|
ret = block_write_begin(mapping, pos, len, flags, pagep,
|
|
|
|
ext2_get_block);
|
2010-05-26 15:05:37 +00:00
|
|
|
if (ret < 0)
|
|
|
|
ext2_write_failed(mapping, pos + len);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int ext2_write_end(struct file *file, struct address_space *mapping,
|
|
|
|
loff_t pos, unsigned len, unsigned copied,
|
|
|
|
struct page *page, void *fsdata)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = generic_write_end(file, mapping, pos, len, copied, page, fsdata);
|
|
|
|
if (ret < len)
|
|
|
|
ext2_write_failed(mapping, pos + len);
|
|
|
|
return ret;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2007-10-16 08:25:25 +00:00
|
|
|
static int
|
|
|
|
ext2_nobh_write_begin(struct file *file, struct address_space *mapping,
|
|
|
|
loff_t pos, unsigned len, unsigned flags,
|
|
|
|
struct page **pagep, void **fsdata)
|
|
|
|
{
|
2010-05-26 15:05:37 +00:00
|
|
|
int ret;
|
|
|
|
|
2010-06-04 09:29:54 +00:00
|
|
|
ret = nobh_write_begin(mapping, pos, len, flags, pagep, fsdata,
|
|
|
|
ext2_get_block);
|
2010-05-26 15:05:37 +00:00
|
|
|
if (ret < 0)
|
|
|
|
ext2_write_failed(mapping, pos + len);
|
|
|
|
return ret;
|
2007-10-16 08:25:25 +00:00
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
static int ext2_nobh_writepage(struct page *page,
|
|
|
|
struct writeback_control *wbc)
|
|
|
|
{
|
|
|
|
return nobh_writepage(page, ext2_get_block, wbc);
|
|
|
|
}
|
|
|
|
|
|
|
|
static sector_t ext2_bmap(struct address_space *mapping, sector_t block)
|
|
|
|
{
|
|
|
|
return generic_block_bmap(mapping,block,ext2_get_block);
|
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t
|
2016-04-07 15:51:58 +00:00
|
|
|
ext2_direct_IO(struct kiocb *iocb, struct iov_iter *iter)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct file *file = iocb->ki_filp;
|
2010-05-26 15:05:37 +00:00
|
|
|
struct address_space *mapping = file->f_mapping;
|
|
|
|
struct inode *inode = mapping->host;
|
2014-03-05 03:38:00 +00:00
|
|
|
size_t count = iov_iter_count(iter);
|
2016-04-07 15:51:58 +00:00
|
|
|
loff_t offset = iocb->ki_pos;
|
2010-05-26 15:05:37 +00:00
|
|
|
ssize_t ret;
|
|
|
|
|
2016-09-19 01:30:29 +00:00
|
|
|
ret = blockdev_direct_IO(iocb, inode, iter, ext2_get_block);
|
2015-03-16 11:33:52 +00:00
|
|
|
if (ret < 0 && iov_iter_rw(iter) == WRITE)
|
2014-03-05 03:38:00 +00:00
|
|
|
ext2_write_failed(mapping, offset + count);
|
2010-05-26 15:05:37 +00:00
|
|
|
return ret;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
ext2_writepages(struct address_space *mapping, struct writeback_control *wbc)
|
|
|
|
{
|
|
|
|
return mpage_writepages(mapping, wbc, ext2_get_block);
|
|
|
|
}
|
|
|
|
|
2017-12-21 20:25:11 +00:00
|
|
|
static int
|
|
|
|
ext2_dax_writepages(struct address_space *mapping, struct writeback_control *wbc)
|
|
|
|
{
|
2020-01-03 18:33:07 +00:00
|
|
|
struct ext2_sb_info *sbi = EXT2_SB(mapping->host->i_sb);
|
|
|
|
|
|
|
|
return dax_writeback_mapping_range(mapping, sbi->s_daxdev, wbc);
|
2017-12-21 20:25:11 +00:00
|
|
|
}
|
|
|
|
|
2006-06-28 11:26:44 +00:00
|
|
|
const struct address_space_operations ext2_aops = {
|
2021-06-29 02:36:12 +00:00
|
|
|
.set_page_dirty = __set_page_dirty_buffers,
|
2005-04-16 22:20:36 +00:00
|
|
|
.readpage = ext2_readpage,
|
fs: convert mpage_readpages to mpage_readahead
Implement the new readahead aop and convert all callers (block_dev,
exfat, ext2, fat, gfs2, hpfs, isofs, jfs, nilfs2, ocfs2, omfs, qnx6,
reiserfs & udf).
The callers are all trivial except for GFS2 & OCFS2.
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Junxiao Bi <junxiao.bi@oracle.com> # ocfs2
Reviewed-by: Joseph Qi <joseph.qi@linux.alibaba.com> # ocfs2
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: John Hubbard <jhubbard@nvidia.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: William Kucharski <william.kucharski@oracle.com>
Cc: Chao Yu <yuchao0@huawei.com>
Cc: Cong Wang <xiyou.wangcong@gmail.com>
Cc: Darrick J. Wong <darrick.wong@oracle.com>
Cc: Eric Biggers <ebiggers@google.com>
Cc: Gao Xiang <gaoxiang25@huawei.com>
Cc: Jaegeuk Kim <jaegeuk@kernel.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Zi Yan <ziy@nvidia.com>
Cc: Johannes Thumshirn <johannes.thumshirn@wdc.com>
Cc: Miklos Szeredi <mszeredi@redhat.com>
Link: http://lkml.kernel.org/r/20200414150233.24495-17-willy@infradead.org
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2020-06-02 04:47:02 +00:00
|
|
|
.readahead = ext2_readahead,
|
2005-04-16 22:20:36 +00:00
|
|
|
.writepage = ext2_writepage,
|
2007-10-16 08:25:04 +00:00
|
|
|
.write_begin = ext2_write_begin,
|
2010-05-26 15:05:37 +00:00
|
|
|
.write_end = ext2_write_end,
|
2005-04-16 22:20:36 +00:00
|
|
|
.bmap = ext2_bmap,
|
|
|
|
.direct_IO = ext2_direct_IO,
|
|
|
|
.writepages = ext2_writepages,
|
2006-02-01 11:05:41 +00:00
|
|
|
.migratepage = buffer_migrate_page,
|
vfs: pagecache usage optimization for pagesize!=blocksize
When we read some part of a file through pagecache, if there is a
pagecache of corresponding index but this page is not uptodate, read IO
is issued and this page will be uptodate.
I think this is good for pagesize == blocksize environment but there is
room for improvement on pagesize != blocksize environment. Because in
this case a page can have multiple buffers and even if a page is not
uptodate, some buffers can be uptodate.
So I suggest that when all buffers which correspond to a part of a file
that we want to read are uptodate, use this pagecache and copy data from
this pagecache to user buffer even if a page is not uptodate. This can
reduce read IO and improve system throughput.
I wrote a benchmark program and got result number with this program.
This benchmark do:
1: mount and open a test file.
2: create a 512MB file.
3: close a file and umount.
4: mount and again open a test file.
5: pwrite randomly 300000 times on a test file. offset is aligned
by IO size(1024bytes).
6: measure time of preading randomly 100000 times on a test file.
The result was:
2.6.26
330 sec
2.6.26-patched
226 sec
Arch:i386
Filesystem:ext3
Blocksize:1024 bytes
Memory: 1GB
On ext3/4, a file is written through buffer/block. So random read/write
mixed workloads or random read after random write workloads are optimized
with this patch under pagesize != blocksize environment. This test result
showed this.
The benchmark program is as follows:
#include <stdio.h>
#include <sys/types.h>
#include <sys/stat.h>
#include <fcntl.h>
#include <unistd.h>
#include <time.h>
#include <stdlib.h>
#include <string.h>
#include <sys/mount.h>
#define LEN 1024
#define LOOP 1024*512 /* 512MB */
main(void)
{
unsigned long i, offset, filesize;
int fd;
char buf[LEN];
time_t t1, t2;
if (mount("/dev/sda1", "/root/test1/", "ext3", 0, 0) < 0) {
perror("cannot mount\n");
exit(1);
}
memset(buf, 0, LEN);
fd = open("/root/test1/testfile", O_CREAT|O_RDWR|O_TRUNC);
if (fd < 0) {
perror("cannot open file\n");
exit(1);
}
for (i = 0; i < LOOP; i++)
write(fd, buf, LEN);
close(fd);
if (umount("/root/test1/") < 0) {
perror("cannot umount\n");
exit(1);
}
if (mount("/dev/sda1", "/root/test1/", "ext3", 0, 0) < 0) {
perror("cannot mount\n");
exit(1);
}
fd = open("/root/test1/testfile", O_RDWR);
if (fd < 0) {
perror("cannot open file\n");
exit(1);
}
filesize = LEN * LOOP;
for (i = 0; i < 300000; i++){
offset = (random() % filesize) & (~(LEN - 1));
pwrite(fd, buf, LEN, offset);
}
printf("start test\n");
time(&t1);
for (i = 0; i < 100000; i++){
offset = (random() % filesize) & (~(LEN - 1));
pread(fd, buf, LEN, offset);
}
time(&t2);
printf("%ld sec\n", t2-t1);
close(fd);
if (umount("/root/test1/") < 0) {
perror("cannot umount\n");
exit(1);
}
}
Signed-off-by: Hisashi Hifumi <hifumi.hisashi@oss.ntt.co.jp>
Cc: Nick Piggin <nickpiggin@yahoo.com.au>
Cc: Christoph Hellwig <hch@infradead.org>
Cc: Jan Kara <jack@ucw.cz>
Cc: <linux-ext4@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2008-07-28 22:46:36 +00:00
|
|
|
.is_partially_uptodate = block_is_partially_uptodate,
|
2009-09-16 09:50:16 +00:00
|
|
|
.error_remove_page = generic_error_remove_page,
|
2005-04-16 22:20:36 +00:00
|
|
|
};
|
|
|
|
|
2006-06-28 11:26:44 +00:00
|
|
|
const struct address_space_operations ext2_nobh_aops = {
|
2021-06-29 02:36:12 +00:00
|
|
|
.set_page_dirty = __set_page_dirty_buffers,
|
2005-04-16 22:20:36 +00:00
|
|
|
.readpage = ext2_readpage,
|
fs: convert mpage_readpages to mpage_readahead
Implement the new readahead aop and convert all callers (block_dev,
exfat, ext2, fat, gfs2, hpfs, isofs, jfs, nilfs2, ocfs2, omfs, qnx6,
reiserfs & udf).
The callers are all trivial except for GFS2 & OCFS2.
Signed-off-by: Matthew Wilcox (Oracle) <willy@infradead.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Reviewed-by: Junxiao Bi <junxiao.bi@oracle.com> # ocfs2
Reviewed-by: Joseph Qi <joseph.qi@linux.alibaba.com> # ocfs2
Reviewed-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: John Hubbard <jhubbard@nvidia.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: William Kucharski <william.kucharski@oracle.com>
Cc: Chao Yu <yuchao0@huawei.com>
Cc: Cong Wang <xiyou.wangcong@gmail.com>
Cc: Darrick J. Wong <darrick.wong@oracle.com>
Cc: Eric Biggers <ebiggers@google.com>
Cc: Gao Xiang <gaoxiang25@huawei.com>
Cc: Jaegeuk Kim <jaegeuk@kernel.org>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Zi Yan <ziy@nvidia.com>
Cc: Johannes Thumshirn <johannes.thumshirn@wdc.com>
Cc: Miklos Szeredi <mszeredi@redhat.com>
Link: http://lkml.kernel.org/r/20200414150233.24495-17-willy@infradead.org
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2020-06-02 04:47:02 +00:00
|
|
|
.readahead = ext2_readahead,
|
2005-04-16 22:20:36 +00:00
|
|
|
.writepage = ext2_nobh_writepage,
|
2007-10-16 08:25:25 +00:00
|
|
|
.write_begin = ext2_nobh_write_begin,
|
|
|
|
.write_end = nobh_write_end,
|
2005-04-16 22:20:36 +00:00
|
|
|
.bmap = ext2_bmap,
|
|
|
|
.direct_IO = ext2_direct_IO,
|
|
|
|
.writepages = ext2_writepages,
|
2006-02-01 11:05:41 +00:00
|
|
|
.migratepage = buffer_migrate_page,
|
2009-09-16 09:50:16 +00:00
|
|
|
.error_remove_page = generic_error_remove_page,
|
2005-04-16 22:20:36 +00:00
|
|
|
};
|
|
|
|
|
2017-12-21 20:25:11 +00:00
|
|
|
static const struct address_space_operations ext2_dax_aops = {
|
|
|
|
.writepages = ext2_dax_writepages,
|
|
|
|
.direct_IO = noop_direct_IO,
|
2021-06-29 02:36:27 +00:00
|
|
|
.set_page_dirty = __set_page_dirty_no_writeback,
|
2017-12-21 20:25:11 +00:00
|
|
|
.invalidatepage = noop_invalidatepage,
|
|
|
|
};
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
/*
|
|
|
|
* Probably it should be a library function... search for first non-zero word
|
|
|
|
* or memcmp with zero_page, whatever is better for particular architecture.
|
|
|
|
* Linus?
|
|
|
|
*/
|
|
|
|
static inline int all_zeroes(__le32 *p, __le32 *q)
|
|
|
|
{
|
|
|
|
while (p < q)
|
|
|
|
if (*p++)
|
|
|
|
return 0;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ext2_find_shared - find the indirect blocks for partial truncation.
|
|
|
|
* @inode: inode in question
|
|
|
|
* @depth: depth of the affected branch
|
|
|
|
* @offsets: offsets of pointers in that branch (see ext2_block_to_path)
|
|
|
|
* @chain: place to store the pointers to partial indirect blocks
|
|
|
|
* @top: place to the (detached) top of branch
|
|
|
|
*
|
|
|
|
* This is a helper function used by ext2_truncate().
|
|
|
|
*
|
|
|
|
* When we do truncate() we may have to clean the ends of several indirect
|
|
|
|
* blocks but leave the blocks themselves alive. Block is partially
|
2011-03-31 01:57:33 +00:00
|
|
|
* truncated if some data below the new i_size is referred from it (and
|
2005-04-16 22:20:36 +00:00
|
|
|
* it is on the path to the first completely truncated data block, indeed).
|
|
|
|
* We have to free the top of that path along with everything to the right
|
|
|
|
* of the path. Since no allocation past the truncation point is possible
|
|
|
|
* until ext2_truncate() finishes, we may safely do the latter, but top
|
|
|
|
* of branch may require special attention - pageout below the truncation
|
|
|
|
* point might try to populate it.
|
|
|
|
*
|
|
|
|
* We atomically detach the top of branch from the tree, store the block
|
|
|
|
* number of its root in *@top, pointers to buffer_heads of partially
|
|
|
|
* truncated blocks - in @chain[].bh and pointers to their last elements
|
|
|
|
* that should not be removed - in @chain[].p. Return value is the pointer
|
|
|
|
* to last filled element of @chain.
|
|
|
|
*
|
|
|
|
* The work left to caller to do the actual freeing of subtrees:
|
|
|
|
* a) free the subtree starting from *@top
|
|
|
|
* b) free the subtrees whose roots are stored in
|
|
|
|
* (@chain[i].p+1 .. end of @chain[i].bh->b_data)
|
|
|
|
* c) free the subtrees growing from the inode past the @chain[0].p
|
|
|
|
* (no partially truncated stuff there).
|
|
|
|
*/
|
|
|
|
|
|
|
|
static Indirect *ext2_find_shared(struct inode *inode,
|
|
|
|
int depth,
|
|
|
|
int offsets[4],
|
|
|
|
Indirect chain[4],
|
|
|
|
__le32 *top)
|
|
|
|
{
|
|
|
|
Indirect *partial, *p;
|
|
|
|
int k, err;
|
|
|
|
|
|
|
|
*top = 0;
|
|
|
|
for (k = depth; k > 1 && !offsets[k-1]; k--)
|
|
|
|
;
|
|
|
|
partial = ext2_get_branch(inode, k, offsets, chain, &err);
|
|
|
|
if (!partial)
|
|
|
|
partial = chain + k-1;
|
|
|
|
/*
|
|
|
|
* If the branch acquired continuation since we've looked at it -
|
|
|
|
* fine, it should all survive and (new) top doesn't belong to us.
|
|
|
|
*/
|
|
|
|
write_lock(&EXT2_I(inode)->i_meta_lock);
|
|
|
|
if (!partial->key && *partial->p) {
|
|
|
|
write_unlock(&EXT2_I(inode)->i_meta_lock);
|
|
|
|
goto no_top;
|
|
|
|
}
|
|
|
|
for (p=partial; p>chain && all_zeroes((__le32*)p->bh->b_data,p->p); p--)
|
|
|
|
;
|
|
|
|
/*
|
|
|
|
* OK, we've found the last block that must survive. The rest of our
|
|
|
|
* branch should be detached before unlocking. However, if that rest
|
|
|
|
* of branch is all ours and does not grow immediately from the inode
|
|
|
|
* it's easier to cheat and just decrement partial->p.
|
|
|
|
*/
|
|
|
|
if (p == chain + k - 1 && p > chain) {
|
|
|
|
p->p--;
|
|
|
|
} else {
|
|
|
|
*top = *p->p;
|
|
|
|
*p->p = 0;
|
|
|
|
}
|
|
|
|
write_unlock(&EXT2_I(inode)->i_meta_lock);
|
|
|
|
|
|
|
|
while(partial > p)
|
|
|
|
{
|
|
|
|
brelse(partial->bh);
|
|
|
|
partial--;
|
|
|
|
}
|
|
|
|
no_top:
|
|
|
|
return partial;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ext2_free_data - free a list of data blocks
|
|
|
|
* @inode: inode we are dealing with
|
|
|
|
* @p: array of block numbers
|
|
|
|
* @q: points immediately past the end of array
|
|
|
|
*
|
2011-03-31 01:57:33 +00:00
|
|
|
* We are freeing all blocks referred from that array (numbers are
|
2005-04-16 22:20:36 +00:00
|
|
|
* stored as little-endian 32-bit) and updating @inode->i_blocks
|
|
|
|
* appropriately.
|
|
|
|
*/
|
|
|
|
static inline void ext2_free_data(struct inode *inode, __le32 *p, __le32 *q)
|
|
|
|
{
|
|
|
|
unsigned long block_to_free = 0, count = 0;
|
|
|
|
unsigned long nr;
|
|
|
|
|
|
|
|
for ( ; p < q ; p++) {
|
|
|
|
nr = le32_to_cpu(*p);
|
|
|
|
if (nr) {
|
|
|
|
*p = 0;
|
|
|
|
/* accumulate blocks to free if they're contiguous */
|
|
|
|
if (count == 0)
|
|
|
|
goto free_this;
|
|
|
|
else if (block_to_free == nr - count)
|
|
|
|
count++;
|
|
|
|
else {
|
|
|
|
ext2_free_blocks (inode, block_to_free, count);
|
2010-07-21 21:19:42 +00:00
|
|
|
mark_inode_dirty(inode);
|
2005-04-16 22:20:36 +00:00
|
|
|
free_this:
|
|
|
|
block_to_free = nr;
|
|
|
|
count = 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (count > 0) {
|
|
|
|
ext2_free_blocks (inode, block_to_free, count);
|
2010-07-21 21:19:42 +00:00
|
|
|
mark_inode_dirty(inode);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ext2_free_branches - free an array of branches
|
|
|
|
* @inode: inode we are dealing with
|
|
|
|
* @p: array of block numbers
|
|
|
|
* @q: pointer immediately past the end of array
|
|
|
|
* @depth: depth of the branches to free
|
|
|
|
*
|
2011-03-31 01:57:33 +00:00
|
|
|
* We are freeing all blocks referred from these branches (numbers are
|
2005-04-16 22:20:36 +00:00
|
|
|
* stored as little-endian 32-bit) and updating @inode->i_blocks
|
|
|
|
* appropriately.
|
|
|
|
*/
|
|
|
|
static void ext2_free_branches(struct inode *inode, __le32 *p, __le32 *q, int depth)
|
|
|
|
{
|
|
|
|
struct buffer_head * bh;
|
|
|
|
unsigned long nr;
|
|
|
|
|
|
|
|
if (depth--) {
|
|
|
|
int addr_per_block = EXT2_ADDR_PER_BLOCK(inode->i_sb);
|
|
|
|
for ( ; p < q ; p++) {
|
|
|
|
nr = le32_to_cpu(*p);
|
|
|
|
if (!nr)
|
|
|
|
continue;
|
|
|
|
*p = 0;
|
|
|
|
bh = sb_bread(inode->i_sb, nr);
|
|
|
|
/*
|
|
|
|
* A read failure? Report error and clear slot
|
|
|
|
* (should be rare).
|
|
|
|
*/
|
|
|
|
if (!bh) {
|
|
|
|
ext2_error(inode->i_sb, "ext2_free_branches",
|
|
|
|
"Read failure, inode=%ld, block=%ld",
|
|
|
|
inode->i_ino, nr);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
ext2_free_branches(inode,
|
|
|
|
(__le32*)bh->b_data,
|
|
|
|
(__le32*)bh->b_data + addr_per_block,
|
|
|
|
depth);
|
|
|
|
bforget(bh);
|
|
|
|
ext2_free_blocks(inode, nr, 1);
|
|
|
|
mark_inode_dirty(inode);
|
|
|
|
}
|
|
|
|
} else
|
|
|
|
ext2_free_data(inode, p, q);
|
|
|
|
}
|
|
|
|
|
2015-10-13 22:25:37 +00:00
|
|
|
/* dax_sem must be held when calling this function */
|
2010-05-26 15:05:37 +00:00
|
|
|
static void __ext2_truncate_blocks(struct inode *inode, loff_t offset)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
__le32 *i_data = EXT2_I(inode)->i_data;
|
2007-10-17 06:30:46 +00:00
|
|
|
struct ext2_inode_info *ei = EXT2_I(inode);
|
2005-04-16 22:20:36 +00:00
|
|
|
int addr_per_block = EXT2_ADDR_PER_BLOCK(inode->i_sb);
|
|
|
|
int offsets[4];
|
|
|
|
Indirect chain[4];
|
|
|
|
Indirect *partial;
|
|
|
|
__le32 nr = 0;
|
|
|
|
int n;
|
|
|
|
long iblock;
|
|
|
|
unsigned blocksize;
|
|
|
|
blocksize = inode->i_sb->s_blocksize;
|
2010-05-26 15:05:37 +00:00
|
|
|
iblock = (offset + blocksize-1) >> EXT2_BLOCK_SIZE_BITS(inode->i_sb);
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2015-10-13 22:25:37 +00:00
|
|
|
#ifdef CONFIG_FS_DAX
|
|
|
|
WARN_ON(!rwsem_is_locked(&ei->dax_sem));
|
|
|
|
#endif
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
n = ext2_block_to_path(inode, iblock, offsets, NULL);
|
|
|
|
if (n == 0)
|
|
|
|
return;
|
|
|
|
|
2007-10-17 06:30:46 +00:00
|
|
|
/*
|
|
|
|
* From here we block out all ext2_get_block() callers who want to
|
|
|
|
* modify the block allocation tree.
|
|
|
|
*/
|
|
|
|
mutex_lock(&ei->truncate_mutex);
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
if (n == 1) {
|
|
|
|
ext2_free_data(inode, i_data+offsets[0],
|
|
|
|
i_data + EXT2_NDIR_BLOCKS);
|
|
|
|
goto do_indirects;
|
|
|
|
}
|
|
|
|
|
|
|
|
partial = ext2_find_shared(inode, n, offsets, chain, &nr);
|
|
|
|
/* Kill the top of shared branch (already detached) */
|
|
|
|
if (nr) {
|
|
|
|
if (partial == chain)
|
|
|
|
mark_inode_dirty(inode);
|
|
|
|
else
|
|
|
|
mark_buffer_dirty_inode(partial->bh, inode);
|
|
|
|
ext2_free_branches(inode, &nr, &nr+1, (chain+n-1) - partial);
|
|
|
|
}
|
|
|
|
/* Clear the ends of indirect blocks on the shared branch */
|
|
|
|
while (partial > chain) {
|
|
|
|
ext2_free_branches(inode,
|
|
|
|
partial->p + 1,
|
|
|
|
(__le32*)partial->bh->b_data+addr_per_block,
|
|
|
|
(chain+n-1) - partial);
|
|
|
|
mark_buffer_dirty_inode(partial->bh, inode);
|
|
|
|
brelse (partial->bh);
|
|
|
|
partial--;
|
|
|
|
}
|
|
|
|
do_indirects:
|
|
|
|
/* Kill the remaining (whole) subtrees */
|
|
|
|
switch (offsets[0]) {
|
|
|
|
default:
|
|
|
|
nr = i_data[EXT2_IND_BLOCK];
|
|
|
|
if (nr) {
|
|
|
|
i_data[EXT2_IND_BLOCK] = 0;
|
|
|
|
mark_inode_dirty(inode);
|
|
|
|
ext2_free_branches(inode, &nr, &nr+1, 1);
|
|
|
|
}
|
2020-08-23 22:36:59 +00:00
|
|
|
fallthrough;
|
2005-04-16 22:20:36 +00:00
|
|
|
case EXT2_IND_BLOCK:
|
|
|
|
nr = i_data[EXT2_DIND_BLOCK];
|
|
|
|
if (nr) {
|
|
|
|
i_data[EXT2_DIND_BLOCK] = 0;
|
|
|
|
mark_inode_dirty(inode);
|
|
|
|
ext2_free_branches(inode, &nr, &nr+1, 2);
|
|
|
|
}
|
2020-08-23 22:36:59 +00:00
|
|
|
fallthrough;
|
2005-04-16 22:20:36 +00:00
|
|
|
case EXT2_DIND_BLOCK:
|
|
|
|
nr = i_data[EXT2_TIND_BLOCK];
|
|
|
|
if (nr) {
|
|
|
|
i_data[EXT2_TIND_BLOCK] = 0;
|
|
|
|
mark_inode_dirty(inode);
|
|
|
|
ext2_free_branches(inode, &nr, &nr+1, 3);
|
|
|
|
}
|
2020-11-20 18:28:25 +00:00
|
|
|
break;
|
2005-04-16 22:20:36 +00:00
|
|
|
case EXT2_TIND_BLOCK:
|
|
|
|
;
|
|
|
|
}
|
2007-10-17 06:30:46 +00:00
|
|
|
|
|
|
|
ext2_discard_reservation(inode);
|
|
|
|
|
|
|
|
mutex_unlock(&ei->truncate_mutex);
|
2010-05-26 15:05:37 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static void ext2_truncate_blocks(struct inode *inode, loff_t offset)
|
|
|
|
{
|
|
|
|
if (!(S_ISREG(inode->i_mode) || S_ISDIR(inode->i_mode) ||
|
|
|
|
S_ISLNK(inode->i_mode)))
|
|
|
|
return;
|
|
|
|
if (ext2_inode_is_fast_symlink(inode))
|
|
|
|
return;
|
2015-10-13 22:25:37 +00:00
|
|
|
|
|
|
|
dax_sem_down_write(EXT2_I(inode));
|
2010-05-26 15:05:37 +00:00
|
|
|
__ext2_truncate_blocks(inode, offset);
|
2015-10-13 22:25:37 +00:00
|
|
|
dax_sem_up_write(EXT2_I(inode));
|
2010-05-26 15:05:37 +00:00
|
|
|
}
|
|
|
|
|
2010-06-04 09:30:04 +00:00
|
|
|
static int ext2_setsize(struct inode *inode, loff_t newsize)
|
2010-05-26 15:05:37 +00:00
|
|
|
{
|
|
|
|
int error;
|
|
|
|
|
|
|
|
if (!(S_ISREG(inode->i_mode) || S_ISDIR(inode->i_mode) ||
|
|
|
|
S_ISLNK(inode->i_mode)))
|
|
|
|
return -EINVAL;
|
|
|
|
if (ext2_inode_is_fast_symlink(inode))
|
|
|
|
return -EINVAL;
|
|
|
|
if (IS_APPEND(inode) || IS_IMMUTABLE(inode))
|
|
|
|
return -EPERM;
|
|
|
|
|
2011-06-24 18:29:45 +00:00
|
|
|
inode_dio_wait(inode);
|
|
|
|
|
2016-11-21 01:47:07 +00:00
|
|
|
if (IS_DAX(inode)) {
|
|
|
|
error = iomap_zero_range(inode, newsize,
|
|
|
|
PAGE_ALIGN(newsize) - newsize, NULL,
|
|
|
|
&ext2_iomap_ops);
|
|
|
|
} else if (test_opt(inode->i_sb, NOBH))
|
2010-05-26 15:05:37 +00:00
|
|
|
error = nobh_truncate_page(inode->i_mapping,
|
|
|
|
newsize, ext2_get_block);
|
|
|
|
else
|
|
|
|
error = block_truncate_page(inode->i_mapping,
|
|
|
|
newsize, ext2_get_block);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
|
2015-10-13 22:25:37 +00:00
|
|
|
dax_sem_down_write(EXT2_I(inode));
|
2010-06-04 09:30:04 +00:00
|
|
|
truncate_setsize(inode, newsize);
|
2010-05-26 15:05:37 +00:00
|
|
|
__ext2_truncate_blocks(inode, newsize);
|
2015-10-13 22:25:37 +00:00
|
|
|
dax_sem_up_write(EXT2_I(inode));
|
2010-05-26 15:05:37 +00:00
|
|
|
|
2016-09-14 14:48:05 +00:00
|
|
|
inode->i_mtime = inode->i_ctime = current_time(inode);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (inode_needs_sync(inode)) {
|
|
|
|
sync_mapping_buffers(inode->i_mapping);
|
2010-10-06 08:48:20 +00:00
|
|
|
sync_inode_metadata(inode, 1);
|
2005-04-16 22:20:36 +00:00
|
|
|
} else {
|
|
|
|
mark_inode_dirty(inode);
|
|
|
|
}
|
2010-05-26 15:05:37 +00:00
|
|
|
|
|
|
|
return 0;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static struct ext2_inode *ext2_get_inode(struct super_block *sb, ino_t ino,
|
|
|
|
struct buffer_head **p)
|
|
|
|
{
|
|
|
|
struct buffer_head * bh;
|
|
|
|
unsigned long block_group;
|
|
|
|
unsigned long block;
|
|
|
|
unsigned long offset;
|
|
|
|
struct ext2_group_desc * gdp;
|
|
|
|
|
|
|
|
*p = NULL;
|
|
|
|
if ((ino != EXT2_ROOT_INO && ino < EXT2_FIRST_INO(sb)) ||
|
|
|
|
ino > le32_to_cpu(EXT2_SB(sb)->s_es->s_inodes_count))
|
|
|
|
goto Einval;
|
|
|
|
|
|
|
|
block_group = (ino - 1) / EXT2_INODES_PER_GROUP(sb);
|
2007-10-17 06:26:30 +00:00
|
|
|
gdp = ext2_get_group_desc(sb, block_group, NULL);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (!gdp)
|
|
|
|
goto Egdp;
|
|
|
|
/*
|
|
|
|
* Figure out the offset within the block group inode table
|
|
|
|
*/
|
|
|
|
offset = ((ino - 1) % EXT2_INODES_PER_GROUP(sb)) * EXT2_INODE_SIZE(sb);
|
|
|
|
block = le32_to_cpu(gdp->bg_inode_table) +
|
|
|
|
(offset >> EXT2_BLOCK_SIZE_BITS(sb));
|
|
|
|
if (!(bh = sb_bread(sb, block)))
|
|
|
|
goto Eio;
|
|
|
|
|
|
|
|
*p = bh;
|
|
|
|
offset &= (EXT2_BLOCK_SIZE(sb) - 1);
|
|
|
|
return (struct ext2_inode *) (bh->b_data + offset);
|
|
|
|
|
|
|
|
Einval:
|
|
|
|
ext2_error(sb, "ext2_get_inode", "bad inode number: %lu",
|
|
|
|
(unsigned long) ino);
|
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
Eio:
|
|
|
|
ext2_error(sb, "ext2_get_inode",
|
|
|
|
"unable to read inode block - inode=%lu, block=%lu",
|
|
|
|
(unsigned long) ino, block);
|
|
|
|
Egdp:
|
|
|
|
return ERR_PTR(-EIO);
|
|
|
|
}
|
|
|
|
|
|
|
|
void ext2_set_inode_flags(struct inode *inode)
|
|
|
|
{
|
|
|
|
unsigned int flags = EXT2_I(inode)->i_flags;
|
|
|
|
|
2015-02-16 23:58:53 +00:00
|
|
|
inode->i_flags &= ~(S_SYNC | S_APPEND | S_IMMUTABLE | S_NOATIME |
|
|
|
|
S_DIRSYNC | S_DAX);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (flags & EXT2_SYNC_FL)
|
|
|
|
inode->i_flags |= S_SYNC;
|
|
|
|
if (flags & EXT2_APPEND_FL)
|
|
|
|
inode->i_flags |= S_APPEND;
|
|
|
|
if (flags & EXT2_IMMUTABLE_FL)
|
|
|
|
inode->i_flags |= S_IMMUTABLE;
|
|
|
|
if (flags & EXT2_NOATIME_FL)
|
|
|
|
inode->i_flags |= S_NOATIME;
|
|
|
|
if (flags & EXT2_DIRSYNC_FL)
|
|
|
|
inode->i_flags |= S_DIRSYNC;
|
2016-02-26 23:19:46 +00:00
|
|
|
if (test_opt(inode->i_sb, DAX) && S_ISREG(inode->i_mode))
|
2015-02-16 23:58:53 +00:00
|
|
|
inode->i_flags |= S_DAX;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2017-12-21 20:25:11 +00:00
|
|
|
void ext2_set_file_ops(struct inode *inode)
|
|
|
|
{
|
|
|
|
inode->i_op = &ext2_file_inode_operations;
|
|
|
|
inode->i_fop = &ext2_file_operations;
|
|
|
|
if (IS_DAX(inode))
|
|
|
|
inode->i_mapping->a_ops = &ext2_dax_aops;
|
|
|
|
else if (test_opt(inode->i_sb, NOBH))
|
|
|
|
inode->i_mapping->a_ops = &ext2_nobh_aops;
|
|
|
|
else
|
|
|
|
inode->i_mapping->a_ops = &ext2_aops;
|
|
|
|
}
|
|
|
|
|
2008-02-07 08:15:35 +00:00
|
|
|
struct inode *ext2_iget (struct super_block *sb, unsigned long ino)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2008-02-07 08:15:35 +00:00
|
|
|
struct ext2_inode_info *ei;
|
2019-06-19 16:29:45 +00:00
|
|
|
struct buffer_head * bh = NULL;
|
2008-02-07 08:15:35 +00:00
|
|
|
struct ext2_inode *raw_inode;
|
|
|
|
struct inode *inode;
|
|
|
|
long ret = -EIO;
|
2005-04-16 22:20:36 +00:00
|
|
|
int n;
|
2012-02-07 23:39:12 +00:00
|
|
|
uid_t i_uid;
|
|
|
|
gid_t i_gid;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
2008-02-07 08:15:35 +00:00
|
|
|
inode = iget_locked(sb, ino);
|
|
|
|
if (!inode)
|
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
if (!(inode->i_state & I_NEW))
|
|
|
|
return inode;
|
|
|
|
|
|
|
|
ei = EXT2_I(inode);
|
2007-10-17 06:30:46 +00:00
|
|
|
ei->i_block_alloc_info = NULL;
|
|
|
|
|
2008-02-07 08:15:35 +00:00
|
|
|
raw_inode = ext2_get_inode(inode->i_sb, ino, &bh);
|
|
|
|
if (IS_ERR(raw_inode)) {
|
|
|
|
ret = PTR_ERR(raw_inode);
|
2005-04-16 22:20:36 +00:00
|
|
|
goto bad_inode;
|
2008-02-07 08:15:35 +00:00
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
inode->i_mode = le16_to_cpu(raw_inode->i_mode);
|
2012-02-07 23:39:12 +00:00
|
|
|
i_uid = (uid_t)le16_to_cpu(raw_inode->i_uid_low);
|
|
|
|
i_gid = (gid_t)le16_to_cpu(raw_inode->i_gid_low);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (!(test_opt (inode->i_sb, NO_UID32))) {
|
2012-02-07 23:39:12 +00:00
|
|
|
i_uid |= le16_to_cpu(raw_inode->i_uid_high) << 16;
|
|
|
|
i_gid |= le16_to_cpu(raw_inode->i_gid_high) << 16;
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
2012-02-07 23:39:12 +00:00
|
|
|
i_uid_write(inode, i_uid);
|
|
|
|
i_gid_write(inode, i_gid);
|
2011-10-28 12:13:29 +00:00
|
|
|
set_nlink(inode, le16_to_cpu(raw_inode->i_links_count));
|
2005-04-16 22:20:36 +00:00
|
|
|
inode->i_size = le32_to_cpu(raw_inode->i_size);
|
ext2/3/4: fix file date underflow on ext2 3 filesystems on 64 bit systems
Taken from http://bugzilla.kernel.org/show_bug.cgi?id=5079
signed long ranges from -2.147.483.648 to 2.147.483.647 on x86 32bit
10000011110110100100111110111101 .. -2,082,844,739
10000011110110100100111110111101 .. 2,212,122,557 <- this currently gets
stored on the disk but when converting it to a 64bit signed long value it loses
its sign and becomes positive.
Cc: Andreas Dilger <adilger@dilger.ca>
Cc: <linux-ext4@vger.kernel.org>
Andreas says:
This patch is now treating timestamps with the high bit set as negative
times (before Jan 1, 1970). This means we lose 1/2 of the possible range
of timestamps (lopping off 68 years before unix timestamp overflow -
now only 30 years away :-) to handle the extremely rare case of setting
timestamps into the distant past.
If we are only interested in fixing the underflow case, we could just
limit the values to 0 instead of storing negative values. At worst this
will skew the timestamp by a few hours for timezones in the far east
(files would still show Jan 1, 1970 in "ls -l" output).
That said, it seems 32-bit systems (mine at least) allow files to be set
into the past (01/01/1907 works fine) so it seems this patch is bringing
the x86_64 behaviour into sync with other kernels.
On the plus side, we have a patch that is ready to add nanosecond timestamps
to ext3 and as an added bonus adds 2 high bits to the on-disk timestamp so
this extends the maximum date to 2242.
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2007-05-08 07:23:39 +00:00
|
|
|
inode->i_atime.tv_sec = (signed)le32_to_cpu(raw_inode->i_atime);
|
|
|
|
inode->i_ctime.tv_sec = (signed)le32_to_cpu(raw_inode->i_ctime);
|
|
|
|
inode->i_mtime.tv_sec = (signed)le32_to_cpu(raw_inode->i_mtime);
|
2005-04-16 22:20:36 +00:00
|
|
|
inode->i_atime.tv_nsec = inode->i_mtime.tv_nsec = inode->i_ctime.tv_nsec = 0;
|
|
|
|
ei->i_dtime = le32_to_cpu(raw_inode->i_dtime);
|
|
|
|
/* We now have enough fields to check if the inode was active or not.
|
|
|
|
* This is needed because nfsd might try to access dead inodes
|
|
|
|
* the test is that same one that e2fsck uses
|
|
|
|
* NeilBrown 1999oct15
|
|
|
|
*/
|
|
|
|
if (inode->i_nlink == 0 && (inode->i_mode == 0 || ei->i_dtime)) {
|
|
|
|
/* this inode is deleted */
|
2008-02-07 08:15:35 +00:00
|
|
|
ret = -ESTALE;
|
2005-04-16 22:20:36 +00:00
|
|
|
goto bad_inode;
|
|
|
|
}
|
|
|
|
inode->i_blocks = le32_to_cpu(raw_inode->i_blocks);
|
|
|
|
ei->i_flags = le32_to_cpu(raw_inode->i_flags);
|
2018-09-14 14:51:14 +00:00
|
|
|
ext2_set_inode_flags(inode);
|
2005-04-16 22:20:36 +00:00
|
|
|
ei->i_faddr = le32_to_cpu(raw_inode->i_faddr);
|
|
|
|
ei->i_frag_no = raw_inode->i_frag;
|
|
|
|
ei->i_frag_size = raw_inode->i_fsize;
|
|
|
|
ei->i_file_acl = le32_to_cpu(raw_inode->i_file_acl);
|
|
|
|
ei->i_dir_acl = 0;
|
ext2: fix filesystem deadlock while reading corrupted xattr block
This bug can be reproducible with fsfuzzer, although, I couldn't reproduce it
100% of my tries, it is quite easily reproducible.
During the deletion of an inode, ext2_xattr_delete_inode() does not check if the
block pointed by EXT2_I(inode)->i_file_acl is a valid data block, this might
lead to a deadlock, when i_file_acl == 1, and the filesystem block size is 1024.
In that situation, ext2_xattr_delete_inode, will load the superblock's buffer
head (instead of a valid i_file_acl block), and then lock that buffer head,
which, ext2_sync_super will also try to lock, making the filesystem deadlock in
the following stack trace:
root 17180 0.0 0.0 113660 660 pts/0 D+ 07:08 0:00 rmdir
/media/test/dir1
[<ffffffff8125da9f>] __sync_dirty_buffer+0xaf/0x100
[<ffffffff8125db03>] sync_dirty_buffer+0x13/0x20
[<ffffffffa03f0d57>] ext2_sync_super+0xb7/0xc0 [ext2]
[<ffffffffa03f10b9>] ext2_error+0x119/0x130 [ext2]
[<ffffffffa03e9d93>] ext2_free_blocks+0x83/0x350 [ext2]
[<ffffffffa03f3d03>] ext2_xattr_delete_inode+0x173/0x190 [ext2]
[<ffffffffa03ee9e9>] ext2_evict_inode+0xc9/0x130 [ext2]
[<ffffffff8123fd23>] evict+0xb3/0x180
[<ffffffff81240008>] iput+0x1b8/0x240
[<ffffffff8123c4ac>] d_delete+0x11c/0x150
[<ffffffff8122fa7e>] vfs_rmdir+0xfe/0x120
[<ffffffff812340ee>] do_rmdir+0x17e/0x1f0
[<ffffffff81234dd6>] SyS_rmdir+0x16/0x20
[<ffffffff81838cf2>] entry_SYSCALL_64_fastpath+0x1a/0xa4
[<ffffffffffffffff>] 0xffffffffffffffff
Fix this by using the same approach ext4 uses to test data blocks validity,
implementing ext2_data_block_valid.
An another possibility when the superblock is very corrupted, is that i_file_acl
is 1, block_count is 1 and first_data_block is 0. For such situations, we might
have i_file_acl pointing to a 'valid' block, but still step over the superblock.
The approach I used was to also test if the superblock is not in the range
described by ext2_data_block_valid() arguments
Signed-off-by: Carlos Maiolino <cmaiolino@redhat.com>
Signed-off-by: Theodore Ts'o <tytso@mit.edu>
2016-07-06 02:02:41 +00:00
|
|
|
|
|
|
|
if (ei->i_file_acl &&
|
|
|
|
!ext2_data_block_valid(EXT2_SB(sb), ei->i_file_acl, 1)) {
|
|
|
|
ext2_error(sb, "ext2_iget", "bad extended attribute block %u",
|
|
|
|
ei->i_file_acl);
|
|
|
|
ret = -EFSCORRUPTED;
|
|
|
|
goto bad_inode;
|
|
|
|
}
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
if (S_ISREG(inode->i_mode))
|
|
|
|
inode->i_size |= ((__u64)le32_to_cpu(raw_inode->i_size_high)) << 32;
|
|
|
|
else
|
|
|
|
ei->i_dir_acl = le32_to_cpu(raw_inode->i_dir_acl);
|
2016-12-06 23:59:01 +00:00
|
|
|
if (i_size_read(inode) < 0) {
|
|
|
|
ret = -EFSCORRUPTED;
|
|
|
|
goto bad_inode;
|
|
|
|
}
|
2005-04-16 22:20:36 +00:00
|
|
|
ei->i_dtime = 0;
|
|
|
|
inode->i_generation = le32_to_cpu(raw_inode->i_generation);
|
|
|
|
ei->i_state = 0;
|
|
|
|
ei->i_block_group = (ino - 1) / EXT2_INODES_PER_GROUP(inode->i_sb);
|
|
|
|
ei->i_dir_start_lookup = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* NOTE! The in-memory inode i_data array is in little-endian order
|
|
|
|
* even on big-endian machines: we do NOT byteswap the block numbers!
|
|
|
|
*/
|
|
|
|
for (n = 0; n < EXT2_N_BLOCKS; n++)
|
|
|
|
ei->i_data[n] = raw_inode->i_block[n];
|
|
|
|
|
|
|
|
if (S_ISREG(inode->i_mode)) {
|
2017-12-21 20:25:11 +00:00
|
|
|
ext2_set_file_ops(inode);
|
2005-04-16 22:20:36 +00:00
|
|
|
} else if (S_ISDIR(inode->i_mode)) {
|
|
|
|
inode->i_op = &ext2_dir_inode_operations;
|
|
|
|
inode->i_fop = &ext2_dir_operations;
|
|
|
|
if (test_opt(inode->i_sb, NOBH))
|
|
|
|
inode->i_mapping->a_ops = &ext2_nobh_aops;
|
|
|
|
else
|
|
|
|
inode->i_mapping->a_ops = &ext2_aops;
|
|
|
|
} else if (S_ISLNK(inode->i_mode)) {
|
2008-12-19 20:47:13 +00:00
|
|
|
if (ext2_inode_is_fast_symlink(inode)) {
|
2015-05-02 14:02:46 +00:00
|
|
|
inode->i_link = (char *)ei->i_data;
|
2005-04-16 22:20:36 +00:00
|
|
|
inode->i_op = &ext2_fast_symlink_inode_operations;
|
2008-12-19 20:47:13 +00:00
|
|
|
nd_terminate_link(ei->i_data, inode->i_size,
|
|
|
|
sizeof(ei->i_data) - 1);
|
|
|
|
} else {
|
2005-04-16 22:20:36 +00:00
|
|
|
inode->i_op = &ext2_symlink_inode_operations;
|
2015-11-17 06:07:57 +00:00
|
|
|
inode_nohighmem(inode);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (test_opt(inode->i_sb, NOBH))
|
|
|
|
inode->i_mapping->a_ops = &ext2_nobh_aops;
|
|
|
|
else
|
|
|
|
inode->i_mapping->a_ops = &ext2_aops;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
inode->i_op = &ext2_special_inode_operations;
|
|
|
|
if (raw_inode->i_block[0])
|
|
|
|
init_special_inode(inode, inode->i_mode,
|
|
|
|
old_decode_dev(le32_to_cpu(raw_inode->i_block[0])));
|
|
|
|
else
|
|
|
|
init_special_inode(inode, inode->i_mode,
|
|
|
|
new_decode_dev(le32_to_cpu(raw_inode->i_block[1])));
|
|
|
|
}
|
|
|
|
brelse (bh);
|
2008-02-07 08:15:35 +00:00
|
|
|
unlock_new_inode(inode);
|
|
|
|
return inode;
|
2005-04-16 22:20:36 +00:00
|
|
|
|
|
|
|
bad_inode:
|
2019-06-19 16:29:45 +00:00
|
|
|
brelse(bh);
|
2008-02-07 08:15:35 +00:00
|
|
|
iget_failed(inode);
|
|
|
|
return ERR_PTR(ret);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
|
2010-03-05 08:21:37 +00:00
|
|
|
static int __ext2_write_inode(struct inode *inode, int do_sync)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
|
|
|
struct ext2_inode_info *ei = EXT2_I(inode);
|
|
|
|
struct super_block *sb = inode->i_sb;
|
|
|
|
ino_t ino = inode->i_ino;
|
2012-02-07 23:39:12 +00:00
|
|
|
uid_t uid = i_uid_read(inode);
|
|
|
|
gid_t gid = i_gid_read(inode);
|
2005-04-16 22:20:36 +00:00
|
|
|
struct buffer_head * bh;
|
|
|
|
struct ext2_inode * raw_inode = ext2_get_inode(sb, ino, &bh);
|
|
|
|
int n;
|
|
|
|
int err = 0;
|
|
|
|
|
|
|
|
if (IS_ERR(raw_inode))
|
|
|
|
return -EIO;
|
|
|
|
|
|
|
|
/* For fields not not tracking in the in-memory inode,
|
|
|
|
* initialise them to zero for new inodes. */
|
|
|
|
if (ei->i_state & EXT2_STATE_NEW)
|
|
|
|
memset(raw_inode, 0, EXT2_SB(sb)->s_inode_size);
|
|
|
|
|
|
|
|
raw_inode->i_mode = cpu_to_le16(inode->i_mode);
|
|
|
|
if (!(test_opt(sb, NO_UID32))) {
|
|
|
|
raw_inode->i_uid_low = cpu_to_le16(low_16_bits(uid));
|
|
|
|
raw_inode->i_gid_low = cpu_to_le16(low_16_bits(gid));
|
|
|
|
/*
|
|
|
|
* Fix up interoperability with old kernels. Otherwise, old inodes get
|
|
|
|
* re-used with the upper 16 bits of the uid/gid intact
|
|
|
|
*/
|
|
|
|
if (!ei->i_dtime) {
|
|
|
|
raw_inode->i_uid_high = cpu_to_le16(high_16_bits(uid));
|
|
|
|
raw_inode->i_gid_high = cpu_to_le16(high_16_bits(gid));
|
|
|
|
} else {
|
|
|
|
raw_inode->i_uid_high = 0;
|
|
|
|
raw_inode->i_gid_high = 0;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
raw_inode->i_uid_low = cpu_to_le16(fs_high2lowuid(uid));
|
|
|
|
raw_inode->i_gid_low = cpu_to_le16(fs_high2lowgid(gid));
|
|
|
|
raw_inode->i_uid_high = 0;
|
|
|
|
raw_inode->i_gid_high = 0;
|
|
|
|
}
|
|
|
|
raw_inode->i_links_count = cpu_to_le16(inode->i_nlink);
|
|
|
|
raw_inode->i_size = cpu_to_le32(inode->i_size);
|
|
|
|
raw_inode->i_atime = cpu_to_le32(inode->i_atime.tv_sec);
|
|
|
|
raw_inode->i_ctime = cpu_to_le32(inode->i_ctime.tv_sec);
|
|
|
|
raw_inode->i_mtime = cpu_to_le32(inode->i_mtime.tv_sec);
|
|
|
|
|
|
|
|
raw_inode->i_blocks = cpu_to_le32(inode->i_blocks);
|
|
|
|
raw_inode->i_dtime = cpu_to_le32(ei->i_dtime);
|
|
|
|
raw_inode->i_flags = cpu_to_le32(ei->i_flags);
|
|
|
|
raw_inode->i_faddr = cpu_to_le32(ei->i_faddr);
|
|
|
|
raw_inode->i_frag = ei->i_frag_no;
|
|
|
|
raw_inode->i_fsize = ei->i_frag_size;
|
|
|
|
raw_inode->i_file_acl = cpu_to_le32(ei->i_file_acl);
|
|
|
|
if (!S_ISREG(inode->i_mode))
|
|
|
|
raw_inode->i_dir_acl = cpu_to_le32(ei->i_dir_acl);
|
|
|
|
else {
|
|
|
|
raw_inode->i_size_high = cpu_to_le32(inode->i_size >> 32);
|
|
|
|
if (inode->i_size > 0x7fffffffULL) {
|
|
|
|
if (!EXT2_HAS_RO_COMPAT_FEATURE(sb,
|
|
|
|
EXT2_FEATURE_RO_COMPAT_LARGE_FILE) ||
|
|
|
|
EXT2_SB(sb)->s_es->s_rev_level ==
|
|
|
|
cpu_to_le32(EXT2_GOOD_OLD_REV)) {
|
|
|
|
/* If this is the first large file
|
|
|
|
* created, add a flag to the superblock.
|
|
|
|
*/
|
2010-04-14 12:38:38 +00:00
|
|
|
spin_lock(&EXT2_SB(sb)->s_lock);
|
2005-04-16 22:20:36 +00:00
|
|
|
ext2_update_dynamic_rev(sb);
|
|
|
|
EXT2_SET_RO_COMPAT_FEATURE(sb,
|
|
|
|
EXT2_FEATURE_RO_COMPAT_LARGE_FILE);
|
2010-04-14 12:38:38 +00:00
|
|
|
spin_unlock(&EXT2_SB(sb)->s_lock);
|
2017-04-05 11:32:52 +00:00
|
|
|
ext2_sync_super(sb, EXT2_SB(sb)->s_es, 1);
|
2005-04-16 22:20:36 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
raw_inode->i_generation = cpu_to_le32(inode->i_generation);
|
|
|
|
if (S_ISCHR(inode->i_mode) || S_ISBLK(inode->i_mode)) {
|
|
|
|
if (old_valid_dev(inode->i_rdev)) {
|
|
|
|
raw_inode->i_block[0] =
|
|
|
|
cpu_to_le32(old_encode_dev(inode->i_rdev));
|
|
|
|
raw_inode->i_block[1] = 0;
|
|
|
|
} else {
|
|
|
|
raw_inode->i_block[0] = 0;
|
|
|
|
raw_inode->i_block[1] =
|
|
|
|
cpu_to_le32(new_encode_dev(inode->i_rdev));
|
|
|
|
raw_inode->i_block[2] = 0;
|
|
|
|
}
|
|
|
|
} else for (n = 0; n < EXT2_N_BLOCKS; n++)
|
|
|
|
raw_inode->i_block[n] = ei->i_data[n];
|
|
|
|
mark_buffer_dirty(bh);
|
|
|
|
if (do_sync) {
|
|
|
|
sync_dirty_buffer(bh);
|
|
|
|
if (buffer_req(bh) && !buffer_uptodate(bh)) {
|
|
|
|
printk ("IO error syncing ext2 inode [%s:%08lx]\n",
|
|
|
|
sb->s_id, (unsigned long) ino);
|
|
|
|
err = -EIO;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
ei->i_state &= ~EXT2_STATE_NEW;
|
|
|
|
brelse (bh);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2010-03-05 08:21:37 +00:00
|
|
|
int ext2_write_inode(struct inode *inode, struct writeback_control *wbc)
|
|
|
|
{
|
|
|
|
return __ext2_write_inode(inode, wbc->sync_mode == WB_SYNC_ALL);
|
|
|
|
}
|
|
|
|
|
2021-01-21 13:19:43 +00:00
|
|
|
int ext2_getattr(struct user_namespace *mnt_userns, const struct path *path,
|
|
|
|
struct kstat *stat, u32 request_mask, unsigned int query_flags)
|
2019-02-18 01:07:02 +00:00
|
|
|
{
|
|
|
|
struct inode *inode = d_inode(path->dentry);
|
|
|
|
struct ext2_inode_info *ei = EXT2_I(inode);
|
|
|
|
unsigned int flags;
|
|
|
|
|
|
|
|
flags = ei->i_flags & EXT2_FL_USER_VISIBLE;
|
|
|
|
if (flags & EXT2_APPEND_FL)
|
|
|
|
stat->attributes |= STATX_ATTR_APPEND;
|
|
|
|
if (flags & EXT2_COMPR_FL)
|
|
|
|
stat->attributes |= STATX_ATTR_COMPRESSED;
|
|
|
|
if (flags & EXT2_IMMUTABLE_FL)
|
|
|
|
stat->attributes |= STATX_ATTR_IMMUTABLE;
|
|
|
|
if (flags & EXT2_NODUMP_FL)
|
|
|
|
stat->attributes |= STATX_ATTR_NODUMP;
|
|
|
|
stat->attributes_mask |= (STATX_ATTR_APPEND |
|
|
|
|
STATX_ATTR_COMPRESSED |
|
|
|
|
STATX_ATTR_ENCRYPTED |
|
|
|
|
STATX_ATTR_IMMUTABLE |
|
|
|
|
STATX_ATTR_NODUMP);
|
|
|
|
|
2021-01-21 13:19:30 +00:00
|
|
|
generic_fillattr(&init_user_ns, inode, stat);
|
2019-02-18 01:07:02 +00:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2021-01-21 13:19:43 +00:00
|
|
|
int ext2_setattr(struct user_namespace *mnt_userns, struct dentry *dentry,
|
|
|
|
struct iattr *iattr)
|
2005-04-16 22:20:36 +00:00
|
|
|
{
|
2015-03-17 22:25:59 +00:00
|
|
|
struct inode *inode = d_inode(dentry);
|
2005-04-16 22:20:36 +00:00
|
|
|
int error;
|
|
|
|
|
2021-01-21 13:19:26 +00:00
|
|
|
error = setattr_prepare(&init_user_ns, dentry, iattr);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (error)
|
|
|
|
return error;
|
2010-03-03 14:05:06 +00:00
|
|
|
|
2015-06-29 14:08:45 +00:00
|
|
|
if (is_quota_modification(inode, iattr)) {
|
|
|
|
error = dquot_initialize(inode);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
}
|
2012-02-07 23:39:12 +00:00
|
|
|
if ((iattr->ia_valid & ATTR_UID && !uid_eq(iattr->ia_uid, inode->i_uid)) ||
|
|
|
|
(iattr->ia_valid & ATTR_GID && !gid_eq(iattr->ia_gid, inode->i_gid))) {
|
2010-03-03 14:05:03 +00:00
|
|
|
error = dquot_transfer(inode, iattr);
|
2005-04-16 22:20:36 +00:00
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
}
|
2010-06-03 12:01:46 +00:00
|
|
|
if (iattr->ia_valid & ATTR_SIZE && iattr->ia_size != inode->i_size) {
|
2010-05-26 15:05:37 +00:00
|
|
|
error = ext2_setsize(inode, iattr->ia_size);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
}
|
2021-01-21 13:19:26 +00:00
|
|
|
setattr_copy(&init_user_ns, inode, iattr);
|
2010-05-26 15:05:37 +00:00
|
|
|
if (iattr->ia_valid & ATTR_MODE)
|
2021-01-21 13:19:27 +00:00
|
|
|
error = posix_acl_chmod(&init_user_ns, inode, inode->i_mode);
|
2010-05-26 15:05:37 +00:00
|
|
|
mark_inode_dirty(inode);
|
|
|
|
|
2005-04-16 22:20:36 +00:00
|
|
|
return error;
|
|
|
|
}
|