mirror of
https://github.com/torvalds/linux.git
synced 2024-11-21 19:41:42 +00:00
ufs cleanups, fixes and folio conversion
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk> -----BEGIN PGP SIGNATURE----- iHUEABYIAB0WIQQqUNBr3gm4hGXdBJlZ7Krx/gZQ6wUCZzdkpwAKCRBZ7Krx/gZQ 632xAP0WukRCsROhDX1LsNXY+SF3E6znsO8VwjjHTDwZ+dGW6gD/f+KooaGG5qBU 4Z7sVuOIpOOA0S1LLDx8XBYidcG23g8= =Aw1m -----END PGP SIGNATURE----- Merge tag 'pull-ufs' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs Pull ufs updates from Al Viro: "ufs cleanups, fixes and folio conversion" * tag 'pull-ufs' of git://git.kernel.org/pub/scm/linux/kernel/git/viro/vfs: ufs: ufs_sb_private_info: remove unused s_{2,3}apb fields ufs: Convert ufs_change_blocknr() to take a folio ufs: Pass a folio to ufs_new_fragments() ufs: Convert ufs_inode_getfrag() to take a folio ufs: Convert ufs_extend_tail() to take a folio ufs: Convert ufs_inode_getblock() to take a folio ufs: take the handling of free block counters into a helper clean ufs_trunc_direct() up a bit... ufs: get rid of ubh_{ubhcpymem,memcpyubh}() ufs_inode_getfrag(): remove junk comment ufs_free_fragments(): fix the braino in sanity check ufs_clusteracct(): switch to passing fragment number ufs: untangle ubh_...block...(), part 3 ufs: untangle ubh_...block...(), part 2 ufs: untangle ubh_...block...() macros, part 1 ufs: fix ufs_read_cylinder() failure handling ufs: missing ->splice_write() ufs: fix handling of delete_entry and set_link failures
This commit is contained in:
commit
9fb2cfa463
107
fs/ufs/balloc.c
107
fs/ufs/balloc.c
@ -33,6 +33,29 @@ static u64 ufs_bitmap_search (struct super_block *, struct ufs_cg_private_info *
|
||||
static unsigned char ufs_fragtable_8fpb[], ufs_fragtable_other[];
|
||||
static void ufs_clusteracct(struct super_block *, struct ufs_cg_private_info *, unsigned, int);
|
||||
|
||||
static void adjust_free_blocks(struct super_block *sb,
|
||||
struct ufs_cylinder_group *ucg,
|
||||
struct ufs_cg_private_info *ucpi,
|
||||
unsigned fragment, int delta)
|
||||
{
|
||||
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
|
||||
|
||||
if ((UFS_SB(sb)->s_flags & UFS_CG_MASK) == UFS_CG_44BSD)
|
||||
ufs_clusteracct(sb, ucpi, fragment, delta);
|
||||
|
||||
fs32_add(sb, &ucg->cg_cs.cs_nbfree, delta);
|
||||
uspi->cs_total.cs_nbfree += delta;
|
||||
fs32_add(sb, &UFS_SB(sb)->fs_cs(ucpi->c_cgx).cs_nbfree, delta);
|
||||
|
||||
if (uspi->fs_magic != UFS2_MAGIC) {
|
||||
unsigned cylno = ufs_cbtocylno(fragment);
|
||||
|
||||
fs16_add(sb, &ubh_cg_blks(ucpi, cylno,
|
||||
ufs_cbtorpos(fragment)), delta);
|
||||
fs32_add(sb, &ubh_cg_blktot(ucpi, cylno), delta);
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* Free 'count' fragments from fragment number 'fragment'
|
||||
*/
|
||||
@ -43,7 +66,6 @@ void ufs_free_fragments(struct inode *inode, u64 fragment, unsigned count)
|
||||
struct ufs_cg_private_info * ucpi;
|
||||
struct ufs_cylinder_group * ucg;
|
||||
unsigned cgno, bit, end_bit, bbase, blkmap, i;
|
||||
u64 blkno;
|
||||
|
||||
sb = inode->i_sb;
|
||||
uspi = UFS_SB(sb)->s_uspi;
|
||||
@ -51,7 +73,7 @@ void ufs_free_fragments(struct inode *inode, u64 fragment, unsigned count)
|
||||
UFSD("ENTER, fragment %llu, count %u\n",
|
||||
(unsigned long long)fragment, count);
|
||||
|
||||
if (ufs_fragnum(fragment) + count > uspi->s_fpg)
|
||||
if (ufs_fragnum(fragment) + count > uspi->s_fpb)
|
||||
ufs_error (sb, "ufs_free_fragments", "internal error");
|
||||
|
||||
mutex_lock(&UFS_SB(sb)->s_lock);
|
||||
@ -94,23 +116,11 @@ void ufs_free_fragments(struct inode *inode, u64 fragment, unsigned count)
|
||||
/*
|
||||
* Trying to reassemble free fragments into block
|
||||
*/
|
||||
blkno = ufs_fragstoblks (bbase);
|
||||
if (ubh_isblockset(UCPI_UBH(ucpi), ucpi->c_freeoff, blkno)) {
|
||||
if (ubh_isblockset(uspi, ucpi, bbase)) {
|
||||
fs32_sub(sb, &ucg->cg_cs.cs_nffree, uspi->s_fpb);
|
||||
uspi->cs_total.cs_nffree -= uspi->s_fpb;
|
||||
fs32_sub(sb, &UFS_SB(sb)->fs_cs(cgno).cs_nffree, uspi->s_fpb);
|
||||
if ((UFS_SB(sb)->s_flags & UFS_CG_MASK) == UFS_CG_44BSD)
|
||||
ufs_clusteracct (sb, ucpi, blkno, 1);
|
||||
fs32_add(sb, &ucg->cg_cs.cs_nbfree, 1);
|
||||
uspi->cs_total.cs_nbfree++;
|
||||
fs32_add(sb, &UFS_SB(sb)->fs_cs(cgno).cs_nbfree, 1);
|
||||
if (uspi->fs_magic != UFS2_MAGIC) {
|
||||
unsigned cylno = ufs_cbtocylno (bbase);
|
||||
|
||||
fs16_add(sb, &ubh_cg_blks(ucpi, cylno,
|
||||
ufs_cbtorpos(bbase)), 1);
|
||||
fs32_add(sb, &ubh_cg_blktot(ucpi, cylno), 1);
|
||||
}
|
||||
adjust_free_blocks(sb, ucg, ucpi, bbase, 1);
|
||||
}
|
||||
|
||||
ubh_mark_buffer_dirty (USPI_UBH(uspi));
|
||||
@ -139,7 +149,6 @@ void ufs_free_blocks(struct inode *inode, u64 fragment, unsigned count)
|
||||
struct ufs_cg_private_info * ucpi;
|
||||
struct ufs_cylinder_group * ucg;
|
||||
unsigned overflow, cgno, bit, end_bit, i;
|
||||
u64 blkno;
|
||||
|
||||
sb = inode->i_sb;
|
||||
uspi = UFS_SB(sb)->s_uspi;
|
||||
@ -181,26 +190,12 @@ do_more:
|
||||
}
|
||||
|
||||
for (i = bit; i < end_bit; i += uspi->s_fpb) {
|
||||
blkno = ufs_fragstoblks(i);
|
||||
if (ubh_isblockset(UCPI_UBH(ucpi), ucpi->c_freeoff, blkno)) {
|
||||
if (ubh_isblockset(uspi, ucpi, i)) {
|
||||
ufs_error(sb, "ufs_free_blocks", "freeing free fragment");
|
||||
}
|
||||
ubh_setblock(UCPI_UBH(ucpi), ucpi->c_freeoff, blkno);
|
||||
ubh_setblock(uspi, ucpi, i);
|
||||
inode_sub_bytes(inode, uspi->s_fpb << uspi->s_fshift);
|
||||
if ((UFS_SB(sb)->s_flags & UFS_CG_MASK) == UFS_CG_44BSD)
|
||||
ufs_clusteracct (sb, ucpi, blkno, 1);
|
||||
|
||||
fs32_add(sb, &ucg->cg_cs.cs_nbfree, 1);
|
||||
uspi->cs_total.cs_nbfree++;
|
||||
fs32_add(sb, &UFS_SB(sb)->fs_cs(cgno).cs_nbfree, 1);
|
||||
|
||||
if (uspi->fs_magic != UFS2_MAGIC) {
|
||||
unsigned cylno = ufs_cbtocylno(i);
|
||||
|
||||
fs16_add(sb, &ubh_cg_blks(ucpi, cylno,
|
||||
ufs_cbtorpos(i)), 1);
|
||||
fs32_add(sb, &ubh_cg_blktot(ucpi, cylno), 1);
|
||||
}
|
||||
adjust_free_blocks(sb, ucg, ucpi, i, 1);
|
||||
}
|
||||
|
||||
ubh_mark_buffer_dirty (USPI_UBH(uspi));
|
||||
@ -234,13 +229,13 @@ failed:
|
||||
* situated at the end of file.
|
||||
*
|
||||
* We can come here from ufs_writepage or ufs_prepare_write,
|
||||
* locked_page is argument of these functions, so we already lock it.
|
||||
* locked_folio is argument of these functions, so we already lock it.
|
||||
*/
|
||||
static void ufs_change_blocknr(struct inode *inode, sector_t beg,
|
||||
unsigned int count, sector_t oldb,
|
||||
sector_t newb, struct page *locked_page)
|
||||
sector_t newb, struct folio *locked_folio)
|
||||
{
|
||||
struct folio *folio, *locked_folio = page_folio(locked_page);
|
||||
struct folio *folio;
|
||||
const unsigned blks_per_page =
|
||||
1 << (PAGE_SHIFT - inode->i_blkbits);
|
||||
const unsigned mask = blks_per_page - 1;
|
||||
@ -337,7 +332,7 @@ static void ufs_clear_frags(struct inode *inode, sector_t beg, unsigned int n,
|
||||
|
||||
u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment,
|
||||
u64 goal, unsigned count, int *err,
|
||||
struct page *locked_page)
|
||||
struct folio *locked_folio)
|
||||
{
|
||||
struct super_block * sb;
|
||||
struct ufs_sb_private_info * uspi;
|
||||
@ -417,7 +412,7 @@ u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment,
|
||||
result = ufs_alloc_fragments (inode, cgno, goal, count, err);
|
||||
if (result) {
|
||||
ufs_clear_frags(inode, result + oldcount,
|
||||
newcount - oldcount, locked_page != NULL);
|
||||
newcount - oldcount, locked_folio != NULL);
|
||||
*err = 0;
|
||||
write_seqlock(&UFS_I(inode)->meta_lock);
|
||||
ufs_cpu_to_data_ptr(sb, p, result);
|
||||
@ -441,7 +436,7 @@ u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment,
|
||||
fragment + count);
|
||||
read_sequnlock_excl(&UFS_I(inode)->meta_lock);
|
||||
ufs_clear_frags(inode, result + oldcount, newcount - oldcount,
|
||||
locked_page != NULL);
|
||||
locked_folio != NULL);
|
||||
mutex_unlock(&UFS_SB(sb)->s_lock);
|
||||
UFSD("EXIT, result %llu\n", (unsigned long long)result);
|
||||
return result;
|
||||
@ -462,11 +457,11 @@ u64 ufs_new_fragments(struct inode *inode, void *p, u64 fragment,
|
||||
result = ufs_alloc_fragments (inode, cgno, goal, request, err);
|
||||
if (result) {
|
||||
ufs_clear_frags(inode, result + oldcount, newcount - oldcount,
|
||||
locked_page != NULL);
|
||||
locked_folio != NULL);
|
||||
mutex_unlock(&UFS_SB(sb)->s_lock);
|
||||
ufs_change_blocknr(inode, fragment - oldcount, oldcount,
|
||||
uspi->s_sbbase + tmp,
|
||||
uspi->s_sbbase + result, locked_page);
|
||||
uspi->s_sbbase + result, locked_folio);
|
||||
*err = 0;
|
||||
write_seqlock(&UFS_I(inode)->meta_lock);
|
||||
ufs_cpu_to_data_ptr(sb, p, result);
|
||||
@ -698,7 +693,7 @@ static u64 ufs_alloccg_block(struct inode *inode,
|
||||
struct super_block * sb;
|
||||
struct ufs_sb_private_info * uspi;
|
||||
struct ufs_cylinder_group * ucg;
|
||||
u64 result, blkno;
|
||||
u64 result;
|
||||
|
||||
UFSD("ENTER, goal %llu\n", (unsigned long long)goal);
|
||||
|
||||
@ -716,7 +711,7 @@ static u64 ufs_alloccg_block(struct inode *inode,
|
||||
/*
|
||||
* If the requested block is available, use it.
|
||||
*/
|
||||
if (ubh_isblockset(UCPI_UBH(ucpi), ucpi->c_freeoff, ufs_fragstoblks(goal))) {
|
||||
if (ubh_isblockset(uspi, ucpi, goal)) {
|
||||
result = goal;
|
||||
goto gotit;
|
||||
}
|
||||
@ -729,22 +724,8 @@ norot:
|
||||
gotit:
|
||||
if (!try_add_frags(inode, uspi->s_fpb))
|
||||
return 0;
|
||||
blkno = ufs_fragstoblks(result);
|
||||
ubh_clrblock (UCPI_UBH(ucpi), ucpi->c_freeoff, blkno);
|
||||
if ((UFS_SB(sb)->s_flags & UFS_CG_MASK) == UFS_CG_44BSD)
|
||||
ufs_clusteracct (sb, ucpi, blkno, -1);
|
||||
|
||||
fs32_sub(sb, &ucg->cg_cs.cs_nbfree, 1);
|
||||
uspi->cs_total.cs_nbfree--;
|
||||
fs32_sub(sb, &UFS_SB(sb)->fs_cs(ucpi->c_cgx).cs_nbfree, 1);
|
||||
|
||||
if (uspi->fs_magic != UFS2_MAGIC) {
|
||||
unsigned cylno = ufs_cbtocylno((unsigned)result);
|
||||
|
||||
fs16_sub(sb, &ubh_cg_blks(ucpi, cylno,
|
||||
ufs_cbtorpos((unsigned)result)), 1);
|
||||
fs32_sub(sb, &ubh_cg_blktot(ucpi, cylno), 1);
|
||||
}
|
||||
ubh_clrblock(uspi, ucpi, result);
|
||||
adjust_free_blocks(sb, ucg, ucpi, result, -1);
|
||||
|
||||
UFSD("EXIT, result %llu\n", (unsigned long long)result);
|
||||
|
||||
@ -863,12 +844,12 @@ static u64 ufs_bitmap_search(struct super_block *sb,
|
||||
}
|
||||
|
||||
static void ufs_clusteracct(struct super_block * sb,
|
||||
struct ufs_cg_private_info * ucpi, unsigned blkno, int cnt)
|
||||
struct ufs_cg_private_info * ucpi, unsigned frag, int cnt)
|
||||
{
|
||||
struct ufs_sb_private_info * uspi;
|
||||
struct ufs_sb_private_info * uspi = UFS_SB(sb)->s_uspi;
|
||||
int i, start, end, forw, back;
|
||||
unsigned blkno = ufs_fragstoblks(frag);
|
||||
|
||||
uspi = UFS_SB(sb)->s_uspi;
|
||||
if (uspi->s_contigsumsize <= 0)
|
||||
return;
|
||||
|
||||
|
@ -26,7 +26,7 @@
|
||||
* Read cylinder group into cache. The memory space for ufs_cg_private_info
|
||||
* structure is already allocated during ufs_read_super.
|
||||
*/
|
||||
static void ufs_read_cylinder (struct super_block * sb,
|
||||
static bool ufs_read_cylinder(struct super_block *sb,
|
||||
unsigned cgno, unsigned bitmap_nr)
|
||||
{
|
||||
struct ufs_sb_info * sbi = UFS_SB(sb);
|
||||
@ -46,9 +46,11 @@ static void ufs_read_cylinder (struct super_block * sb,
|
||||
* We have already the first fragment of cylinder group block in buffer
|
||||
*/
|
||||
UCPI_UBH(ucpi)->bh[0] = sbi->s_ucg[cgno];
|
||||
for (i = 1; i < UCPI_UBH(ucpi)->count; i++)
|
||||
if (!(UCPI_UBH(ucpi)->bh[i] = sb_bread(sb, UCPI_UBH(ucpi)->fragment + i)))
|
||||
for (i = 1; i < UCPI_UBH(ucpi)->count; i++) {
|
||||
UCPI_UBH(ucpi)->bh[i] = sb_bread(sb, UCPI_UBH(ucpi)->fragment + i);
|
||||
if (!UCPI_UBH(ucpi)->bh[i])
|
||||
goto failed;
|
||||
}
|
||||
sbi->s_cgno[bitmap_nr] = cgno;
|
||||
|
||||
ucpi->c_cgx = fs32_to_cpu(sb, ucg->cg_cgx);
|
||||
@ -67,13 +69,14 @@ static void ufs_read_cylinder (struct super_block * sb,
|
||||
ucpi->c_clusteroff = fs32_to_cpu(sb, ucg->cg_u.cg_44.cg_clusteroff);
|
||||
ucpi->c_nclusterblks = fs32_to_cpu(sb, ucg->cg_u.cg_44.cg_nclusterblks);
|
||||
UFSD("EXIT\n");
|
||||
return;
|
||||
return true;
|
||||
|
||||
failed:
|
||||
for (j = 1; j < i; j++)
|
||||
brelse (sbi->s_ucg[j]);
|
||||
brelse(UCPI_UBH(ucpi)->bh[j]);
|
||||
sbi->s_cgno[bitmap_nr] = UFS_CGNO_EMPTY;
|
||||
ufs_error (sb, "ufs_read_cylinder", "can't read cylinder group block %u", cgno);
|
||||
return false;
|
||||
}
|
||||
|
||||
/*
|
||||
@ -156,15 +159,14 @@ struct ufs_cg_private_info * ufs_load_cylinder (
|
||||
UFSD("EXIT (FAILED)\n");
|
||||
return NULL;
|
||||
}
|
||||
else {
|
||||
UFSD("EXIT\n");
|
||||
return sbi->s_ucpi[cgno];
|
||||
}
|
||||
} else {
|
||||
ufs_read_cylinder (sb, cgno, cgno);
|
||||
UFSD("EXIT\n");
|
||||
return sbi->s_ucpi[cgno];
|
||||
if (unlikely(!ufs_read_cylinder (sb, cgno, cgno))) {
|
||||
UFSD("EXIT (FAILED)\n");
|
||||
return NULL;
|
||||
}
|
||||
}
|
||||
UFSD("EXIT\n");
|
||||
return sbi->s_ucpi[cgno];
|
||||
}
|
||||
/*
|
||||
* Cylinder group number cg is in cache but it was not last used,
|
||||
@ -195,7 +197,10 @@ struct ufs_cg_private_info * ufs_load_cylinder (
|
||||
sbi->s_ucpi[j] = sbi->s_ucpi[j-1];
|
||||
}
|
||||
sbi->s_ucpi[0] = ucpi;
|
||||
ufs_read_cylinder (sb, cgno, 0);
|
||||
if (unlikely(!ufs_read_cylinder (sb, cgno, 0))) {
|
||||
UFSD("EXIT (FAILED)\n");
|
||||
return NULL;
|
||||
}
|
||||
}
|
||||
UFSD("EXIT\n");
|
||||
return sbi->s_ucpi[0];
|
||||
|
29
fs/ufs/dir.c
29
fs/ufs/dir.c
@ -81,10 +81,9 @@ ino_t ufs_inode_by_name(struct inode *dir, const struct qstr *qstr)
|
||||
}
|
||||
|
||||
|
||||
/* Releases the page */
|
||||
void ufs_set_link(struct inode *dir, struct ufs_dir_entry *de,
|
||||
struct folio *folio, struct inode *inode,
|
||||
bool update_times)
|
||||
int ufs_set_link(struct inode *dir, struct ufs_dir_entry *de,
|
||||
struct folio *folio, struct inode *inode,
|
||||
bool update_times)
|
||||
{
|
||||
loff_t pos = folio_pos(folio) + offset_in_folio(folio, de);
|
||||
unsigned len = fs16_to_cpu(dir->i_sb, de->d_reclen);
|
||||
@ -92,17 +91,19 @@ void ufs_set_link(struct inode *dir, struct ufs_dir_entry *de,
|
||||
|
||||
folio_lock(folio);
|
||||
err = ufs_prepare_chunk(folio, pos, len);
|
||||
BUG_ON(err);
|
||||
if (unlikely(err)) {
|
||||
folio_unlock(folio);
|
||||
return err;
|
||||
}
|
||||
|
||||
de->d_ino = cpu_to_fs32(dir->i_sb, inode->i_ino);
|
||||
ufs_set_de_type(dir->i_sb, de, inode->i_mode);
|
||||
|
||||
ufs_commit_chunk(folio, pos, len);
|
||||
folio_release_kmap(folio, de);
|
||||
if (update_times)
|
||||
inode_set_mtime_to_ts(dir, inode_set_ctime_current(dir));
|
||||
mark_inode_dirty(dir);
|
||||
ufs_handle_dirsync(dir);
|
||||
return ufs_handle_dirsync(dir);
|
||||
}
|
||||
|
||||
static bool ufs_check_folio(struct folio *folio, char *kaddr)
|
||||
@ -505,8 +506,7 @@ int ufs_delete_entry(struct inode *inode, struct ufs_dir_entry *dir,
|
||||
if (de->d_reclen == 0) {
|
||||
ufs_error(inode->i_sb, __func__,
|
||||
"zero-length directory entry");
|
||||
err = -EIO;
|
||||
goto out;
|
||||
return -EIO;
|
||||
}
|
||||
pde = de;
|
||||
de = ufs_next_entry(sb, de);
|
||||
@ -516,18 +516,17 @@ int ufs_delete_entry(struct inode *inode, struct ufs_dir_entry *dir,
|
||||
pos = folio_pos(folio) + from;
|
||||
folio_lock(folio);
|
||||
err = ufs_prepare_chunk(folio, pos, to - from);
|
||||
BUG_ON(err);
|
||||
if (unlikely(err)) {
|
||||
folio_unlock(folio);
|
||||
return err;
|
||||
}
|
||||
if (pde)
|
||||
pde->d_reclen = cpu_to_fs16(sb, to - from);
|
||||
dir->d_ino = 0;
|
||||
ufs_commit_chunk(folio, pos, to - from);
|
||||
inode_set_mtime_to_ts(inode, inode_set_ctime_current(inode));
|
||||
mark_inode_dirty(inode);
|
||||
err = ufs_handle_dirsync(inode);
|
||||
out:
|
||||
folio_release_kmap(folio, kaddr);
|
||||
UFSD("EXIT\n");
|
||||
return err;
|
||||
return ufs_handle_dirsync(inode);
|
||||
}
|
||||
|
||||
int ufs_make_empty(struct inode * inode, struct inode *dir)
|
||||
|
@ -42,4 +42,5 @@ const struct file_operations ufs_file_operations = {
|
||||
.open = generic_file_open,
|
||||
.fsync = generic_file_fsync,
|
||||
.splice_read = filemap_splice_read,
|
||||
.splice_write = iter_file_splice_write,
|
||||
};
|
||||
|
187
fs/ufs/inode.c
187
fs/ufs/inode.c
@ -220,7 +220,7 @@ changed:
|
||||
*/
|
||||
static bool
|
||||
ufs_extend_tail(struct inode *inode, u64 writes_to,
|
||||
int *err, struct page *locked_page)
|
||||
int *err, struct folio *locked_folio)
|
||||
{
|
||||
struct ufs_inode_info *ufsi = UFS_I(inode);
|
||||
struct super_block *sb = inode->i_sb;
|
||||
@ -239,7 +239,7 @@ ufs_extend_tail(struct inode *inode, u64 writes_to,
|
||||
p = ufs_get_direct_data_ptr(uspi, ufsi, block);
|
||||
tmp = ufs_new_fragments(inode, p, lastfrag, ufs_data_ptr_to_cpu(sb, p),
|
||||
new_size - (lastfrag & uspi->s_fpbmask), err,
|
||||
locked_page);
|
||||
locked_folio);
|
||||
return tmp != 0;
|
||||
}
|
||||
|
||||
@ -250,12 +250,11 @@ ufs_extend_tail(struct inode *inode, u64 writes_to,
|
||||
* @new_fragment: number of new allocated fragment(s)
|
||||
* @err: we set it if something wrong
|
||||
* @new: we set it if we allocate new block
|
||||
* @locked_page: for ufs_new_fragments()
|
||||
* @locked_folio: for ufs_new_fragments()
|
||||
*/
|
||||
static u64
|
||||
ufs_inode_getfrag(struct inode *inode, unsigned index,
|
||||
static u64 ufs_inode_getfrag(struct inode *inode, unsigned index,
|
||||
sector_t new_fragment, int *err,
|
||||
int *new, struct page *locked_page)
|
||||
int *new, struct folio *locked_folio)
|
||||
{
|
||||
struct ufs_inode_info *ufsi = UFS_I(inode);
|
||||
struct super_block *sb = inode->i_sb;
|
||||
@ -264,11 +263,6 @@ ufs_inode_getfrag(struct inode *inode, unsigned index,
|
||||
unsigned nfrags = uspi->s_fpb;
|
||||
void *p;
|
||||
|
||||
/* TODO : to be done for write support
|
||||
if ( (flags & UFS_TYPE_MASK) == UFS_TYPE_UFS2)
|
||||
goto ufs2;
|
||||
*/
|
||||
|
||||
p = ufs_get_direct_data_ptr(uspi, ufsi, index);
|
||||
tmp = ufs_data_ptr_to_cpu(sb, p);
|
||||
if (tmp)
|
||||
@ -288,7 +282,7 @@ ufs_inode_getfrag(struct inode *inode, unsigned index,
|
||||
goal += uspi->s_fpb;
|
||||
}
|
||||
tmp = ufs_new_fragments(inode, p, ufs_blknum(new_fragment),
|
||||
goal, nfrags, err, locked_page);
|
||||
goal, nfrags, err, locked_folio);
|
||||
|
||||
if (!tmp) {
|
||||
*err = -ENOSPC;
|
||||
@ -303,21 +297,6 @@ ufs_inode_getfrag(struct inode *inode, unsigned index,
|
||||
mark_inode_dirty(inode);
|
||||
out:
|
||||
return tmp + uspi->s_sbbase;
|
||||
|
||||
/* This part : To be implemented ....
|
||||
Required only for writing, not required for READ-ONLY.
|
||||
ufs2:
|
||||
|
||||
u2_block = ufs_fragstoblks(fragment);
|
||||
u2_blockoff = ufs_fragnum(fragment);
|
||||
p = ufsi->i_u1.u2_i_data + block;
|
||||
goal = 0;
|
||||
|
||||
repeat2:
|
||||
tmp = fs32_to_cpu(sb, *p);
|
||||
lastfrag = ufsi->i_lastfrag;
|
||||
|
||||
*/
|
||||
}
|
||||
|
||||
/**
|
||||
@ -329,12 +308,11 @@ repeat2:
|
||||
* (block will hold this fragment and also uspi->s_fpb-1)
|
||||
* @err: see ufs_inode_getfrag()
|
||||
* @new: see ufs_inode_getfrag()
|
||||
* @locked_page: see ufs_inode_getfrag()
|
||||
* @locked_folio: see ufs_inode_getfrag()
|
||||
*/
|
||||
static u64
|
||||
ufs_inode_getblock(struct inode *inode, u64 ind_block,
|
||||
unsigned index, sector_t new_fragment, int *err,
|
||||
int *new, struct page *locked_page)
|
||||
static u64 ufs_inode_getblock(struct inode *inode, u64 ind_block,
|
||||
unsigned index, sector_t new_fragment, int *err,
|
||||
int *new, struct folio *locked_folio)
|
||||
{
|
||||
struct super_block *sb = inode->i_sb;
|
||||
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
|
||||
@ -369,7 +347,7 @@ ufs_inode_getblock(struct inode *inode, u64 ind_block,
|
||||
else
|
||||
goal = bh->b_blocknr + uspi->s_fpb;
|
||||
tmp = ufs_new_fragments(inode, p, ufs_blknum(new_fragment), goal,
|
||||
uspi->s_fpb, err, locked_page);
|
||||
uspi->s_fpb, err, locked_folio);
|
||||
if (!tmp)
|
||||
goto out;
|
||||
|
||||
@ -434,14 +412,14 @@ static int ufs_getfrag_block(struct inode *inode, sector_t fragment, struct buff
|
||||
unsigned tailfrags = lastfrag & uspi->s_fpbmask;
|
||||
if (tailfrags && fragment >= lastfrag) {
|
||||
if (!ufs_extend_tail(inode, fragment,
|
||||
&err, bh_result->b_page))
|
||||
&err, bh_result->b_folio))
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
|
||||
if (depth == 1) {
|
||||
phys64 = ufs_inode_getfrag(inode, offsets[0], fragment,
|
||||
&err, &new, bh_result->b_page);
|
||||
&err, &new, bh_result->b_folio);
|
||||
} else {
|
||||
int i;
|
||||
phys64 = ufs_inode_getfrag(inode, offsets[0], fragment,
|
||||
@ -450,7 +428,7 @@ static int ufs_getfrag_block(struct inode *inode, sector_t fragment, struct buff
|
||||
phys64 = ufs_inode_getblock(inode, phys64, offsets[i],
|
||||
fragment, &err, NULL, NULL);
|
||||
phys64 = ufs_inode_getblock(inode, phys64, offsets[depth - 1],
|
||||
fragment, &err, &new, bh_result->b_page);
|
||||
fragment, &err, &new, bh_result->b_folio);
|
||||
}
|
||||
out:
|
||||
if (phys64) {
|
||||
@ -898,91 +876,84 @@ static inline void free_data(struct to_free *ctx, u64 from, unsigned count)
|
||||
|
||||
#define DIRECT_FRAGMENT ((inode->i_size + uspi->s_fsize - 1) >> uspi->s_fshift)
|
||||
|
||||
/*
|
||||
* used only for truncation down to direct blocks.
|
||||
*/
|
||||
static void ufs_trunc_direct(struct inode *inode)
|
||||
{
|
||||
struct ufs_inode_info *ufsi = UFS_I(inode);
|
||||
struct super_block * sb;
|
||||
struct ufs_sb_private_info * uspi;
|
||||
void *p;
|
||||
u64 frag1, frag2, frag3, frag4, block1, block2;
|
||||
struct super_block *sb = inode->i_sb;
|
||||
struct ufs_sb_private_info *uspi = UFS_SB(sb)->s_uspi;
|
||||
unsigned int new_frags, old_frags;
|
||||
unsigned int old_slot, new_slot;
|
||||
unsigned int old_tail, new_tail;
|
||||
struct to_free ctx = {.inode = inode};
|
||||
unsigned i, tmp;
|
||||
|
||||
UFSD("ENTER: ino %lu\n", inode->i_ino);
|
||||
|
||||
sb = inode->i_sb;
|
||||
uspi = UFS_SB(sb)->s_uspi;
|
||||
new_frags = DIRECT_FRAGMENT;
|
||||
// new_frags = first fragment past the new EOF
|
||||
old_frags = min_t(u64, UFS_NDIR_FRAGMENT, ufsi->i_lastfrag);
|
||||
// old_frags = first fragment past the old EOF or covered by indirects
|
||||
|
||||
frag1 = DIRECT_FRAGMENT;
|
||||
frag4 = min_t(u64, UFS_NDIR_FRAGMENT, ufsi->i_lastfrag);
|
||||
frag2 = ((frag1 & uspi->s_fpbmask) ? ((frag1 | uspi->s_fpbmask) + 1) : frag1);
|
||||
frag3 = frag4 & ~uspi->s_fpbmask;
|
||||
block1 = block2 = 0;
|
||||
if (frag2 > frag3) {
|
||||
frag2 = frag4;
|
||||
frag3 = frag4 = 0;
|
||||
} else if (frag2 < frag3) {
|
||||
block1 = ufs_fragstoblks (frag2);
|
||||
block2 = ufs_fragstoblks (frag3);
|
||||
}
|
||||
if (new_frags >= old_frags) // expanding - nothing to free
|
||||
goto done;
|
||||
|
||||
UFSD("ino %lu, frag1 %llu, frag2 %llu, block1 %llu, block2 %llu,"
|
||||
" frag3 %llu, frag4 %llu\n", inode->i_ino,
|
||||
(unsigned long long)frag1, (unsigned long long)frag2,
|
||||
(unsigned long long)block1, (unsigned long long)block2,
|
||||
(unsigned long long)frag3, (unsigned long long)frag4);
|
||||
old_tail = ufs_fragnum(old_frags);
|
||||
old_slot = ufs_fragstoblks(old_frags);
|
||||
new_tail = ufs_fragnum(new_frags);
|
||||
new_slot = ufs_fragstoblks(new_frags);
|
||||
|
||||
if (frag1 >= frag2)
|
||||
goto next1;
|
||||
|
||||
/*
|
||||
* Free first free fragments
|
||||
*/
|
||||
p = ufs_get_direct_data_ptr(uspi, ufsi, ufs_fragstoblks(frag1));
|
||||
tmp = ufs_data_ptr_to_cpu(sb, p);
|
||||
if (!tmp )
|
||||
ufs_panic (sb, "ufs_trunc_direct", "internal error");
|
||||
frag2 -= frag1;
|
||||
frag1 = ufs_fragnum (frag1);
|
||||
|
||||
ufs_free_fragments(inode, tmp + frag1, frag2);
|
||||
|
||||
next1:
|
||||
/*
|
||||
* Free whole blocks
|
||||
*/
|
||||
for (i = block1 ; i < block2; i++) {
|
||||
p = ufs_get_direct_data_ptr(uspi, ufsi, i);
|
||||
tmp = ufs_data_ptr_to_cpu(sb, p);
|
||||
if (old_slot == new_slot) { // old_tail > 0
|
||||
void *p = ufs_get_direct_data_ptr(uspi, ufsi, old_slot);
|
||||
u64 tmp = ufs_data_ptr_to_cpu(sb, p);
|
||||
if (!tmp)
|
||||
continue;
|
||||
write_seqlock(&ufsi->meta_lock);
|
||||
ufs_data_ptr_clear(uspi, p);
|
||||
write_sequnlock(&ufsi->meta_lock);
|
||||
ufs_panic(sb, __func__, "internal error");
|
||||
if (!new_tail) {
|
||||
write_seqlock(&ufsi->meta_lock);
|
||||
ufs_data_ptr_clear(uspi, p);
|
||||
write_sequnlock(&ufsi->meta_lock);
|
||||
}
|
||||
ufs_free_fragments(inode, tmp + new_tail, old_tail - new_tail);
|
||||
} else {
|
||||
unsigned int slot = new_slot;
|
||||
|
||||
free_data(&ctx, tmp, uspi->s_fpb);
|
||||
if (new_tail) {
|
||||
void *p = ufs_get_direct_data_ptr(uspi, ufsi, slot++);
|
||||
u64 tmp = ufs_data_ptr_to_cpu(sb, p);
|
||||
if (!tmp)
|
||||
ufs_panic(sb, __func__, "internal error");
|
||||
|
||||
ufs_free_fragments(inode, tmp + new_tail,
|
||||
uspi->s_fpb - new_tail);
|
||||
}
|
||||
while (slot < old_slot) {
|
||||
void *p = ufs_get_direct_data_ptr(uspi, ufsi, slot++);
|
||||
u64 tmp = ufs_data_ptr_to_cpu(sb, p);
|
||||
if (!tmp)
|
||||
continue;
|
||||
write_seqlock(&ufsi->meta_lock);
|
||||
ufs_data_ptr_clear(uspi, p);
|
||||
write_sequnlock(&ufsi->meta_lock);
|
||||
|
||||
free_data(&ctx, tmp, uspi->s_fpb);
|
||||
}
|
||||
|
||||
free_data(&ctx, 0, 0);
|
||||
|
||||
if (old_tail) {
|
||||
void *p = ufs_get_direct_data_ptr(uspi, ufsi, slot);
|
||||
u64 tmp = ufs_data_ptr_to_cpu(sb, p);
|
||||
if (!tmp)
|
||||
ufs_panic(sb, __func__, "internal error");
|
||||
write_seqlock(&ufsi->meta_lock);
|
||||
ufs_data_ptr_clear(uspi, p);
|
||||
write_sequnlock(&ufsi->meta_lock);
|
||||
|
||||
ufs_free_fragments(inode, tmp, old_tail);
|
||||
}
|
||||
}
|
||||
|
||||
free_data(&ctx, 0, 0);
|
||||
|
||||
if (frag3 >= frag4)
|
||||
goto next3;
|
||||
|
||||
/*
|
||||
* Free last free fragments
|
||||
*/
|
||||
p = ufs_get_direct_data_ptr(uspi, ufsi, ufs_fragstoblks(frag3));
|
||||
tmp = ufs_data_ptr_to_cpu(sb, p);
|
||||
if (!tmp )
|
||||
ufs_panic(sb, "ufs_truncate_direct", "internal error");
|
||||
frag4 = ufs_fragnum (frag4);
|
||||
write_seqlock(&ufsi->meta_lock);
|
||||
ufs_data_ptr_clear(uspi, p);
|
||||
write_sequnlock(&ufsi->meta_lock);
|
||||
|
||||
ufs_free_fragments (inode, tmp, frag4);
|
||||
next3:
|
||||
|
||||
done:
|
||||
UFSD("EXIT: ino %lu\n", inode->i_ino);
|
||||
}
|
||||
|
||||
|
@ -210,20 +210,18 @@ static int ufs_unlink(struct inode *dir, struct dentry *dentry)
|
||||
struct inode * inode = d_inode(dentry);
|
||||
struct ufs_dir_entry *de;
|
||||
struct folio *folio;
|
||||
int err = -ENOENT;
|
||||
int err;
|
||||
|
||||
de = ufs_find_entry(dir, &dentry->d_name, &folio);
|
||||
if (!de)
|
||||
goto out;
|
||||
return -ENOENT;
|
||||
|
||||
err = ufs_delete_entry(dir, de, folio);
|
||||
if (err)
|
||||
goto out;
|
||||
|
||||
inode_set_ctime_to_ts(inode, inode_get_ctime(dir));
|
||||
inode_dec_link_count(inode);
|
||||
err = 0;
|
||||
out:
|
||||
if (!err) {
|
||||
inode_set_ctime_to_ts(inode, inode_get_ctime(dir));
|
||||
inode_dec_link_count(inode);
|
||||
}
|
||||
folio_release_kmap(folio, de);
|
||||
return err;
|
||||
}
|
||||
|
||||
@ -253,14 +251,14 @@ static int ufs_rename(struct mnt_idmap *idmap, struct inode *old_dir,
|
||||
struct ufs_dir_entry * dir_de = NULL;
|
||||
struct folio *old_folio;
|
||||
struct ufs_dir_entry *old_de;
|
||||
int err = -ENOENT;
|
||||
int err;
|
||||
|
||||
if (flags & ~RENAME_NOREPLACE)
|
||||
return -EINVAL;
|
||||
|
||||
old_de = ufs_find_entry(old_dir, &old_dentry->d_name, &old_folio);
|
||||
if (!old_de)
|
||||
goto out;
|
||||
return -ENOENT;
|
||||
|
||||
if (S_ISDIR(old_inode->i_mode)) {
|
||||
err = -EIO;
|
||||
@ -281,7 +279,10 @@ static int ufs_rename(struct mnt_idmap *idmap, struct inode *old_dir,
|
||||
new_de = ufs_find_entry(new_dir, &new_dentry->d_name, &new_folio);
|
||||
if (!new_de)
|
||||
goto out_dir;
|
||||
ufs_set_link(new_dir, new_de, new_folio, old_inode, 1);
|
||||
err = ufs_set_link(new_dir, new_de, new_folio, old_inode, 1);
|
||||
folio_release_kmap(new_folio, new_de);
|
||||
if (err)
|
||||
goto out_dir;
|
||||
inode_set_ctime_current(new_inode);
|
||||
if (dir_de)
|
||||
drop_nlink(new_inode);
|
||||
@ -299,26 +300,20 @@ static int ufs_rename(struct mnt_idmap *idmap, struct inode *old_dir,
|
||||
* rename.
|
||||
*/
|
||||
inode_set_ctime_current(old_inode);
|
||||
|
||||
ufs_delete_entry(old_dir, old_de, old_folio);
|
||||
mark_inode_dirty(old_inode);
|
||||
|
||||
if (dir_de) {
|
||||
err = ufs_delete_entry(old_dir, old_de, old_folio);
|
||||
if (!err && dir_de) {
|
||||
if (old_dir != new_dir)
|
||||
ufs_set_link(old_inode, dir_de, dir_folio, new_dir, 0);
|
||||
else
|
||||
folio_release_kmap(dir_folio, dir_de);
|
||||
err = ufs_set_link(old_inode, dir_de, dir_folio,
|
||||
new_dir, 0);
|
||||
inode_dec_link_count(old_dir);
|
||||
}
|
||||
return 0;
|
||||
|
||||
|
||||
out_dir:
|
||||
if (dir_de)
|
||||
folio_release_kmap(dir_folio, dir_de);
|
||||
out_old:
|
||||
folio_release_kmap(old_folio, old_de);
|
||||
out:
|
||||
return err;
|
||||
}
|
||||
|
||||
|
@ -505,7 +505,6 @@ static int ufs_read_cylinder_structures(struct super_block *sb)
|
||||
{
|
||||
struct ufs_sb_info *sbi = UFS_SB(sb);
|
||||
struct ufs_sb_private_info *uspi = sbi->s_uspi;
|
||||
struct ufs_buffer_head * ubh;
|
||||
unsigned char * base, * space;
|
||||
unsigned size, blks, i;
|
||||
|
||||
@ -521,21 +520,13 @@ static int ufs_read_cylinder_structures(struct super_block *sb)
|
||||
if (!base)
|
||||
goto failed;
|
||||
sbi->s_csp = (struct ufs_csum *)space;
|
||||
for (i = 0; i < blks; i += uspi->s_fpb) {
|
||||
size = uspi->s_bsize;
|
||||
if (i + uspi->s_fpb > blks)
|
||||
size = (blks - i) * uspi->s_fsize;
|
||||
|
||||
ubh = ubh_bread(sb, uspi->s_csaddr + i, size);
|
||||
|
||||
if (!ubh)
|
||||
for (i = 0; i < blks; i++) {
|
||||
struct buffer_head *bh = sb_bread(sb, uspi->s_csaddr + i);
|
||||
if (!bh)
|
||||
goto failed;
|
||||
|
||||
ubh_ubhcpymem (space, ubh, size);
|
||||
|
||||
space += size;
|
||||
ubh_brelse (ubh);
|
||||
ubh = NULL;
|
||||
memcpy(space, bh->b_data, uspi->s_fsize);
|
||||
space += uspi->s_fsize;
|
||||
brelse (bh);
|
||||
}
|
||||
|
||||
/*
|
||||
@ -645,7 +636,6 @@ static void ufs_put_super_internal(struct super_block *sb)
|
||||
{
|
||||
struct ufs_sb_info *sbi = UFS_SB(sb);
|
||||
struct ufs_sb_private_info *uspi = sbi->s_uspi;
|
||||
struct ufs_buffer_head * ubh;
|
||||
unsigned char * base, * space;
|
||||
unsigned blks, size, i;
|
||||
|
||||
@ -656,18 +646,17 @@ static void ufs_put_super_internal(struct super_block *sb)
|
||||
size = uspi->s_cssize;
|
||||
blks = (size + uspi->s_fsize - 1) >> uspi->s_fshift;
|
||||
base = space = (char*) sbi->s_csp;
|
||||
for (i = 0; i < blks; i += uspi->s_fpb) {
|
||||
size = uspi->s_bsize;
|
||||
if (i + uspi->s_fpb > blks)
|
||||
size = (blks - i) * uspi->s_fsize;
|
||||
for (i = 0; i < blks; i++, space += uspi->s_fsize) {
|
||||
struct buffer_head *bh = sb_bread(sb, uspi->s_csaddr + i);
|
||||
|
||||
ubh = ubh_bread(sb, uspi->s_csaddr + i, size);
|
||||
|
||||
ubh_memcpyubh (ubh, space, size);
|
||||
space += size;
|
||||
ubh_mark_buffer_uptodate (ubh, 1);
|
||||
ubh_mark_buffer_dirty (ubh);
|
||||
ubh_brelse (ubh);
|
||||
if (unlikely(!bh)) { // better than an oops...
|
||||
ufs_panic(sb, __func__,
|
||||
"can't write part of cylinder group summary");
|
||||
continue;
|
||||
}
|
||||
memcpy(bh->b_data, space, uspi->s_fsize);
|
||||
mark_buffer_dirty(bh);
|
||||
brelse(bh);
|
||||
}
|
||||
for (i = 0; i < sbi->s_cg_loaded; i++) {
|
||||
ufs_put_cylinder (sb, i);
|
||||
@ -1240,11 +1229,7 @@ magic_found:
|
||||
else
|
||||
uspi->s_apbshift = uspi->s_bshift - 2;
|
||||
|
||||
uspi->s_2apbshift = uspi->s_apbshift * 2;
|
||||
uspi->s_3apbshift = uspi->s_apbshift * 3;
|
||||
uspi->s_apb = 1 << uspi->s_apbshift;
|
||||
uspi->s_2apb = 1 << uspi->s_2apbshift;
|
||||
uspi->s_3apb = 1 << uspi->s_3apbshift;
|
||||
uspi->s_apbmask = uspi->s_apb - 1;
|
||||
uspi->s_nspfshift = uspi->s_fshift - UFS_SECTOR_BITS;
|
||||
uspi->s_nspb = uspi->s_nspf << uspi->s_fpbshift;
|
||||
|
12
fs/ufs/ufs.h
12
fs/ufs/ufs.h
@ -88,10 +88,10 @@ struct ufs_inode_info {
|
||||
#endif
|
||||
|
||||
/* balloc.c */
|
||||
extern void ufs_free_fragments (struct inode *, u64, unsigned);
|
||||
extern void ufs_free_blocks (struct inode *, u64, unsigned);
|
||||
extern u64 ufs_new_fragments(struct inode *, void *, u64, u64,
|
||||
unsigned, int *, struct page *);
|
||||
void ufs_free_fragments (struct inode *, u64 fragment, unsigned count);
|
||||
void ufs_free_blocks (struct inode *, u64 fragment, unsigned count);
|
||||
u64 ufs_new_fragments(struct inode *, void *, u64 fragment, u64 goal,
|
||||
unsigned count, int *err, struct folio *);
|
||||
|
||||
/* cylinder.c */
|
||||
extern struct ufs_cg_private_info * ufs_load_cylinder (struct super_block *, unsigned);
|
||||
@ -108,8 +108,8 @@ struct ufs_dir_entry *ufs_find_entry(struct inode *, const struct qstr *,
|
||||
int ufs_delete_entry(struct inode *, struct ufs_dir_entry *, struct folio *);
|
||||
int ufs_empty_dir(struct inode *);
|
||||
struct ufs_dir_entry *ufs_dotdot(struct inode *, struct folio **);
|
||||
void ufs_set_link(struct inode *dir, struct ufs_dir_entry *de,
|
||||
struct folio *folio, struct inode *inode, bool update_times);
|
||||
int ufs_set_link(struct inode *dir, struct ufs_dir_entry *de,
|
||||
struct folio *folio, struct inode *inode, bool update_times);
|
||||
|
||||
/* file.c */
|
||||
extern const struct inode_operations ufs_file_inode_operations;
|
||||
|
@ -775,12 +775,8 @@ struct ufs_sb_private_info {
|
||||
|
||||
__u32 s_fpbmask; /* fragments per block mask */
|
||||
__u32 s_apb; /* address per block */
|
||||
__u32 s_2apb; /* address per block^2 */
|
||||
__u32 s_3apb; /* address per block^3 */
|
||||
__u32 s_apbmask; /* address per block mask */
|
||||
__u32 s_apbshift; /* address per block shift */
|
||||
__u32 s_2apbshift; /* address per block shift * 2 */
|
||||
__u32 s_3apbshift; /* address per block shift * 3 */
|
||||
__u32 s_nspfshift; /* number of sector per fragment shift */
|
||||
__u32 s_nspb; /* number of sector per block */
|
||||
__u32 s_inopf; /* inodes per fragment */
|
||||
|
@ -99,20 +99,6 @@ void ubh_mark_buffer_dirty (struct ufs_buffer_head * ubh)
|
||||
mark_buffer_dirty (ubh->bh[i]);
|
||||
}
|
||||
|
||||
void ubh_mark_buffer_uptodate (struct ufs_buffer_head * ubh, int flag)
|
||||
{
|
||||
unsigned i;
|
||||
if (!ubh)
|
||||
return;
|
||||
if (flag) {
|
||||
for ( i = 0; i < ubh->count; i++ )
|
||||
set_buffer_uptodate (ubh->bh[i]);
|
||||
} else {
|
||||
for ( i = 0; i < ubh->count; i++ )
|
||||
clear_buffer_uptodate (ubh->bh[i]);
|
||||
}
|
||||
}
|
||||
|
||||
void ubh_sync_block(struct ufs_buffer_head *ubh)
|
||||
{
|
||||
if (ubh) {
|
||||
@ -146,38 +132,6 @@ int ubh_buffer_dirty (struct ufs_buffer_head * ubh)
|
||||
return result;
|
||||
}
|
||||
|
||||
void _ubh_ubhcpymem_(struct ufs_sb_private_info * uspi,
|
||||
unsigned char * mem, struct ufs_buffer_head * ubh, unsigned size)
|
||||
{
|
||||
unsigned len, bhno;
|
||||
if (size > (ubh->count << uspi->s_fshift))
|
||||
size = ubh->count << uspi->s_fshift;
|
||||
bhno = 0;
|
||||
while (size) {
|
||||
len = min_t(unsigned int, size, uspi->s_fsize);
|
||||
memcpy (mem, ubh->bh[bhno]->b_data, len);
|
||||
mem += uspi->s_fsize;
|
||||
size -= len;
|
||||
bhno++;
|
||||
}
|
||||
}
|
||||
|
||||
void _ubh_memcpyubh_(struct ufs_sb_private_info * uspi,
|
||||
struct ufs_buffer_head * ubh, unsigned char * mem, unsigned size)
|
||||
{
|
||||
unsigned len, bhno;
|
||||
if (size > (ubh->count << uspi->s_fshift))
|
||||
size = ubh->count << uspi->s_fshift;
|
||||
bhno = 0;
|
||||
while (size) {
|
||||
len = min_t(unsigned int, size, uspi->s_fsize);
|
||||
memcpy (ubh->bh[bhno]->b_data, mem, len);
|
||||
mem += uspi->s_fsize;
|
||||
size -= len;
|
||||
bhno++;
|
||||
}
|
||||
}
|
||||
|
||||
dev_t
|
||||
ufs_get_inode_dev(struct super_block *sb, struct ufs_inode_info *ufsi)
|
||||
{
|
||||
|
@ -263,14 +263,9 @@ extern struct ufs_buffer_head * ubh_bread_uspi(struct ufs_sb_private_info *, str
|
||||
extern void ubh_brelse (struct ufs_buffer_head *);
|
||||
extern void ubh_brelse_uspi (struct ufs_sb_private_info *);
|
||||
extern void ubh_mark_buffer_dirty (struct ufs_buffer_head *);
|
||||
extern void ubh_mark_buffer_uptodate (struct ufs_buffer_head *, int);
|
||||
extern void ubh_sync_block(struct ufs_buffer_head *);
|
||||
extern void ubh_bforget (struct ufs_buffer_head *);
|
||||
extern int ubh_buffer_dirty (struct ufs_buffer_head *);
|
||||
#define ubh_ubhcpymem(mem,ubh,size) _ubh_ubhcpymem_(uspi,mem,ubh,size)
|
||||
extern void _ubh_ubhcpymem_(struct ufs_sb_private_info *, unsigned char *, struct ufs_buffer_head *, unsigned);
|
||||
#define ubh_memcpyubh(ubh,mem,size) _ubh_memcpyubh_(uspi,ubh,mem,size)
|
||||
extern void _ubh_memcpyubh_(struct ufs_sb_private_info *, struct ufs_buffer_head *, unsigned char *, unsigned);
|
||||
|
||||
/* This functions works with cache pages*/
|
||||
struct folio *ufs_get_locked_folio(struct address_space *mapping, pgoff_t index);
|
||||
@ -455,65 +450,69 @@ static inline unsigned _ubh_find_last_zero_bit_(
|
||||
return (base << uspi->s_bpfshift) + pos - begin;
|
||||
}
|
||||
|
||||
#define ubh_isblockclear(ubh,begin,block) (!_ubh_isblockset_(uspi,ubh,begin,block))
|
||||
|
||||
#define ubh_isblockset(ubh,begin,block) _ubh_isblockset_(uspi,ubh,begin,block)
|
||||
static inline int _ubh_isblockset_(struct ufs_sb_private_info * uspi,
|
||||
struct ufs_buffer_head * ubh, unsigned begin, unsigned block)
|
||||
static inline int ubh_isblockset(struct ufs_sb_private_info *uspi,
|
||||
struct ufs_cg_private_info *ucpi, unsigned int frag)
|
||||
{
|
||||
struct ufs_buffer_head *ubh = UCPI_UBH(ucpi);
|
||||
u8 *p = ubh_get_addr(ubh, ucpi->c_freeoff + (frag >> 3));
|
||||
u8 mask;
|
||||
|
||||
switch (uspi->s_fpb) {
|
||||
case 8:
|
||||
return (*ubh_get_addr (ubh, begin + block) == 0xff);
|
||||
return *p == 0xff;
|
||||
case 4:
|
||||
mask = 0x0f << ((block & 0x01) << 2);
|
||||
return (*ubh_get_addr (ubh, begin + (block >> 1)) & mask) == mask;
|
||||
mask = 0x0f << (frag & 4);
|
||||
return (*p & mask) == mask;
|
||||
case 2:
|
||||
mask = 0x03 << ((block & 0x03) << 1);
|
||||
return (*ubh_get_addr (ubh, begin + (block >> 2)) & mask) == mask;
|
||||
mask = 0x03 << (frag & 6);
|
||||
return (*p & mask) == mask;
|
||||
case 1:
|
||||
mask = 0x01 << (block & 0x07);
|
||||
return (*ubh_get_addr (ubh, begin + (block >> 3)) & mask) == mask;
|
||||
mask = 0x01 << (frag & 7);
|
||||
return (*p & mask) == mask;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
#define ubh_clrblock(ubh,begin,block) _ubh_clrblock_(uspi,ubh,begin,block)
|
||||
static inline void _ubh_clrblock_(struct ufs_sb_private_info * uspi,
|
||||
struct ufs_buffer_head * ubh, unsigned begin, unsigned block)
|
||||
static inline void ubh_clrblock(struct ufs_sb_private_info *uspi,
|
||||
struct ufs_cg_private_info *ucpi, unsigned int frag)
|
||||
{
|
||||
struct ufs_buffer_head *ubh = UCPI_UBH(ucpi);
|
||||
u8 *p = ubh_get_addr(ubh, ucpi->c_freeoff + (frag >> 3));
|
||||
|
||||
switch (uspi->s_fpb) {
|
||||
case 8:
|
||||
*ubh_get_addr (ubh, begin + block) = 0x00;
|
||||
*p = 0x00;
|
||||
return;
|
||||
case 4:
|
||||
*ubh_get_addr (ubh, begin + (block >> 1)) &= ~(0x0f << ((block & 0x01) << 2));
|
||||
*p &= ~(0x0f << (frag & 4));
|
||||
return;
|
||||
case 2:
|
||||
*ubh_get_addr (ubh, begin + (block >> 2)) &= ~(0x03 << ((block & 0x03) << 1));
|
||||
*p &= ~(0x03 << (frag & 6));
|
||||
return;
|
||||
case 1:
|
||||
*ubh_get_addr (ubh, begin + (block >> 3)) &= ~(0x01 << ((block & 0x07)));
|
||||
*p &= ~(0x01 << (frag & 7));
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
#define ubh_setblock(ubh,begin,block) _ubh_setblock_(uspi,ubh,begin,block)
|
||||
static inline void _ubh_setblock_(struct ufs_sb_private_info * uspi,
|
||||
struct ufs_buffer_head * ubh, unsigned begin, unsigned block)
|
||||
static inline void ubh_setblock(struct ufs_sb_private_info * uspi,
|
||||
struct ufs_cg_private_info *ucpi, unsigned int frag)
|
||||
{
|
||||
struct ufs_buffer_head *ubh = UCPI_UBH(ucpi);
|
||||
u8 *p = ubh_get_addr(ubh, ucpi->c_freeoff + (frag >> 3));
|
||||
|
||||
switch (uspi->s_fpb) {
|
||||
case 8:
|
||||
*ubh_get_addr(ubh, begin + block) = 0xff;
|
||||
*p = 0xff;
|
||||
return;
|
||||
case 4:
|
||||
*ubh_get_addr(ubh, begin + (block >> 1)) |= (0x0f << ((block & 0x01) << 2));
|
||||
*p |= 0x0f << (frag & 4);
|
||||
return;
|
||||
case 2:
|
||||
*ubh_get_addr(ubh, begin + (block >> 2)) |= (0x03 << ((block & 0x03) << 1));
|
||||
*p |= 0x03 << (frag & 6);
|
||||
return;
|
||||
case 1:
|
||||
*ubh_get_addr(ubh, begin + (block >> 3)) |= (0x01 << ((block & 0x07)));
|
||||
*p |= 0x01 << (frag & 7);
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
Loading…
Reference in New Issue
Block a user