mirror of
https://github.com/torvalds/linux.git
synced 2024-11-02 10:11:36 +00:00
5be6f7f174
Signed-off-by: Chris Mason <chris.mason@oracle.com>
443 lines
12 KiB
C
443 lines
12 KiB
C
#include <linux/module.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/blkdev.h>
|
|
#include <linux/crypto.h>
|
|
#include <linux/scatterlist.h>
|
|
#include <linux/swap.h>
|
|
#include "ctree.h"
|
|
#include "disk-io.h"
|
|
#include "transaction.h"
|
|
|
|
static int check_tree_block(struct btrfs_root *root, struct buffer_head *buf)
|
|
{
|
|
struct btrfs_node *node = btrfs_buffer_node(buf);
|
|
if (buf->b_blocknr != btrfs_header_blocknr(&node->header)) {
|
|
BUG();
|
|
}
|
|
if (root->node && btrfs_header_parentid(&node->header) !=
|
|
btrfs_header_parentid(btrfs_buffer_header(root->node))) {
|
|
BUG();
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
struct buffer_head *btrfs_find_tree_block(struct btrfs_root *root, u64 blocknr)
|
|
{
|
|
struct address_space *mapping = root->fs_info->btree_inode->i_mapping;
|
|
int blockbits = root->fs_info->sb->s_blocksize_bits;
|
|
unsigned long index = blocknr >> (PAGE_CACHE_SHIFT - blockbits);
|
|
struct page *page;
|
|
struct buffer_head *bh;
|
|
struct buffer_head *head;
|
|
struct buffer_head *ret = NULL;
|
|
|
|
|
|
page = find_lock_page(mapping, index);
|
|
if (!page)
|
|
return NULL;
|
|
|
|
if (!page_has_buffers(page))
|
|
goto out_unlock;
|
|
|
|
head = page_buffers(page);
|
|
bh = head;
|
|
do {
|
|
if (buffer_mapped(bh) && bh->b_blocknr == blocknr) {
|
|
ret = bh;
|
|
get_bh(bh);
|
|
goto out_unlock;
|
|
}
|
|
bh = bh->b_this_page;
|
|
} while (bh != head);
|
|
out_unlock:
|
|
unlock_page(page);
|
|
if (ret) {
|
|
touch_buffer(ret);
|
|
}
|
|
page_cache_release(page);
|
|
return ret;
|
|
}
|
|
|
|
struct buffer_head *btrfs_find_create_tree_block(struct btrfs_root *root,
|
|
u64 blocknr)
|
|
{
|
|
struct address_space *mapping = root->fs_info->btree_inode->i_mapping;
|
|
int blockbits = root->fs_info->sb->s_blocksize_bits;
|
|
unsigned long index = blocknr >> (PAGE_CACHE_SHIFT - blockbits);
|
|
struct page *page;
|
|
struct buffer_head *bh;
|
|
struct buffer_head *head;
|
|
struct buffer_head *ret = NULL;
|
|
u64 first_block = index << (PAGE_CACHE_SHIFT - blockbits);
|
|
|
|
page = grab_cache_page(mapping, index);
|
|
if (!page)
|
|
return NULL;
|
|
|
|
if (!page_has_buffers(page))
|
|
create_empty_buffers(page, root->fs_info->sb->s_blocksize, 0);
|
|
head = page_buffers(page);
|
|
bh = head;
|
|
do {
|
|
if (!buffer_mapped(bh)) {
|
|
bh->b_bdev = root->fs_info->sb->s_bdev;
|
|
bh->b_blocknr = first_block;
|
|
set_buffer_mapped(bh);
|
|
}
|
|
if (bh->b_blocknr == blocknr) {
|
|
ret = bh;
|
|
get_bh(bh);
|
|
goto out_unlock;
|
|
}
|
|
bh = bh->b_this_page;
|
|
first_block++;
|
|
} while (bh != head);
|
|
out_unlock:
|
|
unlock_page(page);
|
|
if (ret)
|
|
touch_buffer(ret);
|
|
page_cache_release(page);
|
|
return ret;
|
|
}
|
|
|
|
static sector_t max_block(struct block_device *bdev)
|
|
{
|
|
sector_t retval = ~((sector_t)0);
|
|
loff_t sz = i_size_read(bdev->bd_inode);
|
|
|
|
if (sz) {
|
|
unsigned int size = block_size(bdev);
|
|
unsigned int sizebits = blksize_bits(size);
|
|
retval = (sz >> sizebits);
|
|
}
|
|
return retval;
|
|
}
|
|
|
|
static int btree_get_block(struct inode *inode, sector_t iblock,
|
|
struct buffer_head *bh, int create)
|
|
{
|
|
if (iblock >= max_block(inode->i_sb->s_bdev)) {
|
|
if (create)
|
|
return -EIO;
|
|
|
|
/*
|
|
* for reads, we're just trying to fill a partial page.
|
|
* return a hole, they will have to call get_block again
|
|
* before they can fill it, and they will get -EIO at that
|
|
* time
|
|
*/
|
|
return 0;
|
|
}
|
|
bh->b_bdev = inode->i_sb->s_bdev;
|
|
bh->b_blocknr = iblock;
|
|
set_buffer_mapped(bh);
|
|
return 0;
|
|
}
|
|
|
|
int btrfs_csum_data(struct btrfs_root * root, char *data, size_t len,
|
|
char *result)
|
|
{
|
|
struct scatterlist sg;
|
|
struct crypto_hash *tfm = root->fs_info->hash_tfm;
|
|
struct hash_desc desc;
|
|
int ret;
|
|
|
|
desc.tfm = tfm;
|
|
desc.flags = 0;
|
|
sg_init_one(&sg, data, len);
|
|
spin_lock(&root->fs_info->hash_lock);
|
|
ret = crypto_hash_digest(&desc, &sg, 1, result);
|
|
spin_unlock(&root->fs_info->hash_lock);
|
|
if (ret) {
|
|
printk("sha256 digest failed\n");
|
|
}
|
|
return ret;
|
|
}
|
|
static int csum_tree_block(struct btrfs_root *root, struct buffer_head *bh,
|
|
int verify)
|
|
{
|
|
char result[BTRFS_CSUM_SIZE];
|
|
int ret;
|
|
struct btrfs_node *node;
|
|
|
|
ret = btrfs_csum_data(root, bh->b_data + BTRFS_CSUM_SIZE,
|
|
bh->b_size - BTRFS_CSUM_SIZE, result);
|
|
if (ret)
|
|
return ret;
|
|
if (verify) {
|
|
if (memcmp(bh->b_data, result, BTRFS_CSUM_SIZE)) {
|
|
printk("checksum verify failed on %lu\n",
|
|
bh->b_blocknr);
|
|
return 1;
|
|
}
|
|
} else {
|
|
node = btrfs_buffer_node(bh);
|
|
memcpy(node->header.csum, result, BTRFS_CSUM_SIZE);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int btree_writepage(struct page *page, struct writeback_control *wbc)
|
|
{
|
|
struct buffer_head *bh;
|
|
struct btrfs_root *root = btrfs_sb(page->mapping->host->i_sb);
|
|
struct buffer_head *head;
|
|
if (!page_has_buffers(page)) {
|
|
create_empty_buffers(page, root->fs_info->sb->s_blocksize,
|
|
(1 << BH_Dirty)|(1 << BH_Uptodate));
|
|
}
|
|
head = page_buffers(page);
|
|
bh = head;
|
|
do {
|
|
if (buffer_dirty(bh))
|
|
csum_tree_block(root, bh, 0);
|
|
bh = bh->b_this_page;
|
|
} while (bh != head);
|
|
return block_write_full_page(page, btree_get_block, wbc);
|
|
}
|
|
|
|
static int btree_readpage(struct file * file, struct page * page)
|
|
{
|
|
return block_read_full_page(page, btree_get_block);
|
|
}
|
|
|
|
static struct address_space_operations btree_aops = {
|
|
.readpage = btree_readpage,
|
|
.writepage = btree_writepage,
|
|
.sync_page = block_sync_page,
|
|
};
|
|
|
|
struct buffer_head *read_tree_block(struct btrfs_root *root, u64 blocknr)
|
|
{
|
|
struct buffer_head *bh = NULL;
|
|
|
|
bh = btrfs_find_create_tree_block(root, blocknr);
|
|
if (!bh)
|
|
return bh;
|
|
if (buffer_uptodate(bh))
|
|
goto uptodate;
|
|
lock_buffer(bh);
|
|
if (!buffer_uptodate(bh)) {
|
|
get_bh(bh);
|
|
bh->b_end_io = end_buffer_read_sync;
|
|
submit_bh(READ, bh);
|
|
wait_on_buffer(bh);
|
|
if (!buffer_uptodate(bh))
|
|
goto fail;
|
|
csum_tree_block(root, bh, 1);
|
|
} else {
|
|
unlock_buffer(bh);
|
|
}
|
|
uptodate:
|
|
if (check_tree_block(root, bh))
|
|
BUG();
|
|
return bh;
|
|
fail:
|
|
brelse(bh);
|
|
return NULL;
|
|
}
|
|
|
|
int dirty_tree_block(struct btrfs_trans_handle *trans, struct btrfs_root *root,
|
|
struct buffer_head *buf)
|
|
{
|
|
WARN_ON(atomic_read(&buf->b_count) == 0);
|
|
mark_buffer_dirty(buf);
|
|
return 0;
|
|
}
|
|
|
|
int clean_tree_block(struct btrfs_trans_handle *trans, struct btrfs_root *root,
|
|
struct buffer_head *buf)
|
|
{
|
|
WARN_ON(atomic_read(&buf->b_count) == 0);
|
|
clear_buffer_dirty(buf);
|
|
return 0;
|
|
}
|
|
|
|
static int __setup_root(int blocksize,
|
|
struct btrfs_root *root,
|
|
struct btrfs_fs_info *fs_info,
|
|
u64 objectid)
|
|
{
|
|
root->node = NULL;
|
|
root->commit_root = NULL;
|
|
root->blocksize = blocksize;
|
|
root->ref_cows = 0;
|
|
root->fs_info = fs_info;
|
|
memset(&root->root_key, 0, sizeof(root->root_key));
|
|
memset(&root->root_item, 0, sizeof(root->root_item));
|
|
return 0;
|
|
}
|
|
|
|
static int find_and_setup_root(int blocksize,
|
|
struct btrfs_root *tree_root,
|
|
struct btrfs_fs_info *fs_info,
|
|
u64 objectid,
|
|
struct btrfs_root *root)
|
|
{
|
|
int ret;
|
|
|
|
__setup_root(blocksize, root, fs_info, objectid);
|
|
ret = btrfs_find_last_root(tree_root, objectid,
|
|
&root->root_item, &root->root_key);
|
|
BUG_ON(ret);
|
|
|
|
root->node = read_tree_block(root,
|
|
btrfs_root_blocknr(&root->root_item));
|
|
BUG_ON(!root->node);
|
|
return 0;
|
|
}
|
|
|
|
struct btrfs_root *open_ctree(struct super_block *sb)
|
|
{
|
|
struct btrfs_root *root = kmalloc(sizeof(struct btrfs_root),
|
|
GFP_NOFS);
|
|
struct btrfs_root *extent_root = kmalloc(sizeof(struct btrfs_root),
|
|
GFP_NOFS);
|
|
struct btrfs_root *tree_root = kmalloc(sizeof(struct btrfs_root),
|
|
GFP_NOFS);
|
|
struct btrfs_root *inode_root = kmalloc(sizeof(struct btrfs_root),
|
|
GFP_NOFS);
|
|
struct btrfs_fs_info *fs_info = kmalloc(sizeof(*fs_info),
|
|
GFP_NOFS);
|
|
int ret;
|
|
struct btrfs_super_block *disk_super;
|
|
|
|
init_bit_radix(&fs_info->pinned_radix);
|
|
init_bit_radix(&fs_info->pending_del_radix);
|
|
sb_set_blocksize(sb, 4096);
|
|
fs_info->running_transaction = NULL;
|
|
fs_info->fs_root = root;
|
|
fs_info->tree_root = tree_root;
|
|
fs_info->extent_root = extent_root;
|
|
fs_info->inode_root = inode_root;
|
|
fs_info->last_inode_alloc = 0;
|
|
fs_info->highest_inode = 0;
|
|
fs_info->sb = sb;
|
|
fs_info->btree_inode = new_inode(sb);
|
|
fs_info->btree_inode->i_ino = 1;
|
|
fs_info->btree_inode->i_nlink = 1;
|
|
fs_info->btree_inode->i_size = sb->s_bdev->bd_inode->i_size;
|
|
fs_info->btree_inode->i_mapping->a_ops = &btree_aops;
|
|
insert_inode_hash(fs_info->btree_inode);
|
|
mapping_set_gfp_mask(fs_info->btree_inode->i_mapping, GFP_NOFS);
|
|
fs_info->hash_tfm = crypto_alloc_hash("sha256", 0, CRYPTO_ALG_ASYNC);
|
|
spin_lock_init(&fs_info->hash_lock);
|
|
if (!fs_info->hash_tfm || IS_ERR(fs_info->hash_tfm)) {
|
|
printk("failed to allocate sha256 hash\n");
|
|
return NULL;
|
|
}
|
|
mutex_init(&fs_info->trans_mutex);
|
|
mutex_init(&fs_info->fs_mutex);
|
|
memset(&fs_info->current_insert, 0, sizeof(fs_info->current_insert));
|
|
memset(&fs_info->last_insert, 0, sizeof(fs_info->last_insert));
|
|
|
|
__setup_root(sb->s_blocksize, tree_root,
|
|
fs_info, BTRFS_ROOT_TREE_OBJECTID);
|
|
fs_info->sb_buffer = read_tree_block(tree_root,
|
|
BTRFS_SUPER_INFO_OFFSET /
|
|
sb->s_blocksize);
|
|
|
|
if (!fs_info->sb_buffer) {
|
|
printk("failed2\n");
|
|
return NULL;
|
|
}
|
|
disk_super = (struct btrfs_super_block *)fs_info->sb_buffer->b_data;
|
|
if (!btrfs_super_root(disk_super)) {
|
|
return NULL;
|
|
}
|
|
fs_info->disk_super = disk_super;
|
|
tree_root->node = read_tree_block(tree_root,
|
|
btrfs_super_root(disk_super));
|
|
BUG_ON(!tree_root->node);
|
|
|
|
mutex_lock(&fs_info->fs_mutex);
|
|
ret = find_and_setup_root(sb->s_blocksize, tree_root, fs_info,
|
|
BTRFS_EXTENT_TREE_OBJECTID, extent_root);
|
|
BUG_ON(ret);
|
|
|
|
ret = find_and_setup_root(sb->s_blocksize, tree_root, fs_info,
|
|
BTRFS_INODE_MAP_OBJECTID, inode_root);
|
|
BUG_ON(ret);
|
|
|
|
ret = find_and_setup_root(sb->s_blocksize, tree_root, fs_info,
|
|
BTRFS_FS_TREE_OBJECTID, root);
|
|
BUG_ON(ret);
|
|
root->commit_root = root->node;
|
|
get_bh(root->node);
|
|
root->ref_cows = 1;
|
|
root->fs_info->generation = root->root_key.offset + 1;
|
|
ret = btrfs_find_highest_inode(root, &root->fs_info->last_inode_alloc);
|
|
if (ret == 0)
|
|
fs_info->highest_inode = fs_info->last_inode_alloc;
|
|
mutex_unlock(&fs_info->fs_mutex);
|
|
return root;
|
|
}
|
|
|
|
int write_ctree_super(struct btrfs_trans_handle *trans, struct btrfs_root
|
|
*root)
|
|
{
|
|
struct buffer_head *bh = root->fs_info->sb_buffer;
|
|
|
|
btrfs_set_super_root(root->fs_info->disk_super,
|
|
root->fs_info->tree_root->node->b_blocknr);
|
|
lock_buffer(bh);
|
|
WARN_ON(atomic_read(&bh->b_count) < 1);
|
|
clear_buffer_dirty(bh);
|
|
csum_tree_block(root, bh, 0);
|
|
bh->b_end_io = end_buffer_write_sync;
|
|
get_bh(bh);
|
|
submit_bh(WRITE, bh);
|
|
wait_on_buffer(bh);
|
|
if (!buffer_uptodate(bh)) {
|
|
WARN_ON(1);
|
|
return -EIO;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
int close_ctree(struct btrfs_root *root)
|
|
{
|
|
int ret;
|
|
struct btrfs_trans_handle *trans;
|
|
|
|
mutex_lock(&root->fs_info->fs_mutex);
|
|
trans = btrfs_start_transaction(root, 1);
|
|
btrfs_commit_transaction(trans, root);
|
|
/* run commit again to drop the original snapshot */
|
|
trans = btrfs_start_transaction(root, 1);
|
|
btrfs_commit_transaction(trans, root);
|
|
ret = btrfs_write_and_wait_transaction(NULL, root);
|
|
BUG_ON(ret);
|
|
write_ctree_super(NULL, root);
|
|
mutex_unlock(&root->fs_info->fs_mutex);
|
|
|
|
if (root->node)
|
|
btrfs_block_release(root, root->node);
|
|
if (root->fs_info->extent_root->node)
|
|
btrfs_block_release(root->fs_info->extent_root,
|
|
root->fs_info->extent_root->node);
|
|
if (root->fs_info->inode_root->node)
|
|
btrfs_block_release(root->fs_info->inode_root,
|
|
root->fs_info->inode_root->node);
|
|
if (root->fs_info->tree_root->node)
|
|
btrfs_block_release(root->fs_info->tree_root,
|
|
root->fs_info->tree_root->node);
|
|
btrfs_block_release(root, root->commit_root);
|
|
btrfs_block_release(root, root->fs_info->sb_buffer);
|
|
crypto_free_hash(root->fs_info->hash_tfm);
|
|
truncate_inode_pages(root->fs_info->btree_inode->i_mapping, 0);
|
|
iput(root->fs_info->btree_inode);
|
|
kfree(root->fs_info->extent_root);
|
|
kfree(root->fs_info->inode_root);
|
|
kfree(root->fs_info->tree_root);
|
|
kfree(root->fs_info);
|
|
kfree(root);
|
|
return 0;
|
|
}
|
|
|
|
void btrfs_block_release(struct btrfs_root *root, struct buffer_head *buf)
|
|
{
|
|
brelse(buf);
|
|
}
|
|
|