-#define _XOPEN_SOURCE 500
-#include <stdio.h>
-#include <stdlib.h>
-#include <sys/types.h>
-#include <sys/stat.h>
-#include <fcntl.h>
-#include <unistd.h>
-#include "kerncompat.h"
-#include "radix-tree.h"
+#include <linux/module.h>
+#include <linux/fs.h>
+#include <linux/blkdev.h>
+#include <linux/crypto.h>
+#include <linux/scatterlist.h>
+#include <linux/swap.h>
+#include <linux/radix-tree.h>
#include "ctree.h"
#include "disk-io.h"
+#include "transaction.h"
+#include "btrfs_inode.h"
-static int allocated_blocks = 0;
+static int check_tree_block(struct btrfs_root *root, struct buffer_head *buf)
+{
+ struct btrfs_node *node = btrfs_buffer_node(buf);
+ if (buf->b_blocknr != btrfs_header_blocknr(&node->header)) {
+ BUG();
+ }
+ return 0;
+}
-static int get_free_block(struct ctree_root *root, u64 *block)
+struct buffer_head *btrfs_find_tree_block(struct btrfs_root *root, u64 blocknr)
{
- struct stat st;
- int ret = 0;
+ struct address_space *mapping = root->fs_info->btree_inode->i_mapping;
+ int blockbits = root->fs_info->sb->s_blocksize_bits;
+ unsigned long index = blocknr >> (PAGE_CACHE_SHIFT - blockbits);
+ struct page *page;
+ struct buffer_head *bh;
+ struct buffer_head *head;
+ struct buffer_head *ret = NULL;
- if (root->alloc_extent->num_used >= root->alloc_extent->num_blocks)
- return -1;
-
- *block = root->alloc_extent->blocknr + root->alloc_extent->num_used;
- root->alloc_extent->num_used += 1;
- if (root->alloc_extent->num_used >= root->alloc_extent->num_blocks) {
- struct alloc_extent *ae = root->alloc_extent;
- root->alloc_extent = root->reserve_extent;
- root->reserve_extent = ae;
- ae->num_blocks = 0;
- }
- st.st_size = 0;
- ret = fstat(root->fp, &st);
- if (st.st_size < (*block + 1) * CTREE_BLOCKSIZE) {
- ret = ftruncate(root->fp,
- (*block + 1) * CTREE_BLOCKSIZE);
- if (ret) {
- perror("ftruncate");
- exit(1);
+
+ page = find_lock_page(mapping, index);
+ if (!page)
+ return NULL;
+
+ if (!page_has_buffers(page))
+ goto out_unlock;
+
+ head = page_buffers(page);
+ bh = head;
+ do {
+ if (buffer_mapped(bh) && bh->b_blocknr == blocknr) {
+ ret = bh;
+ get_bh(bh);
+ goto out_unlock;
}
+ bh = bh->b_this_page;
+ } while (bh != head);
+out_unlock:
+ unlock_page(page);
+ if (ret) {
+ touch_buffer(ret);
}
+ page_cache_release(page);
return ret;
}
-struct tree_buffer *alloc_tree_block(struct ctree_root *root, u64 blocknr)
+struct buffer_head *btrfs_find_create_tree_block(struct btrfs_root *root,
+ u64 blocknr)
{
- struct tree_buffer *buf;
- int ret;
- buf = malloc(sizeof(struct tree_buffer));
- if (!buf)
- return buf;
- allocated_blocks++;
- buf->blocknr = blocknr;
- buf->count = 1;
- radix_tree_preload(GFP_KERNEL);
- ret = radix_tree_insert(&root->cache_radix, blocknr, buf);
- radix_tree_preload_end();
- if (ret) {
- free(buf);
+ struct address_space *mapping = root->fs_info->btree_inode->i_mapping;
+ int blockbits = root->fs_info->sb->s_blocksize_bits;
+ unsigned long index = blocknr >> (PAGE_CACHE_SHIFT - blockbits);
+ struct page *page;
+ struct buffer_head *bh;
+ struct buffer_head *head;
+ struct buffer_head *ret = NULL;
+ u64 first_block = index << (PAGE_CACHE_SHIFT - blockbits);
+
+ page = grab_cache_page(mapping, index);
+ if (!page)
return NULL;
+
+ if (!page_has_buffers(page))
+ create_empty_buffers(page, root->fs_info->sb->s_blocksize, 0);
+ head = page_buffers(page);
+ bh = head;
+ do {
+ if (!buffer_mapped(bh)) {
+ bh->b_bdev = root->fs_info->sb->s_bdev;
+ bh->b_blocknr = first_block;
+ set_buffer_mapped(bh);
+ }
+ if (bh->b_blocknr == blocknr) {
+ ret = bh;
+ get_bh(bh);
+ goto out_unlock;
+ }
+ bh = bh->b_this_page;
+ first_block++;
+ } while (bh != head);
+out_unlock:
+ unlock_page(page);
+ if (ret)
+ touch_buffer(ret);
+ page_cache_release(page);
+ return ret;
+}
+
+static sector_t max_block(struct block_device *bdev)
+{
+ sector_t retval = ~((sector_t)0);
+ loff_t sz = i_size_read(bdev->bd_inode);
+
+ if (sz) {
+ unsigned int size = block_size(bdev);
+ unsigned int sizebits = blksize_bits(size);
+ retval = (sz >> sizebits);
}
- return buf;
+ return retval;
}
-struct tree_buffer *alloc_free_block(struct ctree_root *root)
+static int btree_get_block(struct inode *inode, sector_t iblock,
+ struct buffer_head *bh, int create)
{
- u64 free_block;
+ if (iblock >= max_block(inode->i_sb->s_bdev)) {
+ if (create)
+ return -EIO;
+
+ /*
+ * for reads, we're just trying to fill a partial page.
+ * return a hole, they will have to call get_block again
+ * before they can fill it, and they will get -EIO at that
+ * time
+ */
+ return 0;
+ }
+ bh->b_bdev = inode->i_sb->s_bdev;
+ bh->b_blocknr = iblock;
+ set_buffer_mapped(bh);
+ return 0;
+}
+
+int btrfs_csum_data(struct btrfs_root * root, char *data, size_t len,
+ char *result)
+{
+ struct scatterlist sg;
+ struct crypto_hash *tfm = root->fs_info->hash_tfm;
+ struct hash_desc desc;
int ret;
- struct tree_buffer * buf;
- ret = get_free_block(root, &free_block);
+
+ desc.tfm = tfm;
+ desc.flags = 0;
+ sg_init_one(&sg, data, len);
+ spin_lock(&root->fs_info->hash_lock);
+ ret = crypto_hash_digest(&desc, &sg, 1, result);
+ spin_unlock(&root->fs_info->hash_lock);
if (ret) {
- BUG();
- return NULL;
+ printk("sha256 digest failed\n");
}
- buf = alloc_tree_block(root, free_block);
- if (!buf)
- BUG();
- return buf;
+ return ret;
}
-
-struct tree_buffer *read_tree_block(struct ctree_root *root, u64 blocknr)
+static int csum_tree_block(struct btrfs_root *root, struct buffer_head *bh,
+ int verify)
{
- loff_t offset = blocknr * CTREE_BLOCKSIZE;
- struct tree_buffer *buf;
+ char result[BTRFS_CSUM_SIZE];
int ret;
+ struct btrfs_node *node;
- buf = radix_tree_lookup(&root->cache_radix, blocknr);
- if (buf) {
- buf->count++;
- goto test;
+ ret = btrfs_csum_data(root, bh->b_data + BTRFS_CSUM_SIZE,
+ bh->b_size - BTRFS_CSUM_SIZE, result);
+ if (ret)
+ return ret;
+ if (verify) {
+ if (memcmp(bh->b_data, result, BTRFS_CSUM_SIZE)) {
+ printk("checksum verify failed on %lu\n",
+ bh->b_blocknr);
+ return 1;
+ }
+ } else {
+ node = btrfs_buffer_node(bh);
+ memcpy(node->header.csum, result, BTRFS_CSUM_SIZE);
}
- buf = alloc_tree_block(root, blocknr);
- if (!buf)
- return NULL;
- ret = pread(root->fp, &buf->node, CTREE_BLOCKSIZE, offset);
- if (ret != CTREE_BLOCKSIZE) {
- free(buf);
- return NULL;
+ return 0;
+}
+
+static int btree_writepage(struct page *page, struct writeback_control *wbc)
+{
+ struct buffer_head *bh;
+ struct btrfs_root *root = BTRFS_I(page->mapping->host)->root;
+ struct buffer_head *head;
+ if (!page_has_buffers(page)) {
+ create_empty_buffers(page, root->fs_info->sb->s_blocksize,
+ (1 << BH_Dirty)|(1 << BH_Uptodate));
}
-test:
- if (buf->blocknr != buf->node.header.blocknr)
- BUG();
- if (root->node && buf->node.header.parentid != root->node->node.header.parentid)
- BUG();
- return buf;
+ head = page_buffers(page);
+ bh = head;
+ do {
+ if (buffer_dirty(bh))
+ csum_tree_block(root, bh, 0);
+ bh = bh->b_this_page;
+ } while (bh != head);
+ return block_write_full_page(page, btree_get_block, wbc);
}
-int write_tree_block(struct ctree_root *root, struct tree_buffer *buf)
+static int btree_readpage(struct file * file, struct page * page)
{
- u64 blocknr = buf->blocknr;
- loff_t offset = blocknr * CTREE_BLOCKSIZE;
- int ret;
+ return block_read_full_page(page, btree_get_block);
+}
+
+static struct address_space_operations btree_aops = {
+ .readpage = btree_readpage,
+ .writepage = btree_writepage,
+ .sync_page = block_sync_page,
+};
- if (buf->blocknr != buf->node.header.blocknr)
+struct buffer_head *read_tree_block(struct btrfs_root *root, u64 blocknr)
+{
+ struct buffer_head *bh = NULL;
+
+ bh = btrfs_find_create_tree_block(root, blocknr);
+ if (!bh)
+ return bh;
+ if (buffer_uptodate(bh))
+ goto uptodate;
+ lock_buffer(bh);
+ if (!buffer_uptodate(bh)) {
+ get_bh(bh);
+ bh->b_end_io = end_buffer_read_sync;
+ submit_bh(READ, bh);
+ wait_on_buffer(bh);
+ if (!buffer_uptodate(bh))
+ goto fail;
+ csum_tree_block(root, bh, 1);
+ } else {
+ unlock_buffer(bh);
+ }
+uptodate:
+ if (check_tree_block(root, bh))
BUG();
- ret = pwrite(root->fp, &buf->node, CTREE_BLOCKSIZE, offset);
- if (ret != CTREE_BLOCKSIZE)
- return ret;
+ return bh;
+fail:
+ brelse(bh);
+ return NULL;
+}
+
+int dirty_tree_block(struct btrfs_trans_handle *trans, struct btrfs_root *root,
+ struct buffer_head *buf)
+{
+ WARN_ON(atomic_read(&buf->b_count) == 0);
+ mark_buffer_dirty(buf);
return 0;
}
-static int __setup_root(struct ctree_root *root, struct ctree_root *extent_root,
- struct ctree_root_info *info, int fp)
+int clean_tree_block(struct btrfs_trans_handle *trans, struct btrfs_root *root,
+ struct buffer_head *buf)
+{
+ WARN_ON(atomic_read(&buf->b_count) == 0);
+ clear_buffer_dirty(buf);
+ return 0;
+}
+
+static int __setup_root(int blocksize,
+ struct btrfs_root *root,
+ struct btrfs_fs_info *fs_info,
+ u64 objectid)
{
- INIT_RADIX_TREE(&root->cache_radix, GFP_KERNEL);
- root->fp = fp;
root->node = NULL;
- root->node = read_tree_block(root, info->tree_root);
- root->extent_root = extent_root;
- memcpy(&root->ai1, &info->alloc_extent, sizeof(info->alloc_extent));
- memcpy(&root->ai2, &info->reserve_extent, sizeof(info->reserve_extent));
- root->alloc_extent = &root->ai1;
- root->reserve_extent = &root->ai2;
- printf("setup done reading root %p, used %lu available %lu\n", root, root->alloc_extent->num_used, root->alloc_extent->num_blocks);
- printf("setup done reading root %p, reserve used %lu available %lu\n", root, root->reserve_extent->num_used, root->reserve_extent->num_blocks);
+ root->inode = NULL;
+ root->commit_root = NULL;
+ root->blocksize = blocksize;
+ root->ref_cows = 0;
+ root->fs_info = fs_info;
+ root->objectid = objectid;
+ root->last_trans = 0;
+ memset(&root->root_key, 0, sizeof(root->root_key));
+ memset(&root->root_item, 0, sizeof(root->root_item));
return 0;
}
-struct ctree_root *open_ctree(char *filename, struct ctree_super_block *super)
+static int find_and_setup_root(int blocksize,
+ struct btrfs_root *tree_root,
+ struct btrfs_fs_info *fs_info,
+ u64 objectid,
+ struct btrfs_root *root)
{
- struct ctree_root *root = malloc(sizeof(struct ctree_root));
- struct ctree_root *extent_root = malloc(sizeof(struct ctree_root));
- int fp;
int ret;
- fp = open(filename, O_CREAT | O_RDWR);
- if (fp < 0) {
- free(root);
- return NULL;
+ __setup_root(blocksize, root, fs_info, objectid);
+ ret = btrfs_find_last_root(tree_root, objectid,
+ &root->root_item, &root->root_key);
+ BUG_ON(ret);
+
+ root->node = read_tree_block(root,
+ btrfs_root_blocknr(&root->root_item));
+ BUG_ON(!root->node);
+ return 0;
+}
+
+struct btrfs_root *btrfs_read_fs_root(struct btrfs_fs_info *fs_info,
+ struct btrfs_key *location)
+{
+ struct btrfs_root *root;
+ struct btrfs_root *tree_root = fs_info->tree_root;
+ struct btrfs_path *path;
+ struct btrfs_leaf *l;
+ int ret = 0;
+
+printk("read_fs_root looking for %Lu %Lu %u\n", location->objectid, location->offset, location->flags);
+ root = kmalloc(sizeof(*root), GFP_NOFS);
+ if (!root) {
+printk("failed1\n");
+ return ERR_PTR(-ENOMEM);
+ }
+ if (location->offset == (u64)-1) {
+ ret = find_and_setup_root(fs_info->sb->s_blocksize,
+ fs_info->tree_root, fs_info,
+ location->objectid, root);
+ if (ret) {
+printk("failed2\n");
+ kfree(root);
+ return ERR_PTR(ret);
+ }
+ goto insert;
+ }
+
+ __setup_root(fs_info->sb->s_blocksize, root, fs_info,
+ location->objectid);
+
+ path = btrfs_alloc_path();
+ BUG_ON(!path);
+ ret = btrfs_search_slot(NULL, tree_root, location, path, 0, 0);
+ if (ret != 0) {
+printk("internal search_slot gives us %d\n", ret);
+ if (ret > 0)
+ ret = -ENOENT;
+ goto out;
+ }
+ l = btrfs_buffer_leaf(path->nodes[0]);
+ memcpy(&root->root_item,
+ btrfs_item_ptr(l, path->slots[0], struct btrfs_root_item),
+ sizeof(root->root_item));
+ memcpy(&root->root_key, location, sizeof(*location));
+ ret = 0;
+out:
+ btrfs_release_path(root, path);
+ btrfs_free_path(path);
+ if (ret) {
+ kfree(root);
+ return ERR_PTR(ret);
+ }
+ root->node = read_tree_block(root,
+ btrfs_root_blocknr(&root->root_item));
+ BUG_ON(!root->node);
+insert:
+printk("inserting %p\n", root);
+ root->ref_cows = 1;
+ ret = radix_tree_insert(&fs_info->fs_roots_radix, (unsigned long)root,
+ root);
+ if (ret) {
+printk("radix_tree_insert gives us %d\n", ret);
+ brelse(root->node);
+ kfree(root);
+ return ERR_PTR(ret);
}
- ret = pread(fp, super, sizeof(struct ctree_super_block),
- CTREE_SUPER_INFO_OFFSET(CTREE_BLOCKSIZE));
- if (ret == 0 || super->root_info.tree_root == 0) {
- printf("making new FS!\n");
- ret = mkfs(fp);
- if (ret)
- return NULL;
- ret = pread(fp, super, sizeof(struct ctree_super_block),
- CTREE_SUPER_INFO_OFFSET(CTREE_BLOCKSIZE));
- if (ret != sizeof(struct ctree_super_block))
- return NULL;
- }
- BUG_ON(ret < 0);
- __setup_root(root, extent_root, &super->root_info, fp);
- __setup_root(extent_root, extent_root, &super->extent_info, fp);
+printk("all worked\n");
return root;
}
-static int __update_root(struct ctree_root *root, struct ctree_root_info *info)
+struct btrfs_root *open_ctree(struct super_block *sb)
{
- info->tree_root = root->node->blocknr;
- memcpy(&info->alloc_extent, root->alloc_extent, sizeof(struct alloc_extent));
- memcpy(&info->reserve_extent, root->reserve_extent, sizeof(struct alloc_extent));
+ struct btrfs_root *extent_root = kmalloc(sizeof(struct btrfs_root),
+ GFP_NOFS);
+ struct btrfs_root *tree_root = kmalloc(sizeof(struct btrfs_root),
+ GFP_NOFS);
+ struct btrfs_root *inode_root = kmalloc(sizeof(struct btrfs_root),
+ GFP_NOFS);
+ struct btrfs_fs_info *fs_info = kmalloc(sizeof(*fs_info),
+ GFP_NOFS);
+ int ret;
+ struct btrfs_super_block *disk_super;
+
+ init_bit_radix(&fs_info->pinned_radix);
+ init_bit_radix(&fs_info->pending_del_radix);
+ INIT_RADIX_TREE(&fs_info->fs_roots_radix, GFP_NOFS);
+ sb_set_blocksize(sb, 4096);
+ fs_info->running_transaction = NULL;
+ fs_info->tree_root = tree_root;
+ fs_info->extent_root = extent_root;
+ fs_info->inode_root = inode_root;
+ fs_info->last_inode_alloc = 0;
+ fs_info->highest_inode = 0;
+ fs_info->sb = sb;
+ fs_info->btree_inode = new_inode(sb);
+ fs_info->btree_inode->i_ino = 1;
+ fs_info->btree_inode->i_nlink = 1;
+ fs_info->btree_inode->i_size = sb->s_bdev->bd_inode->i_size;
+ fs_info->btree_inode->i_mapping->a_ops = &btree_aops;
+ BTRFS_I(fs_info->btree_inode)->root = tree_root;
+ memset(&BTRFS_I(fs_info->btree_inode)->location, 0,
+ sizeof(struct btrfs_key));
+ insert_inode_hash(fs_info->btree_inode);
+ mapping_set_gfp_mask(fs_info->btree_inode->i_mapping, GFP_NOFS);
+ fs_info->hash_tfm = crypto_alloc_hash("sha256", 0, CRYPTO_ALG_ASYNC);
+ spin_lock_init(&fs_info->hash_lock);
+ if (!fs_info->hash_tfm || IS_ERR(fs_info->hash_tfm)) {
+ printk("failed to allocate sha256 hash\n");
+ return NULL;
+ }
+ mutex_init(&fs_info->trans_mutex);
+ mutex_init(&fs_info->fs_mutex);
+ memset(&fs_info->current_insert, 0, sizeof(fs_info->current_insert));
+ memset(&fs_info->last_insert, 0, sizeof(fs_info->last_insert));
+
+ __setup_root(sb->s_blocksize, tree_root,
+ fs_info, BTRFS_ROOT_TREE_OBJECTID);
+ fs_info->sb_buffer = read_tree_block(tree_root,
+ BTRFS_SUPER_INFO_OFFSET /
+ sb->s_blocksize);
+
+ if (!fs_info->sb_buffer)
+ return NULL;
+ disk_super = (struct btrfs_super_block *)fs_info->sb_buffer->b_data;
+ if (!btrfs_super_root(disk_super))
+ return NULL;
+
+ fs_info->disk_super = disk_super;
+ tree_root->node = read_tree_block(tree_root,
+ btrfs_super_root(disk_super));
+ BUG_ON(!tree_root->node);
+
+ mutex_lock(&fs_info->fs_mutex);
+ ret = find_and_setup_root(sb->s_blocksize, tree_root, fs_info,
+ BTRFS_EXTENT_TREE_OBJECTID, extent_root);
+ BUG_ON(ret);
+
+ ret = find_and_setup_root(sb->s_blocksize, tree_root, fs_info,
+ BTRFS_INODE_MAP_OBJECTID, inode_root);
+ BUG_ON(ret);
+
+ fs_info->generation = btrfs_super_generation(disk_super) + 1;
+ ret = btrfs_find_highest_inode(tree_root, &fs_info->last_inode_alloc);
+ if (ret == 0)
+ fs_info->highest_inode = fs_info->last_inode_alloc;
+ memset(&fs_info->kobj, 0, sizeof(fs_info->kobj));
+ kobj_set_kset_s(fs_info, btrfs_subsys);
+ kobject_set_name(&fs_info->kobj, "%s", sb->s_id);
+ kobject_register(&fs_info->kobj);
+ mutex_unlock(&fs_info->fs_mutex);
+ return tree_root;
+}
+
+int write_ctree_super(struct btrfs_trans_handle *trans, struct btrfs_root
+ *root)
+{
+ struct buffer_head *bh = root->fs_info->sb_buffer;
+
+ btrfs_set_super_root(root->fs_info->disk_super,
+ root->fs_info->tree_root->node->b_blocknr);
+ lock_buffer(bh);
+ WARN_ON(atomic_read(&bh->b_count) < 1);
+ clear_buffer_dirty(bh);
+ csum_tree_block(root, bh, 0);
+ bh->b_end_io = end_buffer_write_sync;
+ get_bh(bh);
+ submit_bh(WRITE, bh);
+ wait_on_buffer(bh);
+ if (!buffer_uptodate(bh)) {
+ WARN_ON(1);
+ return -EIO;
+ }
return 0;
}
-int write_ctree_super(struct ctree_root *root, struct ctree_super_block *s)
+int del_fs_roots(struct btrfs_fs_info *fs_info)
{
int ret;
- __update_root(root, &s->root_info);
- __update_root(root->extent_root, &s->extent_info);
- ret = pwrite(root->fp, s, sizeof(*s), CTREE_SUPER_INFO_OFFSET(CTREE_BLOCKSIZE));
- if (ret != sizeof(*s)) {
- fprintf(stderr, "failed to write new super block err %d\n", ret);
- return ret;
+ struct btrfs_root *gang[8];
+ int i;
+
+ while(1) {
+ ret = radix_tree_gang_lookup(&fs_info->fs_roots_radix,
+ (void **)gang, 0,
+ ARRAY_SIZE(gang));
+ if (!ret)
+ break;
+ for (i = 0; i < ret; i++) {
+ radix_tree_delete(&fs_info->fs_roots_radix,
+ (unsigned long)gang[i]);
+ if (gang[i]->inode)
+ iput(gang[i]->inode);
+ else
+ printk("no inode for root %p\n", gang[i]);
+ if (gang[i]->node)
+ brelse(gang[i]->node);
+ if (gang[i]->commit_root)
+ brelse(gang[i]->commit_root);
+ kfree(gang[i]);
+ }
}
return 0;
}
-int close_ctree(struct ctree_root *root)
+int close_ctree(struct btrfs_root *root)
{
- close(root->fp);
- if (root->node)
- tree_block_release(root, root->node);
- if (root->extent_root->node)
- tree_block_release(root->extent_root, root->extent_root->node);
- free(root);
- printf("on close %d blocks are allocated\n", allocated_blocks);
+ int ret;
+ struct btrfs_trans_handle *trans;
+ struct btrfs_fs_info *fs_info = root->fs_info;
+
+ mutex_lock(&fs_info->fs_mutex);
+ trans = btrfs_start_transaction(root, 1);
+ btrfs_commit_transaction(trans, root);
+ /* run commit again to drop the original snapshot */
+ trans = btrfs_start_transaction(root, 1);
+ btrfs_commit_transaction(trans, root);
+ ret = btrfs_write_and_wait_transaction(NULL, root);
+ BUG_ON(ret);
+ write_ctree_super(NULL, root);
+ mutex_unlock(&fs_info->fs_mutex);
+
+ if (fs_info->extent_root->node)
+ btrfs_block_release(fs_info->extent_root,
+ fs_info->extent_root->node);
+ if (fs_info->inode_root->node)
+ btrfs_block_release(fs_info->inode_root,
+ fs_info->inode_root->node);
+ if (fs_info->tree_root->node)
+ btrfs_block_release(fs_info->tree_root,
+ fs_info->tree_root->node);
+ btrfs_block_release(root, fs_info->sb_buffer);
+ crypto_free_hash(fs_info->hash_tfm);
+ truncate_inode_pages(fs_info->btree_inode->i_mapping, 0);
+ iput(fs_info->btree_inode);
+ del_fs_roots(fs_info);
+ kfree(fs_info->extent_root);
+ kfree(fs_info->inode_root);
+ kfree(fs_info->tree_root);
+ kobject_unregister(&fs_info->kobj);
return 0;
}
-void tree_block_release(struct ctree_root *root, struct tree_buffer *buf)
+void btrfs_block_release(struct btrfs_root *root, struct buffer_head *buf)
{
- buf->count--;
- if (buf->count < 0)
- BUG();
- if (buf->count == 0) {
- if (!radix_tree_lookup(&root->cache_radix, buf->blocknr))
- BUG();
- radix_tree_delete(&root->cache_radix, buf->blocknr);
- memset(buf, 0, sizeof(*buf));
- free(buf);
- BUG_ON(allocated_blocks == 0);
- allocated_blocks--;
- }
+ brelse(buf);
}