232 lines
5.5 KiB
C
232 lines
5.5 KiB
C
|
/*
|
||
|
* Copyright (C) 2008 Oracle. All rights reserved.
|
||
|
*
|
||
|
* This program is free software; you can redistribute it and/or
|
||
|
* modify it under the terms of the GNU General Public
|
||
|
* License v2 as published by the Free Software Foundation.
|
||
|
*
|
||
|
* This program is distributed in the hope that it will be useful,
|
||
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
||
|
* General Public License for more details.
|
||
|
*
|
||
|
* You should have received a copy of the GNU General Public
|
||
|
* License along with this program; if not, write to the
|
||
|
* Free Software Foundation, Inc., 59 Temple Place - Suite 330,
|
||
|
* Boston, MA 021110-1307, USA.
|
||
|
*/
|
||
|
|
||
|
#include <linux/sched.h>
|
||
|
#include <linux/sort.h>
|
||
|
#include "ctree.h"
|
||
|
#include "ref-cache.h"
|
||
|
#include "transaction.h"
|
||
|
|
||
|
/*
|
||
|
* leaf refs are used to cache the information about which extents
|
||
|
* a given leaf has references on. This allows us to process that leaf
|
||
|
* in btrfs_drop_snapshot without needing to read it back from disk.
|
||
|
*/
|
||
|
|
||
|
/*
|
||
|
* kmalloc a leaf reference struct and update the counters for the
|
||
|
* total ref cache size
|
||
|
*/
|
||
|
struct btrfs_leaf_ref *btrfs_alloc_leaf_ref(struct btrfs_root *root,
|
||
|
int nr_extents)
|
||
|
{
|
||
|
struct btrfs_leaf_ref *ref;
|
||
|
size_t size = btrfs_leaf_ref_size(nr_extents);
|
||
|
|
||
|
ref = kmalloc(size, GFP_NOFS);
|
||
|
if (ref) {
|
||
|
spin_lock(&root->fs_info->ref_cache_lock);
|
||
|
root->fs_info->total_ref_cache_size += size;
|
||
|
spin_unlock(&root->fs_info->ref_cache_lock);
|
||
|
|
||
|
memset(ref, 0, sizeof(*ref));
|
||
|
atomic_set(&ref->usage, 1);
|
||
|
INIT_LIST_HEAD(&ref->list);
|
||
|
}
|
||
|
return ref;
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* free a leaf reference struct and update the counters for the
|
||
|
* total ref cache size
|
||
|
*/
|
||
|
void btrfs_free_leaf_ref(struct btrfs_root *root, struct btrfs_leaf_ref *ref)
|
||
|
{
|
||
|
if (!ref)
|
||
|
return;
|
||
|
WARN_ON(atomic_read(&ref->usage) == 0);
|
||
|
if (atomic_dec_and_test(&ref->usage)) {
|
||
|
size_t size = btrfs_leaf_ref_size(ref->nritems);
|
||
|
|
||
|
BUG_ON(ref->in_tree);
|
||
|
kfree(ref);
|
||
|
|
||
|
spin_lock(&root->fs_info->ref_cache_lock);
|
||
|
root->fs_info->total_ref_cache_size -= size;
|
||
|
spin_unlock(&root->fs_info->ref_cache_lock);
|
||
|
}
|
||
|
}
|
||
|
|
||
|
static struct rb_node *tree_insert(struct rb_root *root, u64 bytenr,
|
||
|
struct rb_node *node)
|
||
|
{
|
||
|
struct rb_node **p = &root->rb_node;
|
||
|
struct rb_node *parent = NULL;
|
||
|
struct btrfs_leaf_ref *entry;
|
||
|
|
||
|
while (*p) {
|
||
|
parent = *p;
|
||
|
entry = rb_entry(parent, struct btrfs_leaf_ref, rb_node);
|
||
|
|
||
|
if (bytenr < entry->bytenr)
|
||
|
p = &(*p)->rb_left;
|
||
|
else if (bytenr > entry->bytenr)
|
||
|
p = &(*p)->rb_right;
|
||
|
else
|
||
|
return parent;
|
||
|
}
|
||
|
|
||
|
entry = rb_entry(node, struct btrfs_leaf_ref, rb_node);
|
||
|
rb_link_node(node, parent, p);
|
||
|
rb_insert_color(node, root);
|
||
|
return NULL;
|
||
|
}
|
||
|
|
||
|
static struct rb_node *tree_search(struct rb_root *root, u64 bytenr)
|
||
|
{
|
||
|
struct rb_node *n = root->rb_node;
|
||
|
struct btrfs_leaf_ref *entry;
|
||
|
|
||
|
while (n) {
|
||
|
entry = rb_entry(n, struct btrfs_leaf_ref, rb_node);
|
||
|
WARN_ON(!entry->in_tree);
|
||
|
|
||
|
if (bytenr < entry->bytenr)
|
||
|
n = n->rb_left;
|
||
|
else if (bytenr > entry->bytenr)
|
||
|
n = n->rb_right;
|
||
|
else
|
||
|
return n;
|
||
|
}
|
||
|
return NULL;
|
||
|
}
|
||
|
|
||
|
int btrfs_remove_leaf_refs(struct btrfs_root *root, u64 max_root_gen,
|
||
|
int shared)
|
||
|
{
|
||
|
struct btrfs_leaf_ref *ref = NULL;
|
||
|
struct btrfs_leaf_ref_tree *tree = root->ref_tree;
|
||
|
|
||
|
if (shared)
|
||
|
tree = &root->fs_info->shared_ref_tree;
|
||
|
if (!tree)
|
||
|
return 0;
|
||
|
|
||
|
spin_lock(&tree->lock);
|
||
|
while (!list_empty(&tree->list)) {
|
||
|
ref = list_entry(tree->list.next, struct btrfs_leaf_ref, list);
|
||
|
BUG_ON(ref->tree != tree);
|
||
|
if (ref->root_gen > max_root_gen)
|
||
|
break;
|
||
|
if (!xchg(&ref->in_tree, 0)) {
|
||
|
cond_resched_lock(&tree->lock);
|
||
|
continue;
|
||
|
}
|
||
|
|
||
|
rb_erase(&ref->rb_node, &tree->root);
|
||
|
list_del_init(&ref->list);
|
||
|
|
||
|
spin_unlock(&tree->lock);
|
||
|
btrfs_free_leaf_ref(root, ref);
|
||
|
cond_resched();
|
||
|
spin_lock(&tree->lock);
|
||
|
}
|
||
|
spin_unlock(&tree->lock);
|
||
|
return 0;
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* find the leaf ref for a given extent. This returns the ref struct with
|
||
|
* a usage reference incremented
|
||
|
*/
|
||
|
struct btrfs_leaf_ref *btrfs_lookup_leaf_ref(struct btrfs_root *root,
|
||
|
u64 bytenr)
|
||
|
{
|
||
|
struct rb_node *rb;
|
||
|
struct btrfs_leaf_ref *ref = NULL;
|
||
|
struct btrfs_leaf_ref_tree *tree = root->ref_tree;
|
||
|
again:
|
||
|
if (tree) {
|
||
|
spin_lock(&tree->lock);
|
||
|
rb = tree_search(&tree->root, bytenr);
|
||
|
if (rb)
|
||
|
ref = rb_entry(rb, struct btrfs_leaf_ref, rb_node);
|
||
|
if (ref)
|
||
|
atomic_inc(&ref->usage);
|
||
|
spin_unlock(&tree->lock);
|
||
|
if (ref)
|
||
|
return ref;
|
||
|
}
|
||
|
if (tree != &root->fs_info->shared_ref_tree) {
|
||
|
tree = &root->fs_info->shared_ref_tree;
|
||
|
goto again;
|
||
|
}
|
||
|
return NULL;
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* add a fully filled in leaf ref struct
|
||
|
* remove all the refs older than a given root generation
|
||
|
*/
|
||
|
int btrfs_add_leaf_ref(struct btrfs_root *root, struct btrfs_leaf_ref *ref,
|
||
|
int shared)
|
||
|
{
|
||
|
int ret = 0;
|
||
|
struct rb_node *rb;
|
||
|
struct btrfs_leaf_ref_tree *tree = root->ref_tree;
|
||
|
|
||
|
if (shared)
|
||
|
tree = &root->fs_info->shared_ref_tree;
|
||
|
|
||
|
spin_lock(&tree->lock);
|
||
|
rb = tree_insert(&tree->root, ref->bytenr, &ref->rb_node);
|
||
|
if (rb) {
|
||
|
ret = -EEXIST;
|
||
|
} else {
|
||
|
atomic_inc(&ref->usage);
|
||
|
ref->tree = tree;
|
||
|
ref->in_tree = 1;
|
||
|
list_add_tail(&ref->list, &tree->list);
|
||
|
}
|
||
|
spin_unlock(&tree->lock);
|
||
|
return ret;
|
||
|
}
|
||
|
|
||
|
/*
|
||
|
* remove a single leaf ref from the tree. This drops the ref held by the tree
|
||
|
* only
|
||
|
*/
|
||
|
int btrfs_remove_leaf_ref(struct btrfs_root *root, struct btrfs_leaf_ref *ref)
|
||
|
{
|
||
|
struct btrfs_leaf_ref_tree *tree;
|
||
|
|
||
|
if (!xchg(&ref->in_tree, 0))
|
||
|
return 0;
|
||
|
|
||
|
tree = ref->tree;
|
||
|
spin_lock(&tree->lock);
|
||
|
|
||
|
rb_erase(&ref->rb_node, &tree->root);
|
||
|
list_del_init(&ref->list);
|
||
|
|
||
|
spin_unlock(&tree->lock);
|
||
|
|
||
|
btrfs_free_leaf_ref(root, ref);
|
||
|
return 0;
|
||
|
}
|