diff options
Diffstat (limited to 'fs/btrfs/ref-cache.c')
-rw-r--r-- | fs/btrfs/ref-cache.c | 230 |
1 files changed, 230 insertions, 0 deletions
diff --git a/fs/btrfs/ref-cache.c b/fs/btrfs/ref-cache.c new file mode 100644 index 000000000000..6f0acc4c9eab --- /dev/null +++ b/fs/btrfs/ref-cache.c | |||
@@ -0,0 +1,230 @@ | |||
1 | /* | ||
2 | * Copyright (C) 2008 Oracle. All rights reserved. | ||
3 | * | ||
4 | * This program is free software; you can redistribute it and/or | ||
5 | * modify it under the terms of the GNU General Public | ||
6 | * License v2 as published by the Free Software Foundation. | ||
7 | * | ||
8 | * This program is distributed in the hope that it will be useful, | ||
9 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | ||
10 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | ||
11 | * General Public License for more details. | ||
12 | * | ||
13 | * You should have received a copy of the GNU General Public | ||
14 | * License along with this program; if not, write to the | ||
15 | * Free Software Foundation, Inc., 59 Temple Place - Suite 330, | ||
16 | * Boston, MA 021110-1307, USA. | ||
17 | */ | ||
18 | |||
19 | #include <linux/sched.h> | ||
20 | #include "ctree.h" | ||
21 | #include "ref-cache.h" | ||
22 | #include "transaction.h" | ||
23 | |||
24 | /* | ||
25 | * leaf refs are used to cache the information about which extents | ||
26 | * a given leaf has references on. This allows us to process that leaf | ||
27 | * in btrfs_drop_snapshot without needing to read it back from disk. | ||
28 | */ | ||
29 | |||
30 | /* | ||
31 | * kmalloc a leaf reference struct and update the counters for the | ||
32 | * total ref cache size | ||
33 | */ | ||
34 | struct btrfs_leaf_ref *btrfs_alloc_leaf_ref(struct btrfs_root *root, | ||
35 | int nr_extents) | ||
36 | { | ||
37 | struct btrfs_leaf_ref *ref; | ||
38 | size_t size = btrfs_leaf_ref_size(nr_extents); | ||
39 | |||
40 | ref = kmalloc(size, GFP_NOFS); | ||
41 | if (ref) { | ||
42 | spin_lock(&root->fs_info->ref_cache_lock); | ||
43 | root->fs_info->total_ref_cache_size += size; | ||
44 | spin_unlock(&root->fs_info->ref_cache_lock); | ||
45 | |||
46 | memset(ref, 0, sizeof(*ref)); | ||
47 | atomic_set(&ref->usage, 1); | ||
48 | INIT_LIST_HEAD(&ref->list); | ||
49 | } | ||
50 | return ref; | ||
51 | } | ||
52 | |||
53 | /* | ||
54 | * free a leaf reference struct and update the counters for the | ||
55 | * total ref cache size | ||
56 | */ | ||
57 | void btrfs_free_leaf_ref(struct btrfs_root *root, struct btrfs_leaf_ref *ref) | ||
58 | { | ||
59 | if (!ref) | ||
60 | return; | ||
61 | WARN_ON(atomic_read(&ref->usage) == 0); | ||
62 | if (atomic_dec_and_test(&ref->usage)) { | ||
63 | size_t size = btrfs_leaf_ref_size(ref->nritems); | ||
64 | |||
65 | BUG_ON(ref->in_tree); | ||
66 | kfree(ref); | ||
67 | |||
68 | spin_lock(&root->fs_info->ref_cache_lock); | ||
69 | root->fs_info->total_ref_cache_size -= size; | ||
70 | spin_unlock(&root->fs_info->ref_cache_lock); | ||
71 | } | ||
72 | } | ||
73 | |||
74 | static struct rb_node *tree_insert(struct rb_root *root, u64 bytenr, | ||
75 | struct rb_node *node) | ||
76 | { | ||
77 | struct rb_node **p = &root->rb_node; | ||
78 | struct rb_node *parent = NULL; | ||
79 | struct btrfs_leaf_ref *entry; | ||
80 | |||
81 | while (*p) { | ||
82 | parent = *p; | ||
83 | entry = rb_entry(parent, struct btrfs_leaf_ref, rb_node); | ||
84 | |||
85 | if (bytenr < entry->bytenr) | ||
86 | p = &(*p)->rb_left; | ||
87 | else if (bytenr > entry->bytenr) | ||
88 | p = &(*p)->rb_right; | ||
89 | else | ||
90 | return parent; | ||
91 | } | ||
92 | |||
93 | entry = rb_entry(node, struct btrfs_leaf_ref, rb_node); | ||
94 | rb_link_node(node, parent, p); | ||
95 | rb_insert_color(node, root); | ||
96 | return NULL; | ||
97 | } | ||
98 | |||
99 | static struct rb_node *tree_search(struct rb_root *root, u64 bytenr) | ||
100 | { | ||
101 | struct rb_node *n = root->rb_node; | ||
102 | struct btrfs_leaf_ref *entry; | ||
103 | |||
104 | while (n) { | ||
105 | entry = rb_entry(n, struct btrfs_leaf_ref, rb_node); | ||
106 | WARN_ON(!entry->in_tree); | ||
107 | |||
108 | if (bytenr < entry->bytenr) | ||
109 | n = n->rb_left; | ||
110 | else if (bytenr > entry->bytenr) | ||
111 | n = n->rb_right; | ||
112 | else | ||
113 | return n; | ||
114 | } | ||
115 | return NULL; | ||
116 | } | ||
117 | |||
118 | int btrfs_remove_leaf_refs(struct btrfs_root *root, u64 max_root_gen, | ||
119 | int shared) | ||
120 | { | ||
121 | struct btrfs_leaf_ref *ref = NULL; | ||
122 | struct btrfs_leaf_ref_tree *tree = root->ref_tree; | ||
123 | |||
124 | if (shared) | ||
125 | tree = &root->fs_info->shared_ref_tree; | ||
126 | if (!tree) | ||
127 | return 0; | ||
128 | |||
129 | spin_lock(&tree->lock); | ||
130 | while (!list_empty(&tree->list)) { | ||
131 | ref = list_entry(tree->list.next, struct btrfs_leaf_ref, list); | ||
132 | BUG_ON(ref->tree != tree); | ||
133 | if (ref->root_gen > max_root_gen) | ||
134 | break; | ||
135 | if (!xchg(&ref->in_tree, 0)) { | ||
136 | cond_resched_lock(&tree->lock); | ||
137 | continue; | ||
138 | } | ||
139 | |||
140 | rb_erase(&ref->rb_node, &tree->root); | ||
141 | list_del_init(&ref->list); | ||
142 | |||
143 | spin_unlock(&tree->lock); | ||
144 | btrfs_free_leaf_ref(root, ref); | ||
145 | cond_resched(); | ||
146 | spin_lock(&tree->lock); | ||
147 | } | ||
148 | spin_unlock(&tree->lock); | ||
149 | return 0; | ||
150 | } | ||
151 | |||
152 | /* | ||
153 | * find the leaf ref for a given extent. This returns the ref struct with | ||
154 | * a usage reference incremented | ||
155 | */ | ||
156 | struct btrfs_leaf_ref *btrfs_lookup_leaf_ref(struct btrfs_root *root, | ||
157 | u64 bytenr) | ||
158 | { | ||
159 | struct rb_node *rb; | ||
160 | struct btrfs_leaf_ref *ref = NULL; | ||
161 | struct btrfs_leaf_ref_tree *tree = root->ref_tree; | ||
162 | again: | ||
163 | if (tree) { | ||
164 | spin_lock(&tree->lock); | ||
165 | rb = tree_search(&tree->root, bytenr); | ||
166 | if (rb) | ||
167 | ref = rb_entry(rb, struct btrfs_leaf_ref, rb_node); | ||
168 | if (ref) | ||
169 | atomic_inc(&ref->usage); | ||
170 | spin_unlock(&tree->lock); | ||
171 | if (ref) | ||
172 | return ref; | ||
173 | } | ||
174 | if (tree != &root->fs_info->shared_ref_tree) { | ||
175 | tree = &root->fs_info->shared_ref_tree; | ||
176 | goto again; | ||
177 | } | ||
178 | return NULL; | ||
179 | } | ||
180 | |||
181 | /* | ||
182 | * add a fully filled in leaf ref struct | ||
183 | * remove all the refs older than a given root generation | ||
184 | */ | ||
185 | int btrfs_add_leaf_ref(struct btrfs_root *root, struct btrfs_leaf_ref *ref, | ||
186 | int shared) | ||
187 | { | ||
188 | int ret = 0; | ||
189 | struct rb_node *rb; | ||
190 | struct btrfs_leaf_ref_tree *tree = root->ref_tree; | ||
191 | |||
192 | if (shared) | ||
193 | tree = &root->fs_info->shared_ref_tree; | ||
194 | |||
195 | spin_lock(&tree->lock); | ||
196 | rb = tree_insert(&tree->root, ref->bytenr, &ref->rb_node); | ||
197 | if (rb) { | ||
198 | ret = -EEXIST; | ||
199 | } else { | ||
200 | atomic_inc(&ref->usage); | ||
201 | ref->tree = tree; | ||
202 | ref->in_tree = 1; | ||
203 | list_add_tail(&ref->list, &tree->list); | ||
204 | } | ||
205 | spin_unlock(&tree->lock); | ||
206 | return ret; | ||
207 | } | ||
208 | |||
209 | /* | ||
210 | * remove a single leaf ref from the tree. This drops the ref held by the tree | ||
211 | * only | ||
212 | */ | ||
213 | int btrfs_remove_leaf_ref(struct btrfs_root *root, struct btrfs_leaf_ref *ref) | ||
214 | { | ||
215 | struct btrfs_leaf_ref_tree *tree; | ||
216 | |||
217 | if (!xchg(&ref->in_tree, 0)) | ||
218 | return 0; | ||
219 | |||
220 | tree = ref->tree; | ||
221 | spin_lock(&tree->lock); | ||
222 | |||
223 | rb_erase(&ref->rb_node, &tree->root); | ||
224 | list_del_init(&ref->list); | ||
225 | |||
226 | spin_unlock(&tree->lock); | ||
227 | |||
228 | btrfs_free_leaf_ref(root, ref); | ||
229 | return 0; | ||
230 | } | ||