diff options
Diffstat (limited to 'fs/notify/inode_mark.c')
-rw-r--r-- | fs/notify/inode_mark.c | 331 |
1 files changed, 103 insertions, 228 deletions
diff --git a/fs/notify/inode_mark.c b/fs/notify/inode_mark.c index 152b83ec005d..33297c005060 100644 --- a/fs/notify/inode_mark.c +++ b/fs/notify/inode_mark.c | |||
@@ -16,72 +16,6 @@ | |||
16 | * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. | 16 | * the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA. |
17 | */ | 17 | */ |
18 | 18 | ||
19 | /* | ||
20 | * fsnotify inode mark locking/lifetime/and refcnting | ||
21 | * | ||
22 | * REFCNT: | ||
23 | * The mark->refcnt tells how many "things" in the kernel currently are | ||
24 | * referencing this object. The object typically will live inside the kernel | ||
25 | * with a refcnt of 2, one for each list it is on (i_list, g_list). Any task | ||
26 | * which can find this object holding the appropriete locks, can take a reference | ||
27 | * and the object itself is guarenteed to survive until the reference is dropped. | ||
28 | * | ||
29 | * LOCKING: | ||
30 | * There are 3 spinlocks involved with fsnotify inode marks and they MUST | ||
31 | * be taken in order as follows: | ||
32 | * | ||
33 | * entry->lock | ||
34 | * group->mark_lock | ||
35 | * inode->i_lock | ||
36 | * | ||
37 | * entry->lock protects 2 things, entry->group and entry->inode. You must hold | ||
38 | * that lock to dereference either of these things (they could be NULL even with | ||
39 | * the lock) | ||
40 | * | ||
41 | * group->mark_lock protects the mark_entries list anchored inside a given group | ||
42 | * and each entry is hooked via the g_list. It also sorta protects the | ||
43 | * free_g_list, which when used is anchored by a private list on the stack of the | ||
44 | * task which held the group->mark_lock. | ||
45 | * | ||
46 | * inode->i_lock protects the i_fsnotify_mark_entries list anchored inside a | ||
47 | * given inode and each entry is hooked via the i_list. (and sorta the | ||
48 | * free_i_list) | ||
49 | * | ||
50 | * | ||
51 | * LIFETIME: | ||
52 | * Inode marks survive between when they are added to an inode and when their | ||
53 | * refcnt==0. | ||
54 | * | ||
55 | * The inode mark can be cleared for a number of different reasons including: | ||
56 | * - The inode is unlinked for the last time. (fsnotify_inode_remove) | ||
57 | * - The inode is being evicted from cache. (fsnotify_inode_delete) | ||
58 | * - The fs the inode is on is unmounted. (fsnotify_inode_delete/fsnotify_unmount_inodes) | ||
59 | * - Something explicitly requests that it be removed. (fsnotify_destroy_mark_by_entry) | ||
60 | * - The fsnotify_group associated with the mark is going away and all such marks | ||
61 | * need to be cleaned up. (fsnotify_clear_marks_by_group) | ||
62 | * | ||
63 | * Worst case we are given an inode and need to clean up all the marks on that | ||
64 | * inode. We take i_lock and walk the i_fsnotify_mark_entries safely. For each | ||
65 | * mark on the list we take a reference (so the mark can't disappear under us). | ||
66 | * We remove that mark form the inode's list of marks and we add this mark to a | ||
67 | * private list anchored on the stack using i_free_list; At this point we no | ||
68 | * longer fear anything finding the mark using the inode's list of marks. | ||
69 | * | ||
70 | * We can safely and locklessly run the private list on the stack of everything | ||
71 | * we just unattached from the original inode. For each mark on the private list | ||
72 | * we grab the mark-> and can thus dereference mark->group and mark->inode. If | ||
73 | * we see the group and inode are not NULL we take those locks. Now holding all | ||
74 | * 3 locks we can completely remove the mark from other tasks finding it in the | ||
75 | * future. Remember, 10 things might already be referencing this mark, but they | ||
76 | * better be holding a ref. We drop our reference we took before we unhooked it | ||
77 | * from the inode. When the ref hits 0 we can free the mark. | ||
78 | * | ||
79 | * Very similarly for freeing by group, except we use free_g_list. | ||
80 | * | ||
81 | * This has the very interesting property of being able to run concurrently with | ||
82 | * any (or all) other directions. | ||
83 | */ | ||
84 | |||
85 | #include <linux/fs.h> | 19 | #include <linux/fs.h> |
86 | #include <linux/init.h> | 20 | #include <linux/init.h> |
87 | #include <linux/kernel.h> | 21 | #include <linux/kernel.h> |
@@ -95,30 +29,19 @@ | |||
95 | #include <linux/fsnotify_backend.h> | 29 | #include <linux/fsnotify_backend.h> |
96 | #include "fsnotify.h" | 30 | #include "fsnotify.h" |
97 | 31 | ||
98 | void fsnotify_get_mark(struct fsnotify_mark_entry *entry) | ||
99 | { | ||
100 | atomic_inc(&entry->refcnt); | ||
101 | } | ||
102 | |||
103 | void fsnotify_put_mark(struct fsnotify_mark_entry *entry) | ||
104 | { | ||
105 | if (atomic_dec_and_test(&entry->refcnt)) | ||
106 | entry->free_mark(entry); | ||
107 | } | ||
108 | |||
109 | /* | 32 | /* |
110 | * Recalculate the mask of events relevant to a given inode locked. | 33 | * Recalculate the mask of events relevant to a given inode locked. |
111 | */ | 34 | */ |
112 | static void fsnotify_recalc_inode_mask_locked(struct inode *inode) | 35 | static void fsnotify_recalc_inode_mask_locked(struct inode *inode) |
113 | { | 36 | { |
114 | struct fsnotify_mark_entry *entry; | 37 | struct fsnotify_mark *mark; |
115 | struct hlist_node *pos; | 38 | struct hlist_node *pos; |
116 | __u32 new_mask = 0; | 39 | __u32 new_mask = 0; |
117 | 40 | ||
118 | assert_spin_locked(&inode->i_lock); | 41 | assert_spin_locked(&inode->i_lock); |
119 | 42 | ||
120 | hlist_for_each_entry(entry, pos, &inode->i_fsnotify_mark_entries, i_list) | 43 | hlist_for_each_entry(mark, pos, &inode->i_fsnotify_marks, i.i_list) |
121 | new_mask |= entry->mask; | 44 | new_mask |= mark->mask; |
122 | inode->i_fsnotify_mask = new_mask; | 45 | inode->i_fsnotify_mask = new_mask; |
123 | } | 46 | } |
124 | 47 | ||
@@ -135,107 +58,26 @@ void fsnotify_recalc_inode_mask(struct inode *inode) | |||
135 | __fsnotify_update_child_dentry_flags(inode); | 58 | __fsnotify_update_child_dentry_flags(inode); |
136 | } | 59 | } |
137 | 60 | ||
138 | /* | 61 | void fsnotify_destroy_inode_mark(struct fsnotify_mark *mark) |
139 | * Any time a mark is getting freed we end up here. | ||
140 | * The caller had better be holding a reference to this mark so we don't actually | ||
141 | * do the final put under the entry->lock | ||
142 | */ | ||
143 | void fsnotify_destroy_mark_by_entry(struct fsnotify_mark_entry *entry) | ||
144 | { | 62 | { |
145 | struct fsnotify_group *group; | 63 | struct inode *inode = mark->i.inode; |
146 | struct inode *inode; | ||
147 | 64 | ||
148 | spin_lock(&entry->lock); | 65 | assert_spin_locked(&mark->lock); |
66 | assert_spin_locked(&mark->group->mark_lock); | ||
149 | 67 | ||
150 | group = entry->group; | ||
151 | inode = entry->inode; | ||
152 | |||
153 | BUG_ON(group && !inode); | ||
154 | BUG_ON(!group && inode); | ||
155 | |||
156 | /* if !group something else already marked this to die */ | ||
157 | if (!group) { | ||
158 | spin_unlock(&entry->lock); | ||
159 | return; | ||
160 | } | ||
161 | |||
162 | /* 1 from caller and 1 for being on i_list/g_list */ | ||
163 | BUG_ON(atomic_read(&entry->refcnt) < 2); | ||
164 | |||
165 | spin_lock(&group->mark_lock); | ||
166 | spin_lock(&inode->i_lock); | 68 | spin_lock(&inode->i_lock); |
167 | 69 | ||
168 | hlist_del_init(&entry->i_list); | 70 | hlist_del_init_rcu(&mark->i.i_list); |
169 | entry->inode = NULL; | 71 | mark->i.inode = NULL; |
170 | |||
171 | list_del_init(&entry->g_list); | ||
172 | entry->group = NULL; | ||
173 | |||
174 | fsnotify_put_mark(entry); /* for i_list and g_list */ | ||
175 | 72 | ||
176 | /* | 73 | /* |
177 | * this mark is now off the inode->i_fsnotify_mark_entries list and we | 74 | * this mark is now off the inode->i_fsnotify_marks list and we |
178 | * hold the inode->i_lock, so this is the perfect time to update the | 75 | * hold the inode->i_lock, so this is the perfect time to update the |
179 | * inode->i_fsnotify_mask | 76 | * inode->i_fsnotify_mask |
180 | */ | 77 | */ |
181 | fsnotify_recalc_inode_mask_locked(inode); | 78 | fsnotify_recalc_inode_mask_locked(inode); |
182 | 79 | ||
183 | spin_unlock(&inode->i_lock); | 80 | spin_unlock(&inode->i_lock); |
184 | spin_unlock(&group->mark_lock); | ||
185 | spin_unlock(&entry->lock); | ||
186 | |||
187 | /* | ||
188 | * Some groups like to know that marks are being freed. This is a | ||
189 | * callback to the group function to let it know that this entry | ||
190 | * is being freed. | ||
191 | */ | ||
192 | if (group->ops->freeing_mark) | ||
193 | group->ops->freeing_mark(entry, group); | ||
194 | |||
195 | /* | ||
196 | * __fsnotify_update_child_dentry_flags(inode); | ||
197 | * | ||
198 | * I really want to call that, but we can't, we have no idea if the inode | ||
199 | * still exists the second we drop the entry->lock. | ||
200 | * | ||
201 | * The next time an event arrive to this inode from one of it's children | ||
202 | * __fsnotify_parent will see that the inode doesn't care about it's | ||
203 | * children and will update all of these flags then. So really this | ||
204 | * is just a lazy update (and could be a perf win...) | ||
205 | */ | ||
206 | |||
207 | |||
208 | iput(inode); | ||
209 | |||
210 | /* | ||
211 | * it's possible that this group tried to destroy itself, but this | ||
212 | * this mark was simultaneously being freed by inode. If that's the | ||
213 | * case, we finish freeing the group here. | ||
214 | */ | ||
215 | if (unlikely(atomic_dec_and_test(&group->num_marks))) | ||
216 | fsnotify_final_destroy_group(group); | ||
217 | } | ||
218 | |||
219 | /* | ||
220 | * Given a group, destroy all of the marks associated with that group. | ||
221 | */ | ||
222 | void fsnotify_clear_marks_by_group(struct fsnotify_group *group) | ||
223 | { | ||
224 | struct fsnotify_mark_entry *lentry, *entry; | ||
225 | LIST_HEAD(free_list); | ||
226 | |||
227 | spin_lock(&group->mark_lock); | ||
228 | list_for_each_entry_safe(entry, lentry, &group->mark_entries, g_list) { | ||
229 | list_add(&entry->free_g_list, &free_list); | ||
230 | list_del_init(&entry->g_list); | ||
231 | fsnotify_get_mark(entry); | ||
232 | } | ||
233 | spin_unlock(&group->mark_lock); | ||
234 | |||
235 | list_for_each_entry_safe(entry, lentry, &free_list, free_g_list) { | ||
236 | fsnotify_destroy_mark_by_entry(entry); | ||
237 | fsnotify_put_mark(entry); | ||
238 | } | ||
239 | } | 81 | } |
240 | 82 | ||
241 | /* | 83 | /* |
@@ -243,112 +85,145 @@ void fsnotify_clear_marks_by_group(struct fsnotify_group *group) | |||
243 | */ | 85 | */ |
244 | void fsnotify_clear_marks_by_inode(struct inode *inode) | 86 | void fsnotify_clear_marks_by_inode(struct inode *inode) |
245 | { | 87 | { |
246 | struct fsnotify_mark_entry *entry, *lentry; | 88 | struct fsnotify_mark *mark, *lmark; |
247 | struct hlist_node *pos, *n; | 89 | struct hlist_node *pos, *n; |
248 | LIST_HEAD(free_list); | 90 | LIST_HEAD(free_list); |
249 | 91 | ||
250 | spin_lock(&inode->i_lock); | 92 | spin_lock(&inode->i_lock); |
251 | hlist_for_each_entry_safe(entry, pos, n, &inode->i_fsnotify_mark_entries, i_list) { | 93 | hlist_for_each_entry_safe(mark, pos, n, &inode->i_fsnotify_marks, i.i_list) { |
252 | list_add(&entry->free_i_list, &free_list); | 94 | list_add(&mark->i.free_i_list, &free_list); |
253 | hlist_del_init(&entry->i_list); | 95 | hlist_del_init_rcu(&mark->i.i_list); |
254 | fsnotify_get_mark(entry); | 96 | fsnotify_get_mark(mark); |
255 | } | 97 | } |
256 | spin_unlock(&inode->i_lock); | 98 | spin_unlock(&inode->i_lock); |
257 | 99 | ||
258 | list_for_each_entry_safe(entry, lentry, &free_list, free_i_list) { | 100 | list_for_each_entry_safe(mark, lmark, &free_list, i.free_i_list) { |
259 | fsnotify_destroy_mark_by_entry(entry); | 101 | fsnotify_destroy_mark(mark); |
260 | fsnotify_put_mark(entry); | 102 | fsnotify_put_mark(mark); |
261 | } | 103 | } |
262 | } | 104 | } |
263 | 105 | ||
264 | /* | 106 | /* |
107 | * Given a group clear all of the inode marks associated with that group. | ||
108 | */ | ||
109 | void fsnotify_clear_inode_marks_by_group(struct fsnotify_group *group) | ||
110 | { | ||
111 | fsnotify_clear_marks_by_group_flags(group, FSNOTIFY_MARK_FLAG_INODE); | ||
112 | } | ||
113 | |||
114 | /* | ||
265 | * given a group and inode, find the mark associated with that combination. | 115 | * given a group and inode, find the mark associated with that combination. |
266 | * if found take a reference to that mark and return it, else return NULL | 116 | * if found take a reference to that mark and return it, else return NULL |
267 | */ | 117 | */ |
268 | struct fsnotify_mark_entry *fsnotify_find_mark_entry(struct fsnotify_group *group, | 118 | struct fsnotify_mark *fsnotify_find_inode_mark_locked(struct fsnotify_group *group, |
269 | struct inode *inode) | 119 | struct inode *inode) |
270 | { | 120 | { |
271 | struct fsnotify_mark_entry *entry; | 121 | struct fsnotify_mark *mark; |
272 | struct hlist_node *pos; | 122 | struct hlist_node *pos; |
273 | 123 | ||
274 | assert_spin_locked(&inode->i_lock); | 124 | assert_spin_locked(&inode->i_lock); |
275 | 125 | ||
276 | hlist_for_each_entry(entry, pos, &inode->i_fsnotify_mark_entries, i_list) { | 126 | hlist_for_each_entry(mark, pos, &inode->i_fsnotify_marks, i.i_list) { |
277 | if (entry->group == group) { | 127 | if (mark->group == group) { |
278 | fsnotify_get_mark(entry); | 128 | fsnotify_get_mark(mark); |
279 | return entry; | 129 | return mark; |
280 | } | 130 | } |
281 | } | 131 | } |
282 | return NULL; | 132 | return NULL; |
283 | } | 133 | } |
284 | 134 | ||
285 | /* | 135 | /* |
286 | * Nothing fancy, just initialize lists and locks and counters. | 136 | * given a group and inode, find the mark associated with that combination. |
137 | * if found take a reference to that mark and return it, else return NULL | ||
287 | */ | 138 | */ |
288 | void fsnotify_init_mark(struct fsnotify_mark_entry *entry, | 139 | struct fsnotify_mark *fsnotify_find_inode_mark(struct fsnotify_group *group, |
289 | void (*free_mark)(struct fsnotify_mark_entry *entry)) | 140 | struct inode *inode) |
141 | { | ||
142 | struct fsnotify_mark *mark; | ||
143 | |||
144 | spin_lock(&inode->i_lock); | ||
145 | mark = fsnotify_find_inode_mark_locked(group, inode); | ||
146 | spin_unlock(&inode->i_lock); | ||
290 | 147 | ||
148 | return mark; | ||
149 | } | ||
150 | |||
151 | /* | ||
152 | * If we are setting a mark mask on an inode mark we should pin the inode | ||
153 | * in memory. | ||
154 | */ | ||
155 | void fsnotify_set_inode_mark_mask_locked(struct fsnotify_mark *mark, | ||
156 | __u32 mask) | ||
291 | { | 157 | { |
292 | spin_lock_init(&entry->lock); | 158 | struct inode *inode; |
293 | atomic_set(&entry->refcnt, 1); | 159 | |
294 | INIT_HLIST_NODE(&entry->i_list); | 160 | assert_spin_locked(&mark->lock); |
295 | entry->group = NULL; | 161 | |
296 | entry->mask = 0; | 162 | if (mask && |
297 | entry->inode = NULL; | 163 | mark->i.inode && |
298 | entry->free_mark = free_mark; | 164 | !(mark->flags & FSNOTIFY_MARK_FLAG_OBJECT_PINNED)) { |
165 | mark->flags |= FSNOTIFY_MARK_FLAG_OBJECT_PINNED; | ||
166 | inode = igrab(mark->i.inode); | ||
167 | /* | ||
168 | * we shouldn't be able to get here if the inode wasn't | ||
169 | * already safely held in memory. But bug in case it | ||
170 | * ever is wrong. | ||
171 | */ | ||
172 | BUG_ON(!inode); | ||
173 | } | ||
299 | } | 174 | } |
300 | 175 | ||
301 | /* | 176 | /* |
302 | * Attach an initialized mark entry to a given group and inode. | 177 | * Attach an initialized mark to a given inode. |
303 | * These marks may be used for the fsnotify backend to determine which | 178 | * These marks may be used for the fsnotify backend to determine which |
304 | * event types should be delivered to which group and for which inodes. | 179 | * event types should be delivered to which group and for which inodes. These |
180 | * marks are ordered according to the group's location in memory. | ||
305 | */ | 181 | */ |
306 | int fsnotify_add_mark(struct fsnotify_mark_entry *entry, | 182 | int fsnotify_add_inode_mark(struct fsnotify_mark *mark, |
307 | struct fsnotify_group *group, struct inode *inode) | 183 | struct fsnotify_group *group, struct inode *inode, |
184 | int allow_dups) | ||
308 | { | 185 | { |
309 | struct fsnotify_mark_entry *lentry; | 186 | struct fsnotify_mark *lmark; |
187 | struct hlist_node *node, *last = NULL; | ||
310 | int ret = 0; | 188 | int ret = 0; |
311 | 189 | ||
312 | inode = igrab(inode); | 190 | mark->flags |= FSNOTIFY_MARK_FLAG_INODE; |
313 | if (unlikely(!inode)) | 191 | |
314 | return -EINVAL; | 192 | assert_spin_locked(&mark->lock); |
193 | assert_spin_locked(&group->mark_lock); | ||
315 | 194 | ||
316 | /* | ||
317 | * LOCKING ORDER!!!! | ||
318 | * entry->lock | ||
319 | * group->mark_lock | ||
320 | * inode->i_lock | ||
321 | */ | ||
322 | spin_lock(&entry->lock); | ||
323 | spin_lock(&group->mark_lock); | ||
324 | spin_lock(&inode->i_lock); | 195 | spin_lock(&inode->i_lock); |
325 | 196 | ||
326 | lentry = fsnotify_find_mark_entry(group, inode); | 197 | mark->i.inode = inode; |
327 | if (!lentry) { | ||
328 | entry->group = group; | ||
329 | entry->inode = inode; | ||
330 | 198 | ||
331 | hlist_add_head(&entry->i_list, &inode->i_fsnotify_mark_entries); | 199 | /* is mark the first mark? */ |
332 | list_add(&entry->g_list, &group->mark_entries); | 200 | if (hlist_empty(&inode->i_fsnotify_marks)) { |
201 | hlist_add_head_rcu(&mark->i.i_list, &inode->i_fsnotify_marks); | ||
202 | goto out; | ||
203 | } | ||
333 | 204 | ||
334 | fsnotify_get_mark(entry); /* for i_list and g_list */ | 205 | /* should mark be in the middle of the current list? */ |
206 | hlist_for_each_entry(lmark, node, &inode->i_fsnotify_marks, i.i_list) { | ||
207 | last = node; | ||
208 | |||
209 | if ((lmark->group == group) && !allow_dups) { | ||
210 | ret = -EEXIST; | ||
211 | goto out; | ||
212 | } | ||
335 | 213 | ||
336 | atomic_inc(&group->num_marks); | 214 | if (mark->group < lmark->group) |
215 | continue; | ||
337 | 216 | ||
338 | fsnotify_recalc_inode_mask_locked(inode); | 217 | hlist_add_before_rcu(&mark->i.i_list, &lmark->i.i_list); |
218 | goto out; | ||
339 | } | 219 | } |
340 | 220 | ||
221 | BUG_ON(last == NULL); | ||
222 | /* mark should be the last entry. last is the current last entry */ | ||
223 | hlist_add_after_rcu(last, &mark->i.i_list); | ||
224 | out: | ||
225 | fsnotify_recalc_inode_mask_locked(inode); | ||
341 | spin_unlock(&inode->i_lock); | 226 | spin_unlock(&inode->i_lock); |
342 | spin_unlock(&group->mark_lock); | ||
343 | spin_unlock(&entry->lock); | ||
344 | |||
345 | if (lentry) { | ||
346 | ret = -EEXIST; | ||
347 | iput(inode); | ||
348 | fsnotify_put_mark(lentry); | ||
349 | } else { | ||
350 | __fsnotify_update_child_dentry_flags(inode); | ||
351 | } | ||
352 | 227 | ||
353 | return ret; | 228 | return ret; |
354 | } | 229 | } |