diff options
author | Jiri Kosina <jkosina@suse.cz> | 2010-08-11 03:36:51 -0400 |
---|---|---|
committer | Jiri Kosina <jkosina@suse.cz> | 2010-08-11 03:36:51 -0400 |
commit | 6396fc3b3ff3f6b942992b653a62df11dcef9bea (patch) | |
tree | db3c7cbe833b43c653adc99f70941431c5ff7c4e /kernel | |
parent | 4785879e4d340e24e54f6de2ccfc42728b912808 (diff) | |
parent | 3d30701b58970425e1d45994d6cb82f828924fdd (diff) |
Merge branch 'master' into for-next
Conflicts:
fs/exofs/inode.c
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/Makefile | 5 | ||||
-rw-r--r-- | kernel/acct.c | 2 | ||||
-rw-r--r-- | kernel/audit.c | 1 | ||||
-rw-r--r-- | kernel/audit.h | 26 | ||||
-rw-r--r-- | kernel/audit_tree.c | 237 | ||||
-rw-r--r-- | kernel/audit_watch.c | 274 | ||||
-rw-r--r-- | kernel/auditfilter.c | 39 | ||||
-rw-r--r-- | kernel/auditsc.c | 10 | ||||
-rw-r--r-- | kernel/compat.c | 17 | ||||
-rw-r--r-- | kernel/posix-cpu-timers.c | 8 | ||||
-rw-r--r-- | kernel/power/block_io.c | 2 | ||||
-rw-r--r-- | kernel/sys.c | 202 | ||||
-rw-r--r-- | kernel/sys_ni.c | 4 | ||||
-rw-r--r-- | kernel/sysctl.c | 7 | ||||
-rw-r--r-- | kernel/timer.c | 1 | ||||
-rw-r--r-- | kernel/trace/blktrace.c | 80 |
16 files changed, 556 insertions, 359 deletions
diff --git a/kernel/Makefile b/kernel/Makefile index c53e491e25a8..0b72d1a74be0 100644 --- a/kernel/Makefile +++ b/kernel/Makefile | |||
@@ -70,10 +70,11 @@ obj-$(CONFIG_IKCONFIG) += configs.o | |||
70 | obj-$(CONFIG_RESOURCE_COUNTERS) += res_counter.o | 70 | obj-$(CONFIG_RESOURCE_COUNTERS) += res_counter.o |
71 | obj-$(CONFIG_SMP) += stop_machine.o | 71 | obj-$(CONFIG_SMP) += stop_machine.o |
72 | obj-$(CONFIG_KPROBES_SANITY_TEST) += test_kprobes.o | 72 | obj-$(CONFIG_KPROBES_SANITY_TEST) += test_kprobes.o |
73 | obj-$(CONFIG_AUDIT) += audit.o auditfilter.o audit_watch.o | 73 | obj-$(CONFIG_AUDIT) += audit.o auditfilter.o |
74 | obj-$(CONFIG_AUDITSYSCALL) += auditsc.o | 74 | obj-$(CONFIG_AUDITSYSCALL) += auditsc.o |
75 | obj-$(CONFIG_GCOV_KERNEL) += gcov/ | 75 | obj-$(CONFIG_AUDIT_WATCH) += audit_watch.o |
76 | obj-$(CONFIG_AUDIT_TREE) += audit_tree.o | 76 | obj-$(CONFIG_AUDIT_TREE) += audit_tree.o |
77 | obj-$(CONFIG_GCOV_KERNEL) += gcov/ | ||
77 | obj-$(CONFIG_KPROBES) += kprobes.o | 78 | obj-$(CONFIG_KPROBES) += kprobes.o |
78 | obj-$(CONFIG_KGDB) += debug/ | 79 | obj-$(CONFIG_KGDB) += debug/ |
79 | obj-$(CONFIG_DETECT_HUNG_TASK) += hung_task.o | 80 | obj-$(CONFIG_DETECT_HUNG_TASK) += hung_task.o |
diff --git a/kernel/acct.c b/kernel/acct.c index 385b88461c29..fa7eb3de2ddc 100644 --- a/kernel/acct.c +++ b/kernel/acct.c | |||
@@ -122,7 +122,7 @@ static int check_free_space(struct bsd_acct_struct *acct, struct file *file) | |||
122 | spin_unlock(&acct_lock); | 122 | spin_unlock(&acct_lock); |
123 | 123 | ||
124 | /* May block */ | 124 | /* May block */ |
125 | if (vfs_statfs(file->f_path.dentry, &sbuf)) | 125 | if (vfs_statfs(&file->f_path, &sbuf)) |
126 | return res; | 126 | return res; |
127 | suspend = sbuf.f_blocks * SUSPEND; | 127 | suspend = sbuf.f_blocks * SUSPEND; |
128 | resume = sbuf.f_blocks * RESUME; | 128 | resume = sbuf.f_blocks * RESUME; |
diff --git a/kernel/audit.c b/kernel/audit.c index 8296aa516c5a..d96045789b54 100644 --- a/kernel/audit.c +++ b/kernel/audit.c | |||
@@ -56,7 +56,6 @@ | |||
56 | #include <net/netlink.h> | 56 | #include <net/netlink.h> |
57 | #include <linux/skbuff.h> | 57 | #include <linux/skbuff.h> |
58 | #include <linux/netlink.h> | 58 | #include <linux/netlink.h> |
59 | #include <linux/inotify.h> | ||
60 | #include <linux/freezer.h> | 59 | #include <linux/freezer.h> |
61 | #include <linux/tty.h> | 60 | #include <linux/tty.h> |
62 | 61 | ||
diff --git a/kernel/audit.h b/kernel/audit.h index 208687be4f30..f7206db4e13d 100644 --- a/kernel/audit.h +++ b/kernel/audit.h | |||
@@ -103,21 +103,27 @@ extern struct mutex audit_filter_mutex; | |||
103 | extern void audit_free_rule_rcu(struct rcu_head *); | 103 | extern void audit_free_rule_rcu(struct rcu_head *); |
104 | extern struct list_head audit_filter_list[]; | 104 | extern struct list_head audit_filter_list[]; |
105 | 105 | ||
106 | extern struct audit_entry *audit_dupe_rule(struct audit_krule *old); | ||
107 | |||
106 | /* audit watch functions */ | 108 | /* audit watch functions */ |
107 | extern unsigned long audit_watch_inode(struct audit_watch *watch); | 109 | #ifdef CONFIG_AUDIT_WATCH |
108 | extern dev_t audit_watch_dev(struct audit_watch *watch); | ||
109 | extern void audit_put_watch(struct audit_watch *watch); | 110 | extern void audit_put_watch(struct audit_watch *watch); |
110 | extern void audit_get_watch(struct audit_watch *watch); | 111 | extern void audit_get_watch(struct audit_watch *watch); |
111 | extern int audit_to_watch(struct audit_krule *krule, char *path, int len, u32 op); | 112 | extern int audit_to_watch(struct audit_krule *krule, char *path, int len, u32 op); |
112 | extern int audit_add_watch(struct audit_krule *krule); | 113 | extern int audit_add_watch(struct audit_krule *krule, struct list_head **list); |
113 | extern void audit_remove_watch(struct audit_watch *watch); | 114 | extern void audit_remove_watch_rule(struct audit_krule *krule); |
114 | extern void audit_remove_watch_rule(struct audit_krule *krule, struct list_head *list); | ||
115 | extern void audit_inotify_unregister(struct list_head *in_list); | ||
116 | extern char *audit_watch_path(struct audit_watch *watch); | 115 | extern char *audit_watch_path(struct audit_watch *watch); |
117 | extern struct list_head *audit_watch_rules(struct audit_watch *watch); | 116 | extern int audit_watch_compare(struct audit_watch *watch, unsigned long ino, dev_t dev); |
118 | 117 | #else | |
119 | extern struct audit_entry *audit_dupe_rule(struct audit_krule *old, | 118 | #define audit_put_watch(w) {} |
120 | struct audit_watch *watch); | 119 | #define audit_get_watch(w) {} |
120 | #define audit_to_watch(k, p, l, o) (-EINVAL) | ||
121 | #define audit_add_watch(k, l) (-EINVAL) | ||
122 | #define audit_remove_watch_rule(k) BUG() | ||
123 | #define audit_watch_path(w) "" | ||
124 | #define audit_watch_compare(w, i, d) 0 | ||
125 | |||
126 | #endif /* CONFIG_AUDIT_WATCH */ | ||
121 | 127 | ||
122 | #ifdef CONFIG_AUDIT_TREE | 128 | #ifdef CONFIG_AUDIT_TREE |
123 | extern struct audit_chunk *audit_tree_lookup(const struct inode *); | 129 | extern struct audit_chunk *audit_tree_lookup(const struct inode *); |
diff --git a/kernel/audit_tree.c b/kernel/audit_tree.c index 46a57b57a335..7f18d3a4527e 100644 --- a/kernel/audit_tree.c +++ b/kernel/audit_tree.c | |||
@@ -1,5 +1,5 @@ | |||
1 | #include "audit.h" | 1 | #include "audit.h" |
2 | #include <linux/inotify.h> | 2 | #include <linux/fsnotify_backend.h> |
3 | #include <linux/namei.h> | 3 | #include <linux/namei.h> |
4 | #include <linux/mount.h> | 4 | #include <linux/mount.h> |
5 | #include <linux/kthread.h> | 5 | #include <linux/kthread.h> |
@@ -22,7 +22,7 @@ struct audit_tree { | |||
22 | 22 | ||
23 | struct audit_chunk { | 23 | struct audit_chunk { |
24 | struct list_head hash; | 24 | struct list_head hash; |
25 | struct inotify_watch watch; | 25 | struct fsnotify_mark mark; |
26 | struct list_head trees; /* with root here */ | 26 | struct list_head trees; /* with root here */ |
27 | int dead; | 27 | int dead; |
28 | int count; | 28 | int count; |
@@ -59,7 +59,7 @@ static LIST_HEAD(prune_list); | |||
59 | * tree is refcounted; one reference for "some rules on rules_list refer to | 59 | * tree is refcounted; one reference for "some rules on rules_list refer to |
60 | * it", one for each chunk with pointer to it. | 60 | * it", one for each chunk with pointer to it. |
61 | * | 61 | * |
62 | * chunk is refcounted by embedded inotify_watch + .refs (non-zero refcount | 62 | * chunk is refcounted by embedded fsnotify_mark + .refs (non-zero refcount |
63 | * of watch contributes 1 to .refs). | 63 | * of watch contributes 1 to .refs). |
64 | * | 64 | * |
65 | * node.index allows to get from node.list to containing chunk. | 65 | * node.index allows to get from node.list to containing chunk. |
@@ -68,7 +68,7 @@ static LIST_HEAD(prune_list); | |||
68 | * that makes a difference. Some. | 68 | * that makes a difference. Some. |
69 | */ | 69 | */ |
70 | 70 | ||
71 | static struct inotify_handle *rtree_ih; | 71 | static struct fsnotify_group *audit_tree_group; |
72 | 72 | ||
73 | static struct audit_tree *alloc_tree(const char *s) | 73 | static struct audit_tree *alloc_tree(const char *s) |
74 | { | 74 | { |
@@ -111,29 +111,6 @@ const char *audit_tree_path(struct audit_tree *tree) | |||
111 | return tree->pathname; | 111 | return tree->pathname; |
112 | } | 112 | } |
113 | 113 | ||
114 | static struct audit_chunk *alloc_chunk(int count) | ||
115 | { | ||
116 | struct audit_chunk *chunk; | ||
117 | size_t size; | ||
118 | int i; | ||
119 | |||
120 | size = offsetof(struct audit_chunk, owners) + count * sizeof(struct node); | ||
121 | chunk = kzalloc(size, GFP_KERNEL); | ||
122 | if (!chunk) | ||
123 | return NULL; | ||
124 | |||
125 | INIT_LIST_HEAD(&chunk->hash); | ||
126 | INIT_LIST_HEAD(&chunk->trees); | ||
127 | chunk->count = count; | ||
128 | atomic_long_set(&chunk->refs, 1); | ||
129 | for (i = 0; i < count; i++) { | ||
130 | INIT_LIST_HEAD(&chunk->owners[i].list); | ||
131 | chunk->owners[i].index = i; | ||
132 | } | ||
133 | inotify_init_watch(&chunk->watch); | ||
134 | return chunk; | ||
135 | } | ||
136 | |||
137 | static void free_chunk(struct audit_chunk *chunk) | 114 | static void free_chunk(struct audit_chunk *chunk) |
138 | { | 115 | { |
139 | int i; | 116 | int i; |
@@ -157,6 +134,35 @@ static void __put_chunk(struct rcu_head *rcu) | |||
157 | audit_put_chunk(chunk); | 134 | audit_put_chunk(chunk); |
158 | } | 135 | } |
159 | 136 | ||
137 | static void audit_tree_destroy_watch(struct fsnotify_mark *entry) | ||
138 | { | ||
139 | struct audit_chunk *chunk = container_of(entry, struct audit_chunk, mark); | ||
140 | call_rcu(&chunk->head, __put_chunk); | ||
141 | } | ||
142 | |||
143 | static struct audit_chunk *alloc_chunk(int count) | ||
144 | { | ||
145 | struct audit_chunk *chunk; | ||
146 | size_t size; | ||
147 | int i; | ||
148 | |||
149 | size = offsetof(struct audit_chunk, owners) + count * sizeof(struct node); | ||
150 | chunk = kzalloc(size, GFP_KERNEL); | ||
151 | if (!chunk) | ||
152 | return NULL; | ||
153 | |||
154 | INIT_LIST_HEAD(&chunk->hash); | ||
155 | INIT_LIST_HEAD(&chunk->trees); | ||
156 | chunk->count = count; | ||
157 | atomic_long_set(&chunk->refs, 1); | ||
158 | for (i = 0; i < count; i++) { | ||
159 | INIT_LIST_HEAD(&chunk->owners[i].list); | ||
160 | chunk->owners[i].index = i; | ||
161 | } | ||
162 | fsnotify_init_mark(&chunk->mark, audit_tree_destroy_watch); | ||
163 | return chunk; | ||
164 | } | ||
165 | |||
160 | enum {HASH_SIZE = 128}; | 166 | enum {HASH_SIZE = 128}; |
161 | static struct list_head chunk_hash_heads[HASH_SIZE]; | 167 | static struct list_head chunk_hash_heads[HASH_SIZE]; |
162 | static __cacheline_aligned_in_smp DEFINE_SPINLOCK(hash_lock); | 168 | static __cacheline_aligned_in_smp DEFINE_SPINLOCK(hash_lock); |
@@ -167,10 +173,15 @@ static inline struct list_head *chunk_hash(const struct inode *inode) | |||
167 | return chunk_hash_heads + n % HASH_SIZE; | 173 | return chunk_hash_heads + n % HASH_SIZE; |
168 | } | 174 | } |
169 | 175 | ||
170 | /* hash_lock is held by caller */ | 176 | /* hash_lock & entry->lock is held by caller */ |
171 | static void insert_hash(struct audit_chunk *chunk) | 177 | static void insert_hash(struct audit_chunk *chunk) |
172 | { | 178 | { |
173 | struct list_head *list = chunk_hash(chunk->watch.inode); | 179 | struct fsnotify_mark *entry = &chunk->mark; |
180 | struct list_head *list; | ||
181 | |||
182 | if (!entry->i.inode) | ||
183 | return; | ||
184 | list = chunk_hash(entry->i.inode); | ||
174 | list_add_rcu(&chunk->hash, list); | 185 | list_add_rcu(&chunk->hash, list); |
175 | } | 186 | } |
176 | 187 | ||
@@ -181,7 +192,8 @@ struct audit_chunk *audit_tree_lookup(const struct inode *inode) | |||
181 | struct audit_chunk *p; | 192 | struct audit_chunk *p; |
182 | 193 | ||
183 | list_for_each_entry_rcu(p, list, hash) { | 194 | list_for_each_entry_rcu(p, list, hash) { |
184 | if (p->watch.inode == inode) { | 195 | /* mark.inode may have gone NULL, but who cares? */ |
196 | if (p->mark.i.inode == inode) { | ||
185 | atomic_long_inc(&p->refs); | 197 | atomic_long_inc(&p->refs); |
186 | return p; | 198 | return p; |
187 | } | 199 | } |
@@ -210,38 +222,19 @@ static struct audit_chunk *find_chunk(struct node *p) | |||
210 | static void untag_chunk(struct node *p) | 222 | static void untag_chunk(struct node *p) |
211 | { | 223 | { |
212 | struct audit_chunk *chunk = find_chunk(p); | 224 | struct audit_chunk *chunk = find_chunk(p); |
225 | struct fsnotify_mark *entry = &chunk->mark; | ||
213 | struct audit_chunk *new; | 226 | struct audit_chunk *new; |
214 | struct audit_tree *owner; | 227 | struct audit_tree *owner; |
215 | int size = chunk->count - 1; | 228 | int size = chunk->count - 1; |
216 | int i, j; | 229 | int i, j; |
217 | 230 | ||
218 | if (!pin_inotify_watch(&chunk->watch)) { | 231 | fsnotify_get_mark(entry); |
219 | /* | ||
220 | * Filesystem is shutting down; all watches are getting | ||
221 | * evicted, just take it off the node list for this | ||
222 | * tree and let the eviction logics take care of the | ||
223 | * rest. | ||
224 | */ | ||
225 | owner = p->owner; | ||
226 | if (owner->root == chunk) { | ||
227 | list_del_init(&owner->same_root); | ||
228 | owner->root = NULL; | ||
229 | } | ||
230 | list_del_init(&p->list); | ||
231 | p->owner = NULL; | ||
232 | put_tree(owner); | ||
233 | return; | ||
234 | } | ||
235 | 232 | ||
236 | spin_unlock(&hash_lock); | 233 | spin_unlock(&hash_lock); |
237 | 234 | ||
238 | /* | 235 | spin_lock(&entry->lock); |
239 | * pin_inotify_watch() succeeded, so the watch won't go away | 236 | if (chunk->dead || !entry->i.inode) { |
240 | * from under us. | 237 | spin_unlock(&entry->lock); |
241 | */ | ||
242 | mutex_lock(&chunk->watch.inode->inotify_mutex); | ||
243 | if (chunk->dead) { | ||
244 | mutex_unlock(&chunk->watch.inode->inotify_mutex); | ||
245 | goto out; | 238 | goto out; |
246 | } | 239 | } |
247 | 240 | ||
@@ -256,16 +249,17 @@ static void untag_chunk(struct node *p) | |||
256 | list_del_init(&p->list); | 249 | list_del_init(&p->list); |
257 | list_del_rcu(&chunk->hash); | 250 | list_del_rcu(&chunk->hash); |
258 | spin_unlock(&hash_lock); | 251 | spin_unlock(&hash_lock); |
259 | inotify_evict_watch(&chunk->watch); | 252 | spin_unlock(&entry->lock); |
260 | mutex_unlock(&chunk->watch.inode->inotify_mutex); | 253 | fsnotify_destroy_mark(entry); |
261 | put_inotify_watch(&chunk->watch); | 254 | fsnotify_put_mark(entry); |
262 | goto out; | 255 | goto out; |
263 | } | 256 | } |
264 | 257 | ||
265 | new = alloc_chunk(size); | 258 | new = alloc_chunk(size); |
266 | if (!new) | 259 | if (!new) |
267 | goto Fallback; | 260 | goto Fallback; |
268 | if (inotify_clone_watch(&chunk->watch, &new->watch) < 0) { | 261 | fsnotify_duplicate_mark(&new->mark, entry); |
262 | if (fsnotify_add_mark(&new->mark, new->mark.group, new->mark.i.inode, NULL, 1)) { | ||
269 | free_chunk(new); | 263 | free_chunk(new); |
270 | goto Fallback; | 264 | goto Fallback; |
271 | } | 265 | } |
@@ -298,9 +292,9 @@ static void untag_chunk(struct node *p) | |||
298 | list_for_each_entry(owner, &new->trees, same_root) | 292 | list_for_each_entry(owner, &new->trees, same_root) |
299 | owner->root = new; | 293 | owner->root = new; |
300 | spin_unlock(&hash_lock); | 294 | spin_unlock(&hash_lock); |
301 | inotify_evict_watch(&chunk->watch); | 295 | spin_unlock(&entry->lock); |
302 | mutex_unlock(&chunk->watch.inode->inotify_mutex); | 296 | fsnotify_destroy_mark(entry); |
303 | put_inotify_watch(&chunk->watch); | 297 | fsnotify_put_mark(entry); |
304 | goto out; | 298 | goto out; |
305 | 299 | ||
306 | Fallback: | 300 | Fallback: |
@@ -314,31 +308,33 @@ Fallback: | |||
314 | p->owner = NULL; | 308 | p->owner = NULL; |
315 | put_tree(owner); | 309 | put_tree(owner); |
316 | spin_unlock(&hash_lock); | 310 | spin_unlock(&hash_lock); |
317 | mutex_unlock(&chunk->watch.inode->inotify_mutex); | 311 | spin_unlock(&entry->lock); |
318 | out: | 312 | out: |
319 | unpin_inotify_watch(&chunk->watch); | 313 | fsnotify_put_mark(entry); |
320 | spin_lock(&hash_lock); | 314 | spin_lock(&hash_lock); |
321 | } | 315 | } |
322 | 316 | ||
323 | static int create_chunk(struct inode *inode, struct audit_tree *tree) | 317 | static int create_chunk(struct inode *inode, struct audit_tree *tree) |
324 | { | 318 | { |
319 | struct fsnotify_mark *entry; | ||
325 | struct audit_chunk *chunk = alloc_chunk(1); | 320 | struct audit_chunk *chunk = alloc_chunk(1); |
326 | if (!chunk) | 321 | if (!chunk) |
327 | return -ENOMEM; | 322 | return -ENOMEM; |
328 | 323 | ||
329 | if (inotify_add_watch(rtree_ih, &chunk->watch, inode, IN_IGNORED | IN_DELETE_SELF) < 0) { | 324 | entry = &chunk->mark; |
325 | if (fsnotify_add_mark(entry, audit_tree_group, inode, NULL, 0)) { | ||
330 | free_chunk(chunk); | 326 | free_chunk(chunk); |
331 | return -ENOSPC; | 327 | return -ENOSPC; |
332 | } | 328 | } |
333 | 329 | ||
334 | mutex_lock(&inode->inotify_mutex); | 330 | spin_lock(&entry->lock); |
335 | spin_lock(&hash_lock); | 331 | spin_lock(&hash_lock); |
336 | if (tree->goner) { | 332 | if (tree->goner) { |
337 | spin_unlock(&hash_lock); | 333 | spin_unlock(&hash_lock); |
338 | chunk->dead = 1; | 334 | chunk->dead = 1; |
339 | inotify_evict_watch(&chunk->watch); | 335 | spin_unlock(&entry->lock); |
340 | mutex_unlock(&inode->inotify_mutex); | 336 | fsnotify_destroy_mark(entry); |
341 | put_inotify_watch(&chunk->watch); | 337 | fsnotify_put_mark(entry); |
342 | return 0; | 338 | return 0; |
343 | } | 339 | } |
344 | chunk->owners[0].index = (1U << 31); | 340 | chunk->owners[0].index = (1U << 31); |
@@ -351,30 +347,31 @@ static int create_chunk(struct inode *inode, struct audit_tree *tree) | |||
351 | } | 347 | } |
352 | insert_hash(chunk); | 348 | insert_hash(chunk); |
353 | spin_unlock(&hash_lock); | 349 | spin_unlock(&hash_lock); |
354 | mutex_unlock(&inode->inotify_mutex); | 350 | spin_unlock(&entry->lock); |
355 | return 0; | 351 | return 0; |
356 | } | 352 | } |
357 | 353 | ||
358 | /* the first tagged inode becomes root of tree */ | 354 | /* the first tagged inode becomes root of tree */ |
359 | static int tag_chunk(struct inode *inode, struct audit_tree *tree) | 355 | static int tag_chunk(struct inode *inode, struct audit_tree *tree) |
360 | { | 356 | { |
361 | struct inotify_watch *watch; | 357 | struct fsnotify_mark *old_entry, *chunk_entry; |
362 | struct audit_tree *owner; | 358 | struct audit_tree *owner; |
363 | struct audit_chunk *chunk, *old; | 359 | struct audit_chunk *chunk, *old; |
364 | struct node *p; | 360 | struct node *p; |
365 | int n; | 361 | int n; |
366 | 362 | ||
367 | if (inotify_find_watch(rtree_ih, inode, &watch) < 0) | 363 | old_entry = fsnotify_find_inode_mark(audit_tree_group, inode); |
364 | if (!old_entry) | ||
368 | return create_chunk(inode, tree); | 365 | return create_chunk(inode, tree); |
369 | 366 | ||
370 | old = container_of(watch, struct audit_chunk, watch); | 367 | old = container_of(old_entry, struct audit_chunk, mark); |
371 | 368 | ||
372 | /* are we already there? */ | 369 | /* are we already there? */ |
373 | spin_lock(&hash_lock); | 370 | spin_lock(&hash_lock); |
374 | for (n = 0; n < old->count; n++) { | 371 | for (n = 0; n < old->count; n++) { |
375 | if (old->owners[n].owner == tree) { | 372 | if (old->owners[n].owner == tree) { |
376 | spin_unlock(&hash_lock); | 373 | spin_unlock(&hash_lock); |
377 | put_inotify_watch(&old->watch); | 374 | fsnotify_put_mark(old_entry); |
378 | return 0; | 375 | return 0; |
379 | } | 376 | } |
380 | } | 377 | } |
@@ -382,25 +379,44 @@ static int tag_chunk(struct inode *inode, struct audit_tree *tree) | |||
382 | 379 | ||
383 | chunk = alloc_chunk(old->count + 1); | 380 | chunk = alloc_chunk(old->count + 1); |
384 | if (!chunk) { | 381 | if (!chunk) { |
385 | put_inotify_watch(&old->watch); | 382 | fsnotify_put_mark(old_entry); |
386 | return -ENOMEM; | 383 | return -ENOMEM; |
387 | } | 384 | } |
388 | 385 | ||
389 | mutex_lock(&inode->inotify_mutex); | 386 | chunk_entry = &chunk->mark; |
390 | if (inotify_clone_watch(&old->watch, &chunk->watch) < 0) { | 387 | |
391 | mutex_unlock(&inode->inotify_mutex); | 388 | spin_lock(&old_entry->lock); |
392 | put_inotify_watch(&old->watch); | 389 | if (!old_entry->i.inode) { |
390 | /* old_entry is being shot, lets just lie */ | ||
391 | spin_unlock(&old_entry->lock); | ||
392 | fsnotify_put_mark(old_entry); | ||
393 | free_chunk(chunk); | 393 | free_chunk(chunk); |
394 | return -ENOENT; | ||
395 | } | ||
396 | |||
397 | fsnotify_duplicate_mark(chunk_entry, old_entry); | ||
398 | if (fsnotify_add_mark(chunk_entry, chunk_entry->group, chunk_entry->i.inode, NULL, 1)) { | ||
399 | spin_unlock(&old_entry->lock); | ||
400 | free_chunk(chunk); | ||
401 | fsnotify_put_mark(old_entry); | ||
394 | return -ENOSPC; | 402 | return -ENOSPC; |
395 | } | 403 | } |
404 | |||
405 | /* even though we hold old_entry->lock, this is safe since chunk_entry->lock could NEVER have been grabbed before */ | ||
406 | spin_lock(&chunk_entry->lock); | ||
396 | spin_lock(&hash_lock); | 407 | spin_lock(&hash_lock); |
408 | |||
409 | /* we now hold old_entry->lock, chunk_entry->lock, and hash_lock */ | ||
397 | if (tree->goner) { | 410 | if (tree->goner) { |
398 | spin_unlock(&hash_lock); | 411 | spin_unlock(&hash_lock); |
399 | chunk->dead = 1; | 412 | chunk->dead = 1; |
400 | inotify_evict_watch(&chunk->watch); | 413 | spin_unlock(&chunk_entry->lock); |
401 | mutex_unlock(&inode->inotify_mutex); | 414 | spin_unlock(&old_entry->lock); |
402 | put_inotify_watch(&old->watch); | 415 | |
403 | put_inotify_watch(&chunk->watch); | 416 | fsnotify_destroy_mark(chunk_entry); |
417 | |||
418 | fsnotify_put_mark(chunk_entry); | ||
419 | fsnotify_put_mark(old_entry); | ||
404 | return 0; | 420 | return 0; |
405 | } | 421 | } |
406 | list_replace_init(&old->trees, &chunk->trees); | 422 | list_replace_init(&old->trees, &chunk->trees); |
@@ -426,10 +442,11 @@ static int tag_chunk(struct inode *inode, struct audit_tree *tree) | |||
426 | list_add(&tree->same_root, &chunk->trees); | 442 | list_add(&tree->same_root, &chunk->trees); |
427 | } | 443 | } |
428 | spin_unlock(&hash_lock); | 444 | spin_unlock(&hash_lock); |
429 | inotify_evict_watch(&old->watch); | 445 | spin_unlock(&chunk_entry->lock); |
430 | mutex_unlock(&inode->inotify_mutex); | 446 | spin_unlock(&old_entry->lock); |
431 | put_inotify_watch(&old->watch); /* pair to inotify_find_watch */ | 447 | fsnotify_destroy_mark(old_entry); |
432 | put_inotify_watch(&old->watch); /* and kill it */ | 448 | fsnotify_put_mark(old_entry); /* pair to fsnotify_find mark_entry */ |
449 | fsnotify_put_mark(old_entry); /* and kill it */ | ||
433 | return 0; | 450 | return 0; |
434 | } | 451 | } |
435 | 452 | ||
@@ -584,7 +601,9 @@ void audit_trim_trees(void) | |||
584 | 601 | ||
585 | spin_lock(&hash_lock); | 602 | spin_lock(&hash_lock); |
586 | list_for_each_entry(node, &tree->chunks, list) { | 603 | list_for_each_entry(node, &tree->chunks, list) { |
587 | struct inode *inode = find_chunk(node)->watch.inode; | 604 | struct audit_chunk *chunk = find_chunk(node); |
605 | /* this could be NULL if the watch is dieing else where... */ | ||
606 | struct inode *inode = chunk->mark.i.inode; | ||
588 | node->index |= 1U<<31; | 607 | node->index |= 1U<<31; |
589 | if (iterate_mounts(compare_root, inode, root_mnt)) | 608 | if (iterate_mounts(compare_root, inode, root_mnt)) |
590 | node->index &= ~(1U<<31); | 609 | node->index &= ~(1U<<31); |
@@ -846,7 +865,6 @@ void audit_kill_trees(struct list_head *list) | |||
846 | * Here comes the stuff asynchronous to auditctl operations | 865 | * Here comes the stuff asynchronous to auditctl operations |
847 | */ | 866 | */ |
848 | 867 | ||
849 | /* inode->inotify_mutex is locked */ | ||
850 | static void evict_chunk(struct audit_chunk *chunk) | 868 | static void evict_chunk(struct audit_chunk *chunk) |
851 | { | 869 | { |
852 | struct audit_tree *owner; | 870 | struct audit_tree *owner; |
@@ -885,35 +903,46 @@ static void evict_chunk(struct audit_chunk *chunk) | |||
885 | mutex_unlock(&audit_filter_mutex); | 903 | mutex_unlock(&audit_filter_mutex); |
886 | } | 904 | } |
887 | 905 | ||
888 | static void handle_event(struct inotify_watch *watch, u32 wd, u32 mask, | 906 | static int audit_tree_handle_event(struct fsnotify_group *group, |
889 | u32 cookie, const char *dname, struct inode *inode) | 907 | struct fsnotify_mark *inode_mark, |
908 | struct fsnotify_mark *vfsmonut_mark, | ||
909 | struct fsnotify_event *event) | ||
910 | { | ||
911 | BUG(); | ||
912 | return -EOPNOTSUPP; | ||
913 | } | ||
914 | |||
915 | static void audit_tree_freeing_mark(struct fsnotify_mark *entry, struct fsnotify_group *group) | ||
890 | { | 916 | { |
891 | struct audit_chunk *chunk = container_of(watch, struct audit_chunk, watch); | 917 | struct audit_chunk *chunk = container_of(entry, struct audit_chunk, mark); |
892 | 918 | ||
893 | if (mask & IN_IGNORED) { | 919 | evict_chunk(chunk); |
894 | evict_chunk(chunk); | 920 | fsnotify_put_mark(entry); |
895 | put_inotify_watch(watch); | ||
896 | } | ||
897 | } | 921 | } |
898 | 922 | ||
899 | static void destroy_watch(struct inotify_watch *watch) | 923 | static bool audit_tree_send_event(struct fsnotify_group *group, struct inode *inode, |
924 | struct fsnotify_mark *inode_mark, | ||
925 | struct fsnotify_mark *vfsmount_mark, | ||
926 | __u32 mask, void *data, int data_type) | ||
900 | { | 927 | { |
901 | struct audit_chunk *chunk = container_of(watch, struct audit_chunk, watch); | 928 | return false; |
902 | call_rcu(&chunk->head, __put_chunk); | ||
903 | } | 929 | } |
904 | 930 | ||
905 | static const struct inotify_operations rtree_inotify_ops = { | 931 | static const struct fsnotify_ops audit_tree_ops = { |
906 | .handle_event = handle_event, | 932 | .handle_event = audit_tree_handle_event, |
907 | .destroy_watch = destroy_watch, | 933 | .should_send_event = audit_tree_send_event, |
934 | .free_group_priv = NULL, | ||
935 | .free_event_priv = NULL, | ||
936 | .freeing_mark = audit_tree_freeing_mark, | ||
908 | }; | 937 | }; |
909 | 938 | ||
910 | static int __init audit_tree_init(void) | 939 | static int __init audit_tree_init(void) |
911 | { | 940 | { |
912 | int i; | 941 | int i; |
913 | 942 | ||
914 | rtree_ih = inotify_init(&rtree_inotify_ops); | 943 | audit_tree_group = fsnotify_alloc_group(&audit_tree_ops); |
915 | if (IS_ERR(rtree_ih)) | 944 | if (IS_ERR(audit_tree_group)) |
916 | audit_panic("cannot initialize inotify handle for rectree watches"); | 945 | audit_panic("cannot initialize fsnotify group for rectree watches"); |
917 | 946 | ||
918 | for (i = 0; i < HASH_SIZE; i++) | 947 | for (i = 0; i < HASH_SIZE; i++) |
919 | INIT_LIST_HEAD(&chunk_hash_heads[i]); | 948 | INIT_LIST_HEAD(&chunk_hash_heads[i]); |
diff --git a/kernel/audit_watch.c b/kernel/audit_watch.c index 8df43696f4ba..6bf2306be7d6 100644 --- a/kernel/audit_watch.c +++ b/kernel/audit_watch.c | |||
@@ -24,18 +24,18 @@ | |||
24 | #include <linux/kthread.h> | 24 | #include <linux/kthread.h> |
25 | #include <linux/mutex.h> | 25 | #include <linux/mutex.h> |
26 | #include <linux/fs.h> | 26 | #include <linux/fs.h> |
27 | #include <linux/fsnotify_backend.h> | ||
27 | #include <linux/namei.h> | 28 | #include <linux/namei.h> |
28 | #include <linux/netlink.h> | 29 | #include <linux/netlink.h> |
29 | #include <linux/sched.h> | 30 | #include <linux/sched.h> |
30 | #include <linux/slab.h> | 31 | #include <linux/slab.h> |
31 | #include <linux/inotify.h> | ||
32 | #include <linux/security.h> | 32 | #include <linux/security.h> |
33 | #include "audit.h" | 33 | #include "audit.h" |
34 | 34 | ||
35 | /* | 35 | /* |
36 | * Reference counting: | 36 | * Reference counting: |
37 | * | 37 | * |
38 | * audit_parent: lifetime is from audit_init_parent() to receipt of an IN_IGNORED | 38 | * audit_parent: lifetime is from audit_init_parent() to receipt of an FS_IGNORED |
39 | * event. Each audit_watch holds a reference to its associated parent. | 39 | * event. Each audit_watch holds a reference to its associated parent. |
40 | * | 40 | * |
41 | * audit_watch: if added to lists, lifetime is from audit_init_watch() to | 41 | * audit_watch: if added to lists, lifetime is from audit_init_watch() to |
@@ -51,40 +51,61 @@ struct audit_watch { | |||
51 | unsigned long ino; /* associated inode number */ | 51 | unsigned long ino; /* associated inode number */ |
52 | struct audit_parent *parent; /* associated parent */ | 52 | struct audit_parent *parent; /* associated parent */ |
53 | struct list_head wlist; /* entry in parent->watches list */ | 53 | struct list_head wlist; /* entry in parent->watches list */ |
54 | struct list_head rules; /* associated rules */ | 54 | struct list_head rules; /* anchor for krule->rlist */ |
55 | }; | 55 | }; |
56 | 56 | ||
57 | struct audit_parent { | 57 | struct audit_parent { |
58 | struct list_head ilist; /* entry in inotify registration list */ | 58 | struct list_head watches; /* anchor for audit_watch->wlist */ |
59 | struct list_head watches; /* associated watches */ | 59 | struct fsnotify_mark mark; /* fsnotify mark on the inode */ |
60 | struct inotify_watch wdata; /* inotify watch data */ | ||
61 | unsigned flags; /* status flags */ | ||
62 | }; | 60 | }; |
63 | 61 | ||
64 | /* Inotify handle. */ | 62 | /* fsnotify handle. */ |
65 | struct inotify_handle *audit_ih; | 63 | struct fsnotify_group *audit_watch_group; |
66 | 64 | ||
67 | /* | 65 | /* fsnotify events we care about. */ |
68 | * audit_parent status flags: | 66 | #define AUDIT_FS_WATCH (FS_MOVE | FS_CREATE | FS_DELETE | FS_DELETE_SELF |\ |
69 | * | 67 | FS_MOVE_SELF | FS_EVENT_ON_CHILD) |
70 | * AUDIT_PARENT_INVALID - set anytime rules/watches are auto-removed due to | ||
71 | * a filesystem event to ensure we're adding audit watches to a valid parent. | ||
72 | * Technically not needed for IN_DELETE_SELF or IN_UNMOUNT events, as we cannot | ||
73 | * receive them while we have nameidata, but must be used for IN_MOVE_SELF which | ||
74 | * we can receive while holding nameidata. | ||
75 | */ | ||
76 | #define AUDIT_PARENT_INVALID 0x001 | ||
77 | 68 | ||
78 | /* Inotify events we care about. */ | 69 | static void audit_free_parent(struct audit_parent *parent) |
79 | #define AUDIT_IN_WATCH IN_MOVE|IN_CREATE|IN_DELETE|IN_DELETE_SELF|IN_MOVE_SELF | 70 | { |
71 | WARN_ON(!list_empty(&parent->watches)); | ||
72 | kfree(parent); | ||
73 | } | ||
80 | 74 | ||
81 | static void audit_free_parent(struct inotify_watch *i_watch) | 75 | static void audit_watch_free_mark(struct fsnotify_mark *entry) |
82 | { | 76 | { |
83 | struct audit_parent *parent; | 77 | struct audit_parent *parent; |
84 | 78 | ||
85 | parent = container_of(i_watch, struct audit_parent, wdata); | 79 | parent = container_of(entry, struct audit_parent, mark); |
86 | WARN_ON(!list_empty(&parent->watches)); | 80 | audit_free_parent(parent); |
87 | kfree(parent); | 81 | } |
82 | |||
83 | static void audit_get_parent(struct audit_parent *parent) | ||
84 | { | ||
85 | if (likely(parent)) | ||
86 | fsnotify_get_mark(&parent->mark); | ||
87 | } | ||
88 | |||
89 | static void audit_put_parent(struct audit_parent *parent) | ||
90 | { | ||
91 | if (likely(parent)) | ||
92 | fsnotify_put_mark(&parent->mark); | ||
93 | } | ||
94 | |||
95 | /* | ||
96 | * Find and return the audit_parent on the given inode. If found a reference | ||
97 | * is taken on this parent. | ||
98 | */ | ||
99 | static inline struct audit_parent *audit_find_parent(struct inode *inode) | ||
100 | { | ||
101 | struct audit_parent *parent = NULL; | ||
102 | struct fsnotify_mark *entry; | ||
103 | |||
104 | entry = fsnotify_find_inode_mark(audit_watch_group, inode); | ||
105 | if (entry) | ||
106 | parent = container_of(entry, struct audit_parent, mark); | ||
107 | |||
108 | return parent; | ||
88 | } | 109 | } |
89 | 110 | ||
90 | void audit_get_watch(struct audit_watch *watch) | 111 | void audit_get_watch(struct audit_watch *watch) |
@@ -105,7 +126,7 @@ void audit_put_watch(struct audit_watch *watch) | |||
105 | void audit_remove_watch(struct audit_watch *watch) | 126 | void audit_remove_watch(struct audit_watch *watch) |
106 | { | 127 | { |
107 | list_del(&watch->wlist); | 128 | list_del(&watch->wlist); |
108 | put_inotify_watch(&watch->parent->wdata); | 129 | audit_put_parent(watch->parent); |
109 | watch->parent = NULL; | 130 | watch->parent = NULL; |
110 | audit_put_watch(watch); /* match initial get */ | 131 | audit_put_watch(watch); /* match initial get */ |
111 | } | 132 | } |
@@ -115,42 +136,32 @@ char *audit_watch_path(struct audit_watch *watch) | |||
115 | return watch->path; | 136 | return watch->path; |
116 | } | 137 | } |
117 | 138 | ||
118 | struct list_head *audit_watch_rules(struct audit_watch *watch) | 139 | int audit_watch_compare(struct audit_watch *watch, unsigned long ino, dev_t dev) |
119 | { | ||
120 | return &watch->rules; | ||
121 | } | ||
122 | |||
123 | unsigned long audit_watch_inode(struct audit_watch *watch) | ||
124 | { | 140 | { |
125 | return watch->ino; | 141 | return (watch->ino != (unsigned long)-1) && |
126 | } | 142 | (watch->ino == ino) && |
127 | 143 | (watch->dev == dev); | |
128 | dev_t audit_watch_dev(struct audit_watch *watch) | ||
129 | { | ||
130 | return watch->dev; | ||
131 | } | 144 | } |
132 | 145 | ||
133 | /* Initialize a parent watch entry. */ | 146 | /* Initialize a parent watch entry. */ |
134 | static struct audit_parent *audit_init_parent(struct nameidata *ndp) | 147 | static struct audit_parent *audit_init_parent(struct nameidata *ndp) |
135 | { | 148 | { |
149 | struct inode *inode = ndp->path.dentry->d_inode; | ||
136 | struct audit_parent *parent; | 150 | struct audit_parent *parent; |
137 | s32 wd; | 151 | int ret; |
138 | 152 | ||
139 | parent = kzalloc(sizeof(*parent), GFP_KERNEL); | 153 | parent = kzalloc(sizeof(*parent), GFP_KERNEL); |
140 | if (unlikely(!parent)) | 154 | if (unlikely(!parent)) |
141 | return ERR_PTR(-ENOMEM); | 155 | return ERR_PTR(-ENOMEM); |
142 | 156 | ||
143 | INIT_LIST_HEAD(&parent->watches); | 157 | INIT_LIST_HEAD(&parent->watches); |
144 | parent->flags = 0; | 158 | |
145 | 159 | fsnotify_init_mark(&parent->mark, audit_watch_free_mark); | |
146 | inotify_init_watch(&parent->wdata); | 160 | parent->mark.mask = AUDIT_FS_WATCH; |
147 | /* grab a ref so inotify watch hangs around until we take audit_filter_mutex */ | 161 | ret = fsnotify_add_mark(&parent->mark, audit_watch_group, inode, NULL, 0); |
148 | get_inotify_watch(&parent->wdata); | 162 | if (ret < 0) { |
149 | wd = inotify_add_watch(audit_ih, &parent->wdata, | 163 | audit_free_parent(parent); |
150 | ndp->path.dentry->d_inode, AUDIT_IN_WATCH); | 164 | return ERR_PTR(ret); |
151 | if (wd < 0) { | ||
152 | audit_free_parent(&parent->wdata); | ||
153 | return ERR_PTR(wd); | ||
154 | } | 165 | } |
155 | 166 | ||
156 | return parent; | 167 | return parent; |
@@ -179,7 +190,7 @@ int audit_to_watch(struct audit_krule *krule, char *path, int len, u32 op) | |||
179 | { | 190 | { |
180 | struct audit_watch *watch; | 191 | struct audit_watch *watch; |
181 | 192 | ||
182 | if (!audit_ih) | 193 | if (!audit_watch_group) |
183 | return -EOPNOTSUPP; | 194 | return -EOPNOTSUPP; |
184 | 195 | ||
185 | if (path[0] != '/' || path[len-1] == '/' || | 196 | if (path[0] != '/' || path[len-1] == '/' || |
@@ -217,7 +228,7 @@ static struct audit_watch *audit_dupe_watch(struct audit_watch *old) | |||
217 | 228 | ||
218 | new->dev = old->dev; | 229 | new->dev = old->dev; |
219 | new->ino = old->ino; | 230 | new->ino = old->ino; |
220 | get_inotify_watch(&old->parent->wdata); | 231 | audit_get_parent(old->parent); |
221 | new->parent = old->parent; | 232 | new->parent = old->parent; |
222 | 233 | ||
223 | out: | 234 | out: |
@@ -251,15 +262,19 @@ static void audit_update_watch(struct audit_parent *parent, | |||
251 | struct audit_entry *oentry, *nentry; | 262 | struct audit_entry *oentry, *nentry; |
252 | 263 | ||
253 | mutex_lock(&audit_filter_mutex); | 264 | mutex_lock(&audit_filter_mutex); |
265 | /* Run all of the watches on this parent looking for the one that | ||
266 | * matches the given dname */ | ||
254 | list_for_each_entry_safe(owatch, nextw, &parent->watches, wlist) { | 267 | list_for_each_entry_safe(owatch, nextw, &parent->watches, wlist) { |
255 | if (audit_compare_dname_path(dname, owatch->path, NULL)) | 268 | if (audit_compare_dname_path(dname, owatch->path, NULL)) |
256 | continue; | 269 | continue; |
257 | 270 | ||
258 | /* If the update involves invalidating rules, do the inode-based | 271 | /* If the update involves invalidating rules, do the inode-based |
259 | * filtering now, so we don't omit records. */ | 272 | * filtering now, so we don't omit records. */ |
260 | if (invalidating && current->audit_context) | 273 | if (invalidating && !audit_dummy_context()) |
261 | audit_filter_inodes(current, current->audit_context); | 274 | audit_filter_inodes(current, current->audit_context); |
262 | 275 | ||
276 | /* updating ino will likely change which audit_hash_list we | ||
277 | * are on so we need a new watch for the new list */ | ||
263 | nwatch = audit_dupe_watch(owatch); | 278 | nwatch = audit_dupe_watch(owatch); |
264 | if (IS_ERR(nwatch)) { | 279 | if (IS_ERR(nwatch)) { |
265 | mutex_unlock(&audit_filter_mutex); | 280 | mutex_unlock(&audit_filter_mutex); |
@@ -275,12 +290,21 @@ static void audit_update_watch(struct audit_parent *parent, | |||
275 | list_del(&oentry->rule.rlist); | 290 | list_del(&oentry->rule.rlist); |
276 | list_del_rcu(&oentry->list); | 291 | list_del_rcu(&oentry->list); |
277 | 292 | ||
278 | nentry = audit_dupe_rule(&oentry->rule, nwatch); | 293 | nentry = audit_dupe_rule(&oentry->rule); |
279 | if (IS_ERR(nentry)) { | 294 | if (IS_ERR(nentry)) { |
280 | list_del(&oentry->rule.list); | 295 | list_del(&oentry->rule.list); |
281 | audit_panic("error updating watch, removing"); | 296 | audit_panic("error updating watch, removing"); |
282 | } else { | 297 | } else { |
283 | int h = audit_hash_ino((u32)ino); | 298 | int h = audit_hash_ino((u32)ino); |
299 | |||
300 | /* | ||
301 | * nentry->rule.watch == oentry->rule.watch so | ||
302 | * we must drop that reference and set it to our | ||
303 | * new watch. | ||
304 | */ | ||
305 | audit_put_watch(nentry->rule.watch); | ||
306 | audit_get_watch(nwatch); | ||
307 | nentry->rule.watch = nwatch; | ||
284 | list_add(&nentry->rule.rlist, &nwatch->rules); | 308 | list_add(&nentry->rule.rlist, &nwatch->rules); |
285 | list_add_rcu(&nentry->list, &audit_inode_hash[h]); | 309 | list_add_rcu(&nentry->list, &audit_inode_hash[h]); |
286 | list_replace(&oentry->rule.list, | 310 | list_replace(&oentry->rule.list, |
@@ -312,7 +336,6 @@ static void audit_remove_parent_watches(struct audit_parent *parent) | |||
312 | struct audit_entry *e; | 336 | struct audit_entry *e; |
313 | 337 | ||
314 | mutex_lock(&audit_filter_mutex); | 338 | mutex_lock(&audit_filter_mutex); |
315 | parent->flags |= AUDIT_PARENT_INVALID; | ||
316 | list_for_each_entry_safe(w, nextw, &parent->watches, wlist) { | 339 | list_for_each_entry_safe(w, nextw, &parent->watches, wlist) { |
317 | list_for_each_entry_safe(r, nextr, &w->rules, rlist) { | 340 | list_for_each_entry_safe(r, nextr, &w->rules, rlist) { |
318 | e = container_of(r, struct audit_entry, rule); | 341 | e = container_of(r, struct audit_entry, rule); |
@@ -325,20 +348,8 @@ static void audit_remove_parent_watches(struct audit_parent *parent) | |||
325 | audit_remove_watch(w); | 348 | audit_remove_watch(w); |
326 | } | 349 | } |
327 | mutex_unlock(&audit_filter_mutex); | 350 | mutex_unlock(&audit_filter_mutex); |
328 | } | ||
329 | |||
330 | /* Unregister inotify watches for parents on in_list. | ||
331 | * Generates an IN_IGNORED event. */ | ||
332 | void audit_inotify_unregister(struct list_head *in_list) | ||
333 | { | ||
334 | struct audit_parent *p, *n; | ||
335 | 351 | ||
336 | list_for_each_entry_safe(p, n, in_list, ilist) { | 352 | fsnotify_destroy_mark(&parent->mark); |
337 | list_del(&p->ilist); | ||
338 | inotify_rm_watch(audit_ih, &p->wdata); | ||
339 | /* the unpin matching the pin in audit_do_del_rule() */ | ||
340 | unpin_inotify_watch(&p->wdata); | ||
341 | } | ||
342 | } | 353 | } |
343 | 354 | ||
344 | /* Get path information necessary for adding watches. */ | 355 | /* Get path information necessary for adding watches. */ |
@@ -389,7 +400,7 @@ static void audit_put_nd(struct nameidata *ndp, struct nameidata *ndw) | |||
389 | } | 400 | } |
390 | } | 401 | } |
391 | 402 | ||
392 | /* Associate the given rule with an existing parent inotify_watch. | 403 | /* Associate the given rule with an existing parent. |
393 | * Caller must hold audit_filter_mutex. */ | 404 | * Caller must hold audit_filter_mutex. */ |
394 | static void audit_add_to_parent(struct audit_krule *krule, | 405 | static void audit_add_to_parent(struct audit_krule *krule, |
395 | struct audit_parent *parent) | 406 | struct audit_parent *parent) |
@@ -397,6 +408,8 @@ static void audit_add_to_parent(struct audit_krule *krule, | |||
397 | struct audit_watch *w, *watch = krule->watch; | 408 | struct audit_watch *w, *watch = krule->watch; |
398 | int watch_found = 0; | 409 | int watch_found = 0; |
399 | 410 | ||
411 | BUG_ON(!mutex_is_locked(&audit_filter_mutex)); | ||
412 | |||
400 | list_for_each_entry(w, &parent->watches, wlist) { | 413 | list_for_each_entry(w, &parent->watches, wlist) { |
401 | if (strcmp(watch->path, w->path)) | 414 | if (strcmp(watch->path, w->path)) |
402 | continue; | 415 | continue; |
@@ -413,7 +426,7 @@ static void audit_add_to_parent(struct audit_krule *krule, | |||
413 | } | 426 | } |
414 | 427 | ||
415 | if (!watch_found) { | 428 | if (!watch_found) { |
416 | get_inotify_watch(&parent->wdata); | 429 | audit_get_parent(parent); |
417 | watch->parent = parent; | 430 | watch->parent = parent; |
418 | 431 | ||
419 | list_add(&watch->wlist, &parent->watches); | 432 | list_add(&watch->wlist, &parent->watches); |
@@ -423,13 +436,12 @@ static void audit_add_to_parent(struct audit_krule *krule, | |||
423 | 436 | ||
424 | /* Find a matching watch entry, or add this one. | 437 | /* Find a matching watch entry, or add this one. |
425 | * Caller must hold audit_filter_mutex. */ | 438 | * Caller must hold audit_filter_mutex. */ |
426 | int audit_add_watch(struct audit_krule *krule) | 439 | int audit_add_watch(struct audit_krule *krule, struct list_head **list) |
427 | { | 440 | { |
428 | struct audit_watch *watch = krule->watch; | 441 | struct audit_watch *watch = krule->watch; |
429 | struct inotify_watch *i_watch; | ||
430 | struct audit_parent *parent; | 442 | struct audit_parent *parent; |
431 | struct nameidata *ndp = NULL, *ndw = NULL; | 443 | struct nameidata *ndp = NULL, *ndw = NULL; |
432 | int ret = 0; | 444 | int h, ret = 0; |
433 | 445 | ||
434 | mutex_unlock(&audit_filter_mutex); | 446 | mutex_unlock(&audit_filter_mutex); |
435 | 447 | ||
@@ -441,47 +453,38 @@ int audit_add_watch(struct audit_krule *krule) | |||
441 | goto error; | 453 | goto error; |
442 | } | 454 | } |
443 | 455 | ||
456 | mutex_lock(&audit_filter_mutex); | ||
457 | |||
444 | /* update watch filter fields */ | 458 | /* update watch filter fields */ |
445 | if (ndw) { | 459 | if (ndw) { |
446 | watch->dev = ndw->path.dentry->d_inode->i_sb->s_dev; | 460 | watch->dev = ndw->path.dentry->d_inode->i_sb->s_dev; |
447 | watch->ino = ndw->path.dentry->d_inode->i_ino; | 461 | watch->ino = ndw->path.dentry->d_inode->i_ino; |
448 | } | 462 | } |
449 | 463 | ||
450 | /* The audit_filter_mutex must not be held during inotify calls because | 464 | /* either find an old parent or attach a new one */ |
451 | * we hold it during inotify event callback processing. If an existing | 465 | parent = audit_find_parent(ndp->path.dentry->d_inode); |
452 | * inotify watch is found, inotify_find_watch() grabs a reference before | 466 | if (!parent) { |
453 | * returning. | ||
454 | */ | ||
455 | if (inotify_find_watch(audit_ih, ndp->path.dentry->d_inode, | ||
456 | &i_watch) < 0) { | ||
457 | parent = audit_init_parent(ndp); | 467 | parent = audit_init_parent(ndp); |
458 | if (IS_ERR(parent)) { | 468 | if (IS_ERR(parent)) { |
459 | /* caller expects mutex locked */ | ||
460 | mutex_lock(&audit_filter_mutex); | ||
461 | ret = PTR_ERR(parent); | 469 | ret = PTR_ERR(parent); |
462 | goto error; | 470 | goto error; |
463 | } | 471 | } |
464 | } else | 472 | } |
465 | parent = container_of(i_watch, struct audit_parent, wdata); | ||
466 | |||
467 | mutex_lock(&audit_filter_mutex); | ||
468 | 473 | ||
469 | /* parent was moved before we took audit_filter_mutex */ | 474 | audit_add_to_parent(krule, parent); |
470 | if (parent->flags & AUDIT_PARENT_INVALID) | ||
471 | ret = -ENOENT; | ||
472 | else | ||
473 | audit_add_to_parent(krule, parent); | ||
474 | 475 | ||
475 | /* match get in audit_init_parent or inotify_find_watch */ | 476 | /* match get in audit_find_parent or audit_init_parent */ |
476 | put_inotify_watch(&parent->wdata); | 477 | audit_put_parent(parent); |
477 | 478 | ||
479 | h = audit_hash_ino((u32)watch->ino); | ||
480 | *list = &audit_inode_hash[h]; | ||
478 | error: | 481 | error: |
479 | audit_put_nd(ndp, ndw); /* NULL args OK */ | 482 | audit_put_nd(ndp, ndw); /* NULL args OK */ |
480 | return ret; | 483 | return ret; |
481 | 484 | ||
482 | } | 485 | } |
483 | 486 | ||
484 | void audit_remove_watch_rule(struct audit_krule *krule, struct list_head *list) | 487 | void audit_remove_watch_rule(struct audit_krule *krule) |
485 | { | 488 | { |
486 | struct audit_watch *watch = krule->watch; | 489 | struct audit_watch *watch = krule->watch; |
487 | struct audit_parent *parent = watch->parent; | 490 | struct audit_parent *parent = watch->parent; |
@@ -492,53 +495,74 @@ void audit_remove_watch_rule(struct audit_krule *krule, struct list_head *list) | |||
492 | audit_remove_watch(watch); | 495 | audit_remove_watch(watch); |
493 | 496 | ||
494 | if (list_empty(&parent->watches)) { | 497 | if (list_empty(&parent->watches)) { |
495 | /* Put parent on the inotify un-registration | 498 | audit_get_parent(parent); |
496 | * list. Grab a reference before releasing | 499 | fsnotify_destroy_mark(&parent->mark); |
497 | * audit_filter_mutex, to be released in | 500 | audit_put_parent(parent); |
498 | * audit_inotify_unregister(). | ||
499 | * If filesystem is going away, just leave | ||
500 | * the sucker alone, eviction will take | ||
501 | * care of it. */ | ||
502 | if (pin_inotify_watch(&parent->wdata)) | ||
503 | list_add(&parent->ilist, list); | ||
504 | } | 501 | } |
505 | } | 502 | } |
506 | } | 503 | } |
507 | 504 | ||
508 | /* Update watch data in audit rules based on inotify events. */ | 505 | static bool audit_watch_should_send_event(struct fsnotify_group *group, struct inode *inode, |
509 | static void audit_handle_ievent(struct inotify_watch *i_watch, u32 wd, u32 mask, | 506 | struct fsnotify_mark *inode_mark, |
510 | u32 cookie, const char *dname, struct inode *inode) | 507 | struct fsnotify_mark *vfsmount_mark, |
508 | __u32 mask, void *data, int data_type) | ||
509 | { | ||
510 | return true; | ||
511 | } | ||
512 | |||
513 | /* Update watch data in audit rules based on fsnotify events. */ | ||
514 | static int audit_watch_handle_event(struct fsnotify_group *group, | ||
515 | struct fsnotify_mark *inode_mark, | ||
516 | struct fsnotify_mark *vfsmount_mark, | ||
517 | struct fsnotify_event *event) | ||
511 | { | 518 | { |
519 | struct inode *inode; | ||
520 | __u32 mask = event->mask; | ||
521 | const char *dname = event->file_name; | ||
512 | struct audit_parent *parent; | 522 | struct audit_parent *parent; |
513 | 523 | ||
514 | parent = container_of(i_watch, struct audit_parent, wdata); | 524 | parent = container_of(inode_mark, struct audit_parent, mark); |
515 | 525 | ||
516 | if (mask & (IN_CREATE|IN_MOVED_TO) && inode) | 526 | BUG_ON(group != audit_watch_group); |
517 | audit_update_watch(parent, dname, inode->i_sb->s_dev, | 527 | |
518 | inode->i_ino, 0); | 528 | switch (event->data_type) { |
519 | else if (mask & (IN_DELETE|IN_MOVED_FROM)) | 529 | case (FSNOTIFY_EVENT_FILE): |
530 | inode = event->file->f_path.dentry->d_inode; | ||
531 | break; | ||
532 | case (FSNOTIFY_EVENT_INODE): | ||
533 | inode = event->inode; | ||
534 | break; | ||
535 | default: | ||
536 | BUG(); | ||
537 | inode = NULL; | ||
538 | break; | ||
539 | }; | ||
540 | |||
541 | if (mask & (FS_CREATE|FS_MOVED_TO) && inode) | ||
542 | audit_update_watch(parent, dname, inode->i_sb->s_dev, inode->i_ino, 0); | ||
543 | else if (mask & (FS_DELETE|FS_MOVED_FROM)) | ||
520 | audit_update_watch(parent, dname, (dev_t)-1, (unsigned long)-1, 1); | 544 | audit_update_watch(parent, dname, (dev_t)-1, (unsigned long)-1, 1); |
521 | /* inotify automatically removes the watch and sends IN_IGNORED */ | 545 | else if (mask & (FS_DELETE_SELF|FS_UNMOUNT|FS_MOVE_SELF)) |
522 | else if (mask & (IN_DELETE_SELF|IN_UNMOUNT)) | ||
523 | audit_remove_parent_watches(parent); | ||
524 | /* inotify does not remove the watch, so remove it manually */ | ||
525 | else if(mask & IN_MOVE_SELF) { | ||
526 | audit_remove_parent_watches(parent); | 546 | audit_remove_parent_watches(parent); |
527 | inotify_remove_watch_locked(audit_ih, i_watch); | 547 | |
528 | } else if (mask & IN_IGNORED) | 548 | return 0; |
529 | put_inotify_watch(i_watch); | ||
530 | } | 549 | } |
531 | 550 | ||
532 | static const struct inotify_operations audit_inotify_ops = { | 551 | static const struct fsnotify_ops audit_watch_fsnotify_ops = { |
533 | .handle_event = audit_handle_ievent, | 552 | .should_send_event = audit_watch_should_send_event, |
534 | .destroy_watch = audit_free_parent, | 553 | .handle_event = audit_watch_handle_event, |
554 | .free_group_priv = NULL, | ||
555 | .freeing_mark = NULL, | ||
556 | .free_event_priv = NULL, | ||
535 | }; | 557 | }; |
536 | 558 | ||
537 | static int __init audit_watch_init(void) | 559 | static int __init audit_watch_init(void) |
538 | { | 560 | { |
539 | audit_ih = inotify_init(&audit_inotify_ops); | 561 | audit_watch_group = fsnotify_alloc_group(&audit_watch_fsnotify_ops); |
540 | if (IS_ERR(audit_ih)) | 562 | if (IS_ERR(audit_watch_group)) { |
541 | audit_panic("cannot initialize inotify handle"); | 563 | audit_watch_group = NULL; |
564 | audit_panic("cannot create audit fsnotify group"); | ||
565 | } | ||
542 | return 0; | 566 | return 0; |
543 | } | 567 | } |
544 | subsys_initcall(audit_watch_init); | 568 | device_initcall(audit_watch_init); |
diff --git a/kernel/auditfilter.c b/kernel/auditfilter.c index ce08041f578d..eb7675499fb5 100644 --- a/kernel/auditfilter.c +++ b/kernel/auditfilter.c | |||
@@ -71,6 +71,7 @@ static inline void audit_free_rule(struct audit_entry *e) | |||
71 | { | 71 | { |
72 | int i; | 72 | int i; |
73 | struct audit_krule *erule = &e->rule; | 73 | struct audit_krule *erule = &e->rule; |
74 | |||
74 | /* some rules don't have associated watches */ | 75 | /* some rules don't have associated watches */ |
75 | if (erule->watch) | 76 | if (erule->watch) |
76 | audit_put_watch(erule->watch); | 77 | audit_put_watch(erule->watch); |
@@ -746,8 +747,7 @@ static inline int audit_dupe_lsm_field(struct audit_field *df, | |||
746 | * rule with the new rule in the filterlist, then free the old rule. | 747 | * rule with the new rule in the filterlist, then free the old rule. |
747 | * The rlist element is undefined; list manipulations are handled apart from | 748 | * The rlist element is undefined; list manipulations are handled apart from |
748 | * the initial copy. */ | 749 | * the initial copy. */ |
749 | struct audit_entry *audit_dupe_rule(struct audit_krule *old, | 750 | struct audit_entry *audit_dupe_rule(struct audit_krule *old) |
750 | struct audit_watch *watch) | ||
751 | { | 751 | { |
752 | u32 fcount = old->field_count; | 752 | u32 fcount = old->field_count; |
753 | struct audit_entry *entry; | 753 | struct audit_entry *entry; |
@@ -769,8 +769,8 @@ struct audit_entry *audit_dupe_rule(struct audit_krule *old, | |||
769 | new->prio = old->prio; | 769 | new->prio = old->prio; |
770 | new->buflen = old->buflen; | 770 | new->buflen = old->buflen; |
771 | new->inode_f = old->inode_f; | 771 | new->inode_f = old->inode_f; |
772 | new->watch = NULL; | ||
773 | new->field_count = old->field_count; | 772 | new->field_count = old->field_count; |
773 | |||
774 | /* | 774 | /* |
775 | * note that we are OK with not refcounting here; audit_match_tree() | 775 | * note that we are OK with not refcounting here; audit_match_tree() |
776 | * never dereferences tree and we can't get false positives there | 776 | * never dereferences tree and we can't get false positives there |
@@ -811,9 +811,9 @@ struct audit_entry *audit_dupe_rule(struct audit_krule *old, | |||
811 | } | 811 | } |
812 | } | 812 | } |
813 | 813 | ||
814 | if (watch) { | 814 | if (old->watch) { |
815 | audit_get_watch(watch); | 815 | audit_get_watch(old->watch); |
816 | new->watch = watch; | 816 | new->watch = old->watch; |
817 | } | 817 | } |
818 | 818 | ||
819 | return entry; | 819 | return entry; |
@@ -866,7 +866,7 @@ static inline int audit_add_rule(struct audit_entry *entry) | |||
866 | struct audit_watch *watch = entry->rule.watch; | 866 | struct audit_watch *watch = entry->rule.watch; |
867 | struct audit_tree *tree = entry->rule.tree; | 867 | struct audit_tree *tree = entry->rule.tree; |
868 | struct list_head *list; | 868 | struct list_head *list; |
869 | int h, err; | 869 | int err; |
870 | #ifdef CONFIG_AUDITSYSCALL | 870 | #ifdef CONFIG_AUDITSYSCALL |
871 | int dont_count = 0; | 871 | int dont_count = 0; |
872 | 872 | ||
@@ -889,15 +889,11 @@ static inline int audit_add_rule(struct audit_entry *entry) | |||
889 | 889 | ||
890 | if (watch) { | 890 | if (watch) { |
891 | /* audit_filter_mutex is dropped and re-taken during this call */ | 891 | /* audit_filter_mutex is dropped and re-taken during this call */ |
892 | err = audit_add_watch(&entry->rule); | 892 | err = audit_add_watch(&entry->rule, &list); |
893 | if (err) { | 893 | if (err) { |
894 | mutex_unlock(&audit_filter_mutex); | 894 | mutex_unlock(&audit_filter_mutex); |
895 | goto error; | 895 | goto error; |
896 | } | 896 | } |
897 | /* entry->rule.watch may have changed during audit_add_watch() */ | ||
898 | watch = entry->rule.watch; | ||
899 | h = audit_hash_ino((u32)audit_watch_inode(watch)); | ||
900 | list = &audit_inode_hash[h]; | ||
901 | } | 897 | } |
902 | if (tree) { | 898 | if (tree) { |
903 | err = audit_add_tree_rule(&entry->rule); | 899 | err = audit_add_tree_rule(&entry->rule); |
@@ -949,7 +945,6 @@ static inline int audit_del_rule(struct audit_entry *entry) | |||
949 | struct audit_watch *watch = entry->rule.watch; | 945 | struct audit_watch *watch = entry->rule.watch; |
950 | struct audit_tree *tree = entry->rule.tree; | 946 | struct audit_tree *tree = entry->rule.tree; |
951 | struct list_head *list; | 947 | struct list_head *list; |
952 | LIST_HEAD(inotify_list); | ||
953 | int ret = 0; | 948 | int ret = 0; |
954 | #ifdef CONFIG_AUDITSYSCALL | 949 | #ifdef CONFIG_AUDITSYSCALL |
955 | int dont_count = 0; | 950 | int dont_count = 0; |
@@ -969,7 +964,7 @@ static inline int audit_del_rule(struct audit_entry *entry) | |||
969 | } | 964 | } |
970 | 965 | ||
971 | if (e->rule.watch) | 966 | if (e->rule.watch) |
972 | audit_remove_watch_rule(&e->rule, &inotify_list); | 967 | audit_remove_watch_rule(&e->rule); |
973 | 968 | ||
974 | if (e->rule.tree) | 969 | if (e->rule.tree) |
975 | audit_remove_tree_rule(&e->rule); | 970 | audit_remove_tree_rule(&e->rule); |
@@ -987,9 +982,6 @@ static inline int audit_del_rule(struct audit_entry *entry) | |||
987 | #endif | 982 | #endif |
988 | mutex_unlock(&audit_filter_mutex); | 983 | mutex_unlock(&audit_filter_mutex); |
989 | 984 | ||
990 | if (!list_empty(&inotify_list)) | ||
991 | audit_inotify_unregister(&inotify_list); | ||
992 | |||
993 | out: | 985 | out: |
994 | if (watch) | 986 | if (watch) |
995 | audit_put_watch(watch); /* match initial get */ | 987 | audit_put_watch(watch); /* match initial get */ |
@@ -1323,30 +1315,23 @@ static int update_lsm_rule(struct audit_krule *r) | |||
1323 | { | 1315 | { |
1324 | struct audit_entry *entry = container_of(r, struct audit_entry, rule); | 1316 | struct audit_entry *entry = container_of(r, struct audit_entry, rule); |
1325 | struct audit_entry *nentry; | 1317 | struct audit_entry *nentry; |
1326 | struct audit_watch *watch; | ||
1327 | struct audit_tree *tree; | ||
1328 | int err = 0; | 1318 | int err = 0; |
1329 | 1319 | ||
1330 | if (!security_audit_rule_known(r)) | 1320 | if (!security_audit_rule_known(r)) |
1331 | return 0; | 1321 | return 0; |
1332 | 1322 | ||
1333 | watch = r->watch; | 1323 | nentry = audit_dupe_rule(r); |
1334 | tree = r->tree; | ||
1335 | nentry = audit_dupe_rule(r, watch); | ||
1336 | if (IS_ERR(nentry)) { | 1324 | if (IS_ERR(nentry)) { |
1337 | /* save the first error encountered for the | 1325 | /* save the first error encountered for the |
1338 | * return value */ | 1326 | * return value */ |
1339 | err = PTR_ERR(nentry); | 1327 | err = PTR_ERR(nentry); |
1340 | audit_panic("error updating LSM filters"); | 1328 | audit_panic("error updating LSM filters"); |
1341 | if (watch) | 1329 | if (r->watch) |
1342 | list_del(&r->rlist); | 1330 | list_del(&r->rlist); |
1343 | list_del_rcu(&entry->list); | 1331 | list_del_rcu(&entry->list); |
1344 | list_del(&r->list); | 1332 | list_del(&r->list); |
1345 | } else { | 1333 | } else { |
1346 | if (watch) { | 1334 | if (r->watch || r->tree) |
1347 | list_add(&nentry->rule.rlist, audit_watch_rules(watch)); | ||
1348 | list_del(&r->rlist); | ||
1349 | } else if (tree) | ||
1350 | list_replace_init(&r->rlist, &nentry->rule.rlist); | 1335 | list_replace_init(&r->rlist, &nentry->rule.rlist); |
1351 | list_replace_rcu(&entry->list, &nentry->list); | 1336 | list_replace_rcu(&entry->list, &nentry->list); |
1352 | list_replace(&r->list, &nentry->rule.list); | 1337 | list_replace(&r->list, &nentry->rule.list); |
diff --git a/kernel/auditsc.c b/kernel/auditsc.c index 3828ad5fb8f1..b87a63beb66c 100644 --- a/kernel/auditsc.c +++ b/kernel/auditsc.c | |||
@@ -65,7 +65,6 @@ | |||
65 | #include <linux/binfmts.h> | 65 | #include <linux/binfmts.h> |
66 | #include <linux/highmem.h> | 66 | #include <linux/highmem.h> |
67 | #include <linux/syscalls.h> | 67 | #include <linux/syscalls.h> |
68 | #include <linux/inotify.h> | ||
69 | #include <linux/capability.h> | 68 | #include <linux/capability.h> |
70 | #include <linux/fs_struct.h> | 69 | #include <linux/fs_struct.h> |
71 | 70 | ||
@@ -549,9 +548,8 @@ static int audit_filter_rules(struct task_struct *tsk, | |||
549 | } | 548 | } |
550 | break; | 549 | break; |
551 | case AUDIT_WATCH: | 550 | case AUDIT_WATCH: |
552 | if (name && audit_watch_inode(rule->watch) != (unsigned long)-1) | 551 | if (name) |
553 | result = (name->dev == audit_watch_dev(rule->watch) && | 552 | result = audit_watch_compare(rule->watch, name->ino, name->dev); |
554 | name->ino == audit_watch_inode(rule->watch)); | ||
555 | break; | 553 | break; |
556 | case AUDIT_DIR: | 554 | case AUDIT_DIR: |
557 | if (ctx) | 555 | if (ctx) |
@@ -1726,7 +1724,7 @@ static inline void handle_one(const struct inode *inode) | |||
1726 | struct audit_tree_refs *p; | 1724 | struct audit_tree_refs *p; |
1727 | struct audit_chunk *chunk; | 1725 | struct audit_chunk *chunk; |
1728 | int count; | 1726 | int count; |
1729 | if (likely(list_empty(&inode->inotify_watches))) | 1727 | if (likely(hlist_empty(&inode->i_fsnotify_marks))) |
1730 | return; | 1728 | return; |
1731 | context = current->audit_context; | 1729 | context = current->audit_context; |
1732 | p = context->trees; | 1730 | p = context->trees; |
@@ -1769,7 +1767,7 @@ retry: | |||
1769 | seq = read_seqbegin(&rename_lock); | 1767 | seq = read_seqbegin(&rename_lock); |
1770 | for(;;) { | 1768 | for(;;) { |
1771 | struct inode *inode = d->d_inode; | 1769 | struct inode *inode = d->d_inode; |
1772 | if (inode && unlikely(!list_empty(&inode->inotify_watches))) { | 1770 | if (inode && unlikely(!hlist_empty(&inode->i_fsnotify_marks))) { |
1773 | struct audit_chunk *chunk; | 1771 | struct audit_chunk *chunk; |
1774 | chunk = audit_tree_lookup(inode); | 1772 | chunk = audit_tree_lookup(inode); |
1775 | if (chunk) { | 1773 | if (chunk) { |
diff --git a/kernel/compat.c b/kernel/compat.c index 5adab05a3172..e167efce8423 100644 --- a/kernel/compat.c +++ b/kernel/compat.c | |||
@@ -279,11 +279,6 @@ asmlinkage long compat_sys_setrlimit(unsigned int resource, | |||
279 | struct compat_rlimit __user *rlim) | 279 | struct compat_rlimit __user *rlim) |
280 | { | 280 | { |
281 | struct rlimit r; | 281 | struct rlimit r; |
282 | int ret; | ||
283 | mm_segment_t old_fs = get_fs (); | ||
284 | |||
285 | if (resource >= RLIM_NLIMITS) | ||
286 | return -EINVAL; | ||
287 | 282 | ||
288 | if (!access_ok(VERIFY_READ, rlim, sizeof(*rlim)) || | 283 | if (!access_ok(VERIFY_READ, rlim, sizeof(*rlim)) || |
289 | __get_user(r.rlim_cur, &rlim->rlim_cur) || | 284 | __get_user(r.rlim_cur, &rlim->rlim_cur) || |
@@ -294,10 +289,7 @@ asmlinkage long compat_sys_setrlimit(unsigned int resource, | |||
294 | r.rlim_cur = RLIM_INFINITY; | 289 | r.rlim_cur = RLIM_INFINITY; |
295 | if (r.rlim_max == COMPAT_RLIM_INFINITY) | 290 | if (r.rlim_max == COMPAT_RLIM_INFINITY) |
296 | r.rlim_max = RLIM_INFINITY; | 291 | r.rlim_max = RLIM_INFINITY; |
297 | set_fs(KERNEL_DS); | 292 | return do_prlimit(current, resource, &r, NULL); |
298 | ret = sys_setrlimit(resource, (struct rlimit __user *) &r); | ||
299 | set_fs(old_fs); | ||
300 | return ret; | ||
301 | } | 293 | } |
302 | 294 | ||
303 | #ifdef COMPAT_RLIM_OLD_INFINITY | 295 | #ifdef COMPAT_RLIM_OLD_INFINITY |
@@ -329,16 +321,13 @@ asmlinkage long compat_sys_old_getrlimit(unsigned int resource, | |||
329 | 321 | ||
330 | #endif | 322 | #endif |
331 | 323 | ||
332 | asmlinkage long compat_sys_getrlimit (unsigned int resource, | 324 | asmlinkage long compat_sys_getrlimit(unsigned int resource, |
333 | struct compat_rlimit __user *rlim) | 325 | struct compat_rlimit __user *rlim) |
334 | { | 326 | { |
335 | struct rlimit r; | 327 | struct rlimit r; |
336 | int ret; | 328 | int ret; |
337 | mm_segment_t old_fs = get_fs(); | ||
338 | 329 | ||
339 | set_fs(KERNEL_DS); | 330 | ret = do_prlimit(current, resource, NULL, &r); |
340 | ret = sys_getrlimit(resource, (struct rlimit __user *) &r); | ||
341 | set_fs(old_fs); | ||
342 | if (!ret) { | 331 | if (!ret) { |
343 | if (r.rlim_cur > COMPAT_RLIM_INFINITY) | 332 | if (r.rlim_cur > COMPAT_RLIM_INFINITY) |
344 | r.rlim_cur = COMPAT_RLIM_INFINITY; | 333 | r.rlim_cur = COMPAT_RLIM_INFINITY; |
diff --git a/kernel/posix-cpu-timers.c b/kernel/posix-cpu-timers.c index f66bdd33a6c6..6842eeba5879 100644 --- a/kernel/posix-cpu-timers.c +++ b/kernel/posix-cpu-timers.c | |||
@@ -16,13 +16,13 @@ | |||
16 | * siglock protection since other code may update expiration cache as | 16 | * siglock protection since other code may update expiration cache as |
17 | * well. | 17 | * well. |
18 | */ | 18 | */ |
19 | void update_rlimit_cpu(unsigned long rlim_new) | 19 | void update_rlimit_cpu(struct task_struct *task, unsigned long rlim_new) |
20 | { | 20 | { |
21 | cputime_t cputime = secs_to_cputime(rlim_new); | 21 | cputime_t cputime = secs_to_cputime(rlim_new); |
22 | 22 | ||
23 | spin_lock_irq(¤t->sighand->siglock); | 23 | spin_lock_irq(&task->sighand->siglock); |
24 | set_process_cpu_timer(current, CPUCLOCK_PROF, &cputime, NULL); | 24 | set_process_cpu_timer(task, CPUCLOCK_PROF, &cputime, NULL); |
25 | spin_unlock_irq(¤t->sighand->siglock); | 25 | spin_unlock_irq(&task->sighand->siglock); |
26 | } | 26 | } |
27 | 27 | ||
28 | static int check_clock(const clockid_t which_clock) | 28 | static int check_clock(const clockid_t which_clock) |
diff --git a/kernel/power/block_io.c b/kernel/power/block_io.c index 97024fd40cd5..83bbc7c02df9 100644 --- a/kernel/power/block_io.c +++ b/kernel/power/block_io.c | |||
@@ -28,7 +28,7 @@ | |||
28 | static int submit(int rw, struct block_device *bdev, sector_t sector, | 28 | static int submit(int rw, struct block_device *bdev, sector_t sector, |
29 | struct page *page, struct bio **bio_chain) | 29 | struct page *page, struct bio **bio_chain) |
30 | { | 30 | { |
31 | const int bio_rw = rw | (1 << BIO_RW_SYNCIO) | (1 << BIO_RW_UNPLUG); | 31 | const int bio_rw = rw | REQ_SYNC | REQ_UNPLUG; |
32 | struct bio *bio; | 32 | struct bio *bio; |
33 | 33 | ||
34 | bio = bio_alloc(__GFP_WAIT | __GFP_HIGH, 1); | 34 | bio = bio_alloc(__GFP_WAIT | __GFP_HIGH, 1); |
diff --git a/kernel/sys.c b/kernel/sys.c index e83ddbbaf89d..e9ad44489828 100644 --- a/kernel/sys.c +++ b/kernel/sys.c | |||
@@ -1236,15 +1236,14 @@ SYSCALL_DEFINE2(setdomainname, char __user *, name, int, len) | |||
1236 | 1236 | ||
1237 | SYSCALL_DEFINE2(getrlimit, unsigned int, resource, struct rlimit __user *, rlim) | 1237 | SYSCALL_DEFINE2(getrlimit, unsigned int, resource, struct rlimit __user *, rlim) |
1238 | { | 1238 | { |
1239 | if (resource >= RLIM_NLIMITS) | 1239 | struct rlimit value; |
1240 | return -EINVAL; | 1240 | int ret; |
1241 | else { | 1241 | |
1242 | struct rlimit value; | 1242 | ret = do_prlimit(current, resource, NULL, &value); |
1243 | task_lock(current->group_leader); | 1243 | if (!ret) |
1244 | value = current->signal->rlim[resource]; | 1244 | ret = copy_to_user(rlim, &value, sizeof(*rlim)) ? -EFAULT : 0; |
1245 | task_unlock(current->group_leader); | 1245 | |
1246 | return copy_to_user(rlim, &value, sizeof(*rlim)) ? -EFAULT : 0; | 1246 | return ret; |
1247 | } | ||
1248 | } | 1247 | } |
1249 | 1248 | ||
1250 | #ifdef __ARCH_WANT_SYS_OLD_GETRLIMIT | 1249 | #ifdef __ARCH_WANT_SYS_OLD_GETRLIMIT |
@@ -1272,44 +1271,89 @@ SYSCALL_DEFINE2(old_getrlimit, unsigned int, resource, | |||
1272 | 1271 | ||
1273 | #endif | 1272 | #endif |
1274 | 1273 | ||
1275 | SYSCALL_DEFINE2(setrlimit, unsigned int, resource, struct rlimit __user *, rlim) | 1274 | static inline bool rlim64_is_infinity(__u64 rlim64) |
1276 | { | 1275 | { |
1277 | struct rlimit new_rlim, *old_rlim; | 1276 | #if BITS_PER_LONG < 64 |
1278 | int retval; | 1277 | return rlim64 >= ULONG_MAX; |
1278 | #else | ||
1279 | return rlim64 == RLIM64_INFINITY; | ||
1280 | #endif | ||
1281 | } | ||
1282 | |||
1283 | static void rlim_to_rlim64(const struct rlimit *rlim, struct rlimit64 *rlim64) | ||
1284 | { | ||
1285 | if (rlim->rlim_cur == RLIM_INFINITY) | ||
1286 | rlim64->rlim_cur = RLIM64_INFINITY; | ||
1287 | else | ||
1288 | rlim64->rlim_cur = rlim->rlim_cur; | ||
1289 | if (rlim->rlim_max == RLIM_INFINITY) | ||
1290 | rlim64->rlim_max = RLIM64_INFINITY; | ||
1291 | else | ||
1292 | rlim64->rlim_max = rlim->rlim_max; | ||
1293 | } | ||
1294 | |||
1295 | static void rlim64_to_rlim(const struct rlimit64 *rlim64, struct rlimit *rlim) | ||
1296 | { | ||
1297 | if (rlim64_is_infinity(rlim64->rlim_cur)) | ||
1298 | rlim->rlim_cur = RLIM_INFINITY; | ||
1299 | else | ||
1300 | rlim->rlim_cur = (unsigned long)rlim64->rlim_cur; | ||
1301 | if (rlim64_is_infinity(rlim64->rlim_max)) | ||
1302 | rlim->rlim_max = RLIM_INFINITY; | ||
1303 | else | ||
1304 | rlim->rlim_max = (unsigned long)rlim64->rlim_max; | ||
1305 | } | ||
1306 | |||
1307 | /* make sure you are allowed to change @tsk limits before calling this */ | ||
1308 | int do_prlimit(struct task_struct *tsk, unsigned int resource, | ||
1309 | struct rlimit *new_rlim, struct rlimit *old_rlim) | ||
1310 | { | ||
1311 | struct rlimit *rlim; | ||
1312 | int retval = 0; | ||
1279 | 1313 | ||
1280 | if (resource >= RLIM_NLIMITS) | 1314 | if (resource >= RLIM_NLIMITS) |
1281 | return -EINVAL; | 1315 | return -EINVAL; |
1282 | if (copy_from_user(&new_rlim, rlim, sizeof(*rlim))) | 1316 | if (new_rlim) { |
1283 | return -EFAULT; | 1317 | if (new_rlim->rlim_cur > new_rlim->rlim_max) |
1284 | if (new_rlim.rlim_cur > new_rlim.rlim_max) | 1318 | return -EINVAL; |
1285 | return -EINVAL; | 1319 | if (resource == RLIMIT_NOFILE && |
1286 | old_rlim = current->signal->rlim + resource; | 1320 | new_rlim->rlim_max > sysctl_nr_open) |
1287 | if ((new_rlim.rlim_max > old_rlim->rlim_max) && | 1321 | return -EPERM; |
1288 | !capable(CAP_SYS_RESOURCE)) | ||
1289 | return -EPERM; | ||
1290 | if (resource == RLIMIT_NOFILE && new_rlim.rlim_max > sysctl_nr_open) | ||
1291 | return -EPERM; | ||
1292 | |||
1293 | retval = security_task_setrlimit(resource, &new_rlim); | ||
1294 | if (retval) | ||
1295 | return retval; | ||
1296 | |||
1297 | if (resource == RLIMIT_CPU && new_rlim.rlim_cur == 0) { | ||
1298 | /* | ||
1299 | * The caller is asking for an immediate RLIMIT_CPU | ||
1300 | * expiry. But we use the zero value to mean "it was | ||
1301 | * never set". So let's cheat and make it one second | ||
1302 | * instead | ||
1303 | */ | ||
1304 | new_rlim.rlim_cur = 1; | ||
1305 | } | 1322 | } |
1306 | 1323 | ||
1307 | task_lock(current->group_leader); | 1324 | /* protect tsk->signal and tsk->sighand from disappearing */ |
1308 | *old_rlim = new_rlim; | 1325 | read_lock(&tasklist_lock); |
1309 | task_unlock(current->group_leader); | 1326 | if (!tsk->sighand) { |
1310 | 1327 | retval = -ESRCH; | |
1311 | if (resource != RLIMIT_CPU) | ||
1312 | goto out; | 1328 | goto out; |
1329 | } | ||
1330 | |||
1331 | rlim = tsk->signal->rlim + resource; | ||
1332 | task_lock(tsk->group_leader); | ||
1333 | if (new_rlim) { | ||
1334 | if (new_rlim->rlim_max > rlim->rlim_max && | ||
1335 | !capable(CAP_SYS_RESOURCE)) | ||
1336 | retval = -EPERM; | ||
1337 | if (!retval) | ||
1338 | retval = security_task_setrlimit(tsk->group_leader, | ||
1339 | resource, new_rlim); | ||
1340 | if (resource == RLIMIT_CPU && new_rlim->rlim_cur == 0) { | ||
1341 | /* | ||
1342 | * The caller is asking for an immediate RLIMIT_CPU | ||
1343 | * expiry. But we use the zero value to mean "it was | ||
1344 | * never set". So let's cheat and make it one second | ||
1345 | * instead | ||
1346 | */ | ||
1347 | new_rlim->rlim_cur = 1; | ||
1348 | } | ||
1349 | } | ||
1350 | if (!retval) { | ||
1351 | if (old_rlim) | ||
1352 | *old_rlim = *rlim; | ||
1353 | if (new_rlim) | ||
1354 | *rlim = *new_rlim; | ||
1355 | } | ||
1356 | task_unlock(tsk->group_leader); | ||
1313 | 1357 | ||
1314 | /* | 1358 | /* |
1315 | * RLIMIT_CPU handling. Note that the kernel fails to return an error | 1359 | * RLIMIT_CPU handling. Note that the kernel fails to return an error |
@@ -1317,14 +1361,84 @@ SYSCALL_DEFINE2(setrlimit, unsigned int, resource, struct rlimit __user *, rlim) | |||
1317 | * very long-standing error, and fixing it now risks breakage of | 1361 | * very long-standing error, and fixing it now risks breakage of |
1318 | * applications, so we live with it | 1362 | * applications, so we live with it |
1319 | */ | 1363 | */ |
1320 | if (new_rlim.rlim_cur == RLIM_INFINITY) | 1364 | if (!retval && new_rlim && resource == RLIMIT_CPU && |
1321 | goto out; | 1365 | new_rlim->rlim_cur != RLIM_INFINITY) |
1322 | 1366 | update_rlimit_cpu(tsk, new_rlim->rlim_cur); | |
1323 | update_rlimit_cpu(new_rlim.rlim_cur); | ||
1324 | out: | 1367 | out: |
1368 | read_unlock(&tasklist_lock); | ||
1369 | return retval; | ||
1370 | } | ||
1371 | |||
1372 | /* rcu lock must be held */ | ||
1373 | static int check_prlimit_permission(struct task_struct *task) | ||
1374 | { | ||
1375 | const struct cred *cred = current_cred(), *tcred; | ||
1376 | |||
1377 | tcred = __task_cred(task); | ||
1378 | if ((cred->uid != tcred->euid || | ||
1379 | cred->uid != tcred->suid || | ||
1380 | cred->uid != tcred->uid || | ||
1381 | cred->gid != tcred->egid || | ||
1382 | cred->gid != tcred->sgid || | ||
1383 | cred->gid != tcred->gid) && | ||
1384 | !capable(CAP_SYS_RESOURCE)) { | ||
1385 | return -EPERM; | ||
1386 | } | ||
1387 | |||
1325 | return 0; | 1388 | return 0; |
1326 | } | 1389 | } |
1327 | 1390 | ||
1391 | SYSCALL_DEFINE4(prlimit64, pid_t, pid, unsigned int, resource, | ||
1392 | const struct rlimit64 __user *, new_rlim, | ||
1393 | struct rlimit64 __user *, old_rlim) | ||
1394 | { | ||
1395 | struct rlimit64 old64, new64; | ||
1396 | struct rlimit old, new; | ||
1397 | struct task_struct *tsk; | ||
1398 | int ret; | ||
1399 | |||
1400 | if (new_rlim) { | ||
1401 | if (copy_from_user(&new64, new_rlim, sizeof(new64))) | ||
1402 | return -EFAULT; | ||
1403 | rlim64_to_rlim(&new64, &new); | ||
1404 | } | ||
1405 | |||
1406 | rcu_read_lock(); | ||
1407 | tsk = pid ? find_task_by_vpid(pid) : current; | ||
1408 | if (!tsk) { | ||
1409 | rcu_read_unlock(); | ||
1410 | return -ESRCH; | ||
1411 | } | ||
1412 | ret = check_prlimit_permission(tsk); | ||
1413 | if (ret) { | ||
1414 | rcu_read_unlock(); | ||
1415 | return ret; | ||
1416 | } | ||
1417 | get_task_struct(tsk); | ||
1418 | rcu_read_unlock(); | ||
1419 | |||
1420 | ret = do_prlimit(tsk, resource, new_rlim ? &new : NULL, | ||
1421 | old_rlim ? &old : NULL); | ||
1422 | |||
1423 | if (!ret && old_rlim) { | ||
1424 | rlim_to_rlim64(&old, &old64); | ||
1425 | if (copy_to_user(old_rlim, &old64, sizeof(old64))) | ||
1426 | ret = -EFAULT; | ||
1427 | } | ||
1428 | |||
1429 | put_task_struct(tsk); | ||
1430 | return ret; | ||
1431 | } | ||
1432 | |||
1433 | SYSCALL_DEFINE2(setrlimit, unsigned int, resource, struct rlimit __user *, rlim) | ||
1434 | { | ||
1435 | struct rlimit new_rlim; | ||
1436 | |||
1437 | if (copy_from_user(&new_rlim, rlim, sizeof(*rlim))) | ||
1438 | return -EFAULT; | ||
1439 | return do_prlimit(current, resource, &new_rlim, NULL); | ||
1440 | } | ||
1441 | |||
1328 | /* | 1442 | /* |
1329 | * It would make sense to put struct rusage in the task_struct, | 1443 | * It would make sense to put struct rusage in the task_struct, |
1330 | * except that would make the task_struct be *really big*. After | 1444 | * except that would make the task_struct be *really big*. After |
diff --git a/kernel/sys_ni.c b/kernel/sys_ni.c index 70f2ea758ffe..bad369ec5403 100644 --- a/kernel/sys_ni.c +++ b/kernel/sys_ni.c | |||
@@ -181,3 +181,7 @@ cond_syscall(sys_eventfd2); | |||
181 | 181 | ||
182 | /* performance counters: */ | 182 | /* performance counters: */ |
183 | cond_syscall(sys_perf_event_open); | 183 | cond_syscall(sys_perf_event_open); |
184 | |||
185 | /* fanotify! */ | ||
186 | cond_syscall(sys_fanotify_init); | ||
187 | cond_syscall(sys_fanotify_mark); | ||
diff --git a/kernel/sysctl.c b/kernel/sysctl.c index 6b005e4912b5..ca38e8e3e907 100644 --- a/kernel/sysctl.c +++ b/kernel/sysctl.c | |||
@@ -44,6 +44,7 @@ | |||
44 | #include <linux/times.h> | 44 | #include <linux/times.h> |
45 | #include <linux/limits.h> | 45 | #include <linux/limits.h> |
46 | #include <linux/dcache.h> | 46 | #include <linux/dcache.h> |
47 | #include <linux/dnotify.h> | ||
47 | #include <linux/syscalls.h> | 48 | #include <linux/syscalls.h> |
48 | #include <linux/vmstat.h> | 49 | #include <linux/vmstat.h> |
49 | #include <linux/nfs_fs.h> | 50 | #include <linux/nfs_fs.h> |
@@ -131,6 +132,9 @@ static int min_percpu_pagelist_fract = 8; | |||
131 | 132 | ||
132 | static int ngroups_max = NGROUPS_MAX; | 133 | static int ngroups_max = NGROUPS_MAX; |
133 | 134 | ||
135 | #ifdef CONFIG_INOTIFY_USER | ||
136 | #include <linux/inotify.h> | ||
137 | #endif | ||
134 | #ifdef CONFIG_SPARC | 138 | #ifdef CONFIG_SPARC |
135 | #include <asm/system.h> | 139 | #include <asm/system.h> |
136 | #endif | 140 | #endif |
@@ -207,9 +211,6 @@ static struct ctl_table fs_table[]; | |||
207 | static struct ctl_table debug_table[]; | 211 | static struct ctl_table debug_table[]; |
208 | static struct ctl_table dev_table[]; | 212 | static struct ctl_table dev_table[]; |
209 | extern struct ctl_table random_table[]; | 213 | extern struct ctl_table random_table[]; |
210 | #ifdef CONFIG_INOTIFY_USER | ||
211 | extern struct ctl_table inotify_table[]; | ||
212 | #endif | ||
213 | #ifdef CONFIG_EPOLL | 214 | #ifdef CONFIG_EPOLL |
214 | extern struct ctl_table epoll_table[]; | 215 | extern struct ctl_table epoll_table[]; |
215 | #endif | 216 | #endif |
diff --git a/kernel/timer.c b/kernel/timer.c index f1b8afe1ad86..97bf05baade7 100644 --- a/kernel/timer.c +++ b/kernel/timer.c | |||
@@ -326,6 +326,7 @@ EXPORT_SYMBOL_GPL(round_jiffies_up_relative); | |||
326 | 326 | ||
327 | /** | 327 | /** |
328 | * set_timer_slack - set the allowed slack for a timer | 328 | * set_timer_slack - set the allowed slack for a timer |
329 | * @timer: the timer to be modified | ||
329 | * @slack_hz: the amount of time (in jiffies) allowed for rounding | 330 | * @slack_hz: the amount of time (in jiffies) allowed for rounding |
330 | * | 331 | * |
331 | * Set the amount of time, in jiffies, that a certain timer has | 332 | * Set the amount of time, in jiffies, that a certain timer has |
diff --git a/kernel/trace/blktrace.c b/kernel/trace/blktrace.c index 638711c17504..82499a5bdcb7 100644 --- a/kernel/trace/blktrace.c +++ b/kernel/trace/blktrace.c | |||
@@ -169,9 +169,12 @@ static int act_log_check(struct blk_trace *bt, u32 what, sector_t sector, | |||
169 | static const u32 ddir_act[2] = { BLK_TC_ACT(BLK_TC_READ), | 169 | static const u32 ddir_act[2] = { BLK_TC_ACT(BLK_TC_READ), |
170 | BLK_TC_ACT(BLK_TC_WRITE) }; | 170 | BLK_TC_ACT(BLK_TC_WRITE) }; |
171 | 171 | ||
172 | #define BLK_TC_HARDBARRIER BLK_TC_BARRIER | ||
173 | #define BLK_TC_RAHEAD BLK_TC_AHEAD | ||
174 | |||
172 | /* The ilog2() calls fall out because they're constant */ | 175 | /* The ilog2() calls fall out because they're constant */ |
173 | #define MASK_TC_BIT(rw, __name) ((rw & (1 << BIO_RW_ ## __name)) << \ | 176 | #define MASK_TC_BIT(rw, __name) ((rw & REQ_ ## __name) << \ |
174 | (ilog2(BLK_TC_ ## __name) + BLK_TC_SHIFT - BIO_RW_ ## __name)) | 177 | (ilog2(BLK_TC_ ## __name) + BLK_TC_SHIFT - __REQ_ ## __name)) |
175 | 178 | ||
176 | /* | 179 | /* |
177 | * The worker for the various blk_add_trace*() types. Fills out a | 180 | * The worker for the various blk_add_trace*() types. Fills out a |
@@ -194,9 +197,9 @@ static void __blk_add_trace(struct blk_trace *bt, sector_t sector, int bytes, | |||
194 | return; | 197 | return; |
195 | 198 | ||
196 | what |= ddir_act[rw & WRITE]; | 199 | what |= ddir_act[rw & WRITE]; |
197 | what |= MASK_TC_BIT(rw, BARRIER); | 200 | what |= MASK_TC_BIT(rw, HARDBARRIER); |
198 | what |= MASK_TC_BIT(rw, SYNCIO); | 201 | what |= MASK_TC_BIT(rw, SYNC); |
199 | what |= MASK_TC_BIT(rw, AHEAD); | 202 | what |= MASK_TC_BIT(rw, RAHEAD); |
200 | what |= MASK_TC_BIT(rw, META); | 203 | what |= MASK_TC_BIT(rw, META); |
201 | what |= MASK_TC_BIT(rw, DISCARD); | 204 | what |= MASK_TC_BIT(rw, DISCARD); |
202 | 205 | ||
@@ -549,6 +552,41 @@ int blk_trace_setup(struct request_queue *q, char *name, dev_t dev, | |||
549 | } | 552 | } |
550 | EXPORT_SYMBOL_GPL(blk_trace_setup); | 553 | EXPORT_SYMBOL_GPL(blk_trace_setup); |
551 | 554 | ||
555 | #if defined(CONFIG_COMPAT) && defined(CONFIG_X86_64) | ||
556 | static int compat_blk_trace_setup(struct request_queue *q, char *name, | ||
557 | dev_t dev, struct block_device *bdev, | ||
558 | char __user *arg) | ||
559 | { | ||
560 | struct blk_user_trace_setup buts; | ||
561 | struct compat_blk_user_trace_setup cbuts; | ||
562 | int ret; | ||
563 | |||
564 | if (copy_from_user(&cbuts, arg, sizeof(cbuts))) | ||
565 | return -EFAULT; | ||
566 | |||
567 | buts = (struct blk_user_trace_setup) { | ||
568 | .act_mask = cbuts.act_mask, | ||
569 | .buf_size = cbuts.buf_size, | ||
570 | .buf_nr = cbuts.buf_nr, | ||
571 | .start_lba = cbuts.start_lba, | ||
572 | .end_lba = cbuts.end_lba, | ||
573 | .pid = cbuts.pid, | ||
574 | }; | ||
575 | memcpy(&buts.name, &cbuts.name, 32); | ||
576 | |||
577 | ret = do_blk_trace_setup(q, name, dev, bdev, &buts); | ||
578 | if (ret) | ||
579 | return ret; | ||
580 | |||
581 | if (copy_to_user(arg, &buts.name, 32)) { | ||
582 | blk_trace_remove(q); | ||
583 | return -EFAULT; | ||
584 | } | ||
585 | |||
586 | return 0; | ||
587 | } | ||
588 | #endif | ||
589 | |||
552 | int blk_trace_startstop(struct request_queue *q, int start) | 590 | int blk_trace_startstop(struct request_queue *q, int start) |
553 | { | 591 | { |
554 | int ret; | 592 | int ret; |
@@ -601,6 +639,7 @@ int blk_trace_ioctl(struct block_device *bdev, unsigned cmd, char __user *arg) | |||
601 | if (!q) | 639 | if (!q) |
602 | return -ENXIO; | 640 | return -ENXIO; |
603 | 641 | ||
642 | lock_kernel(); | ||
604 | mutex_lock(&bdev->bd_mutex); | 643 | mutex_lock(&bdev->bd_mutex); |
605 | 644 | ||
606 | switch (cmd) { | 645 | switch (cmd) { |
@@ -608,6 +647,12 @@ int blk_trace_ioctl(struct block_device *bdev, unsigned cmd, char __user *arg) | |||
608 | bdevname(bdev, b); | 647 | bdevname(bdev, b); |
609 | ret = blk_trace_setup(q, b, bdev->bd_dev, bdev, arg); | 648 | ret = blk_trace_setup(q, b, bdev->bd_dev, bdev, arg); |
610 | break; | 649 | break; |
650 | #if defined(CONFIG_COMPAT) && defined(CONFIG_X86_64) | ||
651 | case BLKTRACESETUP32: | ||
652 | bdevname(bdev, b); | ||
653 | ret = compat_blk_trace_setup(q, b, bdev->bd_dev, bdev, arg); | ||
654 | break; | ||
655 | #endif | ||
611 | case BLKTRACESTART: | 656 | case BLKTRACESTART: |
612 | start = 1; | 657 | start = 1; |
613 | case BLKTRACESTOP: | 658 | case BLKTRACESTOP: |
@@ -622,6 +667,7 @@ int blk_trace_ioctl(struct block_device *bdev, unsigned cmd, char __user *arg) | |||
622 | } | 667 | } |
623 | 668 | ||
624 | mutex_unlock(&bdev->bd_mutex); | 669 | mutex_unlock(&bdev->bd_mutex); |
670 | unlock_kernel(); | ||
625 | return ret; | 671 | return ret; |
626 | } | 672 | } |
627 | 673 | ||
@@ -661,10 +707,10 @@ static void blk_add_trace_rq(struct request_queue *q, struct request *rq, | |||
661 | if (likely(!bt)) | 707 | if (likely(!bt)) |
662 | return; | 708 | return; |
663 | 709 | ||
664 | if (blk_discard_rq(rq)) | 710 | if (rq->cmd_flags & REQ_DISCARD) |
665 | rw |= (1 << BIO_RW_DISCARD); | 711 | rw |= REQ_DISCARD; |
666 | 712 | ||
667 | if (blk_pc_request(rq)) { | 713 | if (rq->cmd_type == REQ_TYPE_BLOCK_PC) { |
668 | what |= BLK_TC_ACT(BLK_TC_PC); | 714 | what |= BLK_TC_ACT(BLK_TC_PC); |
669 | __blk_add_trace(bt, 0, blk_rq_bytes(rq), rw, | 715 | __blk_add_trace(bt, 0, blk_rq_bytes(rq), rw, |
670 | what, rq->errors, rq->cmd_len, rq->cmd); | 716 | what, rq->errors, rq->cmd_len, rq->cmd); |
@@ -925,7 +971,7 @@ void blk_add_driver_data(struct request_queue *q, | |||
925 | if (likely(!bt)) | 971 | if (likely(!bt)) |
926 | return; | 972 | return; |
927 | 973 | ||
928 | if (blk_pc_request(rq)) | 974 | if (rq->cmd_type == REQ_TYPE_BLOCK_PC) |
929 | __blk_add_trace(bt, 0, blk_rq_bytes(rq), 0, | 975 | __blk_add_trace(bt, 0, blk_rq_bytes(rq), 0, |
930 | BLK_TA_DRV_DATA, rq->errors, len, data); | 976 | BLK_TA_DRV_DATA, rq->errors, len, data); |
931 | else | 977 | else |
@@ -1730,7 +1776,7 @@ void blk_dump_cmd(char *buf, struct request *rq) | |||
1730 | int len = rq->cmd_len; | 1776 | int len = rq->cmd_len; |
1731 | unsigned char *cmd = rq->cmd; | 1777 | unsigned char *cmd = rq->cmd; |
1732 | 1778 | ||
1733 | if (!blk_pc_request(rq)) { | 1779 | if (rq->cmd_type != REQ_TYPE_BLOCK_PC) { |
1734 | buf[0] = '\0'; | 1780 | buf[0] = '\0'; |
1735 | return; | 1781 | return; |
1736 | } | 1782 | } |
@@ -1755,20 +1801,20 @@ void blk_fill_rwbs(char *rwbs, u32 rw, int bytes) | |||
1755 | 1801 | ||
1756 | if (rw & WRITE) | 1802 | if (rw & WRITE) |
1757 | rwbs[i++] = 'W'; | 1803 | rwbs[i++] = 'W'; |
1758 | else if (rw & 1 << BIO_RW_DISCARD) | 1804 | else if (rw & REQ_DISCARD) |
1759 | rwbs[i++] = 'D'; | 1805 | rwbs[i++] = 'D'; |
1760 | else if (bytes) | 1806 | else if (bytes) |
1761 | rwbs[i++] = 'R'; | 1807 | rwbs[i++] = 'R'; |
1762 | else | 1808 | else |
1763 | rwbs[i++] = 'N'; | 1809 | rwbs[i++] = 'N'; |
1764 | 1810 | ||
1765 | if (rw & 1 << BIO_RW_AHEAD) | 1811 | if (rw & REQ_RAHEAD) |
1766 | rwbs[i++] = 'A'; | 1812 | rwbs[i++] = 'A'; |
1767 | if (rw & 1 << BIO_RW_BARRIER) | 1813 | if (rw & REQ_HARDBARRIER) |
1768 | rwbs[i++] = 'B'; | 1814 | rwbs[i++] = 'B'; |
1769 | if (rw & 1 << BIO_RW_SYNCIO) | 1815 | if (rw & REQ_SYNC) |
1770 | rwbs[i++] = 'S'; | 1816 | rwbs[i++] = 'S'; |
1771 | if (rw & 1 << BIO_RW_META) | 1817 | if (rw & REQ_META) |
1772 | rwbs[i++] = 'M'; | 1818 | rwbs[i++] = 'M'; |
1773 | 1819 | ||
1774 | rwbs[i] = '\0'; | 1820 | rwbs[i] = '\0'; |
@@ -1779,8 +1825,8 @@ void blk_fill_rwbs_rq(char *rwbs, struct request *rq) | |||
1779 | int rw = rq->cmd_flags & 0x03; | 1825 | int rw = rq->cmd_flags & 0x03; |
1780 | int bytes; | 1826 | int bytes; |
1781 | 1827 | ||
1782 | if (blk_discard_rq(rq)) | 1828 | if (rq->cmd_flags & REQ_DISCARD) |
1783 | rw |= (1 << BIO_RW_DISCARD); | 1829 | rw |= REQ_DISCARD; |
1784 | 1830 | ||
1785 | bytes = blk_rq_bytes(rq); | 1831 | bytes = blk_rq_bytes(rq); |
1786 | 1832 | ||