diff options
Diffstat (limited to 'kernel')
| -rw-r--r-- | kernel/Makefile | 5 | ||||
| -rw-r--r-- | kernel/audit.c | 1 | ||||
| -rw-r--r-- | kernel/audit.h | 26 | ||||
| -rw-r--r-- | kernel/audit_tree.c | 237 | ||||
| -rw-r--r-- | kernel/audit_watch.c | 274 | ||||
| -rw-r--r-- | kernel/auditfilter.c | 39 | ||||
| -rw-r--r-- | kernel/auditsc.c | 10 | ||||
| -rw-r--r-- | kernel/sys_ni.c | 4 | ||||
| -rw-r--r-- | kernel/sysctl.c | 7 |
9 files changed, 325 insertions, 278 deletions
diff --git a/kernel/Makefile b/kernel/Makefile index c53e491e25a8..0b72d1a74be0 100644 --- a/kernel/Makefile +++ b/kernel/Makefile | |||
| @@ -70,10 +70,11 @@ obj-$(CONFIG_IKCONFIG) += configs.o | |||
| 70 | obj-$(CONFIG_RESOURCE_COUNTERS) += res_counter.o | 70 | obj-$(CONFIG_RESOURCE_COUNTERS) += res_counter.o |
| 71 | obj-$(CONFIG_SMP) += stop_machine.o | 71 | obj-$(CONFIG_SMP) += stop_machine.o |
| 72 | obj-$(CONFIG_KPROBES_SANITY_TEST) += test_kprobes.o | 72 | obj-$(CONFIG_KPROBES_SANITY_TEST) += test_kprobes.o |
| 73 | obj-$(CONFIG_AUDIT) += audit.o auditfilter.o audit_watch.o | 73 | obj-$(CONFIG_AUDIT) += audit.o auditfilter.o |
| 74 | obj-$(CONFIG_AUDITSYSCALL) += auditsc.o | 74 | obj-$(CONFIG_AUDITSYSCALL) += auditsc.o |
| 75 | obj-$(CONFIG_GCOV_KERNEL) += gcov/ | 75 | obj-$(CONFIG_AUDIT_WATCH) += audit_watch.o |
| 76 | obj-$(CONFIG_AUDIT_TREE) += audit_tree.o | 76 | obj-$(CONFIG_AUDIT_TREE) += audit_tree.o |
| 77 | obj-$(CONFIG_GCOV_KERNEL) += gcov/ | ||
| 77 | obj-$(CONFIG_KPROBES) += kprobes.o | 78 | obj-$(CONFIG_KPROBES) += kprobes.o |
| 78 | obj-$(CONFIG_KGDB) += debug/ | 79 | obj-$(CONFIG_KGDB) += debug/ |
| 79 | obj-$(CONFIG_DETECT_HUNG_TASK) += hung_task.o | 80 | obj-$(CONFIG_DETECT_HUNG_TASK) += hung_task.o |
diff --git a/kernel/audit.c b/kernel/audit.c index 8296aa516c5a..d96045789b54 100644 --- a/kernel/audit.c +++ b/kernel/audit.c | |||
| @@ -56,7 +56,6 @@ | |||
| 56 | #include <net/netlink.h> | 56 | #include <net/netlink.h> |
| 57 | #include <linux/skbuff.h> | 57 | #include <linux/skbuff.h> |
| 58 | #include <linux/netlink.h> | 58 | #include <linux/netlink.h> |
| 59 | #include <linux/inotify.h> | ||
| 60 | #include <linux/freezer.h> | 59 | #include <linux/freezer.h> |
| 61 | #include <linux/tty.h> | 60 | #include <linux/tty.h> |
| 62 | 61 | ||
diff --git a/kernel/audit.h b/kernel/audit.h index 208687be4f30..f7206db4e13d 100644 --- a/kernel/audit.h +++ b/kernel/audit.h | |||
| @@ -103,21 +103,27 @@ extern struct mutex audit_filter_mutex; | |||
| 103 | extern void audit_free_rule_rcu(struct rcu_head *); | 103 | extern void audit_free_rule_rcu(struct rcu_head *); |
| 104 | extern struct list_head audit_filter_list[]; | 104 | extern struct list_head audit_filter_list[]; |
| 105 | 105 | ||
| 106 | extern struct audit_entry *audit_dupe_rule(struct audit_krule *old); | ||
| 107 | |||
| 106 | /* audit watch functions */ | 108 | /* audit watch functions */ |
| 107 | extern unsigned long audit_watch_inode(struct audit_watch *watch); | 109 | #ifdef CONFIG_AUDIT_WATCH |
| 108 | extern dev_t audit_watch_dev(struct audit_watch *watch); | ||
| 109 | extern void audit_put_watch(struct audit_watch *watch); | 110 | extern void audit_put_watch(struct audit_watch *watch); |
| 110 | extern void audit_get_watch(struct audit_watch *watch); | 111 | extern void audit_get_watch(struct audit_watch *watch); |
| 111 | extern int audit_to_watch(struct audit_krule *krule, char *path, int len, u32 op); | 112 | extern int audit_to_watch(struct audit_krule *krule, char *path, int len, u32 op); |
| 112 | extern int audit_add_watch(struct audit_krule *krule); | 113 | extern int audit_add_watch(struct audit_krule *krule, struct list_head **list); |
| 113 | extern void audit_remove_watch(struct audit_watch *watch); | 114 | extern void audit_remove_watch_rule(struct audit_krule *krule); |
| 114 | extern void audit_remove_watch_rule(struct audit_krule *krule, struct list_head *list); | ||
| 115 | extern void audit_inotify_unregister(struct list_head *in_list); | ||
| 116 | extern char *audit_watch_path(struct audit_watch *watch); | 115 | extern char *audit_watch_path(struct audit_watch *watch); |
| 117 | extern struct list_head *audit_watch_rules(struct audit_watch *watch); | 116 | extern int audit_watch_compare(struct audit_watch *watch, unsigned long ino, dev_t dev); |
| 118 | 117 | #else | |
| 119 | extern struct audit_entry *audit_dupe_rule(struct audit_krule *old, | 118 | #define audit_put_watch(w) {} |
| 120 | struct audit_watch *watch); | 119 | #define audit_get_watch(w) {} |
| 120 | #define audit_to_watch(k, p, l, o) (-EINVAL) | ||
| 121 | #define audit_add_watch(k, l) (-EINVAL) | ||
| 122 | #define audit_remove_watch_rule(k) BUG() | ||
| 123 | #define audit_watch_path(w) "" | ||
| 124 | #define audit_watch_compare(w, i, d) 0 | ||
| 125 | |||
| 126 | #endif /* CONFIG_AUDIT_WATCH */ | ||
| 121 | 127 | ||
| 122 | #ifdef CONFIG_AUDIT_TREE | 128 | #ifdef CONFIG_AUDIT_TREE |
| 123 | extern struct audit_chunk *audit_tree_lookup(const struct inode *); | 129 | extern struct audit_chunk *audit_tree_lookup(const struct inode *); |
diff --git a/kernel/audit_tree.c b/kernel/audit_tree.c index 46a57b57a335..7f18d3a4527e 100644 --- a/kernel/audit_tree.c +++ b/kernel/audit_tree.c | |||
| @@ -1,5 +1,5 @@ | |||
| 1 | #include "audit.h" | 1 | #include "audit.h" |
| 2 | #include <linux/inotify.h> | 2 | #include <linux/fsnotify_backend.h> |
| 3 | #include <linux/namei.h> | 3 | #include <linux/namei.h> |
| 4 | #include <linux/mount.h> | 4 | #include <linux/mount.h> |
| 5 | #include <linux/kthread.h> | 5 | #include <linux/kthread.h> |
| @@ -22,7 +22,7 @@ struct audit_tree { | |||
| 22 | 22 | ||
| 23 | struct audit_chunk { | 23 | struct audit_chunk { |
| 24 | struct list_head hash; | 24 | struct list_head hash; |
| 25 | struct inotify_watch watch; | 25 | struct fsnotify_mark mark; |
| 26 | struct list_head trees; /* with root here */ | 26 | struct list_head trees; /* with root here */ |
| 27 | int dead; | 27 | int dead; |
| 28 | int count; | 28 | int count; |
| @@ -59,7 +59,7 @@ static LIST_HEAD(prune_list); | |||
| 59 | * tree is refcounted; one reference for "some rules on rules_list refer to | 59 | * tree is refcounted; one reference for "some rules on rules_list refer to |
| 60 | * it", one for each chunk with pointer to it. | 60 | * it", one for each chunk with pointer to it. |
| 61 | * | 61 | * |
| 62 | * chunk is refcounted by embedded inotify_watch + .refs (non-zero refcount | 62 | * chunk is refcounted by embedded fsnotify_mark + .refs (non-zero refcount |
| 63 | * of watch contributes 1 to .refs). | 63 | * of watch contributes 1 to .refs). |
| 64 | * | 64 | * |
| 65 | * node.index allows to get from node.list to containing chunk. | 65 | * node.index allows to get from node.list to containing chunk. |
| @@ -68,7 +68,7 @@ static LIST_HEAD(prune_list); | |||
| 68 | * that makes a difference. Some. | 68 | * that makes a difference. Some. |
| 69 | */ | 69 | */ |
| 70 | 70 | ||
| 71 | static struct inotify_handle *rtree_ih; | 71 | static struct fsnotify_group *audit_tree_group; |
| 72 | 72 | ||
| 73 | static struct audit_tree *alloc_tree(const char *s) | 73 | static struct audit_tree *alloc_tree(const char *s) |
| 74 | { | 74 | { |
| @@ -111,29 +111,6 @@ const char *audit_tree_path(struct audit_tree *tree) | |||
| 111 | return tree->pathname; | 111 | return tree->pathname; |
| 112 | } | 112 | } |
| 113 | 113 | ||
| 114 | static struct audit_chunk *alloc_chunk(int count) | ||
| 115 | { | ||
| 116 | struct audit_chunk *chunk; | ||
| 117 | size_t size; | ||
| 118 | int i; | ||
| 119 | |||
| 120 | size = offsetof(struct audit_chunk, owners) + count * sizeof(struct node); | ||
| 121 | chunk = kzalloc(size, GFP_KERNEL); | ||
| 122 | if (!chunk) | ||
| 123 | return NULL; | ||
| 124 | |||
| 125 | INIT_LIST_HEAD(&chunk->hash); | ||
| 126 | INIT_LIST_HEAD(&chunk->trees); | ||
| 127 | chunk->count = count; | ||
| 128 | atomic_long_set(&chunk->refs, 1); | ||
| 129 | for (i = 0; i < count; i++) { | ||
| 130 | INIT_LIST_HEAD(&chunk->owners[i].list); | ||
| 131 | chunk->owners[i].index = i; | ||
| 132 | } | ||
| 133 | inotify_init_watch(&chunk->watch); | ||
| 134 | return chunk; | ||
| 135 | } | ||
| 136 | |||
| 137 | static void free_chunk(struct audit_chunk *chunk) | 114 | static void free_chunk(struct audit_chunk *chunk) |
| 138 | { | 115 | { |
| 139 | int i; | 116 | int i; |
| @@ -157,6 +134,35 @@ static void __put_chunk(struct rcu_head *rcu) | |||
| 157 | audit_put_chunk(chunk); | 134 | audit_put_chunk(chunk); |
| 158 | } | 135 | } |
| 159 | 136 | ||
| 137 | static void audit_tree_destroy_watch(struct fsnotify_mark *entry) | ||
| 138 | { | ||
| 139 | struct audit_chunk *chunk = container_of(entry, struct audit_chunk, mark); | ||
| 140 | call_rcu(&chunk->head, __put_chunk); | ||
| 141 | } | ||
| 142 | |||
| 143 | static struct audit_chunk *alloc_chunk(int count) | ||
| 144 | { | ||
| 145 | struct audit_chunk *chunk; | ||
| 146 | size_t size; | ||
| 147 | int i; | ||
| 148 | |||
| 149 | size = offsetof(struct audit_chunk, owners) + count * sizeof(struct node); | ||
| 150 | chunk = kzalloc(size, GFP_KERNEL); | ||
| 151 | if (!chunk) | ||
| 152 | return NULL; | ||
| 153 | |||
| 154 | INIT_LIST_HEAD(&chunk->hash); | ||
| 155 | INIT_LIST_HEAD(&chunk->trees); | ||
| 156 | chunk->count = count; | ||
| 157 | atomic_long_set(&chunk->refs, 1); | ||
| 158 | for (i = 0; i < count; i++) { | ||
| 159 | INIT_LIST_HEAD(&chunk->owners[i].list); | ||
| 160 | chunk->owners[i].index = i; | ||
| 161 | } | ||
| 162 | fsnotify_init_mark(&chunk->mark, audit_tree_destroy_watch); | ||
| 163 | return chunk; | ||
| 164 | } | ||
| 165 | |||
| 160 | enum {HASH_SIZE = 128}; | 166 | enum {HASH_SIZE = 128}; |
| 161 | static struct list_head chunk_hash_heads[HASH_SIZE]; | 167 | static struct list_head chunk_hash_heads[HASH_SIZE]; |
| 162 | static __cacheline_aligned_in_smp DEFINE_SPINLOCK(hash_lock); | 168 | static __cacheline_aligned_in_smp DEFINE_SPINLOCK(hash_lock); |
| @@ -167,10 +173,15 @@ static inline struct list_head *chunk_hash(const struct inode *inode) | |||
| 167 | return chunk_hash_heads + n % HASH_SIZE; | 173 | return chunk_hash_heads + n % HASH_SIZE; |
| 168 | } | 174 | } |
| 169 | 175 | ||
| 170 | /* hash_lock is held by caller */ | 176 | /* hash_lock & entry->lock is held by caller */ |
| 171 | static void insert_hash(struct audit_chunk *chunk) | 177 | static void insert_hash(struct audit_chunk *chunk) |
| 172 | { | 178 | { |
| 173 | struct list_head *list = chunk_hash(chunk->watch.inode); | 179 | struct fsnotify_mark *entry = &chunk->mark; |
| 180 | struct list_head *list; | ||
| 181 | |||
| 182 | if (!entry->i.inode) | ||
| 183 | return; | ||
| 184 | list = chunk_hash(entry->i.inode); | ||
| 174 | list_add_rcu(&chunk->hash, list); | 185 | list_add_rcu(&chunk->hash, list); |
| 175 | } | 186 | } |
| 176 | 187 | ||
| @@ -181,7 +192,8 @@ struct audit_chunk *audit_tree_lookup(const struct inode *inode) | |||
| 181 | struct audit_chunk *p; | 192 | struct audit_chunk *p; |
| 182 | 193 | ||
| 183 | list_for_each_entry_rcu(p, list, hash) { | 194 | list_for_each_entry_rcu(p, list, hash) { |
| 184 | if (p->watch.inode == inode) { | 195 | /* mark.inode may have gone NULL, but who cares? */ |
| 196 | if (p->mark.i.inode == inode) { | ||
| 185 | atomic_long_inc(&p->refs); | 197 | atomic_long_inc(&p->refs); |
| 186 | return p; | 198 | return p; |
| 187 | } | 199 | } |
| @@ -210,38 +222,19 @@ static struct audit_chunk *find_chunk(struct node *p) | |||
| 210 | static void untag_chunk(struct node *p) | 222 | static void untag_chunk(struct node *p) |
| 211 | { | 223 | { |
| 212 | struct audit_chunk *chunk = find_chunk(p); | 224 | struct audit_chunk *chunk = find_chunk(p); |
| 225 | struct fsnotify_mark *entry = &chunk->mark; | ||
| 213 | struct audit_chunk *new; | 226 | struct audit_chunk *new; |
| 214 | struct audit_tree *owner; | 227 | struct audit_tree *owner; |
| 215 | int size = chunk->count - 1; | 228 | int size = chunk->count - 1; |
| 216 | int i, j; | 229 | int i, j; |
| 217 | 230 | ||
| 218 | if (!pin_inotify_watch(&chunk->watch)) { | 231 | fsnotify_get_mark(entry); |
| 219 | /* | ||
| 220 | * Filesystem is shutting down; all watches are getting | ||
| 221 | * evicted, just take it off the node list for this | ||
| 222 | * tree and let the eviction logics take care of the | ||
| 223 | * rest. | ||
| 224 | */ | ||
| 225 | owner = p->owner; | ||
| 226 | if (owner->root == chunk) { | ||
| 227 | list_del_init(&owner->same_root); | ||
| 228 | owner->root = NULL; | ||
| 229 | } | ||
| 230 | list_del_init(&p->list); | ||
| 231 | p->owner = NULL; | ||
| 232 | put_tree(owner); | ||
| 233 | return; | ||
| 234 | } | ||
| 235 | 232 | ||
| 236 | spin_unlock(&hash_lock); | 233 | spin_unlock(&hash_lock); |
| 237 | 234 | ||
| 238 | /* | 235 | spin_lock(&entry->lock); |
| 239 | * pin_inotify_watch() succeeded, so the watch won't go away | 236 | if (chunk->dead || !entry->i.inode) { |
| 240 | * from under us. | 237 | spin_unlock(&entry->lock); |
| 241 | */ | ||
| 242 | mutex_lock(&chunk->watch.inode->inotify_mutex); | ||
| 243 | if (chunk->dead) { | ||
| 244 | mutex_unlock(&chunk->watch.inode->inotify_mutex); | ||
| 245 | goto out; | 238 | goto out; |
| 246 | } | 239 | } |
| 247 | 240 | ||
| @@ -256,16 +249,17 @@ static void untag_chunk(struct node *p) | |||
| 256 | list_del_init(&p->list); | 249 | list_del_init(&p->list); |
| 257 | list_del_rcu(&chunk->hash); | 250 | list_del_rcu(&chunk->hash); |
| 258 | spin_unlock(&hash_lock); | 251 | spin_unlock(&hash_lock); |
| 259 | inotify_evict_watch(&chunk->watch); | 252 | spin_unlock(&entry->lock); |
| 260 | mutex_unlock(&chunk->watch.inode->inotify_mutex); | 253 | fsnotify_destroy_mark(entry); |
| 261 | put_inotify_watch(&chunk->watch); | 254 | fsnotify_put_mark(entry); |
| 262 | goto out; | 255 | goto out; |
| 263 | } | 256 | } |
| 264 | 257 | ||
| 265 | new = alloc_chunk(size); | 258 | new = alloc_chunk(size); |
| 266 | if (!new) | 259 | if (!new) |
| 267 | goto Fallback; | 260 | goto Fallback; |
| 268 | if (inotify_clone_watch(&chunk->watch, &new->watch) < 0) { | 261 | fsnotify_duplicate_mark(&new->mark, entry); |
| 262 | if (fsnotify_add_mark(&new->mark, new->mark.group, new->mark.i.inode, NULL, 1)) { | ||
| 269 | free_chunk(new); | 263 | free_chunk(new); |
| 270 | goto Fallback; | 264 | goto Fallback; |
| 271 | } | 265 | } |
| @@ -298,9 +292,9 @@ static void untag_chunk(struct node *p) | |||
| 298 | list_for_each_entry(owner, &new->trees, same_root) | 292 | list_for_each_entry(owner, &new->trees, same_root) |
| 299 | owner->root = new; | 293 | owner->root = new; |
| 300 | spin_unlock(&hash_lock); | 294 | spin_unlock(&hash_lock); |
| 301 | inotify_evict_watch(&chunk->watch); | 295 | spin_unlock(&entry->lock); |
| 302 | mutex_unlock(&chunk->watch.inode->inotify_mutex); | 296 | fsnotify_destroy_mark(entry); |
| 303 | put_inotify_watch(&chunk->watch); | 297 | fsnotify_put_mark(entry); |
| 304 | goto out; | 298 | goto out; |
| 305 | 299 | ||
| 306 | Fallback: | 300 | Fallback: |
| @@ -314,31 +308,33 @@ Fallback: | |||
| 314 | p->owner = NULL; | 308 | p->owner = NULL; |
| 315 | put_tree(owner); | 309 | put_tree(owner); |
| 316 | spin_unlock(&hash_lock); | 310 | spin_unlock(&hash_lock); |
| 317 | mutex_unlock(&chunk->watch.inode->inotify_mutex); | 311 | spin_unlock(&entry->lock); |
| 318 | out: | 312 | out: |
| 319 | unpin_inotify_watch(&chunk->watch); | 313 | fsnotify_put_mark(entry); |
| 320 | spin_lock(&hash_lock); | 314 | spin_lock(&hash_lock); |
| 321 | } | 315 | } |
| 322 | 316 | ||
| 323 | static int create_chunk(struct inode *inode, struct audit_tree *tree) | 317 | static int create_chunk(struct inode *inode, struct audit_tree *tree) |
| 324 | { | 318 | { |
| 319 | struct fsnotify_mark *entry; | ||
| 325 | struct audit_chunk *chunk = alloc_chunk(1); | 320 | struct audit_chunk *chunk = alloc_chunk(1); |
| 326 | if (!chunk) | 321 | if (!chunk) |
| 327 | return -ENOMEM; | 322 | return -ENOMEM; |
| 328 | 323 | ||
| 329 | if (inotify_add_watch(rtree_ih, &chunk->watch, inode, IN_IGNORED | IN_DELETE_SELF) < 0) { | 324 | entry = &chunk->mark; |
| 325 | if (fsnotify_add_mark(entry, audit_tree_group, inode, NULL, 0)) { | ||
| 330 | free_chunk(chunk); | 326 | free_chunk(chunk); |
| 331 | return -ENOSPC; | 327 | return -ENOSPC; |
| 332 | } | 328 | } |
| 333 | 329 | ||
| 334 | mutex_lock(&inode->inotify_mutex); | 330 | spin_lock(&entry->lock); |
| 335 | spin_lock(&hash_lock); | 331 | spin_lock(&hash_lock); |
| 336 | if (tree->goner) { | 332 | if (tree->goner) { |
| 337 | spin_unlock(&hash_lock); | 333 | spin_unlock(&hash_lock); |
| 338 | chunk->dead = 1; | 334 | chunk->dead = 1; |
| 339 | inotify_evict_watch(&chunk->watch); | 335 | spin_unlock(&entry->lock); |
| 340 | mutex_unlock(&inode->inotify_mutex); | 336 | fsnotify_destroy_mark(entry); |
| 341 | put_inotify_watch(&chunk->watch); | 337 | fsnotify_put_mark(entry); |
| 342 | return 0; | 338 | return 0; |
| 343 | } | 339 | } |
| 344 | chunk->owners[0].index = (1U << 31); | 340 | chunk->owners[0].index = (1U << 31); |
| @@ -351,30 +347,31 @@ static int create_chunk(struct inode *inode, struct audit_tree *tree) | |||
| 351 | } | 347 | } |
| 352 | insert_hash(chunk); | 348 | insert_hash(chunk); |
| 353 | spin_unlock(&hash_lock); | 349 | spin_unlock(&hash_lock); |
| 354 | mutex_unlock(&inode->inotify_mutex); | 350 | spin_unlock(&entry->lock); |
| 355 | return 0; | 351 | return 0; |
| 356 | } | 352 | } |
| 357 | 353 | ||
| 358 | /* the first tagged inode becomes root of tree */ | 354 | /* the first tagged inode becomes root of tree */ |
| 359 | static int tag_chunk(struct inode *inode, struct audit_tree *tree) | 355 | static int tag_chunk(struct inode *inode, struct audit_tree *tree) |
| 360 | { | 356 | { |
| 361 | struct inotify_watch *watch; | 357 | struct fsnotify_mark *old_entry, *chunk_entry; |
| 362 | struct audit_tree *owner; | 358 | struct audit_tree *owner; |
| 363 | struct audit_chunk *chunk, *old; | 359 | struct audit_chunk *chunk, *old; |
| 364 | struct node *p; | 360 | struct node *p; |
| 365 | int n; | 361 | int n; |
| 366 | 362 | ||
| 367 | if (inotify_find_watch(rtree_ih, inode, &watch) < 0) | 363 | old_entry = fsnotify_find_inode_mark(audit_tree_group, inode); |
| 364 | if (!old_entry) | ||
| 368 | return create_chunk(inode, tree); | 365 | return create_chunk(inode, tree); |
| 369 | 366 | ||
| 370 | old = container_of(watch, struct audit_chunk, watch); | 367 | old = container_of(old_entry, struct audit_chunk, mark); |
| 371 | 368 | ||
| 372 | /* are we already there? */ | 369 | /* are we already there? */ |
| 373 | spin_lock(&hash_lock); | 370 | spin_lock(&hash_lock); |
| 374 | for (n = 0; n < old->count; n++) { | 371 | for (n = 0; n < old->count; n++) { |
| 375 | if (old->owners[n].owner == tree) { | 372 | if (old->owners[n].owner == tree) { |
| 376 | spin_unlock(&hash_lock); | 373 | spin_unlock(&hash_lock); |
| 377 | put_inotify_watch(&old->watch); | 374 | fsnotify_put_mark(old_entry); |
| 378 | return 0; | 375 | return 0; |
| 379 | } | 376 | } |
| 380 | } | 377 | } |
| @@ -382,25 +379,44 @@ static int tag_chunk(struct inode *inode, struct audit_tree *tree) | |||
| 382 | 379 | ||
| 383 | chunk = alloc_chunk(old->count + 1); | 380 | chunk = alloc_chunk(old->count + 1); |
| 384 | if (!chunk) { | 381 | if (!chunk) { |
| 385 | put_inotify_watch(&old->watch); | 382 | fsnotify_put_mark(old_entry); |
| 386 | return -ENOMEM; | 383 | return -ENOMEM; |
| 387 | } | 384 | } |
| 388 | 385 | ||
| 389 | mutex_lock(&inode->inotify_mutex); | 386 | chunk_entry = &chunk->mark; |
| 390 | if (inotify_clone_watch(&old->watch, &chunk->watch) < 0) { | 387 | |
| 391 | mutex_unlock(&inode->inotify_mutex); | 388 | spin_lock(&old_entry->lock); |
| 392 | put_inotify_watch(&old->watch); | 389 | if (!old_entry->i.inode) { |
| 390 | /* old_entry is being shot, lets just lie */ | ||
| 391 | spin_unlock(&old_entry->lock); | ||
| 392 | fsnotify_put_mark(old_entry); | ||
| 393 | free_chunk(chunk); | 393 | free_chunk(chunk); |
| 394 | return -ENOENT; | ||
| 395 | } | ||
| 396 | |||
| 397 | fsnotify_duplicate_mark(chunk_entry, old_entry); | ||
| 398 | if (fsnotify_add_mark(chunk_entry, chunk_entry->group, chunk_entry->i.inode, NULL, 1)) { | ||
| 399 | spin_unlock(&old_entry->lock); | ||
| 400 | free_chunk(chunk); | ||
| 401 | fsnotify_put_mark(old_entry); | ||
| 394 | return -ENOSPC; | 402 | return -ENOSPC; |
| 395 | } | 403 | } |
| 404 | |||
| 405 | /* even though we hold old_entry->lock, this is safe since chunk_entry->lock could NEVER have been grabbed before */ | ||
| 406 | spin_lock(&chunk_entry->lock); | ||
| 396 | spin_lock(&hash_lock); | 407 | spin_lock(&hash_lock); |
| 408 | |||
| 409 | /* we now hold old_entry->lock, chunk_entry->lock, and hash_lock */ | ||
| 397 | if (tree->goner) { | 410 | if (tree->goner) { |
| 398 | spin_unlock(&hash_lock); | 411 | spin_unlock(&hash_lock); |
| 399 | chunk->dead = 1; | 412 | chunk->dead = 1; |
| 400 | inotify_evict_watch(&chunk->watch); | 413 | spin_unlock(&chunk_entry->lock); |
| 401 | mutex_unlock(&inode->inotify_mutex); | 414 | spin_unlock(&old_entry->lock); |
| 402 | put_inotify_watch(&old->watch); | 415 | |
| 403 | put_inotify_watch(&chunk->watch); | 416 | fsnotify_destroy_mark(chunk_entry); |
| 417 | |||
| 418 | fsnotify_put_mark(chunk_entry); | ||
| 419 | fsnotify_put_mark(old_entry); | ||
| 404 | return 0; | 420 | return 0; |
| 405 | } | 421 | } |
| 406 | list_replace_init(&old->trees, &chunk->trees); | 422 | list_replace_init(&old->trees, &chunk->trees); |
| @@ -426,10 +442,11 @@ static int tag_chunk(struct inode *inode, struct audit_tree *tree) | |||
| 426 | list_add(&tree->same_root, &chunk->trees); | 442 | list_add(&tree->same_root, &chunk->trees); |
| 427 | } | 443 | } |
| 428 | spin_unlock(&hash_lock); | 444 | spin_unlock(&hash_lock); |
| 429 | inotify_evict_watch(&old->watch); | 445 | spin_unlock(&chunk_entry->lock); |
| 430 | mutex_unlock(&inode->inotify_mutex); | 446 | spin_unlock(&old_entry->lock); |
| 431 | put_inotify_watch(&old->watch); /* pair to inotify_find_watch */ | 447 | fsnotify_destroy_mark(old_entry); |
| 432 | put_inotify_watch(&old->watch); /* and kill it */ | 448 | fsnotify_put_mark(old_entry); /* pair to fsnotify_find mark_entry */ |
| 449 | fsnotify_put_mark(old_entry); /* and kill it */ | ||
| 433 | return 0; | 450 | return 0; |
| 434 | } | 451 | } |
| 435 | 452 | ||
| @@ -584,7 +601,9 @@ void audit_trim_trees(void) | |||
| 584 | 601 | ||
| 585 | spin_lock(&hash_lock); | 602 | spin_lock(&hash_lock); |
| 586 | list_for_each_entry(node, &tree->chunks, list) { | 603 | list_for_each_entry(node, &tree->chunks, list) { |
| 587 | struct inode *inode = find_chunk(node)->watch.inode; | 604 | struct audit_chunk *chunk = find_chunk(node); |
| 605 | /* this could be NULL if the watch is dieing else where... */ | ||
| 606 | struct inode *inode = chunk->mark.i.inode; | ||
| 588 | node->index |= 1U<<31; | 607 | node->index |= 1U<<31; |
| 589 | if (iterate_mounts(compare_root, inode, root_mnt)) | 608 | if (iterate_mounts(compare_root, inode, root_mnt)) |
| 590 | node->index &= ~(1U<<31); | 609 | node->index &= ~(1U<<31); |
| @@ -846,7 +865,6 @@ void audit_kill_trees(struct list_head *list) | |||
| 846 | * Here comes the stuff asynchronous to auditctl operations | 865 | * Here comes the stuff asynchronous to auditctl operations |
| 847 | */ | 866 | */ |
| 848 | 867 | ||
| 849 | /* inode->inotify_mutex is locked */ | ||
| 850 | static void evict_chunk(struct audit_chunk *chunk) | 868 | static void evict_chunk(struct audit_chunk *chunk) |
| 851 | { | 869 | { |
| 852 | struct audit_tree *owner; | 870 | struct audit_tree *owner; |
| @@ -885,35 +903,46 @@ static void evict_chunk(struct audit_chunk *chunk) | |||
| 885 | mutex_unlock(&audit_filter_mutex); | 903 | mutex_unlock(&audit_filter_mutex); |
| 886 | } | 904 | } |
| 887 | 905 | ||
| 888 | static void handle_event(struct inotify_watch *watch, u32 wd, u32 mask, | 906 | static int audit_tree_handle_event(struct fsnotify_group *group, |
| 889 | u32 cookie, const char *dname, struct inode *inode) | 907 | struct fsnotify_mark *inode_mark, |
| 908 | struct fsnotify_mark *vfsmonut_mark, | ||
| 909 | struct fsnotify_event *event) | ||
| 910 | { | ||
| 911 | BUG(); | ||
| 912 | return -EOPNOTSUPP; | ||
| 913 | } | ||
| 914 | |||
| 915 | static void audit_tree_freeing_mark(struct fsnotify_mark *entry, struct fsnotify_group *group) | ||
| 890 | { | 916 | { |
| 891 | struct audit_chunk *chunk = container_of(watch, struct audit_chunk, watch); | 917 | struct audit_chunk *chunk = container_of(entry, struct audit_chunk, mark); |
| 892 | 918 | ||
| 893 | if (mask & IN_IGNORED) { | 919 | evict_chunk(chunk); |
| 894 | evict_chunk(chunk); | 920 | fsnotify_put_mark(entry); |
| 895 | put_inotify_watch(watch); | ||
| 896 | } | ||
| 897 | } | 921 | } |
| 898 | 922 | ||
| 899 | static void destroy_watch(struct inotify_watch *watch) | 923 | static bool audit_tree_send_event(struct fsnotify_group *group, struct inode *inode, |
| 924 | struct fsnotify_mark *inode_mark, | ||
| 925 | struct fsnotify_mark *vfsmount_mark, | ||
| 926 | __u32 mask, void *data, int data_type) | ||
| 900 | { | 927 | { |
| 901 | struct audit_chunk *chunk = container_of(watch, struct audit_chunk, watch); | 928 | return false; |
| 902 | call_rcu(&chunk->head, __put_chunk); | ||
| 903 | } | 929 | } |
| 904 | 930 | ||
| 905 | static const struct inotify_operations rtree_inotify_ops = { | 931 | static const struct fsnotify_ops audit_tree_ops = { |
| 906 | .handle_event = handle_event, | 932 | .handle_event = audit_tree_handle_event, |
| 907 | .destroy_watch = destroy_watch, | 933 | .should_send_event = audit_tree_send_event, |
| 934 | .free_group_priv = NULL, | ||
| 935 | .free_event_priv = NULL, | ||
| 936 | .freeing_mark = audit_tree_freeing_mark, | ||
| 908 | }; | 937 | }; |
| 909 | 938 | ||
| 910 | static int __init audit_tree_init(void) | 939 | static int __init audit_tree_init(void) |
| 911 | { | 940 | { |
| 912 | int i; | 941 | int i; |
| 913 | 942 | ||
| 914 | rtree_ih = inotify_init(&rtree_inotify_ops); | 943 | audit_tree_group = fsnotify_alloc_group(&audit_tree_ops); |
| 915 | if (IS_ERR(rtree_ih)) | 944 | if (IS_ERR(audit_tree_group)) |
| 916 | audit_panic("cannot initialize inotify handle for rectree watches"); | 945 | audit_panic("cannot initialize fsnotify group for rectree watches"); |
| 917 | 946 | ||
| 918 | for (i = 0; i < HASH_SIZE; i++) | 947 | for (i = 0; i < HASH_SIZE; i++) |
| 919 | INIT_LIST_HEAD(&chunk_hash_heads[i]); | 948 | INIT_LIST_HEAD(&chunk_hash_heads[i]); |
diff --git a/kernel/audit_watch.c b/kernel/audit_watch.c index 8df43696f4ba..6bf2306be7d6 100644 --- a/kernel/audit_watch.c +++ b/kernel/audit_watch.c | |||
| @@ -24,18 +24,18 @@ | |||
| 24 | #include <linux/kthread.h> | 24 | #include <linux/kthread.h> |
| 25 | #include <linux/mutex.h> | 25 | #include <linux/mutex.h> |
| 26 | #include <linux/fs.h> | 26 | #include <linux/fs.h> |
| 27 | #include <linux/fsnotify_backend.h> | ||
| 27 | #include <linux/namei.h> | 28 | #include <linux/namei.h> |
| 28 | #include <linux/netlink.h> | 29 | #include <linux/netlink.h> |
| 29 | #include <linux/sched.h> | 30 | #include <linux/sched.h> |
| 30 | #include <linux/slab.h> | 31 | #include <linux/slab.h> |
| 31 | #include <linux/inotify.h> | ||
| 32 | #include <linux/security.h> | 32 | #include <linux/security.h> |
| 33 | #include "audit.h" | 33 | #include "audit.h" |
| 34 | 34 | ||
| 35 | /* | 35 | /* |
| 36 | * Reference counting: | 36 | * Reference counting: |
| 37 | * | 37 | * |
| 38 | * audit_parent: lifetime is from audit_init_parent() to receipt of an IN_IGNORED | 38 | * audit_parent: lifetime is from audit_init_parent() to receipt of an FS_IGNORED |
| 39 | * event. Each audit_watch holds a reference to its associated parent. | 39 | * event. Each audit_watch holds a reference to its associated parent. |
| 40 | * | 40 | * |
| 41 | * audit_watch: if added to lists, lifetime is from audit_init_watch() to | 41 | * audit_watch: if added to lists, lifetime is from audit_init_watch() to |
| @@ -51,40 +51,61 @@ struct audit_watch { | |||
| 51 | unsigned long ino; /* associated inode number */ | 51 | unsigned long ino; /* associated inode number */ |
| 52 | struct audit_parent *parent; /* associated parent */ | 52 | struct audit_parent *parent; /* associated parent */ |
| 53 | struct list_head wlist; /* entry in parent->watches list */ | 53 | struct list_head wlist; /* entry in parent->watches list */ |
| 54 | struct list_head rules; /* associated rules */ | 54 | struct list_head rules; /* anchor for krule->rlist */ |
| 55 | }; | 55 | }; |
| 56 | 56 | ||
| 57 | struct audit_parent { | 57 | struct audit_parent { |
| 58 | struct list_head ilist; /* entry in inotify registration list */ | 58 | struct list_head watches; /* anchor for audit_watch->wlist */ |
| 59 | struct list_head watches; /* associated watches */ | 59 | struct fsnotify_mark mark; /* fsnotify mark on the inode */ |
| 60 | struct inotify_watch wdata; /* inotify watch data */ | ||
| 61 | unsigned flags; /* status flags */ | ||
| 62 | }; | 60 | }; |
| 63 | 61 | ||
| 64 | /* Inotify handle. */ | 62 | /* fsnotify handle. */ |
| 65 | struct inotify_handle *audit_ih; | 63 | struct fsnotify_group *audit_watch_group; |
| 66 | 64 | ||
| 67 | /* | 65 | /* fsnotify events we care about. */ |
| 68 | * audit_parent status flags: | 66 | #define AUDIT_FS_WATCH (FS_MOVE | FS_CREATE | FS_DELETE | FS_DELETE_SELF |\ |
| 69 | * | 67 | FS_MOVE_SELF | FS_EVENT_ON_CHILD) |
| 70 | * AUDIT_PARENT_INVALID - set anytime rules/watches are auto-removed due to | ||
| 71 | * a filesystem event to ensure we're adding audit watches to a valid parent. | ||
| 72 | * Technically not needed for IN_DELETE_SELF or IN_UNMOUNT events, as we cannot | ||
| 73 | * receive them while we have nameidata, but must be used for IN_MOVE_SELF which | ||
| 74 | * we can receive while holding nameidata. | ||
| 75 | */ | ||
| 76 | #define AUDIT_PARENT_INVALID 0x001 | ||
| 77 | 68 | ||
| 78 | /* Inotify events we care about. */ | 69 | static void audit_free_parent(struct audit_parent *parent) |
| 79 | #define AUDIT_IN_WATCH IN_MOVE|IN_CREATE|IN_DELETE|IN_DELETE_SELF|IN_MOVE_SELF | 70 | { |
| 71 | WARN_ON(!list_empty(&parent->watches)); | ||
| 72 | kfree(parent); | ||
| 73 | } | ||
| 80 | 74 | ||
| 81 | static void audit_free_parent(struct inotify_watch *i_watch) | 75 | static void audit_watch_free_mark(struct fsnotify_mark *entry) |
| 82 | { | 76 | { |
| 83 | struct audit_parent *parent; | 77 | struct audit_parent *parent; |
| 84 | 78 | ||
| 85 | parent = container_of(i_watch, struct audit_parent, wdata); | 79 | parent = container_of(entry, struct audit_parent, mark); |
| 86 | WARN_ON(!list_empty(&parent->watches)); | 80 | audit_free_parent(parent); |
| 87 | kfree(parent); | 81 | } |
| 82 | |||
| 83 | static void audit_get_parent(struct audit_parent *parent) | ||
| 84 | { | ||
| 85 | if (likely(parent)) | ||
| 86 | fsnotify_get_mark(&parent->mark); | ||
| 87 | } | ||
| 88 | |||
| 89 | static void audit_put_parent(struct audit_parent *parent) | ||
| 90 | { | ||
| 91 | if (likely(parent)) | ||
| 92 | fsnotify_put_mark(&parent->mark); | ||
| 93 | } | ||
| 94 | |||
| 95 | /* | ||
| 96 | * Find and return the audit_parent on the given inode. If found a reference | ||
| 97 | * is taken on this parent. | ||
| 98 | */ | ||
| 99 | static inline struct audit_parent *audit_find_parent(struct inode *inode) | ||
| 100 | { | ||
| 101 | struct audit_parent *parent = NULL; | ||
| 102 | struct fsnotify_mark *entry; | ||
| 103 | |||
| 104 | entry = fsnotify_find_inode_mark(audit_watch_group, inode); | ||
| 105 | if (entry) | ||
| 106 | parent = container_of(entry, struct audit_parent, mark); | ||
| 107 | |||
| 108 | return parent; | ||
| 88 | } | 109 | } |
| 89 | 110 | ||
| 90 | void audit_get_watch(struct audit_watch *watch) | 111 | void audit_get_watch(struct audit_watch *watch) |
| @@ -105,7 +126,7 @@ void audit_put_watch(struct audit_watch *watch) | |||
| 105 | void audit_remove_watch(struct audit_watch *watch) | 126 | void audit_remove_watch(struct audit_watch *watch) |
| 106 | { | 127 | { |
| 107 | list_del(&watch->wlist); | 128 | list_del(&watch->wlist); |
| 108 | put_inotify_watch(&watch->parent->wdata); | 129 | audit_put_parent(watch->parent); |
| 109 | watch->parent = NULL; | 130 | watch->parent = NULL; |
| 110 | audit_put_watch(watch); /* match initial get */ | 131 | audit_put_watch(watch); /* match initial get */ |
| 111 | } | 132 | } |
| @@ -115,42 +136,32 @@ char *audit_watch_path(struct audit_watch *watch) | |||
| 115 | return watch->path; | 136 | return watch->path; |
| 116 | } | 137 | } |
| 117 | 138 | ||
| 118 | struct list_head *audit_watch_rules(struct audit_watch *watch) | 139 | int audit_watch_compare(struct audit_watch *watch, unsigned long ino, dev_t dev) |
| 119 | { | ||
| 120 | return &watch->rules; | ||
| 121 | } | ||
| 122 | |||
| 123 | unsigned long audit_watch_inode(struct audit_watch *watch) | ||
| 124 | { | 140 | { |
| 125 | return watch->ino; | 141 | return (watch->ino != (unsigned long)-1) && |
| 126 | } | 142 | (watch->ino == ino) && |
| 127 | 143 | (watch->dev == dev); | |
| 128 | dev_t audit_watch_dev(struct audit_watch *watch) | ||
| 129 | { | ||
| 130 | return watch->dev; | ||
| 131 | } | 144 | } |
| 132 | 145 | ||
| 133 | /* Initialize a parent watch entry. */ | 146 | /* Initialize a parent watch entry. */ |
| 134 | static struct audit_parent *audit_init_parent(struct nameidata *ndp) | 147 | static struct audit_parent *audit_init_parent(struct nameidata *ndp) |
| 135 | { | 148 | { |
| 149 | struct inode *inode = ndp->path.dentry->d_inode; | ||
| 136 | struct audit_parent *parent; | 150 | struct audit_parent *parent; |
| 137 | s32 wd; | 151 | int ret; |
| 138 | 152 | ||
| 139 | parent = kzalloc(sizeof(*parent), GFP_KERNEL); | 153 | parent = kzalloc(sizeof(*parent), GFP_KERNEL); |
| 140 | if (unlikely(!parent)) | 154 | if (unlikely(!parent)) |
| 141 | return ERR_PTR(-ENOMEM); | 155 | return ERR_PTR(-ENOMEM); |
| 142 | 156 | ||
| 143 | INIT_LIST_HEAD(&parent->watches); | 157 | INIT_LIST_HEAD(&parent->watches); |
| 144 | parent->flags = 0; | 158 | |
| 145 | 159 | fsnotify_init_mark(&parent->mark, audit_watch_free_mark); | |
| 146 | inotify_init_watch(&parent->wdata); | 160 | parent->mark.mask = AUDIT_FS_WATCH; |
| 147 | /* grab a ref so inotify watch hangs around until we take audit_filter_mutex */ | 161 | ret = fsnotify_add_mark(&parent->mark, audit_watch_group, inode, NULL, 0); |
| 148 | get_inotify_watch(&parent->wdata); | 162 | if (ret < 0) { |
| 149 | wd = inotify_add_watch(audit_ih, &parent->wdata, | 163 | audit_free_parent(parent); |
| 150 | ndp->path.dentry->d_inode, AUDIT_IN_WATCH); | 164 | return ERR_PTR(ret); |
| 151 | if (wd < 0) { | ||
| 152 | audit_free_parent(&parent->wdata); | ||
| 153 | return ERR_PTR(wd); | ||
| 154 | } | 165 | } |
| 155 | 166 | ||
| 156 | return parent; | 167 | return parent; |
| @@ -179,7 +190,7 @@ int audit_to_watch(struct audit_krule *krule, char *path, int len, u32 op) | |||
| 179 | { | 190 | { |
| 180 | struct audit_watch *watch; | 191 | struct audit_watch *watch; |
| 181 | 192 | ||
| 182 | if (!audit_ih) | 193 | if (!audit_watch_group) |
| 183 | return -EOPNOTSUPP; | 194 | return -EOPNOTSUPP; |
| 184 | 195 | ||
| 185 | if (path[0] != '/' || path[len-1] == '/' || | 196 | if (path[0] != '/' || path[len-1] == '/' || |
| @@ -217,7 +228,7 @@ static struct audit_watch *audit_dupe_watch(struct audit_watch *old) | |||
| 217 | 228 | ||
| 218 | new->dev = old->dev; | 229 | new->dev = old->dev; |
| 219 | new->ino = old->ino; | 230 | new->ino = old->ino; |
| 220 | get_inotify_watch(&old->parent->wdata); | 231 | audit_get_parent(old->parent); |
| 221 | new->parent = old->parent; | 232 | new->parent = old->parent; |
| 222 | 233 | ||
| 223 | out: | 234 | out: |
| @@ -251,15 +262,19 @@ static void audit_update_watch(struct audit_parent *parent, | |||
| 251 | struct audit_entry *oentry, *nentry; | 262 | struct audit_entry *oentry, *nentry; |
| 252 | 263 | ||
| 253 | mutex_lock(&audit_filter_mutex); | 264 | mutex_lock(&audit_filter_mutex); |
| 265 | /* Run all of the watches on this parent looking for the one that | ||
| 266 | * matches the given dname */ | ||
| 254 | list_for_each_entry_safe(owatch, nextw, &parent->watches, wlist) { | 267 | list_for_each_entry_safe(owatch, nextw, &parent->watches, wlist) { |
| 255 | if (audit_compare_dname_path(dname, owatch->path, NULL)) | 268 | if (audit_compare_dname_path(dname, owatch->path, NULL)) |
| 256 | continue; | 269 | continue; |
| 257 | 270 | ||
| 258 | /* If the update involves invalidating rules, do the inode-based | 271 | /* If the update involves invalidating rules, do the inode-based |
| 259 | * filtering now, so we don't omit records. */ | 272 | * filtering now, so we don't omit records. */ |
| 260 | if (invalidating && current->audit_context) | 273 | if (invalidating && !audit_dummy_context()) |
| 261 | audit_filter_inodes(current, current->audit_context); | 274 | audit_filter_inodes(current, current->audit_context); |
| 262 | 275 | ||
| 276 | /* updating ino will likely change which audit_hash_list we | ||
| 277 | * are on so we need a new watch for the new list */ | ||
| 263 | nwatch = audit_dupe_watch(owatch); | 278 | nwatch = audit_dupe_watch(owatch); |
| 264 | if (IS_ERR(nwatch)) { | 279 | if (IS_ERR(nwatch)) { |
| 265 | mutex_unlock(&audit_filter_mutex); | 280 | mutex_unlock(&audit_filter_mutex); |
| @@ -275,12 +290,21 @@ static void audit_update_watch(struct audit_parent *parent, | |||
| 275 | list_del(&oentry->rule.rlist); | 290 | list_del(&oentry->rule.rlist); |
| 276 | list_del_rcu(&oentry->list); | 291 | list_del_rcu(&oentry->list); |
| 277 | 292 | ||
| 278 | nentry = audit_dupe_rule(&oentry->rule, nwatch); | 293 | nentry = audit_dupe_rule(&oentry->rule); |
| 279 | if (IS_ERR(nentry)) { | 294 | if (IS_ERR(nentry)) { |
| 280 | list_del(&oentry->rule.list); | 295 | list_del(&oentry->rule.list); |
| 281 | audit_panic("error updating watch, removing"); | 296 | audit_panic("error updating watch, removing"); |
| 282 | } else { | 297 | } else { |
| 283 | int h = audit_hash_ino((u32)ino); | 298 | int h = audit_hash_ino((u32)ino); |
| 299 | |||
| 300 | /* | ||
| 301 | * nentry->rule.watch == oentry->rule.watch so | ||
| 302 | * we must drop that reference and set it to our | ||
| 303 | * new watch. | ||
| 304 | */ | ||
| 305 | audit_put_watch(nentry->rule.watch); | ||
| 306 | audit_get_watch(nwatch); | ||
| 307 | nentry->rule.watch = nwatch; | ||
| 284 | list_add(&nentry->rule.rlist, &nwatch->rules); | 308 | list_add(&nentry->rule.rlist, &nwatch->rules); |
| 285 | list_add_rcu(&nentry->list, &audit_inode_hash[h]); | 309 | list_add_rcu(&nentry->list, &audit_inode_hash[h]); |
| 286 | list_replace(&oentry->rule.list, | 310 | list_replace(&oentry->rule.list, |
| @@ -312,7 +336,6 @@ static void audit_remove_parent_watches(struct audit_parent *parent) | |||
| 312 | struct audit_entry *e; | 336 | struct audit_entry *e; |
| 313 | 337 | ||
| 314 | mutex_lock(&audit_filter_mutex); | 338 | mutex_lock(&audit_filter_mutex); |
| 315 | parent->flags |= AUDIT_PARENT_INVALID; | ||
| 316 | list_for_each_entry_safe(w, nextw, &parent->watches, wlist) { | 339 | list_for_each_entry_safe(w, nextw, &parent->watches, wlist) { |
| 317 | list_for_each_entry_safe(r, nextr, &w->rules, rlist) { | 340 | list_for_each_entry_safe(r, nextr, &w->rules, rlist) { |
| 318 | e = container_of(r, struct audit_entry, rule); | 341 | e = container_of(r, struct audit_entry, rule); |
| @@ -325,20 +348,8 @@ static void audit_remove_parent_watches(struct audit_parent *parent) | |||
| 325 | audit_remove_watch(w); | 348 | audit_remove_watch(w); |
| 326 | } | 349 | } |
| 327 | mutex_unlock(&audit_filter_mutex); | 350 | mutex_unlock(&audit_filter_mutex); |
| 328 | } | ||
| 329 | |||
| 330 | /* Unregister inotify watches for parents on in_list. | ||
| 331 | * Generates an IN_IGNORED event. */ | ||
| 332 | void audit_inotify_unregister(struct list_head *in_list) | ||
| 333 | { | ||
| 334 | struct audit_parent *p, *n; | ||
| 335 | 351 | ||
| 336 | list_for_each_entry_safe(p, n, in_list, ilist) { | 352 | fsnotify_destroy_mark(&parent->mark); |
| 337 | list_del(&p->ilist); | ||
| 338 | inotify_rm_watch(audit_ih, &p->wdata); | ||
| 339 | /* the unpin matching the pin in audit_do_del_rule() */ | ||
| 340 | unpin_inotify_watch(&p->wdata); | ||
| 341 | } | ||
| 342 | } | 353 | } |
| 343 | 354 | ||
| 344 | /* Get path information necessary for adding watches. */ | 355 | /* Get path information necessary for adding watches. */ |
| @@ -389,7 +400,7 @@ static void audit_put_nd(struct nameidata *ndp, struct nameidata *ndw) | |||
| 389 | } | 400 | } |
| 390 | } | 401 | } |
| 391 | 402 | ||
| 392 | /* Associate the given rule with an existing parent inotify_watch. | 403 | /* Associate the given rule with an existing parent. |
| 393 | * Caller must hold audit_filter_mutex. */ | 404 | * Caller must hold audit_filter_mutex. */ |
| 394 | static void audit_add_to_parent(struct audit_krule *krule, | 405 | static void audit_add_to_parent(struct audit_krule *krule, |
| 395 | struct audit_parent *parent) | 406 | struct audit_parent *parent) |
| @@ -397,6 +408,8 @@ static void audit_add_to_parent(struct audit_krule *krule, | |||
| 397 | struct audit_watch *w, *watch = krule->watch; | 408 | struct audit_watch *w, *watch = krule->watch; |
| 398 | int watch_found = 0; | 409 | int watch_found = 0; |
| 399 | 410 | ||
| 411 | BUG_ON(!mutex_is_locked(&audit_filter_mutex)); | ||
| 412 | |||
| 400 | list_for_each_entry(w, &parent->watches, wlist) { | 413 | list_for_each_entry(w, &parent->watches, wlist) { |
| 401 | if (strcmp(watch->path, w->path)) | 414 | if (strcmp(watch->path, w->path)) |
| 402 | continue; | 415 | continue; |
| @@ -413,7 +426,7 @@ static void audit_add_to_parent(struct audit_krule *krule, | |||
| 413 | } | 426 | } |
| 414 | 427 | ||
| 415 | if (!watch_found) { | 428 | if (!watch_found) { |
| 416 | get_inotify_watch(&parent->wdata); | 429 | audit_get_parent(parent); |
| 417 | watch->parent = parent; | 430 | watch->parent = parent; |
| 418 | 431 | ||
| 419 | list_add(&watch->wlist, &parent->watches); | 432 | list_add(&watch->wlist, &parent->watches); |
| @@ -423,13 +436,12 @@ static void audit_add_to_parent(struct audit_krule *krule, | |||
| 423 | 436 | ||
| 424 | /* Find a matching watch entry, or add this one. | 437 | /* Find a matching watch entry, or add this one. |
| 425 | * Caller must hold audit_filter_mutex. */ | 438 | * Caller must hold audit_filter_mutex. */ |
| 426 | int audit_add_watch(struct audit_krule *krule) | 439 | int audit_add_watch(struct audit_krule *krule, struct list_head **list) |
| 427 | { | 440 | { |
| 428 | struct audit_watch *watch = krule->watch; | 441 | struct audit_watch *watch = krule->watch; |
| 429 | struct inotify_watch *i_watch; | ||
| 430 | struct audit_parent *parent; | 442 | struct audit_parent *parent; |
| 431 | struct nameidata *ndp = NULL, *ndw = NULL; | 443 | struct nameidata *ndp = NULL, *ndw = NULL; |
| 432 | int ret = 0; | 444 | int h, ret = 0; |
| 433 | 445 | ||
| 434 | mutex_unlock(&audit_filter_mutex); | 446 | mutex_unlock(&audit_filter_mutex); |
| 435 | 447 | ||
| @@ -441,47 +453,38 @@ int audit_add_watch(struct audit_krule *krule) | |||
| 441 | goto error; | 453 | goto error; |
| 442 | } | 454 | } |
| 443 | 455 | ||
| 456 | mutex_lock(&audit_filter_mutex); | ||
| 457 | |||
| 444 | /* update watch filter fields */ | 458 | /* update watch filter fields */ |
| 445 | if (ndw) { | 459 | if (ndw) { |
| 446 | watch->dev = ndw->path.dentry->d_inode->i_sb->s_dev; | 460 | watch->dev = ndw->path.dentry->d_inode->i_sb->s_dev; |
| 447 | watch->ino = ndw->path.dentry->d_inode->i_ino; | 461 | watch->ino = ndw->path.dentry->d_inode->i_ino; |
| 448 | } | 462 | } |
| 449 | 463 | ||
| 450 | /* The audit_filter_mutex must not be held during inotify calls because | 464 | /* either find an old parent or attach a new one */ |
| 451 | * we hold it during inotify event callback processing. If an existing | 465 | parent = audit_find_parent(ndp->path.dentry->d_inode); |
| 452 | * inotify watch is found, inotify_find_watch() grabs a reference before | 466 | if (!parent) { |
| 453 | * returning. | ||
| 454 | */ | ||
| 455 | if (inotify_find_watch(audit_ih, ndp->path.dentry->d_inode, | ||
| 456 | &i_watch) < 0) { | ||
| 457 | parent = audit_init_parent(ndp); | 467 | parent = audit_init_parent(ndp); |
| 458 | if (IS_ERR(parent)) { | 468 | if (IS_ERR(parent)) { |
| 459 | /* caller expects mutex locked */ | ||
| 460 | mutex_lock(&audit_filter_mutex); | ||
| 461 | ret = PTR_ERR(parent); | 469 | ret = PTR_ERR(parent); |
| 462 | goto error; | 470 | goto error; |
| 463 | } | 471 | } |
| 464 | } else | 472 | } |
| 465 | parent = container_of(i_watch, struct audit_parent, wdata); | ||
| 466 | |||
| 467 | mutex_lock(&audit_filter_mutex); | ||
| 468 | 473 | ||
| 469 | /* parent was moved before we took audit_filter_mutex */ | 474 | audit_add_to_parent(krule, parent); |
| 470 | if (parent->flags & AUDIT_PARENT_INVALID) | ||
| 471 | ret = -ENOENT; | ||
| 472 | else | ||
| 473 | audit_add_to_parent(krule, parent); | ||
| 474 | 475 | ||
| 475 | /* match get in audit_init_parent or inotify_find_watch */ | 476 | /* match get in audit_find_parent or audit_init_parent */ |
| 476 | put_inotify_watch(&parent->wdata); | 477 | audit_put_parent(parent); |
| 477 | 478 | ||
| 479 | h = audit_hash_ino((u32)watch->ino); | ||
| 480 | *list = &audit_inode_hash[h]; | ||
| 478 | error: | 481 | error: |
| 479 | audit_put_nd(ndp, ndw); /* NULL args OK */ | 482 | audit_put_nd(ndp, ndw); /* NULL args OK */ |
| 480 | return ret; | 483 | return ret; |
| 481 | 484 | ||
| 482 | } | 485 | } |
| 483 | 486 | ||
| 484 | void audit_remove_watch_rule(struct audit_krule *krule, struct list_head *list) | 487 | void audit_remove_watch_rule(struct audit_krule *krule) |
| 485 | { | 488 | { |
| 486 | struct audit_watch *watch = krule->watch; | 489 | struct audit_watch *watch = krule->watch; |
| 487 | struct audit_parent *parent = watch->parent; | 490 | struct audit_parent *parent = watch->parent; |
| @@ -492,53 +495,74 @@ void audit_remove_watch_rule(struct audit_krule *krule, struct list_head *list) | |||
| 492 | audit_remove_watch(watch); | 495 | audit_remove_watch(watch); |
| 493 | 496 | ||
| 494 | if (list_empty(&parent->watches)) { | 497 | if (list_empty(&parent->watches)) { |
| 495 | /* Put parent on the inotify un-registration | 498 | audit_get_parent(parent); |
| 496 | * list. Grab a reference before releasing | 499 | fsnotify_destroy_mark(&parent->mark); |
| 497 | * audit_filter_mutex, to be released in | 500 | audit_put_parent(parent); |
| 498 | * audit_inotify_unregister(). | ||
| 499 | * If filesystem is going away, just leave | ||
| 500 | * the sucker alone, eviction will take | ||
| 501 | * care of it. */ | ||
| 502 | if (pin_inotify_watch(&parent->wdata)) | ||
| 503 | list_add(&parent->ilist, list); | ||
| 504 | } | 501 | } |
| 505 | } | 502 | } |
| 506 | } | 503 | } |
| 507 | 504 | ||
| 508 | /* Update watch data in audit rules based on inotify events. */ | 505 | static bool audit_watch_should_send_event(struct fsnotify_group *group, struct inode *inode, |
| 509 | static void audit_handle_ievent(struct inotify_watch *i_watch, u32 wd, u32 mask, | 506 | struct fsnotify_mark *inode_mark, |
| 510 | u32 cookie, const char *dname, struct inode *inode) | 507 | struct fsnotify_mark *vfsmount_mark, |
| 508 | __u32 mask, void *data, int data_type) | ||
| 509 | { | ||
| 510 | return true; | ||
| 511 | } | ||
| 512 | |||
| 513 | /* Update watch data in audit rules based on fsnotify events. */ | ||
| 514 | static int audit_watch_handle_event(struct fsnotify_group *group, | ||
| 515 | struct fsnotify_mark *inode_mark, | ||
| 516 | struct fsnotify_mark *vfsmount_mark, | ||
| 517 | struct fsnotify_event *event) | ||
| 511 | { | 518 | { |
| 519 | struct inode *inode; | ||
| 520 | __u32 mask = event->mask; | ||
| 521 | const char *dname = event->file_name; | ||
| 512 | struct audit_parent *parent; | 522 | struct audit_parent *parent; |
| 513 | 523 | ||
| 514 | parent = container_of(i_watch, struct audit_parent, wdata); | 524 | parent = container_of(inode_mark, struct audit_parent, mark); |
| 515 | 525 | ||
| 516 | if (mask & (IN_CREATE|IN_MOVED_TO) && inode) | 526 | BUG_ON(group != audit_watch_group); |
| 517 | audit_update_watch(parent, dname, inode->i_sb->s_dev, | 527 | |
| 518 | inode->i_ino, 0); | 528 | switch (event->data_type) { |
| 519 | else if (mask & (IN_DELETE|IN_MOVED_FROM)) | 529 | case (FSNOTIFY_EVENT_FILE): |
| 530 | inode = event->file->f_path.dentry->d_inode; | ||
| 531 | break; | ||
| 532 | case (FSNOTIFY_EVENT_INODE): | ||
| 533 | inode = event->inode; | ||
| 534 | break; | ||
| 535 | default: | ||
| 536 | BUG(); | ||
| 537 | inode = NULL; | ||
| 538 | break; | ||
| 539 | }; | ||
| 540 | |||
| 541 | if (mask & (FS_CREATE|FS_MOVED_TO) && inode) | ||
| 542 | audit_update_watch(parent, dname, inode->i_sb->s_dev, inode->i_ino, 0); | ||
| 543 | else if (mask & (FS_DELETE|FS_MOVED_FROM)) | ||
| 520 | audit_update_watch(parent, dname, (dev_t)-1, (unsigned long)-1, 1); | 544 | audit_update_watch(parent, dname, (dev_t)-1, (unsigned long)-1, 1); |
| 521 | /* inotify automatically removes the watch and sends IN_IGNORED */ | 545 | else if (mask & (FS_DELETE_SELF|FS_UNMOUNT|FS_MOVE_SELF)) |
| 522 | else if (mask & (IN_DELETE_SELF|IN_UNMOUNT)) | ||
| 523 | audit_remove_parent_watches(parent); | ||
| 524 | /* inotify does not remove the watch, so remove it manually */ | ||
| 525 | else if(mask & IN_MOVE_SELF) { | ||
| 526 | audit_remove_parent_watches(parent); | 546 | audit_remove_parent_watches(parent); |
| 527 | inotify_remove_watch_locked(audit_ih, i_watch); | 547 | |
| 528 | } else if (mask & IN_IGNORED) | 548 | return 0; |
| 529 | put_inotify_watch(i_watch); | ||
| 530 | } | 549 | } |
| 531 | 550 | ||
| 532 | static const struct inotify_operations audit_inotify_ops = { | 551 | static const struct fsnotify_ops audit_watch_fsnotify_ops = { |
| 533 | .handle_event = audit_handle_ievent, | 552 | .should_send_event = audit_watch_should_send_event, |
| 534 | .destroy_watch = audit_free_parent, | 553 | .handle_event = audit_watch_handle_event, |
| 554 | .free_group_priv = NULL, | ||
| 555 | .freeing_mark = NULL, | ||
| 556 | .free_event_priv = NULL, | ||
| 535 | }; | 557 | }; |
| 536 | 558 | ||
| 537 | static int __init audit_watch_init(void) | 559 | static int __init audit_watch_init(void) |
| 538 | { | 560 | { |
| 539 | audit_ih = inotify_init(&audit_inotify_ops); | 561 | audit_watch_group = fsnotify_alloc_group(&audit_watch_fsnotify_ops); |
| 540 | if (IS_ERR(audit_ih)) | 562 | if (IS_ERR(audit_watch_group)) { |
| 541 | audit_panic("cannot initialize inotify handle"); | 563 | audit_watch_group = NULL; |
| 564 | audit_panic("cannot create audit fsnotify group"); | ||
| 565 | } | ||
| 542 | return 0; | 566 | return 0; |
| 543 | } | 567 | } |
| 544 | subsys_initcall(audit_watch_init); | 568 | device_initcall(audit_watch_init); |
diff --git a/kernel/auditfilter.c b/kernel/auditfilter.c index ce08041f578d..eb7675499fb5 100644 --- a/kernel/auditfilter.c +++ b/kernel/auditfilter.c | |||
| @@ -71,6 +71,7 @@ static inline void audit_free_rule(struct audit_entry *e) | |||
| 71 | { | 71 | { |
| 72 | int i; | 72 | int i; |
| 73 | struct audit_krule *erule = &e->rule; | 73 | struct audit_krule *erule = &e->rule; |
| 74 | |||
| 74 | /* some rules don't have associated watches */ | 75 | /* some rules don't have associated watches */ |
| 75 | if (erule->watch) | 76 | if (erule->watch) |
| 76 | audit_put_watch(erule->watch); | 77 | audit_put_watch(erule->watch); |
| @@ -746,8 +747,7 @@ static inline int audit_dupe_lsm_field(struct audit_field *df, | |||
| 746 | * rule with the new rule in the filterlist, then free the old rule. | 747 | * rule with the new rule in the filterlist, then free the old rule. |
| 747 | * The rlist element is undefined; list manipulations are handled apart from | 748 | * The rlist element is undefined; list manipulations are handled apart from |
| 748 | * the initial copy. */ | 749 | * the initial copy. */ |
| 749 | struct audit_entry *audit_dupe_rule(struct audit_krule *old, | 750 | struct audit_entry *audit_dupe_rule(struct audit_krule *old) |
| 750 | struct audit_watch *watch) | ||
| 751 | { | 751 | { |
| 752 | u32 fcount = old->field_count; | 752 | u32 fcount = old->field_count; |
| 753 | struct audit_entry *entry; | 753 | struct audit_entry *entry; |
| @@ -769,8 +769,8 @@ struct audit_entry *audit_dupe_rule(struct audit_krule *old, | |||
| 769 | new->prio = old->prio; | 769 | new->prio = old->prio; |
| 770 | new->buflen = old->buflen; | 770 | new->buflen = old->buflen; |
| 771 | new->inode_f = old->inode_f; | 771 | new->inode_f = old->inode_f; |
| 772 | new->watch = NULL; | ||
| 773 | new->field_count = old->field_count; | 772 | new->field_count = old->field_count; |
| 773 | |||
| 774 | /* | 774 | /* |
| 775 | * note that we are OK with not refcounting here; audit_match_tree() | 775 | * note that we are OK with not refcounting here; audit_match_tree() |
| 776 | * never dereferences tree and we can't get false positives there | 776 | * never dereferences tree and we can't get false positives there |
| @@ -811,9 +811,9 @@ struct audit_entry *audit_dupe_rule(struct audit_krule *old, | |||
| 811 | } | 811 | } |
| 812 | } | 812 | } |
| 813 | 813 | ||
| 814 | if (watch) { | 814 | if (old->watch) { |
| 815 | audit_get_watch(watch); | 815 | audit_get_watch(old->watch); |
| 816 | new->watch = watch; | 816 | new->watch = old->watch; |
| 817 | } | 817 | } |
| 818 | 818 | ||
| 819 | return entry; | 819 | return entry; |
| @@ -866,7 +866,7 @@ static inline int audit_add_rule(struct audit_entry *entry) | |||
| 866 | struct audit_watch *watch = entry->rule.watch; | 866 | struct audit_watch *watch = entry->rule.watch; |
| 867 | struct audit_tree *tree = entry->rule.tree; | 867 | struct audit_tree *tree = entry->rule.tree; |
| 868 | struct list_head *list; | 868 | struct list_head *list; |
| 869 | int h, err; | 869 | int err; |
| 870 | #ifdef CONFIG_AUDITSYSCALL | 870 | #ifdef CONFIG_AUDITSYSCALL |
| 871 | int dont_count = 0; | 871 | int dont_count = 0; |
| 872 | 872 | ||
| @@ -889,15 +889,11 @@ static inline int audit_add_rule(struct audit_entry *entry) | |||
| 889 | 889 | ||
| 890 | if (watch) { | 890 | if (watch) { |
| 891 | /* audit_filter_mutex is dropped and re-taken during this call */ | 891 | /* audit_filter_mutex is dropped and re-taken during this call */ |
| 892 | err = audit_add_watch(&entry->rule); | 892 | err = audit_add_watch(&entry->rule, &list); |
| 893 | if (err) { | 893 | if (err) { |
| 894 | mutex_unlock(&audit_filter_mutex); | 894 | mutex_unlock(&audit_filter_mutex); |
| 895 | goto error; | 895 | goto error; |
| 896 | } | 896 | } |
| 897 | /* entry->rule.watch may have changed during audit_add_watch() */ | ||
| 898 | watch = entry->rule.watch; | ||
| 899 | h = audit_hash_ino((u32)audit_watch_inode(watch)); | ||
| 900 | list = &audit_inode_hash[h]; | ||
| 901 | } | 897 | } |
| 902 | if (tree) { | 898 | if (tree) { |
| 903 | err = audit_add_tree_rule(&entry->rule); | 899 | err = audit_add_tree_rule(&entry->rule); |
| @@ -949,7 +945,6 @@ static inline int audit_del_rule(struct audit_entry *entry) | |||
| 949 | struct audit_watch *watch = entry->rule.watch; | 945 | struct audit_watch *watch = entry->rule.watch; |
| 950 | struct audit_tree *tree = entry->rule.tree; | 946 | struct audit_tree *tree = entry->rule.tree; |
| 951 | struct list_head *list; | 947 | struct list_head *list; |
| 952 | LIST_HEAD(inotify_list); | ||
| 953 | int ret = 0; | 948 | int ret = 0; |
| 954 | #ifdef CONFIG_AUDITSYSCALL | 949 | #ifdef CONFIG_AUDITSYSCALL |
| 955 | int dont_count = 0; | 950 | int dont_count = 0; |
| @@ -969,7 +964,7 @@ static inline int audit_del_rule(struct audit_entry *entry) | |||
| 969 | } | 964 | } |
| 970 | 965 | ||
| 971 | if (e->rule.watch) | 966 | if (e->rule.watch) |
| 972 | audit_remove_watch_rule(&e->rule, &inotify_list); | 967 | audit_remove_watch_rule(&e->rule); |
| 973 | 968 | ||
| 974 | if (e->rule.tree) | 969 | if (e->rule.tree) |
| 975 | audit_remove_tree_rule(&e->rule); | 970 | audit_remove_tree_rule(&e->rule); |
| @@ -987,9 +982,6 @@ static inline int audit_del_rule(struct audit_entry *entry) | |||
| 987 | #endif | 982 | #endif |
| 988 | mutex_unlock(&audit_filter_mutex); | 983 | mutex_unlock(&audit_filter_mutex); |
| 989 | 984 | ||
| 990 | if (!list_empty(&inotify_list)) | ||
| 991 | audit_inotify_unregister(&inotify_list); | ||
| 992 | |||
| 993 | out: | 985 | out: |
| 994 | if (watch) | 986 | if (watch) |
| 995 | audit_put_watch(watch); /* match initial get */ | 987 | audit_put_watch(watch); /* match initial get */ |
| @@ -1323,30 +1315,23 @@ static int update_lsm_rule(struct audit_krule *r) | |||
| 1323 | { | 1315 | { |
| 1324 | struct audit_entry *entry = container_of(r, struct audit_entry, rule); | 1316 | struct audit_entry *entry = container_of(r, struct audit_entry, rule); |
| 1325 | struct audit_entry *nentry; | 1317 | struct audit_entry *nentry; |
| 1326 | struct audit_watch *watch; | ||
| 1327 | struct audit_tree *tree; | ||
| 1328 | int err = 0; | 1318 | int err = 0; |
| 1329 | 1319 | ||
| 1330 | if (!security_audit_rule_known(r)) | 1320 | if (!security_audit_rule_known(r)) |
| 1331 | return 0; | 1321 | return 0; |
| 1332 | 1322 | ||
| 1333 | watch = r->watch; | 1323 | nentry = audit_dupe_rule(r); |
| 1334 | tree = r->tree; | ||
| 1335 | nentry = audit_dupe_rule(r, watch); | ||
| 1336 | if (IS_ERR(nentry)) { | 1324 | if (IS_ERR(nentry)) { |
| 1337 | /* save the first error encountered for the | 1325 | /* save the first error encountered for the |
| 1338 | * return value */ | 1326 | * return value */ |
| 1339 | err = PTR_ERR(nentry); | 1327 | err = PTR_ERR(nentry); |
| 1340 | audit_panic("error updating LSM filters"); | 1328 | audit_panic("error updating LSM filters"); |
| 1341 | if (watch) | 1329 | if (r->watch) |
| 1342 | list_del(&r->rlist); | 1330 | list_del(&r->rlist); |
| 1343 | list_del_rcu(&entry->list); | 1331 | list_del_rcu(&entry->list); |
| 1344 | list_del(&r->list); | 1332 | list_del(&r->list); |
| 1345 | } else { | 1333 | } else { |
| 1346 | if (watch) { | 1334 | if (r->watch || r->tree) |
| 1347 | list_add(&nentry->rule.rlist, audit_watch_rules(watch)); | ||
| 1348 | list_del(&r->rlist); | ||
| 1349 | } else if (tree) | ||
| 1350 | list_replace_init(&r->rlist, &nentry->rule.rlist); | 1335 | list_replace_init(&r->rlist, &nentry->rule.rlist); |
| 1351 | list_replace_rcu(&entry->list, &nentry->list); | 1336 | list_replace_rcu(&entry->list, &nentry->list); |
| 1352 | list_replace(&r->list, &nentry->rule.list); | 1337 | list_replace(&r->list, &nentry->rule.list); |
diff --git a/kernel/auditsc.c b/kernel/auditsc.c index 3828ad5fb8f1..b87a63beb66c 100644 --- a/kernel/auditsc.c +++ b/kernel/auditsc.c | |||
| @@ -65,7 +65,6 @@ | |||
| 65 | #include <linux/binfmts.h> | 65 | #include <linux/binfmts.h> |
| 66 | #include <linux/highmem.h> | 66 | #include <linux/highmem.h> |
| 67 | #include <linux/syscalls.h> | 67 | #include <linux/syscalls.h> |
| 68 | #include <linux/inotify.h> | ||
| 69 | #include <linux/capability.h> | 68 | #include <linux/capability.h> |
| 70 | #include <linux/fs_struct.h> | 69 | #include <linux/fs_struct.h> |
| 71 | 70 | ||
| @@ -549,9 +548,8 @@ static int audit_filter_rules(struct task_struct *tsk, | |||
| 549 | } | 548 | } |
| 550 | break; | 549 | break; |
| 551 | case AUDIT_WATCH: | 550 | case AUDIT_WATCH: |
| 552 | if (name && audit_watch_inode(rule->watch) != (unsigned long)-1) | 551 | if (name) |
| 553 | result = (name->dev == audit_watch_dev(rule->watch) && | 552 | result = audit_watch_compare(rule->watch, name->ino, name->dev); |
| 554 | name->ino == audit_watch_inode(rule->watch)); | ||
| 555 | break; | 553 | break; |
| 556 | case AUDIT_DIR: | 554 | case AUDIT_DIR: |
| 557 | if (ctx) | 555 | if (ctx) |
| @@ -1726,7 +1724,7 @@ static inline void handle_one(const struct inode *inode) | |||
| 1726 | struct audit_tree_refs *p; | 1724 | struct audit_tree_refs *p; |
| 1727 | struct audit_chunk *chunk; | 1725 | struct audit_chunk *chunk; |
| 1728 | int count; | 1726 | int count; |
| 1729 | if (likely(list_empty(&inode->inotify_watches))) | 1727 | if (likely(hlist_empty(&inode->i_fsnotify_marks))) |
| 1730 | return; | 1728 | return; |
| 1731 | context = current->audit_context; | 1729 | context = current->audit_context; |
| 1732 | p = context->trees; | 1730 | p = context->trees; |
| @@ -1769,7 +1767,7 @@ retry: | |||
| 1769 | seq = read_seqbegin(&rename_lock); | 1767 | seq = read_seqbegin(&rename_lock); |
| 1770 | for(;;) { | 1768 | for(;;) { |
| 1771 | struct inode *inode = d->d_inode; | 1769 | struct inode *inode = d->d_inode; |
| 1772 | if (inode && unlikely(!list_empty(&inode->inotify_watches))) { | 1770 | if (inode && unlikely(!hlist_empty(&inode->i_fsnotify_marks))) { |
| 1773 | struct audit_chunk *chunk; | 1771 | struct audit_chunk *chunk; |
| 1774 | chunk = audit_tree_lookup(inode); | 1772 | chunk = audit_tree_lookup(inode); |
| 1775 | if (chunk) { | 1773 | if (chunk) { |
diff --git a/kernel/sys_ni.c b/kernel/sys_ni.c index 70f2ea758ffe..bad369ec5403 100644 --- a/kernel/sys_ni.c +++ b/kernel/sys_ni.c | |||
| @@ -181,3 +181,7 @@ cond_syscall(sys_eventfd2); | |||
| 181 | 181 | ||
| 182 | /* performance counters: */ | 182 | /* performance counters: */ |
| 183 | cond_syscall(sys_perf_event_open); | 183 | cond_syscall(sys_perf_event_open); |
| 184 | |||
| 185 | /* fanotify! */ | ||
| 186 | cond_syscall(sys_fanotify_init); | ||
| 187 | cond_syscall(sys_fanotify_mark); | ||
diff --git a/kernel/sysctl.c b/kernel/sysctl.c index 6b005e4912b5..ca38e8e3e907 100644 --- a/kernel/sysctl.c +++ b/kernel/sysctl.c | |||
| @@ -44,6 +44,7 @@ | |||
| 44 | #include <linux/times.h> | 44 | #include <linux/times.h> |
| 45 | #include <linux/limits.h> | 45 | #include <linux/limits.h> |
| 46 | #include <linux/dcache.h> | 46 | #include <linux/dcache.h> |
| 47 | #include <linux/dnotify.h> | ||
| 47 | #include <linux/syscalls.h> | 48 | #include <linux/syscalls.h> |
| 48 | #include <linux/vmstat.h> | 49 | #include <linux/vmstat.h> |
| 49 | #include <linux/nfs_fs.h> | 50 | #include <linux/nfs_fs.h> |
| @@ -131,6 +132,9 @@ static int min_percpu_pagelist_fract = 8; | |||
| 131 | 132 | ||
| 132 | static int ngroups_max = NGROUPS_MAX; | 133 | static int ngroups_max = NGROUPS_MAX; |
| 133 | 134 | ||
| 135 | #ifdef CONFIG_INOTIFY_USER | ||
| 136 | #include <linux/inotify.h> | ||
| 137 | #endif | ||
| 134 | #ifdef CONFIG_SPARC | 138 | #ifdef CONFIG_SPARC |
| 135 | #include <asm/system.h> | 139 | #include <asm/system.h> |
| 136 | #endif | 140 | #endif |
| @@ -207,9 +211,6 @@ static struct ctl_table fs_table[]; | |||
| 207 | static struct ctl_table debug_table[]; | 211 | static struct ctl_table debug_table[]; |
| 208 | static struct ctl_table dev_table[]; | 212 | static struct ctl_table dev_table[]; |
| 209 | extern struct ctl_table random_table[]; | 213 | extern struct ctl_table random_table[]; |
| 210 | #ifdef CONFIG_INOTIFY_USER | ||
| 211 | extern struct ctl_table inotify_table[]; | ||
| 212 | #endif | ||
| 213 | #ifdef CONFIG_EPOLL | 214 | #ifdef CONFIG_EPOLL |
| 214 | extern struct ctl_table epoll_table[]; | 215 | extern struct ctl_table epoll_table[]; |
| 215 | #endif | 216 | #endif |
