diff options
| author | Jeff Garzik <jgarzik@pobox.com> | 2005-10-06 05:51:07 -0400 |
|---|---|---|
| committer | Jeff Garzik <jgarzik@pobox.com> | 2005-10-06 05:51:07 -0400 |
| commit | 012e060c95e547eceea4a12c6f58592473bf4011 (patch) | |
| tree | b57d3eafb50ce517577d2cf366c9ef0b4b286589 /kernel | |
| parent | 923f122573851d18a3832ca808269fa2d5046fb1 (diff) | |
| parent | ed39f731ab2e77e58122232f6e27333331d7793d (diff) | |
Merge branch 'master'
Diffstat (limited to 'kernel')
| -rw-r--r-- | kernel/audit.c | 128 | ||||
| -rw-r--r-- | kernel/auditsc.c | 327 | ||||
| -rw-r--r-- | kernel/cpuset.c | 11 | ||||
| -rw-r--r-- | kernel/exit.c | 8 | ||||
| -rw-r--r-- | kernel/fork.c | 3 | ||||
| -rw-r--r-- | kernel/module.c | 11 | ||||
| -rw-r--r-- | kernel/params.c | 10 | ||||
| -rw-r--r-- | kernel/power/Kconfig | 2 | ||||
| -rw-r--r-- | kernel/power/disk.c | 6 | ||||
| -rw-r--r-- | kernel/power/power.h | 7 | ||||
| -rw-r--r-- | kernel/power/swsusp.c | 29 | ||||
| -rw-r--r-- | kernel/printk.c | 7 | ||||
| -rw-r--r-- | kernel/sched.c | 8 | ||||
| -rw-r--r-- | kernel/signal.c | 34 | ||||
| -rw-r--r-- | kernel/sys.c | 55 | ||||
| -rw-r--r-- | kernel/timer.c | 9 |
16 files changed, 437 insertions, 218 deletions
diff --git a/kernel/audit.c b/kernel/audit.c index 7f0699790d46..83096b67510a 100644 --- a/kernel/audit.c +++ b/kernel/audit.c | |||
| @@ -79,6 +79,8 @@ static int audit_rate_limit; | |||
| 79 | 79 | ||
| 80 | /* Number of outstanding audit_buffers allowed. */ | 80 | /* Number of outstanding audit_buffers allowed. */ |
| 81 | static int audit_backlog_limit = 64; | 81 | static int audit_backlog_limit = 64; |
| 82 | static int audit_backlog_wait_time = 60 * HZ; | ||
| 83 | static int audit_backlog_wait_overflow = 0; | ||
| 82 | 84 | ||
| 83 | /* The identity of the user shutting down the audit system. */ | 85 | /* The identity of the user shutting down the audit system. */ |
| 84 | uid_t audit_sig_uid = -1; | 86 | uid_t audit_sig_uid = -1; |
| @@ -106,18 +108,12 @@ static LIST_HEAD(audit_freelist); | |||
| 106 | static struct sk_buff_head audit_skb_queue; | 108 | static struct sk_buff_head audit_skb_queue; |
| 107 | static struct task_struct *kauditd_task; | 109 | static struct task_struct *kauditd_task; |
| 108 | static DECLARE_WAIT_QUEUE_HEAD(kauditd_wait); | 110 | static DECLARE_WAIT_QUEUE_HEAD(kauditd_wait); |
| 109 | 111 | static DECLARE_WAIT_QUEUE_HEAD(audit_backlog_wait); | |
| 110 | /* There are three lists of rules -- one to search at task creation | ||
| 111 | * time, one to search at syscall entry time, and another to search at | ||
| 112 | * syscall exit time. */ | ||
| 113 | static LIST_HEAD(audit_tsklist); | ||
| 114 | static LIST_HEAD(audit_entlist); | ||
| 115 | static LIST_HEAD(audit_extlist); | ||
| 116 | 112 | ||
| 117 | /* The netlink socket is only to be read by 1 CPU, which lets us assume | 113 | /* The netlink socket is only to be read by 1 CPU, which lets us assume |
| 118 | * that list additions and deletions never happen simultaneously in | 114 | * that list additions and deletions never happen simultaneously in |
| 119 | * auditsc.c */ | 115 | * auditsc.c */ |
| 120 | static DECLARE_MUTEX(audit_netlink_sem); | 116 | DECLARE_MUTEX(audit_netlink_sem); |
| 121 | 117 | ||
| 122 | /* AUDIT_BUFSIZ is the size of the temporary buffer used for formatting | 118 | /* AUDIT_BUFSIZ is the size of the temporary buffer used for formatting |
| 123 | * audit records. Since printk uses a 1024 byte buffer, this buffer | 119 | * audit records. Since printk uses a 1024 byte buffer, this buffer |
| @@ -137,6 +133,7 @@ struct audit_buffer { | |||
| 137 | struct list_head list; | 133 | struct list_head list; |
| 138 | struct sk_buff *skb; /* formatted skb ready to send */ | 134 | struct sk_buff *skb; /* formatted skb ready to send */ |
| 139 | struct audit_context *ctx; /* NULL or associated context */ | 135 | struct audit_context *ctx; /* NULL or associated context */ |
| 136 | int gfp_mask; | ||
| 140 | }; | 137 | }; |
| 141 | 138 | ||
| 142 | static void audit_set_pid(struct audit_buffer *ab, pid_t pid) | 139 | static void audit_set_pid(struct audit_buffer *ab, pid_t pid) |
| @@ -145,11 +142,6 @@ static void audit_set_pid(struct audit_buffer *ab, pid_t pid) | |||
| 145 | nlh->nlmsg_pid = pid; | 142 | nlh->nlmsg_pid = pid; |
| 146 | } | 143 | } |
| 147 | 144 | ||
| 148 | struct audit_entry { | ||
| 149 | struct list_head list; | ||
| 150 | struct audit_rule rule; | ||
| 151 | }; | ||
| 152 | |||
| 153 | static void audit_panic(const char *message) | 145 | static void audit_panic(const char *message) |
| 154 | { | 146 | { |
| 155 | switch (audit_failure) | 147 | switch (audit_failure) |
| @@ -233,7 +225,7 @@ static int audit_set_rate_limit(int limit, uid_t loginuid) | |||
| 233 | { | 225 | { |
| 234 | int old = audit_rate_limit; | 226 | int old = audit_rate_limit; |
| 235 | audit_rate_limit = limit; | 227 | audit_rate_limit = limit; |
| 236 | audit_log(NULL, AUDIT_CONFIG_CHANGE, | 228 | audit_log(NULL, GFP_KERNEL, AUDIT_CONFIG_CHANGE, |
| 237 | "audit_rate_limit=%d old=%d by auid=%u", | 229 | "audit_rate_limit=%d old=%d by auid=%u", |
| 238 | audit_rate_limit, old, loginuid); | 230 | audit_rate_limit, old, loginuid); |
| 239 | return old; | 231 | return old; |
| @@ -243,7 +235,7 @@ static int audit_set_backlog_limit(int limit, uid_t loginuid) | |||
| 243 | { | 235 | { |
| 244 | int old = audit_backlog_limit; | 236 | int old = audit_backlog_limit; |
| 245 | audit_backlog_limit = limit; | 237 | audit_backlog_limit = limit; |
| 246 | audit_log(NULL, AUDIT_CONFIG_CHANGE, | 238 | audit_log(NULL, GFP_KERNEL, AUDIT_CONFIG_CHANGE, |
| 247 | "audit_backlog_limit=%d old=%d by auid=%u", | 239 | "audit_backlog_limit=%d old=%d by auid=%u", |
| 248 | audit_backlog_limit, old, loginuid); | 240 | audit_backlog_limit, old, loginuid); |
| 249 | return old; | 241 | return old; |
| @@ -255,7 +247,7 @@ static int audit_set_enabled(int state, uid_t loginuid) | |||
| 255 | if (state != 0 && state != 1) | 247 | if (state != 0 && state != 1) |
| 256 | return -EINVAL; | 248 | return -EINVAL; |
| 257 | audit_enabled = state; | 249 | audit_enabled = state; |
| 258 | audit_log(NULL, AUDIT_CONFIG_CHANGE, | 250 | audit_log(NULL, GFP_KERNEL, AUDIT_CONFIG_CHANGE, |
| 259 | "audit_enabled=%d old=%d by auid=%u", | 251 | "audit_enabled=%d old=%d by auid=%u", |
| 260 | audit_enabled, old, loginuid); | 252 | audit_enabled, old, loginuid); |
| 261 | return old; | 253 | return old; |
| @@ -269,7 +261,7 @@ static int audit_set_failure(int state, uid_t loginuid) | |||
| 269 | && state != AUDIT_FAIL_PANIC) | 261 | && state != AUDIT_FAIL_PANIC) |
| 270 | return -EINVAL; | 262 | return -EINVAL; |
| 271 | audit_failure = state; | 263 | audit_failure = state; |
| 272 | audit_log(NULL, AUDIT_CONFIG_CHANGE, | 264 | audit_log(NULL, GFP_KERNEL, AUDIT_CONFIG_CHANGE, |
| 273 | "audit_failure=%d old=%d by auid=%u", | 265 | "audit_failure=%d old=%d by auid=%u", |
| 274 | audit_failure, old, loginuid); | 266 | audit_failure, old, loginuid); |
| 275 | return old; | 267 | return old; |
| @@ -281,6 +273,7 @@ int kauditd_thread(void *dummy) | |||
| 281 | 273 | ||
| 282 | while (1) { | 274 | while (1) { |
| 283 | skb = skb_dequeue(&audit_skb_queue); | 275 | skb = skb_dequeue(&audit_skb_queue); |
| 276 | wake_up(&audit_backlog_wait); | ||
| 284 | if (skb) { | 277 | if (skb) { |
| 285 | if (audit_pid) { | 278 | if (audit_pid) { |
| 286 | int err = netlink_unicast(audit_sock, skb, audit_pid, 0); | 279 | int err = netlink_unicast(audit_sock, skb, audit_pid, 0); |
| @@ -290,7 +283,7 @@ int kauditd_thread(void *dummy) | |||
| 290 | audit_pid = 0; | 283 | audit_pid = 0; |
| 291 | } | 284 | } |
| 292 | } else { | 285 | } else { |
| 293 | printk(KERN_ERR "%s\n", skb->data + NLMSG_SPACE(0)); | 286 | printk(KERN_NOTICE "%s\n", skb->data + NLMSG_SPACE(0)); |
| 294 | kfree_skb(skb); | 287 | kfree_skb(skb); |
| 295 | } | 288 | } |
| 296 | } else { | 289 | } else { |
| @@ -423,7 +416,7 @@ static int audit_receive_msg(struct sk_buff *skb, struct nlmsghdr *nlh) | |||
| 423 | if (status_get->mask & AUDIT_STATUS_PID) { | 416 | if (status_get->mask & AUDIT_STATUS_PID) { |
| 424 | int old = audit_pid; | 417 | int old = audit_pid; |
| 425 | audit_pid = status_get->pid; | 418 | audit_pid = status_get->pid; |
| 426 | audit_log(NULL, AUDIT_CONFIG_CHANGE, | 419 | audit_log(NULL, GFP_KERNEL, AUDIT_CONFIG_CHANGE, |
| 427 | "audit_pid=%d old=%d by auid=%u", | 420 | "audit_pid=%d old=%d by auid=%u", |
| 428 | audit_pid, old, loginuid); | 421 | audit_pid, old, loginuid); |
| 429 | } | 422 | } |
| @@ -435,15 +428,21 @@ static int audit_receive_msg(struct sk_buff *skb, struct nlmsghdr *nlh) | |||
| 435 | break; | 428 | break; |
| 436 | case AUDIT_USER: | 429 | case AUDIT_USER: |
| 437 | case AUDIT_FIRST_USER_MSG...AUDIT_LAST_USER_MSG: | 430 | case AUDIT_FIRST_USER_MSG...AUDIT_LAST_USER_MSG: |
| 438 | ab = audit_log_start(NULL, msg_type); | 431 | if (!audit_enabled && msg_type != AUDIT_USER_AVC) |
| 439 | if (!ab) | 432 | return 0; |
| 440 | break; /* audit_panic has been called */ | 433 | |
| 441 | audit_log_format(ab, | 434 | err = audit_filter_user(&NETLINK_CB(skb), msg_type); |
| 442 | "user pid=%d uid=%u auid=%u" | 435 | if (err == 1) { |
| 443 | " msg='%.1024s'", | 436 | err = 0; |
| 444 | pid, uid, loginuid, (char *)data); | 437 | ab = audit_log_start(NULL, GFP_KERNEL, msg_type); |
| 445 | audit_set_pid(ab, pid); | 438 | if (ab) { |
| 446 | audit_log_end(ab); | 439 | audit_log_format(ab, |
| 440 | "user pid=%d uid=%u auid=%u msg='%.1024s'", | ||
| 441 | pid, uid, loginuid, (char *)data); | ||
| 442 | audit_set_pid(ab, pid); | ||
| 443 | audit_log_end(ab); | ||
| 444 | } | ||
| 445 | } | ||
| 447 | break; | 446 | break; |
| 448 | case AUDIT_ADD: | 447 | case AUDIT_ADD: |
| 449 | case AUDIT_DEL: | 448 | case AUDIT_DEL: |
| @@ -523,7 +522,7 @@ static int __init audit_init(void) | |||
| 523 | skb_queue_head_init(&audit_skb_queue); | 522 | skb_queue_head_init(&audit_skb_queue); |
| 524 | audit_initialized = 1; | 523 | audit_initialized = 1; |
| 525 | audit_enabled = audit_default; | 524 | audit_enabled = audit_default; |
| 526 | audit_log(NULL, AUDIT_KERNEL, "initialized"); | 525 | audit_log(NULL, GFP_KERNEL, AUDIT_KERNEL, "initialized"); |
| 527 | return 0; | 526 | return 0; |
| 528 | } | 527 | } |
| 529 | __initcall(audit_init); | 528 | __initcall(audit_init); |
| @@ -561,7 +560,7 @@ static void audit_buffer_free(struct audit_buffer *ab) | |||
| 561 | } | 560 | } |
| 562 | 561 | ||
| 563 | static struct audit_buffer * audit_buffer_alloc(struct audit_context *ctx, | 562 | static struct audit_buffer * audit_buffer_alloc(struct audit_context *ctx, |
| 564 | int gfp_mask, int type) | 563 | unsigned int __nocast gfp_mask, int type) |
| 565 | { | 564 | { |
| 566 | unsigned long flags; | 565 | unsigned long flags; |
| 567 | struct audit_buffer *ab = NULL; | 566 | struct audit_buffer *ab = NULL; |
| @@ -587,6 +586,7 @@ static struct audit_buffer * audit_buffer_alloc(struct audit_context *ctx, | |||
| 587 | goto err; | 586 | goto err; |
| 588 | 587 | ||
| 589 | ab->ctx = ctx; | 588 | ab->ctx = ctx; |
| 589 | ab->gfp_mask = gfp_mask; | ||
| 590 | nlh = (struct nlmsghdr *)skb_put(ab->skb, NLMSG_SPACE(0)); | 590 | nlh = (struct nlmsghdr *)skb_put(ab->skb, NLMSG_SPACE(0)); |
| 591 | nlh->nlmsg_type = type; | 591 | nlh->nlmsg_type = type; |
| 592 | nlh->nlmsg_flags = 0; | 592 | nlh->nlmsg_flags = 0; |
| @@ -606,26 +606,27 @@ err: | |||
| 606 | * (timestamp,serial) tuple is unique for each syscall and is live from | 606 | * (timestamp,serial) tuple is unique for each syscall and is live from |
| 607 | * syscall entry to syscall exit. | 607 | * syscall entry to syscall exit. |
| 608 | * | 608 | * |
| 609 | * Atomic values are only guaranteed to be 24-bit, so we count down. | ||
| 610 | * | ||
| 611 | * NOTE: Another possibility is to store the formatted records off the | 609 | * NOTE: Another possibility is to store the formatted records off the |
| 612 | * audit context (for those records that have a context), and emit them | 610 | * audit context (for those records that have a context), and emit them |
| 613 | * all at syscall exit. However, this could delay the reporting of | 611 | * all at syscall exit. However, this could delay the reporting of |
| 614 | * significant errors until syscall exit (or never, if the system | 612 | * significant errors until syscall exit (or never, if the system |
| 615 | * halts). */ | 613 | * halts). */ |
| 614 | |||
| 616 | unsigned int audit_serial(void) | 615 | unsigned int audit_serial(void) |
| 617 | { | 616 | { |
| 618 | static atomic_t serial = ATOMIC_INIT(0xffffff); | 617 | static spinlock_t serial_lock = SPIN_LOCK_UNLOCKED; |
| 619 | unsigned int a, b; | 618 | static unsigned int serial = 0; |
| 619 | |||
| 620 | unsigned long flags; | ||
| 621 | unsigned int ret; | ||
| 620 | 622 | ||
| 623 | spin_lock_irqsave(&serial_lock, flags); | ||
| 621 | do { | 624 | do { |
| 622 | a = atomic_read(&serial); | 625 | ret = ++serial; |
| 623 | if (atomic_dec_and_test(&serial)) | 626 | } while (unlikely(!ret)); |
| 624 | atomic_set(&serial, 0xffffff); | 627 | spin_unlock_irqrestore(&serial_lock, flags); |
| 625 | b = atomic_read(&serial); | ||
| 626 | } while (b != a - 1); | ||
| 627 | 628 | ||
| 628 | return 0xffffff - b; | 629 | return ret; |
| 629 | } | 630 | } |
| 630 | 631 | ||
| 631 | static inline void audit_get_stamp(struct audit_context *ctx, | 632 | static inline void audit_get_stamp(struct audit_context *ctx, |
| @@ -645,17 +646,43 @@ static inline void audit_get_stamp(struct audit_context *ctx, | |||
| 645 | * syscall, then the syscall is marked as auditable and an audit record | 646 | * syscall, then the syscall is marked as auditable and an audit record |
| 646 | * will be written at syscall exit. If there is no associated task, tsk | 647 | * will be written at syscall exit. If there is no associated task, tsk |
| 647 | * should be NULL. */ | 648 | * should be NULL. */ |
| 648 | struct audit_buffer *audit_log_start(struct audit_context *ctx, int type) | 649 | |
| 650 | struct audit_buffer *audit_log_start(struct audit_context *ctx, int gfp_mask, | ||
| 651 | int type) | ||
| 649 | { | 652 | { |
| 650 | struct audit_buffer *ab = NULL; | 653 | struct audit_buffer *ab = NULL; |
| 651 | struct timespec t; | 654 | struct timespec t; |
| 652 | unsigned int serial; | 655 | unsigned int serial; |
| 656 | int reserve; | ||
| 657 | unsigned long timeout_start = jiffies; | ||
| 653 | 658 | ||
| 654 | if (!audit_initialized) | 659 | if (!audit_initialized) |
| 655 | return NULL; | 660 | return NULL; |
| 656 | 661 | ||
| 657 | if (audit_backlog_limit | 662 | if (gfp_mask & __GFP_WAIT) |
| 658 | && skb_queue_len(&audit_skb_queue) > audit_backlog_limit) { | 663 | reserve = 0; |
| 664 | else | ||
| 665 | reserve = 5; /* Allow atomic callers to go up to five | ||
| 666 | entries over the normal backlog limit */ | ||
| 667 | |||
| 668 | while (audit_backlog_limit | ||
| 669 | && skb_queue_len(&audit_skb_queue) > audit_backlog_limit + reserve) { | ||
| 670 | if (gfp_mask & __GFP_WAIT && audit_backlog_wait_time | ||
| 671 | && time_before(jiffies, timeout_start + audit_backlog_wait_time)) { | ||
| 672 | |||
| 673 | /* Wait for auditd to drain the queue a little */ | ||
| 674 | DECLARE_WAITQUEUE(wait, current); | ||
| 675 | set_current_state(TASK_INTERRUPTIBLE); | ||
| 676 | add_wait_queue(&audit_backlog_wait, &wait); | ||
| 677 | |||
| 678 | if (audit_backlog_limit && | ||
| 679 | skb_queue_len(&audit_skb_queue) > audit_backlog_limit) | ||
| 680 | schedule_timeout(timeout_start + audit_backlog_wait_time - jiffies); | ||
| 681 | |||
| 682 | __set_current_state(TASK_RUNNING); | ||
| 683 | remove_wait_queue(&audit_backlog_wait, &wait); | ||
| 684 | continue; | ||
| 685 | } | ||
| 659 | if (audit_rate_check()) | 686 | if (audit_rate_check()) |
| 660 | printk(KERN_WARNING | 687 | printk(KERN_WARNING |
| 661 | "audit: audit_backlog=%d > " | 688 | "audit: audit_backlog=%d > " |
| @@ -663,10 +690,12 @@ struct audit_buffer *audit_log_start(struct audit_context *ctx, int type) | |||
| 663 | skb_queue_len(&audit_skb_queue), | 690 | skb_queue_len(&audit_skb_queue), |
| 664 | audit_backlog_limit); | 691 | audit_backlog_limit); |
| 665 | audit_log_lost("backlog limit exceeded"); | 692 | audit_log_lost("backlog limit exceeded"); |
| 693 | audit_backlog_wait_time = audit_backlog_wait_overflow; | ||
| 694 | wake_up(&audit_backlog_wait); | ||
| 666 | return NULL; | 695 | return NULL; |
| 667 | } | 696 | } |
| 668 | 697 | ||
| 669 | ab = audit_buffer_alloc(ctx, GFP_ATOMIC, type); | 698 | ab = audit_buffer_alloc(ctx, gfp_mask, type); |
| 670 | if (!ab) { | 699 | if (!ab) { |
| 671 | audit_log_lost("out of memory in audit_log_start"); | 700 | audit_log_lost("out of memory in audit_log_start"); |
| 672 | return NULL; | 701 | return NULL; |
| @@ -690,7 +719,7 @@ static inline int audit_expand(struct audit_buffer *ab, int extra) | |||
| 690 | { | 719 | { |
| 691 | struct sk_buff *skb = ab->skb; | 720 | struct sk_buff *skb = ab->skb; |
| 692 | int ret = pskb_expand_head(skb, skb_headroom(skb), extra, | 721 | int ret = pskb_expand_head(skb, skb_headroom(skb), extra, |
| 693 | GFP_ATOMIC); | 722 | ab->gfp_mask); |
| 694 | if (ret < 0) { | 723 | if (ret < 0) { |
| 695 | audit_log_lost("out of memory in audit_expand"); | 724 | audit_log_lost("out of memory in audit_expand"); |
| 696 | return 0; | 725 | return 0; |
| @@ -809,7 +838,7 @@ void audit_log_d_path(struct audit_buffer *ab, const char *prefix, | |||
| 809 | audit_log_format(ab, " %s", prefix); | 838 | audit_log_format(ab, " %s", prefix); |
| 810 | 839 | ||
| 811 | /* We will allow 11 spaces for ' (deleted)' to be appended */ | 840 | /* We will allow 11 spaces for ' (deleted)' to be appended */ |
| 812 | path = kmalloc(PATH_MAX+11, GFP_KERNEL); | 841 | path = kmalloc(PATH_MAX+11, ab->gfp_mask); |
| 813 | if (!path) { | 842 | if (!path) { |
| 814 | audit_log_format(ab, "<no memory>"); | 843 | audit_log_format(ab, "<no memory>"); |
| 815 | return; | 844 | return; |
| @@ -841,7 +870,7 @@ void audit_log_end(struct audit_buffer *ab) | |||
| 841 | ab->skb = NULL; | 870 | ab->skb = NULL; |
| 842 | wake_up_interruptible(&kauditd_wait); | 871 | wake_up_interruptible(&kauditd_wait); |
| 843 | } else { | 872 | } else { |
| 844 | printk("%s\n", ab->skb->data + NLMSG_SPACE(0)); | 873 | printk(KERN_NOTICE "%s\n", ab->skb->data + NLMSG_SPACE(0)); |
| 845 | } | 874 | } |
| 846 | } | 875 | } |
| 847 | audit_buffer_free(ab); | 876 | audit_buffer_free(ab); |
| @@ -850,12 +879,13 @@ void audit_log_end(struct audit_buffer *ab) | |||
| 850 | /* Log an audit record. This is a convenience function that calls | 879 | /* Log an audit record. This is a convenience function that calls |
| 851 | * audit_log_start, audit_log_vformat, and audit_log_end. It may be | 880 | * audit_log_start, audit_log_vformat, and audit_log_end. It may be |
| 852 | * called in any context. */ | 881 | * called in any context. */ |
| 853 | void audit_log(struct audit_context *ctx, int type, const char *fmt, ...) | 882 | void audit_log(struct audit_context *ctx, int gfp_mask, int type, |
| 883 | const char *fmt, ...) | ||
| 854 | { | 884 | { |
| 855 | struct audit_buffer *ab; | 885 | struct audit_buffer *ab; |
| 856 | va_list args; | 886 | va_list args; |
| 857 | 887 | ||
| 858 | ab = audit_log_start(ctx, type); | 888 | ab = audit_log_start(ctx, gfp_mask, type); |
| 859 | if (ab) { | 889 | if (ab) { |
| 860 | va_start(args, fmt); | 890 | va_start(args, fmt); |
| 861 | audit_log_vformat(ab, fmt, args); | 891 | audit_log_vformat(ab, fmt, args); |
diff --git a/kernel/auditsc.c b/kernel/auditsc.c index e75f84e1a1a0..88696f639aab 100644 --- a/kernel/auditsc.c +++ b/kernel/auditsc.c | |||
| @@ -39,6 +39,9 @@ | |||
| 39 | #include <linux/audit.h> | 39 | #include <linux/audit.h> |
| 40 | #include <linux/personality.h> | 40 | #include <linux/personality.h> |
| 41 | #include <linux/time.h> | 41 | #include <linux/time.h> |
| 42 | #include <linux/kthread.h> | ||
| 43 | #include <linux/netlink.h> | ||
| 44 | #include <linux/compiler.h> | ||
| 42 | #include <asm/unistd.h> | 45 | #include <asm/unistd.h> |
| 43 | 46 | ||
| 44 | /* 0 = no checking | 47 | /* 0 = no checking |
| @@ -95,6 +98,7 @@ struct audit_names { | |||
| 95 | uid_t uid; | 98 | uid_t uid; |
| 96 | gid_t gid; | 99 | gid_t gid; |
| 97 | dev_t rdev; | 100 | dev_t rdev; |
| 101 | unsigned flags; | ||
| 98 | }; | 102 | }; |
| 99 | 103 | ||
| 100 | struct audit_aux_data { | 104 | struct audit_aux_data { |
| @@ -167,9 +171,16 @@ struct audit_context { | |||
| 167 | /* There are three lists of rules -- one to search at task creation | 171 | /* There are three lists of rules -- one to search at task creation |
| 168 | * time, one to search at syscall entry time, and another to search at | 172 | * time, one to search at syscall entry time, and another to search at |
| 169 | * syscall exit time. */ | 173 | * syscall exit time. */ |
| 170 | static LIST_HEAD(audit_tsklist); | 174 | static struct list_head audit_filter_list[AUDIT_NR_FILTERS] = { |
| 171 | static LIST_HEAD(audit_entlist); | 175 | LIST_HEAD_INIT(audit_filter_list[0]), |
| 172 | static LIST_HEAD(audit_extlist); | 176 | LIST_HEAD_INIT(audit_filter_list[1]), |
| 177 | LIST_HEAD_INIT(audit_filter_list[2]), | ||
| 178 | LIST_HEAD_INIT(audit_filter_list[3]), | ||
| 179 | LIST_HEAD_INIT(audit_filter_list[4]), | ||
| 180 | #if AUDIT_NR_FILTERS != 5 | ||
| 181 | #error Fix audit_filter_list initialiser | ||
| 182 | #endif | ||
| 183 | }; | ||
| 173 | 184 | ||
| 174 | struct audit_entry { | 185 | struct audit_entry { |
| 175 | struct list_head list; | 186 | struct list_head list; |
| @@ -179,9 +190,36 @@ struct audit_entry { | |||
| 179 | 190 | ||
| 180 | extern int audit_pid; | 191 | extern int audit_pid; |
| 181 | 192 | ||
| 193 | /* Copy rule from user-space to kernel-space. Called from | ||
| 194 | * audit_add_rule during AUDIT_ADD. */ | ||
| 195 | static inline int audit_copy_rule(struct audit_rule *d, struct audit_rule *s) | ||
| 196 | { | ||
| 197 | int i; | ||
| 198 | |||
| 199 | if (s->action != AUDIT_NEVER | ||
| 200 | && s->action != AUDIT_POSSIBLE | ||
| 201 | && s->action != AUDIT_ALWAYS) | ||
| 202 | return -1; | ||
| 203 | if (s->field_count < 0 || s->field_count > AUDIT_MAX_FIELDS) | ||
| 204 | return -1; | ||
| 205 | if ((s->flags & ~AUDIT_FILTER_PREPEND) >= AUDIT_NR_FILTERS) | ||
| 206 | return -1; | ||
| 207 | |||
| 208 | d->flags = s->flags; | ||
| 209 | d->action = s->action; | ||
| 210 | d->field_count = s->field_count; | ||
| 211 | for (i = 0; i < d->field_count; i++) { | ||
| 212 | d->fields[i] = s->fields[i]; | ||
| 213 | d->values[i] = s->values[i]; | ||
| 214 | } | ||
| 215 | for (i = 0; i < AUDIT_BITMASK_SIZE; i++) d->mask[i] = s->mask[i]; | ||
| 216 | return 0; | ||
| 217 | } | ||
| 218 | |||
| 182 | /* Check to see if two rules are identical. It is called from | 219 | /* Check to see if two rules are identical. It is called from |
| 220 | * audit_add_rule during AUDIT_ADD and | ||
| 183 | * audit_del_rule during AUDIT_DEL. */ | 221 | * audit_del_rule during AUDIT_DEL. */ |
| 184 | static int audit_compare_rule(struct audit_rule *a, struct audit_rule *b) | 222 | static inline int audit_compare_rule(struct audit_rule *a, struct audit_rule *b) |
| 185 | { | 223 | { |
| 186 | int i; | 224 | int i; |
| 187 | 225 | ||
| @@ -210,19 +248,37 @@ static int audit_compare_rule(struct audit_rule *a, struct audit_rule *b) | |||
| 210 | /* Note that audit_add_rule and audit_del_rule are called via | 248 | /* Note that audit_add_rule and audit_del_rule are called via |
| 211 | * audit_receive() in audit.c, and are protected by | 249 | * audit_receive() in audit.c, and are protected by |
| 212 | * audit_netlink_sem. */ | 250 | * audit_netlink_sem. */ |
| 213 | static inline int audit_add_rule(struct audit_entry *entry, | 251 | static inline int audit_add_rule(struct audit_rule *rule, |
| 214 | struct list_head *list) | 252 | struct list_head *list) |
| 215 | { | 253 | { |
| 216 | if (entry->rule.flags & AUDIT_PREPEND) { | 254 | struct audit_entry *entry; |
| 217 | entry->rule.flags &= ~AUDIT_PREPEND; | 255 | |
| 256 | /* Do not use the _rcu iterator here, since this is the only | ||
| 257 | * addition routine. */ | ||
| 258 | list_for_each_entry(entry, list, list) { | ||
| 259 | if (!audit_compare_rule(rule, &entry->rule)) { | ||
| 260 | return -EEXIST; | ||
| 261 | } | ||
| 262 | } | ||
| 263 | |||
| 264 | if (!(entry = kmalloc(sizeof(*entry), GFP_KERNEL))) | ||
| 265 | return -ENOMEM; | ||
| 266 | if (audit_copy_rule(&entry->rule, rule)) { | ||
| 267 | kfree(entry); | ||
| 268 | return -EINVAL; | ||
| 269 | } | ||
| 270 | |||
| 271 | if (entry->rule.flags & AUDIT_FILTER_PREPEND) { | ||
| 272 | entry->rule.flags &= ~AUDIT_FILTER_PREPEND; | ||
| 218 | list_add_rcu(&entry->list, list); | 273 | list_add_rcu(&entry->list, list); |
| 219 | } else { | 274 | } else { |
| 220 | list_add_tail_rcu(&entry->list, list); | 275 | list_add_tail_rcu(&entry->list, list); |
| 221 | } | 276 | } |
| 277 | |||
| 222 | return 0; | 278 | return 0; |
| 223 | } | 279 | } |
| 224 | 280 | ||
| 225 | static void audit_free_rule(struct rcu_head *head) | 281 | static inline void audit_free_rule(struct rcu_head *head) |
| 226 | { | 282 | { |
| 227 | struct audit_entry *e = container_of(head, struct audit_entry, rcu); | 283 | struct audit_entry *e = container_of(head, struct audit_entry, rcu); |
| 228 | kfree(e); | 284 | kfree(e); |
| @@ -245,82 +301,82 @@ static inline int audit_del_rule(struct audit_rule *rule, | |||
| 245 | return 0; | 301 | return 0; |
| 246 | } | 302 | } |
| 247 | } | 303 | } |
| 248 | return -EFAULT; /* No matching rule */ | 304 | return -ENOENT; /* No matching rule */ |
| 249 | } | 305 | } |
| 250 | 306 | ||
| 251 | /* Copy rule from user-space to kernel-space. Called during | 307 | static int audit_list_rules(void *_dest) |
| 252 | * AUDIT_ADD. */ | ||
| 253 | static int audit_copy_rule(struct audit_rule *d, struct audit_rule *s) | ||
| 254 | { | 308 | { |
| 309 | int pid, seq; | ||
| 310 | int *dest = _dest; | ||
| 311 | struct audit_entry *entry; | ||
| 255 | int i; | 312 | int i; |
| 256 | 313 | ||
| 257 | if (s->action != AUDIT_NEVER | 314 | pid = dest[0]; |
| 258 | && s->action != AUDIT_POSSIBLE | 315 | seq = dest[1]; |
| 259 | && s->action != AUDIT_ALWAYS) | 316 | kfree(dest); |
| 260 | return -1; | ||
| 261 | if (s->field_count < 0 || s->field_count > AUDIT_MAX_FIELDS) | ||
| 262 | return -1; | ||
| 263 | 317 | ||
| 264 | d->flags = s->flags; | 318 | down(&audit_netlink_sem); |
| 265 | d->action = s->action; | 319 | |
| 266 | d->field_count = s->field_count; | 320 | /* The *_rcu iterators not needed here because we are |
| 267 | for (i = 0; i < d->field_count; i++) { | 321 | always called with audit_netlink_sem held. */ |
| 268 | d->fields[i] = s->fields[i]; | 322 | for (i=0; i<AUDIT_NR_FILTERS; i++) { |
| 269 | d->values[i] = s->values[i]; | 323 | list_for_each_entry(entry, &audit_filter_list[i], list) |
| 324 | audit_send_reply(pid, seq, AUDIT_LIST, 0, 1, | ||
| 325 | &entry->rule, sizeof(entry->rule)); | ||
| 270 | } | 326 | } |
| 271 | for (i = 0; i < AUDIT_BITMASK_SIZE; i++) d->mask[i] = s->mask[i]; | 327 | audit_send_reply(pid, seq, AUDIT_LIST, 1, 1, NULL, 0); |
| 328 | |||
| 329 | up(&audit_netlink_sem); | ||
| 272 | return 0; | 330 | return 0; |
| 273 | } | 331 | } |
| 274 | 332 | ||
| 275 | int audit_receive_filter(int type, int pid, int uid, int seq, void *data, | 333 | int audit_receive_filter(int type, int pid, int uid, int seq, void *data, |
| 276 | uid_t loginuid) | 334 | uid_t loginuid) |
| 277 | { | 335 | { |
| 278 | u32 flags; | 336 | struct task_struct *tsk; |
| 279 | struct audit_entry *entry; | 337 | int *dest; |
| 280 | int err = 0; | 338 | int err = 0; |
| 339 | unsigned listnr; | ||
| 281 | 340 | ||
| 282 | switch (type) { | 341 | switch (type) { |
| 283 | case AUDIT_LIST: | 342 | case AUDIT_LIST: |
| 284 | /* The *_rcu iterators not needed here because we are | 343 | /* We can't just spew out the rules here because we might fill |
| 285 | always called with audit_netlink_sem held. */ | 344 | * the available socket buffer space and deadlock waiting for |
| 286 | list_for_each_entry(entry, &audit_tsklist, list) | 345 | * auditctl to read from it... which isn't ever going to |
| 287 | audit_send_reply(pid, seq, AUDIT_LIST, 0, 1, | 346 | * happen if we're actually running in the context of auditctl |
| 288 | &entry->rule, sizeof(entry->rule)); | 347 | * trying to _send_ the stuff */ |
| 289 | list_for_each_entry(entry, &audit_entlist, list) | 348 | |
| 290 | audit_send_reply(pid, seq, AUDIT_LIST, 0, 1, | 349 | dest = kmalloc(2 * sizeof(int), GFP_KERNEL); |
| 291 | &entry->rule, sizeof(entry->rule)); | 350 | if (!dest) |
| 292 | list_for_each_entry(entry, &audit_extlist, list) | 351 | return -ENOMEM; |
| 293 | audit_send_reply(pid, seq, AUDIT_LIST, 0, 1, | 352 | dest[0] = pid; |
| 294 | &entry->rule, sizeof(entry->rule)); | 353 | dest[1] = seq; |
| 295 | audit_send_reply(pid, seq, AUDIT_LIST, 1, 1, NULL, 0); | 354 | |
| 355 | tsk = kthread_run(audit_list_rules, dest, "audit_list_rules"); | ||
| 356 | if (IS_ERR(tsk)) { | ||
| 357 | kfree(dest); | ||
| 358 | err = PTR_ERR(tsk); | ||
| 359 | } | ||
| 296 | break; | 360 | break; |
| 297 | case AUDIT_ADD: | 361 | case AUDIT_ADD: |
| 298 | if (!(entry = kmalloc(sizeof(*entry), GFP_KERNEL))) | 362 | listnr =((struct audit_rule *)data)->flags & ~AUDIT_FILTER_PREPEND; |
| 299 | return -ENOMEM; | 363 | if (listnr >= AUDIT_NR_FILTERS) |
| 300 | if (audit_copy_rule(&entry->rule, data)) { | ||
| 301 | kfree(entry); | ||
| 302 | return -EINVAL; | 364 | return -EINVAL; |
| 303 | } | 365 | |
| 304 | flags = entry->rule.flags; | 366 | err = audit_add_rule(data, &audit_filter_list[listnr]); |
| 305 | if (!err && (flags & AUDIT_PER_TASK)) | 367 | if (!err) |
| 306 | err = audit_add_rule(entry, &audit_tsklist); | 368 | audit_log(NULL, GFP_KERNEL, AUDIT_CONFIG_CHANGE, |
| 307 | if (!err && (flags & AUDIT_AT_ENTRY)) | 369 | "auid=%u added an audit rule\n", loginuid); |
| 308 | err = audit_add_rule(entry, &audit_entlist); | ||
| 309 | if (!err && (flags & AUDIT_AT_EXIT)) | ||
| 310 | err = audit_add_rule(entry, &audit_extlist); | ||
| 311 | audit_log(NULL, AUDIT_CONFIG_CHANGE, | ||
| 312 | "auid=%u added an audit rule\n", loginuid); | ||
| 313 | break; | 370 | break; |
| 314 | case AUDIT_DEL: | 371 | case AUDIT_DEL: |
| 315 | flags =((struct audit_rule *)data)->flags; | 372 | listnr =((struct audit_rule *)data)->flags & ~AUDIT_FILTER_PREPEND; |
| 316 | if (!err && (flags & AUDIT_PER_TASK)) | 373 | if (listnr >= AUDIT_NR_FILTERS) |
| 317 | err = audit_del_rule(data, &audit_tsklist); | 374 | return -EINVAL; |
| 318 | if (!err && (flags & AUDIT_AT_ENTRY)) | 375 | |
| 319 | err = audit_del_rule(data, &audit_entlist); | 376 | err = audit_del_rule(data, &audit_filter_list[listnr]); |
| 320 | if (!err && (flags & AUDIT_AT_EXIT)) | 377 | if (!err) |
| 321 | err = audit_del_rule(data, &audit_extlist); | 378 | audit_log(NULL, GFP_KERNEL, AUDIT_CONFIG_CHANGE, |
| 322 | audit_log(NULL, AUDIT_CONFIG_CHANGE, | 379 | "auid=%u removed an audit rule\n", loginuid); |
| 323 | "auid=%u removed an audit rule\n", loginuid); | ||
| 324 | break; | 380 | break; |
| 325 | default: | 381 | default: |
| 326 | return -EINVAL; | 382 | return -EINVAL; |
| @@ -384,8 +440,12 @@ static int audit_filter_rules(struct task_struct *tsk, | |||
| 384 | result = (ctx->return_code == value); | 440 | result = (ctx->return_code == value); |
| 385 | break; | 441 | break; |
| 386 | case AUDIT_SUCCESS: | 442 | case AUDIT_SUCCESS: |
| 387 | if (ctx && ctx->return_valid) | 443 | if (ctx && ctx->return_valid) { |
| 388 | result = (ctx->return_valid == AUDITSC_SUCCESS); | 444 | if (value) |
| 445 | result = (ctx->return_valid == AUDITSC_SUCCESS); | ||
| 446 | else | ||
| 447 | result = (ctx->return_valid == AUDITSC_FAILURE); | ||
| 448 | } | ||
| 389 | break; | 449 | break; |
| 390 | case AUDIT_DEVMAJOR: | 450 | case AUDIT_DEVMAJOR: |
| 391 | if (ctx) { | 451 | if (ctx) { |
| @@ -454,7 +514,7 @@ static enum audit_state audit_filter_task(struct task_struct *tsk) | |||
| 454 | enum audit_state state; | 514 | enum audit_state state; |
| 455 | 515 | ||
| 456 | rcu_read_lock(); | 516 | rcu_read_lock(); |
| 457 | list_for_each_entry_rcu(e, &audit_tsklist, list) { | 517 | list_for_each_entry_rcu(e, &audit_filter_list[AUDIT_FILTER_TASK], list) { |
| 458 | if (audit_filter_rules(tsk, &e->rule, NULL, &state)) { | 518 | if (audit_filter_rules(tsk, &e->rule, NULL, &state)) { |
| 459 | rcu_read_unlock(); | 519 | rcu_read_unlock(); |
| 460 | return state; | 520 | return state; |
| @@ -474,20 +534,84 @@ static enum audit_state audit_filter_syscall(struct task_struct *tsk, | |||
| 474 | struct list_head *list) | 534 | struct list_head *list) |
| 475 | { | 535 | { |
| 476 | struct audit_entry *e; | 536 | struct audit_entry *e; |
| 537 | enum audit_state state; | ||
| 538 | |||
| 539 | if (audit_pid && tsk->tgid == audit_pid) | ||
| 540 | return AUDIT_DISABLED; | ||
| 541 | |||
| 542 | rcu_read_lock(); | ||
| 543 | if (!list_empty(list)) { | ||
| 544 | int word = AUDIT_WORD(ctx->major); | ||
| 545 | int bit = AUDIT_BIT(ctx->major); | ||
| 546 | |||
| 547 | list_for_each_entry_rcu(e, list, list) { | ||
| 548 | if ((e->rule.mask[word] & bit) == bit | ||
| 549 | && audit_filter_rules(tsk, &e->rule, ctx, &state)) { | ||
| 550 | rcu_read_unlock(); | ||
| 551 | return state; | ||
| 552 | } | ||
| 553 | } | ||
| 554 | } | ||
| 555 | rcu_read_unlock(); | ||
| 556 | return AUDIT_BUILD_CONTEXT; | ||
| 557 | } | ||
| 558 | |||
| 559 | static int audit_filter_user_rules(struct netlink_skb_parms *cb, | ||
| 560 | struct audit_rule *rule, | ||
| 561 | enum audit_state *state) | ||
| 562 | { | ||
| 563 | int i; | ||
| 564 | |||
| 565 | for (i = 0; i < rule->field_count; i++) { | ||
| 566 | u32 field = rule->fields[i] & ~AUDIT_NEGATE; | ||
| 567 | u32 value = rule->values[i]; | ||
| 568 | int result = 0; | ||
| 569 | |||
| 570 | switch (field) { | ||
| 571 | case AUDIT_PID: | ||
| 572 | result = (cb->creds.pid == value); | ||
| 573 | break; | ||
| 574 | case AUDIT_UID: | ||
| 575 | result = (cb->creds.uid == value); | ||
| 576 | break; | ||
| 577 | case AUDIT_GID: | ||
| 578 | result = (cb->creds.gid == value); | ||
| 579 | break; | ||
| 580 | case AUDIT_LOGINUID: | ||
| 581 | result = (cb->loginuid == value); | ||
| 582 | break; | ||
| 583 | } | ||
| 584 | |||
| 585 | if (rule->fields[i] & AUDIT_NEGATE) | ||
| 586 | result = !result; | ||
| 587 | if (!result) | ||
| 588 | return 0; | ||
| 589 | } | ||
| 590 | switch (rule->action) { | ||
| 591 | case AUDIT_NEVER: *state = AUDIT_DISABLED; break; | ||
| 592 | case AUDIT_POSSIBLE: *state = AUDIT_BUILD_CONTEXT; break; | ||
| 593 | case AUDIT_ALWAYS: *state = AUDIT_RECORD_CONTEXT; break; | ||
| 594 | } | ||
| 595 | return 1; | ||
| 596 | } | ||
| 597 | |||
| 598 | int audit_filter_user(struct netlink_skb_parms *cb, int type) | ||
| 599 | { | ||
| 600 | struct audit_entry *e; | ||
| 477 | enum audit_state state; | 601 | enum audit_state state; |
| 478 | int word = AUDIT_WORD(ctx->major); | 602 | int ret = 1; |
| 479 | int bit = AUDIT_BIT(ctx->major); | ||
| 480 | 603 | ||
| 481 | rcu_read_lock(); | 604 | rcu_read_lock(); |
| 482 | list_for_each_entry_rcu(e, list, list) { | 605 | list_for_each_entry_rcu(e, &audit_filter_list[AUDIT_FILTER_USER], list) { |
| 483 | if ((e->rule.mask[word] & bit) == bit | 606 | if (audit_filter_user_rules(cb, &e->rule, &state)) { |
| 484 | && audit_filter_rules(tsk, &e->rule, ctx, &state)) { | 607 | if (state == AUDIT_DISABLED) |
| 485 | rcu_read_unlock(); | 608 | ret = 0; |
| 486 | return state; | 609 | break; |
| 487 | } | 610 | } |
| 488 | } | 611 | } |
| 489 | rcu_read_unlock(); | 612 | rcu_read_unlock(); |
| 490 | return AUDIT_BUILD_CONTEXT; | 613 | |
| 614 | return ret; /* Audit by default */ | ||
| 491 | } | 615 | } |
| 492 | 616 | ||
| 493 | /* This should be called with task_lock() held. */ | 617 | /* This should be called with task_lock() held. */ |
| @@ -504,7 +628,7 @@ static inline struct audit_context *audit_get_context(struct task_struct *tsk, | |||
| 504 | 628 | ||
| 505 | if (context->in_syscall && !context->auditable) { | 629 | if (context->in_syscall && !context->auditable) { |
| 506 | enum audit_state state; | 630 | enum audit_state state; |
| 507 | state = audit_filter_syscall(tsk, context, &audit_extlist); | 631 | state = audit_filter_syscall(tsk, context, &audit_filter_list[AUDIT_FILTER_EXIT]); |
| 508 | if (state == AUDIT_RECORD_CONTEXT) | 632 | if (state == AUDIT_RECORD_CONTEXT) |
| 509 | context->auditable = 1; | 633 | context->auditable = 1; |
| 510 | } | 634 | } |
| @@ -679,13 +803,13 @@ static void audit_log_task_info(struct audit_buffer *ab) | |||
| 679 | up_read(&mm->mmap_sem); | 803 | up_read(&mm->mmap_sem); |
| 680 | } | 804 | } |
| 681 | 805 | ||
| 682 | static void audit_log_exit(struct audit_context *context) | 806 | static void audit_log_exit(struct audit_context *context, unsigned int gfp_mask) |
| 683 | { | 807 | { |
| 684 | int i; | 808 | int i; |
| 685 | struct audit_buffer *ab; | 809 | struct audit_buffer *ab; |
| 686 | struct audit_aux_data *aux; | 810 | struct audit_aux_data *aux; |
| 687 | 811 | ||
| 688 | ab = audit_log_start(context, AUDIT_SYSCALL); | 812 | ab = audit_log_start(context, gfp_mask, AUDIT_SYSCALL); |
| 689 | if (!ab) | 813 | if (!ab) |
| 690 | return; /* audit_panic has been called */ | 814 | return; /* audit_panic has been called */ |
| 691 | audit_log_format(ab, "arch=%x syscall=%d", | 815 | audit_log_format(ab, "arch=%x syscall=%d", |
| @@ -717,7 +841,7 @@ static void audit_log_exit(struct audit_context *context) | |||
| 717 | 841 | ||
| 718 | for (aux = context->aux; aux; aux = aux->next) { | 842 | for (aux = context->aux; aux; aux = aux->next) { |
| 719 | 843 | ||
| 720 | ab = audit_log_start(context, aux->type); | 844 | ab = audit_log_start(context, GFP_KERNEL, aux->type); |
| 721 | if (!ab) | 845 | if (!ab) |
| 722 | continue; /* audit_panic has been called */ | 846 | continue; /* audit_panic has been called */ |
| 723 | 847 | ||
| @@ -754,14 +878,14 @@ static void audit_log_exit(struct audit_context *context) | |||
| 754 | } | 878 | } |
| 755 | 879 | ||
| 756 | if (context->pwd && context->pwdmnt) { | 880 | if (context->pwd && context->pwdmnt) { |
| 757 | ab = audit_log_start(context, AUDIT_CWD); | 881 | ab = audit_log_start(context, GFP_KERNEL, AUDIT_CWD); |
| 758 | if (ab) { | 882 | if (ab) { |
| 759 | audit_log_d_path(ab, "cwd=", context->pwd, context->pwdmnt); | 883 | audit_log_d_path(ab, "cwd=", context->pwd, context->pwdmnt); |
| 760 | audit_log_end(ab); | 884 | audit_log_end(ab); |
| 761 | } | 885 | } |
| 762 | } | 886 | } |
| 763 | for (i = 0; i < context->name_count; i++) { | 887 | for (i = 0; i < context->name_count; i++) { |
| 764 | ab = audit_log_start(context, AUDIT_PATH); | 888 | ab = audit_log_start(context, GFP_KERNEL, AUDIT_PATH); |
| 765 | if (!ab) | 889 | if (!ab) |
| 766 | continue; /* audit_panic has been called */ | 890 | continue; /* audit_panic has been called */ |
| 767 | 891 | ||
| @@ -770,6 +894,8 @@ static void audit_log_exit(struct audit_context *context) | |||
| 770 | audit_log_format(ab, " name="); | 894 | audit_log_format(ab, " name="); |
| 771 | audit_log_untrustedstring(ab, context->names[i].name); | 895 | audit_log_untrustedstring(ab, context->names[i].name); |
| 772 | } | 896 | } |
| 897 | audit_log_format(ab, " flags=%x\n", context->names[i].flags); | ||
| 898 | |||
| 773 | if (context->names[i].ino != (unsigned long)-1) | 899 | if (context->names[i].ino != (unsigned long)-1) |
| 774 | audit_log_format(ab, " inode=%lu dev=%02x:%02x mode=%#o" | 900 | audit_log_format(ab, " inode=%lu dev=%02x:%02x mode=%#o" |
| 775 | " ouid=%u ogid=%u rdev=%02x:%02x", | 901 | " ouid=%u ogid=%u rdev=%02x:%02x", |
| @@ -799,9 +925,11 @@ void audit_free(struct task_struct *tsk) | |||
| 799 | return; | 925 | return; |
| 800 | 926 | ||
| 801 | /* Check for system calls that do not go through the exit | 927 | /* Check for system calls that do not go through the exit |
| 802 | * function (e.g., exit_group), then free context block. */ | 928 | * function (e.g., exit_group), then free context block. |
| 803 | if (context->in_syscall && context->auditable && context->pid != audit_pid) | 929 | * We use GFP_ATOMIC here because we might be doing this |
| 804 | audit_log_exit(context); | 930 | * in the context of the idle thread */ |
| 931 | if (context->in_syscall && context->auditable) | ||
| 932 | audit_log_exit(context, GFP_ATOMIC); | ||
| 805 | 933 | ||
| 806 | audit_free_context(context); | 934 | audit_free_context(context); |
| 807 | } | 935 | } |
| @@ -876,11 +1004,11 @@ void audit_syscall_entry(struct task_struct *tsk, int arch, int major, | |||
| 876 | 1004 | ||
| 877 | state = context->state; | 1005 | state = context->state; |
| 878 | if (state == AUDIT_SETUP_CONTEXT || state == AUDIT_BUILD_CONTEXT) | 1006 | if (state == AUDIT_SETUP_CONTEXT || state == AUDIT_BUILD_CONTEXT) |
| 879 | state = audit_filter_syscall(tsk, context, &audit_entlist); | 1007 | state = audit_filter_syscall(tsk, context, &audit_filter_list[AUDIT_FILTER_ENTRY]); |
| 880 | if (likely(state == AUDIT_DISABLED)) | 1008 | if (likely(state == AUDIT_DISABLED)) |
| 881 | return; | 1009 | return; |
| 882 | 1010 | ||
| 883 | context->serial = audit_serial(); | 1011 | context->serial = 0; |
| 884 | context->ctime = CURRENT_TIME; | 1012 | context->ctime = CURRENT_TIME; |
| 885 | context->in_syscall = 1; | 1013 | context->in_syscall = 1; |
| 886 | context->auditable = !!(state == AUDIT_RECORD_CONTEXT); | 1014 | context->auditable = !!(state == AUDIT_RECORD_CONTEXT); |
| @@ -903,10 +1031,10 @@ void audit_syscall_exit(struct task_struct *tsk, int valid, long return_code) | |||
| 903 | /* Not having a context here is ok, since the parent may have | 1031 | /* Not having a context here is ok, since the parent may have |
| 904 | * called __put_task_struct. */ | 1032 | * called __put_task_struct. */ |
| 905 | if (likely(!context)) | 1033 | if (likely(!context)) |
| 906 | return; | 1034 | goto out; |
| 907 | 1035 | ||
| 908 | if (context->in_syscall && context->auditable && context->pid != audit_pid) | 1036 | if (context->in_syscall && context->auditable) |
| 909 | audit_log_exit(context); | 1037 | audit_log_exit(context, GFP_KERNEL); |
| 910 | 1038 | ||
| 911 | context->in_syscall = 0; | 1039 | context->in_syscall = 0; |
| 912 | context->auditable = 0; | 1040 | context->auditable = 0; |
| @@ -919,9 +1047,9 @@ void audit_syscall_exit(struct task_struct *tsk, int valid, long return_code) | |||
| 919 | } else { | 1047 | } else { |
| 920 | audit_free_names(context); | 1048 | audit_free_names(context); |
| 921 | audit_free_aux(context); | 1049 | audit_free_aux(context); |
| 922 | audit_zero_context(context, context->state); | ||
| 923 | tsk->audit_context = context; | 1050 | tsk->audit_context = context; |
| 924 | } | 1051 | } |
| 1052 | out: | ||
| 925 | put_task_struct(tsk); | 1053 | put_task_struct(tsk); |
| 926 | } | 1054 | } |
| 927 | 1055 | ||
| @@ -996,7 +1124,7 @@ void audit_putname(const char *name) | |||
| 996 | 1124 | ||
| 997 | /* Store the inode and device from a lookup. Called from | 1125 | /* Store the inode and device from a lookup. Called from |
| 998 | * fs/namei.c:path_lookup(). */ | 1126 | * fs/namei.c:path_lookup(). */ |
| 999 | void audit_inode(const char *name, const struct inode *inode) | 1127 | void audit_inode(const char *name, const struct inode *inode, unsigned flags) |
| 1000 | { | 1128 | { |
| 1001 | int idx; | 1129 | int idx; |
| 1002 | struct audit_context *context = current->audit_context; | 1130 | struct audit_context *context = current->audit_context; |
| @@ -1022,17 +1150,20 @@ void audit_inode(const char *name, const struct inode *inode) | |||
| 1022 | ++context->ino_count; | 1150 | ++context->ino_count; |
| 1023 | #endif | 1151 | #endif |
| 1024 | } | 1152 | } |
| 1025 | context->names[idx].ino = inode->i_ino; | 1153 | context->names[idx].flags = flags; |
| 1026 | context->names[idx].dev = inode->i_sb->s_dev; | 1154 | context->names[idx].ino = inode->i_ino; |
| 1027 | context->names[idx].mode = inode->i_mode; | 1155 | context->names[idx].dev = inode->i_sb->s_dev; |
| 1028 | context->names[idx].uid = inode->i_uid; | 1156 | context->names[idx].mode = inode->i_mode; |
| 1029 | context->names[idx].gid = inode->i_gid; | 1157 | context->names[idx].uid = inode->i_uid; |
| 1030 | context->names[idx].rdev = inode->i_rdev; | 1158 | context->names[idx].gid = inode->i_gid; |
| 1159 | context->names[idx].rdev = inode->i_rdev; | ||
| 1031 | } | 1160 | } |
| 1032 | 1161 | ||
| 1033 | void auditsc_get_stamp(struct audit_context *ctx, | 1162 | void auditsc_get_stamp(struct audit_context *ctx, |
| 1034 | struct timespec *t, unsigned int *serial) | 1163 | struct timespec *t, unsigned int *serial) |
| 1035 | { | 1164 | { |
| 1165 | if (!ctx->serial) | ||
| 1166 | ctx->serial = audit_serial(); | ||
| 1036 | t->tv_sec = ctx->ctime.tv_sec; | 1167 | t->tv_sec = ctx->ctime.tv_sec; |
| 1037 | t->tv_nsec = ctx->ctime.tv_nsec; | 1168 | t->tv_nsec = ctx->ctime.tv_nsec; |
| 1038 | *serial = ctx->serial; | 1169 | *serial = ctx->serial; |
| @@ -1044,7 +1175,7 @@ int audit_set_loginuid(struct task_struct *task, uid_t loginuid) | |||
| 1044 | if (task->audit_context) { | 1175 | if (task->audit_context) { |
| 1045 | struct audit_buffer *ab; | 1176 | struct audit_buffer *ab; |
| 1046 | 1177 | ||
| 1047 | ab = audit_log_start(NULL, AUDIT_LOGIN); | 1178 | ab = audit_log_start(NULL, GFP_KERNEL, AUDIT_LOGIN); |
| 1048 | if (ab) { | 1179 | if (ab) { |
| 1049 | audit_log_format(ab, "login pid=%d uid=%u " | 1180 | audit_log_format(ab, "login pid=%d uid=%u " |
| 1050 | "old auid=%u new auid=%u", | 1181 | "old auid=%u new auid=%u", |
| @@ -1153,7 +1284,7 @@ void audit_signal_info(int sig, struct task_struct *t) | |||
| 1153 | extern pid_t audit_sig_pid; | 1284 | extern pid_t audit_sig_pid; |
| 1154 | extern uid_t audit_sig_uid; | 1285 | extern uid_t audit_sig_uid; |
| 1155 | 1286 | ||
| 1156 | if (unlikely(audit_pid && t->pid == audit_pid)) { | 1287 | if (unlikely(audit_pid && t->tgid == audit_pid)) { |
| 1157 | if (sig == SIGTERM || sig == SIGHUP) { | 1288 | if (sig == SIGTERM || sig == SIGHUP) { |
| 1158 | struct audit_context *ctx = current->audit_context; | 1289 | struct audit_context *ctx = current->audit_context; |
| 1159 | audit_sig_pid = current->pid; | 1290 | audit_sig_pid = current->pid; |
diff --git a/kernel/cpuset.c b/kernel/cpuset.c index 79866bc6b3a1..45a5719a0104 100644 --- a/kernel/cpuset.c +++ b/kernel/cpuset.c | |||
| @@ -968,8 +968,6 @@ static ssize_t cpuset_common_file_read(struct file *file, char __user *buf, | |||
| 968 | char *page; | 968 | char *page; |
| 969 | ssize_t retval = 0; | 969 | ssize_t retval = 0; |
| 970 | char *s; | 970 | char *s; |
| 971 | char *start; | ||
| 972 | size_t n; | ||
| 973 | 971 | ||
| 974 | if (!(page = (char *)__get_free_page(GFP_KERNEL))) | 972 | if (!(page = (char *)__get_free_page(GFP_KERNEL))) |
| 975 | return -ENOMEM; | 973 | return -ENOMEM; |
| @@ -999,14 +997,7 @@ static ssize_t cpuset_common_file_read(struct file *file, char __user *buf, | |||
| 999 | *s++ = '\n'; | 997 | *s++ = '\n'; |
| 1000 | *s = '\0'; | 998 | *s = '\0'; |
| 1001 | 999 | ||
| 1002 | /* Do nothing if *ppos is at the eof or beyond the eof. */ | 1000 | retval = simple_read_from_buffer(buf, nbytes, ppos, page, s - page); |
| 1003 | if (s - page <= *ppos) | ||
| 1004 | return 0; | ||
| 1005 | |||
| 1006 | start = page + *ppos; | ||
| 1007 | n = s - start; | ||
| 1008 | retval = n - copy_to_user(buf, start, min(n, nbytes)); | ||
| 1009 | *ppos += retval; | ||
| 1010 | out: | 1001 | out: |
| 1011 | free_page((unsigned long)page); | 1002 | free_page((unsigned long)page); |
| 1012 | return retval; | 1003 | return retval; |
diff --git a/kernel/exit.c b/kernel/exit.c index 6d2089a1bce7..43077732619b 100644 --- a/kernel/exit.c +++ b/kernel/exit.c | |||
| @@ -371,6 +371,12 @@ static inline void close_files(struct files_struct * files) | |||
| 371 | struct fdtable *fdt; | 371 | struct fdtable *fdt; |
| 372 | 372 | ||
| 373 | j = 0; | 373 | j = 0; |
| 374 | |||
| 375 | /* | ||
| 376 | * It is safe to dereference the fd table without RCU or | ||
| 377 | * ->file_lock because this is the last reference to the | ||
| 378 | * files structure. | ||
| 379 | */ | ||
| 374 | fdt = files_fdtable(files); | 380 | fdt = files_fdtable(files); |
| 375 | for (;;) { | 381 | for (;;) { |
| 376 | unsigned long set; | 382 | unsigned long set; |
| @@ -1197,7 +1203,7 @@ static int wait_task_stopped(task_t *p, int delayed_group_leader, int noreap, | |||
| 1197 | 1203 | ||
| 1198 | exit_code = p->exit_code; | 1204 | exit_code = p->exit_code; |
| 1199 | if (unlikely(!exit_code) || | 1205 | if (unlikely(!exit_code) || |
| 1200 | unlikely(p->state > TASK_STOPPED)) | 1206 | unlikely(p->state & TASK_TRACED)) |
| 1201 | goto bail_ref; | 1207 | goto bail_ref; |
| 1202 | return wait_noreap_copyout(p, pid, uid, | 1208 | return wait_noreap_copyout(p, pid, uid, |
| 1203 | why, (exit_code << 8) | 0x7f, | 1209 | why, (exit_code << 8) | 0x7f, |
diff --git a/kernel/fork.c b/kernel/fork.c index 8149f3602881..533ce27f4b2c 100644 --- a/kernel/fork.c +++ b/kernel/fork.c | |||
| @@ -1062,7 +1062,8 @@ static task_t *copy_process(unsigned long clone_flags, | |||
| 1062 | * parent's CPU). This avoids alot of nasty races. | 1062 | * parent's CPU). This avoids alot of nasty races. |
| 1063 | */ | 1063 | */ |
| 1064 | p->cpus_allowed = current->cpus_allowed; | 1064 | p->cpus_allowed = current->cpus_allowed; |
| 1065 | if (unlikely(!cpu_isset(task_cpu(p), p->cpus_allowed))) | 1065 | if (unlikely(!cpu_isset(task_cpu(p), p->cpus_allowed) || |
| 1066 | !cpu_online(task_cpu(p)))) | ||
| 1066 | set_task_cpu(p, smp_processor_id()); | 1067 | set_task_cpu(p, smp_processor_id()); |
| 1067 | 1068 | ||
| 1068 | /* | 1069 | /* |
diff --git a/kernel/module.c b/kernel/module.c index 4b39d3793c72..ff5c500ab625 100644 --- a/kernel/module.c +++ b/kernel/module.c | |||
| @@ -20,6 +20,7 @@ | |||
| 20 | #include <linux/module.h> | 20 | #include <linux/module.h> |
| 21 | #include <linux/moduleloader.h> | 21 | #include <linux/moduleloader.h> |
| 22 | #include <linux/init.h> | 22 | #include <linux/init.h> |
| 23 | #include <linux/kernel.h> | ||
| 23 | #include <linux/slab.h> | 24 | #include <linux/slab.h> |
| 24 | #include <linux/vmalloc.h> | 25 | #include <linux/vmalloc.h> |
| 25 | #include <linux/elf.h> | 26 | #include <linux/elf.h> |
| @@ -498,7 +499,7 @@ static inline int try_force(unsigned int flags) | |||
| 498 | { | 499 | { |
| 499 | int ret = (flags & O_TRUNC); | 500 | int ret = (flags & O_TRUNC); |
| 500 | if (ret) | 501 | if (ret) |
| 501 | tainted |= TAINT_FORCED_MODULE; | 502 | add_taint(TAINT_FORCED_MODULE); |
| 502 | return ret; | 503 | return ret; |
| 503 | } | 504 | } |
| 504 | #else | 505 | #else |
| @@ -897,7 +898,7 @@ static int check_version(Elf_Shdr *sechdrs, | |||
| 897 | if (!(tainted & TAINT_FORCED_MODULE)) { | 898 | if (!(tainted & TAINT_FORCED_MODULE)) { |
| 898 | printk("%s: no version for \"%s\" found: kernel tainted.\n", | 899 | printk("%s: no version for \"%s\" found: kernel tainted.\n", |
| 899 | mod->name, symname); | 900 | mod->name, symname); |
| 900 | tainted |= TAINT_FORCED_MODULE; | 901 | add_taint(TAINT_FORCED_MODULE); |
| 901 | } | 902 | } |
| 902 | return 1; | 903 | return 1; |
| 903 | } | 904 | } |
| @@ -1352,7 +1353,7 @@ static void set_license(struct module *mod, const char *license) | |||
| 1352 | if (!mod->license_gplok && !(tainted & TAINT_PROPRIETARY_MODULE)) { | 1353 | if (!mod->license_gplok && !(tainted & TAINT_PROPRIETARY_MODULE)) { |
| 1353 | printk(KERN_WARNING "%s: module license '%s' taints kernel.\n", | 1354 | printk(KERN_WARNING "%s: module license '%s' taints kernel.\n", |
| 1354 | mod->name, license); | 1355 | mod->name, license); |
| 1355 | tainted |= TAINT_PROPRIETARY_MODULE; | 1356 | add_taint(TAINT_PROPRIETARY_MODULE); |
| 1356 | } | 1357 | } |
| 1357 | } | 1358 | } |
| 1358 | 1359 | ||
| @@ -1610,7 +1611,7 @@ static struct module *load_module(void __user *umod, | |||
| 1610 | modmagic = get_modinfo(sechdrs, infoindex, "vermagic"); | 1611 | modmagic = get_modinfo(sechdrs, infoindex, "vermagic"); |
| 1611 | /* This is allowed: modprobe --force will invalidate it. */ | 1612 | /* This is allowed: modprobe --force will invalidate it. */ |
| 1612 | if (!modmagic) { | 1613 | if (!modmagic) { |
| 1613 | tainted |= TAINT_FORCED_MODULE; | 1614 | add_taint(TAINT_FORCED_MODULE); |
| 1614 | printk(KERN_WARNING "%s: no version magic, tainting kernel.\n", | 1615 | printk(KERN_WARNING "%s: no version magic, tainting kernel.\n", |
| 1615 | mod->name); | 1616 | mod->name); |
| 1616 | } else if (!same_magic(modmagic, vermagic)) { | 1617 | } else if (!same_magic(modmagic, vermagic)) { |
| @@ -1739,7 +1740,7 @@ static struct module *load_module(void __user *umod, | |||
| 1739 | (mod->num_gpl_syms && !gplcrcindex)) { | 1740 | (mod->num_gpl_syms && !gplcrcindex)) { |
| 1740 | printk(KERN_WARNING "%s: No versions for exported symbols." | 1741 | printk(KERN_WARNING "%s: No versions for exported symbols." |
| 1741 | " Tainting kernel.\n", mod->name); | 1742 | " Tainting kernel.\n", mod->name); |
| 1742 | tainted |= TAINT_FORCED_MODULE; | 1743 | add_taint(TAINT_FORCED_MODULE); |
| 1743 | } | 1744 | } |
| 1744 | #endif | 1745 | #endif |
| 1745 | 1746 | ||
diff --git a/kernel/params.c b/kernel/params.c index fbf173215fd2..1a8614bac5d5 100644 --- a/kernel/params.c +++ b/kernel/params.c | |||
| @@ -80,8 +80,6 @@ static char *next_arg(char *args, char **param, char **val) | |||
| 80 | int in_quote = 0, quoted = 0; | 80 | int in_quote = 0, quoted = 0; |
| 81 | char *next; | 81 | char *next; |
| 82 | 82 | ||
| 83 | /* Chew any extra spaces */ | ||
| 84 | while (*args == ' ') args++; | ||
| 85 | if (*args == '"') { | 83 | if (*args == '"') { |
| 86 | args++; | 84 | args++; |
| 87 | in_quote = 1; | 85 | in_quote = 1; |
| @@ -121,6 +119,10 @@ static char *next_arg(char *args, char **param, char **val) | |||
| 121 | next = args + i + 1; | 119 | next = args + i + 1; |
| 122 | } else | 120 | } else |
| 123 | next = args + i; | 121 | next = args + i; |
| 122 | |||
| 123 | /* Chew up trailing spaces. */ | ||
| 124 | while (*next == ' ') | ||
| 125 | next++; | ||
| 124 | return next; | 126 | return next; |
| 125 | } | 127 | } |
| 126 | 128 | ||
| @@ -135,6 +137,10 @@ int parse_args(const char *name, | |||
| 135 | 137 | ||
| 136 | DEBUGP("Parsing ARGS: %s\n", args); | 138 | DEBUGP("Parsing ARGS: %s\n", args); |
| 137 | 139 | ||
| 140 | /* Chew leading spaces */ | ||
| 141 | while (*args == ' ') | ||
| 142 | args++; | ||
| 143 | |||
| 138 | while (*args) { | 144 | while (*args) { |
| 139 | int ret; | 145 | int ret; |
| 140 | 146 | ||
diff --git a/kernel/power/Kconfig b/kernel/power/Kconfig index 396c7873e804..46a5e5acff97 100644 --- a/kernel/power/Kconfig +++ b/kernel/power/Kconfig | |||
| @@ -29,7 +29,7 @@ config PM_DEBUG | |||
| 29 | 29 | ||
| 30 | config SOFTWARE_SUSPEND | 30 | config SOFTWARE_SUSPEND |
| 31 | bool "Software Suspend" | 31 | bool "Software Suspend" |
| 32 | depends on PM && SWAP && (X86 || ((FVR || PPC32) && !SMP)) | 32 | depends on PM && SWAP && (X86 && (!SMP || SUSPEND_SMP)) || ((FVR || PPC32) && !SMP) |
| 33 | ---help--- | 33 | ---help--- |
| 34 | Enable the possibility of suspending the machine. | 34 | Enable the possibility of suspending the machine. |
| 35 | It doesn't need APM. | 35 | It doesn't need APM. |
diff --git a/kernel/power/disk.c b/kernel/power/disk.c index 2d8bf054d036..761956e813f5 100644 --- a/kernel/power/disk.c +++ b/kernel/power/disk.c | |||
| @@ -17,12 +17,12 @@ | |||
| 17 | #include <linux/delay.h> | 17 | #include <linux/delay.h> |
| 18 | #include <linux/fs.h> | 18 | #include <linux/fs.h> |
| 19 | #include <linux/mount.h> | 19 | #include <linux/mount.h> |
| 20 | #include <linux/pm.h> | ||
| 20 | 21 | ||
| 21 | #include "power.h" | 22 | #include "power.h" |
| 22 | 23 | ||
| 23 | 24 | ||
| 24 | extern suspend_disk_method_t pm_disk_mode; | 25 | extern suspend_disk_method_t pm_disk_mode; |
| 25 | extern struct pm_ops * pm_ops; | ||
| 26 | 26 | ||
| 27 | extern int swsusp_suspend(void); | 27 | extern int swsusp_suspend(void); |
| 28 | extern int swsusp_write(void); | 28 | extern int swsusp_write(void); |
| @@ -49,13 +49,11 @@ dev_t swsusp_resume_device; | |||
| 49 | 49 | ||
| 50 | static void power_down(suspend_disk_method_t mode) | 50 | static void power_down(suspend_disk_method_t mode) |
| 51 | { | 51 | { |
| 52 | unsigned long flags; | ||
| 53 | int error = 0; | 52 | int error = 0; |
| 54 | 53 | ||
| 55 | local_irq_save(flags); | ||
| 56 | switch(mode) { | 54 | switch(mode) { |
| 57 | case PM_DISK_PLATFORM: | 55 | case PM_DISK_PLATFORM: |
| 58 | device_shutdown(); | 56 | kernel_power_off_prepare(); |
| 59 | error = pm_ops->enter(PM_SUSPEND_DISK); | 57 | error = pm_ops->enter(PM_SUSPEND_DISK); |
| 60 | break; | 58 | break; |
| 61 | case PM_DISK_SHUTDOWN: | 59 | case PM_DISK_SHUTDOWN: |
diff --git a/kernel/power/power.h b/kernel/power/power.h index cd6a3493cc0d..6748de23e83c 100644 --- a/kernel/power/power.h +++ b/kernel/power/power.h | |||
| @@ -1,7 +1,7 @@ | |||
| 1 | #include <linux/suspend.h> | 1 | #include <linux/suspend.h> |
| 2 | #include <linux/utsname.h> | 2 | #include <linux/utsname.h> |
| 3 | 3 | ||
| 4 | /* With SUSPEND_CONSOLE defined, it suspend looks *really* cool, but | 4 | /* With SUSPEND_CONSOLE defined suspend looks *really* cool, but |
| 5 | we probably do not take enough locks for switching consoles, etc, | 5 | we probably do not take enough locks for switching consoles, etc, |
| 6 | so bad things might happen. | 6 | so bad things might happen. |
| 7 | */ | 7 | */ |
| @@ -9,6 +9,9 @@ | |||
| 9 | #define SUSPEND_CONSOLE (MAX_NR_CONSOLES-1) | 9 | #define SUSPEND_CONSOLE (MAX_NR_CONSOLES-1) |
| 10 | #endif | 10 | #endif |
| 11 | 11 | ||
| 12 | #define MAX_PBES ((PAGE_SIZE - sizeof(struct new_utsname) \ | ||
| 13 | - 4 - 3*sizeof(unsigned long) - sizeof(int) \ | ||
| 14 | - sizeof(void *)) / sizeof(swp_entry_t)) | ||
| 12 | 15 | ||
| 13 | struct swsusp_info { | 16 | struct swsusp_info { |
| 14 | struct new_utsname uts; | 17 | struct new_utsname uts; |
| @@ -18,7 +21,7 @@ struct swsusp_info { | |||
| 18 | unsigned long image_pages; | 21 | unsigned long image_pages; |
| 19 | unsigned long pagedir_pages; | 22 | unsigned long pagedir_pages; |
| 20 | suspend_pagedir_t * suspend_pagedir; | 23 | suspend_pagedir_t * suspend_pagedir; |
| 21 | swp_entry_t pagedir[768]; | 24 | swp_entry_t pagedir[MAX_PBES]; |
| 22 | } __attribute__((aligned(PAGE_SIZE))); | 25 | } __attribute__((aligned(PAGE_SIZE))); |
| 23 | 26 | ||
| 24 | 27 | ||
diff --git a/kernel/power/swsusp.c b/kernel/power/swsusp.c index d967e875ee82..acf79ac1cb6d 100644 --- a/kernel/power/swsusp.c +++ b/kernel/power/swsusp.c | |||
| @@ -363,7 +363,7 @@ static void lock_swapdevices(void) | |||
| 363 | } | 363 | } |
| 364 | 364 | ||
| 365 | /** | 365 | /** |
| 366 | * write_swap_page - Write one page to a fresh swap location. | 366 | * write_page - Write one page to a fresh swap location. |
| 367 | * @addr: Address we're writing. | 367 | * @addr: Address we're writing. |
| 368 | * @loc: Place to store the entry we used. | 368 | * @loc: Place to store the entry we used. |
| 369 | * | 369 | * |
| @@ -402,15 +402,14 @@ static int write_page(unsigned long addr, swp_entry_t * loc) | |||
| 402 | static void data_free(void) | 402 | static void data_free(void) |
| 403 | { | 403 | { |
| 404 | swp_entry_t entry; | 404 | swp_entry_t entry; |
| 405 | int i; | 405 | struct pbe * p; |
| 406 | 406 | ||
| 407 | for (i = 0; i < nr_copy_pages; i++) { | 407 | for_each_pbe(p, pagedir_nosave) { |
| 408 | entry = (pagedir_nosave + i)->swap_address; | 408 | entry = p->swap_address; |
| 409 | if (entry.val) | 409 | if (entry.val) |
| 410 | swap_free(entry); | 410 | swap_free(entry); |
| 411 | else | 411 | else |
| 412 | break; | 412 | break; |
| 413 | (pagedir_nosave + i)->swap_address = (swp_entry_t){0}; | ||
| 414 | } | 413 | } |
| 415 | } | 414 | } |
| 416 | 415 | ||
| @@ -863,6 +862,9 @@ static int alloc_image_pages(void) | |||
| 863 | return 0; | 862 | return 0; |
| 864 | } | 863 | } |
| 865 | 864 | ||
| 865 | /* Free pages we allocated for suspend. Suspend pages are alocated | ||
| 866 | * before atomic copy, so we need to free them after resume. | ||
| 867 | */ | ||
| 866 | void swsusp_free(void) | 868 | void swsusp_free(void) |
| 867 | { | 869 | { |
| 868 | BUG_ON(PageNosave(virt_to_page(pagedir_save))); | 870 | BUG_ON(PageNosave(virt_to_page(pagedir_save))); |
| @@ -918,6 +920,7 @@ static int swsusp_alloc(void) | |||
| 918 | 920 | ||
| 919 | pagedir_nosave = NULL; | 921 | pagedir_nosave = NULL; |
| 920 | nr_copy_pages = calc_nr(nr_copy_pages); | 922 | nr_copy_pages = calc_nr(nr_copy_pages); |
| 923 | nr_copy_pages_check = nr_copy_pages; | ||
| 921 | 924 | ||
| 922 | pr_debug("suspend: (pages needed: %d + %d free: %d)\n", | 925 | pr_debug("suspend: (pages needed: %d + %d free: %d)\n", |
| 923 | nr_copy_pages, PAGES_FOR_IO, nr_free_pages()); | 926 | nr_copy_pages, PAGES_FOR_IO, nr_free_pages()); |
| @@ -928,6 +931,10 @@ static int swsusp_alloc(void) | |||
| 928 | if (!enough_swap()) | 931 | if (!enough_swap()) |
| 929 | return -ENOSPC; | 932 | return -ENOSPC; |
| 930 | 933 | ||
| 934 | if (MAX_PBES < nr_copy_pages / PBES_PER_PAGE + | ||
| 935 | !!(nr_copy_pages % PBES_PER_PAGE)) | ||
| 936 | return -ENOSPC; | ||
| 937 | |||
| 931 | if (!(pagedir_save = alloc_pagedir(nr_copy_pages))) { | 938 | if (!(pagedir_save = alloc_pagedir(nr_copy_pages))) { |
| 932 | printk(KERN_ERR "suspend: Allocating pagedir failed.\n"); | 939 | printk(KERN_ERR "suspend: Allocating pagedir failed.\n"); |
| 933 | return -ENOMEM; | 940 | return -ENOMEM; |
| @@ -940,7 +947,6 @@ static int swsusp_alloc(void) | |||
| 940 | return error; | 947 | return error; |
| 941 | } | 948 | } |
| 942 | 949 | ||
| 943 | nr_copy_pages_check = nr_copy_pages; | ||
| 944 | return 0; | 950 | return 0; |
| 945 | } | 951 | } |
| 946 | 952 | ||
| @@ -1213,8 +1219,9 @@ static struct pbe * swsusp_pagedir_relocate(struct pbe *pblist) | |||
| 1213 | free_pagedir(pblist); | 1219 | free_pagedir(pblist); |
| 1214 | free_eaten_memory(); | 1220 | free_eaten_memory(); |
| 1215 | pblist = NULL; | 1221 | pblist = NULL; |
| 1216 | } | 1222 | /* Is this even worth handling? It should never ever happen, and we |
| 1217 | else | 1223 | have just lost user's state, anyway... */ |
| 1224 | } else | ||
| 1218 | printk("swsusp: Relocated %d pages\n", rel); | 1225 | printk("swsusp: Relocated %d pages\n", rel); |
| 1219 | 1226 | ||
| 1220 | return pblist; | 1227 | return pblist; |
| @@ -1434,9 +1441,9 @@ static int read_pagedir(struct pbe *pblist) | |||
| 1434 | } | 1441 | } |
| 1435 | 1442 | ||
| 1436 | if (error) | 1443 | if (error) |
| 1437 | free_page((unsigned long)pblist); | 1444 | free_pagedir(pblist); |
| 1438 | 1445 | else | |
| 1439 | BUG_ON(i != swsusp_info.pagedir_pages); | 1446 | BUG_ON(i != swsusp_info.pagedir_pages); |
| 1440 | 1447 | ||
| 1441 | return error; | 1448 | return error; |
| 1442 | } | 1449 | } |
diff --git a/kernel/printk.c b/kernel/printk.c index a967605bc2e3..4b8f0f9230a4 100644 --- a/kernel/printk.c +++ b/kernel/printk.c | |||
| @@ -488,6 +488,11 @@ static int __init printk_time_setup(char *str) | |||
| 488 | 488 | ||
| 489 | __setup("time", printk_time_setup); | 489 | __setup("time", printk_time_setup); |
| 490 | 490 | ||
| 491 | __attribute__((weak)) unsigned long long printk_clock(void) | ||
| 492 | { | ||
| 493 | return sched_clock(); | ||
| 494 | } | ||
| 495 | |||
| 491 | /* | 496 | /* |
| 492 | * This is printk. It can be called from any context. We want it to work. | 497 | * This is printk. It can be called from any context. We want it to work. |
| 493 | * | 498 | * |
| @@ -565,7 +570,7 @@ asmlinkage int vprintk(const char *fmt, va_list args) | |||
| 565 | loglev_char = default_message_loglevel | 570 | loglev_char = default_message_loglevel |
| 566 | + '0'; | 571 | + '0'; |
| 567 | } | 572 | } |
| 568 | t = sched_clock(); | 573 | t = printk_clock(); |
| 569 | nanosec_rem = do_div(t, 1000000000); | 574 | nanosec_rem = do_div(t, 1000000000); |
| 570 | tlen = sprintf(tbuf, | 575 | tlen = sprintf(tbuf, |
| 571 | "<%c>[%5lu.%06lu] ", | 576 | "<%c>[%5lu.%06lu] ", |
diff --git a/kernel/sched.c b/kernel/sched.c index 81b3a96ed2d0..1f31a528fdba 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
| @@ -294,6 +294,10 @@ static inline void prepare_lock_switch(runqueue_t *rq, task_t *next) | |||
| 294 | 294 | ||
| 295 | static inline void finish_lock_switch(runqueue_t *rq, task_t *prev) | 295 | static inline void finish_lock_switch(runqueue_t *rq, task_t *prev) |
| 296 | { | 296 | { |
| 297 | #ifdef CONFIG_DEBUG_SPINLOCK | ||
| 298 | /* this is a valid case when another task releases the spinlock */ | ||
| 299 | rq->lock.owner = current; | ||
| 300 | #endif | ||
| 297 | spin_unlock_irq(&rq->lock); | 301 | spin_unlock_irq(&rq->lock); |
| 298 | } | 302 | } |
| 299 | 303 | ||
| @@ -1529,10 +1533,6 @@ static inline void finish_task_switch(runqueue_t *rq, task_t *prev) | |||
| 1529 | * Manfred Spraul <manfred@colorfullife.com> | 1533 | * Manfred Spraul <manfred@colorfullife.com> |
| 1530 | */ | 1534 | */ |
| 1531 | prev_task_flags = prev->flags; | 1535 | prev_task_flags = prev->flags; |
| 1532 | #ifdef CONFIG_DEBUG_SPINLOCK | ||
| 1533 | /* this is a valid case when another task releases the spinlock */ | ||
| 1534 | rq->lock.owner = current; | ||
| 1535 | #endif | ||
| 1536 | finish_arch_switch(prev); | 1536 | finish_arch_switch(prev); |
| 1537 | finish_lock_switch(rq, prev); | 1537 | finish_lock_switch(rq, prev); |
| 1538 | if (mm) | 1538 | if (mm) |
diff --git a/kernel/signal.c b/kernel/signal.c index b92c3c9f8b9a..619b027e92b5 100644 --- a/kernel/signal.c +++ b/kernel/signal.c | |||
| @@ -936,34 +936,31 @@ force_sig_specific(int sig, struct task_struct *t) | |||
| 936 | * as soon as they're available, so putting the signal on the shared queue | 936 | * as soon as they're available, so putting the signal on the shared queue |
| 937 | * will be equivalent to sending it to one such thread. | 937 | * will be equivalent to sending it to one such thread. |
| 938 | */ | 938 | */ |
| 939 | #define wants_signal(sig, p, mask) \ | 939 | static inline int wants_signal(int sig, struct task_struct *p) |
| 940 | (!sigismember(&(p)->blocked, sig) \ | 940 | { |
| 941 | && !((p)->state & mask) \ | 941 | if (sigismember(&p->blocked, sig)) |
| 942 | && !((p)->flags & PF_EXITING) \ | 942 | return 0; |
| 943 | && (task_curr(p) || !signal_pending(p))) | 943 | if (p->flags & PF_EXITING) |
| 944 | 944 | return 0; | |
| 945 | if (sig == SIGKILL) | ||
| 946 | return 1; | ||
| 947 | if (p->state & (TASK_STOPPED | TASK_TRACED)) | ||
| 948 | return 0; | ||
| 949 | return task_curr(p) || !signal_pending(p); | ||
| 950 | } | ||
| 945 | 951 | ||
| 946 | static void | 952 | static void |
| 947 | __group_complete_signal(int sig, struct task_struct *p) | 953 | __group_complete_signal(int sig, struct task_struct *p) |
| 948 | { | 954 | { |
| 949 | unsigned int mask; | ||
| 950 | struct task_struct *t; | 955 | struct task_struct *t; |
| 951 | 956 | ||
| 952 | /* | 957 | /* |
| 953 | * Don't bother traced and stopped tasks (but | ||
| 954 | * SIGKILL will punch through that). | ||
| 955 | */ | ||
| 956 | mask = TASK_STOPPED | TASK_TRACED; | ||
| 957 | if (sig == SIGKILL) | ||
| 958 | mask = 0; | ||
| 959 | |||
| 960 | /* | ||
| 961 | * Now find a thread we can wake up to take the signal off the queue. | 958 | * Now find a thread we can wake up to take the signal off the queue. |
| 962 | * | 959 | * |
| 963 | * If the main thread wants the signal, it gets first crack. | 960 | * If the main thread wants the signal, it gets first crack. |
| 964 | * Probably the least surprising to the average bear. | 961 | * Probably the least surprising to the average bear. |
| 965 | */ | 962 | */ |
| 966 | if (wants_signal(sig, p, mask)) | 963 | if (wants_signal(sig, p)) |
| 967 | t = p; | 964 | t = p; |
| 968 | else if (thread_group_empty(p)) | 965 | else if (thread_group_empty(p)) |
| 969 | /* | 966 | /* |
| @@ -981,7 +978,7 @@ __group_complete_signal(int sig, struct task_struct *p) | |||
| 981 | t = p->signal->curr_target = p; | 978 | t = p->signal->curr_target = p; |
| 982 | BUG_ON(t->tgid != p->tgid); | 979 | BUG_ON(t->tgid != p->tgid); |
| 983 | 980 | ||
| 984 | while (!wants_signal(sig, t, mask)) { | 981 | while (!wants_signal(sig, t)) { |
| 985 | t = next_thread(t); | 982 | t = next_thread(t); |
| 986 | if (t == p->signal->curr_target) | 983 | if (t == p->signal->curr_target) |
| 987 | /* | 984 | /* |
| @@ -1766,7 +1763,8 @@ do_signal_stop(int signr) | |||
| 1766 | * stop is always done with the siglock held, | 1763 | * stop is always done with the siglock held, |
| 1767 | * so this check has no races. | 1764 | * so this check has no races. |
| 1768 | */ | 1765 | */ |
| 1769 | if (t->state < TASK_STOPPED) { | 1766 | if (!t->exit_state && |
| 1767 | !(t->state & (TASK_STOPPED|TASK_TRACED))) { | ||
| 1770 | stop_count++; | 1768 | stop_count++; |
| 1771 | signal_wake_up(t, 0); | 1769 | signal_wake_up(t, 0); |
| 1772 | } | 1770 | } |
diff --git a/kernel/sys.c b/kernel/sys.c index c80412be2302..2fa1ed18123c 100644 --- a/kernel/sys.c +++ b/kernel/sys.c | |||
| @@ -361,17 +361,35 @@ out_unlock: | |||
| 361 | return retval; | 361 | return retval; |
| 362 | } | 362 | } |
| 363 | 363 | ||
| 364 | /** | ||
| 365 | * emergency_restart - reboot the system | ||
| 366 | * | ||
| 367 | * Without shutting down any hardware or taking any locks | ||
| 368 | * reboot the system. This is called when we know we are in | ||
| 369 | * trouble so this is our best effort to reboot. This is | ||
| 370 | * safe to call in interrupt context. | ||
| 371 | */ | ||
| 364 | void emergency_restart(void) | 372 | void emergency_restart(void) |
| 365 | { | 373 | { |
| 366 | machine_emergency_restart(); | 374 | machine_emergency_restart(); |
| 367 | } | 375 | } |
| 368 | EXPORT_SYMBOL_GPL(emergency_restart); | 376 | EXPORT_SYMBOL_GPL(emergency_restart); |
| 369 | 377 | ||
| 370 | void kernel_restart(char *cmd) | 378 | /** |
| 379 | * kernel_restart - reboot the system | ||
| 380 | * | ||
| 381 | * Shutdown everything and perform a clean reboot. | ||
| 382 | * This is not safe to call in interrupt context. | ||
| 383 | */ | ||
| 384 | void kernel_restart_prepare(char *cmd) | ||
| 371 | { | 385 | { |
| 372 | notifier_call_chain(&reboot_notifier_list, SYS_RESTART, cmd); | 386 | notifier_call_chain(&reboot_notifier_list, SYS_RESTART, cmd); |
| 373 | system_state = SYSTEM_RESTART; | 387 | system_state = SYSTEM_RESTART; |
| 374 | device_shutdown(); | 388 | device_shutdown(); |
| 389 | } | ||
| 390 | void kernel_restart(char *cmd) | ||
| 391 | { | ||
| 392 | kernel_restart_prepare(cmd); | ||
| 375 | if (!cmd) { | 393 | if (!cmd) { |
| 376 | printk(KERN_EMERG "Restarting system.\n"); | 394 | printk(KERN_EMERG "Restarting system.\n"); |
| 377 | } else { | 395 | } else { |
| @@ -382,6 +400,12 @@ void kernel_restart(char *cmd) | |||
| 382 | } | 400 | } |
| 383 | EXPORT_SYMBOL_GPL(kernel_restart); | 401 | EXPORT_SYMBOL_GPL(kernel_restart); |
| 384 | 402 | ||
| 403 | /** | ||
| 404 | * kernel_kexec - reboot the system | ||
| 405 | * | ||
| 406 | * Move into place and start executing a preloaded standalone | ||
| 407 | * executable. If nothing was preloaded return an error. | ||
| 408 | */ | ||
| 385 | void kernel_kexec(void) | 409 | void kernel_kexec(void) |
| 386 | { | 410 | { |
| 387 | #ifdef CONFIG_KEXEC | 411 | #ifdef CONFIG_KEXEC |
| @@ -390,9 +414,7 @@ void kernel_kexec(void) | |||
| 390 | if (!image) { | 414 | if (!image) { |
| 391 | return; | 415 | return; |
| 392 | } | 416 | } |
| 393 | notifier_call_chain(&reboot_notifier_list, SYS_RESTART, NULL); | 417 | kernel_restart_prepare(NULL); |
| 394 | system_state = SYSTEM_RESTART; | ||
| 395 | device_shutdown(); | ||
| 396 | printk(KERN_EMERG "Starting new kernel\n"); | 418 | printk(KERN_EMERG "Starting new kernel\n"); |
| 397 | machine_shutdown(); | 419 | machine_shutdown(); |
| 398 | machine_kexec(image); | 420 | machine_kexec(image); |
| @@ -400,21 +422,39 @@ void kernel_kexec(void) | |||
| 400 | } | 422 | } |
| 401 | EXPORT_SYMBOL_GPL(kernel_kexec); | 423 | EXPORT_SYMBOL_GPL(kernel_kexec); |
| 402 | 424 | ||
| 403 | void kernel_halt(void) | 425 | /** |
| 426 | * kernel_halt - halt the system | ||
| 427 | * | ||
| 428 | * Shutdown everything and perform a clean system halt. | ||
| 429 | */ | ||
| 430 | void kernel_halt_prepare(void) | ||
| 404 | { | 431 | { |
| 405 | notifier_call_chain(&reboot_notifier_list, SYS_HALT, NULL); | 432 | notifier_call_chain(&reboot_notifier_list, SYS_HALT, NULL); |
| 406 | system_state = SYSTEM_HALT; | 433 | system_state = SYSTEM_HALT; |
| 407 | device_shutdown(); | 434 | device_shutdown(); |
| 435 | } | ||
| 436 | void kernel_halt(void) | ||
| 437 | { | ||
| 438 | kernel_halt_prepare(); | ||
| 408 | printk(KERN_EMERG "System halted.\n"); | 439 | printk(KERN_EMERG "System halted.\n"); |
| 409 | machine_halt(); | 440 | machine_halt(); |
| 410 | } | 441 | } |
| 411 | EXPORT_SYMBOL_GPL(kernel_halt); | 442 | EXPORT_SYMBOL_GPL(kernel_halt); |
| 412 | 443 | ||
| 413 | void kernel_power_off(void) | 444 | /** |
| 445 | * kernel_power_off - power_off the system | ||
| 446 | * | ||
| 447 | * Shutdown everything and perform a clean system power_off. | ||
| 448 | */ | ||
| 449 | void kernel_power_off_prepare(void) | ||
| 414 | { | 450 | { |
| 415 | notifier_call_chain(&reboot_notifier_list, SYS_POWER_OFF, NULL); | 451 | notifier_call_chain(&reboot_notifier_list, SYS_POWER_OFF, NULL); |
| 416 | system_state = SYSTEM_POWER_OFF; | 452 | system_state = SYSTEM_POWER_OFF; |
| 417 | device_shutdown(); | 453 | device_shutdown(); |
| 454 | } | ||
| 455 | void kernel_power_off(void) | ||
| 456 | { | ||
| 457 | kernel_power_off_prepare(); | ||
| 418 | printk(KERN_EMERG "Power down.\n"); | 458 | printk(KERN_EMERG "Power down.\n"); |
| 419 | machine_power_off(); | 459 | machine_power_off(); |
| 420 | } | 460 | } |
| @@ -1728,8 +1768,7 @@ asmlinkage long sys_prctl(int option, unsigned long arg2, unsigned long arg3, | |||
| 1728 | error = put_user(current->pdeath_signal, (int __user *)arg2); | 1768 | error = put_user(current->pdeath_signal, (int __user *)arg2); |
| 1729 | break; | 1769 | break; |
| 1730 | case PR_GET_DUMPABLE: | 1770 | case PR_GET_DUMPABLE: |
| 1731 | if (current->mm->dumpable) | 1771 | error = current->mm->dumpable; |
| 1732 | error = 1; | ||
| 1733 | break; | 1772 | break; |
| 1734 | case PR_SET_DUMPABLE: | 1773 | case PR_SET_DUMPABLE: |
| 1735 | if (arg2 < 0 || arg2 > 2) { | 1774 | if (arg2 < 0 || arg2 > 2) { |
diff --git a/kernel/timer.c b/kernel/timer.c index f4152fcd9f8e..3ba10fa35b60 100644 --- a/kernel/timer.c +++ b/kernel/timer.c | |||
| @@ -1151,19 +1151,22 @@ fastcall signed long __sched schedule_timeout(signed long timeout) | |||
| 1151 | out: | 1151 | out: |
| 1152 | return timeout < 0 ? 0 : timeout; | 1152 | return timeout < 0 ? 0 : timeout; |
| 1153 | } | 1153 | } |
| 1154 | |||
| 1155 | EXPORT_SYMBOL(schedule_timeout); | 1154 | EXPORT_SYMBOL(schedule_timeout); |
| 1156 | 1155 | ||
| 1156 | /* | ||
| 1157 | * We can use __set_current_state() here because schedule_timeout() calls | ||
| 1158 | * schedule() unconditionally. | ||
| 1159 | */ | ||
| 1157 | signed long __sched schedule_timeout_interruptible(signed long timeout) | 1160 | signed long __sched schedule_timeout_interruptible(signed long timeout) |
| 1158 | { | 1161 | { |
| 1159 | set_current_state(TASK_INTERRUPTIBLE); | 1162 | __set_current_state(TASK_INTERRUPTIBLE); |
| 1160 | return schedule_timeout(timeout); | 1163 | return schedule_timeout(timeout); |
| 1161 | } | 1164 | } |
| 1162 | EXPORT_SYMBOL(schedule_timeout_interruptible); | 1165 | EXPORT_SYMBOL(schedule_timeout_interruptible); |
| 1163 | 1166 | ||
| 1164 | signed long __sched schedule_timeout_uninterruptible(signed long timeout) | 1167 | signed long __sched schedule_timeout_uninterruptible(signed long timeout) |
| 1165 | { | 1168 | { |
| 1166 | set_current_state(TASK_UNINTERRUPTIBLE); | 1169 | __set_current_state(TASK_UNINTERRUPTIBLE); |
| 1167 | return schedule_timeout(timeout); | 1170 | return schedule_timeout(timeout); |
| 1168 | } | 1171 | } |
| 1169 | EXPORT_SYMBOL(schedule_timeout_uninterruptible); | 1172 | EXPORT_SYMBOL(schedule_timeout_uninterruptible); |
