diff options
Diffstat (limited to 'kernel')
| -rw-r--r-- | kernel/audit.c | 6 | ||||
| -rw-r--r-- | kernel/auditsc.c | 2 | ||||
| -rw-r--r-- | kernel/kexec.c | 7 | ||||
| -rw-r--r-- | kernel/posix-cpu-timers.c | 74 | ||||
| -rw-r--r-- | kernel/power/swsusp.c | 2 | ||||
| -rw-r--r-- | kernel/sched.c | 1 |
6 files changed, 47 insertions, 45 deletions
diff --git a/kernel/audit.c b/kernel/audit.c index aefa73a8a586..0c56320d38dc 100644 --- a/kernel/audit.c +++ b/kernel/audit.c | |||
| @@ -133,7 +133,7 @@ struct audit_buffer { | |||
| 133 | struct list_head list; | 133 | struct list_head list; |
| 134 | struct sk_buff *skb; /* formatted skb ready to send */ | 134 | struct sk_buff *skb; /* formatted skb ready to send */ |
| 135 | struct audit_context *ctx; /* NULL or associated context */ | 135 | struct audit_context *ctx; /* NULL or associated context */ |
| 136 | int gfp_mask; | 136 | gfp_t gfp_mask; |
| 137 | }; | 137 | }; |
| 138 | 138 | ||
| 139 | static void audit_set_pid(struct audit_buffer *ab, pid_t pid) | 139 | static void audit_set_pid(struct audit_buffer *ab, pid_t pid) |
| @@ -647,7 +647,7 @@ static inline void audit_get_stamp(struct audit_context *ctx, | |||
| 647 | * will be written at syscall exit. If there is no associated task, tsk | 647 | * will be written at syscall exit. If there is no associated task, tsk |
| 648 | * should be NULL. */ | 648 | * should be NULL. */ |
| 649 | 649 | ||
| 650 | struct audit_buffer *audit_log_start(struct audit_context *ctx, int gfp_mask, | 650 | struct audit_buffer *audit_log_start(struct audit_context *ctx, gfp_t gfp_mask, |
| 651 | int type) | 651 | int type) |
| 652 | { | 652 | { |
| 653 | struct audit_buffer *ab = NULL; | 653 | struct audit_buffer *ab = NULL; |
| @@ -879,7 +879,7 @@ void audit_log_end(struct audit_buffer *ab) | |||
| 879 | /* Log an audit record. This is a convenience function that calls | 879 | /* Log an audit record. This is a convenience function that calls |
| 880 | * audit_log_start, audit_log_vformat, and audit_log_end. It may be | 880 | * audit_log_start, audit_log_vformat, and audit_log_end. It may be |
| 881 | * called in any context. */ | 881 | * called in any context. */ |
| 882 | void audit_log(struct audit_context *ctx, int gfp_mask, int type, | 882 | void audit_log(struct audit_context *ctx, gfp_t gfp_mask, int type, |
| 883 | const char *fmt, ...) | 883 | const char *fmt, ...) |
| 884 | { | 884 | { |
| 885 | struct audit_buffer *ab; | 885 | struct audit_buffer *ab; |
diff --git a/kernel/auditsc.c b/kernel/auditsc.c index 88696f639aab..d8a68509e729 100644 --- a/kernel/auditsc.c +++ b/kernel/auditsc.c | |||
| @@ -803,7 +803,7 @@ static void audit_log_task_info(struct audit_buffer *ab) | |||
| 803 | up_read(&mm->mmap_sem); | 803 | up_read(&mm->mmap_sem); |
| 804 | } | 804 | } |
| 805 | 805 | ||
| 806 | static void audit_log_exit(struct audit_context *context, unsigned int gfp_mask) | 806 | static void audit_log_exit(struct audit_context *context, gfp_t gfp_mask) |
| 807 | { | 807 | { |
| 808 | int i; | 808 | int i; |
| 809 | struct audit_buffer *ab; | 809 | struct audit_buffer *ab; |
diff --git a/kernel/kexec.c b/kernel/kexec.c index cdd4dcd8fb63..36c5d9cd4cc1 100644 --- a/kernel/kexec.c +++ b/kernel/kexec.c | |||
| @@ -90,7 +90,7 @@ int kexec_should_crash(struct task_struct *p) | |||
| 90 | static int kimage_is_destination_range(struct kimage *image, | 90 | static int kimage_is_destination_range(struct kimage *image, |
| 91 | unsigned long start, unsigned long end); | 91 | unsigned long start, unsigned long end); |
| 92 | static struct page *kimage_alloc_page(struct kimage *image, | 92 | static struct page *kimage_alloc_page(struct kimage *image, |
| 93 | unsigned int gfp_mask, | 93 | gfp_t gfp_mask, |
| 94 | unsigned long dest); | 94 | unsigned long dest); |
| 95 | 95 | ||
| 96 | static int do_kimage_alloc(struct kimage **rimage, unsigned long entry, | 96 | static int do_kimage_alloc(struct kimage **rimage, unsigned long entry, |
| @@ -326,8 +326,7 @@ static int kimage_is_destination_range(struct kimage *image, | |||
| 326 | return 0; | 326 | return 0; |
| 327 | } | 327 | } |
| 328 | 328 | ||
| 329 | static struct page *kimage_alloc_pages(unsigned int gfp_mask, | 329 | static struct page *kimage_alloc_pages(gfp_t gfp_mask, unsigned int order) |
| 330 | unsigned int order) | ||
| 331 | { | 330 | { |
| 332 | struct page *pages; | 331 | struct page *pages; |
| 333 | 332 | ||
| @@ -654,7 +653,7 @@ static kimage_entry_t *kimage_dst_used(struct kimage *image, | |||
| 654 | } | 653 | } |
| 655 | 654 | ||
| 656 | static struct page *kimage_alloc_page(struct kimage *image, | 655 | static struct page *kimage_alloc_page(struct kimage *image, |
| 657 | unsigned int gfp_mask, | 656 | gfp_t gfp_mask, |
| 658 | unsigned long destination) | 657 | unsigned long destination) |
| 659 | { | 658 | { |
| 660 | /* | 659 | /* |
diff --git a/kernel/posix-cpu-timers.c b/kernel/posix-cpu-timers.c index d30b304a3384..bf374fceb39c 100644 --- a/kernel/posix-cpu-timers.c +++ b/kernel/posix-cpu-timers.c | |||
| @@ -91,7 +91,7 @@ static inline union cpu_time_count cpu_time_sub(clockid_t which_clock, | |||
| 91 | * Update expiry time from increment, and increase overrun count, | 91 | * Update expiry time from increment, and increase overrun count, |
| 92 | * given the current clock sample. | 92 | * given the current clock sample. |
| 93 | */ | 93 | */ |
| 94 | static inline void bump_cpu_timer(struct k_itimer *timer, | 94 | static void bump_cpu_timer(struct k_itimer *timer, |
| 95 | union cpu_time_count now) | 95 | union cpu_time_count now) |
| 96 | { | 96 | { |
| 97 | int i; | 97 | int i; |
| @@ -110,7 +110,7 @@ static inline void bump_cpu_timer(struct k_itimer *timer, | |||
| 110 | for (i = 0; incr < delta - incr; i++) | 110 | for (i = 0; incr < delta - incr; i++) |
| 111 | incr = incr << 1; | 111 | incr = incr << 1; |
| 112 | for (; i >= 0; incr >>= 1, i--) { | 112 | for (; i >= 0; incr >>= 1, i--) { |
| 113 | if (delta <= incr) | 113 | if (delta < incr) |
| 114 | continue; | 114 | continue; |
| 115 | timer->it.cpu.expires.sched += incr; | 115 | timer->it.cpu.expires.sched += incr; |
| 116 | timer->it_overrun += 1 << i; | 116 | timer->it_overrun += 1 << i; |
| @@ -128,7 +128,7 @@ static inline void bump_cpu_timer(struct k_itimer *timer, | |||
| 128 | for (i = 0; cputime_lt(incr, cputime_sub(delta, incr)); i++) | 128 | for (i = 0; cputime_lt(incr, cputime_sub(delta, incr)); i++) |
| 129 | incr = cputime_add(incr, incr); | 129 | incr = cputime_add(incr, incr); |
| 130 | for (; i >= 0; incr = cputime_halve(incr), i--) { | 130 | for (; i >= 0; incr = cputime_halve(incr), i--) { |
| 131 | if (cputime_le(delta, incr)) | 131 | if (cputime_lt(delta, incr)) |
| 132 | continue; | 132 | continue; |
| 133 | timer->it.cpu.expires.cpu = | 133 | timer->it.cpu.expires.cpu = |
| 134 | cputime_add(timer->it.cpu.expires.cpu, incr); | 134 | cputime_add(timer->it.cpu.expires.cpu, incr); |
| @@ -380,14 +380,9 @@ int posix_cpu_timer_create(struct k_itimer *new_timer) | |||
| 380 | int posix_cpu_timer_del(struct k_itimer *timer) | 380 | int posix_cpu_timer_del(struct k_itimer *timer) |
| 381 | { | 381 | { |
| 382 | struct task_struct *p = timer->it.cpu.task; | 382 | struct task_struct *p = timer->it.cpu.task; |
| 383 | int ret = 0; | ||
| 383 | 384 | ||
| 384 | if (timer->it.cpu.firing) | 385 | if (likely(p != NULL)) { |
| 385 | return TIMER_RETRY; | ||
| 386 | |||
| 387 | if (unlikely(p == NULL)) | ||
| 388 | return 0; | ||
| 389 | |||
| 390 | if (!list_empty(&timer->it.cpu.entry)) { | ||
| 391 | read_lock(&tasklist_lock); | 386 | read_lock(&tasklist_lock); |
| 392 | if (unlikely(p->signal == NULL)) { | 387 | if (unlikely(p->signal == NULL)) { |
| 393 | /* | 388 | /* |
| @@ -396,18 +391,20 @@ int posix_cpu_timer_del(struct k_itimer *timer) | |||
| 396 | */ | 391 | */ |
| 397 | BUG_ON(!list_empty(&timer->it.cpu.entry)); | 392 | BUG_ON(!list_empty(&timer->it.cpu.entry)); |
| 398 | } else { | 393 | } else { |
| 399 | /* | ||
| 400 | * Take us off the task's timer list. | ||
| 401 | */ | ||
| 402 | spin_lock(&p->sighand->siglock); | 394 | spin_lock(&p->sighand->siglock); |
| 403 | list_del(&timer->it.cpu.entry); | 395 | if (timer->it.cpu.firing) |
| 396 | ret = TIMER_RETRY; | ||
| 397 | else | ||
| 398 | list_del(&timer->it.cpu.entry); | ||
| 404 | spin_unlock(&p->sighand->siglock); | 399 | spin_unlock(&p->sighand->siglock); |
| 405 | } | 400 | } |
| 406 | read_unlock(&tasklist_lock); | 401 | read_unlock(&tasklist_lock); |
| 402 | |||
| 403 | if (!ret) | ||
| 404 | put_task_struct(p); | ||
| 407 | } | 405 | } |
| 408 | put_task_struct(p); | ||
| 409 | 406 | ||
| 410 | return 0; | 407 | return ret; |
| 411 | } | 408 | } |
| 412 | 409 | ||
| 413 | /* | 410 | /* |
| @@ -424,8 +421,6 @@ static void cleanup_timers(struct list_head *head, | |||
| 424 | cputime_t ptime = cputime_add(utime, stime); | 421 | cputime_t ptime = cputime_add(utime, stime); |
| 425 | 422 | ||
| 426 | list_for_each_entry_safe(timer, next, head, entry) { | 423 | list_for_each_entry_safe(timer, next, head, entry) { |
| 427 | put_task_struct(timer->task); | ||
| 428 | timer->task = NULL; | ||
| 429 | list_del_init(&timer->entry); | 424 | list_del_init(&timer->entry); |
| 430 | if (cputime_lt(timer->expires.cpu, ptime)) { | 425 | if (cputime_lt(timer->expires.cpu, ptime)) { |
| 431 | timer->expires.cpu = cputime_zero; | 426 | timer->expires.cpu = cputime_zero; |
| @@ -437,8 +432,6 @@ static void cleanup_timers(struct list_head *head, | |||
| 437 | 432 | ||
| 438 | ++head; | 433 | ++head; |
| 439 | list_for_each_entry_safe(timer, next, head, entry) { | 434 | list_for_each_entry_safe(timer, next, head, entry) { |
| 440 | put_task_struct(timer->task); | ||
| 441 | timer->task = NULL; | ||
| 442 | list_del_init(&timer->entry); | 435 | list_del_init(&timer->entry); |
| 443 | if (cputime_lt(timer->expires.cpu, utime)) { | 436 | if (cputime_lt(timer->expires.cpu, utime)) { |
| 444 | timer->expires.cpu = cputime_zero; | 437 | timer->expires.cpu = cputime_zero; |
| @@ -450,8 +443,6 @@ static void cleanup_timers(struct list_head *head, | |||
| 450 | 443 | ||
| 451 | ++head; | 444 | ++head; |
| 452 | list_for_each_entry_safe(timer, next, head, entry) { | 445 | list_for_each_entry_safe(timer, next, head, entry) { |
| 453 | put_task_struct(timer->task); | ||
| 454 | timer->task = NULL; | ||
| 455 | list_del_init(&timer->entry); | 446 | list_del_init(&timer->entry); |
| 456 | if (timer->expires.sched < sched_time) { | 447 | if (timer->expires.sched < sched_time) { |
| 457 | timer->expires.sched = 0; | 448 | timer->expires.sched = 0; |
| @@ -495,6 +486,9 @@ static void process_timer_rebalance(struct task_struct *p, | |||
| 495 | struct task_struct *t = p; | 486 | struct task_struct *t = p; |
| 496 | unsigned int nthreads = atomic_read(&p->signal->live); | 487 | unsigned int nthreads = atomic_read(&p->signal->live); |
| 497 | 488 | ||
| 489 | if (!nthreads) | ||
| 490 | return; | ||
| 491 | |||
| 498 | switch (clock_idx) { | 492 | switch (clock_idx) { |
| 499 | default: | 493 | default: |
| 500 | BUG(); | 494 | BUG(); |
| @@ -503,7 +497,7 @@ static void process_timer_rebalance(struct task_struct *p, | |||
| 503 | left = cputime_div(cputime_sub(expires.cpu, val.cpu), | 497 | left = cputime_div(cputime_sub(expires.cpu, val.cpu), |
| 504 | nthreads); | 498 | nthreads); |
| 505 | do { | 499 | do { |
| 506 | if (!unlikely(t->exit_state)) { | 500 | if (!unlikely(t->flags & PF_EXITING)) { |
| 507 | ticks = cputime_add(prof_ticks(t), left); | 501 | ticks = cputime_add(prof_ticks(t), left); |
| 508 | if (cputime_eq(t->it_prof_expires, | 502 | if (cputime_eq(t->it_prof_expires, |
| 509 | cputime_zero) || | 503 | cputime_zero) || |
| @@ -518,7 +512,7 @@ static void process_timer_rebalance(struct task_struct *p, | |||
| 518 | left = cputime_div(cputime_sub(expires.cpu, val.cpu), | 512 | left = cputime_div(cputime_sub(expires.cpu, val.cpu), |
| 519 | nthreads); | 513 | nthreads); |
| 520 | do { | 514 | do { |
| 521 | if (!unlikely(t->exit_state)) { | 515 | if (!unlikely(t->flags & PF_EXITING)) { |
| 522 | ticks = cputime_add(virt_ticks(t), left); | 516 | ticks = cputime_add(virt_ticks(t), left); |
| 523 | if (cputime_eq(t->it_virt_expires, | 517 | if (cputime_eq(t->it_virt_expires, |
| 524 | cputime_zero) || | 518 | cputime_zero) || |
| @@ -533,7 +527,7 @@ static void process_timer_rebalance(struct task_struct *p, | |||
| 533 | nsleft = expires.sched - val.sched; | 527 | nsleft = expires.sched - val.sched; |
| 534 | do_div(nsleft, nthreads); | 528 | do_div(nsleft, nthreads); |
| 535 | do { | 529 | do { |
| 536 | if (!unlikely(t->exit_state)) { | 530 | if (!unlikely(t->flags & PF_EXITING)) { |
| 537 | ns = t->sched_time + nsleft; | 531 | ns = t->sched_time + nsleft; |
| 538 | if (t->it_sched_expires == 0 || | 532 | if (t->it_sched_expires == 0 || |
| 539 | t->it_sched_expires > ns) { | 533 | t->it_sched_expires > ns) { |
| @@ -572,6 +566,9 @@ static void arm_timer(struct k_itimer *timer, union cpu_time_count now) | |||
| 572 | struct cpu_timer_list *next; | 566 | struct cpu_timer_list *next; |
| 573 | unsigned long i; | 567 | unsigned long i; |
| 574 | 568 | ||
| 569 | if (CPUCLOCK_PERTHREAD(timer->it_clock) && (p->flags & PF_EXITING)) | ||
| 570 | return; | ||
| 571 | |||
| 575 | head = (CPUCLOCK_PERTHREAD(timer->it_clock) ? | 572 | head = (CPUCLOCK_PERTHREAD(timer->it_clock) ? |
| 576 | p->cpu_timers : p->signal->cpu_timers); | 573 | p->cpu_timers : p->signal->cpu_timers); |
| 577 | head += CPUCLOCK_WHICH(timer->it_clock); | 574 | head += CPUCLOCK_WHICH(timer->it_clock); |
| @@ -582,17 +579,15 @@ static void arm_timer(struct k_itimer *timer, union cpu_time_count now) | |||
| 582 | listpos = head; | 579 | listpos = head; |
| 583 | if (CPUCLOCK_WHICH(timer->it_clock) == CPUCLOCK_SCHED) { | 580 | if (CPUCLOCK_WHICH(timer->it_clock) == CPUCLOCK_SCHED) { |
| 584 | list_for_each_entry(next, head, entry) { | 581 | list_for_each_entry(next, head, entry) { |
| 585 | if (next->expires.sched > nt->expires.sched) { | 582 | if (next->expires.sched > nt->expires.sched) |
| 586 | listpos = &next->entry; | ||
| 587 | break; | 583 | break; |
| 588 | } | 584 | listpos = &next->entry; |
| 589 | } | 585 | } |
| 590 | } else { | 586 | } else { |
| 591 | list_for_each_entry(next, head, entry) { | 587 | list_for_each_entry(next, head, entry) { |
| 592 | if (cputime_gt(next->expires.cpu, nt->expires.cpu)) { | 588 | if (cputime_gt(next->expires.cpu, nt->expires.cpu)) |
| 593 | listpos = &next->entry; | ||
| 594 | break; | 589 | break; |
| 595 | } | 590 | listpos = &next->entry; |
| 596 | } | 591 | } |
| 597 | } | 592 | } |
| 598 | list_add(&nt->entry, listpos); | 593 | list_add(&nt->entry, listpos); |
| @@ -736,9 +731,15 @@ int posix_cpu_timer_set(struct k_itimer *timer, int flags, | |||
| 736 | * Disarm any old timer after extracting its expiry time. | 731 | * Disarm any old timer after extracting its expiry time. |
| 737 | */ | 732 | */ |
| 738 | BUG_ON(!irqs_disabled()); | 733 | BUG_ON(!irqs_disabled()); |
| 734 | |||
| 735 | ret = 0; | ||
| 739 | spin_lock(&p->sighand->siglock); | 736 | spin_lock(&p->sighand->siglock); |
| 740 | old_expires = timer->it.cpu.expires; | 737 | old_expires = timer->it.cpu.expires; |
| 741 | list_del_init(&timer->it.cpu.entry); | 738 | if (unlikely(timer->it.cpu.firing)) { |
| 739 | timer->it.cpu.firing = -1; | ||
| 740 | ret = TIMER_RETRY; | ||
| 741 | } else | ||
| 742 | list_del_init(&timer->it.cpu.entry); | ||
| 742 | spin_unlock(&p->sighand->siglock); | 743 | spin_unlock(&p->sighand->siglock); |
| 743 | 744 | ||
| 744 | /* | 745 | /* |
| @@ -786,7 +787,7 @@ int posix_cpu_timer_set(struct k_itimer *timer, int flags, | |||
| 786 | } | 787 | } |
| 787 | } | 788 | } |
| 788 | 789 | ||
| 789 | if (unlikely(timer->it.cpu.firing)) { | 790 | if (unlikely(ret)) { |
| 790 | /* | 791 | /* |
| 791 | * We are colliding with the timer actually firing. | 792 | * We are colliding with the timer actually firing. |
| 792 | * Punt after filling in the timer's old value, and | 793 | * Punt after filling in the timer's old value, and |
| @@ -794,8 +795,6 @@ int posix_cpu_timer_set(struct k_itimer *timer, int flags, | |||
| 794 | * it as an overrun (thanks to bump_cpu_timer above). | 795 | * it as an overrun (thanks to bump_cpu_timer above). |
| 795 | */ | 796 | */ |
| 796 | read_unlock(&tasklist_lock); | 797 | read_unlock(&tasklist_lock); |
| 797 | timer->it.cpu.firing = -1; | ||
| 798 | ret = TIMER_RETRY; | ||
| 799 | goto out; | 798 | goto out; |
| 800 | } | 799 | } |
| 801 | 800 | ||
| @@ -1169,6 +1168,9 @@ static void check_process_timers(struct task_struct *tsk, | |||
| 1169 | unsigned long long sched_left, sched; | 1168 | unsigned long long sched_left, sched; |
| 1170 | const unsigned int nthreads = atomic_read(&sig->live); | 1169 | const unsigned int nthreads = atomic_read(&sig->live); |
| 1171 | 1170 | ||
| 1171 | if (!nthreads) | ||
| 1172 | return; | ||
| 1173 | |||
| 1172 | prof_left = cputime_sub(prof_expires, utime); | 1174 | prof_left = cputime_sub(prof_expires, utime); |
| 1173 | prof_left = cputime_sub(prof_left, stime); | 1175 | prof_left = cputime_sub(prof_left, stime); |
| 1174 | prof_left = cputime_div(prof_left, nthreads); | 1176 | prof_left = cputime_div(prof_left, nthreads); |
| @@ -1205,7 +1207,7 @@ static void check_process_timers(struct task_struct *tsk, | |||
| 1205 | 1207 | ||
| 1206 | do { | 1208 | do { |
| 1207 | t = next_thread(t); | 1209 | t = next_thread(t); |
| 1208 | } while (unlikely(t->exit_state)); | 1210 | } while (unlikely(t->flags & PF_EXITING)); |
| 1209 | } while (t != tsk); | 1211 | } while (t != tsk); |
| 1210 | } | 1212 | } |
| 1211 | } | 1213 | } |
diff --git a/kernel/power/swsusp.c b/kernel/power/swsusp.c index 2d5c45676442..10bc5ec496d7 100644 --- a/kernel/power/swsusp.c +++ b/kernel/power/swsusp.c | |||
| @@ -1095,7 +1095,7 @@ static inline void eat_page(void *page) | |||
| 1095 | *eaten_memory = c; | 1095 | *eaten_memory = c; |
| 1096 | } | 1096 | } |
| 1097 | 1097 | ||
| 1098 | unsigned long get_usable_page(unsigned gfp_mask) | 1098 | unsigned long get_usable_page(gfp_t gfp_mask) |
| 1099 | { | 1099 | { |
| 1100 | unsigned long m; | 1100 | unsigned long m; |
| 1101 | 1101 | ||
diff --git a/kernel/sched.c b/kernel/sched.c index 1f31a528fdba..1e5cafdf4e27 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
| @@ -3879,6 +3879,7 @@ EXPORT_SYMBOL(cpu_present_map); | |||
| 3879 | 3879 | ||
| 3880 | #ifndef CONFIG_SMP | 3880 | #ifndef CONFIG_SMP |
| 3881 | cpumask_t cpu_online_map = CPU_MASK_ALL; | 3881 | cpumask_t cpu_online_map = CPU_MASK_ALL; |
| 3882 | EXPORT_SYMBOL_GPL(cpu_online_map); | ||
| 3882 | cpumask_t cpu_possible_map = CPU_MASK_ALL; | 3883 | cpumask_t cpu_possible_map = CPU_MASK_ALL; |
| 3883 | #endif | 3884 | #endif |
| 3884 | 3885 | ||
