aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/posix-cpu-timers.c
diff options
context:
space:
mode:
authorAnton Altaparmakov <aia21@cantab.net>2005-10-30 16:00:04 -0500
committerAnton Altaparmakov <aia21@cantab.net>2005-10-30 16:00:04 -0500
commit07b188ab773e183871e57b33ae37bf635c9f12ba (patch)
tree311df8a0dd12fb7bd3e9b5b1a5ca500f0428d679 /kernel/posix-cpu-timers.c
parent47c564e10f219f867bdb49225972749a43485a47 (diff)
parent9f75e1eff3edb2bb07349b94c28f4f2a6c66ca43 (diff)
Merge branch 'master' of /usr/src/ntfs-2.6/
Diffstat (limited to 'kernel/posix-cpu-timers.c')
-rw-r--r--kernel/posix-cpu-timers.c74
1 files changed, 38 insertions, 36 deletions
diff --git a/kernel/posix-cpu-timers.c b/kernel/posix-cpu-timers.c
index d30b304a3384..bf374fceb39c 100644
--- a/kernel/posix-cpu-timers.c
+++ b/kernel/posix-cpu-timers.c
@@ -91,7 +91,7 @@ static inline union cpu_time_count cpu_time_sub(clockid_t which_clock,
91 * Update expiry time from increment, and increase overrun count, 91 * Update expiry time from increment, and increase overrun count,
92 * given the current clock sample. 92 * given the current clock sample.
93 */ 93 */
94static inline void bump_cpu_timer(struct k_itimer *timer, 94static void bump_cpu_timer(struct k_itimer *timer,
95 union cpu_time_count now) 95 union cpu_time_count now)
96{ 96{
97 int i; 97 int i;
@@ -110,7 +110,7 @@ static inline void bump_cpu_timer(struct k_itimer *timer,
110 for (i = 0; incr < delta - incr; i++) 110 for (i = 0; incr < delta - incr; i++)
111 incr = incr << 1; 111 incr = incr << 1;
112 for (; i >= 0; incr >>= 1, i--) { 112 for (; i >= 0; incr >>= 1, i--) {
113 if (delta <= incr) 113 if (delta < incr)
114 continue; 114 continue;
115 timer->it.cpu.expires.sched += incr; 115 timer->it.cpu.expires.sched += incr;
116 timer->it_overrun += 1 << i; 116 timer->it_overrun += 1 << i;
@@ -128,7 +128,7 @@ static inline void bump_cpu_timer(struct k_itimer *timer,
128 for (i = 0; cputime_lt(incr, cputime_sub(delta, incr)); i++) 128 for (i = 0; cputime_lt(incr, cputime_sub(delta, incr)); i++)
129 incr = cputime_add(incr, incr); 129 incr = cputime_add(incr, incr);
130 for (; i >= 0; incr = cputime_halve(incr), i--) { 130 for (; i >= 0; incr = cputime_halve(incr), i--) {
131 if (cputime_le(delta, incr)) 131 if (cputime_lt(delta, incr))
132 continue; 132 continue;
133 timer->it.cpu.expires.cpu = 133 timer->it.cpu.expires.cpu =
134 cputime_add(timer->it.cpu.expires.cpu, incr); 134 cputime_add(timer->it.cpu.expires.cpu, incr);
@@ -380,14 +380,9 @@ int posix_cpu_timer_create(struct k_itimer *new_timer)
380int posix_cpu_timer_del(struct k_itimer *timer) 380int posix_cpu_timer_del(struct k_itimer *timer)
381{ 381{
382 struct task_struct *p = timer->it.cpu.task; 382 struct task_struct *p = timer->it.cpu.task;
383 int ret = 0;
383 384
384 if (timer->it.cpu.firing) 385 if (likely(p != NULL)) {
385 return TIMER_RETRY;
386
387 if (unlikely(p == NULL))
388 return 0;
389
390 if (!list_empty(&timer->it.cpu.entry)) {
391 read_lock(&tasklist_lock); 386 read_lock(&tasklist_lock);
392 if (unlikely(p->signal == NULL)) { 387 if (unlikely(p->signal == NULL)) {
393 /* 388 /*
@@ -396,18 +391,20 @@ int posix_cpu_timer_del(struct k_itimer *timer)
396 */ 391 */
397 BUG_ON(!list_empty(&timer->it.cpu.entry)); 392 BUG_ON(!list_empty(&timer->it.cpu.entry));
398 } else { 393 } else {
399 /*
400 * Take us off the task's timer list.
401 */
402 spin_lock(&p->sighand->siglock); 394 spin_lock(&p->sighand->siglock);
403 list_del(&timer->it.cpu.entry); 395 if (timer->it.cpu.firing)
396 ret = TIMER_RETRY;
397 else
398 list_del(&timer->it.cpu.entry);
404 spin_unlock(&p->sighand->siglock); 399 spin_unlock(&p->sighand->siglock);
405 } 400 }
406 read_unlock(&tasklist_lock); 401 read_unlock(&tasklist_lock);
402
403 if (!ret)
404 put_task_struct(p);
407 } 405 }
408 put_task_struct(p);
409 406
410 return 0; 407 return ret;
411} 408}
412 409
413/* 410/*
@@ -424,8 +421,6 @@ static void cleanup_timers(struct list_head *head,
424 cputime_t ptime = cputime_add(utime, stime); 421 cputime_t ptime = cputime_add(utime, stime);
425 422
426 list_for_each_entry_safe(timer, next, head, entry) { 423 list_for_each_entry_safe(timer, next, head, entry) {
427 put_task_struct(timer->task);
428 timer->task = NULL;
429 list_del_init(&timer->entry); 424 list_del_init(&timer->entry);
430 if (cputime_lt(timer->expires.cpu, ptime)) { 425 if (cputime_lt(timer->expires.cpu, ptime)) {
431 timer->expires.cpu = cputime_zero; 426 timer->expires.cpu = cputime_zero;
@@ -437,8 +432,6 @@ static void cleanup_timers(struct list_head *head,
437 432
438 ++head; 433 ++head;
439 list_for_each_entry_safe(timer, next, head, entry) { 434 list_for_each_entry_safe(timer, next, head, entry) {
440 put_task_struct(timer->task);
441 timer->task = NULL;
442 list_del_init(&timer->entry); 435 list_del_init(&timer->entry);
443 if (cputime_lt(timer->expires.cpu, utime)) { 436 if (cputime_lt(timer->expires.cpu, utime)) {
444 timer->expires.cpu = cputime_zero; 437 timer->expires.cpu = cputime_zero;
@@ -450,8 +443,6 @@ static void cleanup_timers(struct list_head *head,
450 443
451 ++head; 444 ++head;
452 list_for_each_entry_safe(timer, next, head, entry) { 445 list_for_each_entry_safe(timer, next, head, entry) {
453 put_task_struct(timer->task);
454 timer->task = NULL;
455 list_del_init(&timer->entry); 446 list_del_init(&timer->entry);
456 if (timer->expires.sched < sched_time) { 447 if (timer->expires.sched < sched_time) {
457 timer->expires.sched = 0; 448 timer->expires.sched = 0;
@@ -495,6 +486,9 @@ static void process_timer_rebalance(struct task_struct *p,
495 struct task_struct *t = p; 486 struct task_struct *t = p;
496 unsigned int nthreads = atomic_read(&p->signal->live); 487 unsigned int nthreads = atomic_read(&p->signal->live);
497 488
489 if (!nthreads)
490 return;
491
498 switch (clock_idx) { 492 switch (clock_idx) {
499 default: 493 default:
500 BUG(); 494 BUG();
@@ -503,7 +497,7 @@ static void process_timer_rebalance(struct task_struct *p,
503 left = cputime_div(cputime_sub(expires.cpu, val.cpu), 497 left = cputime_div(cputime_sub(expires.cpu, val.cpu),
504 nthreads); 498 nthreads);
505 do { 499 do {
506 if (!unlikely(t->exit_state)) { 500 if (!unlikely(t->flags & PF_EXITING)) {
507 ticks = cputime_add(prof_ticks(t), left); 501 ticks = cputime_add(prof_ticks(t), left);
508 if (cputime_eq(t->it_prof_expires, 502 if (cputime_eq(t->it_prof_expires,
509 cputime_zero) || 503 cputime_zero) ||
@@ -518,7 +512,7 @@ static void process_timer_rebalance(struct task_struct *p,
518 left = cputime_div(cputime_sub(expires.cpu, val.cpu), 512 left = cputime_div(cputime_sub(expires.cpu, val.cpu),
519 nthreads); 513 nthreads);
520 do { 514 do {
521 if (!unlikely(t->exit_state)) { 515 if (!unlikely(t->flags & PF_EXITING)) {
522 ticks = cputime_add(virt_ticks(t), left); 516 ticks = cputime_add(virt_ticks(t), left);
523 if (cputime_eq(t->it_virt_expires, 517 if (cputime_eq(t->it_virt_expires,
524 cputime_zero) || 518 cputime_zero) ||
@@ -533,7 +527,7 @@ static void process_timer_rebalance(struct task_struct *p,
533 nsleft = expires.sched - val.sched; 527 nsleft = expires.sched - val.sched;
534 do_div(nsleft, nthreads); 528 do_div(nsleft, nthreads);
535 do { 529 do {
536 if (!unlikely(t->exit_state)) { 530 if (!unlikely(t->flags & PF_EXITING)) {
537 ns = t->sched_time + nsleft; 531 ns = t->sched_time + nsleft;
538 if (t->it_sched_expires == 0 || 532 if (t->it_sched_expires == 0 ||
539 t->it_sched_expires > ns) { 533 t->it_sched_expires > ns) {
@@ -572,6 +566,9 @@ static void arm_timer(struct k_itimer *timer, union cpu_time_count now)
572 struct cpu_timer_list *next; 566 struct cpu_timer_list *next;
573 unsigned long i; 567 unsigned long i;
574 568
569 if (CPUCLOCK_PERTHREAD(timer->it_clock) && (p->flags & PF_EXITING))
570 return;
571
575 head = (CPUCLOCK_PERTHREAD(timer->it_clock) ? 572 head = (CPUCLOCK_PERTHREAD(timer->it_clock) ?
576 p->cpu_timers : p->signal->cpu_timers); 573 p->cpu_timers : p->signal->cpu_timers);
577 head += CPUCLOCK_WHICH(timer->it_clock); 574 head += CPUCLOCK_WHICH(timer->it_clock);
@@ -582,17 +579,15 @@ static void arm_timer(struct k_itimer *timer, union cpu_time_count now)
582 listpos = head; 579 listpos = head;
583 if (CPUCLOCK_WHICH(timer->it_clock) == CPUCLOCK_SCHED) { 580 if (CPUCLOCK_WHICH(timer->it_clock) == CPUCLOCK_SCHED) {
584 list_for_each_entry(next, head, entry) { 581 list_for_each_entry(next, head, entry) {
585 if (next->expires.sched > nt->expires.sched) { 582 if (next->expires.sched > nt->expires.sched)
586 listpos = &next->entry;
587 break; 583 break;
588 } 584 listpos = &next->entry;
589 } 585 }
590 } else { 586 } else {
591 list_for_each_entry(next, head, entry) { 587 list_for_each_entry(next, head, entry) {
592 if (cputime_gt(next->expires.cpu, nt->expires.cpu)) { 588 if (cputime_gt(next->expires.cpu, nt->expires.cpu))
593 listpos = &next->entry;
594 break; 589 break;
595 } 590 listpos = &next->entry;
596 } 591 }
597 } 592 }
598 list_add(&nt->entry, listpos); 593 list_add(&nt->entry, listpos);
@@ -736,9 +731,15 @@ int posix_cpu_timer_set(struct k_itimer *timer, int flags,
736 * Disarm any old timer after extracting its expiry time. 731 * Disarm any old timer after extracting its expiry time.
737 */ 732 */
738 BUG_ON(!irqs_disabled()); 733 BUG_ON(!irqs_disabled());
734
735 ret = 0;
739 spin_lock(&p->sighand->siglock); 736 spin_lock(&p->sighand->siglock);
740 old_expires = timer->it.cpu.expires; 737 old_expires = timer->it.cpu.expires;
741 list_del_init(&timer->it.cpu.entry); 738 if (unlikely(timer->it.cpu.firing)) {
739 timer->it.cpu.firing = -1;
740 ret = TIMER_RETRY;
741 } else
742 list_del_init(&timer->it.cpu.entry);
742 spin_unlock(&p->sighand->siglock); 743 spin_unlock(&p->sighand->siglock);
743 744
744 /* 745 /*
@@ -786,7 +787,7 @@ int posix_cpu_timer_set(struct k_itimer *timer, int flags,
786 } 787 }
787 } 788 }
788 789
789 if (unlikely(timer->it.cpu.firing)) { 790 if (unlikely(ret)) {
790 /* 791 /*
791 * We are colliding with the timer actually firing. 792 * We are colliding with the timer actually firing.
792 * Punt after filling in the timer's old value, and 793 * Punt after filling in the timer's old value, and
@@ -794,8 +795,6 @@ int posix_cpu_timer_set(struct k_itimer *timer, int flags,
794 * it as an overrun (thanks to bump_cpu_timer above). 795 * it as an overrun (thanks to bump_cpu_timer above).
795 */ 796 */
796 read_unlock(&tasklist_lock); 797 read_unlock(&tasklist_lock);
797 timer->it.cpu.firing = -1;
798 ret = TIMER_RETRY;
799 goto out; 798 goto out;
800 } 799 }
801 800
@@ -1169,6 +1168,9 @@ static void check_process_timers(struct task_struct *tsk,
1169 unsigned long long sched_left, sched; 1168 unsigned long long sched_left, sched;
1170 const unsigned int nthreads = atomic_read(&sig->live); 1169 const unsigned int nthreads = atomic_read(&sig->live);
1171 1170
1171 if (!nthreads)
1172 return;
1173
1172 prof_left = cputime_sub(prof_expires, utime); 1174 prof_left = cputime_sub(prof_expires, utime);
1173 prof_left = cputime_sub(prof_left, stime); 1175 prof_left = cputime_sub(prof_left, stime);
1174 prof_left = cputime_div(prof_left, nthreads); 1176 prof_left = cputime_div(prof_left, nthreads);
@@ -1205,7 +1207,7 @@ static void check_process_timers(struct task_struct *tsk,
1205 1207
1206 do { 1208 do {
1207 t = next_thread(t); 1209 t = next_thread(t);
1208 } while (unlikely(t->exit_state)); 1210 } while (unlikely(t->flags & PF_EXITING));
1209 } while (t != tsk); 1211 } while (t != tsk);
1210 } 1212 }
1211} 1213}