diff options
author | Ingo Molnar <mingo@elte.hu> | 2007-08-09 05:16:51 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2007-08-09 05:16:51 -0400 |
commit | bdd4dfa89c1e3e1379729b9edec1526b3ecc25ec (patch) | |
tree | 9bf9002699184a1171522735b2be92bead6df02e /kernel | |
parent | 2e1cb74a501c4b1bca5e55dabff24f267349193c (diff) |
sched: remove the 'u64 now' local variables
final step: remove all (now superfluous) 'u64 now' variables.
( identity transformation that causes no change in functionality. )
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/sched.c | 16 | ||||
-rw-r--r-- | kernel/sched_fair.c | 6 |
2 files changed, 0 insertions, 22 deletions
diff --git a/kernel/sched.c b/kernel/sched.c index 2dc5d2f7b39..b78b9d9ffd1 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -910,10 +910,7 @@ static int effective_prio(struct task_struct *p) | |||
910 | */ | 910 | */ |
911 | static void activate_task(struct rq *rq, struct task_struct *p, int wakeup) | 911 | static void activate_task(struct rq *rq, struct task_struct *p, int wakeup) |
912 | { | 912 | { |
913 | u64 now; | ||
914 | |||
915 | update_rq_clock(rq); | 913 | update_rq_clock(rq); |
916 | now = rq->clock; | ||
917 | 914 | ||
918 | if (p->state == TASK_UNINTERRUPTIBLE) | 915 | if (p->state == TASK_UNINTERRUPTIBLE) |
919 | rq->nr_uninterruptible--; | 916 | rq->nr_uninterruptible--; |
@@ -927,10 +924,7 @@ static void activate_task(struct rq *rq, struct task_struct *p, int wakeup) | |||
927 | */ | 924 | */ |
928 | static inline void activate_idle_task(struct task_struct *p, struct rq *rq) | 925 | static inline void activate_idle_task(struct task_struct *p, struct rq *rq) |
929 | { | 926 | { |
930 | u64 now; | ||
931 | |||
932 | update_rq_clock(rq); | 927 | update_rq_clock(rq); |
933 | now = rq->clock; | ||
934 | 928 | ||
935 | if (p->state == TASK_UNINTERRUPTIBLE) | 929 | if (p->state == TASK_UNINTERRUPTIBLE) |
936 | rq->nr_uninterruptible--; | 930 | rq->nr_uninterruptible--; |
@@ -1647,13 +1641,11 @@ void fastcall wake_up_new_task(struct task_struct *p, unsigned long clone_flags) | |||
1647 | unsigned long flags; | 1641 | unsigned long flags; |
1648 | struct rq *rq; | 1642 | struct rq *rq; |
1649 | int this_cpu; | 1643 | int this_cpu; |
1650 | u64 now; | ||
1651 | 1644 | ||
1652 | rq = task_rq_lock(p, &flags); | 1645 | rq = task_rq_lock(p, &flags); |
1653 | BUG_ON(p->state != TASK_RUNNING); | 1646 | BUG_ON(p->state != TASK_RUNNING); |
1654 | this_cpu = smp_processor_id(); /* parent's CPU */ | 1647 | this_cpu = smp_processor_id(); /* parent's CPU */ |
1655 | update_rq_clock(rq); | 1648 | update_rq_clock(rq); |
1656 | now = rq->clock; | ||
1657 | 1649 | ||
1658 | p->prio = effective_prio(p); | 1650 | p->prio = effective_prio(p); |
1659 | 1651 | ||
@@ -1955,11 +1947,9 @@ static void update_cpu_load(struct rq *this_rq) | |||
1955 | unsigned long total_load = this_rq->ls.load.weight; | 1947 | unsigned long total_load = this_rq->ls.load.weight; |
1956 | unsigned long this_load = total_load; | 1948 | unsigned long this_load = total_load; |
1957 | struct load_stat *ls = &this_rq->ls; | 1949 | struct load_stat *ls = &this_rq->ls; |
1958 | u64 now; | ||
1959 | int i, scale; | 1950 | int i, scale; |
1960 | 1951 | ||
1961 | __update_rq_clock(this_rq); | 1952 | __update_rq_clock(this_rq); |
1962 | now = this_rq->clock; | ||
1963 | 1953 | ||
1964 | this_rq->nr_load_updates++; | 1954 | this_rq->nr_load_updates++; |
1965 | if (unlikely(!(sysctl_sched_features & SCHED_FEAT_PRECISE_CPU_LOAD))) | 1955 | if (unlikely(!(sysctl_sched_features & SCHED_FEAT_PRECISE_CPU_LOAD))) |
@@ -3431,7 +3421,6 @@ asmlinkage void __sched schedule(void) | |||
3431 | struct task_struct *prev, *next; | 3421 | struct task_struct *prev, *next; |
3432 | long *switch_count; | 3422 | long *switch_count; |
3433 | struct rq *rq; | 3423 | struct rq *rq; |
3434 | u64 now; | ||
3435 | int cpu; | 3424 | int cpu; |
3436 | 3425 | ||
3437 | need_resched: | 3426 | need_resched: |
@@ -3450,7 +3439,6 @@ need_resched_nonpreemptible: | |||
3450 | spin_lock_irq(&rq->lock); | 3439 | spin_lock_irq(&rq->lock); |
3451 | clear_tsk_need_resched(prev); | 3440 | clear_tsk_need_resched(prev); |
3452 | __update_rq_clock(rq); | 3441 | __update_rq_clock(rq); |
3453 | now = rq->clock; | ||
3454 | 3442 | ||
3455 | if (prev->state && !(preempt_count() & PREEMPT_ACTIVE)) { | 3443 | if (prev->state && !(preempt_count() & PREEMPT_ACTIVE)) { |
3456 | if (unlikely((prev->state & TASK_INTERRUPTIBLE) && | 3444 | if (unlikely((prev->state & TASK_INTERRUPTIBLE) && |
@@ -3909,13 +3897,11 @@ void rt_mutex_setprio(struct task_struct *p, int prio) | |||
3909 | unsigned long flags; | 3897 | unsigned long flags; |
3910 | int oldprio, on_rq; | 3898 | int oldprio, on_rq; |
3911 | struct rq *rq; | 3899 | struct rq *rq; |
3912 | u64 now; | ||
3913 | 3900 | ||
3914 | BUG_ON(prio < 0 || prio > MAX_PRIO); | 3901 | BUG_ON(prio < 0 || prio > MAX_PRIO); |
3915 | 3902 | ||
3916 | rq = task_rq_lock(p, &flags); | 3903 | rq = task_rq_lock(p, &flags); |
3917 | update_rq_clock(rq); | 3904 | update_rq_clock(rq); |
3918 | now = rq->clock; | ||
3919 | 3905 | ||
3920 | oldprio = p->prio; | 3906 | oldprio = p->prio; |
3921 | on_rq = p->se.on_rq; | 3907 | on_rq = p->se.on_rq; |
@@ -3953,7 +3939,6 @@ void set_user_nice(struct task_struct *p, long nice) | |||
3953 | int old_prio, delta, on_rq; | 3939 | int old_prio, delta, on_rq; |
3954 | unsigned long flags; | 3940 | unsigned long flags; |
3955 | struct rq *rq; | 3941 | struct rq *rq; |
3956 | u64 now; | ||
3957 | 3942 | ||
3958 | if (TASK_NICE(p) == nice || nice < -20 || nice > 19) | 3943 | if (TASK_NICE(p) == nice || nice < -20 || nice > 19) |
3959 | return; | 3944 | return; |
@@ -3963,7 +3948,6 @@ void set_user_nice(struct task_struct *p, long nice) | |||
3963 | */ | 3948 | */ |
3964 | rq = task_rq_lock(p, &flags); | 3949 | rq = task_rq_lock(p, &flags); |
3965 | update_rq_clock(rq); | 3950 | update_rq_clock(rq); |
3966 | now = rq->clock; | ||
3967 | /* | 3951 | /* |
3968 | * The RT priorities are set via sched_setscheduler(), but we still | 3952 | * The RT priorities are set via sched_setscheduler(), but we still |
3969 | * allow the 'normal' nice value to be set - but as expected | 3953 | * allow the 'normal' nice value to be set - but as expected |
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c index 4a2cbde1057..eb7ca49c326 100644 --- a/kernel/sched_fair.c +++ b/kernel/sched_fair.c | |||
@@ -667,10 +667,8 @@ static void entity_tick(struct cfs_rq *cfs_rq, struct sched_entity *curr) | |||
667 | { | 667 | { |
668 | struct rq *rq = rq_of(cfs_rq); | 668 | struct rq *rq = rq_of(cfs_rq); |
669 | struct sched_entity *next; | 669 | struct sched_entity *next; |
670 | u64 now; | ||
671 | 670 | ||
672 | __update_rq_clock(rq); | 671 | __update_rq_clock(rq); |
673 | now = rq->clock; | ||
674 | 672 | ||
675 | /* | 673 | /* |
676 | * Dequeue and enqueue the task to update its | 674 | * Dequeue and enqueue the task to update its |
@@ -820,10 +818,8 @@ static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int sleep) | |||
820 | static void yield_task_fair(struct rq *rq, struct task_struct *p) | 818 | static void yield_task_fair(struct rq *rq, struct task_struct *p) |
821 | { | 819 | { |
822 | struct cfs_rq *cfs_rq = task_cfs_rq(p); | 820 | struct cfs_rq *cfs_rq = task_cfs_rq(p); |
823 | u64 now; | ||
824 | 821 | ||
825 | __update_rq_clock(rq); | 822 | __update_rq_clock(rq); |
826 | now = rq->clock; | ||
827 | /* | 823 | /* |
828 | * Dequeue and enqueue the task to update its | 824 | * Dequeue and enqueue the task to update its |
829 | * position within the tree: | 825 | * position within the tree: |
@@ -1062,11 +1058,9 @@ static void set_curr_task_fair(struct rq *rq) | |||
1062 | { | 1058 | { |
1063 | struct task_struct *curr = rq->curr; | 1059 | struct task_struct *curr = rq->curr; |
1064 | struct sched_entity *se = &curr->se; | 1060 | struct sched_entity *se = &curr->se; |
1065 | u64 now; | ||
1066 | struct cfs_rq *cfs_rq; | 1061 | struct cfs_rq *cfs_rq; |
1067 | 1062 | ||
1068 | update_rq_clock(rq); | 1063 | update_rq_clock(rq); |
1069 | now = rq->clock; | ||
1070 | 1064 | ||
1071 | for_each_sched_entity(se) { | 1065 | for_each_sched_entity(se) { |
1072 | cfs_rq = cfs_rq_of(se); | 1066 | cfs_rq = cfs_rq_of(se); |