diff options
| -rw-r--r-- | kernel/sched.c | 37 | ||||
| -rw-r--r-- | kernel/sched_fair.c | 6 | ||||
| -rw-r--r-- | kernel/sched_stoptask.c | 2 |
3 files changed, 27 insertions, 18 deletions
diff --git a/kernel/sched.c b/kernel/sched.c index 36a088018fe0..dc91a4d09ac3 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
| @@ -560,18 +560,8 @@ struct rq { | |||
| 560 | 560 | ||
| 561 | static DEFINE_PER_CPU_SHARED_ALIGNED(struct rq, runqueues); | 561 | static DEFINE_PER_CPU_SHARED_ALIGNED(struct rq, runqueues); |
| 562 | 562 | ||
| 563 | static inline | ||
| 564 | void check_preempt_curr(struct rq *rq, struct task_struct *p, int flags) | ||
| 565 | { | ||
| 566 | rq->curr->sched_class->check_preempt_curr(rq, p, flags); | ||
| 567 | 563 | ||
| 568 | /* | 564 | static void check_preempt_curr(struct rq *rq, struct task_struct *p, int flags); |
| 569 | * A queue event has occurred, and we're going to schedule. In | ||
| 570 | * this case, we can save a useless back to back clock update. | ||
| 571 | */ | ||
| 572 | if (test_tsk_need_resched(p)) | ||
| 573 | rq->skip_clock_update = 1; | ||
| 574 | } | ||
| 575 | 565 | ||
| 576 | static inline int cpu_of(struct rq *rq) | 566 | static inline int cpu_of(struct rq *rq) |
| 577 | { | 567 | { |
| @@ -2118,6 +2108,31 @@ static inline void check_class_changed(struct rq *rq, struct task_struct *p, | |||
| 2118 | p->sched_class->prio_changed(rq, p, oldprio, running); | 2108 | p->sched_class->prio_changed(rq, p, oldprio, running); |
| 2119 | } | 2109 | } |
| 2120 | 2110 | ||
| 2111 | static void check_preempt_curr(struct rq *rq, struct task_struct *p, int flags) | ||
| 2112 | { | ||
| 2113 | const struct sched_class *class; | ||
| 2114 | |||
| 2115 | if (p->sched_class == rq->curr->sched_class) { | ||
| 2116 | rq->curr->sched_class->check_preempt_curr(rq, p, flags); | ||
| 2117 | } else { | ||
| 2118 | for_each_class(class) { | ||
| 2119 | if (class == rq->curr->sched_class) | ||
| 2120 | break; | ||
| 2121 | if (class == p->sched_class) { | ||
| 2122 | resched_task(rq->curr); | ||
| 2123 | break; | ||
| 2124 | } | ||
| 2125 | } | ||
| 2126 | } | ||
| 2127 | |||
| 2128 | /* | ||
| 2129 | * A queue event has occurred, and we're going to schedule. In | ||
| 2130 | * this case, we can save a useless back to back clock update. | ||
| 2131 | */ | ||
| 2132 | if (test_tsk_need_resched(rq->curr)) | ||
| 2133 | rq->skip_clock_update = 1; | ||
| 2134 | } | ||
| 2135 | |||
| 2121 | #ifdef CONFIG_SMP | 2136 | #ifdef CONFIG_SMP |
| 2122 | /* | 2137 | /* |
| 2123 | * Is this task likely cache-hot: | 2138 | * Is this task likely cache-hot: |
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c index 034c4f410b36..52ab113d8bb9 100644 --- a/kernel/sched_fair.c +++ b/kernel/sched_fair.c | |||
| @@ -1654,12 +1654,6 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p, int wake_ | |||
| 1654 | struct cfs_rq *cfs_rq = task_cfs_rq(curr); | 1654 | struct cfs_rq *cfs_rq = task_cfs_rq(curr); |
| 1655 | int scale = cfs_rq->nr_running >= sched_nr_latency; | 1655 | int scale = cfs_rq->nr_running >= sched_nr_latency; |
| 1656 | 1656 | ||
| 1657 | if (unlikely(rt_prio(p->prio))) | ||
| 1658 | goto preempt; | ||
| 1659 | |||
| 1660 | if (unlikely(p->sched_class != &fair_sched_class)) | ||
| 1661 | return; | ||
| 1662 | |||
| 1663 | if (unlikely(se == pse)) | 1657 | if (unlikely(se == pse)) |
| 1664 | return; | 1658 | return; |
| 1665 | 1659 | ||
diff --git a/kernel/sched_stoptask.c b/kernel/sched_stoptask.c index 755483b2a2ad..2bf6b47058c1 100644 --- a/kernel/sched_stoptask.c +++ b/kernel/sched_stoptask.c | |||
| @@ -19,7 +19,7 @@ select_task_rq_stop(struct rq *rq, struct task_struct *p, | |||
| 19 | static void | 19 | static void |
| 20 | check_preempt_curr_stop(struct rq *rq, struct task_struct *p, int flags) | 20 | check_preempt_curr_stop(struct rq *rq, struct task_struct *p, int flags) |
| 21 | { | 21 | { |
| 22 | resched_task(rq->curr); /* we preempt everything */ | 22 | /* we're never preempted */ |
| 23 | } | 23 | } |
| 24 | 24 | ||
| 25 | static struct task_struct *pick_next_task_stop(struct rq *rq) | 25 | static struct task_struct *pick_next_task_stop(struct rq *rq) |
