diff options
author | Gregory Haskins <ghaskins@novell.com> | 2008-06-27 16:30:00 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2008-07-04 06:50:22 -0400 |
commit | 2087a1ad822cd3a68b73338457047fcc54da726b (patch) | |
tree | 1ebc45b2499034ffe37b29bf7a748b8e992cd870 /kernel/sched.c | |
parent | c4acb2c0669c5c5c9b28e9d02a34b5c67edf7092 (diff) |
sched: add avg-overlap support to RT tasks
We have the notion of tracking process-coupling (a.k.a. buddy-wake) via
the p->se.last_wake / p->se.avg_overlap facilities, but it is only used
for cfs to cfs interactions. There is no reason why an rt to cfs
interaction cannot share in establishing a relationhip in a similar
manner.
Because PREEMPT_RT runs many kernel threads as FIFO priority, we often
times have heavy interaction between RT threads waking CFS applications.
This patch offers a substantial boost (50-60%+) in perfomance under those
circumstances.
Signed-off-by: Gregory Haskins <ghaskins@novell.com>
Cc: npiggin@suse.de
Cc: rostedt@goodmis.org
Acked-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/sched.c')
-rw-r--r-- | kernel/sched.c | 14 |
1 files changed, 14 insertions, 0 deletions
diff --git a/kernel/sched.c b/kernel/sched.c index d99aeabeb72f..bbc40c3a0657 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -1693,6 +1693,12 @@ static void set_load_weight(struct task_struct *p) | |||
1693 | p->se.load.inv_weight = prio_to_wmult[p->static_prio - MAX_RT_PRIO]; | 1693 | p->se.load.inv_weight = prio_to_wmult[p->static_prio - MAX_RT_PRIO]; |
1694 | } | 1694 | } |
1695 | 1695 | ||
1696 | static void update_avg(u64 *avg, u64 sample) | ||
1697 | { | ||
1698 | s64 diff = sample - *avg; | ||
1699 | *avg += diff >> 3; | ||
1700 | } | ||
1701 | |||
1696 | static void enqueue_task(struct rq *rq, struct task_struct *p, int wakeup) | 1702 | static void enqueue_task(struct rq *rq, struct task_struct *p, int wakeup) |
1697 | { | 1703 | { |
1698 | sched_info_queued(p); | 1704 | sched_info_queued(p); |
@@ -1702,6 +1708,12 @@ static void enqueue_task(struct rq *rq, struct task_struct *p, int wakeup) | |||
1702 | 1708 | ||
1703 | static void dequeue_task(struct rq *rq, struct task_struct *p, int sleep) | 1709 | static void dequeue_task(struct rq *rq, struct task_struct *p, int sleep) |
1704 | { | 1710 | { |
1711 | if (sleep && p->se.last_wakeup) { | ||
1712 | update_avg(&p->se.avg_overlap, | ||
1713 | p->se.sum_exec_runtime - p->se.last_wakeup); | ||
1714 | p->se.last_wakeup = 0; | ||
1715 | } | ||
1716 | |||
1705 | p->sched_class->dequeue_task(rq, p, sleep); | 1717 | p->sched_class->dequeue_task(rq, p, sleep); |
1706 | p->se.on_rq = 0; | 1718 | p->se.on_rq = 0; |
1707 | } | 1719 | } |
@@ -2313,6 +2325,8 @@ out_running: | |||
2313 | p->sched_class->task_wake_up(rq, p); | 2325 | p->sched_class->task_wake_up(rq, p); |
2314 | #endif | 2326 | #endif |
2315 | out: | 2327 | out: |
2328 | current->se.last_wakeup = current->se.sum_exec_runtime; | ||
2329 | |||
2316 | task_rq_unlock(rq, &flags); | 2330 | task_rq_unlock(rq, &flags); |
2317 | 2331 | ||
2318 | return success; | 2332 | return success; |