aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/sched_fair.c
diff options
context:
space:
mode:
authorPeter Zijlstra <a.p.zijlstra@chello.nl>2008-04-19 13:44:57 -0400
committerIngo Molnar <mingo@elte.hu>2008-04-19 13:44:57 -0400
commit0bbd3336eee1e712a290e0dfd1a64cbbdd63a508 (patch)
treea44ea1d2a5bf1e9f2975bc55a88167b3fc2104fe /kernel/sched_fair.c
parent15934a37324f32e0fda633dc7984a671ea81cd75 (diff)
sched: fix wakeup granularity for buddies
The wakeup buddy logic didn't use the same wakeup granularity logic as the wakeup preemption did, this might cause the ->next buddy to be selected past the point where we would have preempted had the task been a single running instance. Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/sched_fair.c')
-rw-r--r--kernel/sched_fair.c69
1 files changed, 49 insertions, 20 deletions
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c
index 86a93376282c..b01f8e77f2ac 100644
--- a/kernel/sched_fair.c
+++ b/kernel/sched_fair.c
@@ -73,13 +73,13 @@ unsigned int sysctl_sched_batch_wakeup_granularity = 10000000UL;
73 73
74/* 74/*
75 * SCHED_OTHER wake-up granularity. 75 * SCHED_OTHER wake-up granularity.
76 * (default: 5 msec * (1 + ilog(ncpus)), units: nanoseconds) 76 * (default: 10 msec * (1 + ilog(ncpus)), units: nanoseconds)
77 * 77 *
78 * This option delays the preemption effects of decoupled workloads 78 * This option delays the preemption effects of decoupled workloads
79 * and reduces their over-scheduling. Synchronous workloads will still 79 * and reduces their over-scheduling. Synchronous workloads will still
80 * have immediate wakeup/sleep latencies. 80 * have immediate wakeup/sleep latencies.
81 */ 81 */
82unsigned int sysctl_sched_wakeup_granularity = 5000000UL; 82unsigned int sysctl_sched_wakeup_granularity = 10000000UL;
83 83
84const_debug unsigned int sysctl_sched_migration_cost = 500000UL; 84const_debug unsigned int sysctl_sched_migration_cost = 500000UL;
85 85
@@ -629,20 +629,16 @@ set_next_entity(struct cfs_rq *cfs_rq, struct sched_entity *se)
629 se->prev_sum_exec_runtime = se->sum_exec_runtime; 629 se->prev_sum_exec_runtime = se->sum_exec_runtime;
630} 630}
631 631
632static int
633wakeup_preempt_entity(struct sched_entity *curr, struct sched_entity *se);
634
632static struct sched_entity * 635static struct sched_entity *
633pick_next(struct cfs_rq *cfs_rq, struct sched_entity *se) 636pick_next(struct cfs_rq *cfs_rq, struct sched_entity *se)
634{ 637{
635 s64 diff, gran;
636
637 if (!cfs_rq->next) 638 if (!cfs_rq->next)
638 return se; 639 return se;
639 640
640 diff = cfs_rq->next->vruntime - se->vruntime; 641 if (wakeup_preempt_entity(cfs_rq->next, se) != 0)
641 if (diff < 0)
642 return se;
643
644 gran = calc_delta_fair(sysctl_sched_wakeup_granularity, &cfs_rq->load);
645 if (diff > gran)
646 return se; 642 return se;
647 643
648 return cfs_rq->next; 644 return cfs_rq->next;
@@ -1101,6 +1097,48 @@ out:
1101} 1097}
1102#endif /* CONFIG_SMP */ 1098#endif /* CONFIG_SMP */
1103 1099
1100static unsigned long wakeup_gran(struct sched_entity *se)
1101{
1102 unsigned long gran = sysctl_sched_wakeup_granularity;
1103
1104 /*
1105 * More easily preempt - nice tasks, while not making
1106 * it harder for + nice tasks.
1107 */
1108 if (unlikely(se->load.weight > NICE_0_LOAD))
1109 gran = calc_delta_fair(gran, &se->load);
1110
1111 return gran;
1112}
1113
1114/*
1115 * Should 'se' preempt 'curr'.
1116 *
1117 * |s1
1118 * |s2
1119 * |s3
1120 * g
1121 * |<--->|c
1122 *
1123 * w(c, s1) = -1
1124 * w(c, s2) = 0
1125 * w(c, s3) = 1
1126 *
1127 */
1128static int
1129wakeup_preempt_entity(struct sched_entity *curr, struct sched_entity *se)
1130{
1131 s64 gran, vdiff = curr->vruntime - se->vruntime;
1132
1133 if (vdiff < 0)
1134 return -1;
1135
1136 gran = wakeup_gran(curr);
1137 if (vdiff > gran)
1138 return 1;
1139
1140 return 0;
1141}
1104 1142
1105/* 1143/*
1106 * Preempt the current task with a newly woken task if needed: 1144 * Preempt the current task with a newly woken task if needed:
@@ -1110,7 +1148,6 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p)
1110 struct task_struct *curr = rq->curr; 1148 struct task_struct *curr = rq->curr;
1111 struct cfs_rq *cfs_rq = task_cfs_rq(curr); 1149 struct cfs_rq *cfs_rq = task_cfs_rq(curr);
1112 struct sched_entity *se = &curr->se, *pse = &p->se; 1150 struct sched_entity *se = &curr->se, *pse = &p->se;
1113 unsigned long gran;
1114 1151
1115 if (unlikely(rt_prio(p->prio))) { 1152 if (unlikely(rt_prio(p->prio))) {
1116 update_rq_clock(rq); 1153 update_rq_clock(rq);
@@ -1140,15 +1177,7 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p)
1140 pse = parent_entity(pse); 1177 pse = parent_entity(pse);
1141 } 1178 }
1142 1179
1143 gran = sysctl_sched_wakeup_granularity; 1180 if (wakeup_preempt_entity(se, pse) == 1)
1144 /*
1145 * More easily preempt - nice tasks, while not making
1146 * it harder for + nice tasks.
1147 */
1148 if (unlikely(se->load.weight > NICE_0_LOAD))
1149 gran = calc_delta_fair(gran, &se->load);
1150
1151 if (pse->vruntime + gran < se->vruntime)
1152 resched_task(curr); 1181 resched_task(curr);
1153} 1182}
1154 1183