diff options
author | Peter Zijlstra <a.p.zijlstra@chello.nl> | 2008-02-13 09:45:40 -0500 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2008-02-13 09:45:40 -0500 |
commit | 052f1dc7eb02300b05170ae341ccd03b76207778 (patch) | |
tree | f58630b7876ae9e4308c0577e36aa13318b7bcfc /kernel/sched_rt.c | |
parent | 9f0c1e560c43327b70998e6c702b2f01321130d9 (diff) |
sched: rt-group: make rt groups scheduling configurable
Make the rt group scheduler compile time configurable.
Keep it experimental for now.
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/sched_rt.c')
-rw-r--r-- | kernel/sched_rt.c | 12 |
1 files changed, 6 insertions, 6 deletions
diff --git a/kernel/sched_rt.c b/kernel/sched_rt.c index 35825b28e429..f54792b175b2 100644 --- a/kernel/sched_rt.c +++ b/kernel/sched_rt.c | |||
@@ -55,7 +55,7 @@ static inline int on_rt_rq(struct sched_rt_entity *rt_se) | |||
55 | return !list_empty(&rt_se->run_list); | 55 | return !list_empty(&rt_se->run_list); |
56 | } | 56 | } |
57 | 57 | ||
58 | #ifdef CONFIG_FAIR_GROUP_SCHED | 58 | #ifdef CONFIG_RT_GROUP_SCHED |
59 | 59 | ||
60 | static inline u64 sched_rt_runtime(struct rt_rq *rt_rq) | 60 | static inline u64 sched_rt_runtime(struct rt_rq *rt_rq) |
61 | { | 61 | { |
@@ -177,7 +177,7 @@ static inline int rt_rq_throttled(struct rt_rq *rt_rq) | |||
177 | 177 | ||
178 | static inline int rt_se_prio(struct sched_rt_entity *rt_se) | 178 | static inline int rt_se_prio(struct sched_rt_entity *rt_se) |
179 | { | 179 | { |
180 | #ifdef CONFIG_FAIR_GROUP_SCHED | 180 | #ifdef CONFIG_RT_GROUP_SCHED |
181 | struct rt_rq *rt_rq = group_rt_rq(rt_se); | 181 | struct rt_rq *rt_rq = group_rt_rq(rt_se); |
182 | 182 | ||
183 | if (rt_rq) | 183 | if (rt_rq) |
@@ -269,7 +269,7 @@ void inc_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) | |||
269 | { | 269 | { |
270 | WARN_ON(!rt_prio(rt_se_prio(rt_se))); | 270 | WARN_ON(!rt_prio(rt_se_prio(rt_se))); |
271 | rt_rq->rt_nr_running++; | 271 | rt_rq->rt_nr_running++; |
272 | #if defined CONFIG_SMP || defined CONFIG_FAIR_GROUP_SCHED | 272 | #if defined CONFIG_SMP || defined CONFIG_RT_GROUP_SCHED |
273 | if (rt_se_prio(rt_se) < rt_rq->highest_prio) | 273 | if (rt_se_prio(rt_se) < rt_rq->highest_prio) |
274 | rt_rq->highest_prio = rt_se_prio(rt_se); | 274 | rt_rq->highest_prio = rt_se_prio(rt_se); |
275 | #endif | 275 | #endif |
@@ -281,7 +281,7 @@ void inc_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) | |||
281 | 281 | ||
282 | update_rt_migration(rq_of_rt_rq(rt_rq)); | 282 | update_rt_migration(rq_of_rt_rq(rt_rq)); |
283 | #endif | 283 | #endif |
284 | #ifdef CONFIG_FAIR_GROUP_SCHED | 284 | #ifdef CONFIG_RT_GROUP_SCHED |
285 | if (rt_se_boosted(rt_se)) | 285 | if (rt_se_boosted(rt_se)) |
286 | rt_rq->rt_nr_boosted++; | 286 | rt_rq->rt_nr_boosted++; |
287 | #endif | 287 | #endif |
@@ -293,7 +293,7 @@ void dec_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) | |||
293 | WARN_ON(!rt_prio(rt_se_prio(rt_se))); | 293 | WARN_ON(!rt_prio(rt_se_prio(rt_se))); |
294 | WARN_ON(!rt_rq->rt_nr_running); | 294 | WARN_ON(!rt_rq->rt_nr_running); |
295 | rt_rq->rt_nr_running--; | 295 | rt_rq->rt_nr_running--; |
296 | #if defined CONFIG_SMP || defined CONFIG_FAIR_GROUP_SCHED | 296 | #if defined CONFIG_SMP || defined CONFIG_RT_GROUP_SCHED |
297 | if (rt_rq->rt_nr_running) { | 297 | if (rt_rq->rt_nr_running) { |
298 | struct rt_prio_array *array; | 298 | struct rt_prio_array *array; |
299 | 299 | ||
@@ -315,7 +315,7 @@ void dec_rt_tasks(struct sched_rt_entity *rt_se, struct rt_rq *rt_rq) | |||
315 | 315 | ||
316 | update_rt_migration(rq_of_rt_rq(rt_rq)); | 316 | update_rt_migration(rq_of_rt_rq(rt_rq)); |
317 | #endif /* CONFIG_SMP */ | 317 | #endif /* CONFIG_SMP */ |
318 | #ifdef CONFIG_FAIR_GROUP_SCHED | 318 | #ifdef CONFIG_RT_GROUP_SCHED |
319 | if (rt_se_boosted(rt_se)) | 319 | if (rt_se_boosted(rt_se)) |
320 | rt_rq->rt_nr_boosted--; | 320 | rt_rq->rt_nr_boosted--; |
321 | 321 | ||