diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2012-03-29 17:46:05 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2012-03-29 17:46:05 -0400 |
commit | 7fda0412c5f7afdd1a5ff518f98dee5157266d8a (patch) | |
tree | d312af46758fa9b59431a479d258b54184a00591 /kernel/sched | |
parent | 6b8212a313dae341ef3a2e413dfec5c4dea59617 (diff) | |
parent | 160594e99dbbb0a5600ad922c630952c7c1c14bf (diff) |
Merge branch 'sched-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
Pull scheduler fixes from Ingo Molnar.
* 'sched-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip:
cpusets: Remove an unused variable
sched/rt: Improve pick_next_highest_task_rt()
sched: Fix select_fallback_rq() vs cpu_active/cpu_online
sched/x86/smp: Do not enable IRQs over calibrate_delay()
sched: Fix compiler warning about declared inline after use
MAINTAINERS: Update email address for SCHEDULER and PERF EVENTS
Diffstat (limited to 'kernel/sched')
-rw-r--r-- | kernel/sched/core.c | 62 | ||||
-rw-r--r-- | kernel/sched/fair.c | 16 | ||||
-rw-r--r-- | kernel/sched/rt.c | 2 |
3 files changed, 55 insertions, 25 deletions
diff --git a/kernel/sched/core.c b/kernel/sched/core.c index 157fb9b2b186..e3ed0ecee7c7 100644 --- a/kernel/sched/core.c +++ b/kernel/sched/core.c | |||
@@ -1265,29 +1265,59 @@ EXPORT_SYMBOL_GPL(kick_process); | |||
1265 | */ | 1265 | */ |
1266 | static int select_fallback_rq(int cpu, struct task_struct *p) | 1266 | static int select_fallback_rq(int cpu, struct task_struct *p) |
1267 | { | 1267 | { |
1268 | int dest_cpu; | ||
1269 | const struct cpumask *nodemask = cpumask_of_node(cpu_to_node(cpu)); | 1268 | const struct cpumask *nodemask = cpumask_of_node(cpu_to_node(cpu)); |
1269 | enum { cpuset, possible, fail } state = cpuset; | ||
1270 | int dest_cpu; | ||
1270 | 1271 | ||
1271 | /* Look for allowed, online CPU in same node. */ | 1272 | /* Look for allowed, online CPU in same node. */ |
1272 | for_each_cpu_and(dest_cpu, nodemask, cpu_active_mask) | 1273 | for_each_cpu_mask(dest_cpu, *nodemask) { |
1274 | if (!cpu_online(dest_cpu)) | ||
1275 | continue; | ||
1276 | if (!cpu_active(dest_cpu)) | ||
1277 | continue; | ||
1273 | if (cpumask_test_cpu(dest_cpu, tsk_cpus_allowed(p))) | 1278 | if (cpumask_test_cpu(dest_cpu, tsk_cpus_allowed(p))) |
1274 | return dest_cpu; | 1279 | return dest_cpu; |
1280 | } | ||
1275 | 1281 | ||
1276 | /* Any allowed, online CPU? */ | 1282 | for (;;) { |
1277 | dest_cpu = cpumask_any_and(tsk_cpus_allowed(p), cpu_active_mask); | 1283 | /* Any allowed, online CPU? */ |
1278 | if (dest_cpu < nr_cpu_ids) | 1284 | for_each_cpu_mask(dest_cpu, *tsk_cpus_allowed(p)) { |
1279 | return dest_cpu; | 1285 | if (!cpu_online(dest_cpu)) |
1286 | continue; | ||
1287 | if (!cpu_active(dest_cpu)) | ||
1288 | continue; | ||
1289 | goto out; | ||
1290 | } | ||
1280 | 1291 | ||
1281 | /* No more Mr. Nice Guy. */ | 1292 | switch (state) { |
1282 | dest_cpu = cpuset_cpus_allowed_fallback(p); | 1293 | case cpuset: |
1283 | /* | 1294 | /* No more Mr. Nice Guy. */ |
1284 | * Don't tell them about moving exiting tasks or | 1295 | cpuset_cpus_allowed_fallback(p); |
1285 | * kernel threads (both mm NULL), since they never | 1296 | state = possible; |
1286 | * leave kernel. | 1297 | break; |
1287 | */ | 1298 | |
1288 | if (p->mm && printk_ratelimit()) { | 1299 | case possible: |
1289 | printk_sched("process %d (%s) no longer affine to cpu%d\n", | 1300 | do_set_cpus_allowed(p, cpu_possible_mask); |
1290 | task_pid_nr(p), p->comm, cpu); | 1301 | state = fail; |
1302 | break; | ||
1303 | |||
1304 | case fail: | ||
1305 | BUG(); | ||
1306 | break; | ||
1307 | } | ||
1308 | } | ||
1309 | |||
1310 | out: | ||
1311 | if (state != cpuset) { | ||
1312 | /* | ||
1313 | * Don't tell them about moving exiting tasks or | ||
1314 | * kernel threads (both mm NULL), since they never | ||
1315 | * leave kernel. | ||
1316 | */ | ||
1317 | if (p->mm && printk_ratelimit()) { | ||
1318 | printk_sched("process %d (%s) no longer affine to cpu%d\n", | ||
1319 | task_pid_nr(p), p->comm, cpu); | ||
1320 | } | ||
1291 | } | 1321 | } |
1292 | 1322 | ||
1293 | return dest_cpu; | 1323 | return dest_cpu; |
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 94340c7544a9..0d97ebdc58f0 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c | |||
@@ -416,8 +416,8 @@ find_matching_se(struct sched_entity **se, struct sched_entity **pse) | |||
416 | 416 | ||
417 | #endif /* CONFIG_FAIR_GROUP_SCHED */ | 417 | #endif /* CONFIG_FAIR_GROUP_SCHED */ |
418 | 418 | ||
419 | static void account_cfs_rq_runtime(struct cfs_rq *cfs_rq, | 419 | static __always_inline |
420 | unsigned long delta_exec); | 420 | void account_cfs_rq_runtime(struct cfs_rq *cfs_rq, unsigned long delta_exec); |
421 | 421 | ||
422 | /************************************************************** | 422 | /************************************************************** |
423 | * Scheduling class tree data structure manipulation methods: | 423 | * Scheduling class tree data structure manipulation methods: |
@@ -1162,7 +1162,7 @@ static void clear_buddies(struct cfs_rq *cfs_rq, struct sched_entity *se) | |||
1162 | __clear_buddies_skip(se); | 1162 | __clear_buddies_skip(se); |
1163 | } | 1163 | } |
1164 | 1164 | ||
1165 | static void return_cfs_rq_runtime(struct cfs_rq *cfs_rq); | 1165 | static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq); |
1166 | 1166 | ||
1167 | static void | 1167 | static void |
1168 | dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) | 1168 | dequeue_entity(struct cfs_rq *cfs_rq, struct sched_entity *se, int flags) |
@@ -1546,8 +1546,8 @@ static void __account_cfs_rq_runtime(struct cfs_rq *cfs_rq, | |||
1546 | resched_task(rq_of(cfs_rq)->curr); | 1546 | resched_task(rq_of(cfs_rq)->curr); |
1547 | } | 1547 | } |
1548 | 1548 | ||
1549 | static __always_inline void account_cfs_rq_runtime(struct cfs_rq *cfs_rq, | 1549 | static __always_inline |
1550 | unsigned long delta_exec) | 1550 | void account_cfs_rq_runtime(struct cfs_rq *cfs_rq, unsigned long delta_exec) |
1551 | { | 1551 | { |
1552 | if (!cfs_bandwidth_used() || !cfs_rq->runtime_enabled) | 1552 | if (!cfs_bandwidth_used() || !cfs_rq->runtime_enabled) |
1553 | return; | 1553 | return; |
@@ -2073,11 +2073,11 @@ void unthrottle_offline_cfs_rqs(struct rq *rq) | |||
2073 | } | 2073 | } |
2074 | 2074 | ||
2075 | #else /* CONFIG_CFS_BANDWIDTH */ | 2075 | #else /* CONFIG_CFS_BANDWIDTH */ |
2076 | static void account_cfs_rq_runtime(struct cfs_rq *cfs_rq, | 2076 | static __always_inline |
2077 | unsigned long delta_exec) {} | 2077 | void account_cfs_rq_runtime(struct cfs_rq *cfs_rq, unsigned long delta_exec) {} |
2078 | static void check_cfs_rq_runtime(struct cfs_rq *cfs_rq) {} | 2078 | static void check_cfs_rq_runtime(struct cfs_rq *cfs_rq) {} |
2079 | static void check_enqueue_throttle(struct cfs_rq *cfs_rq) {} | 2079 | static void check_enqueue_throttle(struct cfs_rq *cfs_rq) {} |
2080 | static void return_cfs_rq_runtime(struct cfs_rq *cfs_rq) {} | 2080 | static __always_inline void return_cfs_rq_runtime(struct cfs_rq *cfs_rq) {} |
2081 | 2081 | ||
2082 | static inline int cfs_rq_throttled(struct cfs_rq *cfs_rq) | 2082 | static inline int cfs_rq_throttled(struct cfs_rq *cfs_rq) |
2083 | { | 2083 | { |
diff --git a/kernel/sched/rt.c b/kernel/sched/rt.c index b60dad720173..44af55e6d5d0 100644 --- a/kernel/sched/rt.c +++ b/kernel/sched/rt.c | |||
@@ -1428,7 +1428,7 @@ static struct task_struct *pick_next_highest_task_rt(struct rq *rq, int cpu) | |||
1428 | next_idx: | 1428 | next_idx: |
1429 | if (idx >= MAX_RT_PRIO) | 1429 | if (idx >= MAX_RT_PRIO) |
1430 | continue; | 1430 | continue; |
1431 | if (next && next->prio < idx) | 1431 | if (next && next->prio <= idx) |
1432 | continue; | 1432 | continue; |
1433 | list_for_each_entry(rt_se, array->queue + idx, run_list) { | 1433 | list_for_each_entry(rt_se, array->queue + idx, run_list) { |
1434 | struct task_struct *p; | 1434 | struct task_struct *p; |