aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/sched/deadline.c
diff options
context:
space:
mode:
authorJuri Lelli <juri.lelli@arm.com>2014-09-19 05:22:41 -0400
committerIngo Molnar <mingo@kernel.org>2014-09-24 08:46:57 -0400
commit91ec6778ec4f963fcb2c2793610919b572f633b0 (patch)
tree5f83a51e7a10de536ad1b83fecf3e5d4344d0d17 /kernel/sched/deadline.c
parenta5e7be3b28a235108c59561bea55eea1072b23b0 (diff)
sched/deadline: Fix inter- exclusive cpusets migrations
Users can perform clustered scheduling using the cpuset facility. After an exclusive cpuset is created, task migrations happen only between CPUs belonging to the same cpuset. Inter- cpuset migrations can only happen when the user requires so, moving a task between different cpusets. This behaviour is broken in SCHED_DEADLINE, as currently spurious inter- cpuset migration may happen without user intervention. This patch fix the problem (and shuffles the code a bit to improve clarity). Signed-off-by: Juri Lelli <juri.lelli@arm.com> Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Cc: raistlin@linux.it Cc: michael@amarulasolutions.com Cc: fchecconi@gmail.com Cc: daniel.wagner@bmw-carit.de Cc: vincent@legout.info Cc: luca.abeni@unitn.it Cc: Linus Torvalds <torvalds@linux-foundation.org> Link: http://lkml.kernel.org/r/1411118561-26323-4-git-send-email-juri.lelli@arm.com Signed-off-by: Ingo Molnar <mingo@kernel.org>
Diffstat (limited to 'kernel/sched/deadline.c')
-rw-r--r--kernel/sched/deadline.c7
1 files changed, 7 insertions, 0 deletions
diff --git a/kernel/sched/deadline.c b/kernel/sched/deadline.c
index efb94124420d..abfaf3d9a29f 100644
--- a/kernel/sched/deadline.c
+++ b/kernel/sched/deadline.c
@@ -1164,6 +1164,13 @@ static int find_later_rq(struct task_struct *task)
1164 if (task->nr_cpus_allowed == 1) 1164 if (task->nr_cpus_allowed == 1)
1165 return -1; 1165 return -1;
1166 1166
1167 /*
1168 * We have to consider system topology and task affinity
1169 * first, then we can look for a suitable cpu.
1170 */
1171 cpumask_copy(later_mask, task_rq(task)->rd->span);
1172 cpumask_and(later_mask, later_mask, cpu_active_mask);
1173 cpumask_and(later_mask, later_mask, &task->cpus_allowed);
1167 best_cpu = cpudl_find(&task_rq(task)->rd->cpudl, 1174 best_cpu = cpudl_find(&task_rq(task)->rd->cpudl,
1168 task, later_mask); 1175 task, later_mask);
1169 if (best_cpu == -1) 1176 if (best_cpu == -1)