aboutsummaryrefslogtreecommitdiffstats
path: root/kernel
diff options
context:
space:
mode:
Diffstat (limited to 'kernel')
-rw-r--r--kernel/sched.c45
1 files changed, 23 insertions, 22 deletions
diff --git a/kernel/sched.c b/kernel/sched.c
index 33d7965f63f0..63e55ac242d1 100644
--- a/kernel/sched.c
+++ b/kernel/sched.c
@@ -2322,7 +2322,7 @@ void task_oncpu_function_call(struct task_struct *p,
2322 * 2322 *
2323 * - fork, @p is stable because it isn't on the tasklist yet 2323 * - fork, @p is stable because it isn't on the tasklist yet
2324 * 2324 *
2325 * - exec, @p is unstable XXX 2325 * - exec, @p is unstable, retry loop
2326 * 2326 *
2327 * - wake-up, we serialize ->cpus_allowed against TASK_WAKING so 2327 * - wake-up, we serialize ->cpus_allowed against TASK_WAKING so
2328 * we should be good. 2328 * we should be good.
@@ -3132,21 +3132,36 @@ static void double_rq_unlock(struct rq *rq1, struct rq *rq2)
3132} 3132}
3133 3133
3134/* 3134/*
3135 * If dest_cpu is allowed for this process, migrate the task to it. 3135 * sched_exec - execve() is a valuable balancing opportunity, because at
3136 * This is accomplished by forcing the cpu_allowed mask to only 3136 * this point the task has the smallest effective memory and cache footprint.
3137 * allow dest_cpu, which will force the cpu onto dest_cpu. Then
3138 * the cpu_allowed mask is restored.
3139 */ 3137 */
3140static void sched_migrate_task(struct task_struct *p, int dest_cpu) 3138void sched_exec(void)
3141{ 3139{
3140 struct task_struct *p = current;
3142 struct migration_req req; 3141 struct migration_req req;
3142 int dest_cpu, this_cpu;
3143 unsigned long flags; 3143 unsigned long flags;
3144 struct rq *rq; 3144 struct rq *rq;
3145 3145
3146again:
3147 this_cpu = get_cpu();
3148 dest_cpu = select_task_rq(p, SD_BALANCE_EXEC, 0);
3149 if (dest_cpu == this_cpu) {
3150 put_cpu();
3151 return;
3152 }
3153
3146 rq = task_rq_lock(p, &flags); 3154 rq = task_rq_lock(p, &flags);
3155 put_cpu();
3156
3157 /*
3158 * select_task_rq() can race against ->cpus_allowed
3159 */
3147 if (!cpumask_test_cpu(dest_cpu, &p->cpus_allowed) 3160 if (!cpumask_test_cpu(dest_cpu, &p->cpus_allowed)
3148 || unlikely(!cpu_active(dest_cpu))) 3161 || unlikely(!cpu_active(dest_cpu))) {
3149 goto out; 3162 task_rq_unlock(rq, &flags);
3163 goto again;
3164 }
3150 3165
3151 /* force the process onto the specified CPU */ 3166 /* force the process onto the specified CPU */
3152 if (migrate_task(p, dest_cpu, &req)) { 3167 if (migrate_task(p, dest_cpu, &req)) {
@@ -3161,24 +3176,10 @@ static void sched_migrate_task(struct task_struct *p, int dest_cpu)
3161 3176
3162 return; 3177 return;
3163 } 3178 }
3164out:
3165 task_rq_unlock(rq, &flags); 3179 task_rq_unlock(rq, &flags);
3166} 3180}
3167 3181
3168/* 3182/*
3169 * sched_exec - execve() is a valuable balancing opportunity, because at
3170 * this point the task has the smallest effective memory and cache footprint.
3171 */
3172void sched_exec(void)
3173{
3174 int new_cpu, this_cpu = get_cpu();
3175 new_cpu = select_task_rq(current, SD_BALANCE_EXEC, 0);
3176 put_cpu();
3177 if (new_cpu != this_cpu)
3178 sched_migrate_task(current, new_cpu);
3179}
3180
3181/*
3182 * pull_task - move a task from a remote runqueue to the local runqueue. 3183 * pull_task - move a task from a remote runqueue to the local runqueue.
3183 * Both runqueues must be locked. 3184 * Both runqueues must be locked.
3184 */ 3185 */