diff options
| author | Xiaotian Feng <dfeng@redhat.com> | 2009-12-16 12:04:32 -0500 |
|---|---|---|
| committer | Ingo Molnar <mingo@elte.hu> | 2009-12-16 13:01:53 -0500 |
| commit | 9ee349ad6d326df3633d43f54202427295999c47 (patch) | |
| tree | 06f4795cb1b3f493b2421b65af0fd870875c8168 | |
| parent | 933b0618d8b2a59c7a0742e43836544e02f1e9bd (diff) | |
sched: Fix set_cpu_active() in cpu_down()
Sachin found cpu hotplug test failures on powerpc, which made
the kernel hang on his POWER box.
The problem is that we fail to re-activate a cpu when a
hot-unplug fails. Fix this by moving the de-activation into
_cpu_down after doing the initial checks.
Remove the synchronize_sched() calls and rely on those implied
by rebuilding the sched domains using the new mask.
Reported-by: Sachin Sant <sachinp@in.ibm.com>
Signed-off-by: Xiaotian Feng <dfeng@redhat.com>
Tested-by: Sachin Sant <sachinp@in.ibm.com>
Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Cc: Mike Galbraith <efault@gmx.de>
LKML-Reference: <20091216170517.500272612@chello.nl>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
| -rw-r--r-- | kernel/cpu.c | 24 |
1 files changed, 3 insertions, 21 deletions
diff --git a/kernel/cpu.c b/kernel/cpu.c index 291ac586f37f..1c8ddd6ee940 100644 --- a/kernel/cpu.c +++ b/kernel/cpu.c | |||
| @@ -209,6 +209,7 @@ static int __ref _cpu_down(unsigned int cpu, int tasks_frozen) | |||
| 209 | return -ENOMEM; | 209 | return -ENOMEM; |
| 210 | 210 | ||
| 211 | cpu_hotplug_begin(); | 211 | cpu_hotplug_begin(); |
| 212 | set_cpu_active(cpu, false); | ||
| 212 | err = __raw_notifier_call_chain(&cpu_chain, CPU_DOWN_PREPARE | mod, | 213 | err = __raw_notifier_call_chain(&cpu_chain, CPU_DOWN_PREPARE | mod, |
| 213 | hcpu, -1, &nr_calls); | 214 | hcpu, -1, &nr_calls); |
| 214 | if (err == NOTIFY_BAD) { | 215 | if (err == NOTIFY_BAD) { |
| @@ -280,18 +281,6 @@ int __ref cpu_down(unsigned int cpu) | |||
| 280 | goto out; | 281 | goto out; |
| 281 | } | 282 | } |
| 282 | 283 | ||
| 283 | set_cpu_active(cpu, false); | ||
| 284 | |||
| 285 | /* | ||
| 286 | * Make sure the all cpus did the reschedule and are not | ||
| 287 | * using stale version of the cpu_active_mask. | ||
| 288 | * This is not strictly necessary becuase stop_machine() | ||
| 289 | * that we run down the line already provides the required | ||
| 290 | * synchronization. But it's really a side effect and we do not | ||
| 291 | * want to depend on the innards of the stop_machine here. | ||
| 292 | */ | ||
| 293 | synchronize_sched(); | ||
| 294 | |||
| 295 | err = _cpu_down(cpu, 0); | 284 | err = _cpu_down(cpu, 0); |
| 296 | 285 | ||
| 297 | out: | 286 | out: |
| @@ -382,19 +371,12 @@ int disable_nonboot_cpus(void) | |||
| 382 | return error; | 371 | return error; |
| 383 | cpu_maps_update_begin(); | 372 | cpu_maps_update_begin(); |
| 384 | first_cpu = cpumask_first(cpu_online_mask); | 373 | first_cpu = cpumask_first(cpu_online_mask); |
| 385 | /* We take down all of the non-boot CPUs in one shot to avoid races | 374 | /* |
| 375 | * We take down all of the non-boot CPUs in one shot to avoid races | ||
| 386 | * with the userspace trying to use the CPU hotplug at the same time | 376 | * with the userspace trying to use the CPU hotplug at the same time |
| 387 | */ | 377 | */ |
| 388 | cpumask_clear(frozen_cpus); | 378 | cpumask_clear(frozen_cpus); |
| 389 | 379 | ||
| 390 | for_each_online_cpu(cpu) { | ||
| 391 | if (cpu == first_cpu) | ||
| 392 | continue; | ||
| 393 | set_cpu_active(cpu, false); | ||
| 394 | } | ||
| 395 | |||
| 396 | synchronize_sched(); | ||
| 397 | |||
| 398 | printk("Disabling non-boot CPUs ...\n"); | 380 | printk("Disabling non-boot CPUs ...\n"); |
| 399 | for_each_online_cpu(cpu) { | 381 | for_each_online_cpu(cpu) { |
| 400 | if (cpu == first_cpu) | 382 | if (cpu == first_cpu) |
