diff options
| author | Len Brown <len.brown@intel.com> | 2009-04-05 02:14:15 -0400 |
|---|---|---|
| committer | Len Brown <len.brown@intel.com> | 2009-04-05 02:14:15 -0400 |
| commit | 478c6a43fcbc6c11609f8cee7c7b57223907754f (patch) | |
| tree | a7f7952099da60d33032aed6de9c0c56c9f8779e /kernel/rcupdate.c | |
| parent | 8a3f257c704e02aee9869decd069a806b45be3f1 (diff) | |
| parent | 6bb597507f9839b13498781e481f5458aea33620 (diff) | |
Merge branch 'linus' into release
Conflicts:
arch/x86/kernel/cpu/cpufreq/longhaul.c
Signed-off-by: Len Brown <len.brown@intel.com>
Diffstat (limited to 'kernel/rcupdate.c')
| -rw-r--r-- | kernel/rcupdate.c | 44 |
1 files changed, 44 insertions, 0 deletions
diff --git a/kernel/rcupdate.c b/kernel/rcupdate.c index cae8a059cf47..2c7b8457d0d2 100644 --- a/kernel/rcupdate.c +++ b/kernel/rcupdate.c | |||
| @@ -122,6 +122,8 @@ static void rcu_barrier_func(void *type) | |||
| 122 | } | 122 | } |
| 123 | } | 123 | } |
| 124 | 124 | ||
| 125 | static inline void wait_migrated_callbacks(void); | ||
| 126 | |||
| 125 | /* | 127 | /* |
| 126 | * Orchestrate the specified type of RCU barrier, waiting for all | 128 | * Orchestrate the specified type of RCU barrier, waiting for all |
| 127 | * RCU callbacks of the specified type to complete. | 129 | * RCU callbacks of the specified type to complete. |
| @@ -147,6 +149,7 @@ static void _rcu_barrier(enum rcu_barrier type) | |||
| 147 | complete(&rcu_barrier_completion); | 149 | complete(&rcu_barrier_completion); |
| 148 | wait_for_completion(&rcu_barrier_completion); | 150 | wait_for_completion(&rcu_barrier_completion); |
| 149 | mutex_unlock(&rcu_barrier_mutex); | 151 | mutex_unlock(&rcu_barrier_mutex); |
| 152 | wait_migrated_callbacks(); | ||
| 150 | } | 153 | } |
| 151 | 154 | ||
| 152 | /** | 155 | /** |
| @@ -176,9 +179,50 @@ void rcu_barrier_sched(void) | |||
| 176 | } | 179 | } |
| 177 | EXPORT_SYMBOL_GPL(rcu_barrier_sched); | 180 | EXPORT_SYMBOL_GPL(rcu_barrier_sched); |
| 178 | 181 | ||
| 182 | static atomic_t rcu_migrate_type_count = ATOMIC_INIT(0); | ||
| 183 | static struct rcu_head rcu_migrate_head[3]; | ||
| 184 | static DECLARE_WAIT_QUEUE_HEAD(rcu_migrate_wq); | ||
| 185 | |||
| 186 | static void rcu_migrate_callback(struct rcu_head *notused) | ||
| 187 | { | ||
| 188 | if (atomic_dec_and_test(&rcu_migrate_type_count)) | ||
| 189 | wake_up(&rcu_migrate_wq); | ||
| 190 | } | ||
| 191 | |||
| 192 | static inline void wait_migrated_callbacks(void) | ||
| 193 | { | ||
| 194 | wait_event(rcu_migrate_wq, !atomic_read(&rcu_migrate_type_count)); | ||
| 195 | } | ||
| 196 | |||
| 197 | static int __cpuinit rcu_barrier_cpu_hotplug(struct notifier_block *self, | ||
| 198 | unsigned long action, void *hcpu) | ||
| 199 | { | ||
| 200 | if (action == CPU_DYING) { | ||
| 201 | /* | ||
| 202 | * preempt_disable() in on_each_cpu() prevents stop_machine(), | ||
| 203 | * so when "on_each_cpu(rcu_barrier_func, (void *)type, 1);" | ||
| 204 | * returns, all online cpus have queued rcu_barrier_func(), | ||
| 205 | * and the dead cpu(if it exist) queues rcu_migrate_callback()s. | ||
| 206 | * | ||
| 207 | * These callbacks ensure _rcu_barrier() waits for all | ||
| 208 | * RCU callbacks of the specified type to complete. | ||
| 209 | */ | ||
| 210 | atomic_set(&rcu_migrate_type_count, 3); | ||
| 211 | call_rcu_bh(rcu_migrate_head, rcu_migrate_callback); | ||
| 212 | call_rcu_sched(rcu_migrate_head + 1, rcu_migrate_callback); | ||
| 213 | call_rcu(rcu_migrate_head + 2, rcu_migrate_callback); | ||
| 214 | } else if (action == CPU_POST_DEAD) { | ||
| 215 | /* rcu_migrate_head is protected by cpu_add_remove_lock */ | ||
| 216 | wait_migrated_callbacks(); | ||
| 217 | } | ||
| 218 | |||
| 219 | return NOTIFY_OK; | ||
| 220 | } | ||
| 221 | |||
| 179 | void __init rcu_init(void) | 222 | void __init rcu_init(void) |
| 180 | { | 223 | { |
| 181 | __rcu_init(); | 224 | __rcu_init(); |
| 225 | hotcpu_notifier(rcu_barrier_cpu_hotplug, 0); | ||
| 182 | } | 226 | } |
| 183 | 227 | ||
| 184 | void rcu_scheduler_starting(void) | 228 | void rcu_scheduler_starting(void) |
