diff options
Diffstat (limited to 'kernel/smp.c')
-rw-r--r-- | kernel/smp.c | 139 |
1 files changed, 49 insertions, 90 deletions
diff --git a/kernel/smp.c b/kernel/smp.c index 75c8dde58c55..9f0eafed1399 100644 --- a/kernel/smp.c +++ b/kernel/smp.c | |||
@@ -24,8 +24,8 @@ struct call_function_data { | |||
24 | struct call_single_data csd; | 24 | struct call_single_data csd; |
25 | spinlock_t lock; | 25 | spinlock_t lock; |
26 | unsigned int refs; | 26 | unsigned int refs; |
27 | cpumask_t cpumask; | ||
28 | struct rcu_head rcu_head; | 27 | struct rcu_head rcu_head; |
28 | unsigned long cpumask_bits[]; | ||
29 | }; | 29 | }; |
30 | 30 | ||
31 | struct call_single_queue { | 31 | struct call_single_queue { |
@@ -110,13 +110,13 @@ void generic_smp_call_function_interrupt(void) | |||
110 | list_for_each_entry_rcu(data, &call_function_queue, csd.list) { | 110 | list_for_each_entry_rcu(data, &call_function_queue, csd.list) { |
111 | int refs; | 111 | int refs; |
112 | 112 | ||
113 | if (!cpu_isset(cpu, data->cpumask)) | 113 | if (!cpumask_test_cpu(cpu, to_cpumask(data->cpumask_bits))) |
114 | continue; | 114 | continue; |
115 | 115 | ||
116 | data->csd.func(data->csd.info); | 116 | data->csd.func(data->csd.info); |
117 | 117 | ||
118 | spin_lock(&data->lock); | 118 | spin_lock(&data->lock); |
119 | cpu_clear(cpu, data->cpumask); | 119 | cpumask_clear_cpu(cpu, to_cpumask(data->cpumask_bits)); |
120 | WARN_ON(data->refs == 0); | 120 | WARN_ON(data->refs == 0); |
121 | data->refs--; | 121 | data->refs--; |
122 | refs = data->refs; | 122 | refs = data->refs; |
@@ -266,51 +266,13 @@ void __smp_call_function_single(int cpu, struct call_single_data *data) | |||
266 | generic_exec_single(cpu, data); | 266 | generic_exec_single(cpu, data); |
267 | } | 267 | } |
268 | 268 | ||
269 | /* Dummy function */ | ||
270 | static void quiesce_dummy(void *unused) | ||
271 | { | ||
272 | } | ||
273 | |||
274 | /* | ||
275 | * Ensure stack based data used in call function mask is safe to free. | ||
276 | * | ||
277 | * This is needed by smp_call_function_mask when using on-stack data, because | ||
278 | * a single call function queue is shared by all CPUs, and any CPU may pick up | ||
279 | * the data item on the queue at any time before it is deleted. So we need to | ||
280 | * ensure that all CPUs have transitioned through a quiescent state after | ||
281 | * this call. | ||
282 | * | ||
283 | * This is a very slow function, implemented by sending synchronous IPIs to | ||
284 | * all possible CPUs. For this reason, we have to alloc data rather than use | ||
285 | * stack based data even in the case of synchronous calls. The stack based | ||
286 | * data is then just used for deadlock/oom fallback which will be very rare. | ||
287 | * | ||
288 | * If a faster scheme can be made, we could go back to preferring stack based | ||
289 | * data -- the data allocation/free is non-zero cost. | ||
290 | */ | ||
291 | static void smp_call_function_mask_quiesce_stack(cpumask_t mask) | ||
292 | { | ||
293 | struct call_single_data data; | ||
294 | int cpu; | ||
295 | |||
296 | data.func = quiesce_dummy; | ||
297 | data.info = NULL; | ||
298 | |||
299 | for_each_cpu_mask(cpu, mask) { | ||
300 | data.flags = CSD_FLAG_WAIT; | ||
301 | generic_exec_single(cpu, &data); | ||
302 | } | ||
303 | } | ||
304 | |||
305 | /** | 269 | /** |
306 | * smp_call_function_mask(): Run a function on a set of other CPUs. | 270 | * smp_call_function_many(): Run a function on a set of other CPUs. |
307 | * @mask: The set of cpus to run on. | 271 | * @mask: The set of cpus to run on (only runs on online subset). |
308 | * @func: The function to run. This must be fast and non-blocking. | 272 | * @func: The function to run. This must be fast and non-blocking. |
309 | * @info: An arbitrary pointer to pass to the function. | 273 | * @info: An arbitrary pointer to pass to the function. |
310 | * @wait: If true, wait (atomically) until function has completed on other CPUs. | 274 | * @wait: If true, wait (atomically) until function has completed on other CPUs. |
311 | * | 275 | * |
312 | * Returns 0 on success, else a negative status code. | ||
313 | * | ||
314 | * If @wait is true, then returns once @func has returned. Note that @wait | 276 | * If @wait is true, then returns once @func has returned. Note that @wait |
315 | * will be implicitly turned on in case of allocation failures, since | 277 | * will be implicitly turned on in case of allocation failures, since |
316 | * we fall back to on-stack allocation. | 278 | * we fall back to on-stack allocation. |
@@ -319,53 +281,57 @@ static void smp_call_function_mask_quiesce_stack(cpumask_t mask) | |||
319 | * hardware interrupt handler or from a bottom half handler. Preemption | 281 | * hardware interrupt handler or from a bottom half handler. Preemption |
320 | * must be disabled when calling this function. | 282 | * must be disabled when calling this function. |
321 | */ | 283 | */ |
322 | int smp_call_function_mask(cpumask_t mask, void (*func)(void *), void *info, | 284 | void smp_call_function_many(const struct cpumask *mask, |
323 | int wait) | 285 | void (*func)(void *), void *info, |
286 | bool wait) | ||
324 | { | 287 | { |
325 | struct call_function_data d; | 288 | struct call_function_data *data; |
326 | struct call_function_data *data = NULL; | ||
327 | cpumask_t allbutself; | ||
328 | unsigned long flags; | 289 | unsigned long flags; |
329 | int cpu, num_cpus; | 290 | int cpu, next_cpu; |
330 | int slowpath = 0; | ||
331 | 291 | ||
332 | /* Can deadlock when called with interrupts disabled */ | 292 | /* Can deadlock when called with interrupts disabled */ |
333 | WARN_ON(irqs_disabled()); | 293 | WARN_ON(irqs_disabled()); |
334 | 294 | ||
335 | cpu = smp_processor_id(); | 295 | /* So, what's a CPU they want? Ignoring this one. */ |
336 | allbutself = cpu_online_map; | 296 | cpu = cpumask_first_and(mask, cpu_online_mask); |
337 | cpu_clear(cpu, allbutself); | 297 | if (cpu == smp_processor_id()) |
338 | cpus_and(mask, mask, allbutself); | 298 | cpu = cpumask_next_and(cpu, mask, cpu_online_mask); |
339 | num_cpus = cpus_weight(mask); | 299 | /* No online cpus? We're done. */ |
340 | 300 | if (cpu >= nr_cpu_ids) | |
341 | /* | 301 | return; |
342 | * If zero CPUs, return. If just a single CPU, turn this request | 302 | |
343 | * into a targetted single call instead since it's faster. | 303 | /* Do we have another CPU which isn't us? */ |
344 | */ | 304 | next_cpu = cpumask_next_and(cpu, mask, cpu_online_mask); |
345 | if (!num_cpus) | 305 | if (next_cpu == smp_processor_id()) |
346 | return 0; | 306 | next_cpu = cpumask_next_and(next_cpu, mask, cpu_online_mask); |
347 | else if (num_cpus == 1) { | 307 | |
348 | cpu = first_cpu(mask); | 308 | /* Fastpath: do that cpu by itself. */ |
349 | return smp_call_function_single(cpu, func, info, wait); | 309 | if (next_cpu >= nr_cpu_ids) { |
310 | smp_call_function_single(cpu, func, info, wait); | ||
311 | return; | ||
350 | } | 312 | } |
351 | 313 | ||
352 | data = kmalloc(sizeof(*data), GFP_ATOMIC); | 314 | data = kmalloc(sizeof(*data) + cpumask_size(), GFP_ATOMIC); |
353 | if (data) { | 315 | if (unlikely(!data)) { |
354 | data->csd.flags = CSD_FLAG_ALLOC; | 316 | /* Slow path. */ |
355 | if (wait) | 317 | for_each_online_cpu(cpu) { |
356 | data->csd.flags |= CSD_FLAG_WAIT; | 318 | if (cpu == smp_processor_id()) |
357 | } else { | 319 | continue; |
358 | data = &d; | 320 | if (cpumask_test_cpu(cpu, mask)) |
359 | data->csd.flags = CSD_FLAG_WAIT; | 321 | smp_call_function_single(cpu, func, info, wait); |
360 | wait = 1; | 322 | } |
361 | slowpath = 1; | 323 | return; |
362 | } | 324 | } |
363 | 325 | ||
364 | spin_lock_init(&data->lock); | 326 | spin_lock_init(&data->lock); |
327 | data->csd.flags = CSD_FLAG_ALLOC; | ||
328 | if (wait) | ||
329 | data->csd.flags |= CSD_FLAG_WAIT; | ||
365 | data->csd.func = func; | 330 | data->csd.func = func; |
366 | data->csd.info = info; | 331 | data->csd.info = info; |
367 | data->refs = num_cpus; | 332 | cpumask_and(to_cpumask(data->cpumask_bits), mask, cpu_online_mask); |
368 | data->cpumask = mask; | 333 | cpumask_clear_cpu(smp_processor_id(), to_cpumask(data->cpumask_bits)); |
334 | data->refs = cpumask_weight(to_cpumask(data->cpumask_bits)); | ||
369 | 335 | ||
370 | spin_lock_irqsave(&call_function_lock, flags); | 336 | spin_lock_irqsave(&call_function_lock, flags); |
371 | list_add_tail_rcu(&data->csd.list, &call_function_queue); | 337 | list_add_tail_rcu(&data->csd.list, &call_function_queue); |
@@ -377,18 +343,13 @@ int smp_call_function_mask(cpumask_t mask, void (*func)(void *), void *info, | |||
377 | smp_mb(); | 343 | smp_mb(); |
378 | 344 | ||
379 | /* Send a message to all CPUs in the map */ | 345 | /* Send a message to all CPUs in the map */ |
380 | arch_send_call_function_ipi(mask); | 346 | arch_send_call_function_ipi(*to_cpumask(data->cpumask_bits)); |
381 | 347 | ||
382 | /* optionally wait for the CPUs to complete */ | 348 | /* optionally wait for the CPUs to complete */ |
383 | if (wait) { | 349 | if (wait) |
384 | csd_flag_wait(&data->csd); | 350 | csd_flag_wait(&data->csd); |
385 | if (unlikely(slowpath)) | ||
386 | smp_call_function_mask_quiesce_stack(mask); | ||
387 | } | ||
388 | |||
389 | return 0; | ||
390 | } | 351 | } |
391 | EXPORT_SYMBOL(smp_call_function_mask); | 352 | EXPORT_SYMBOL(smp_call_function_many); |
392 | 353 | ||
393 | /** | 354 | /** |
394 | * smp_call_function(): Run a function on all other CPUs. | 355 | * smp_call_function(): Run a function on all other CPUs. |
@@ -396,7 +357,7 @@ EXPORT_SYMBOL(smp_call_function_mask); | |||
396 | * @info: An arbitrary pointer to pass to the function. | 357 | * @info: An arbitrary pointer to pass to the function. |
397 | * @wait: If true, wait (atomically) until function has completed on other CPUs. | 358 | * @wait: If true, wait (atomically) until function has completed on other CPUs. |
398 | * | 359 | * |
399 | * Returns 0 on success, else a negative status code. | 360 | * Returns 0. |
400 | * | 361 | * |
401 | * If @wait is true, then returns once @func has returned; otherwise | 362 | * If @wait is true, then returns once @func has returned; otherwise |
402 | * it returns just before the target cpu calls @func. In case of allocation | 363 | * it returns just before the target cpu calls @func. In case of allocation |
@@ -407,12 +368,10 @@ EXPORT_SYMBOL(smp_call_function_mask); | |||
407 | */ | 368 | */ |
408 | int smp_call_function(void (*func)(void *), void *info, int wait) | 369 | int smp_call_function(void (*func)(void *), void *info, int wait) |
409 | { | 370 | { |
410 | int ret; | ||
411 | |||
412 | preempt_disable(); | 371 | preempt_disable(); |
413 | ret = smp_call_function_mask(cpu_online_map, func, info, wait); | 372 | smp_call_function_many(cpu_online_mask, func, info, wait); |
414 | preempt_enable(); | 373 | preempt_enable(); |
415 | return ret; | 374 | return 0; |
416 | } | 375 | } |
417 | EXPORT_SYMBOL(smp_call_function); | 376 | EXPORT_SYMBOL(smp_call_function); |
418 | 377 | ||