aboutsummaryrefslogtreecommitdiffstats
path: root/arch
diff options
context:
space:
mode:
Diffstat (limited to 'arch')
-rw-r--r--arch/powerpc/platforms/pseries/setup.c190
1 files changed, 70 insertions, 120 deletions
diff --git a/arch/powerpc/platforms/pseries/setup.c b/arch/powerpc/platforms/pseries/setup.c
index 44d5c7fdcd97..213bf983242f 100644
--- a/arch/powerpc/platforms/pseries/setup.c
+++ b/arch/powerpc/platforms/pseries/setup.c
@@ -81,8 +81,8 @@ extern void find_udbg_vterm(void);
81 81
82int fwnmi_active; /* TRUE if an FWNMI handler is present */ 82int fwnmi_active; /* TRUE if an FWNMI handler is present */
83 83
84static void pseries_shared_idle(void); 84static void pseries_shared_idle_sleep(void);
85static void pseries_dedicated_idle(void); 85static void pseries_dedicated_idle_sleep(void);
86 86
87struct mpic *pSeries_mpic; 87struct mpic *pSeries_mpic;
88 88
@@ -236,14 +236,13 @@ static void __init pSeries_setup_arch(void)
236 vpa_init(boot_cpuid); 236 vpa_init(boot_cpuid);
237 if (get_lppaca()->shared_proc) { 237 if (get_lppaca()->shared_proc) {
238 printk(KERN_INFO "Using shared processor idle loop\n"); 238 printk(KERN_INFO "Using shared processor idle loop\n");
239 ppc_md.idle_loop = pseries_shared_idle; 239 ppc_md.power_save = pseries_shared_idle_sleep;
240 } else { 240 } else {
241 printk(KERN_INFO "Using dedicated idle loop\n"); 241 printk(KERN_INFO "Using dedicated idle loop\n");
242 ppc_md.idle_loop = pseries_dedicated_idle; 242 ppc_md.power_save = pseries_dedicated_idle_sleep;
243 } 243 }
244 } else { 244 } else {
245 printk(KERN_INFO "Using default idle loop\n"); 245 printk(KERN_INFO "Using default idle loop\n");
246 ppc_md.idle_loop = default_idle;
247 } 246 }
248 247
249 if (firmware_has_feature(FW_FEATURE_LPAR)) 248 if (firmware_has_feature(FW_FEATURE_LPAR))
@@ -393,136 +392,87 @@ static int __init pSeries_probe(int platform)
393 392
394DECLARE_PER_CPU(unsigned long, smt_snooze_delay); 393DECLARE_PER_CPU(unsigned long, smt_snooze_delay);
395 394
396static inline void dedicated_idle_sleep(unsigned int cpu) 395static void pseries_dedicated_idle_sleep(void)
397{
398 struct lppaca *plppaca = &lppaca[cpu ^ 1];
399
400 /* Only sleep if the other thread is not idle */
401 if (!(plppaca->idle)) {
402 local_irq_disable();
403
404 /*
405 * We are about to sleep the thread and so wont be polling any
406 * more.
407 */
408 clear_thread_flag(TIF_POLLING_NRFLAG);
409 smp_mb__after_clear_bit();
410
411 /*
412 * SMT dynamic mode. Cede will result in this thread going
413 * dormant, if the partner thread is still doing work. Thread
414 * wakes up if partner goes idle, an interrupt is presented, or
415 * a prod occurs. Returning from the cede enables external
416 * interrupts.
417 */
418 if (!need_resched())
419 cede_processor();
420 else
421 local_irq_enable();
422 set_thread_flag(TIF_POLLING_NRFLAG);
423 } else {
424 /*
425 * Give the HV an opportunity at the processor, since we are
426 * not doing any work.
427 */
428 poll_pending();
429 }
430}
431
432static void pseries_dedicated_idle(void)
433{ 396{
434 unsigned int cpu = smp_processor_id(); 397 unsigned int cpu = smp_processor_id();
435 unsigned long start_snooze; 398 unsigned long start_snooze;
436 unsigned long *smt_snooze_delay = &__get_cpu_var(smt_snooze_delay); 399 unsigned long *smt_snooze_delay = &__get_cpu_var(smt_snooze_delay);
437 set_thread_flag(TIF_POLLING_NRFLAG);
438
439 while (1) {
440 /*
441 * Indicate to the HV that we are idle. Now would be
442 * a good time to find other work to dispatch.
443 */
444 get_lppaca()->idle = 1;
445
446 if (!need_resched()) {
447 start_snooze = get_tb() +
448 *smt_snooze_delay * tb_ticks_per_usec;
449
450 while (!need_resched() && !cpu_is_offline(cpu)) {
451 ppc64_runlatch_off();
452
453 /*
454 * Go into low thread priority and possibly
455 * low power mode.
456 */
457 HMT_low();
458 HMT_very_low();
459
460 if (*smt_snooze_delay != 0 &&
461 get_tb() > start_snooze) {
462 HMT_medium();
463 dedicated_idle_sleep(cpu);
464 }
465
466 }
467
468 HMT_medium();
469 }
470 400
471 get_lppaca()->idle = 0; 401 /*
472 ppc64_runlatch_on(); 402 * Indicate to the HV that we are idle. Now would be
403 * a good time to find other work to dispatch.
404 */
405 get_lppaca()->idle = 1;
473 406
474 preempt_enable_no_resched(); 407 /*
475 schedule(); 408 * We come in with interrupts disabled, and need_resched()
476 preempt_disable(); 409 * has been checked recently. If we should poll for a little
410 * while, do so.
411 */
412 if (*smt_snooze_delay) {
413 start_snooze = get_tb() +
414 *smt_snooze_delay * tb_ticks_per_usec;
415 local_irq_enable();
416 set_thread_flag(TIF_POLLING_NRFLAG);
477 417
478 if (cpu_is_offline(cpu) && system_state == SYSTEM_RUNNING) 418 while (get_tb() < start_snooze) {
479 cpu_die(); 419 if (need_resched() || cpu_is_offline(cpu))
420 goto out;
421 ppc64_runlatch_off();
422 HMT_low();
423 HMT_very_low();
424 }
425
426 HMT_medium();
427 clear_thread_flag(TIF_POLLING_NRFLAG);
428 smp_mb();
429 local_irq_disable();
430 if (need_resched() || cpu_is_offline(cpu))
431 goto out;
480 } 432 }
433
434 /*
435 * Cede if the other thread is not idle, so that it can
436 * go single-threaded. If the other thread is idle,
437 * we ask the hypervisor if it has pending work it
438 * wants to do and cede if it does. Otherwise we keep
439 * polling in order to reduce interrupt latency.
440 *
441 * Doing the cede when the other thread is active will
442 * result in this thread going dormant, meaning the other
443 * thread gets to run in single-threaded (ST) mode, which
444 * is slightly faster than SMT mode with this thread at
445 * very low priority. The cede enables interrupts, which
446 * doesn't matter here.
447 */
448 if (!lppaca[cpu ^ 1].idle || poll_pending() == H_Pending)
449 cede_processor();
450
451out:
452 HMT_medium();
453 get_lppaca()->idle = 0;
481} 454}
482 455
483static void pseries_shared_idle(void) 456static void pseries_shared_idle_sleep(void)
484{ 457{
485 unsigned int cpu = smp_processor_id(); 458 unsigned int cpu = smp_processor_id();
486 459
487 while (1) { 460 /*
488 /* 461 * Indicate to the HV that we are idle. Now would be
489 * Indicate to the HV that we are idle. Now would be 462 * a good time to find other work to dispatch.
490 * a good time to find other work to dispatch. 463 */
491 */ 464 get_lppaca()->idle = 1;
492 get_lppaca()->idle = 1;
493
494 while (!need_resched() && !cpu_is_offline(cpu)) {
495 local_irq_disable();
496 ppc64_runlatch_off();
497
498 /*
499 * Yield the processor to the hypervisor. We return if
500 * an external interrupt occurs (which are driven prior
501 * to returning here) or if a prod occurs from another
502 * processor. When returning here, external interrupts
503 * are enabled.
504 *
505 * Check need_resched() again with interrupts disabled
506 * to avoid a race.
507 */
508 if (!need_resched())
509 cede_processor();
510 else
511 local_irq_enable();
512
513 HMT_medium();
514 }
515
516 get_lppaca()->idle = 0;
517 ppc64_runlatch_on();
518 465
519 preempt_enable_no_resched(); 466 /*
520 schedule(); 467 * Yield the processor to the hypervisor. We return if
521 preempt_disable(); 468 * an external interrupt occurs (which are driven prior
469 * to returning here) or if a prod occurs from another
470 * processor. When returning here, external interrupts
471 * are enabled.
472 */
473 cede_processor();
522 474
523 if (cpu_is_offline(cpu) && system_state == SYSTEM_RUNNING) 475 get_lppaca()->idle = 0;
524 cpu_die();
525 }
526} 476}
527 477
528static int pSeries_pci_probe_mode(struct pci_bus *bus) 478static int pSeries_pci_probe_mode(struct pci_bus *bus)