aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--arch/i386/kernel/cpu/cpufreq/speedstep-centrino.c246
1 files changed, 180 insertions, 66 deletions
diff --git a/arch/i386/kernel/cpu/cpufreq/speedstep-centrino.c b/arch/i386/kernel/cpu/cpufreq/speedstep-centrino.c
index c173c0fa117a..37dee862f0d3 100644
--- a/arch/i386/kernel/cpu/cpufreq/speedstep-centrino.c
+++ b/arch/i386/kernel/cpu/cpufreq/speedstep-centrino.c
@@ -351,7 +351,36 @@ static unsigned int get_cur_freq(unsigned int cpu)
351 351
352#ifdef CONFIG_X86_SPEEDSTEP_CENTRINO_ACPI 352#ifdef CONFIG_X86_SPEEDSTEP_CENTRINO_ACPI
353 353
354static struct acpi_processor_performance p; 354static struct acpi_processor_performance *acpi_perf_data[NR_CPUS];
355
356/*
357 * centrino_cpu_early_init_acpi - Do the preregistering with ACPI P-States
358 * library
359 *
360 * Before doing the actual init, we need to do _PSD related setup whenever
361 * supported by the BIOS. These are handled by this early_init routine.
362 */
363static int centrino_cpu_early_init_acpi(void)
364{
365 unsigned int i, j;
366 struct acpi_processor_performance *data;
367
368 for_each_cpu(i) {
369 data = kzalloc(sizeof(struct acpi_processor_performance),
370 GFP_KERNEL);
371 if (!data) {
372 for_each_cpu(j) {
373 kfree(acpi_perf_data[j]);
374 acpi_perf_data[j] = NULL;
375 }
376 return (-ENOMEM);
377 }
378 acpi_perf_data[i] = data;
379 }
380
381 acpi_processor_preregister_performance(acpi_perf_data);
382 return 0;
383}
355 384
356/* 385/*
357 * centrino_cpu_init_acpi - register with ACPI P-States library 386 * centrino_cpu_init_acpi - register with ACPI P-States library
@@ -365,46 +394,51 @@ static int centrino_cpu_init_acpi(struct cpufreq_policy *policy)
365 unsigned long cur_freq; 394 unsigned long cur_freq;
366 int result = 0, i; 395 int result = 0, i;
367 unsigned int cpu = policy->cpu; 396 unsigned int cpu = policy->cpu;
397 struct acpi_processor_performance *p;
398
399 p = acpi_perf_data[cpu];
368 400
369 /* register with ACPI core */ 401 /* register with ACPI core */
370 if (acpi_processor_register_performance(&p, cpu)) { 402 if (acpi_processor_register_performance(p, cpu)) {
371 dprintk(KERN_INFO PFX "obtaining ACPI data failed\n"); 403 dprintk(KERN_INFO PFX "obtaining ACPI data failed\n");
372 return -EIO; 404 return -EIO;
373 } 405 }
406 policy->cpus = p->shared_cpu_map;
407 policy->shared_type = p->shared_type;
374 408
375 /* verify the acpi_data */ 409 /* verify the acpi_data */
376 if (p.state_count <= 1) { 410 if (p->state_count <= 1) {
377 dprintk("No P-States\n"); 411 dprintk("No P-States\n");
378 result = -ENODEV; 412 result = -ENODEV;
379 goto err_unreg; 413 goto err_unreg;
380 } 414 }
381 415
382 if ((p.control_register.space_id != ACPI_ADR_SPACE_FIXED_HARDWARE) || 416 if ((p->control_register.space_id != ACPI_ADR_SPACE_FIXED_HARDWARE) ||
383 (p.status_register.space_id != ACPI_ADR_SPACE_FIXED_HARDWARE)) { 417 (p->status_register.space_id != ACPI_ADR_SPACE_FIXED_HARDWARE)) {
384 dprintk("Invalid control/status registers (%x - %x)\n", 418 dprintk("Invalid control/status registers (%x - %x)\n",
385 p.control_register.space_id, p.status_register.space_id); 419 p->control_register.space_id, p->status_register.space_id);
386 result = -EIO; 420 result = -EIO;
387 goto err_unreg; 421 goto err_unreg;
388 } 422 }
389 423
390 for (i=0; i<p.state_count; i++) { 424 for (i=0; i<p->state_count; i++) {
391 if (p.states[i].control != p.states[i].status) { 425 if (p->states[i].control != p->states[i].status) {
392 dprintk("Different control (%llu) and status values (%llu)\n", 426 dprintk("Different control (%llu) and status values (%llu)\n",
393 p.states[i].control, p.states[i].status); 427 p->states[i].control, p->states[i].status);
394 result = -EINVAL; 428 result = -EINVAL;
395 goto err_unreg; 429 goto err_unreg;
396 } 430 }
397 431
398 if (!p.states[i].core_frequency) { 432 if (!p->states[i].core_frequency) {
399 dprintk("Zero core frequency for state %u\n", i); 433 dprintk("Zero core frequency for state %u\n", i);
400 result = -EINVAL; 434 result = -EINVAL;
401 goto err_unreg; 435 goto err_unreg;
402 } 436 }
403 437
404 if (p.states[i].core_frequency > p.states[0].core_frequency) { 438 if (p->states[i].core_frequency > p->states[0].core_frequency) {
405 dprintk("P%u has larger frequency (%llu) than P0 (%llu), skipping\n", i, 439 dprintk("P%u has larger frequency (%llu) than P0 (%llu), skipping\n", i,
406 p.states[i].core_frequency, p.states[0].core_frequency); 440 p->states[i].core_frequency, p->states[0].core_frequency);
407 p.states[i].core_frequency = 0; 441 p->states[i].core_frequency = 0;
408 continue; 442 continue;
409 } 443 }
410 } 444 }
@@ -416,26 +450,26 @@ static int centrino_cpu_init_acpi(struct cpufreq_policy *policy)
416 } 450 }
417 451
418 centrino_model[cpu]->model_name=NULL; 452 centrino_model[cpu]->model_name=NULL;
419 centrino_model[cpu]->max_freq = p.states[0].core_frequency * 1000; 453 centrino_model[cpu]->max_freq = p->states[0].core_frequency * 1000;
420 centrino_model[cpu]->op_points = kmalloc(sizeof(struct cpufreq_frequency_table) * 454 centrino_model[cpu]->op_points = kmalloc(sizeof(struct cpufreq_frequency_table) *
421 (p.state_count + 1), GFP_KERNEL); 455 (p->state_count + 1), GFP_KERNEL);
422 if (!centrino_model[cpu]->op_points) { 456 if (!centrino_model[cpu]->op_points) {
423 result = -ENOMEM; 457 result = -ENOMEM;
424 goto err_kfree; 458 goto err_kfree;
425 } 459 }
426 460
427 for (i=0; i<p.state_count; i++) { 461 for (i=0; i<p->state_count; i++) {
428 centrino_model[cpu]->op_points[i].index = p.states[i].control; 462 centrino_model[cpu]->op_points[i].index = p->states[i].control;
429 centrino_model[cpu]->op_points[i].frequency = p.states[i].core_frequency * 1000; 463 centrino_model[cpu]->op_points[i].frequency = p->states[i].core_frequency * 1000;
430 dprintk("adding state %i with frequency %u and control value %04x\n", 464 dprintk("adding state %i with frequency %u and control value %04x\n",
431 i, centrino_model[cpu]->op_points[i].frequency, centrino_model[cpu]->op_points[i].index); 465 i, centrino_model[cpu]->op_points[i].frequency, centrino_model[cpu]->op_points[i].index);
432 } 466 }
433 centrino_model[cpu]->op_points[p.state_count].frequency = CPUFREQ_TABLE_END; 467 centrino_model[cpu]->op_points[p->state_count].frequency = CPUFREQ_TABLE_END;
434 468
435 cur_freq = get_cur_freq(cpu); 469 cur_freq = get_cur_freq(cpu);
436 470
437 for (i=0; i<p.state_count; i++) { 471 for (i=0; i<p->state_count; i++) {
438 if (!p.states[i].core_frequency) { 472 if (!p->states[i].core_frequency) {
439 dprintk("skipping state %u\n", i); 473 dprintk("skipping state %u\n", i);
440 centrino_model[cpu]->op_points[i].frequency = CPUFREQ_ENTRY_INVALID; 474 centrino_model[cpu]->op_points[i].frequency = CPUFREQ_ENTRY_INVALID;
441 continue; 475 continue;
@@ -451,7 +485,7 @@ static int centrino_cpu_init_acpi(struct cpufreq_policy *policy)
451 } 485 }
452 486
453 if (cur_freq == centrino_model[cpu]->op_points[i].frequency) 487 if (cur_freq == centrino_model[cpu]->op_points[i].frequency)
454 p.state = i; 488 p->state = i;
455 } 489 }
456 490
457 /* notify BIOS that we exist */ 491 /* notify BIOS that we exist */
@@ -464,12 +498,13 @@ static int centrino_cpu_init_acpi(struct cpufreq_policy *policy)
464 err_kfree: 498 err_kfree:
465 kfree(centrino_model[cpu]); 499 kfree(centrino_model[cpu]);
466 err_unreg: 500 err_unreg:
467 acpi_processor_unregister_performance(&p, cpu); 501 acpi_processor_unregister_performance(p, cpu);
468 dprintk(KERN_INFO PFX "invalid ACPI data\n"); 502 dprintk(KERN_INFO PFX "invalid ACPI data\n");
469 return (result); 503 return (result);
470} 504}
471#else 505#else
472static inline int centrino_cpu_init_acpi(struct cpufreq_policy *policy) { return -ENODEV; } 506static inline int centrino_cpu_init_acpi(struct cpufreq_policy *policy) { return -ENODEV; }
507static inline int centrino_cpu_early_init_acpi(void) { return 0; }
473#endif 508#endif
474 509
475static int centrino_cpu_init(struct cpufreq_policy *policy) 510static int centrino_cpu_init(struct cpufreq_policy *policy)
@@ -557,10 +592,15 @@ static int centrino_cpu_exit(struct cpufreq_policy *policy)
557 592
558#ifdef CONFIG_X86_SPEEDSTEP_CENTRINO_ACPI 593#ifdef CONFIG_X86_SPEEDSTEP_CENTRINO_ACPI
559 if (!centrino_model[cpu]->model_name) { 594 if (!centrino_model[cpu]->model_name) {
560 dprintk("unregistering and freeing ACPI data\n"); 595 static struct acpi_processor_performance *p;
561 acpi_processor_unregister_performance(&p, cpu); 596
562 kfree(centrino_model[cpu]->op_points); 597 if (acpi_perf_data[cpu]) {
563 kfree(centrino_model[cpu]); 598 p = acpi_perf_data[cpu];
599 dprintk("unregistering and freeing ACPI data\n");
600 acpi_processor_unregister_performance(p, cpu);
601 kfree(centrino_model[cpu]->op_points);
602 kfree(centrino_model[cpu]);
603 }
564 } 604 }
565#endif 605#endif
566 606
@@ -594,63 +634,124 @@ static int centrino_target (struct cpufreq_policy *policy,
594 unsigned int relation) 634 unsigned int relation)
595{ 635{
596 unsigned int newstate = 0; 636 unsigned int newstate = 0;
597 unsigned int msr, oldmsr, h, cpu = policy->cpu; 637 unsigned int msr, oldmsr = 0, h = 0, cpu = policy->cpu;
598 struct cpufreq_freqs freqs; 638 struct cpufreq_freqs freqs;
639 cpumask_t online_policy_cpus;
599 cpumask_t saved_mask; 640 cpumask_t saved_mask;
600 int retval; 641 cpumask_t set_mask;
642 cpumask_t covered_cpus;
643 int retval = 0;
644 unsigned int j, k, first_cpu, tmp;
601 645
602 if (centrino_model[cpu] == NULL) 646 if (unlikely(centrino_model[cpu] == NULL))
603 return -ENODEV; 647 return -ENODEV;
604 648
605 /* 649 if (unlikely(cpufreq_frequency_table_target(policy,
606 * Support for SMP systems. 650 centrino_model[cpu]->op_points,
607 * Make sure we are running on the CPU that wants to change frequency 651 target_freq,
608 */ 652 relation,
609 saved_mask = current->cpus_allowed; 653 &newstate))) {
610 set_cpus_allowed(current, policy->cpus); 654 return -EINVAL;
611 if (!cpu_isset(smp_processor_id(), policy->cpus)) {
612 dprintk("couldn't limit to CPUs in this domain\n");
613 return(-EAGAIN);
614 } 655 }
615 656
616 if (cpufreq_frequency_table_target(policy, centrino_model[cpu]->op_points, target_freq, 657 /* cpufreq holds the hotplug lock, so we are safe from here on */
617 relation, &newstate)) { 658 cpus_and(online_policy_cpus, cpu_online_map, policy->cpus);
618 retval = -EINVAL;
619 goto migrate_end;
620 }
621 659
622 msr = centrino_model[cpu]->op_points[newstate].index; 660 saved_mask = current->cpus_allowed;
623 rdmsr(MSR_IA32_PERF_CTL, oldmsr, h); 661 first_cpu = 1;
662 cpus_clear(covered_cpus);
663 for_each_cpu_mask(j, online_policy_cpus) {
664 /*
665 * Support for SMP systems.
666 * Make sure we are running on CPU that wants to change freq
667 */
668 cpus_clear(set_mask);
669 if (policy->shared_type == CPUFREQ_SHARED_TYPE_ANY)
670 cpus_or(set_mask, set_mask, online_policy_cpus);
671 else
672 cpu_set(j, set_mask);
673
674 set_cpus_allowed(current, set_mask);
675 if (unlikely(!cpu_isset(smp_processor_id(), set_mask))) {
676 dprintk("couldn't limit to CPUs in this domain\n");
677 retval = -EAGAIN;
678 if (first_cpu) {
679 /* We haven't started the transition yet. */
680 goto migrate_end;
681 }
682 break;
683 }
624 684
625 if (msr == (oldmsr & 0xffff)) { 685 msr = centrino_model[cpu]->op_points[newstate].index;
626 retval = 0; 686
627 dprintk("no change needed - msr was and needs to be %x\n", oldmsr); 687 if (first_cpu) {
628 goto migrate_end; 688 rdmsr(MSR_IA32_PERF_CTL, oldmsr, h);
629 } 689 if (msr == (oldmsr & 0xffff)) {
690 dprintk("no change needed - msr was and needs "
691 "to be %x\n", oldmsr);
692 retval = 0;
693 goto migrate_end;
694 }
695
696 freqs.old = extract_clock(oldmsr, cpu, 0);
697 freqs.new = extract_clock(msr, cpu, 0);
698
699 dprintk("target=%dkHz old=%d new=%d msr=%04x\n",
700 target_freq, freqs.old, freqs.new, msr);
701
702 for_each_cpu_mask(k, online_policy_cpus) {
703 freqs.cpu = k;
704 cpufreq_notify_transition(&freqs,
705 CPUFREQ_PRECHANGE);
706 }
707
708 first_cpu = 0;
709 /* all but 16 LSB are reserved, treat them with care */
710 oldmsr &= ~0xffff;
711 msr &= 0xffff;
712 oldmsr |= msr;
713 }
630 714
631 freqs.cpu = cpu; 715 wrmsr(MSR_IA32_PERF_CTL, oldmsr, h);
632 freqs.old = extract_clock(oldmsr, cpu, 0); 716 if (policy->shared_type == CPUFREQ_SHARED_TYPE_ANY)
633 freqs.new = extract_clock(msr, cpu, 0); 717 break;
634 718
635 dprintk("target=%dkHz old=%d new=%d msr=%04x\n", 719 cpu_set(j, covered_cpus);
636 target_freq, freqs.old, freqs.new, msr); 720 }
637 721
638 cpufreq_notify_transition(&freqs, CPUFREQ_PRECHANGE); 722 for_each_cpu_mask(k, online_policy_cpus) {
723 freqs.cpu = k;
724 cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE);
725 }
639 726
640 /* all but 16 LSB are "reserved", so treat them with 727 if (unlikely(retval)) {
641 care */ 728 /*
642 oldmsr &= ~0xffff; 729 * We have failed halfway through the frequency change.
643 msr &= 0xffff; 730 * We have sent callbacks to policy->cpus and
644 oldmsr |= msr; 731 * MSRs have already been written on coverd_cpus.
732 * Best effort undo..
733 */
645 734
646 wrmsr(MSR_IA32_PERF_CTL, oldmsr, h); 735 if (!cpus_empty(covered_cpus)) {
736 for_each_cpu_mask(j, covered_cpus) {
737 set_cpus_allowed(current, cpumask_of_cpu(j));
738 wrmsr(MSR_IA32_PERF_CTL, oldmsr, h);
739 }
740 }
647 741
648 cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE); 742 tmp = freqs.new;
743 freqs.new = freqs.old;
744 freqs.old = tmp;
745 for_each_cpu_mask(j, online_policy_cpus) {
746 freqs.cpu = j;
747 cpufreq_notify_transition(&freqs, CPUFREQ_PRECHANGE);
748 cpufreq_notify_transition(&freqs, CPUFREQ_POSTCHANGE);
749 }
750 }
649 751
650 retval = 0;
651migrate_end: 752migrate_end:
652 set_cpus_allowed(current, saved_mask); 753 set_cpus_allowed(current, saved_mask);
653 return (retval); 754 return 0;
654} 755}
655 756
656static struct freq_attr* centrino_attr[] = { 757static struct freq_attr* centrino_attr[] = {
@@ -692,12 +793,25 @@ static int __init centrino_init(void)
692 if (!cpu_has(cpu, X86_FEATURE_EST)) 793 if (!cpu_has(cpu, X86_FEATURE_EST))
693 return -ENODEV; 794 return -ENODEV;
694 795
796 centrino_cpu_early_init_acpi();
797
695 return cpufreq_register_driver(&centrino_driver); 798 return cpufreq_register_driver(&centrino_driver);
696} 799}
697 800
698static void __exit centrino_exit(void) 801static void __exit centrino_exit(void)
699{ 802{
803#ifdef CONFIG_X86_SPEEDSTEP_CENTRINO_ACPI
804 unsigned int j;
805#endif
806
700 cpufreq_unregister_driver(&centrino_driver); 807 cpufreq_unregister_driver(&centrino_driver);
808
809#ifdef CONFIG_X86_SPEEDSTEP_CENTRINO_ACPI
810 for_each_cpu(j) {
811 kfree(acpi_perf_data[j]);
812 acpi_perf_data[j] = NULL;
813 }
814#endif
701} 815}
702 816
703MODULE_AUTHOR ("Jeremy Fitzhardinge <jeremy@goop.org>"); 817MODULE_AUTHOR ("Jeremy Fitzhardinge <jeremy@goop.org>");