diff options
Diffstat (limited to 'arch/x86_64')
-rw-r--r-- | arch/x86_64/kernel/setup.c | 69 | ||||
-rw-r--r-- | arch/x86_64/kernel/smpboot.c | 69 |
2 files changed, 84 insertions, 54 deletions
diff --git a/arch/x86_64/kernel/setup.c b/arch/x86_64/kernel/setup.c index 476ee034fca2..40c77f6fe4b0 100644 --- a/arch/x86_64/kernel/setup.c +++ b/arch/x86_64/kernel/setup.c | |||
@@ -795,7 +795,7 @@ static void __init amd_detect_cmp(struct cpuinfo_x86 *c) | |||
795 | #endif | 795 | #endif |
796 | 796 | ||
797 | bits = 0; | 797 | bits = 0; |
798 | while ((1 << bits) < c->x86_num_cores) | 798 | while ((1 << bits) < c->x86_max_cores) |
799 | bits++; | 799 | bits++; |
800 | 800 | ||
801 | /* Low order bits define the core id (index of core in socket) */ | 801 | /* Low order bits define the core id (index of core in socket) */ |
@@ -828,7 +828,7 @@ static void __init amd_detect_cmp(struct cpuinfo_x86 *c) | |||
828 | numa_set_node(cpu, node); | 828 | numa_set_node(cpu, node); |
829 | 829 | ||
830 | printk(KERN_INFO "CPU %d(%d) -> Node %d -> Core %d\n", | 830 | printk(KERN_INFO "CPU %d(%d) -> Node %d -> Core %d\n", |
831 | cpu, c->x86_num_cores, node, cpu_core_id[cpu]); | 831 | cpu, c->x86_max_cores, node, cpu_core_id[cpu]); |
832 | #endif | 832 | #endif |
833 | #endif | 833 | #endif |
834 | } | 834 | } |
@@ -877,9 +877,9 @@ static int __init init_amd(struct cpuinfo_x86 *c) | |||
877 | display_cacheinfo(c); | 877 | display_cacheinfo(c); |
878 | 878 | ||
879 | if (c->extended_cpuid_level >= 0x80000008) { | 879 | if (c->extended_cpuid_level >= 0x80000008) { |
880 | c->x86_num_cores = (cpuid_ecx(0x80000008) & 0xff) + 1; | 880 | c->x86_max_cores = (cpuid_ecx(0x80000008) & 0xff) + 1; |
881 | if (c->x86_num_cores & (c->x86_num_cores - 1)) | 881 | if (c->x86_max_cores & (c->x86_max_cores - 1)) |
882 | c->x86_num_cores = 1; | 882 | c->x86_max_cores = 1; |
883 | 883 | ||
884 | amd_detect_cmp(c); | 884 | amd_detect_cmp(c); |
885 | } | 885 | } |
@@ -891,54 +891,44 @@ static void __cpuinit detect_ht(struct cpuinfo_x86 *c) | |||
891 | { | 891 | { |
892 | #ifdef CONFIG_SMP | 892 | #ifdef CONFIG_SMP |
893 | u32 eax, ebx, ecx, edx; | 893 | u32 eax, ebx, ecx, edx; |
894 | int index_msb, tmp; | 894 | int index_msb, core_bits; |
895 | int cpu = smp_processor_id(); | 895 | int cpu = smp_processor_id(); |
896 | 896 | ||
897 | cpuid(1, &eax, &ebx, &ecx, &edx); | ||
898 | |||
899 | c->apicid = phys_pkg_id(0); | ||
900 | |||
897 | if (!cpu_has(c, X86_FEATURE_HT) || cpu_has(c, X86_FEATURE_CMP_LEGACY)) | 901 | if (!cpu_has(c, X86_FEATURE_HT) || cpu_has(c, X86_FEATURE_CMP_LEGACY)) |
898 | return; | 902 | return; |
899 | 903 | ||
900 | cpuid(1, &eax, &ebx, &ecx, &edx); | ||
901 | smp_num_siblings = (ebx & 0xff0000) >> 16; | 904 | smp_num_siblings = (ebx & 0xff0000) >> 16; |
902 | 905 | ||
903 | if (smp_num_siblings == 1) { | 906 | if (smp_num_siblings == 1) { |
904 | printk(KERN_INFO "CPU: Hyper-Threading is disabled\n"); | 907 | printk(KERN_INFO "CPU: Hyper-Threading is disabled\n"); |
905 | } else if (smp_num_siblings > 1) { | 908 | } else if (smp_num_siblings > 1 ) { |
906 | index_msb = 31; | 909 | |
907 | /* | ||
908 | * At this point we only support two siblings per | ||
909 | * processor package. | ||
910 | */ | ||
911 | if (smp_num_siblings > NR_CPUS) { | 910 | if (smp_num_siblings > NR_CPUS) { |
912 | printk(KERN_WARNING "CPU: Unsupported number of the siblings %d", smp_num_siblings); | 911 | printk(KERN_WARNING "CPU: Unsupported number of the siblings %d", smp_num_siblings); |
913 | smp_num_siblings = 1; | 912 | smp_num_siblings = 1; |
914 | return; | 913 | return; |
915 | } | 914 | } |
916 | tmp = smp_num_siblings; | 915 | |
917 | while ((tmp & 0x80000000 ) == 0) { | 916 | index_msb = get_count_order(smp_num_siblings); |
918 | tmp <<=1 ; | ||
919 | index_msb--; | ||
920 | } | ||
921 | if (smp_num_siblings & (smp_num_siblings - 1)) | ||
922 | index_msb++; | ||
923 | phys_proc_id[cpu] = phys_pkg_id(index_msb); | 917 | phys_proc_id[cpu] = phys_pkg_id(index_msb); |
924 | 918 | ||
925 | printk(KERN_INFO "CPU: Physical Processor ID: %d\n", | 919 | printk(KERN_INFO "CPU: Physical Processor ID: %d\n", |
926 | phys_proc_id[cpu]); | 920 | phys_proc_id[cpu]); |
927 | 921 | ||
928 | smp_num_siblings = smp_num_siblings / c->x86_num_cores; | 922 | smp_num_siblings = smp_num_siblings / c->x86_max_cores; |
929 | 923 | ||
930 | tmp = smp_num_siblings; | 924 | index_msb = get_count_order(smp_num_siblings) ; |
931 | index_msb = 31; | 925 | |
932 | while ((tmp & 0x80000000) == 0) { | 926 | core_bits = get_count_order(c->x86_max_cores); |
933 | tmp <<=1 ; | ||
934 | index_msb--; | ||
935 | } | ||
936 | if (smp_num_siblings & (smp_num_siblings - 1)) | ||
937 | index_msb++; | ||
938 | 927 | ||
939 | cpu_core_id[cpu] = phys_pkg_id(index_msb); | 928 | cpu_core_id[cpu] = phys_pkg_id(index_msb) & |
929 | ((1 << core_bits) - 1); | ||
940 | 930 | ||
941 | if (c->x86_num_cores > 1) | 931 | if (c->x86_max_cores > 1) |
942 | printk(KERN_INFO "CPU: Processor Core ID: %d\n", | 932 | printk(KERN_INFO "CPU: Processor Core ID: %d\n", |
943 | cpu_core_id[cpu]); | 933 | cpu_core_id[cpu]); |
944 | } | 934 | } |
@@ -1006,7 +996,7 @@ static void __cpuinit init_intel(struct cpuinfo_x86 *c) | |||
1006 | c->x86_cache_alignment = c->x86_clflush_size * 2; | 996 | c->x86_cache_alignment = c->x86_clflush_size * 2; |
1007 | if (c->x86 >= 15) | 997 | if (c->x86 >= 15) |
1008 | set_bit(X86_FEATURE_CONSTANT_TSC, &c->x86_capability); | 998 | set_bit(X86_FEATURE_CONSTANT_TSC, &c->x86_capability); |
1009 | c->x86_num_cores = intel_num_cpu_cores(c); | 999 | c->x86_max_cores = intel_num_cpu_cores(c); |
1010 | 1000 | ||
1011 | srat_detect_node(); | 1001 | srat_detect_node(); |
1012 | } | 1002 | } |
@@ -1044,7 +1034,7 @@ void __cpuinit early_identify_cpu(struct cpuinfo_x86 *c) | |||
1044 | c->x86_model_id[0] = '\0'; /* Unset */ | 1034 | c->x86_model_id[0] = '\0'; /* Unset */ |
1045 | c->x86_clflush_size = 64; | 1035 | c->x86_clflush_size = 64; |
1046 | c->x86_cache_alignment = c->x86_clflush_size; | 1036 | c->x86_cache_alignment = c->x86_clflush_size; |
1047 | c->x86_num_cores = 1; | 1037 | c->x86_max_cores = 1; |
1048 | c->extended_cpuid_level = 0; | 1038 | c->extended_cpuid_level = 0; |
1049 | memset(&c->x86_capability, 0, sizeof c->x86_capability); | 1039 | memset(&c->x86_capability, 0, sizeof c->x86_capability); |
1050 | 1040 | ||
@@ -1278,13 +1268,12 @@ static int show_cpuinfo(struct seq_file *m, void *v) | |||
1278 | seq_printf(m, "cache size\t: %d KB\n", c->x86_cache_size); | 1268 | seq_printf(m, "cache size\t: %d KB\n", c->x86_cache_size); |
1279 | 1269 | ||
1280 | #ifdef CONFIG_SMP | 1270 | #ifdef CONFIG_SMP |
1281 | if (smp_num_siblings * c->x86_num_cores > 1) { | 1271 | if (smp_num_siblings * c->x86_max_cores > 1) { |
1282 | int cpu = c - cpu_data; | 1272 | int cpu = c - cpu_data; |
1283 | seq_printf(m, "physical id\t: %d\n", phys_proc_id[cpu]); | 1273 | seq_printf(m, "physical id\t: %d\n", phys_proc_id[cpu]); |
1284 | seq_printf(m, "siblings\t: %d\n", | 1274 | seq_printf(m, "siblings\t: %d\n", cpus_weight(cpu_core_map[cpu])); |
1285 | c->x86_num_cores * smp_num_siblings); | ||
1286 | seq_printf(m, "core id\t\t: %d\n", cpu_core_id[cpu]); | 1275 | seq_printf(m, "core id\t\t: %d\n", cpu_core_id[cpu]); |
1287 | seq_printf(m, "cpu cores\t: %d\n", c->x86_num_cores); | 1276 | seq_printf(m, "cpu cores\t: %d\n", c->booted_cores); |
1288 | } | 1277 | } |
1289 | #endif | 1278 | #endif |
1290 | 1279 | ||
diff --git a/arch/x86_64/kernel/smpboot.c b/arch/x86_64/kernel/smpboot.c index f74319a80659..2b9ddba61b37 100644 --- a/arch/x86_64/kernel/smpboot.c +++ b/arch/x86_64/kernel/smpboot.c | |||
@@ -64,6 +64,7 @@ | |||
64 | int smp_num_siblings = 1; | 64 | int smp_num_siblings = 1; |
65 | /* Package ID of each logical CPU */ | 65 | /* Package ID of each logical CPU */ |
66 | u8 phys_proc_id[NR_CPUS] __read_mostly = { [0 ... NR_CPUS-1] = BAD_APICID }; | 66 | u8 phys_proc_id[NR_CPUS] __read_mostly = { [0 ... NR_CPUS-1] = BAD_APICID }; |
67 | /* core ID of each logical CPU */ | ||
67 | u8 cpu_core_id[NR_CPUS] __read_mostly = { [0 ... NR_CPUS-1] = BAD_APICID }; | 68 | u8 cpu_core_id[NR_CPUS] __read_mostly = { [0 ... NR_CPUS-1] = BAD_APICID }; |
68 | EXPORT_SYMBOL(phys_proc_id); | 69 | EXPORT_SYMBOL(phys_proc_id); |
69 | EXPORT_SYMBOL(cpu_core_id); | 70 | EXPORT_SYMBOL(cpu_core_id); |
@@ -89,7 +90,10 @@ struct cpuinfo_x86 cpu_data[NR_CPUS] __cacheline_aligned; | |||
89 | /* Set when the idlers are all forked */ | 90 | /* Set when the idlers are all forked */ |
90 | int smp_threads_ready; | 91 | int smp_threads_ready; |
91 | 92 | ||
93 | /* representing HT siblings of each logical CPU */ | ||
92 | cpumask_t cpu_sibling_map[NR_CPUS] __read_mostly; | 94 | cpumask_t cpu_sibling_map[NR_CPUS] __read_mostly; |
95 | |||
96 | /* representing HT and core siblings of each logical CPU */ | ||
93 | cpumask_t cpu_core_map[NR_CPUS] __read_mostly; | 97 | cpumask_t cpu_core_map[NR_CPUS] __read_mostly; |
94 | EXPORT_SYMBOL(cpu_core_map); | 98 | EXPORT_SYMBOL(cpu_core_map); |
95 | 99 | ||
@@ -436,30 +440,59 @@ void __cpuinit smp_callin(void) | |||
436 | cpu_set(cpuid, cpu_callin_map); | 440 | cpu_set(cpuid, cpu_callin_map); |
437 | } | 441 | } |
438 | 442 | ||
443 | /* representing cpus for which sibling maps can be computed */ | ||
444 | static cpumask_t cpu_sibling_setup_map; | ||
445 | |||
439 | static inline void set_cpu_sibling_map(int cpu) | 446 | static inline void set_cpu_sibling_map(int cpu) |
440 | { | 447 | { |
441 | int i; | 448 | int i; |
449 | struct cpuinfo_x86 *c = cpu_data; | ||
450 | |||
451 | cpu_set(cpu, cpu_sibling_setup_map); | ||
442 | 452 | ||
443 | if (smp_num_siblings > 1) { | 453 | if (smp_num_siblings > 1) { |
444 | for_each_cpu(i) { | 454 | for_each_cpu_mask(i, cpu_sibling_setup_map) { |
445 | if (cpu_core_id[cpu] == cpu_core_id[i]) { | 455 | if (phys_proc_id[cpu] == phys_proc_id[i] && |
456 | cpu_core_id[cpu] == cpu_core_id[i]) { | ||
446 | cpu_set(i, cpu_sibling_map[cpu]); | 457 | cpu_set(i, cpu_sibling_map[cpu]); |
447 | cpu_set(cpu, cpu_sibling_map[i]); | 458 | cpu_set(cpu, cpu_sibling_map[i]); |
459 | cpu_set(i, cpu_core_map[cpu]); | ||
460 | cpu_set(cpu, cpu_core_map[i]); | ||
448 | } | 461 | } |
449 | } | 462 | } |
450 | } else { | 463 | } else { |
451 | cpu_set(cpu, cpu_sibling_map[cpu]); | 464 | cpu_set(cpu, cpu_sibling_map[cpu]); |
452 | } | 465 | } |
453 | 466 | ||
454 | if (current_cpu_data.x86_num_cores > 1) { | 467 | if (current_cpu_data.x86_max_cores == 1) { |
455 | for_each_cpu(i) { | ||
456 | if (phys_proc_id[cpu] == phys_proc_id[i]) { | ||
457 | cpu_set(i, cpu_core_map[cpu]); | ||
458 | cpu_set(cpu, cpu_core_map[i]); | ||
459 | } | ||
460 | } | ||
461 | } else { | ||
462 | cpu_core_map[cpu] = cpu_sibling_map[cpu]; | 468 | cpu_core_map[cpu] = cpu_sibling_map[cpu]; |
469 | c[cpu].booted_cores = 1; | ||
470 | return; | ||
471 | } | ||
472 | |||
473 | for_each_cpu_mask(i, cpu_sibling_setup_map) { | ||
474 | if (phys_proc_id[cpu] == phys_proc_id[i]) { | ||
475 | cpu_set(i, cpu_core_map[cpu]); | ||
476 | cpu_set(cpu, cpu_core_map[i]); | ||
477 | /* | ||
478 | * Does this new cpu bringup a new core? | ||
479 | */ | ||
480 | if (cpus_weight(cpu_sibling_map[cpu]) == 1) { | ||
481 | /* | ||
482 | * for each core in package, increment | ||
483 | * the booted_cores for this new cpu | ||
484 | */ | ||
485 | if (first_cpu(cpu_sibling_map[i]) == i) | ||
486 | c[cpu].booted_cores++; | ||
487 | /* | ||
488 | * increment the core count for all | ||
489 | * the other cpus in this package | ||
490 | */ | ||
491 | if (i != cpu) | ||
492 | c[i].booted_cores++; | ||
493 | } else if (i != cpu && !c[cpu].booted_cores) | ||
494 | c[cpu].booted_cores = c[i].booted_cores; | ||
495 | } | ||
463 | } | 496 | } |
464 | } | 497 | } |
465 | 498 | ||
@@ -993,6 +1026,7 @@ void __init smp_prepare_cpus(unsigned int max_cpus) | |||
993 | nmi_watchdog_default(); | 1026 | nmi_watchdog_default(); |
994 | current_cpu_data = boot_cpu_data; | 1027 | current_cpu_data = boot_cpu_data; |
995 | current_thread_info()->cpu = 0; /* needed? */ | 1028 | current_thread_info()->cpu = 0; /* needed? */ |
1029 | set_cpu_sibling_map(0); | ||
996 | 1030 | ||
997 | if (smp_sanity_check(max_cpus) < 0) { | 1031 | if (smp_sanity_check(max_cpus) < 0) { |
998 | printk(KERN_INFO "SMP disabled\n"); | 1032 | printk(KERN_INFO "SMP disabled\n"); |
@@ -1036,8 +1070,6 @@ void __init smp_prepare_boot_cpu(void) | |||
1036 | int me = smp_processor_id(); | 1070 | int me = smp_processor_id(); |
1037 | cpu_set(me, cpu_online_map); | 1071 | cpu_set(me, cpu_online_map); |
1038 | cpu_set(me, cpu_callout_map); | 1072 | cpu_set(me, cpu_callout_map); |
1039 | cpu_set(0, cpu_sibling_map[0]); | ||
1040 | cpu_set(0, cpu_core_map[0]); | ||
1041 | per_cpu(cpu_state, me) = CPU_ONLINE; | 1073 | per_cpu(cpu_state, me) = CPU_ONLINE; |
1042 | } | 1074 | } |
1043 | 1075 | ||
@@ -1106,15 +1138,24 @@ void __init smp_cpus_done(unsigned int max_cpus) | |||
1106 | static void remove_siblinginfo(int cpu) | 1138 | static void remove_siblinginfo(int cpu) |
1107 | { | 1139 | { |
1108 | int sibling; | 1140 | int sibling; |
1141 | struct cpuinfo_x86 *c = cpu_data; | ||
1109 | 1142 | ||
1143 | for_each_cpu_mask(sibling, cpu_core_map[cpu]) { | ||
1144 | cpu_clear(cpu, cpu_core_map[sibling]); | ||
1145 | /* | ||
1146 | * last thread sibling in this cpu core going down | ||
1147 | */ | ||
1148 | if (cpus_weight(cpu_sibling_map[cpu]) == 1) | ||
1149 | c[sibling].booted_cores--; | ||
1150 | } | ||
1151 | |||
1110 | for_each_cpu_mask(sibling, cpu_sibling_map[cpu]) | 1152 | for_each_cpu_mask(sibling, cpu_sibling_map[cpu]) |
1111 | cpu_clear(cpu, cpu_sibling_map[sibling]); | 1153 | cpu_clear(cpu, cpu_sibling_map[sibling]); |
1112 | for_each_cpu_mask(sibling, cpu_core_map[cpu]) | ||
1113 | cpu_clear(cpu, cpu_core_map[sibling]); | ||
1114 | cpus_clear(cpu_sibling_map[cpu]); | 1154 | cpus_clear(cpu_sibling_map[cpu]); |
1115 | cpus_clear(cpu_core_map[cpu]); | 1155 | cpus_clear(cpu_core_map[cpu]); |
1116 | phys_proc_id[cpu] = BAD_APICID; | 1156 | phys_proc_id[cpu] = BAD_APICID; |
1117 | cpu_core_id[cpu] = BAD_APICID; | 1157 | cpu_core_id[cpu] = BAD_APICID; |
1158 | cpu_clear(cpu, cpu_sibling_setup_map); | ||
1118 | } | 1159 | } |
1119 | 1160 | ||
1120 | void remove_cpu_from_maps(void) | 1161 | void remove_cpu_from_maps(void) |