diff options
author | Mike Travis <travis@sgi.com> | 2008-05-12 15:21:12 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2008-07-08 05:31:20 -0400 |
commit | 23ca4bba3e20c6c3cb11c1bb0ab4770b724d39ac (patch) | |
tree | 39ba5f7705e48717d7a6f2621b8ca7e7015c9802 /include/asm-x86/topology.h | |
parent | 1184dc2ffe2c8fb9afb766d870850f2c3165ef25 (diff) |
x86: cleanup early per cpu variables/accesses v4
* Introduce a new PER_CPU macro called "EARLY_PER_CPU". This is
used by some per_cpu variables that are initialized and accessed
before there are per_cpu areas allocated.
["Early" in respect to per_cpu variables is "earlier than the per_cpu
areas have been setup".]
This patchset adds these new macros:
DEFINE_EARLY_PER_CPU(_type, _name, _initvalue)
EXPORT_EARLY_PER_CPU_SYMBOL(_name)
DECLARE_EARLY_PER_CPU(_type, _name)
early_per_cpu_ptr(_name)
early_per_cpu_map(_name, _idx)
early_per_cpu(_name, _cpu)
The DEFINE macro defines the per_cpu variable as well as the early
map and pointer. It also initializes the per_cpu variable and map
elements to "_initvalue". The early_* macros provide access to
the initial map (usually setup during system init) and the early
pointer. This pointer is initialized to point to the early map
but is then NULL'ed when the actual per_cpu areas are setup. After
that the per_cpu variable is the correct access to the variable.
The early_per_cpu() macro is not very efficient but does show how to
access the variable if you have a function that can be called both
"early" and "late". It tests the early ptr to be NULL, and if not
then it's still valid. Otherwise, the per_cpu variable is used
instead:
#define early_per_cpu(_name, _cpu) \
(early_per_cpu_ptr(_name) ? \
early_per_cpu_ptr(_name)[_cpu] : \
per_cpu(_name, _cpu))
A better method is to actually check the pointer manually. In the
case below, numa_set_node can be called both "early" and "late":
void __cpuinit numa_set_node(int cpu, int node)
{
int *cpu_to_node_map = early_per_cpu_ptr(x86_cpu_to_node_map);
if (cpu_to_node_map)
cpu_to_node_map[cpu] = node;
else
per_cpu(x86_cpu_to_node_map, cpu) = node;
}
* Add a flag "arch_provides_topology_pointers" that indicates pointers
to topology cpumask_t maps are available. Otherwise, use the function
returning the cpumask_t value. This is useful if cpumask_t set size
is very large to avoid copying data on to/off of the stack.
* The coverage of CONFIG_DEBUG_PER_CPU_MAPS has been increased while
the non-debug case has been optimized a bit.
* Remove an unreferenced compiler warning in drivers/base/topology.c
* Clean up #ifdef in setup.c
For inclusion into sched-devel/latest tree.
Based on:
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux-2.6.git
+ sched-devel/latest .../mingo/linux-2.6-sched-devel.git
Signed-off-by: Mike Travis <travis@sgi.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Diffstat (limited to 'include/asm-x86/topology.h')
-rw-r--r-- | include/asm-x86/topology.h | 143 |
1 files changed, 82 insertions, 61 deletions
diff --git a/include/asm-x86/topology.h b/include/asm-x86/topology.h index dcf3f8131d6b..dac09cb66dca 100644 --- a/include/asm-x86/topology.h +++ b/include/asm-x86/topology.h | |||
@@ -35,87 +35,67 @@ | |||
35 | # endif | 35 | # endif |
36 | #endif | 36 | #endif |
37 | 37 | ||
38 | /* Node not present */ | ||
39 | #define NUMA_NO_NODE (-1) | ||
40 | |||
38 | #ifdef CONFIG_NUMA | 41 | #ifdef CONFIG_NUMA |
39 | #include <linux/cpumask.h> | 42 | #include <linux/cpumask.h> |
40 | #include <asm/mpspec.h> | 43 | #include <asm/mpspec.h> |
41 | 44 | ||
42 | /* Mappings between logical cpu number and node number */ | ||
43 | #ifdef CONFIG_X86_32 | 45 | #ifdef CONFIG_X86_32 |
44 | extern int cpu_to_node_map[]; | ||
45 | #else | ||
46 | /* Returns the number of the current Node. */ | ||
47 | #define numa_node_id() (early_cpu_to_node(raw_smp_processor_id())) | ||
48 | #endif | ||
49 | |||
50 | DECLARE_PER_CPU(int, x86_cpu_to_node_map); | ||
51 | |||
52 | #ifdef CONFIG_SMP | ||
53 | extern int x86_cpu_to_node_map_init[]; | ||
54 | extern void *x86_cpu_to_node_map_early_ptr; | ||
55 | #else | ||
56 | #define x86_cpu_to_node_map_early_ptr NULL | ||
57 | #endif | ||
58 | 46 | ||
47 | /* Mappings between node number and cpus on that node. */ | ||
59 | extern cpumask_t node_to_cpumask_map[]; | 48 | extern cpumask_t node_to_cpumask_map[]; |
60 | 49 | ||
61 | #define NUMA_NO_NODE (-1) | 50 | /* Mappings between logical cpu number and node number */ |
51 | extern int cpu_to_node_map[]; | ||
62 | 52 | ||
63 | /* Returns the number of the node containing CPU 'cpu' */ | 53 | /* Returns the number of the node containing CPU 'cpu' */ |
64 | #ifdef CONFIG_X86_32 | ||
65 | #define early_cpu_to_node(cpu) cpu_to_node(cpu) | ||
66 | static inline int cpu_to_node(int cpu) | 54 | static inline int cpu_to_node(int cpu) |
67 | { | 55 | { |
68 | return cpu_to_node_map[cpu]; | 56 | return cpu_to_node_map[cpu]; |
69 | } | 57 | } |
58 | #define early_cpu_to_node(cpu) cpu_to_node(cpu) | ||
70 | 59 | ||
71 | #else /* CONFIG_X86_64 */ | 60 | #else /* CONFIG_X86_64 */ |
72 | 61 | ||
73 | #ifdef CONFIG_SMP | 62 | /* Mappings between node number and cpus on that node. */ |
74 | static inline int early_cpu_to_node(int cpu) | 63 | extern cpumask_t node_to_cpumask_map[]; |
75 | { | 64 | |
76 | int *cpu_to_node_map = x86_cpu_to_node_map_early_ptr; | 65 | /* Mappings between logical cpu number and node number */ |
77 | 66 | DECLARE_EARLY_PER_CPU(int, x86_cpu_to_node_map); | |
78 | if (cpu_to_node_map) | 67 | |
79 | return cpu_to_node_map[cpu]; | 68 | /* Returns the number of the current Node. */ |
80 | else if (per_cpu_offset(cpu)) | 69 | #define numa_node_id() (per_cpu(x86_cpu_to_node_map, raw_smp_processor_id())) |
81 | return per_cpu(x86_cpu_to_node_map, cpu); | 70 | |
82 | else | 71 | #ifdef CONFIG_DEBUG_PER_CPU_MAPS |
83 | return NUMA_NO_NODE; | 72 | extern int cpu_to_node(int cpu); |
84 | } | 73 | extern int early_cpu_to_node(int cpu); |
85 | #else | 74 | extern cpumask_t *_node_to_cpumask_ptr(int node); |
86 | #define early_cpu_to_node(cpu) cpu_to_node(cpu) | 75 | extern cpumask_t node_to_cpumask(int node); |
87 | #endif | ||
88 | 76 | ||
77 | #else /* !CONFIG_DEBUG_PER_CPU_MAPS */ | ||
78 | |||
79 | /* Returns the number of the node containing CPU 'cpu' */ | ||
89 | static inline int cpu_to_node(int cpu) | 80 | static inline int cpu_to_node(int cpu) |
90 | { | 81 | { |
91 | #ifdef CONFIG_DEBUG_PER_CPU_MAPS | ||
92 | if (x86_cpu_to_node_map_early_ptr) { | ||
93 | printk("KERN_NOTICE cpu_to_node(%d): usage too early!\n", | ||
94 | (int)cpu); | ||
95 | dump_stack(); | ||
96 | return ((int *)x86_cpu_to_node_map_early_ptr)[cpu]; | ||
97 | } | ||
98 | #endif | ||
99 | return per_cpu(x86_cpu_to_node_map, cpu); | 82 | return per_cpu(x86_cpu_to_node_map, cpu); |
100 | } | 83 | } |
101 | 84 | ||
102 | #ifdef CONFIG_NUMA | 85 | /* Same function but used if called before per_cpu areas are setup */ |
103 | 86 | static inline int early_cpu_to_node(int cpu) | |
104 | /* Returns a pointer to the cpumask of CPUs on Node 'node'. */ | 87 | { |
105 | #define node_to_cpumask_ptr(v, node) \ | 88 | if (early_per_cpu_ptr(x86_cpu_to_node_map)) |
106 | cpumask_t *v = &(node_to_cpumask_map[node]) | 89 | return early_per_cpu_ptr(x86_cpu_to_node_map)[cpu]; |
107 | |||
108 | #define node_to_cpumask_ptr_next(v, node) \ | ||
109 | v = &(node_to_cpumask_map[node]) | ||
110 | #endif | ||
111 | 90 | ||
112 | #endif /* CONFIG_X86_64 */ | 91 | return per_cpu(x86_cpu_to_node_map, cpu); |
92 | } | ||
113 | 93 | ||
114 | /* | 94 | /* Returns a pointer to the cpumask of CPUs on Node 'node'. */ |
115 | * Returns the number of the node containing Node 'node'. This | 95 | static inline cpumask_t *_node_to_cpumask_ptr(int node) |
116 | * architecture is flat, so it is a pretty simple function! | 96 | { |
117 | */ | 97 | return &node_to_cpumask_map[node]; |
118 | #define parent_node(node) (node) | 98 | } |
119 | 99 | ||
120 | /* Returns a bitmask of CPUs on Node 'node'. */ | 100 | /* Returns a bitmask of CPUs on Node 'node'. */ |
121 | static inline cpumask_t node_to_cpumask(int node) | 101 | static inline cpumask_t node_to_cpumask(int node) |
@@ -123,14 +103,29 @@ static inline cpumask_t node_to_cpumask(int node) | |||
123 | return node_to_cpumask_map[node]; | 103 | return node_to_cpumask_map[node]; |
124 | } | 104 | } |
125 | 105 | ||
106 | #endif /* !CONFIG_DEBUG_PER_CPU_MAPS */ | ||
107 | #endif /* CONFIG_X86_64 */ | ||
108 | |||
109 | /* Replace default node_to_cpumask_ptr with optimized version */ | ||
110 | #define node_to_cpumask_ptr(v, node) \ | ||
111 | cpumask_t *v = _node_to_cpumask_ptr(node) | ||
112 | |||
113 | #define node_to_cpumask_ptr_next(v, node) \ | ||
114 | v = _node_to_cpumask_ptr(node) | ||
115 | |||
126 | /* Returns the number of the first CPU on Node 'node'. */ | 116 | /* Returns the number of the first CPU on Node 'node'. */ |
127 | static inline int node_to_first_cpu(int node) | 117 | static inline int node_to_first_cpu(int node) |
128 | { | 118 | { |
129 | cpumask_t mask = node_to_cpumask(node); | 119 | node_to_cpumask_ptr(mask, node); |
130 | 120 | return first_cpu(*mask); | |
131 | return first_cpu(mask); | ||
132 | } | 121 | } |
133 | 122 | ||
123 | /* | ||
124 | * Returns the number of the node containing Node 'node'. This | ||
125 | * architecture is flat, so it is a pretty simple function! | ||
126 | */ | ||
127 | #define parent_node(node) (node) | ||
128 | |||
134 | #define pcibus_to_node(bus) __pcibus_to_node(bus) | 129 | #define pcibus_to_node(bus) __pcibus_to_node(bus) |
135 | #define pcibus_to_cpumask(bus) __pcibus_to_cpumask(bus) | 130 | #define pcibus_to_cpumask(bus) __pcibus_to_cpumask(bus) |
136 | 131 | ||
@@ -180,8 +175,31 @@ extern int __node_distance(int, int); | |||
180 | #define node_distance(a, b) __node_distance(a, b) | 175 | #define node_distance(a, b) __node_distance(a, b) |
181 | #endif | 176 | #endif |
182 | 177 | ||
183 | #else /* CONFIG_NUMA */ | 178 | #else /* !CONFIG_NUMA */ |
179 | |||
180 | #define numa_node_id() 0 | ||
181 | #define cpu_to_node(cpu) 0 | ||
182 | #define early_cpu_to_node(cpu) 0 | ||
183 | |||
184 | static inline cpumask_t *_node_to_cpumask_ptr(int node) | ||
185 | { | ||
186 | return &cpu_online_map; | ||
187 | } | ||
188 | static inline cpumask_t node_to_cpumask(int node) | ||
189 | { | ||
190 | return cpu_online_map; | ||
191 | } | ||
192 | static inline int node_to_first_cpu(int node) | ||
193 | { | ||
194 | return first_cpu(cpu_online_map); | ||
195 | } | ||
196 | |||
197 | /* Replace default node_to_cpumask_ptr with optimized version */ | ||
198 | #define node_to_cpumask_ptr(v, node) \ | ||
199 | cpumask_t *v = _node_to_cpumask_ptr(node) | ||
184 | 200 | ||
201 | #define node_to_cpumask_ptr_next(v, node) \ | ||
202 | v = _node_to_cpumask_ptr(node) | ||
185 | #endif | 203 | #endif |
186 | 204 | ||
187 | #include <asm-generic/topology.h> | 205 | #include <asm-generic/topology.h> |
@@ -193,6 +211,9 @@ extern cpumask_t cpu_coregroup_map(int cpu); | |||
193 | #define topology_core_id(cpu) (cpu_data(cpu).cpu_core_id) | 211 | #define topology_core_id(cpu) (cpu_data(cpu).cpu_core_id) |
194 | #define topology_core_siblings(cpu) (per_cpu(cpu_core_map, cpu)) | 212 | #define topology_core_siblings(cpu) (per_cpu(cpu_core_map, cpu)) |
195 | #define topology_thread_siblings(cpu) (per_cpu(cpu_sibling_map, cpu)) | 213 | #define topology_thread_siblings(cpu) (per_cpu(cpu_sibling_map, cpu)) |
214 | |||
215 | /* indicates that pointers to the topology cpumask_t maps are valid */ | ||
216 | #define arch_provides_topology_pointers yes | ||
196 | #endif | 217 | #endif |
197 | 218 | ||
198 | static inline void arch_fix_phys_package_id(int num, u32 slot) | 219 | static inline void arch_fix_phys_package_id(int num, u32 slot) |
@@ -220,4 +241,4 @@ static inline void set_mp_bus_to_node(int busnum, int node) | |||
220 | } | 241 | } |
221 | #endif | 242 | #endif |
222 | 243 | ||
223 | #endif | 244 | #endif /* _ASM_X86_TOPOLOGY_H */ |