diff options
author | Paul Mackerras <paulus@samba.org> | 2006-06-25 08:47:14 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@g5.osdl.org> | 2006-06-25 13:01:01 -0400 |
commit | bfe5d834195b3089b8846577311340376cc0f450 (patch) | |
tree | 52470de0fe87ff8372700e3472735cd5c14cee9d | |
parent | 6ceab8a936c302c0cea2bfe55617c76e2f5746fa (diff) |
[PATCH] Define __raw_get_cpu_var and use it
There are several instances of per_cpu(foo, raw_smp_processor_id()), which
is semantically equivalent to __get_cpu_var(foo) but without the warning
that smp_processor_id() can give if CONFIG_DEBUG_PREEMPT is enabled. For
those architectures with optimized per-cpu implementations, namely ia64,
powerpc, s390, sparc64 and x86_64, per_cpu() turns into more and slower
code than __get_cpu_var(), so it would be preferable to use __get_cpu_var
on those platforms.
This defines a __raw_get_cpu_var(x) macro which turns into per_cpu(x,
raw_smp_processor_id()) on architectures that use the generic per-cpu
implementation, and turns into __get_cpu_var(x) on the architectures that
have an optimized per-cpu implementation.
Signed-off-by: Paul Mackerras <paulus@samba.org>
Acked-by: David S. Miller <davem@davemloft.net>
Acked-by: Ingo Molnar <mingo@elte.hu>
Acked-by: Martin Schwidefsky <schwidefsky@de.ibm.com>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
-rw-r--r-- | include/asm-generic/percpu.h | 2 | ||||
-rw-r--r-- | include/asm-ia64/percpu.h | 2 | ||||
-rw-r--r-- | include/asm-powerpc/percpu.h | 2 | ||||
-rw-r--r-- | include/asm-s390/percpu.h | 2 | ||||
-rw-r--r-- | include/asm-sparc64/percpu.h | 2 | ||||
-rw-r--r-- | include/asm-x86_64/percpu.h | 2 | ||||
-rw-r--r-- | kernel/hrtimer.c | 4 | ||||
-rw-r--r-- | kernel/sched.c | 4 | ||||
-rw-r--r-- | kernel/softlockup.c | 2 | ||||
-rw-r--r-- | kernel/timer.c | 2 | ||||
-rw-r--r-- | net/ipv4/route.c | 2 |
11 files changed, 19 insertions, 7 deletions
diff --git a/include/asm-generic/percpu.h b/include/asm-generic/percpu.h index c0caf433a7d7..c74521157461 100644 --- a/include/asm-generic/percpu.h +++ b/include/asm-generic/percpu.h | |||
@@ -14,6 +14,7 @@ extern unsigned long __per_cpu_offset[NR_CPUS]; | |||
14 | /* var is in discarded region: offset to particular copy we want */ | 14 | /* var is in discarded region: offset to particular copy we want */ |
15 | #define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset[cpu])) | 15 | #define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset[cpu])) |
16 | #define __get_cpu_var(var) per_cpu(var, smp_processor_id()) | 16 | #define __get_cpu_var(var) per_cpu(var, smp_processor_id()) |
17 | #define __raw_get_cpu_var(var) per_cpu(var, raw_smp_processor_id()) | ||
17 | 18 | ||
18 | /* A macro to avoid #include hell... */ | 19 | /* A macro to avoid #include hell... */ |
19 | #define percpu_modcopy(pcpudst, src, size) \ | 20 | #define percpu_modcopy(pcpudst, src, size) \ |
@@ -30,6 +31,7 @@ do { \ | |||
30 | 31 | ||
31 | #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu__##var)) | 32 | #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu__##var)) |
32 | #define __get_cpu_var(var) per_cpu__##var | 33 | #define __get_cpu_var(var) per_cpu__##var |
34 | #define __raw_get_cpu_var(var) per_cpu__##var | ||
33 | 35 | ||
34 | #endif /* SMP */ | 36 | #endif /* SMP */ |
35 | 37 | ||
diff --git a/include/asm-ia64/percpu.h b/include/asm-ia64/percpu.h index ae357d504fba..24d898b650c5 100644 --- a/include/asm-ia64/percpu.h +++ b/include/asm-ia64/percpu.h | |||
@@ -42,6 +42,7 @@ DECLARE_PER_CPU(unsigned long, local_per_cpu_offset); | |||
42 | 42 | ||
43 | #define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset[cpu])) | 43 | #define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset[cpu])) |
44 | #define __get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __ia64_per_cpu_var(local_per_cpu_offset))) | 44 | #define __get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __ia64_per_cpu_var(local_per_cpu_offset))) |
45 | #define __raw_get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __ia64_per_cpu_var(local_per_cpu_offset))) | ||
45 | 46 | ||
46 | extern void percpu_modcopy(void *pcpudst, const void *src, unsigned long size); | 47 | extern void percpu_modcopy(void *pcpudst, const void *src, unsigned long size); |
47 | extern void setup_per_cpu_areas (void); | 48 | extern void setup_per_cpu_areas (void); |
@@ -51,6 +52,7 @@ extern void *per_cpu_init(void); | |||
51 | 52 | ||
52 | #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu__##var)) | 53 | #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu__##var)) |
53 | #define __get_cpu_var(var) per_cpu__##var | 54 | #define __get_cpu_var(var) per_cpu__##var |
55 | #define __raw_get_cpu_var(var) per_cpu__##var | ||
54 | #define per_cpu_init() (__phys_per_cpu_start) | 56 | #define per_cpu_init() (__phys_per_cpu_start) |
55 | 57 | ||
56 | #endif /* SMP */ | 58 | #endif /* SMP */ |
diff --git a/include/asm-powerpc/percpu.h b/include/asm-powerpc/percpu.h index 184a7a4d2fdf..faa1fc703053 100644 --- a/include/asm-powerpc/percpu.h +++ b/include/asm-powerpc/percpu.h | |||
@@ -22,6 +22,7 @@ | |||
22 | /* var is in discarded region: offset to particular copy we want */ | 22 | /* var is in discarded region: offset to particular copy we want */ |
23 | #define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset(cpu))) | 23 | #define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset(cpu))) |
24 | #define __get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __my_cpu_offset())) | 24 | #define __get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __my_cpu_offset())) |
25 | #define __raw_get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __my_cpu_offset())) | ||
25 | 26 | ||
26 | /* A macro to avoid #include hell... */ | 27 | /* A macro to avoid #include hell... */ |
27 | #define percpu_modcopy(pcpudst, src, size) \ | 28 | #define percpu_modcopy(pcpudst, src, size) \ |
@@ -41,6 +42,7 @@ extern void setup_per_cpu_areas(void); | |||
41 | 42 | ||
42 | #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu__##var)) | 43 | #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu__##var)) |
43 | #define __get_cpu_var(var) per_cpu__##var | 44 | #define __get_cpu_var(var) per_cpu__##var |
45 | #define __raw_get_cpu_var(var) per_cpu__##var | ||
44 | 46 | ||
45 | #endif /* SMP */ | 47 | #endif /* SMP */ |
46 | 48 | ||
diff --git a/include/asm-s390/percpu.h b/include/asm-s390/percpu.h index 436d216601e5..d9a8cca9b653 100644 --- a/include/asm-s390/percpu.h +++ b/include/asm-s390/percpu.h | |||
@@ -40,6 +40,7 @@ extern unsigned long __per_cpu_offset[NR_CPUS]; | |||
40 | __typeof__(type) per_cpu__##name | 40 | __typeof__(type) per_cpu__##name |
41 | 41 | ||
42 | #define __get_cpu_var(var) __reloc_hide(var,S390_lowcore.percpu_offset) | 42 | #define __get_cpu_var(var) __reloc_hide(var,S390_lowcore.percpu_offset) |
43 | #define __raw_get_cpu_var(var) __reloc_hide(var,S390_lowcore.percpu_offset) | ||
43 | #define per_cpu(var,cpu) __reloc_hide(var,__per_cpu_offset[cpu]) | 44 | #define per_cpu(var,cpu) __reloc_hide(var,__per_cpu_offset[cpu]) |
44 | 45 | ||
45 | /* A macro to avoid #include hell... */ | 46 | /* A macro to avoid #include hell... */ |
@@ -57,6 +58,7 @@ do { \ | |||
57 | __typeof__(type) per_cpu__##name | 58 | __typeof__(type) per_cpu__##name |
58 | 59 | ||
59 | #define __get_cpu_var(var) __reloc_hide(var,0) | 60 | #define __get_cpu_var(var) __reloc_hide(var,0) |
61 | #define __raw_get_cpu_var(var) __reloc_hide(var,0) | ||
60 | #define per_cpu(var,cpu) __reloc_hide(var,0) | 62 | #define per_cpu(var,cpu) __reloc_hide(var,0) |
61 | 63 | ||
62 | #endif /* SMP */ | 64 | #endif /* SMP */ |
diff --git a/include/asm-sparc64/percpu.h b/include/asm-sparc64/percpu.h index baef13b58952..a6ece06b83db 100644 --- a/include/asm-sparc64/percpu.h +++ b/include/asm-sparc64/percpu.h | |||
@@ -21,6 +21,7 @@ register unsigned long __local_per_cpu_offset asm("g5"); | |||
21 | /* var is in discarded region: offset to particular copy we want */ | 21 | /* var is in discarded region: offset to particular copy we want */ |
22 | #define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset(cpu))) | 22 | #define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset(cpu))) |
23 | #define __get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __local_per_cpu_offset)) | 23 | #define __get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __local_per_cpu_offset)) |
24 | #define __raw_get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __local_per_cpu_offset)) | ||
24 | 25 | ||
25 | /* A macro to avoid #include hell... */ | 26 | /* A macro to avoid #include hell... */ |
26 | #define percpu_modcopy(pcpudst, src, size) \ | 27 | #define percpu_modcopy(pcpudst, src, size) \ |
@@ -37,6 +38,7 @@ do { \ | |||
37 | 38 | ||
38 | #define per_cpu(var, cpu) (*((void)cpu, &per_cpu__##var)) | 39 | #define per_cpu(var, cpu) (*((void)cpu, &per_cpu__##var)) |
39 | #define __get_cpu_var(var) per_cpu__##var | 40 | #define __get_cpu_var(var) per_cpu__##var |
41 | #define __raw_get_cpu_var(var) per_cpu__##var | ||
40 | 42 | ||
41 | #endif /* SMP */ | 43 | #endif /* SMP */ |
42 | 44 | ||
diff --git a/include/asm-x86_64/percpu.h b/include/asm-x86_64/percpu.h index 7f33aaf9f7b1..549eb929b2c0 100644 --- a/include/asm-x86_64/percpu.h +++ b/include/asm-x86_64/percpu.h | |||
@@ -21,6 +21,7 @@ | |||
21 | /* var is in discarded region: offset to particular copy we want */ | 21 | /* var is in discarded region: offset to particular copy we want */ |
22 | #define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset(cpu))) | 22 | #define per_cpu(var, cpu) (*RELOC_HIDE(&per_cpu__##var, __per_cpu_offset(cpu))) |
23 | #define __get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __my_cpu_offset())) | 23 | #define __get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __my_cpu_offset())) |
24 | #define __raw_get_cpu_var(var) (*RELOC_HIDE(&per_cpu__##var, __my_cpu_offset())) | ||
24 | 25 | ||
25 | /* A macro to avoid #include hell... */ | 26 | /* A macro to avoid #include hell... */ |
26 | #define percpu_modcopy(pcpudst, src, size) \ | 27 | #define percpu_modcopy(pcpudst, src, size) \ |
@@ -40,6 +41,7 @@ extern void setup_per_cpu_areas(void); | |||
40 | 41 | ||
41 | #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu__##var)) | 42 | #define per_cpu(var, cpu) (*((void)(cpu), &per_cpu__##var)) |
42 | #define __get_cpu_var(var) per_cpu__##var | 43 | #define __get_cpu_var(var) per_cpu__##var |
44 | #define __raw_get_cpu_var(var) per_cpu__##var | ||
43 | 45 | ||
44 | #endif /* SMP */ | 46 | #endif /* SMP */ |
45 | 47 | ||
diff --git a/kernel/hrtimer.c b/kernel/hrtimer.c index 18324305724a..9587aac72f4d 100644 --- a/kernel/hrtimer.c +++ b/kernel/hrtimer.c | |||
@@ -576,7 +576,7 @@ void hrtimer_init(struct hrtimer *timer, clockid_t clock_id, | |||
576 | 576 | ||
577 | memset(timer, 0, sizeof(struct hrtimer)); | 577 | memset(timer, 0, sizeof(struct hrtimer)); |
578 | 578 | ||
579 | bases = per_cpu(hrtimer_bases, raw_smp_processor_id()); | 579 | bases = __raw_get_cpu_var(hrtimer_bases); |
580 | 580 | ||
581 | if (clock_id == CLOCK_REALTIME && mode != HRTIMER_ABS) | 581 | if (clock_id == CLOCK_REALTIME && mode != HRTIMER_ABS) |
582 | clock_id = CLOCK_MONOTONIC; | 582 | clock_id = CLOCK_MONOTONIC; |
@@ -599,7 +599,7 @@ int hrtimer_get_res(const clockid_t which_clock, struct timespec *tp) | |||
599 | { | 599 | { |
600 | struct hrtimer_base *bases; | 600 | struct hrtimer_base *bases; |
601 | 601 | ||
602 | bases = per_cpu(hrtimer_bases, raw_smp_processor_id()); | 602 | bases = __raw_get_cpu_var(hrtimer_bases); |
603 | *tp = ktime_to_timespec(bases[which_clock].resolution); | 603 | *tp = ktime_to_timespec(bases[which_clock].resolution); |
604 | 604 | ||
605 | return 0; | 605 | return 0; |
diff --git a/kernel/sched.c b/kernel/sched.c index 5dbc42694477..f8d540b324ca 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -4152,7 +4152,7 @@ EXPORT_SYMBOL(yield); | |||
4152 | */ | 4152 | */ |
4153 | void __sched io_schedule(void) | 4153 | void __sched io_schedule(void) |
4154 | { | 4154 | { |
4155 | struct runqueue *rq = &per_cpu(runqueues, raw_smp_processor_id()); | 4155 | struct runqueue *rq = &__raw_get_cpu_var(runqueues); |
4156 | 4156 | ||
4157 | atomic_inc(&rq->nr_iowait); | 4157 | atomic_inc(&rq->nr_iowait); |
4158 | schedule(); | 4158 | schedule(); |
@@ -4163,7 +4163,7 @@ EXPORT_SYMBOL(io_schedule); | |||
4163 | 4163 | ||
4164 | long __sched io_schedule_timeout(long timeout) | 4164 | long __sched io_schedule_timeout(long timeout) |
4165 | { | 4165 | { |
4166 | struct runqueue *rq = &per_cpu(runqueues, raw_smp_processor_id()); | 4166 | struct runqueue *rq = &__raw_get_cpu_var(runqueues); |
4167 | long ret; | 4167 | long ret; |
4168 | 4168 | ||
4169 | atomic_inc(&rq->nr_iowait); | 4169 | atomic_inc(&rq->nr_iowait); |
diff --git a/kernel/softlockup.c b/kernel/softlockup.c index 14c7faf02909..2c1be1163edc 100644 --- a/kernel/softlockup.c +++ b/kernel/softlockup.c | |||
@@ -36,7 +36,7 @@ static struct notifier_block panic_block = { | |||
36 | 36 | ||
37 | void touch_softlockup_watchdog(void) | 37 | void touch_softlockup_watchdog(void) |
38 | { | 38 | { |
39 | per_cpu(touch_timestamp, raw_smp_processor_id()) = jiffies; | 39 | __raw_get_cpu_var(touch_timestamp) = jiffies; |
40 | } | 40 | } |
41 | EXPORT_SYMBOL(touch_softlockup_watchdog); | 41 | EXPORT_SYMBOL(touch_softlockup_watchdog); |
42 | 42 | ||
diff --git a/kernel/timer.c b/kernel/timer.c index f35b3939e937..eb97371b87d8 100644 --- a/kernel/timer.c +++ b/kernel/timer.c | |||
@@ -146,7 +146,7 @@ static void internal_add_timer(tvec_base_t *base, struct timer_list *timer) | |||
146 | void fastcall init_timer(struct timer_list *timer) | 146 | void fastcall init_timer(struct timer_list *timer) |
147 | { | 147 | { |
148 | timer->entry.next = NULL; | 148 | timer->entry.next = NULL; |
149 | timer->base = per_cpu(tvec_bases, raw_smp_processor_id()); | 149 | timer->base = __raw_get_cpu_var(tvec_bases); |
150 | } | 150 | } |
151 | EXPORT_SYMBOL(init_timer); | 151 | EXPORT_SYMBOL(init_timer); |
152 | 152 | ||
diff --git a/net/ipv4/route.c b/net/ipv4/route.c index cc9423de7311..60b11aece5c3 100644 --- a/net/ipv4/route.c +++ b/net/ipv4/route.c | |||
@@ -244,7 +244,7 @@ static unsigned int rt_hash_rnd; | |||
244 | 244 | ||
245 | static DEFINE_PER_CPU(struct rt_cache_stat, rt_cache_stat); | 245 | static DEFINE_PER_CPU(struct rt_cache_stat, rt_cache_stat); |
246 | #define RT_CACHE_STAT_INC(field) \ | 246 | #define RT_CACHE_STAT_INC(field) \ |
247 | (per_cpu(rt_cache_stat, raw_smp_processor_id()).field++) | 247 | (__raw_get_cpu_var(rt_cache_stat).field++) |
248 | 248 | ||
249 | static int rt_intern_hash(unsigned hash, struct rtable *rth, | 249 | static int rt_intern_hash(unsigned hash, struct rtable *rth, |
250 | struct rtable **res); | 250 | struct rtable **res); |