diff options
author | Paul Mackerras <paulus@samba.org> | 2006-06-25 08:47:14 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@g5.osdl.org> | 2006-06-25 13:01:01 -0400 |
commit | bfe5d834195b3089b8846577311340376cc0f450 (patch) | |
tree | 52470de0fe87ff8372700e3472735cd5c14cee9d /kernel | |
parent | 6ceab8a936c302c0cea2bfe55617c76e2f5746fa (diff) |
[PATCH] Define __raw_get_cpu_var and use it
There are several instances of per_cpu(foo, raw_smp_processor_id()), which
is semantically equivalent to __get_cpu_var(foo) but without the warning
that smp_processor_id() can give if CONFIG_DEBUG_PREEMPT is enabled. For
those architectures with optimized per-cpu implementations, namely ia64,
powerpc, s390, sparc64 and x86_64, per_cpu() turns into more and slower
code than __get_cpu_var(), so it would be preferable to use __get_cpu_var
on those platforms.
This defines a __raw_get_cpu_var(x) macro which turns into per_cpu(x,
raw_smp_processor_id()) on architectures that use the generic per-cpu
implementation, and turns into __get_cpu_var(x) on the architectures that
have an optimized per-cpu implementation.
Signed-off-by: Paul Mackerras <paulus@samba.org>
Acked-by: David S. Miller <davem@davemloft.net>
Acked-by: Ingo Molnar <mingo@elte.hu>
Acked-by: Martin Schwidefsky <schwidefsky@de.ibm.com>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/hrtimer.c | 4 | ||||
-rw-r--r-- | kernel/sched.c | 4 | ||||
-rw-r--r-- | kernel/softlockup.c | 2 | ||||
-rw-r--r-- | kernel/timer.c | 2 |
4 files changed, 6 insertions, 6 deletions
diff --git a/kernel/hrtimer.c b/kernel/hrtimer.c index 18324305724a..9587aac72f4d 100644 --- a/kernel/hrtimer.c +++ b/kernel/hrtimer.c | |||
@@ -576,7 +576,7 @@ void hrtimer_init(struct hrtimer *timer, clockid_t clock_id, | |||
576 | 576 | ||
577 | memset(timer, 0, sizeof(struct hrtimer)); | 577 | memset(timer, 0, sizeof(struct hrtimer)); |
578 | 578 | ||
579 | bases = per_cpu(hrtimer_bases, raw_smp_processor_id()); | 579 | bases = __raw_get_cpu_var(hrtimer_bases); |
580 | 580 | ||
581 | if (clock_id == CLOCK_REALTIME && mode != HRTIMER_ABS) | 581 | if (clock_id == CLOCK_REALTIME && mode != HRTIMER_ABS) |
582 | clock_id = CLOCK_MONOTONIC; | 582 | clock_id = CLOCK_MONOTONIC; |
@@ -599,7 +599,7 @@ int hrtimer_get_res(const clockid_t which_clock, struct timespec *tp) | |||
599 | { | 599 | { |
600 | struct hrtimer_base *bases; | 600 | struct hrtimer_base *bases; |
601 | 601 | ||
602 | bases = per_cpu(hrtimer_bases, raw_smp_processor_id()); | 602 | bases = __raw_get_cpu_var(hrtimer_bases); |
603 | *tp = ktime_to_timespec(bases[which_clock].resolution); | 603 | *tp = ktime_to_timespec(bases[which_clock].resolution); |
604 | 604 | ||
605 | return 0; | 605 | return 0; |
diff --git a/kernel/sched.c b/kernel/sched.c index 5dbc42694477..f8d540b324ca 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -4152,7 +4152,7 @@ EXPORT_SYMBOL(yield); | |||
4152 | */ | 4152 | */ |
4153 | void __sched io_schedule(void) | 4153 | void __sched io_schedule(void) |
4154 | { | 4154 | { |
4155 | struct runqueue *rq = &per_cpu(runqueues, raw_smp_processor_id()); | 4155 | struct runqueue *rq = &__raw_get_cpu_var(runqueues); |
4156 | 4156 | ||
4157 | atomic_inc(&rq->nr_iowait); | 4157 | atomic_inc(&rq->nr_iowait); |
4158 | schedule(); | 4158 | schedule(); |
@@ -4163,7 +4163,7 @@ EXPORT_SYMBOL(io_schedule); | |||
4163 | 4163 | ||
4164 | long __sched io_schedule_timeout(long timeout) | 4164 | long __sched io_schedule_timeout(long timeout) |
4165 | { | 4165 | { |
4166 | struct runqueue *rq = &per_cpu(runqueues, raw_smp_processor_id()); | 4166 | struct runqueue *rq = &__raw_get_cpu_var(runqueues); |
4167 | long ret; | 4167 | long ret; |
4168 | 4168 | ||
4169 | atomic_inc(&rq->nr_iowait); | 4169 | atomic_inc(&rq->nr_iowait); |
diff --git a/kernel/softlockup.c b/kernel/softlockup.c index 14c7faf02909..2c1be1163edc 100644 --- a/kernel/softlockup.c +++ b/kernel/softlockup.c | |||
@@ -36,7 +36,7 @@ static struct notifier_block panic_block = { | |||
36 | 36 | ||
37 | void touch_softlockup_watchdog(void) | 37 | void touch_softlockup_watchdog(void) |
38 | { | 38 | { |
39 | per_cpu(touch_timestamp, raw_smp_processor_id()) = jiffies; | 39 | __raw_get_cpu_var(touch_timestamp) = jiffies; |
40 | } | 40 | } |
41 | EXPORT_SYMBOL(touch_softlockup_watchdog); | 41 | EXPORT_SYMBOL(touch_softlockup_watchdog); |
42 | 42 | ||
diff --git a/kernel/timer.c b/kernel/timer.c index f35b3939e937..eb97371b87d8 100644 --- a/kernel/timer.c +++ b/kernel/timer.c | |||
@@ -146,7 +146,7 @@ static void internal_add_timer(tvec_base_t *base, struct timer_list *timer) | |||
146 | void fastcall init_timer(struct timer_list *timer) | 146 | void fastcall init_timer(struct timer_list *timer) |
147 | { | 147 | { |
148 | timer->entry.next = NULL; | 148 | timer->entry.next = NULL; |
149 | timer->base = per_cpu(tvec_bases, raw_smp_processor_id()); | 149 | timer->base = __raw_get_cpu_var(tvec_bases); |
150 | } | 150 | } |
151 | EXPORT_SYMBOL(init_timer); | 151 | EXPORT_SYMBOL(init_timer); |
152 | 152 | ||