aboutsummaryrefslogtreecommitdiffstats
path: root/arch
diff options
context:
space:
mode:
authorDavid S. Miller <davem@davemloft.net>2012-10-28 16:04:47 -0400
committerDavid S. Miller <davem@davemloft.net>2012-10-28 16:04:47 -0400
commit187818cd6a5ab6343eac47e52da2f3e40c544b98 (patch)
tree09b55dad78359ae7e047f3eb533844adaa4094c6 /arch
parente9b9eb59ffcdee09ec96b040f85c919618f4043e (diff)
sparc64: Improvde documentation and readability of atomic backoff code.
Document what's going on in asm/backoff.h with a large and descriptive comment. Refer to it above the cpu_relax() definition in asm/processor_64.h Rename the pause patching section to have "3insn" in it's name like the other patching sections do. Based upon feedback from Sam Ravnborg. Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'arch')
-rw-r--r--arch/sparc/include/asm/backoff.h42
-rw-r--r--arch/sparc/include/asm/processor_64.h7
-rw-r--r--arch/sparc/kernel/entry.h4
-rw-r--r--arch/sparc/kernel/setup_64.c4
-rw-r--r--arch/sparc/kernel/vmlinux.lds.S8
5 files changed, 55 insertions, 10 deletions
diff --git a/arch/sparc/include/asm/backoff.h b/arch/sparc/include/asm/backoff.h
index 20f01df0871b..4e02086b839c 100644
--- a/arch/sparc/include/asm/backoff.h
+++ b/arch/sparc/include/asm/backoff.h
@@ -1,6 +1,46 @@
1#ifndef _SPARC64_BACKOFF_H 1#ifndef _SPARC64_BACKOFF_H
2#define _SPARC64_BACKOFF_H 2#define _SPARC64_BACKOFF_H
3 3
4/* The macros in this file implement an exponential backoff facility
5 * for atomic operations.
6 *
7 * When multiple threads compete on an atomic operation, it is
8 * possible for one thread to be continually denied a successful
9 * completion of the compare-and-swap instruction. Heavily
10 * threaded cpu implementations like Niagara can compound this
11 * problem even further.
12 *
13 * When an atomic operation fails and needs to be retried, we spin a
14 * certain number of times. At each subsequent failure of the same
15 * operation we double the spin count, realizing an exponential
16 * backoff.
17 *
18 * When we spin, we try to use an operation that will cause the
19 * current cpu strand to block, and therefore make the core fully
20 * available to any other other runnable strands. There are two
21 * options, based upon cpu capabilities.
22 *
23 * On all cpus prior to SPARC-T4 we do three dummy reads of the
24 * condition code register. Each read blocks the strand for something
25 * between 40 and 50 cpu cycles.
26 *
27 * For SPARC-T4 and later we have a special "pause" instruction
28 * available. This is implemented using writes to register %asr27.
29 * The cpu will block the number of cycles written into the register,
30 * unless a disrupting trap happens first. SPARC-T4 specifically
31 * implements pause with a granularity of 8 cycles. Each strand has
32 * an internal pause counter which decrements every 8 cycles. So the
33 * chip shifts the %asr27 value down by 3 bits, and writes the result
34 * into the pause counter. If a value smaller than 8 is written, the
35 * chip blocks for 1 cycle.
36 *
37 * To achieve the same amount of backoff as the three %ccr reads give
38 * on earlier chips, we shift the backoff value up by 7 bits. (Three
39 * %ccr reads block for about 128 cycles, 1 << 7 == 128) We write the
40 * whole amount we want to block into the pause register, rather than
41 * loop writing 128 each time.
42 */
43
4#define BACKOFF_LIMIT (4 * 1024) 44#define BACKOFF_LIMIT (4 * 1024)
5 45
6#ifdef CONFIG_SMP 46#ifdef CONFIG_SMP
@@ -16,7 +56,7 @@
1688: rd %ccr, %g0; \ 5688: rd %ccr, %g0; \
17 rd %ccr, %g0; \ 57 rd %ccr, %g0; \
18 rd %ccr, %g0; \ 58 rd %ccr, %g0; \
19 .section .pause_patch,"ax"; \ 59 .section .pause_3insn_patch,"ax";\
20 .word 88b; \ 60 .word 88b; \
21 sllx tmp, 7, tmp; \ 61 sllx tmp, 7, tmp; \
22 wr tmp, 0, %asr27; \ 62 wr tmp, 0, %asr27; \
diff --git a/arch/sparc/include/asm/processor_64.h b/arch/sparc/include/asm/processor_64.h
index 9cdf52eec48a..721e25f0e2ea 100644
--- a/arch/sparc/include/asm/processor_64.h
+++ b/arch/sparc/include/asm/processor_64.h
@@ -196,11 +196,16 @@ extern unsigned long get_wchan(struct task_struct *task);
196#define KSTK_EIP(tsk) (task_pt_regs(tsk)->tpc) 196#define KSTK_EIP(tsk) (task_pt_regs(tsk)->tpc)
197#define KSTK_ESP(tsk) (task_pt_regs(tsk)->u_regs[UREG_FP]) 197#define KSTK_ESP(tsk) (task_pt_regs(tsk)->u_regs[UREG_FP])
198 198
199/* Please see the commentary in asm/backoff.h for a description of
200 * what these instructions are doing and how they have been choosen.
201 * To make a long story short, we are trying to yield the current cpu
202 * strand during busy loops.
203 */
199#define cpu_relax() asm volatile("\n99:\n\t" \ 204#define cpu_relax() asm volatile("\n99:\n\t" \
200 "rd %%ccr, %%g0\n\t" \ 205 "rd %%ccr, %%g0\n\t" \
201 "rd %%ccr, %%g0\n\t" \ 206 "rd %%ccr, %%g0\n\t" \
202 "rd %%ccr, %%g0\n\t" \ 207 "rd %%ccr, %%g0\n\t" \
203 ".section .pause_patch,\"ax\"\n\t"\ 208 ".section .pause_3insn_patch,\"ax\"\n\t"\
204 ".word 99b\n\t" \ 209 ".word 99b\n\t" \
205 "wr %%g0, 128, %%asr27\n\t" \ 210 "wr %%g0, 128, %%asr27\n\t" \
206 "nop\n\t" \ 211 "nop\n\t" \
diff --git a/arch/sparc/kernel/entry.h b/arch/sparc/kernel/entry.h
index 51742df63c75..cc3c5cb47cda 100644
--- a/arch/sparc/kernel/entry.h
+++ b/arch/sparc/kernel/entry.h
@@ -63,8 +63,8 @@ struct pause_patch_entry {
63 unsigned int addr; 63 unsigned int addr;
64 unsigned int insns[3]; 64 unsigned int insns[3];
65}; 65};
66extern struct pause_patch_entry __pause_patch, 66extern struct pause_patch_entry __pause_3insn_patch,
67 __pause_patch_end; 67 __pause_3insn_patch_end;
68 68
69extern void __init per_cpu_patch(void); 69extern void __init per_cpu_patch(void);
70extern void sun4v_patch_1insn_range(struct sun4v_1insn_patch_entry *, 70extern void sun4v_patch_1insn_range(struct sun4v_1insn_patch_entry *,
diff --git a/arch/sparc/kernel/setup_64.c b/arch/sparc/kernel/setup_64.c
index b45cff408de3..0eaf0059aaef 100644
--- a/arch/sparc/kernel/setup_64.c
+++ b/arch/sparc/kernel/setup_64.c
@@ -320,8 +320,8 @@ static void __init pause_patch(void)
320{ 320{
321 struct pause_patch_entry *p; 321 struct pause_patch_entry *p;
322 322
323 p = &__pause_patch; 323 p = &__pause_3insn_patch;
324 while (p < &__pause_patch_end) { 324 while (p < &__pause_3insn_patch_end) {
325 unsigned long i, addr = p->addr; 325 unsigned long i, addr = p->addr;
326 326
327 for (i = 0; i < 3; i++) { 327 for (i = 0; i < 3; i++) {
diff --git a/arch/sparc/kernel/vmlinux.lds.S b/arch/sparc/kernel/vmlinux.lds.S
index 847f9f793618..0bacceb19150 100644
--- a/arch/sparc/kernel/vmlinux.lds.S
+++ b/arch/sparc/kernel/vmlinux.lds.S
@@ -132,10 +132,10 @@ SECTIONS
132 *(.popc_6insn_patch) 132 *(.popc_6insn_patch)
133 __popc_6insn_patch_end = .; 133 __popc_6insn_patch_end = .;
134 } 134 }
135 .pause_patch : { 135 .pause_3insn_patch : {
136 __pause_patch = .; 136 __pause_3insn_patch = .;
137 *(.pause_patch) 137 *(.pause_3insn_patch)
138 __pause_patch_end = .; 138 __pause_3insn_patch_end = .;
139 } 139 }
140 PERCPU_SECTION(SMP_CACHE_BYTES) 140 PERCPU_SECTION(SMP_CACHE_BYTES)
141 141