aboutsummaryrefslogtreecommitdiffstats
path: root/arch/powerpc/lib
diff options
context:
space:
mode:
authorAnton Blanchard <anton@samba.org>2012-05-29 15:33:12 -0400
committerBenjamin Herrenschmidt <benh@kernel.crashing.org>2012-07-03 00:14:44 -0400
commitfde69282b7ba2701560764b81ebb756deb98cf2b (patch)
treee350f7d55f90885e9ea9ee26dcb020f0fbcbaa4e /arch/powerpc/lib
parent6f7839e542ee18770288be75114bd2e6771e1421 (diff)
powerpc: POWER7 optimised copy_page using VMX and enhanced prefetch
Implement a POWER7 optimised copy_page using VMX and enhanced prefetch instructions. We use enhanced prefetch hints to prefetch both the load and store side. We copy a cacheline at a time and fall back to regular loads and stores if we are unable to use VMX (eg we are in an interrupt). The following microbenchmark was used to assess the impact of the patch: http://ozlabs.org/~anton/junkcode/page_fault_file.c We test MAP_PRIVATE page faults across a 1GB file, 100 times: # time ./page_fault_file -p -l 1G -i 100 Before: 22.25s After: 18.89s 17% faster Signed-off-by: Anton Blanchard <anton@samba.org> Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Diffstat (limited to 'arch/powerpc/lib')
-rw-r--r--arch/powerpc/lib/Makefile2
-rw-r--r--arch/powerpc/lib/copypage_64.S4
-rw-r--r--arch/powerpc/lib/copypage_power7.S168
-rw-r--r--arch/powerpc/lib/vmx-helper.c23
4 files changed, 196 insertions, 1 deletions
diff --git a/arch/powerpc/lib/Makefile b/arch/powerpc/lib/Makefile
index 1eb94c7af6c7..873805ec3ed6 100644
--- a/arch/powerpc/lib/Makefile
+++ b/arch/powerpc/lib/Makefile
@@ -17,7 +17,7 @@ obj-$(CONFIG_HAS_IOMEM) += devres.o
17obj-$(CONFIG_PPC64) += copypage_64.o copyuser_64.o \ 17obj-$(CONFIG_PPC64) += copypage_64.o copyuser_64.o \
18 memcpy_64.o usercopy_64.o mem_64.o string.o \ 18 memcpy_64.o usercopy_64.o mem_64.o string.o \
19 checksum_wrappers_64.o hweight_64.o \ 19 checksum_wrappers_64.o hweight_64.o \
20 copyuser_power7.o string_64.o 20 copyuser_power7.o string_64.o copypage_power7.o
21obj-$(CONFIG_XMON) += sstep.o ldstfp.o 21obj-$(CONFIG_XMON) += sstep.o ldstfp.o
22obj-$(CONFIG_KPROBES) += sstep.o ldstfp.o 22obj-$(CONFIG_KPROBES) += sstep.o ldstfp.o
23obj-$(CONFIG_HAVE_HW_BREAKPOINT) += sstep.o ldstfp.o 23obj-$(CONFIG_HAVE_HW_BREAKPOINT) += sstep.o ldstfp.o
diff --git a/arch/powerpc/lib/copypage_64.S b/arch/powerpc/lib/copypage_64.S
index 53dcb6b1b708..9f9434a85264 100644
--- a/arch/powerpc/lib/copypage_64.S
+++ b/arch/powerpc/lib/copypage_64.S
@@ -17,7 +17,11 @@ PPC64_CACHES:
17 .section ".text" 17 .section ".text"
18 18
19_GLOBAL(copy_page) 19_GLOBAL(copy_page)
20BEGIN_FTR_SECTION
20 lis r5,PAGE_SIZE@h 21 lis r5,PAGE_SIZE@h
22FTR_SECTION_ELSE
23 b .copypage_power7
24ALT_FTR_SECTION_END_IFCLR(CPU_FTR_VMX_COPY)
21 ori r5,r5,PAGE_SIZE@l 25 ori r5,r5,PAGE_SIZE@l
22BEGIN_FTR_SECTION 26BEGIN_FTR_SECTION
23 ld r10,PPC64_CACHES@toc(r2) 27 ld r10,PPC64_CACHES@toc(r2)
diff --git a/arch/powerpc/lib/copypage_power7.S b/arch/powerpc/lib/copypage_power7.S
new file mode 100644
index 000000000000..01e2b5db325f
--- /dev/null
+++ b/arch/powerpc/lib/copypage_power7.S
@@ -0,0 +1,168 @@
1/*
2 * This program is free software; you can redistribute it and/or modify
3 * it under the terms of the GNU General Public License as published by
4 * the Free Software Foundation; either version 2 of the License, or
5 * (at your option) any later version.
6 *
7 * This program is distributed in the hope that it will be useful,
8 * but WITHOUT ANY WARRANTY; without even the implied warranty of
9 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
10 * GNU General Public License for more details.
11 *
12 * You should have received a copy of the GNU General Public License
13 * along with this program; if not, write to the Free Software
14 * Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
15 *
16 * Copyright (C) IBM Corporation, 2012
17 *
18 * Author: Anton Blanchard <anton@au.ibm.com>
19 */
20#include <asm/page.h>
21#include <asm/ppc_asm.h>
22
23#define STACKFRAMESIZE 256
24#define STK_REG(i) (112 + ((i)-14)*8)
25
26_GLOBAL(copypage_power7)
27 /*
28 * We prefetch both the source and destination using enhanced touch
29 * instructions. We use a stream ID of 0 for the load side and
30 * 1 for the store side. Since source and destination are page
31 * aligned we don't need to clear the bottom 7 bits of either
32 * address.
33 */
34 ori r9,r3,1 /* stream=1 */
35
36#ifdef CONFIG_PPC_64K_PAGES
37 lis r7,0x0E01 /* depth=7, units=512 */
38#else
39 lis r7,0x0E00 /* depth=7 */
40 ori r7,r7,0x1000 /* units=32 */
41#endif
42 ori r10,r7,1 /* stream=1 */
43
44 lis r8,0x8000 /* GO=1 */
45 clrldi r8,r8,32
46
47.machine push
48.machine "power4"
49 dcbt r0,r4,0b01000
50 dcbt r0,r7,0b01010
51 dcbtst r0,r9,0b01000
52 dcbtst r0,r10,0b01010
53 eieio
54 dcbt r0,r8,0b01010 /* GO */
55.machine pop
56
57#ifdef CONFIG_ALTIVEC
58 mflr r0
59 std r3,48(r1)
60 std r4,56(r1)
61 std r0,16(r1)
62 stdu r1,-STACKFRAMESIZE(r1)
63 bl .enter_vmx_copy
64 cmpwi r3,0
65 ld r0,STACKFRAMESIZE+16(r1)
66 ld r3,STACKFRAMESIZE+48(r1)
67 ld r4,STACKFRAMESIZE+56(r1)
68 mtlr r0
69
70 li r0,(PAGE_SIZE/128)
71 mtctr r0
72
73 beq .Lnonvmx_copy
74
75 addi r1,r1,STACKFRAMESIZE
76
77 li r6,16
78 li r7,32
79 li r8,48
80 li r9,64
81 li r10,80
82 li r11,96
83 li r12,112
84
85 .align 5
861: lvx vr7,r0,r4
87 lvx vr6,r4,r6
88 lvx vr5,r4,r7
89 lvx vr4,r4,r8
90 lvx vr3,r4,r9
91 lvx vr2,r4,r10
92 lvx vr1,r4,r11
93 lvx vr0,r4,r12
94 addi r4,r4,128
95 stvx vr7,r0,r3
96 stvx vr6,r3,r6
97 stvx vr5,r3,r7
98 stvx vr4,r3,r8
99 stvx vr3,r3,r9
100 stvx vr2,r3,r10
101 stvx vr1,r3,r11
102 stvx vr0,r3,r12
103 addi r3,r3,128
104 bdnz 1b
105
106 b .exit_vmx_copy /* tail call optimise */
107
108#else
109 li r0,(PAGE_SIZE/128)
110 mtctr r0
111
112 stdu r1,-STACKFRAMESIZE(r1)
113#endif
114
115.Lnonvmx_copy:
116 std r14,STK_REG(r14)(r1)
117 std r15,STK_REG(r15)(r1)
118 std r16,STK_REG(r16)(r1)
119 std r17,STK_REG(r17)(r1)
120 std r18,STK_REG(r18)(r1)
121 std r19,STK_REG(r19)(r1)
122 std r20,STK_REG(r20)(r1)
123
1241: ld r0,0(r4)
125 ld r5,8(r4)
126 ld r6,16(r4)
127 ld r7,24(r4)
128 ld r8,32(r4)
129 ld r9,40(r4)
130 ld r10,48(r4)
131 ld r11,56(r4)
132 ld r12,64(r4)
133 ld r14,72(r4)
134 ld r15,80(r4)
135 ld r16,88(r4)
136 ld r17,96(r4)
137 ld r18,104(r4)
138 ld r19,112(r4)
139 ld r20,120(r4)
140 addi r4,r4,128
141 std r0,0(r3)
142 std r5,8(r3)
143 std r6,16(r3)
144 std r7,24(r3)
145 std r8,32(r3)
146 std r9,40(r3)
147 std r10,48(r3)
148 std r11,56(r3)
149 std r12,64(r3)
150 std r14,72(r3)
151 std r15,80(r3)
152 std r16,88(r3)
153 std r17,96(r3)
154 std r18,104(r3)
155 std r19,112(r3)
156 std r20,120(r3)
157 addi r3,r3,128
158 bdnz 1b
159
160 ld r14,STK_REG(r14)(r1)
161 ld r15,STK_REG(r15)(r1)
162 ld r16,STK_REG(r16)(r1)
163 ld r17,STK_REG(r17)(r1)
164 ld r18,STK_REG(r18)(r1)
165 ld r19,STK_REG(r19)(r1)
166 ld r20,STK_REG(r20)(r1)
167 addi r1,r1,STACKFRAMESIZE
168 blr
diff --git a/arch/powerpc/lib/vmx-helper.c b/arch/powerpc/lib/vmx-helper.c
index 753a839f4a14..3cf529ceec5b 100644
--- a/arch/powerpc/lib/vmx-helper.c
+++ b/arch/powerpc/lib/vmx-helper.c
@@ -49,3 +49,26 @@ int exit_vmx_usercopy(void)
49 pagefault_enable(); 49 pagefault_enable();
50 return 0; 50 return 0;
51} 51}
52
53int enter_vmx_copy(void)
54{
55 if (in_interrupt())
56 return 0;
57
58 preempt_disable();
59
60 enable_kernel_altivec();
61
62 return 1;
63}
64
65/*
66 * All calls to this function will be optimised into tail calls. We are
67 * passed a pointer to the destination which we return as required by a
68 * memcpy implementation.
69 */
70void *exit_vmx_copy(void *dest)
71{
72 preempt_enable();
73 return dest;
74}