diff options
author | Anton Blanchard <anton@samba.org> | 2012-08-07 13:51:41 -0400 |
---|---|---|
committer | Benjamin Herrenschmidt <benh@kernel.crashing.org> | 2012-08-24 06:26:09 -0400 |
commit | 2fae7cdb60240e2e2d9b378afbf6d9fcce8a3890 (patch) | |
tree | b866b45dcd53a21646ab1bc10148988c158b21f9 /arch/powerpc | |
parent | dad477ccd65f05bf3b5a874e0118bf4156a1fcbb (diff) |
powerpc: Fix VMX in interrupt check in POWER7 copy loops
The enhanced prefetch hint patches corrupt the condition register
that was used to check if we are in interrupt. Fix this by using cr1.
Signed-off-by: Anton Blanchard <anton@samba.org>
Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
Diffstat (limited to 'arch/powerpc')
-rw-r--r-- | arch/powerpc/lib/copyuser_power7.S | 4 | ||||
-rw-r--r-- | arch/powerpc/lib/memcpy_power7.S | 4 |
2 files changed, 4 insertions, 4 deletions
diff --git a/arch/powerpc/lib/copyuser_power7.S b/arch/powerpc/lib/copyuser_power7.S index 4a4a46fbad0e..0d24ff15f5f6 100644 --- a/arch/powerpc/lib/copyuser_power7.S +++ b/arch/powerpc/lib/copyuser_power7.S | |||
@@ -288,7 +288,7 @@ err1; stb r0,0(r3) | |||
288 | std r0,16(r1) | 288 | std r0,16(r1) |
289 | stdu r1,-STACKFRAMESIZE(r1) | 289 | stdu r1,-STACKFRAMESIZE(r1) |
290 | bl .enter_vmx_usercopy | 290 | bl .enter_vmx_usercopy |
291 | cmpwi r3,0 | 291 | cmpwi cr1,r3,0 |
292 | ld r0,STACKFRAMESIZE+16(r1) | 292 | ld r0,STACKFRAMESIZE+16(r1) |
293 | ld r3,STACKFRAMESIZE+48(r1) | 293 | ld r3,STACKFRAMESIZE+48(r1) |
294 | ld r4,STACKFRAMESIZE+56(r1) | 294 | ld r4,STACKFRAMESIZE+56(r1) |
@@ -326,7 +326,7 @@ err1; stb r0,0(r3) | |||
326 | dcbt r0,r8,0b01010 /* GO */ | 326 | dcbt r0,r8,0b01010 /* GO */ |
327 | .machine pop | 327 | .machine pop |
328 | 328 | ||
329 | beq .Lunwind_stack_nonvmx_copy | 329 | beq cr1,.Lunwind_stack_nonvmx_copy |
330 | 330 | ||
331 | /* | 331 | /* |
332 | * If source and destination are not relatively aligned we use a | 332 | * If source and destination are not relatively aligned we use a |
diff --git a/arch/powerpc/lib/memcpy_power7.S b/arch/powerpc/lib/memcpy_power7.S index 0efdc51bc716..7ba6c96de778 100644 --- a/arch/powerpc/lib/memcpy_power7.S +++ b/arch/powerpc/lib/memcpy_power7.S | |||
@@ -222,7 +222,7 @@ _GLOBAL(memcpy_power7) | |||
222 | std r0,16(r1) | 222 | std r0,16(r1) |
223 | stdu r1,-STACKFRAMESIZE(r1) | 223 | stdu r1,-STACKFRAMESIZE(r1) |
224 | bl .enter_vmx_copy | 224 | bl .enter_vmx_copy |
225 | cmpwi r3,0 | 225 | cmpwi cr1,r3,0 |
226 | ld r0,STACKFRAMESIZE+16(r1) | 226 | ld r0,STACKFRAMESIZE+16(r1) |
227 | ld r3,STACKFRAMESIZE+48(r1) | 227 | ld r3,STACKFRAMESIZE+48(r1) |
228 | ld r4,STACKFRAMESIZE+56(r1) | 228 | ld r4,STACKFRAMESIZE+56(r1) |
@@ -260,7 +260,7 @@ _GLOBAL(memcpy_power7) | |||
260 | dcbt r0,r8,0b01010 /* GO */ | 260 | dcbt r0,r8,0b01010 /* GO */ |
261 | .machine pop | 261 | .machine pop |
262 | 262 | ||
263 | beq .Lunwind_stack_nonvmx_copy | 263 | beq cr1,.Lunwind_stack_nonvmx_copy |
264 | 264 | ||
265 | /* | 265 | /* |
266 | * If source and destination are not relatively aligned we use a | 266 | * If source and destination are not relatively aligned we use a |