diff options
| author | David S. Miller <davem@davemloft.net> | 2016-08-15 19:07:50 -0400 |
|---|---|---|
| committer | David S. Miller <davem@davemloft.net> | 2016-10-24 14:32:12 -0400 |
| commit | cb736fdbb208eb3420f1a2eb2bfc024a6e9dcada (patch) | |
| tree | fe745222354d12a358e35e5df8d6ac4afebff3ee /arch/sparc/lib | |
| parent | d0796b555ba60c22eb41ae39a8362156cb08eee9 (diff) | |
sparc64: Convert U1copy_{from,to}_user to accurate exception reporting.
Report the exact number of bytes which have not been successfully
copied when an exception occurs, using the running remaining length.
Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'arch/sparc/lib')
| -rw-r--r-- | arch/sparc/lib/U1copy_from_user.S | 8 | ||||
| -rw-r--r-- | arch/sparc/lib/U1copy_to_user.S | 8 | ||||
| -rw-r--r-- | arch/sparc/lib/U1memcpy.S | 345 |
3 files changed, 237 insertions, 124 deletions
diff --git a/arch/sparc/lib/U1copy_from_user.S b/arch/sparc/lib/U1copy_from_user.S index 1ad59fbac7a7..bb6ff73229e3 100644 --- a/arch/sparc/lib/U1copy_from_user.S +++ b/arch/sparc/lib/U1copy_from_user.S | |||
| @@ -3,19 +3,19 @@ | |||
| 3 | * Copyright (C) 1999, 2000, 2004 David S. Miller (davem@redhat.com) | 3 | * Copyright (C) 1999, 2000, 2004 David S. Miller (davem@redhat.com) |
| 4 | */ | 4 | */ |
| 5 | 5 | ||
| 6 | #define EX_LD(x) \ | 6 | #define EX_LD(x,y) \ |
| 7 | 98: x; \ | 7 | 98: x; \ |
| 8 | .section __ex_table,"a";\ | 8 | .section __ex_table,"a";\ |
| 9 | .align 4; \ | 9 | .align 4; \ |
| 10 | .word 98b, __retl_mone; \ | 10 | .word 98b, y; \ |
| 11 | .text; \ | 11 | .text; \ |
| 12 | .align 4; | 12 | .align 4; |
| 13 | 13 | ||
| 14 | #define EX_LD_FP(x) \ | 14 | #define EX_LD_FP(x,y) \ |
| 15 | 98: x; \ | 15 | 98: x; \ |
| 16 | .section __ex_table,"a";\ | 16 | .section __ex_table,"a";\ |
| 17 | .align 4; \ | 17 | .align 4; \ |
| 18 | .word 98b, __retl_mone_fp;\ | 18 | .word 98b, y; \ |
| 19 | .text; \ | 19 | .text; \ |
| 20 | .align 4; | 20 | .align 4; |
| 21 | 21 | ||
diff --git a/arch/sparc/lib/U1copy_to_user.S b/arch/sparc/lib/U1copy_to_user.S index adcc3a510185..ed92ce739558 100644 --- a/arch/sparc/lib/U1copy_to_user.S +++ b/arch/sparc/lib/U1copy_to_user.S | |||
| @@ -3,19 +3,19 @@ | |||
| 3 | * Copyright (C) 1999, 2000, 2004 David S. Miller (davem@redhat.com) | 3 | * Copyright (C) 1999, 2000, 2004 David S. Miller (davem@redhat.com) |
| 4 | */ | 4 | */ |
| 5 | 5 | ||
| 6 | #define EX_ST(x) \ | 6 | #define EX_ST(x,y) \ |
| 7 | 98: x; \ | 7 | 98: x; \ |
| 8 | .section __ex_table,"a";\ | 8 | .section __ex_table,"a";\ |
| 9 | .align 4; \ | 9 | .align 4; \ |
| 10 | .word 98b, __retl_mone; \ | 10 | .word 98b, y; \ |
| 11 | .text; \ | 11 | .text; \ |
| 12 | .align 4; | 12 | .align 4; |
| 13 | 13 | ||
| 14 | #define EX_ST_FP(x) \ | 14 | #define EX_ST_FP(x,y) \ |
| 15 | 98: x; \ | 15 | 98: x; \ |
| 16 | .section __ex_table,"a";\ | 16 | .section __ex_table,"a";\ |
| 17 | .align 4; \ | 17 | .align 4; \ |
| 18 | .word 98b, __retl_mone_fp;\ | 18 | .word 98b, y; \ |
| 19 | .text; \ | 19 | .text; \ |
| 20 | .align 4; | 20 | .align 4; |
| 21 | 21 | ||
diff --git a/arch/sparc/lib/U1memcpy.S b/arch/sparc/lib/U1memcpy.S index 97e1b211090c..4f0d50b33a72 100644 --- a/arch/sparc/lib/U1memcpy.S +++ b/arch/sparc/lib/U1memcpy.S | |||
| @@ -5,6 +5,7 @@ | |||
| 5 | */ | 5 | */ |
| 6 | 6 | ||
| 7 | #ifdef __KERNEL__ | 7 | #ifdef __KERNEL__ |
| 8 | #include <linux/linkage.h> | ||
| 8 | #include <asm/visasm.h> | 9 | #include <asm/visasm.h> |
| 9 | #include <asm/asi.h> | 10 | #include <asm/asi.h> |
| 10 | #include <asm/export.h> | 11 | #include <asm/export.h> |
| @@ -24,21 +25,17 @@ | |||
| 24 | #endif | 25 | #endif |
| 25 | 26 | ||
| 26 | #ifndef EX_LD | 27 | #ifndef EX_LD |
| 27 | #define EX_LD(x) x | 28 | #define EX_LD(x,y) x |
| 28 | #endif | 29 | #endif |
| 29 | #ifndef EX_LD_FP | 30 | #ifndef EX_LD_FP |
| 30 | #define EX_LD_FP(x) x | 31 | #define EX_LD_FP(x,y) x |
| 31 | #endif | 32 | #endif |
| 32 | 33 | ||
| 33 | #ifndef EX_ST | 34 | #ifndef EX_ST |
| 34 | #define EX_ST(x) x | 35 | #define EX_ST(x,y) x |
| 35 | #endif | 36 | #endif |
| 36 | #ifndef EX_ST_FP | 37 | #ifndef EX_ST_FP |
| 37 | #define EX_ST_FP(x) x | 38 | #define EX_ST_FP(x,y) x |
| 38 | #endif | ||
| 39 | |||
| 40 | #ifndef EX_RETVAL | ||
| 41 | #define EX_RETVAL(x) x | ||
| 42 | #endif | 39 | #endif |
| 43 | 40 | ||
| 44 | #ifndef LOAD | 41 | #ifndef LOAD |
| @@ -79,53 +76,169 @@ | |||
| 79 | faligndata %f7, %f8, %f60; \ | 76 | faligndata %f7, %f8, %f60; \ |
| 80 | faligndata %f8, %f9, %f62; | 77 | faligndata %f8, %f9, %f62; |
| 81 | 78 | ||
| 82 | #define MAIN_LOOP_CHUNK(src, dest, fdest, fsrc, len, jmptgt) \ | 79 | #define MAIN_LOOP_CHUNK(src, dest, fdest, fsrc, jmptgt) \ |
| 83 | EX_LD_FP(LOAD_BLK(%src, %fdest)); \ | 80 | EX_LD_FP(LOAD_BLK(%src, %fdest), U1_gs_80_fp); \ |
| 84 | EX_ST_FP(STORE_BLK(%fsrc, %dest)); \ | 81 | EX_ST_FP(STORE_BLK(%fsrc, %dest), U1_gs_80_fp); \ |
| 85 | add %src, 0x40, %src; \ | 82 | add %src, 0x40, %src; \ |
| 86 | subcc %len, 0x40, %len; \ | 83 | subcc %GLOBAL_SPARE, 0x40, %GLOBAL_SPARE; \ |
| 87 | be,pn %xcc, jmptgt; \ | 84 | be,pn %xcc, jmptgt; \ |
| 88 | add %dest, 0x40, %dest; \ | 85 | add %dest, 0x40, %dest; \ |
| 89 | 86 | ||
| 90 | #define LOOP_CHUNK1(src, dest, len, branch_dest) \ | 87 | #define LOOP_CHUNK1(src, dest, branch_dest) \ |
| 91 | MAIN_LOOP_CHUNK(src, dest, f0, f48, len, branch_dest) | 88 | MAIN_LOOP_CHUNK(src, dest, f0, f48, branch_dest) |
| 92 | #define LOOP_CHUNK2(src, dest, len, branch_dest) \ | 89 | #define LOOP_CHUNK2(src, dest, branch_dest) \ |
| 93 | MAIN_LOOP_CHUNK(src, dest, f16, f48, len, branch_dest) | 90 | MAIN_LOOP_CHUNK(src, dest, f16, f48, branch_dest) |
| 94 | #define LOOP_CHUNK3(src, dest, len, branch_dest) \ | 91 | #define LOOP_CHUNK3(src, dest, branch_dest) \ |
| 95 | MAIN_LOOP_CHUNK(src, dest, f32, f48, len, branch_dest) | 92 | MAIN_LOOP_CHUNK(src, dest, f32, f48, branch_dest) |
| 96 | 93 | ||
| 97 | #define DO_SYNC membar #Sync; | 94 | #define DO_SYNC membar #Sync; |
| 98 | #define STORE_SYNC(dest, fsrc) \ | 95 | #define STORE_SYNC(dest, fsrc) \ |
| 99 | EX_ST_FP(STORE_BLK(%fsrc, %dest)); \ | 96 | EX_ST_FP(STORE_BLK(%fsrc, %dest), U1_gs_80_fp); \ |
| 100 | add %dest, 0x40, %dest; \ | 97 | add %dest, 0x40, %dest; \ |
| 101 | DO_SYNC | 98 | DO_SYNC |
| 102 | 99 | ||
| 103 | #define STORE_JUMP(dest, fsrc, target) \ | 100 | #define STORE_JUMP(dest, fsrc, target) \ |
| 104 | EX_ST_FP(STORE_BLK(%fsrc, %dest)); \ | 101 | EX_ST_FP(STORE_BLK(%fsrc, %dest), U1_gs_40_fp); \ |
| 105 | add %dest, 0x40, %dest; \ | 102 | add %dest, 0x40, %dest; \ |
| 106 | ba,pt %xcc, target; \ | 103 | ba,pt %xcc, target; \ |
| 107 | nop; | 104 | nop; |
| 108 | 105 | ||
| 109 | #define FINISH_VISCHUNK(dest, f0, f1, left) \ | 106 | #define FINISH_VISCHUNK(dest, f0, f1) \ |
| 110 | subcc %left, 8, %left;\ | 107 | subcc %g3, 8, %g3; \ |
| 111 | bl,pn %xcc, 95f; \ | 108 | bl,pn %xcc, 95f; \ |
| 112 | faligndata %f0, %f1, %f48; \ | 109 | faligndata %f0, %f1, %f48; \ |
| 113 | EX_ST_FP(STORE(std, %f48, %dest)); \ | 110 | EX_ST_FP(STORE(std, %f48, %dest), U1_g3_8_fp); \ |
| 114 | add %dest, 8, %dest; | 111 | add %dest, 8, %dest; |
| 115 | 112 | ||
| 116 | #define UNEVEN_VISCHUNK_LAST(dest, f0, f1, left) \ | 113 | #define UNEVEN_VISCHUNK_LAST(dest, f0, f1) \ |
| 117 | subcc %left, 8, %left; \ | 114 | subcc %g3, 8, %g3; \ |
| 118 | bl,pn %xcc, 95f; \ | 115 | bl,pn %xcc, 95f; \ |
| 119 | fsrc2 %f0, %f1; | 116 | fsrc2 %f0, %f1; |
| 120 | 117 | ||
| 121 | #define UNEVEN_VISCHUNK(dest, f0, f1, left) \ | 118 | #define UNEVEN_VISCHUNK(dest, f0, f1) \ |
| 122 | UNEVEN_VISCHUNK_LAST(dest, f0, f1, left) \ | 119 | UNEVEN_VISCHUNK_LAST(dest, f0, f1) \ |
| 123 | ba,a,pt %xcc, 93f; | 120 | ba,a,pt %xcc, 93f; |
| 124 | 121 | ||
| 125 | .register %g2,#scratch | 122 | .register %g2,#scratch |
| 126 | .register %g3,#scratch | 123 | .register %g3,#scratch |
| 127 | 124 | ||
| 128 | .text | 125 | .text |
| 126 | #ifndef EX_RETVAL | ||
| 127 | #define EX_RETVAL(x) x | ||
| 128 | ENTRY(U1_g1_1_fp) | ||
| 129 | VISExitHalf | ||
| 130 | add %g1, 1, %g1 | ||
| 131 | add %g1, %g2, %g1 | ||
| 132 | retl | ||
| 133 | add %g1, %o2, %o0 | ||
| 134 | ENDPROC(U1_g1_1_fp) | ||
| 135 | ENTRY(U1_g2_0_fp) | ||
| 136 | VISExitHalf | ||
| 137 | retl | ||
| 138 | add %g2, %o2, %o0 | ||
| 139 | ENDPROC(U1_g2_0_fp) | ||
| 140 | ENTRY(U1_g2_8_fp) | ||
| 141 | VISExitHalf | ||
| 142 | add %g2, 8, %g2 | ||
| 143 | retl | ||
| 144 | add %g2, %o2, %o0 | ||
| 145 | ENDPROC(U1_g2_8_fp) | ||
| 146 | ENTRY(U1_gs_0_fp) | ||
| 147 | VISExitHalf | ||
| 148 | add %GLOBAL_SPARE, %g3, %o0 | ||
| 149 | retl | ||
| 150 | add %o0, %o2, %o0 | ||
| 151 | ENDPROC(U1_gs_0_fp) | ||
| 152 | ENTRY(U1_gs_80_fp) | ||
| 153 | VISExitHalf | ||
| 154 | add %GLOBAL_SPARE, 0x80, %GLOBAL_SPARE | ||
| 155 | add %GLOBAL_SPARE, %g3, %o0 | ||
| 156 | retl | ||
| 157 | add %o0, %o2, %o0 | ||
| 158 | ENDPROC(U1_gs_80_fp) | ||
| 159 | ENTRY(U1_gs_40_fp) | ||
| 160 | VISExitHalf | ||
| 161 | add %GLOBAL_SPARE, 0x40, %GLOBAL_SPARE | ||
| 162 | add %GLOBAL_SPARE, %g3, %o0 | ||
| 163 | retl | ||
| 164 | add %o0, %o2, %o0 | ||
| 165 | ENDPROC(U1_gs_40_fp) | ||
| 166 | ENTRY(U1_g3_0_fp) | ||
| 167 | VISExitHalf | ||
| 168 | retl | ||
| 169 | add %g3, %o2, %o0 | ||
| 170 | ENDPROC(U1_g3_0_fp) | ||
| 171 | ENTRY(U1_g3_8_fp) | ||
| 172 | VISExitHalf | ||
| 173 | add %g3, 8, %g3 | ||
| 174 | retl | ||
| 175 | add %g3, %o2, %o0 | ||
| 176 | ENDPROC(U1_g3_8_fp) | ||
| 177 | ENTRY(U1_o2_0_fp) | ||
| 178 | VISExitHalf | ||
| 179 | retl | ||
| 180 | mov %o2, %o0 | ||
| 181 | ENDPROC(U1_o2_0_fp) | ||
| 182 | ENTRY(U1_o2_1_fp) | ||
| 183 | VISExitHalf | ||
| 184 | retl | ||
| 185 | add %o2, 1, %o0 | ||
| 186 | ENDPROC(U1_o2_1_fp) | ||
| 187 | ENTRY(U1_gs_0) | ||
| 188 | VISExitHalf | ||
| 189 | retl | ||
| 190 | add %GLOBAL_SPARE, %o2, %o0 | ||
| 191 | ENDPROC(U1_gs_0) | ||
| 192 | ENTRY(U1_gs_8) | ||
| 193 | VISExitHalf | ||
| 194 | add %GLOBAL_SPARE, %o2, %GLOBAL_SPARE | ||
| 195 | retl | ||
| 196 | add %GLOBAL_SPARE, 0x8, %o0 | ||
| 197 | ENDPROC(U1_gs_8) | ||
| 198 | ENTRY(U1_gs_10) | ||
| 199 | VISExitHalf | ||
| 200 | add %GLOBAL_SPARE, %o2, %GLOBAL_SPARE | ||
| 201 | retl | ||
| 202 | add %GLOBAL_SPARE, 0x10, %o0 | ||
| 203 | ENDPROC(U1_gs_10) | ||
| 204 | ENTRY(U1_o2_0) | ||
| 205 | retl | ||
| 206 | mov %o2, %o0 | ||
| 207 | ENDPROC(U1_o2_0) | ||
| 208 | ENTRY(U1_o2_8) | ||
| 209 | retl | ||
| 210 | add %o2, 8, %o0 | ||
| 211 | ENDPROC(U1_o2_8) | ||
| 212 | ENTRY(U1_o2_4) | ||
| 213 | retl | ||
| 214 | add %o2, 4, %o0 | ||
| 215 | ENDPROC(U1_o2_4) | ||
| 216 | ENTRY(U1_o2_1) | ||
| 217 | retl | ||
| 218 | add %o2, 1, %o0 | ||
| 219 | ENDPROC(U1_o2_1) | ||
| 220 | ENTRY(U1_g1_0) | ||
| 221 | retl | ||
| 222 | add %g1, %o2, %o0 | ||
| 223 | ENDPROC(U1_g1_0) | ||
| 224 | ENTRY(U1_g1_1) | ||
| 225 | add %g1, 1, %g1 | ||
| 226 | retl | ||
| 227 | add %g1, %o2, %o0 | ||
| 228 | ENDPROC(U1_g1_1) | ||
| 229 | ENTRY(U1_gs_0_o2_adj) | ||
| 230 | and %o2, 7, %o2 | ||
| 231 | retl | ||
| 232 | add %GLOBAL_SPARE, %o2, %o0 | ||
| 233 | ENDPROC(U1_gs_0_o2_adj) | ||
| 234 | ENTRY(U1_gs_8_o2_adj) | ||
| 235 | and %o2, 7, %o2 | ||
| 236 | add %GLOBAL_SPARE, 8, %GLOBAL_SPARE | ||
| 237 | retl | ||
| 238 | add %GLOBAL_SPARE, %o2, %o0 | ||
| 239 | ENDPROC(U1_gs_8_o2_adj) | ||
| 240 | #endif | ||
| 241 | |||
| 129 | .align 64 | 242 | .align 64 |
| 130 | 243 | ||
| 131 | .globl FUNC_NAME | 244 | .globl FUNC_NAME |
| @@ -167,8 +280,8 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 167 | and %g2, 0x38, %g2 | 280 | and %g2, 0x38, %g2 |
| 168 | 281 | ||
| 169 | 1: subcc %g1, 0x1, %g1 | 282 | 1: subcc %g1, 0x1, %g1 |
| 170 | EX_LD_FP(LOAD(ldub, %o1 + 0x00, %o3)) | 283 | EX_LD_FP(LOAD(ldub, %o1 + 0x00, %o3), U1_g1_1_fp) |
| 171 | EX_ST_FP(STORE(stb, %o3, %o1 + %GLOBAL_SPARE)) | 284 | EX_ST_FP(STORE(stb, %o3, %o1 + %GLOBAL_SPARE), U1_g1_1_fp) |
| 172 | bgu,pt %XCC, 1b | 285 | bgu,pt %XCC, 1b |
| 173 | add %o1, 0x1, %o1 | 286 | add %o1, 0x1, %o1 |
| 174 | 287 | ||
| @@ -179,20 +292,20 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 179 | be,pt %icc, 3f | 292 | be,pt %icc, 3f |
| 180 | alignaddr %o1, %g0, %o1 | 293 | alignaddr %o1, %g0, %o1 |
| 181 | 294 | ||
| 182 | EX_LD_FP(LOAD(ldd, %o1, %f4)) | 295 | EX_LD_FP(LOAD(ldd, %o1, %f4), U1_g2_0_fp) |
| 183 | 1: EX_LD_FP(LOAD(ldd, %o1 + 0x8, %f6)) | 296 | 1: EX_LD_FP(LOAD(ldd, %o1 + 0x8, %f6), U1_g2_0_fp) |
| 184 | add %o1, 0x8, %o1 | 297 | add %o1, 0x8, %o1 |
| 185 | subcc %g2, 0x8, %g2 | 298 | subcc %g2, 0x8, %g2 |
| 186 | faligndata %f4, %f6, %f0 | 299 | faligndata %f4, %f6, %f0 |
| 187 | EX_ST_FP(STORE(std, %f0, %o0)) | 300 | EX_ST_FP(STORE(std, %f0, %o0), U1_g2_8_fp) |
| 188 | be,pn %icc, 3f | 301 | be,pn %icc, 3f |
| 189 | add %o0, 0x8, %o0 | 302 | add %o0, 0x8, %o0 |
| 190 | 303 | ||
| 191 | EX_LD_FP(LOAD(ldd, %o1 + 0x8, %f4)) | 304 | EX_LD_FP(LOAD(ldd, %o1 + 0x8, %f4), U1_g2_0_fp) |
| 192 | add %o1, 0x8, %o1 | 305 | add %o1, 0x8, %o1 |
| 193 | subcc %g2, 0x8, %g2 | 306 | subcc %g2, 0x8, %g2 |
| 194 | faligndata %f6, %f4, %f0 | 307 | faligndata %f6, %f4, %f0 |
| 195 | EX_ST_FP(STORE(std, %f0, %o0)) | 308 | EX_ST_FP(STORE(std, %f0, %o0), U1_g2_8_fp) |
| 196 | bne,pt %icc, 1b | 309 | bne,pt %icc, 1b |
| 197 | add %o0, 0x8, %o0 | 310 | add %o0, 0x8, %o0 |
| 198 | 311 | ||
| @@ -215,13 +328,13 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 215 | add %g1, %GLOBAL_SPARE, %g1 | 328 | add %g1, %GLOBAL_SPARE, %g1 |
| 216 | subcc %o2, %g3, %o2 | 329 | subcc %o2, %g3, %o2 |
| 217 | 330 | ||
| 218 | EX_LD_FP(LOAD_BLK(%o1, %f0)) | 331 | EX_LD_FP(LOAD_BLK(%o1, %f0), U1_gs_0_fp) |
| 219 | add %o1, 0x40, %o1 | 332 | add %o1, 0x40, %o1 |
| 220 | add %g1, %g3, %g1 | 333 | add %g1, %g3, %g1 |
| 221 | EX_LD_FP(LOAD_BLK(%o1, %f16)) | 334 | EX_LD_FP(LOAD_BLK(%o1, %f16), U1_gs_0_fp) |
| 222 | add %o1, 0x40, %o1 | 335 | add %o1, 0x40, %o1 |
| 223 | sub %GLOBAL_SPARE, 0x80, %GLOBAL_SPARE | 336 | sub %GLOBAL_SPARE, 0x80, %GLOBAL_SPARE |
| 224 | EX_LD_FP(LOAD_BLK(%o1, %f32)) | 337 | EX_LD_FP(LOAD_BLK(%o1, %f32), U1_gs_80_fp) |
| 225 | add %o1, 0x40, %o1 | 338 | add %o1, 0x40, %o1 |
| 226 | 339 | ||
| 227 | /* There are 8 instances of the unrolled loop, | 340 | /* There are 8 instances of the unrolled loop, |
| @@ -241,11 +354,11 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 241 | 354 | ||
| 242 | .align 64 | 355 | .align 64 |
| 243 | 1: FREG_FROB(f0, f2, f4, f6, f8, f10,f12,f14,f16) | 356 | 1: FREG_FROB(f0, f2, f4, f6, f8, f10,f12,f14,f16) |
| 244 | LOOP_CHUNK1(o1, o0, GLOBAL_SPARE, 1f) | 357 | LOOP_CHUNK1(o1, o0, 1f) |
| 245 | FREG_FROB(f16,f18,f20,f22,f24,f26,f28,f30,f32) | 358 | FREG_FROB(f16,f18,f20,f22,f24,f26,f28,f30,f32) |
| 246 | LOOP_CHUNK2(o1, o0, GLOBAL_SPARE, 2f) | 359 | LOOP_CHUNK2(o1, o0, 2f) |
| 247 | FREG_FROB(f32,f34,f36,f38,f40,f42,f44,f46,f0) | 360 | FREG_FROB(f32,f34,f36,f38,f40,f42,f44,f46,f0) |
| 248 | LOOP_CHUNK3(o1, o0, GLOBAL_SPARE, 3f) | 361 | LOOP_CHUNK3(o1, o0, 3f) |
| 249 | ba,pt %xcc, 1b+4 | 362 | ba,pt %xcc, 1b+4 |
| 250 | faligndata %f0, %f2, %f48 | 363 | faligndata %f0, %f2, %f48 |
| 251 | 1: FREG_FROB(f16,f18,f20,f22,f24,f26,f28,f30,f32) | 364 | 1: FREG_FROB(f16,f18,f20,f22,f24,f26,f28,f30,f32) |
| @@ -262,11 +375,11 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 262 | STORE_JUMP(o0, f48, 56f) | 375 | STORE_JUMP(o0, f48, 56f) |
| 263 | 376 | ||
| 264 | 1: FREG_FROB(f2, f4, f6, f8, f10,f12,f14,f16,f18) | 377 | 1: FREG_FROB(f2, f4, f6, f8, f10,f12,f14,f16,f18) |
| 265 | LOOP_CHUNK1(o1, o0, GLOBAL_SPARE, 1f) | 378 | LOOP_CHUNK1(o1, o0, 1f) |
| 266 | FREG_FROB(f18,f20,f22,f24,f26,f28,f30,f32,f34) | 379 | FREG_FROB(f18,f20,f22,f24,f26,f28,f30,f32,f34) |
| 267 | LOOP_CHUNK2(o1, o0, GLOBAL_SPARE, 2f) | 380 | LOOP_CHUNK2(o1, o0, 2f) |
| 268 | FREG_FROB(f34,f36,f38,f40,f42,f44,f46,f0, f2) | 381 | FREG_FROB(f34,f36,f38,f40,f42,f44,f46,f0, f2) |
| 269 | LOOP_CHUNK3(o1, o0, GLOBAL_SPARE, 3f) | 382 | LOOP_CHUNK3(o1, o0, 3f) |
| 270 | ba,pt %xcc, 1b+4 | 383 | ba,pt %xcc, 1b+4 |
| 271 | faligndata %f2, %f4, %f48 | 384 | faligndata %f2, %f4, %f48 |
| 272 | 1: FREG_FROB(f18,f20,f22,f24,f26,f28,f30,f32,f34) | 385 | 1: FREG_FROB(f18,f20,f22,f24,f26,f28,f30,f32,f34) |
| @@ -283,11 +396,11 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 283 | STORE_JUMP(o0, f48, 57f) | 396 | STORE_JUMP(o0, f48, 57f) |
| 284 | 397 | ||
| 285 | 1: FREG_FROB(f4, f6, f8, f10,f12,f14,f16,f18,f20) | 398 | 1: FREG_FROB(f4, f6, f8, f10,f12,f14,f16,f18,f20) |
| 286 | LOOP_CHUNK1(o1, o0, GLOBAL_SPARE, 1f) | 399 | LOOP_CHUNK1(o1, o0, 1f) |
| 287 | FREG_FROB(f20,f22,f24,f26,f28,f30,f32,f34,f36) | 400 | FREG_FROB(f20,f22,f24,f26,f28,f30,f32,f34,f36) |
| 288 | LOOP_CHUNK2(o1, o0, GLOBAL_SPARE, 2f) | 401 | LOOP_CHUNK2(o1, o0, 2f) |
| 289 | FREG_FROB(f36,f38,f40,f42,f44,f46,f0, f2, f4) | 402 | FREG_FROB(f36,f38,f40,f42,f44,f46,f0, f2, f4) |
| 290 | LOOP_CHUNK3(o1, o0, GLOBAL_SPARE, 3f) | 403 | LOOP_CHUNK3(o1, o0, 3f) |
| 291 | ba,pt %xcc, 1b+4 | 404 | ba,pt %xcc, 1b+4 |
| 292 | faligndata %f4, %f6, %f48 | 405 | faligndata %f4, %f6, %f48 |
| 293 | 1: FREG_FROB(f20,f22,f24,f26,f28,f30,f32,f34,f36) | 406 | 1: FREG_FROB(f20,f22,f24,f26,f28,f30,f32,f34,f36) |
| @@ -304,11 +417,11 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 304 | STORE_JUMP(o0, f48, 58f) | 417 | STORE_JUMP(o0, f48, 58f) |
| 305 | 418 | ||
| 306 | 1: FREG_FROB(f6, f8, f10,f12,f14,f16,f18,f20,f22) | 419 | 1: FREG_FROB(f6, f8, f10,f12,f14,f16,f18,f20,f22) |
| 307 | LOOP_CHUNK1(o1, o0, GLOBAL_SPARE, 1f) | 420 | LOOP_CHUNK1(o1, o0, 1f) |
| 308 | FREG_FROB(f22,f24,f26,f28,f30,f32,f34,f36,f38) | 421 | FREG_FROB(f22,f24,f26,f28,f30,f32,f34,f36,f38) |
| 309 | LOOP_CHUNK2(o1, o0, GLOBAL_SPARE, 2f) | 422 | LOOP_CHUNK2(o1, o0, 2f) |
| 310 | FREG_FROB(f38,f40,f42,f44,f46,f0, f2, f4, f6) | 423 | FREG_FROB(f38,f40,f42,f44,f46,f0, f2, f4, f6) |
| 311 | LOOP_CHUNK3(o1, o0, GLOBAL_SPARE, 3f) | 424 | LOOP_CHUNK3(o1, o0, 3f) |
| 312 | ba,pt %xcc, 1b+4 | 425 | ba,pt %xcc, 1b+4 |
| 313 | faligndata %f6, %f8, %f48 | 426 | faligndata %f6, %f8, %f48 |
| 314 | 1: FREG_FROB(f22,f24,f26,f28,f30,f32,f34,f36,f38) | 427 | 1: FREG_FROB(f22,f24,f26,f28,f30,f32,f34,f36,f38) |
| @@ -325,11 +438,11 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 325 | STORE_JUMP(o0, f48, 59f) | 438 | STORE_JUMP(o0, f48, 59f) |
| 326 | 439 | ||
| 327 | 1: FREG_FROB(f8, f10,f12,f14,f16,f18,f20,f22,f24) | 440 | 1: FREG_FROB(f8, f10,f12,f14,f16,f18,f20,f22,f24) |
| 328 | LOOP_CHUNK1(o1, o0, GLOBAL_SPARE, 1f) | 441 | LOOP_CHUNK1(o1, o0, 1f) |
| 329 | FREG_FROB(f24,f26,f28,f30,f32,f34,f36,f38,f40) | 442 | FREG_FROB(f24,f26,f28,f30,f32,f34,f36,f38,f40) |
| 330 | LOOP_CHUNK2(o1, o0, GLOBAL_SPARE, 2f) | 443 | LOOP_CHUNK2(o1, o0, 2f) |
| 331 | FREG_FROB(f40,f42,f44,f46,f0, f2, f4, f6, f8) | 444 | FREG_FROB(f40,f42,f44,f46,f0, f2, f4, f6, f8) |
| 332 | LOOP_CHUNK3(o1, o0, GLOBAL_SPARE, 3f) | 445 | LOOP_CHUNK3(o1, o0, 3f) |
| 333 | ba,pt %xcc, 1b+4 | 446 | ba,pt %xcc, 1b+4 |
| 334 | faligndata %f8, %f10, %f48 | 447 | faligndata %f8, %f10, %f48 |
| 335 | 1: FREG_FROB(f24,f26,f28,f30,f32,f34,f36,f38,f40) | 448 | 1: FREG_FROB(f24,f26,f28,f30,f32,f34,f36,f38,f40) |
| @@ -346,11 +459,11 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 346 | STORE_JUMP(o0, f48, 60f) | 459 | STORE_JUMP(o0, f48, 60f) |
| 347 | 460 | ||
| 348 | 1: FREG_FROB(f10,f12,f14,f16,f18,f20,f22,f24,f26) | 461 | 1: FREG_FROB(f10,f12,f14,f16,f18,f20,f22,f24,f26) |
| 349 | LOOP_CHUNK1(o1, o0, GLOBAL_SPARE, 1f) | 462 | LOOP_CHUNK1(o1, o0, 1f) |
| 350 | FREG_FROB(f26,f28,f30,f32,f34,f36,f38,f40,f42) | 463 | FREG_FROB(f26,f28,f30,f32,f34,f36,f38,f40,f42) |
| 351 | LOOP_CHUNK2(o1, o0, GLOBAL_SPARE, 2f) | 464 | LOOP_CHUNK2(o1, o0, 2f) |
| 352 | FREG_FROB(f42,f44,f46,f0, f2, f4, f6, f8, f10) | 465 | FREG_FROB(f42,f44,f46,f0, f2, f4, f6, f8, f10) |
| 353 | LOOP_CHUNK3(o1, o0, GLOBAL_SPARE, 3f) | 466 | LOOP_CHUNK3(o1, o0, 3f) |
| 354 | ba,pt %xcc, 1b+4 | 467 | ba,pt %xcc, 1b+4 |
| 355 | faligndata %f10, %f12, %f48 | 468 | faligndata %f10, %f12, %f48 |
| 356 | 1: FREG_FROB(f26,f28,f30,f32,f34,f36,f38,f40,f42) | 469 | 1: FREG_FROB(f26,f28,f30,f32,f34,f36,f38,f40,f42) |
| @@ -367,11 +480,11 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 367 | STORE_JUMP(o0, f48, 61f) | 480 | STORE_JUMP(o0, f48, 61f) |
| 368 | 481 | ||
| 369 | 1: FREG_FROB(f12,f14,f16,f18,f20,f22,f24,f26,f28) | 482 | 1: FREG_FROB(f12,f14,f16,f18,f20,f22,f24,f26,f28) |
| 370 | LOOP_CHUNK1(o1, o0, GLOBAL_SPARE, 1f) | 483 | LOOP_CHUNK1(o1, o0, 1f) |
| 371 | FREG_FROB(f28,f30,f32,f34,f36,f38,f40,f42,f44) | 484 | FREG_FROB(f28,f30,f32,f34,f36,f38,f40,f42,f44) |
| 372 | LOOP_CHUNK2(o1, o0, GLOBAL_SPARE, 2f) | 485 | LOOP_CHUNK2(o1, o0, 2f) |
| 373 | FREG_FROB(f44,f46,f0, f2, f4, f6, f8, f10,f12) | 486 | FREG_FROB(f44,f46,f0, f2, f4, f6, f8, f10,f12) |
| 374 | LOOP_CHUNK3(o1, o0, GLOBAL_SPARE, 3f) | 487 | LOOP_CHUNK3(o1, o0, 3f) |
| 375 | ba,pt %xcc, 1b+4 | 488 | ba,pt %xcc, 1b+4 |
| 376 | faligndata %f12, %f14, %f48 | 489 | faligndata %f12, %f14, %f48 |
| 377 | 1: FREG_FROB(f28,f30,f32,f34,f36,f38,f40,f42,f44) | 490 | 1: FREG_FROB(f28,f30,f32,f34,f36,f38,f40,f42,f44) |
| @@ -388,11 +501,11 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 388 | STORE_JUMP(o0, f48, 62f) | 501 | STORE_JUMP(o0, f48, 62f) |
| 389 | 502 | ||
| 390 | 1: FREG_FROB(f14,f16,f18,f20,f22,f24,f26,f28,f30) | 503 | 1: FREG_FROB(f14,f16,f18,f20,f22,f24,f26,f28,f30) |
| 391 | LOOP_CHUNK1(o1, o0, GLOBAL_SPARE, 1f) | 504 | LOOP_CHUNK1(o1, o0, 1f) |
| 392 | FREG_FROB(f30,f32,f34,f36,f38,f40,f42,f44,f46) | 505 | FREG_FROB(f30,f32,f34,f36,f38,f40,f42,f44,f46) |
| 393 | LOOP_CHUNK2(o1, o0, GLOBAL_SPARE, 2f) | 506 | LOOP_CHUNK2(o1, o0, 2f) |
| 394 | FREG_FROB(f46,f0, f2, f4, f6, f8, f10,f12,f14) | 507 | FREG_FROB(f46,f0, f2, f4, f6, f8, f10,f12,f14) |
| 395 | LOOP_CHUNK3(o1, o0, GLOBAL_SPARE, 3f) | 508 | LOOP_CHUNK3(o1, o0, 3f) |
| 396 | ba,pt %xcc, 1b+4 | 509 | ba,pt %xcc, 1b+4 |
| 397 | faligndata %f14, %f16, %f48 | 510 | faligndata %f14, %f16, %f48 |
| 398 | 1: FREG_FROB(f30,f32,f34,f36,f38,f40,f42,f44,f46) | 511 | 1: FREG_FROB(f30,f32,f34,f36,f38,f40,f42,f44,f46) |
| @@ -408,53 +521,53 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 408 | FREG_FROB(f30,f32,f34,f36,f38,f40,f42,f44,f46) | 521 | FREG_FROB(f30,f32,f34,f36,f38,f40,f42,f44,f46) |
| 409 | STORE_JUMP(o0, f48, 63f) | 522 | STORE_JUMP(o0, f48, 63f) |
| 410 | 523 | ||
| 411 | 40: FINISH_VISCHUNK(o0, f0, f2, g3) | 524 | 40: FINISH_VISCHUNK(o0, f0, f2) |
| 412 | 41: FINISH_VISCHUNK(o0, f2, f4, g3) | 525 | 41: FINISH_VISCHUNK(o0, f2, f4) |
| 413 | 42: FINISH_VISCHUNK(o0, f4, f6, g3) | 526 | 42: FINISH_VISCHUNK(o0, f4, f6) |
| 414 | 43: FINISH_VISCHUNK(o0, f6, f8, g3) | 527 | 43: FINISH_VISCHUNK(o0, f6, f8) |
| 415 | 44: FINISH_VISCHUNK(o0, f8, f10, g3) | 528 | 44: FINISH_VISCHUNK(o0, f8, f10) |
| 416 | 45: FINISH_VISCHUNK(o0, f10, f12, g3) | 529 | 45: FINISH_VISCHUNK(o0, f10, f12) |
| 417 | 46: FINISH_VISCHUNK(o0, f12, f14, g3) | 530 | 46: FINISH_VISCHUNK(o0, f12, f14) |
| 418 | 47: UNEVEN_VISCHUNK(o0, f14, f0, g3) | 531 | 47: UNEVEN_VISCHUNK(o0, f14, f0) |
| 419 | 48: FINISH_VISCHUNK(o0, f16, f18, g3) | 532 | 48: FINISH_VISCHUNK(o0, f16, f18) |
| 420 | 49: FINISH_VISCHUNK(o0, f18, f20, g3) | 533 | 49: FINISH_VISCHUNK(o0, f18, f20) |
| 421 | 50: FINISH_VISCHUNK(o0, f20, f22, g3) | 534 | 50: FINISH_VISCHUNK(o0, f20, f22) |
| 422 | 51: FINISH_VISCHUNK(o0, f22, f24, g3) | 535 | 51: FINISH_VISCHUNK(o0, f22, f24) |
| 423 | 52: FINISH_VISCHUNK(o0, f24, f26, g3) | 536 | 52: FINISH_VISCHUNK(o0, f24, f26) |
| 424 | 53: FINISH_VISCHUNK(o0, f26, f28, g3) | 537 | 53: FINISH_VISCHUNK(o0, f26, f28) |
| 425 | 54: FINISH_VISCHUNK(o0, f28, f30, g3) | 538 | 54: FINISH_VISCHUNK(o0, f28, f30) |
| 426 | 55: UNEVEN_VISCHUNK(o0, f30, f0, g3) | 539 | 55: UNEVEN_VISCHUNK(o0, f30, f0) |
| 427 | 56: FINISH_VISCHUNK(o0, f32, f34, g3) | 540 | 56: FINISH_VISCHUNK(o0, f32, f34) |
| 428 | 57: FINISH_VISCHUNK(o0, f34, f36, g3) | 541 | 57: FINISH_VISCHUNK(o0, f34, f36) |
| 429 | 58: FINISH_VISCHUNK(o0, f36, f38, g3) | 542 | 58: FINISH_VISCHUNK(o0, f36, f38) |
| 430 | 59: FINISH_VISCHUNK(o0, f38, f40, g3) | 543 | 59: FINISH_VISCHUNK(o0, f38, f40) |
| 431 | 60: FINISH_VISCHUNK(o0, f40, f42, g3) | 544 | 60: FINISH_VISCHUNK(o0, f40, f42) |
| 432 | 61: FINISH_VISCHUNK(o0, f42, f44, g3) | 545 | 61: FINISH_VISCHUNK(o0, f42, f44) |
| 433 | 62: FINISH_VISCHUNK(o0, f44, f46, g3) | 546 | 62: FINISH_VISCHUNK(o0, f44, f46) |
| 434 | 63: UNEVEN_VISCHUNK_LAST(o0, f46, f0, g3) | 547 | 63: UNEVEN_VISCHUNK_LAST(o0, f46, f0) |
| 435 | 548 | ||
| 436 | 93: EX_LD_FP(LOAD(ldd, %o1, %f2)) | 549 | 93: EX_LD_FP(LOAD(ldd, %o1, %f2), U1_g3_0_fp) |
| 437 | add %o1, 8, %o1 | 550 | add %o1, 8, %o1 |
| 438 | subcc %g3, 8, %g3 | 551 | subcc %g3, 8, %g3 |
| 439 | faligndata %f0, %f2, %f8 | 552 | faligndata %f0, %f2, %f8 |
| 440 | EX_ST_FP(STORE(std, %f8, %o0)) | 553 | EX_ST_FP(STORE(std, %f8, %o0), U1_g3_8_fp) |
| 441 | bl,pn %xcc, 95f | 554 | bl,pn %xcc, 95f |
| 442 | add %o0, 8, %o0 | 555 | add %o0, 8, %o0 |
| 443 | EX_LD_FP(LOAD(ldd, %o1, %f0)) | 556 | EX_LD_FP(LOAD(ldd, %o1, %f0), U1_g3_0_fp) |
| 444 | add %o1, 8, %o1 | 557 | add %o1, 8, %o1 |
| 445 | subcc %g3, 8, %g3 | 558 | subcc %g3, 8, %g3 |
| 446 | faligndata %f2, %f0, %f8 | 559 | faligndata %f2, %f0, %f8 |
| 447 | EX_ST_FP(STORE(std, %f8, %o0)) | 560 | EX_ST_FP(STORE(std, %f8, %o0), U1_g3_8_fp) |
| 448 | bge,pt %xcc, 93b | 561 | bge,pt %xcc, 93b |
| 449 | add %o0, 8, %o0 | 562 | add %o0, 8, %o0 |
| 450 | 563 | ||
| 451 | 95: brz,pt %o2, 2f | 564 | 95: brz,pt %o2, 2f |
| 452 | mov %g1, %o1 | 565 | mov %g1, %o1 |
| 453 | 566 | ||
| 454 | 1: EX_LD_FP(LOAD(ldub, %o1, %o3)) | 567 | 1: EX_LD_FP(LOAD(ldub, %o1, %o3), U1_o2_0_fp) |
| 455 | add %o1, 1, %o1 | 568 | add %o1, 1, %o1 |
| 456 | subcc %o2, 1, %o2 | 569 | subcc %o2, 1, %o2 |
| 457 | EX_ST_FP(STORE(stb, %o3, %o0)) | 570 | EX_ST_FP(STORE(stb, %o3, %o0), U1_o2_1_fp) |
| 458 | bne,pt %xcc, 1b | 571 | bne,pt %xcc, 1b |
| 459 | add %o0, 1, %o0 | 572 | add %o0, 1, %o0 |
| 460 | 573 | ||
| @@ -470,27 +583,27 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 470 | 583 | ||
| 471 | 72: andn %o2, 0xf, %GLOBAL_SPARE | 584 | 72: andn %o2, 0xf, %GLOBAL_SPARE |
| 472 | and %o2, 0xf, %o2 | 585 | and %o2, 0xf, %o2 |
| 473 | 1: EX_LD(LOAD(ldx, %o1 + 0x00, %o5)) | 586 | 1: EX_LD(LOAD(ldx, %o1 + 0x00, %o5), U1_gs_0) |
| 474 | EX_LD(LOAD(ldx, %o1 + 0x08, %g1)) | 587 | EX_LD(LOAD(ldx, %o1 + 0x08, %g1), U1_gs_0) |
| 475 | subcc %GLOBAL_SPARE, 0x10, %GLOBAL_SPARE | 588 | subcc %GLOBAL_SPARE, 0x10, %GLOBAL_SPARE |
| 476 | EX_ST(STORE(stx, %o5, %o1 + %o3)) | 589 | EX_ST(STORE(stx, %o5, %o1 + %o3), U1_gs_10) |
| 477 | add %o1, 0x8, %o1 | 590 | add %o1, 0x8, %o1 |
| 478 | EX_ST(STORE(stx, %g1, %o1 + %o3)) | 591 | EX_ST(STORE(stx, %g1, %o1 + %o3), U1_gs_8) |
| 479 | bgu,pt %XCC, 1b | 592 | bgu,pt %XCC, 1b |
| 480 | add %o1, 0x8, %o1 | 593 | add %o1, 0x8, %o1 |
| 481 | 73: andcc %o2, 0x8, %g0 | 594 | 73: andcc %o2, 0x8, %g0 |
| 482 | be,pt %XCC, 1f | 595 | be,pt %XCC, 1f |
| 483 | nop | 596 | nop |
| 484 | EX_LD(LOAD(ldx, %o1, %o5)) | 597 | EX_LD(LOAD(ldx, %o1, %o5), U1_o2_0) |
| 485 | sub %o2, 0x8, %o2 | 598 | sub %o2, 0x8, %o2 |
| 486 | EX_ST(STORE(stx, %o5, %o1 + %o3)) | 599 | EX_ST(STORE(stx, %o5, %o1 + %o3), U1_o2_8) |
| 487 | add %o1, 0x8, %o1 | 600 | add %o1, 0x8, %o1 |
| 488 | 1: andcc %o2, 0x4, %g0 | 601 | 1: andcc %o2, 0x4, %g0 |
| 489 | be,pt %XCC, 1f | 602 | be,pt %XCC, 1f |
| 490 | nop | 603 | nop |
| 491 | EX_LD(LOAD(lduw, %o1, %o5)) | 604 | EX_LD(LOAD(lduw, %o1, %o5), U1_o2_0) |
| 492 | sub %o2, 0x4, %o2 | 605 | sub %o2, 0x4, %o2 |
| 493 | EX_ST(STORE(stw, %o5, %o1 + %o3)) | 606 | EX_ST(STORE(stw, %o5, %o1 + %o3), U1_o2_4) |
| 494 | add %o1, 0x4, %o1 | 607 | add %o1, 0x4, %o1 |
| 495 | 1: cmp %o2, 0 | 608 | 1: cmp %o2, 0 |
| 496 | be,pt %XCC, 85f | 609 | be,pt %XCC, 85f |
| @@ -504,9 +617,9 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 504 | sub %g0, %g1, %g1 | 617 | sub %g0, %g1, %g1 |
| 505 | sub %o2, %g1, %o2 | 618 | sub %o2, %g1, %o2 |
| 506 | 619 | ||
| 507 | 1: EX_LD(LOAD(ldub, %o1, %o5)) | 620 | 1: EX_LD(LOAD(ldub, %o1, %o5), U1_g1_0) |
| 508 | subcc %g1, 1, %g1 | 621 | subcc %g1, 1, %g1 |
| 509 | EX_ST(STORE(stb, %o5, %o1 + %o3)) | 622 | EX_ST(STORE(stb, %o5, %o1 + %o3), U1_g1_1) |
| 510 | bgu,pt %icc, 1b | 623 | bgu,pt %icc, 1b |
| 511 | add %o1, 1, %o1 | 624 | add %o1, 1, %o1 |
| 512 | 625 | ||
| @@ -522,16 +635,16 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 522 | 635 | ||
| 523 | 8: mov 64, %o3 | 636 | 8: mov 64, %o3 |
| 524 | andn %o1, 0x7, %o1 | 637 | andn %o1, 0x7, %o1 |
| 525 | EX_LD(LOAD(ldx, %o1, %g2)) | 638 | EX_LD(LOAD(ldx, %o1, %g2), U1_o2_0) |
| 526 | sub %o3, %g1, %o3 | 639 | sub %o3, %g1, %o3 |
| 527 | andn %o2, 0x7, %GLOBAL_SPARE | 640 | andn %o2, 0x7, %GLOBAL_SPARE |
| 528 | sllx %g2, %g1, %g2 | 641 | sllx %g2, %g1, %g2 |
| 529 | 1: EX_LD(LOAD(ldx, %o1 + 0x8, %g3)) | 642 | 1: EX_LD(LOAD(ldx, %o1 + 0x8, %g3), U1_gs_0_o2_adj) |
| 530 | subcc %GLOBAL_SPARE, 0x8, %GLOBAL_SPARE | 643 | subcc %GLOBAL_SPARE, 0x8, %GLOBAL_SPARE |
| 531 | add %o1, 0x8, %o1 | 644 | add %o1, 0x8, %o1 |
| 532 | srlx %g3, %o3, %o5 | 645 | srlx %g3, %o3, %o5 |
| 533 | or %o5, %g2, %o5 | 646 | or %o5, %g2, %o5 |
| 534 | EX_ST(STORE(stx, %o5, %o0)) | 647 | EX_ST(STORE(stx, %o5, %o0), U1_gs_8_o2_adj) |
| 535 | add %o0, 0x8, %o0 | 648 | add %o0, 0x8, %o0 |
| 536 | bgu,pt %icc, 1b | 649 | bgu,pt %icc, 1b |
| 537 | sllx %g3, %g1, %g2 | 650 | sllx %g3, %g1, %g2 |
| @@ -549,9 +662,9 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 549 | bne,pn %XCC, 90f | 662 | bne,pn %XCC, 90f |
| 550 | sub %o0, %o1, %o3 | 663 | sub %o0, %o1, %o3 |
| 551 | 664 | ||
| 552 | 1: EX_LD(LOAD(lduw, %o1, %g1)) | 665 | 1: EX_LD(LOAD(lduw, %o1, %g1), U1_o2_0) |
| 553 | subcc %o2, 4, %o2 | 666 | subcc %o2, 4, %o2 |
| 554 | EX_ST(STORE(stw, %g1, %o1 + %o3)) | 667 | EX_ST(STORE(stw, %g1, %o1 + %o3), U1_o2_4) |
| 555 | bgu,pt %XCC, 1b | 668 | bgu,pt %XCC, 1b |
| 556 | add %o1, 4, %o1 | 669 | add %o1, 4, %o1 |
| 557 | 670 | ||
| @@ -559,9 +672,9 @@ FUNC_NAME: /* %o0=dst, %o1=src, %o2=len */ | |||
| 559 | mov EX_RETVAL(%o4), %o0 | 672 | mov EX_RETVAL(%o4), %o0 |
| 560 | 673 | ||
| 561 | .align 32 | 674 | .align 32 |
| 562 | 90: EX_LD(LOAD(ldub, %o1, %g1)) | 675 | 90: EX_LD(LOAD(ldub, %o1, %g1), U1_o2_0) |
| 563 | subcc %o2, 1, %o2 | 676 | subcc %o2, 1, %o2 |
| 564 | EX_ST(STORE(stb, %g1, %o1 + %o3)) | 677 | EX_ST(STORE(stb, %g1, %o1 + %o3), U1_o2_1) |
| 565 | bgu,pt %XCC, 90b | 678 | bgu,pt %XCC, 90b |
| 566 | add %o1, 1, %o1 | 679 | add %o1, 1, %o1 |
| 567 | retl | 680 | retl |
