diff options
author | Paul Mundt <lethal@linux-sh.org> | 2008-11-18 03:33:48 -0500 |
---|---|---|
committer | Paul Mundt <lethal@linux-sh.org> | 2008-12-22 04:42:53 -0500 |
commit | 1aad54a99b6ce316c851ba99b2efe41998cfd37d (patch) | |
tree | 287f90771c7da18a0841d40db1c0ce92ea7f5820 /arch/sh/lib/movmem.S | |
parent | e9bf51e5ccc7703226c79888603e157066213700 (diff) |
sh: Migrate necessary libgcc bits in to arch/sh/lib for SUPERH32.
This moves in the necessary libgcc bits for SUPERH32 and drops the
libgcc linking for the regular targets. This in turn allows us to rip
out quite a few hacks both in sh_ksyms_32 and arch/sh/Makefile.
Signed-off-by: Paul Mundt <lethal@linux-sh.org>
Diffstat (limited to 'arch/sh/lib/movmem.S')
-rw-r--r-- | arch/sh/lib/movmem.S | 238 |
1 files changed, 238 insertions, 0 deletions
diff --git a/arch/sh/lib/movmem.S b/arch/sh/lib/movmem.S new file mode 100644 index 000000000000..62075f6bc67c --- /dev/null +++ b/arch/sh/lib/movmem.S | |||
@@ -0,0 +1,238 @@ | |||
1 | /* Copyright (C) 1994, 1995, 1997, 1998, 1999, 2000, 2001, 2002, 2003, | ||
2 | 2004, 2005, 2006 | ||
3 | Free Software Foundation, Inc. | ||
4 | |||
5 | This file is free software; you can redistribute it and/or modify it | ||
6 | under the terms of the GNU General Public License as published by the | ||
7 | Free Software Foundation; either version 2, or (at your option) any | ||
8 | later version. | ||
9 | |||
10 | In addition to the permissions in the GNU General Public License, the | ||
11 | Free Software Foundation gives you unlimited permission to link the | ||
12 | compiled version of this file into combinations with other programs, | ||
13 | and to distribute those combinations without any restriction coming | ||
14 | from the use of this file. (The General Public License restrictions | ||
15 | do apply in other respects; for example, they cover modification of | ||
16 | the file, and distribution when not linked into a combine | ||
17 | executable.) | ||
18 | |||
19 | This file is distributed in the hope that it will be useful, but | ||
20 | WITHOUT ANY WARRANTY; without even the implied warranty of | ||
21 | MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU | ||
22 | General Public License for more details. | ||
23 | |||
24 | You should have received a copy of the GNU General Public License | ||
25 | along with this program; see the file COPYING. If not, write to | ||
26 | the Free Software Foundation, 51 Franklin Street, Fifth Floor, | ||
27 | Boston, MA 02110-1301, USA. */ | ||
28 | |||
29 | !! libgcc routines for the Renesas / SuperH SH CPUs. | ||
30 | !! Contributed by Steve Chamberlain. | ||
31 | !! sac@cygnus.com | ||
32 | |||
33 | !! ashiftrt_r4_x, ___ashrsi3, ___ashlsi3, ___lshrsi3 routines | ||
34 | !! recoded in assembly by Toshiyasu Morita | ||
35 | !! tm@netcom.com | ||
36 | |||
37 | /* SH2 optimizations for ___ashrsi3, ___ashlsi3, ___lshrsi3 and | ||
38 | ELF local label prefixes by J"orn Rennecke | ||
39 | amylaar@cygnus.com */ | ||
40 | |||
41 | .text | ||
42 | .balign 4 | ||
43 | .global __movmem | ||
44 | .global __movstr | ||
45 | .set __movstr, __movmem | ||
46 | /* This would be a lot simpler if r6 contained the byte count | ||
47 | minus 64, and we wouldn't be called here for a byte count of 64. */ | ||
48 | __movmem: | ||
49 | sts.l pr,@-r15 | ||
50 | shll2 r6 | ||
51 | bsr __movmemSI52+2 | ||
52 | mov.l @(48,r5),r0 | ||
53 | .balign 4 | ||
54 | movmem_loop: /* Reached with rts */ | ||
55 | mov.l @(60,r5),r0 | ||
56 | add #-64,r6 | ||
57 | mov.l r0,@(60,r4) | ||
58 | tst r6,r6 | ||
59 | mov.l @(56,r5),r0 | ||
60 | bt movmem_done | ||
61 | mov.l r0,@(56,r4) | ||
62 | cmp/pl r6 | ||
63 | mov.l @(52,r5),r0 | ||
64 | add #64,r5 | ||
65 | mov.l r0,@(52,r4) | ||
66 | add #64,r4 | ||
67 | bt __movmemSI52 | ||
68 | ! done all the large groups, do the remainder | ||
69 | ! jump to movmem+ | ||
70 | mova __movmemSI4+4,r0 | ||
71 | add r6,r0 | ||
72 | jmp @r0 | ||
73 | movmem_done: ! share slot insn, works out aligned. | ||
74 | lds.l @r15+,pr | ||
75 | mov.l r0,@(56,r4) | ||
76 | mov.l @(52,r5),r0 | ||
77 | rts | ||
78 | mov.l r0,@(52,r4) | ||
79 | .balign 4 | ||
80 | |||
81 | .global __movmemSI64 | ||
82 | .global __movstrSI64 | ||
83 | .set __movstrSI64, __movmemSI64 | ||
84 | __movmemSI64: | ||
85 | mov.l @(60,r5),r0 | ||
86 | mov.l r0,@(60,r4) | ||
87 | .global __movmemSI60 | ||
88 | .global __movstrSI60 | ||
89 | .set __movstrSI60, __movmemSI60 | ||
90 | __movmemSI60: | ||
91 | mov.l @(56,r5),r0 | ||
92 | mov.l r0,@(56,r4) | ||
93 | .global __movmemSI56 | ||
94 | .global __movstrSI56 | ||
95 | .set __movstrSI56, __movmemSI56 | ||
96 | __movmemSI56: | ||
97 | mov.l @(52,r5),r0 | ||
98 | mov.l r0,@(52,r4) | ||
99 | .global __movmemSI52 | ||
100 | .global __movstrSI52 | ||
101 | .set __movstrSI52, __movmemSI52 | ||
102 | __movmemSI52: | ||
103 | mov.l @(48,r5),r0 | ||
104 | mov.l r0,@(48,r4) | ||
105 | .global __movmemSI48 | ||
106 | .global __movstrSI48 | ||
107 | .set __movstrSI48, __movmemSI48 | ||
108 | __movmemSI48: | ||
109 | mov.l @(44,r5),r0 | ||
110 | mov.l r0,@(44,r4) | ||
111 | .global __movmemSI44 | ||
112 | .global __movstrSI44 | ||
113 | .set __movstrSI44, __movmemSI44 | ||
114 | __movmemSI44: | ||
115 | mov.l @(40,r5),r0 | ||
116 | mov.l r0,@(40,r4) | ||
117 | .global __movmemSI40 | ||
118 | .global __movstrSI40 | ||
119 | .set __movstrSI40, __movmemSI40 | ||
120 | __movmemSI40: | ||
121 | mov.l @(36,r5),r0 | ||
122 | mov.l r0,@(36,r4) | ||
123 | .global __movmemSI36 | ||
124 | .global __movstrSI36 | ||
125 | .set __movstrSI36, __movmemSI36 | ||
126 | __movmemSI36: | ||
127 | mov.l @(32,r5),r0 | ||
128 | mov.l r0,@(32,r4) | ||
129 | .global __movmemSI32 | ||
130 | .global __movstrSI32 | ||
131 | .set __movstrSI32, __movmemSI32 | ||
132 | __movmemSI32: | ||
133 | mov.l @(28,r5),r0 | ||
134 | mov.l r0,@(28,r4) | ||
135 | .global __movmemSI28 | ||
136 | .global __movstrSI28 | ||
137 | .set __movstrSI28, __movmemSI28 | ||
138 | __movmemSI28: | ||
139 | mov.l @(24,r5),r0 | ||
140 | mov.l r0,@(24,r4) | ||
141 | .global __movmemSI24 | ||
142 | .global __movstrSI24 | ||
143 | .set __movstrSI24, __movmemSI24 | ||
144 | __movmemSI24: | ||
145 | mov.l @(20,r5),r0 | ||
146 | mov.l r0,@(20,r4) | ||
147 | .global __movmemSI20 | ||
148 | .global __movstrSI20 | ||
149 | .set __movstrSI20, __movmemSI20 | ||
150 | __movmemSI20: | ||
151 | mov.l @(16,r5),r0 | ||
152 | mov.l r0,@(16,r4) | ||
153 | .global __movmemSI16 | ||
154 | .global __movstrSI16 | ||
155 | .set __movstrSI16, __movmemSI16 | ||
156 | __movmemSI16: | ||
157 | mov.l @(12,r5),r0 | ||
158 | mov.l r0,@(12,r4) | ||
159 | .global __movmemSI12 | ||
160 | .global __movstrSI12 | ||
161 | .set __movstrSI12, __movmemSI12 | ||
162 | __movmemSI12: | ||
163 | mov.l @(8,r5),r0 | ||
164 | mov.l r0,@(8,r4) | ||
165 | .global __movmemSI8 | ||
166 | .global __movstrSI8 | ||
167 | .set __movstrSI8, __movmemSI8 | ||
168 | __movmemSI8: | ||
169 | mov.l @(4,r5),r0 | ||
170 | mov.l r0,@(4,r4) | ||
171 | .global __movmemSI4 | ||
172 | .global __movstrSI4 | ||
173 | .set __movstrSI4, __movmemSI4 | ||
174 | __movmemSI4: | ||
175 | mov.l @(0,r5),r0 | ||
176 | rts | ||
177 | mov.l r0,@(0,r4) | ||
178 | |||
179 | .global __movmem_i4_even | ||
180 | .global __movstr_i4_even | ||
181 | .set __movstr_i4_even, __movmem_i4_even | ||
182 | |||
183 | .global __movmem_i4_odd | ||
184 | .global __movstr_i4_odd | ||
185 | .set __movstr_i4_odd, __movmem_i4_odd | ||
186 | |||
187 | .global __movmemSI12_i4 | ||
188 | .global __movstrSI12_i4 | ||
189 | .set __movstrSI12_i4, __movmemSI12_i4 | ||
190 | |||
191 | .p2align 5 | ||
192 | L_movmem_2mod4_end: | ||
193 | mov.l r0,@(16,r4) | ||
194 | rts | ||
195 | mov.l r1,@(20,r4) | ||
196 | |||
197 | .p2align 2 | ||
198 | |||
199 | __movmem_i4_even: | ||
200 | mov.l @r5+,r0 | ||
201 | bra L_movmem_start_even | ||
202 | mov.l @r5+,r1 | ||
203 | |||
204 | __movmem_i4_odd: | ||
205 | mov.l @r5+,r1 | ||
206 | add #-4,r4 | ||
207 | mov.l @r5+,r2 | ||
208 | mov.l @r5+,r3 | ||
209 | mov.l r1,@(4,r4) | ||
210 | mov.l r2,@(8,r4) | ||
211 | |||
212 | L_movmem_loop: | ||
213 | mov.l r3,@(12,r4) | ||
214 | dt r6 | ||
215 | mov.l @r5+,r0 | ||
216 | bt/s L_movmem_2mod4_end | ||
217 | mov.l @r5+,r1 | ||
218 | add #16,r4 | ||
219 | L_movmem_start_even: | ||
220 | mov.l @r5+,r2 | ||
221 | mov.l @r5+,r3 | ||
222 | mov.l r0,@r4 | ||
223 | dt r6 | ||
224 | mov.l r1,@(4,r4) | ||
225 | bf/s L_movmem_loop | ||
226 | mov.l r2,@(8,r4) | ||
227 | rts | ||
228 | mov.l r3,@(12,r4) | ||
229 | |||
230 | .p2align 4 | ||
231 | __movmemSI12_i4: | ||
232 | mov.l @r5,r0 | ||
233 | mov.l @(4,r5),r1 | ||
234 | mov.l @(8,r5),r2 | ||
235 | mov.l r0,@r4 | ||
236 | mov.l r1,@(4,r4) | ||
237 | rts | ||
238 | mov.l r2,@(8,r4) | ||