1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
|
/* MN10300 Optimised simple memory to memory copy
*
* Copyright (C) 2007 Red Hat, Inc. All Rights Reserved.
* Written by David Howells (dhowells@redhat.com)
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public Licence
* as published by the Free Software Foundation; either version
* 2 of the Licence, or (at your option) any later version.
*/
#include <asm/cache.h>
.section .text
.balign L1_CACHE_BYTES
###############################################################################
#
# void *memcpy(void *dst, const void *src, size_t n)
#
###############################################################################
.globl memcpy
.type memcpy,@function
memcpy:
movm [d2,d3],(sp)
mov d0,(12,sp)
mov d1,(16,sp)
mov (20,sp),d2 # count
mov d0,a0 # dst
mov d1,a1 # src
mov d0,e3 # the return value
cmp +0,d2
beq memcpy_done # return if zero-length copy
# see if the three parameters are all four-byte aligned
or d0,d1,d3
or d2,d3
and +3,d3
bne memcpy_1 # jump if not
# we want to transfer as much as we can in chunks of 32 bytes
cmp +31,d2
bls memcpy_4_remainder # 4-byte aligned remainder
movm [exreg1],(sp)
add -32,d2
mov +32,d3
memcpy_4_loop:
mov (a1+),d0
mov (a1+),d1
mov (a1+),e0
mov (a1+),e1
mov (a1+),e4
mov (a1+),e5
mov (a1+),e6
mov (a1+),e7
mov d0,(a0+)
mov d1,(a0+)
mov e0,(a0+)
mov e1,(a0+)
mov e4,(a0+)
mov e5,(a0+)
mov e6,(a0+)
mov e7,(a0+)
sub d3,d2
bcc memcpy_4_loop
movm (sp),[exreg1]
add d3,d2
beq memcpy_4_no_remainder
memcpy_4_remainder:
# cut 4-7 words down to 0-3
cmp +16,d2
bcs memcpy_4_three_or_fewer_words
mov (a1+),d0
mov (a1+),d1
mov (a1+),e0
mov (a1+),e1
mov d0,(a0+)
mov d1,(a0+)
mov e0,(a0+)
mov e1,(a0+)
add -16,d2
beq memcpy_4_no_remainder
# copy the remaining 1, 2 or 3 words
memcpy_4_three_or_fewer_words:
cmp +8,d2
bcs memcpy_4_one_word
beq memcpy_4_two_words
mov (a1+),d0
mov d0,(a0+)
memcpy_4_two_words:
mov (a1+),d0
mov d0,(a0+)
memcpy_4_one_word:
mov (a1+),d0
mov d0,(a0+)
memcpy_4_no_remainder:
# check we copied the correct amount
# TODO: REMOVE CHECK
sub e3,a0,d2
mov (20,sp),d1
cmp d2,d1
beq memcpy_done
break
break
break
memcpy_done:
mov e3,a0
ret [d2,d3],8
# handle misaligned copying
memcpy_1:
add -1,d2
mov +1,d3
setlb # setlb requires the next insns
# to occupy exactly 4 bytes
sub d3,d2
movbu (a1),d0
movbu d0,(a0)
add_add d3,a1,d3,a0
lcc
mov e3,a0
ret [d2,d3],8
memcpy_end:
.size memcpy, memcpy_end-memcpy
|