aboutsummaryrefslogtreecommitdiffstats
path: root/arch/frv/include/asm
diff options
context:
space:
mode:
authorDavid Howells <dhowells@redhat.com>2012-03-28 13:30:02 -0400
committerDavid Howells <dhowells@redhat.com>2012-03-28 13:30:02 -0400
commit40ab3317493609c9a8199492588ae7bf88eb12d6 (patch)
tree5758d163e818debc0ac63506be16b858db85a1e8 /arch/frv/include/asm
parentb1a154dbf9ddbf396578642299ce75aa73d01763 (diff)
Disintegrate asm/system.h for FRV
Disintegrate asm/system.h for FRV. Signed-off-by: David Howells <dhowells@redhat.com>
Diffstat (limited to 'arch/frv/include/asm')
-rw-r--r--arch/frv/include/asm/atomic.h57
-rw-r--r--arch/frv/include/asm/barrier.h29
-rw-r--r--arch/frv/include/asm/bug.h2
-rw-r--r--arch/frv/include/asm/cmpxchg.h172
-rw-r--r--arch/frv/include/asm/exec.h17
-rw-r--r--arch/frv/include/asm/switch_to.h35
-rw-r--r--arch/frv/include/asm/system.h162
7 files changed, 261 insertions, 213 deletions
diff --git a/arch/frv/include/asm/atomic.h b/arch/frv/include/asm/atomic.h
index 0d8a7d661740..b86329d0e316 100644
--- a/arch/frv/include/asm/atomic.h
+++ b/arch/frv/include/asm/atomic.h
@@ -16,7 +16,7 @@
16 16
17#include <linux/types.h> 17#include <linux/types.h>
18#include <asm/spr-regs.h> 18#include <asm/spr-regs.h>
19#include <asm/system.h> 19#include <asm/cmpxchg.h>
20 20
21#ifdef CONFIG_SMP 21#ifdef CONFIG_SMP
22#error not SMP safe 22#error not SMP safe
@@ -181,61 +181,6 @@ static inline void atomic64_dec(atomic64_t *v)
181#define atomic64_dec_and_test(v) (atomic64_dec_return((v)) == 0) 181#define atomic64_dec_and_test(v) (atomic64_dec_return((v)) == 0)
182#define atomic64_inc_and_test(v) (atomic64_inc_return((v)) == 0) 182#define atomic64_inc_and_test(v) (atomic64_inc_return((v)) == 0)
183 183
184/*****************************************************************************/
185/*
186 * exchange value with memory
187 */
188extern uint64_t __xchg_64(uint64_t i, volatile void *v);
189
190#ifndef CONFIG_FRV_OUTOFLINE_ATOMIC_OPS
191
192#define xchg(ptr, x) \
193({ \
194 __typeof__(ptr) __xg_ptr = (ptr); \
195 __typeof__(*(ptr)) __xg_orig; \
196 \
197 switch (sizeof(__xg_orig)) { \
198 case 4: \
199 asm volatile( \
200 "swap%I0 %M0,%1" \
201 : "+m"(*__xg_ptr), "=r"(__xg_orig) \
202 : "1"(x) \
203 : "memory" \
204 ); \
205 break; \
206 \
207 default: \
208 __xg_orig = (__typeof__(__xg_orig))0; \
209 asm volatile("break"); \
210 break; \
211 } \
212 \
213 __xg_orig; \
214})
215
216#else
217
218extern uint32_t __xchg_32(uint32_t i, volatile void *v);
219
220#define xchg(ptr, x) \
221({ \
222 __typeof__(ptr) __xg_ptr = (ptr); \
223 __typeof__(*(ptr)) __xg_orig; \
224 \
225 switch (sizeof(__xg_orig)) { \
226 case 4: __xg_orig = (__typeof__(*(ptr))) __xchg_32((uint32_t) x, __xg_ptr); break; \
227 default: \
228 __xg_orig = (__typeof__(__xg_orig))0; \
229 asm volatile("break"); \
230 break; \
231 } \
232 __xg_orig; \
233})
234
235#endif
236
237#define tas(ptr) (xchg((ptr), 1))
238
239#define atomic_cmpxchg(v, old, new) (cmpxchg(&(v)->counter, old, new)) 184#define atomic_cmpxchg(v, old, new) (cmpxchg(&(v)->counter, old, new))
240#define atomic_xchg(v, new) (xchg(&(v)->counter, new)) 185#define atomic_xchg(v, new) (xchg(&(v)->counter, new))
241#define atomic64_cmpxchg(v, old, new) (__cmpxchg_64(old, new, &(v)->counter)) 186#define atomic64_cmpxchg(v, old, new) (__cmpxchg_64(old, new, &(v)->counter))
diff --git a/arch/frv/include/asm/barrier.h b/arch/frv/include/asm/barrier.h
new file mode 100644
index 000000000000..06776ad9f5e9
--- /dev/null
+++ b/arch/frv/include/asm/barrier.h
@@ -0,0 +1,29 @@
1/* FR-V CPU memory barrier definitions
2 *
3 * Copyright (C) 2003 Red Hat, Inc. All Rights Reserved.
4 * Written by David Howells (dhowells@redhat.com)
5 *
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
10 */
11
12#ifndef _ASM_BARRIER_H
13#define _ASM_BARRIER_H
14
15#define nop() asm volatile ("nop"::)
16
17#define mb() asm volatile ("membar" : : :"memory")
18#define rmb() asm volatile ("membar" : : :"memory")
19#define wmb() asm volatile ("membar" : : :"memory")
20#define read_barrier_depends() do { } while (0)
21
22#define smp_mb() barrier()
23#define smp_rmb() barrier()
24#define smp_wmb() barrier()
25#define smp_read_barrier_depends() do {} while(0)
26#define set_mb(var, value) \
27 do { var = (value); barrier(); } while (0)
28
29#endif /* _ASM_BARRIER_H */
diff --git a/arch/frv/include/asm/bug.h b/arch/frv/include/asm/bug.h
index 2e054508a2f6..dd01bcf42ee6 100644
--- a/arch/frv/include/asm/bug.h
+++ b/arch/frv/include/asm/bug.h
@@ -51,4 +51,6 @@ do { \
51 51
52#include <asm-generic/bug.h> 52#include <asm-generic/bug.h>
53 53
54extern void die_if_kernel(const char *, ...) __attribute__((format(printf, 1, 2)));
55
54#endif 56#endif
diff --git a/arch/frv/include/asm/cmpxchg.h b/arch/frv/include/asm/cmpxchg.h
new file mode 100644
index 000000000000..5b04dd0aecab
--- /dev/null
+++ b/arch/frv/include/asm/cmpxchg.h
@@ -0,0 +1,172 @@
1/* xchg and cmpxchg operation emulation for FR-V
2 *
3 * For an explanation of how atomic ops work in this arch, see:
4 * Documentation/frv/atomic-ops.txt
5 *
6 * Copyright (C) 2004 Red Hat, Inc. All Rights Reserved.
7 * Written by David Howells (dhowells@redhat.com)
8 *
9 * This program is free software; you can redistribute it and/or
10 * modify it under the terms of the GNU General Public License
11 * as published by the Free Software Foundation; either version
12 * 2 of the License, or (at your option) any later version.
13 */
14#ifndef _ASM_CMPXCHG_H
15#define _ASM_CMPXCHG_H
16
17#include <linux/types.h>
18
19/*****************************************************************************/
20/*
21 * exchange value with memory
22 */
23extern uint64_t __xchg_64(uint64_t i, volatile void *v);
24
25#ifndef CONFIG_FRV_OUTOFLINE_ATOMIC_OPS
26
27#define xchg(ptr, x) \
28({ \
29 __typeof__(ptr) __xg_ptr = (ptr); \
30 __typeof__(*(ptr)) __xg_orig; \
31 \
32 switch (sizeof(__xg_orig)) { \
33 case 4: \
34 asm volatile( \
35 "swap%I0 %M0,%1" \
36 : "+m"(*__xg_ptr), "=r"(__xg_orig) \
37 : "1"(x) \
38 : "memory" \
39 ); \
40 break; \
41 \
42 default: \
43 __xg_orig = (__typeof__(__xg_orig))0; \
44 asm volatile("break"); \
45 break; \
46 } \
47 \
48 __xg_orig; \
49})
50
51#else
52
53extern uint32_t __xchg_32(uint32_t i, volatile void *v);
54
55#define xchg(ptr, x) \
56({ \
57 __typeof__(ptr) __xg_ptr = (ptr); \
58 __typeof__(*(ptr)) __xg_orig; \
59 \
60 switch (sizeof(__xg_orig)) { \
61 case 4: __xg_orig = (__typeof__(*(ptr))) __xchg_32((uint32_t) x, __xg_ptr); break; \
62 default: \
63 __xg_orig = (__typeof__(__xg_orig))0; \
64 asm volatile("break"); \
65 break; \
66 } \
67 __xg_orig; \
68})
69
70#endif
71
72#define tas(ptr) (xchg((ptr), 1))
73
74/*****************************************************************************/
75/*
76 * compare and conditionally exchange value with memory
77 * - if (*ptr == test) then orig = *ptr; *ptr = test;
78 * - if (*ptr != test) then orig = *ptr;
79 */
80extern uint64_t __cmpxchg_64(uint64_t test, uint64_t new, volatile uint64_t *v);
81
82#ifndef CONFIG_FRV_OUTOFLINE_ATOMIC_OPS
83
84#define cmpxchg(ptr, test, new) \
85({ \
86 __typeof__(ptr) __xg_ptr = (ptr); \
87 __typeof__(*(ptr)) __xg_orig, __xg_tmp; \
88 __typeof__(*(ptr)) __xg_test = (test); \
89 __typeof__(*(ptr)) __xg_new = (new); \
90 \
91 switch (sizeof(__xg_orig)) { \
92 case 4: \
93 asm volatile( \
94 "0: \n" \
95 " orcc gr0,gr0,gr0,icc3 \n" \
96 " ckeq icc3,cc7 \n" \
97 " ld.p %M0,%1 \n" \
98 " orcr cc7,cc7,cc3 \n" \
99 " sub%I4cc %1,%4,%2,icc0 \n" \
100 " bne icc0,#0,1f \n" \
101 " cst.p %3,%M0 ,cc3,#1 \n" \
102 " corcc gr29,gr29,gr0 ,cc3,#1 \n" \
103 " beq icc3,#0,0b \n" \
104 "1: \n" \
105 : "+U"(*__xg_ptr), "=&r"(__xg_orig), "=&r"(__xg_tmp) \
106 : "r"(__xg_new), "NPr"(__xg_test) \
107 : "memory", "cc7", "cc3", "icc3", "icc0" \
108 ); \
109 break; \
110 \
111 default: \
112 __xg_orig = (__typeof__(__xg_orig))0; \
113 asm volatile("break"); \
114 break; \
115 } \
116 \
117 __xg_orig; \
118})
119
120#else
121
122extern uint32_t __cmpxchg_32(uint32_t *v, uint32_t test, uint32_t new);
123
124#define cmpxchg(ptr, test, new) \
125({ \
126 __typeof__(ptr) __xg_ptr = (ptr); \
127 __typeof__(*(ptr)) __xg_orig; \
128 __typeof__(*(ptr)) __xg_test = (test); \
129 __typeof__(*(ptr)) __xg_new = (new); \
130 \
131 switch (sizeof(__xg_orig)) { \
132 case 4: __xg_orig = (__force __typeof__(*ptr)) \
133 __cmpxchg_32((__force uint32_t *)__xg_ptr, \
134 (__force uint32_t)__xg_test, \
135 (__force uint32_t)__xg_new); break; \
136 default: \
137 __xg_orig = (__typeof__(__xg_orig))0; \
138 asm volatile("break"); \
139 break; \
140 } \
141 \
142 __xg_orig; \
143})
144
145#endif
146
147#include <asm-generic/cmpxchg-local.h>
148
149static inline unsigned long __cmpxchg_local(volatile void *ptr,
150 unsigned long old,
151 unsigned long new, int size)
152{
153 switch (size) {
154 case 4:
155 return cmpxchg((unsigned long *)ptr, old, new);
156 default:
157 return __cmpxchg_local_generic(ptr, old, new, size);
158 }
159
160 return old;
161}
162
163/*
164 * cmpxchg_local and cmpxchg64_local are atomic wrt current CPU. Always make
165 * them available.
166 */
167#define cmpxchg_local(ptr, o, n) \
168 ((__typeof__(*(ptr)))__cmpxchg_local((ptr), (unsigned long)(o), \
169 (unsigned long)(n), sizeof(*(ptr))))
170#define cmpxchg64_local(ptr, o, n) __cmpxchg64_local_generic((ptr), (o), (n))
171
172#endif /* _ASM_CMPXCHG_H */
diff --git a/arch/frv/include/asm/exec.h b/arch/frv/include/asm/exec.h
new file mode 100644
index 000000000000..65c91305d4a7
--- /dev/null
+++ b/arch/frv/include/asm/exec.h
@@ -0,0 +1,17 @@
1/* FR-V CPU executable handling
2 *
3 * Copyright (C) 2003 Red Hat, Inc. All Rights Reserved.
4 * Written by David Howells (dhowells@redhat.com)
5 *
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
10 */
11
12#ifndef _ASM_EXEC_H
13#define _ASM_EXEC_H
14
15#define arch_align_stack(x) (x)
16
17#endif /* _ASM_EXEC_H */
diff --git a/arch/frv/include/asm/switch_to.h b/arch/frv/include/asm/switch_to.h
new file mode 100644
index 000000000000..2cf0f6a7fbb1
--- /dev/null
+++ b/arch/frv/include/asm/switch_to.h
@@ -0,0 +1,35 @@
1/* FR-V CPU basic task switching
2 *
3 * Copyright (C) 2003 Red Hat, Inc. All Rights Reserved.
4 * Written by David Howells (dhowells@redhat.com)
5 *
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
10 */
11
12#ifndef _ASM_SWITCH_TO_H
13#define _ASM_SWITCH_TO_H
14
15#include <linux/thread_info.h>
16
17/*
18 * switch_to(prev, next) should switch from task `prev' to `next'
19 * `prev' will never be the same as `next'.
20 * The `mb' is to tell GCC not to cache `current' across this call.
21 */
22extern asmlinkage
23struct task_struct *__switch_to(struct thread_struct *prev_thread,
24 struct thread_struct *next_thread,
25 struct task_struct *prev);
26
27#define switch_to(prev, next, last) \
28do { \
29 (prev)->thread.sched_lr = \
30 (unsigned long) __builtin_return_address(0); \
31 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
32 mb(); \
33} while(0)
34
35#endif /* _ASM_SWITCH_TO_H */
diff --git a/arch/frv/include/asm/system.h b/arch/frv/include/asm/system.h
index 6c10fd2c626d..5c707a235403 100644
--- a/arch/frv/include/asm/system.h
+++ b/arch/frv/include/asm/system.h
@@ -1,158 +1,6 @@
1/* system.h: FR-V CPU control definitions 1/* FILE TO BE DELETED. DO NOT ADD STUFF HERE! */
2 * 2#include <asm/barrier.h>
3 * Copyright (C) 2003 Red Hat, Inc. All Rights Reserved. 3#include <asm/cmpxchg.h>
4 * Written by David Howells (dhowells@redhat.com) 4#include <asm/exec.h>
5 * 5#include <asm/switch_to.h>
6 * This program is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU General Public License
8 * as published by the Free Software Foundation; either version
9 * 2 of the License, or (at your option) any later version.
10 */
11
12#ifndef _ASM_SYSTEM_H
13#define _ASM_SYSTEM_H
14
15#include <linux/types.h>
16#include <linux/linkage.h>
17#include <linux/kernel.h>
18
19struct thread_struct;
20
21/*
22 * switch_to(prev, next) should switch from task `prev' to `next'
23 * `prev' will never be the same as `next'.
24 * The `mb' is to tell GCC not to cache `current' across this call.
25 */
26extern asmlinkage
27struct task_struct *__switch_to(struct thread_struct *prev_thread,
28 struct thread_struct *next_thread,
29 struct task_struct *prev);
30
31#define switch_to(prev, next, last) \
32do { \
33 (prev)->thread.sched_lr = \
34 (unsigned long) __builtin_return_address(0); \
35 (last) = __switch_to(&(prev)->thread, &(next)->thread, (prev)); \
36 mb(); \
37} while(0)
38
39/*
40 * Force strict CPU ordering.
41 */
42#define nop() asm volatile ("nop"::)
43#define mb() asm volatile ("membar" : : :"memory")
44#define rmb() asm volatile ("membar" : : :"memory")
45#define wmb() asm volatile ("membar" : : :"memory")
46#define read_barrier_depends() do { } while (0)
47
48#define smp_mb() barrier()
49#define smp_rmb() barrier()
50#define smp_wmb() barrier()
51#define smp_read_barrier_depends() do {} while(0)
52#define set_mb(var, value) \
53 do { var = (value); barrier(); } while (0)
54
55extern void die_if_kernel(const char *, ...) __attribute__((format(printf, 1, 2)));
56extern void free_initmem(void); 6extern void free_initmem(void);
57
58#define arch_align_stack(x) (x)
59
60/*****************************************************************************/
61/*
62 * compare and conditionally exchange value with memory
63 * - if (*ptr == test) then orig = *ptr; *ptr = test;
64 * - if (*ptr != test) then orig = *ptr;
65 */
66extern uint64_t __cmpxchg_64(uint64_t test, uint64_t new, volatile uint64_t *v);
67
68#ifndef CONFIG_FRV_OUTOFLINE_ATOMIC_OPS
69
70#define cmpxchg(ptr, test, new) \
71({ \
72 __typeof__(ptr) __xg_ptr = (ptr); \
73 __typeof__(*(ptr)) __xg_orig, __xg_tmp; \
74 __typeof__(*(ptr)) __xg_test = (test); \
75 __typeof__(*(ptr)) __xg_new = (new); \
76 \
77 switch (sizeof(__xg_orig)) { \
78 case 4: \
79 asm volatile( \
80 "0: \n" \
81 " orcc gr0,gr0,gr0,icc3 \n" \
82 " ckeq icc3,cc7 \n" \
83 " ld.p %M0,%1 \n" \
84 " orcr cc7,cc7,cc3 \n" \
85 " sub%I4cc %1,%4,%2,icc0 \n" \
86 " bne icc0,#0,1f \n" \
87 " cst.p %3,%M0 ,cc3,#1 \n" \
88 " corcc gr29,gr29,gr0 ,cc3,#1 \n" \
89 " beq icc3,#0,0b \n" \
90 "1: \n" \
91 : "+U"(*__xg_ptr), "=&r"(__xg_orig), "=&r"(__xg_tmp) \
92 : "r"(__xg_new), "NPr"(__xg_test) \
93 : "memory", "cc7", "cc3", "icc3", "icc0" \
94 ); \
95 break; \
96 \
97 default: \
98 __xg_orig = (__typeof__(__xg_orig))0; \
99 asm volatile("break"); \
100 break; \
101 } \
102 \
103 __xg_orig; \
104})
105
106#else
107
108extern uint32_t __cmpxchg_32(uint32_t *v, uint32_t test, uint32_t new);
109
110#define cmpxchg(ptr, test, new) \
111({ \
112 __typeof__(ptr) __xg_ptr = (ptr); \
113 __typeof__(*(ptr)) __xg_orig; \
114 __typeof__(*(ptr)) __xg_test = (test); \
115 __typeof__(*(ptr)) __xg_new = (new); \
116 \
117 switch (sizeof(__xg_orig)) { \
118 case 4: __xg_orig = (__force __typeof__(*ptr)) \
119 __cmpxchg_32((__force uint32_t *)__xg_ptr, \
120 (__force uint32_t)__xg_test, \
121 (__force uint32_t)__xg_new); break; \
122 default: \
123 __xg_orig = (__typeof__(__xg_orig))0; \
124 asm volatile("break"); \
125 break; \
126 } \
127 \
128 __xg_orig; \
129})
130
131#endif
132
133#include <asm-generic/cmpxchg-local.h>
134
135static inline unsigned long __cmpxchg_local(volatile void *ptr,
136 unsigned long old,
137 unsigned long new, int size)
138{
139 switch (size) {
140 case 4:
141 return cmpxchg((unsigned long *)ptr, old, new);
142 default:
143 return __cmpxchg_local_generic(ptr, old, new, size);
144 }
145
146 return old;
147}
148
149/*
150 * cmpxchg_local and cmpxchg64_local are atomic wrt current CPU. Always make
151 * them available.
152 */
153#define cmpxchg_local(ptr, o, n) \
154 ((__typeof__(*(ptr)))__cmpxchg_local((ptr), (unsigned long)(o), \
155 (unsigned long)(n), sizeof(*(ptr))))
156#define cmpxchg64_local(ptr, o, n) __cmpxchg64_local_generic((ptr), (o), (n))
157
158#endif /* _ASM_SYSTEM_H */