diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2014-01-30 20:20:32 -0500 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2014-01-30 20:20:32 -0500 |
commit | cdfc83075fb76369a31e6c187d0cebcab9f8b9c8 (patch) | |
tree | 33d1cdca3e2cb610451ed30943189f55652bac4c /arch/mips/kernel/vpe.c | |
parent | 04a24ae45d018e177db7e4ae2d03a70f79149782 (diff) | |
parent | b26a21c1eacdb7daf22a304fa857413df2650cfe (diff) |
Merge branch 'upstream' of git://git.linux-mips.org/pub/scm/ralf/upstream-linus
Pull MIPS updates from Ralf Baechle:
"The most notable new addition inside this pull request is the support
for MIPS's latest and greatest core called "inter/proAptiv". The
patch series describes this core as follows.
"The interAptiv is a power-efficient multi-core microprocessor
for use in system-on-chip (SoC) applications. The interAptiv combines
a multi-threading pipeline with a coherence manager to deliver improved
computational throughput and power efficiency. The interAptiv can
contain one to four MIPS32R3 interAptiv cores, system level
coherence manager with L2 cache, optional coherent I/O port,
and optional floating point unit."
The platform specific patches touch all 3 Broadcom families. It adds
support for the new Broadcom/Netlogix XLP9xx Soc, building a common
BCM63XX SMP kernel for all BCM63XX SoCs regardless of core type/count
and full gpio button/led descriptions for BCM47xx.
The rest of the series are cleanups and bug fixes that are MIPS
generic and consist largely of changes that Imgtec/MIPS had published
in their linux-mti-3.10.git stable tree. Random other cleanups and
patches preparing code to be merged in 3.15"
* 'upstream' of git://git.linux-mips.org/pub/scm/ralf/upstream-linus: (139 commits)
mips: select ARCH_MIGHT_HAVE_PC_SERIO
mips: delete non-required instances of include <linux/init.h>
MIPS: KVM: remove shadow_tlb code
MIPS: KVM: use common EHINV aware UNIQUE_ENTRYHI
mips/ide: flush dcache also if icache does not snoop dcache
MIPS: BCM47XX: fix position of cpu_wait disabling
MIPS: BCM63XX: select correct MIPS_L1_CACHE_SHIFT value
MIPS: update MIPS_L1_CACHE_SHIFT based on MIPS_L1_CACHE_SHIFT_<N>
MIPS: introduce MIPS_L1_CACHE_SHIFT_<N>
MIPS: ZBOOT: gather string functions into string.c
arch/mips/pci: don't check resource with devm_ioremap_resource
arch/mips/lantiq/xway: don't check resource with devm_ioremap_resource
bcma: gpio: don't cast u32 to unsigned long
ssb: gpio: add own IRQ domain
MIPS: BCM47XX: fix sparse warnings in board.c
MIPS: BCM47XX: add board detection for Linksys WRT54GS V1
MIPS: BCM47XX: fix detection for some boards
MIPS: BCM47XX: Enable buttons support on SSB
MIPS: BCM47XX: Convert WNDR4500 to new syntax
MIPS: BCM47XX: Use "timer" trigger for status LEDs
...
Diffstat (limited to 'arch/mips/kernel/vpe.c')
-rw-r--r-- | arch/mips/kernel/vpe.c | 882 |
1 files changed, 124 insertions, 758 deletions
diff --git a/arch/mips/kernel/vpe.c b/arch/mips/kernel/vpe.c index 2d5c142bad67..11da314565cc 100644 --- a/arch/mips/kernel/vpe.c +++ b/arch/mips/kernel/vpe.c | |||
@@ -1,37 +1,22 @@ | |||
1 | /* | 1 | /* |
2 | * Copyright (C) 2004, 2005 MIPS Technologies, Inc. All rights reserved. | 2 | * This file is subject to the terms and conditions of the GNU General Public |
3 | * | 3 | * License. See the file "COPYING" in the main directory of this archive |
4 | * This program is free software; you can distribute it and/or modify it | 4 | * for more details. |
5 | * under the terms of the GNU General Public License (Version 2) as | ||
6 | * published by the Free Software Foundation. | ||
7 | * | ||
8 | * This program is distributed in the hope it will be useful, but WITHOUT | ||
9 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | ||
10 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License | ||
11 | * for more details. | ||
12 | * | ||
13 | * You should have received a copy of the GNU General Public License along | ||
14 | * with this program; if not, write to the Free Software Foundation, Inc., | ||
15 | * 59 Temple Place - Suite 330, Boston MA 02111-1307, USA. | ||
16 | */ | ||
17 | |||
18 | /* | ||
19 | * VPE support module | ||
20 | * | 5 | * |
21 | * Provides support for loading a MIPS SP program on VPE1. | 6 | * Copyright (C) 2004, 2005 MIPS Technologies, Inc. All rights reserved. |
22 | * The SP environment is rather simple, no tlb's. It needs to be relocatable | 7 | * Copyright (C) 2013 Imagination Technologies Ltd. |
23 | * (or partially linked). You should initialise your stack in the startup | ||
24 | * code. This loader looks for the symbol __start and sets up | ||
25 | * execution to resume from there. The MIPS SDE kit contains suitable examples. | ||
26 | * | 8 | * |
27 | * To load and run, simply cat a SP 'program file' to /dev/vpe1. | 9 | * VPE spport module for loading a MIPS SP program into VPE1. The SP |
28 | * i.e cat spapp >/dev/vpe1. | 10 | * environment is rather simple since there are no TLBs. It needs |
11 | * to be relocatable (or partiall linked). Initialize your stack in | ||
12 | * the startup-code. The loader looks for the symbol __start and sets | ||
13 | * up the execution to resume from there. To load and run, simply do | ||
14 | * a cat SP 'binary' to the /dev/vpe1 device. | ||
29 | */ | 15 | */ |
30 | #include <linux/kernel.h> | 16 | #include <linux/kernel.h> |
31 | #include <linux/device.h> | 17 | #include <linux/device.h> |
32 | #include <linux/fs.h> | 18 | #include <linux/fs.h> |
33 | #include <linux/init.h> | 19 | #include <linux/init.h> |
34 | #include <asm/uaccess.h> | ||
35 | #include <linux/slab.h> | 20 | #include <linux/slab.h> |
36 | #include <linux/list.h> | 21 | #include <linux/list.h> |
37 | #include <linux/vmalloc.h> | 22 | #include <linux/vmalloc.h> |
@@ -46,13 +31,10 @@ | |||
46 | #include <asm/mipsmtregs.h> | 31 | #include <asm/mipsmtregs.h> |
47 | #include <asm/cacheflush.h> | 32 | #include <asm/cacheflush.h> |
48 | #include <linux/atomic.h> | 33 | #include <linux/atomic.h> |
49 | #include <asm/cpu.h> | ||
50 | #include <asm/mips_mt.h> | 34 | #include <asm/mips_mt.h> |
51 | #include <asm/processor.h> | 35 | #include <asm/processor.h> |
52 | #include <asm/vpe.h> | 36 | #include <asm/vpe.h> |
53 | 37 | ||
54 | typedef void *vpe_handle; | ||
55 | |||
56 | #ifndef ARCH_SHF_SMALL | 38 | #ifndef ARCH_SHF_SMALL |
57 | #define ARCH_SHF_SMALL 0 | 39 | #define ARCH_SHF_SMALL 0 |
58 | #endif | 40 | #endif |
@@ -60,95 +42,15 @@ typedef void *vpe_handle; | |||
60 | /* If this is set, the section belongs in the init part of the module */ | 42 | /* If this is set, the section belongs in the init part of the module */ |
61 | #define INIT_OFFSET_MASK (1UL << (BITS_PER_LONG-1)) | 43 | #define INIT_OFFSET_MASK (1UL << (BITS_PER_LONG-1)) |
62 | 44 | ||
63 | /* | 45 | struct vpe_control vpecontrol = { |
64 | * The number of TCs and VPEs physically available on the core | ||
65 | */ | ||
66 | static int hw_tcs, hw_vpes; | ||
67 | static char module_name[] = "vpe"; | ||
68 | static int major; | ||
69 | static const int minor = 1; /* fixed for now */ | ||
70 | |||
71 | /* grab the likely amount of memory we will need. */ | ||
72 | #ifdef CONFIG_MIPS_VPE_LOADER_TOM | ||
73 | #define P_SIZE (2 * 1024 * 1024) | ||
74 | #else | ||
75 | /* add an overhead to the max kmalloc size for non-striped symbols/etc */ | ||
76 | #define P_SIZE (256 * 1024) | ||
77 | #endif | ||
78 | |||
79 | extern unsigned long physical_memsize; | ||
80 | |||
81 | #define MAX_VPES 16 | ||
82 | #define VPE_PATH_MAX 256 | ||
83 | |||
84 | enum vpe_state { | ||
85 | VPE_STATE_UNUSED = 0, | ||
86 | VPE_STATE_INUSE, | ||
87 | VPE_STATE_RUNNING | ||
88 | }; | ||
89 | |||
90 | enum tc_state { | ||
91 | TC_STATE_UNUSED = 0, | ||
92 | TC_STATE_INUSE, | ||
93 | TC_STATE_RUNNING, | ||
94 | TC_STATE_DYNAMIC | ||
95 | }; | ||
96 | |||
97 | struct vpe { | ||
98 | enum vpe_state state; | ||
99 | |||
100 | /* (device) minor associated with this vpe */ | ||
101 | int minor; | ||
102 | |||
103 | /* elfloader stuff */ | ||
104 | void *load_addr; | ||
105 | unsigned long len; | ||
106 | char *pbuffer; | ||
107 | unsigned long plen; | ||
108 | char cwd[VPE_PATH_MAX]; | ||
109 | |||
110 | unsigned long __start; | ||
111 | |||
112 | /* tc's associated with this vpe */ | ||
113 | struct list_head tc; | ||
114 | |||
115 | /* The list of vpe's */ | ||
116 | struct list_head list; | ||
117 | |||
118 | /* shared symbol address */ | ||
119 | void *shared_ptr; | ||
120 | |||
121 | /* the list of who wants to know when something major happens */ | ||
122 | struct list_head notify; | ||
123 | |||
124 | unsigned int ntcs; | ||
125 | }; | ||
126 | |||
127 | struct tc { | ||
128 | enum tc_state state; | ||
129 | int index; | ||
130 | |||
131 | struct vpe *pvpe; /* parent VPE */ | ||
132 | struct list_head tc; /* The list of TC's with this VPE */ | ||
133 | struct list_head list; /* The global list of tc's */ | ||
134 | }; | ||
135 | |||
136 | struct { | ||
137 | spinlock_t vpe_list_lock; | ||
138 | struct list_head vpe_list; /* Virtual processing elements */ | ||
139 | spinlock_t tc_list_lock; | ||
140 | struct list_head tc_list; /* Thread contexts */ | ||
141 | } vpecontrol = { | ||
142 | .vpe_list_lock = __SPIN_LOCK_UNLOCKED(vpe_list_lock), | 46 | .vpe_list_lock = __SPIN_LOCK_UNLOCKED(vpe_list_lock), |
143 | .vpe_list = LIST_HEAD_INIT(vpecontrol.vpe_list), | 47 | .vpe_list = LIST_HEAD_INIT(vpecontrol.vpe_list), |
144 | .tc_list_lock = __SPIN_LOCK_UNLOCKED(tc_list_lock), | 48 | .tc_list_lock = __SPIN_LOCK_UNLOCKED(tc_list_lock), |
145 | .tc_list = LIST_HEAD_INIT(vpecontrol.tc_list) | 49 | .tc_list = LIST_HEAD_INIT(vpecontrol.tc_list) |
146 | }; | 50 | }; |
147 | 51 | ||
148 | static void release_progmem(void *ptr); | ||
149 | |||
150 | /* get the vpe associated with this minor */ | 52 | /* get the vpe associated with this minor */ |
151 | static struct vpe *get_vpe(int minor) | 53 | struct vpe *get_vpe(int minor) |
152 | { | 54 | { |
153 | struct vpe *res, *v; | 55 | struct vpe *res, *v; |
154 | 56 | ||
@@ -158,7 +60,7 @@ static struct vpe *get_vpe(int minor) | |||
158 | res = NULL; | 60 | res = NULL; |
159 | spin_lock(&vpecontrol.vpe_list_lock); | 61 | spin_lock(&vpecontrol.vpe_list_lock); |
160 | list_for_each_entry(v, &vpecontrol.vpe_list, list) { | 62 | list_for_each_entry(v, &vpecontrol.vpe_list, list) { |
161 | if (v->minor == minor) { | 63 | if (v->minor == VPE_MODULE_MINOR) { |
162 | res = v; | 64 | res = v; |
163 | break; | 65 | break; |
164 | } | 66 | } |
@@ -169,7 +71,7 @@ static struct vpe *get_vpe(int minor) | |||
169 | } | 71 | } |
170 | 72 | ||
171 | /* get the vpe associated with this minor */ | 73 | /* get the vpe associated with this minor */ |
172 | static struct tc *get_tc(int index) | 74 | struct tc *get_tc(int index) |
173 | { | 75 | { |
174 | struct tc *res, *t; | 76 | struct tc *res, *t; |
175 | 77 | ||
@@ -187,12 +89,13 @@ static struct tc *get_tc(int index) | |||
187 | } | 89 | } |
188 | 90 | ||
189 | /* allocate a vpe and associate it with this minor (or index) */ | 91 | /* allocate a vpe and associate it with this minor (or index) */ |
190 | static struct vpe *alloc_vpe(int minor) | 92 | struct vpe *alloc_vpe(int minor) |
191 | { | 93 | { |
192 | struct vpe *v; | 94 | struct vpe *v; |
193 | 95 | ||
194 | if ((v = kzalloc(sizeof(struct vpe), GFP_KERNEL)) == NULL) | 96 | v = kzalloc(sizeof(struct vpe), GFP_KERNEL); |
195 | return NULL; | 97 | if (v == NULL) |
98 | goto out; | ||
196 | 99 | ||
197 | INIT_LIST_HEAD(&v->tc); | 100 | INIT_LIST_HEAD(&v->tc); |
198 | spin_lock(&vpecontrol.vpe_list_lock); | 101 | spin_lock(&vpecontrol.vpe_list_lock); |
@@ -200,17 +103,19 @@ static struct vpe *alloc_vpe(int minor) | |||
200 | spin_unlock(&vpecontrol.vpe_list_lock); | 103 | spin_unlock(&vpecontrol.vpe_list_lock); |
201 | 104 | ||
202 | INIT_LIST_HEAD(&v->notify); | 105 | INIT_LIST_HEAD(&v->notify); |
203 | v->minor = minor; | 106 | v->minor = VPE_MODULE_MINOR; |
204 | 107 | ||
108 | out: | ||
205 | return v; | 109 | return v; |
206 | } | 110 | } |
207 | 111 | ||
208 | /* allocate a tc. At startup only tc0 is running, all other can be halted. */ | 112 | /* allocate a tc. At startup only tc0 is running, all other can be halted. */ |
209 | static struct tc *alloc_tc(int index) | 113 | struct tc *alloc_tc(int index) |
210 | { | 114 | { |
211 | struct tc *tc; | 115 | struct tc *tc; |
212 | 116 | ||
213 | if ((tc = kzalloc(sizeof(struct tc), GFP_KERNEL)) == NULL) | 117 | tc = kzalloc(sizeof(struct tc), GFP_KERNEL); |
118 | if (tc == NULL) | ||
214 | goto out; | 119 | goto out; |
215 | 120 | ||
216 | INIT_LIST_HEAD(&tc->tc); | 121 | INIT_LIST_HEAD(&tc->tc); |
@@ -225,7 +130,7 @@ out: | |||
225 | } | 130 | } |
226 | 131 | ||
227 | /* clean up and free everything */ | 132 | /* clean up and free everything */ |
228 | static void release_vpe(struct vpe *v) | 133 | void release_vpe(struct vpe *v) |
229 | { | 134 | { |
230 | list_del(&v->list); | 135 | list_del(&v->list); |
231 | if (v->load_addr) | 136 | if (v->load_addr) |
@@ -233,28 +138,8 @@ static void release_vpe(struct vpe *v) | |||
233 | kfree(v); | 138 | kfree(v); |
234 | } | 139 | } |
235 | 140 | ||
236 | static void __maybe_unused dump_mtregs(void) | 141 | /* Find some VPE program space */ |
237 | { | 142 | void *alloc_progmem(unsigned long len) |
238 | unsigned long val; | ||
239 | |||
240 | val = read_c0_config3(); | ||
241 | printk("config3 0x%lx MT %ld\n", val, | ||
242 | (val & CONFIG3_MT) >> CONFIG3_MT_SHIFT); | ||
243 | |||
244 | val = read_c0_mvpcontrol(); | ||
245 | printk("MVPControl 0x%lx, STLB %ld VPC %ld EVP %ld\n", val, | ||
246 | (val & MVPCONTROL_STLB) >> MVPCONTROL_STLB_SHIFT, | ||
247 | (val & MVPCONTROL_VPC) >> MVPCONTROL_VPC_SHIFT, | ||
248 | (val & MVPCONTROL_EVP)); | ||
249 | |||
250 | val = read_c0_mvpconf0(); | ||
251 | printk("mvpconf0 0x%lx, PVPE %ld PTC %ld M %ld\n", val, | ||
252 | (val & MVPCONF0_PVPE) >> MVPCONF0_PVPE_SHIFT, | ||
253 | val & MVPCONF0_PTC, (val & MVPCONF0_M) >> MVPCONF0_M_SHIFT); | ||
254 | } | ||
255 | |||
256 | /* Find some VPE program space */ | ||
257 | static void *alloc_progmem(unsigned long len) | ||
258 | { | 143 | { |
259 | void *addr; | 144 | void *addr; |
260 | 145 | ||
@@ -273,7 +158,7 @@ static void *alloc_progmem(unsigned long len) | |||
273 | return addr; | 158 | return addr; |
274 | } | 159 | } |
275 | 160 | ||
276 | static void release_progmem(void *ptr) | 161 | void release_progmem(void *ptr) |
277 | { | 162 | { |
278 | #ifndef CONFIG_MIPS_VPE_LOADER_TOM | 163 | #ifndef CONFIG_MIPS_VPE_LOADER_TOM |
279 | kfree(ptr); | 164 | kfree(ptr); |
@@ -281,7 +166,7 @@ static void release_progmem(void *ptr) | |||
281 | } | 166 | } |
282 | 167 | ||
283 | /* Update size with this section: return offset. */ | 168 | /* Update size with this section: return offset. */ |
284 | static long get_offset(unsigned long *size, Elf_Shdr * sechdr) | 169 | static long get_offset(unsigned long *size, Elf_Shdr *sechdr) |
285 | { | 170 | { |
286 | long ret; | 171 | long ret; |
287 | 172 | ||
@@ -294,8 +179,8 @@ static long get_offset(unsigned long *size, Elf_Shdr * sechdr) | |||
294 | might -- code, read-only data, read-write data, small data. Tally | 179 | might -- code, read-only data, read-write data, small data. Tally |
295 | sizes, and place the offsets into sh_entsize fields: high bit means it | 180 | sizes, and place the offsets into sh_entsize fields: high bit means it |
296 | belongs in init. */ | 181 | belongs in init. */ |
297 | static void layout_sections(struct module *mod, const Elf_Ehdr * hdr, | 182 | static void layout_sections(struct module *mod, const Elf_Ehdr *hdr, |
298 | Elf_Shdr * sechdrs, const char *secstrings) | 183 | Elf_Shdr *sechdrs, const char *secstrings) |
299 | { | 184 | { |
300 | static unsigned long const masks[][2] = { | 185 | static unsigned long const masks[][2] = { |
301 | /* NOTE: all executable code must be the first section | 186 | /* NOTE: all executable code must be the first section |
@@ -315,7 +200,6 @@ static void layout_sections(struct module *mod, const Elf_Ehdr * hdr, | |||
315 | for (i = 0; i < hdr->e_shnum; ++i) { | 200 | for (i = 0; i < hdr->e_shnum; ++i) { |
316 | Elf_Shdr *s = &sechdrs[i]; | 201 | Elf_Shdr *s = &sechdrs[i]; |
317 | 202 | ||
318 | // || strncmp(secstrings + s->sh_name, ".init", 5) == 0) | ||
319 | if ((s->sh_flags & masks[m][0]) != masks[m][0] | 203 | if ((s->sh_flags & masks[m][0]) != masks[m][0] |
320 | || (s->sh_flags & masks[m][1]) | 204 | || (s->sh_flags & masks[m][1]) |
321 | || s->sh_entsize != ~0UL) | 205 | || s->sh_entsize != ~0UL) |
@@ -330,7 +214,6 @@ static void layout_sections(struct module *mod, const Elf_Ehdr * hdr, | |||
330 | } | 214 | } |
331 | } | 215 | } |
332 | 216 | ||
333 | |||
334 | /* from module-elf32.c, but subverted a little */ | 217 | /* from module-elf32.c, but subverted a little */ |
335 | 218 | ||
336 | struct mips_hi16 { | 219 | struct mips_hi16 { |
@@ -353,20 +236,18 @@ static int apply_r_mips_gprel16(struct module *me, uint32_t *location, | |||
353 | { | 236 | { |
354 | int rel; | 237 | int rel; |
355 | 238 | ||
356 | if( !(*location & 0xffff) ) { | 239 | if (!(*location & 0xffff)) { |
357 | rel = (int)v - gp_addr; | 240 | rel = (int)v - gp_addr; |
358 | } | 241 | } else { |
359 | else { | ||
360 | /* .sbss + gp(relative) + offset */ | 242 | /* .sbss + gp(relative) + offset */ |
361 | /* kludge! */ | 243 | /* kludge! */ |
362 | rel = (int)(short)((int)v + gp_offs + | 244 | rel = (int)(short)((int)v + gp_offs + |
363 | (int)(short)(*location & 0xffff) - gp_addr); | 245 | (int)(short)(*location & 0xffff) - gp_addr); |
364 | } | 246 | } |
365 | 247 | ||
366 | if( (rel > 32768) || (rel < -32768) ) { | 248 | if ((rel > 32768) || (rel < -32768)) { |
367 | printk(KERN_DEBUG "VPE loader: apply_r_mips_gprel16: " | 249 | pr_debug("VPE loader: apply_r_mips_gprel16: relative address 0x%x out of range of gp register\n", |
368 | "relative address 0x%x out of range of gp register\n", | 250 | rel); |
369 | rel); | ||
370 | return -ENOEXEC; | 251 | return -ENOEXEC; |
371 | } | 252 | } |
372 | 253 | ||
@@ -380,12 +261,12 @@ static int apply_r_mips_pc16(struct module *me, uint32_t *location, | |||
380 | { | 261 | { |
381 | int rel; | 262 | int rel; |
382 | rel = (((unsigned int)v - (unsigned int)location)); | 263 | rel = (((unsigned int)v - (unsigned int)location)); |
383 | rel >>= 2; // because the offset is in _instructions_ not bytes. | 264 | rel >>= 2; /* because the offset is in _instructions_ not bytes. */ |
384 | rel -= 1; // and one instruction less due to the branch delay slot. | 265 | rel -= 1; /* and one instruction less due to the branch delay slot. */ |
385 | 266 | ||
386 | if( (rel > 32768) || (rel < -32768) ) { | 267 | if ((rel > 32768) || (rel < -32768)) { |
387 | printk(KERN_DEBUG "VPE loader: " | 268 | pr_debug("VPE loader: apply_r_mips_pc16: relative address out of range 0x%x\n", |
388 | "apply_r_mips_pc16: relative address out of range 0x%x\n", rel); | 269 | rel); |
389 | return -ENOEXEC; | 270 | return -ENOEXEC; |
390 | } | 271 | } |
391 | 272 | ||
@@ -406,8 +287,7 @@ static int apply_r_mips_26(struct module *me, uint32_t *location, | |||
406 | Elf32_Addr v) | 287 | Elf32_Addr v) |
407 | { | 288 | { |
408 | if (v % 4) { | 289 | if (v % 4) { |
409 | printk(KERN_DEBUG "VPE loader: apply_r_mips_26 " | 290 | pr_debug("VPE loader: apply_r_mips_26: unaligned relocation\n"); |
410 | " unaligned relocation\n"); | ||
411 | return -ENOEXEC; | 291 | return -ENOEXEC; |
412 | } | 292 | } |
413 | 293 | ||
@@ -438,7 +318,7 @@ static int apply_r_mips_hi16(struct module *me, uint32_t *location, | |||
438 | * the carry we need to add. Save the information, and let LO16 do the | 318 | * the carry we need to add. Save the information, and let LO16 do the |
439 | * actual relocation. | 319 | * actual relocation. |
440 | */ | 320 | */ |
441 | n = kmalloc(sizeof *n, GFP_KERNEL); | 321 | n = kmalloc(sizeof(*n), GFP_KERNEL); |
442 | if (!n) | 322 | if (!n) |
443 | return -ENOMEM; | 323 | return -ENOMEM; |
444 | 324 | ||
@@ -470,9 +350,7 @@ static int apply_r_mips_lo16(struct module *me, uint32_t *location, | |||
470 | * The value for the HI16 had best be the same. | 350 | * The value for the HI16 had best be the same. |
471 | */ | 351 | */ |
472 | if (v != l->value) { | 352 | if (v != l->value) { |
473 | printk(KERN_DEBUG "VPE loader: " | 353 | pr_debug("VPE loader: apply_r_mips_lo16/hi16: inconsistent value information\n"); |
474 | "apply_r_mips_lo16/hi16: \t" | ||
475 | "inconsistent value information\n"); | ||
476 | goto out_free; | 354 | goto out_free; |
477 | } | 355 | } |
478 | 356 | ||
@@ -568,20 +446,19 @@ static int apply_relocations(Elf32_Shdr *sechdrs, | |||
568 | + ELF32_R_SYM(r_info); | 446 | + ELF32_R_SYM(r_info); |
569 | 447 | ||
570 | if (!sym->st_value) { | 448 | if (!sym->st_value) { |
571 | printk(KERN_DEBUG "%s: undefined weak symbol %s\n", | 449 | pr_debug("%s: undefined weak symbol %s\n", |
572 | me->name, strtab + sym->st_name); | 450 | me->name, strtab + sym->st_name); |
573 | /* just print the warning, dont barf */ | 451 | /* just print the warning, dont barf */ |
574 | } | 452 | } |
575 | 453 | ||
576 | v = sym->st_value; | 454 | v = sym->st_value; |
577 | 455 | ||
578 | res = reloc_handlers[ELF32_R_TYPE(r_info)](me, location, v); | 456 | res = reloc_handlers[ELF32_R_TYPE(r_info)](me, location, v); |
579 | if( res ) { | 457 | if (res) { |
580 | char *r = rstrs[ELF32_R_TYPE(r_info)]; | 458 | char *r = rstrs[ELF32_R_TYPE(r_info)]; |
581 | printk(KERN_WARNING "VPE loader: .text+0x%x " | 459 | pr_warn("VPE loader: .text+0x%x relocation type %s for symbol \"%s\" failed\n", |
582 | "relocation type %s for symbol \"%s\" failed\n", | 460 | rel[i].r_offset, r ? r : "UNKNOWN", |
583 | rel[i].r_offset, r ? r : "UNKNOWN", | 461 | strtab + sym->st_name); |
584 | strtab + sym->st_name); | ||
585 | return res; | 462 | return res; |
586 | } | 463 | } |
587 | } | 464 | } |
@@ -596,10 +473,8 @@ static inline void save_gp_address(unsigned int secbase, unsigned int rel) | |||
596 | } | 473 | } |
597 | /* end module-elf32.c */ | 474 | /* end module-elf32.c */ |
598 | 475 | ||
599 | |||
600 | |||
601 | /* Change all symbols so that sh_value encodes the pointer directly. */ | 476 | /* Change all symbols so that sh_value encodes the pointer directly. */ |
602 | static void simplify_symbols(Elf_Shdr * sechdrs, | 477 | static void simplify_symbols(Elf_Shdr *sechdrs, |
603 | unsigned int symindex, | 478 | unsigned int symindex, |
604 | const char *strtab, | 479 | const char *strtab, |
605 | const char *secstrings, | 480 | const char *secstrings, |
@@ -640,18 +515,16 @@ static void simplify_symbols(Elf_Shdr * sechdrs, | |||
640 | break; | 515 | break; |
641 | 516 | ||
642 | case SHN_MIPS_SCOMMON: | 517 | case SHN_MIPS_SCOMMON: |
643 | printk(KERN_DEBUG "simplify_symbols: ignoring SHN_MIPS_SCOMMON " | 518 | pr_debug("simplify_symbols: ignoring SHN_MIPS_SCOMMON symbol <%s> st_shndx %d\n", |
644 | "symbol <%s> st_shndx %d\n", strtab + sym[i].st_name, | 519 | strtab + sym[i].st_name, sym[i].st_shndx); |
645 | sym[i].st_shndx); | 520 | /* .sbss section */ |
646 | // .sbss section | ||
647 | break; | 521 | break; |
648 | 522 | ||
649 | default: | 523 | default: |
650 | secbase = sechdrs[sym[i].st_shndx].sh_addr; | 524 | secbase = sechdrs[sym[i].st_shndx].sh_addr; |
651 | 525 | ||
652 | if (strncmp(strtab + sym[i].st_name, "_gp", 3) == 0) { | 526 | if (strncmp(strtab + sym[i].st_name, "_gp", 3) == 0) |
653 | save_gp_address(secbase, sym[i].st_value); | 527 | save_gp_address(secbase, sym[i].st_value); |
654 | } | ||
655 | 528 | ||
656 | sym[i].st_value += secbase; | 529 | sym[i].st_value += secbase; |
657 | break; | 530 | break; |
@@ -660,142 +533,21 @@ static void simplify_symbols(Elf_Shdr * sechdrs, | |||
660 | } | 533 | } |
661 | 534 | ||
662 | #ifdef DEBUG_ELFLOADER | 535 | #ifdef DEBUG_ELFLOADER |
663 | static void dump_elfsymbols(Elf_Shdr * sechdrs, unsigned int symindex, | 536 | static void dump_elfsymbols(Elf_Shdr *sechdrs, unsigned int symindex, |
664 | const char *strtab, struct module *mod) | 537 | const char *strtab, struct module *mod) |
665 | { | 538 | { |
666 | Elf_Sym *sym = (void *)sechdrs[symindex].sh_addr; | 539 | Elf_Sym *sym = (void *)sechdrs[symindex].sh_addr; |
667 | unsigned int i, n = sechdrs[symindex].sh_size / sizeof(Elf_Sym); | 540 | unsigned int i, n = sechdrs[symindex].sh_size / sizeof(Elf_Sym); |
668 | 541 | ||
669 | printk(KERN_DEBUG "dump_elfsymbols: n %d\n", n); | 542 | pr_debug("dump_elfsymbols: n %d\n", n); |
670 | for (i = 1; i < n; i++) { | 543 | for (i = 1; i < n; i++) { |
671 | printk(KERN_DEBUG " i %d name <%s> 0x%x\n", i, | 544 | pr_debug(" i %d name <%s> 0x%x\n", i, strtab + sym[i].st_name, |
672 | strtab + sym[i].st_name, sym[i].st_value); | 545 | sym[i].st_value); |
673 | } | 546 | } |
674 | } | 547 | } |
675 | #endif | 548 | #endif |
676 | 549 | ||
677 | /* We are prepared so configure and start the VPE... */ | 550 | static int find_vpe_symbols(struct vpe *v, Elf_Shdr *sechdrs, |
678 | static int vpe_run(struct vpe * v) | ||
679 | { | ||
680 | unsigned long flags, val, dmt_flag; | ||
681 | struct vpe_notifications *n; | ||
682 | unsigned int vpeflags; | ||
683 | struct tc *t; | ||
684 | |||
685 | /* check we are the Master VPE */ | ||
686 | local_irq_save(flags); | ||
687 | val = read_c0_vpeconf0(); | ||
688 | if (!(val & VPECONF0_MVP)) { | ||
689 | printk(KERN_WARNING | ||
690 | "VPE loader: only Master VPE's are allowed to configure MT\n"); | ||
691 | local_irq_restore(flags); | ||
692 | |||
693 | return -1; | ||
694 | } | ||
695 | |||
696 | dmt_flag = dmt(); | ||
697 | vpeflags = dvpe(); | ||
698 | |||
699 | if (list_empty(&v->tc)) { | ||
700 | evpe(vpeflags); | ||
701 | emt(dmt_flag); | ||
702 | local_irq_restore(flags); | ||
703 | |||
704 | printk(KERN_WARNING | ||
705 | "VPE loader: No TC's associated with VPE %d\n", | ||
706 | v->minor); | ||
707 | |||
708 | return -ENOEXEC; | ||
709 | } | ||
710 | |||
711 | t = list_first_entry(&v->tc, struct tc, tc); | ||
712 | |||
713 | /* Put MVPE's into 'configuration state' */ | ||
714 | set_c0_mvpcontrol(MVPCONTROL_VPC); | ||
715 | |||
716 | settc(t->index); | ||
717 | |||
718 | /* should check it is halted, and not activated */ | ||
719 | if ((read_tc_c0_tcstatus() & TCSTATUS_A) || !(read_tc_c0_tchalt() & TCHALT_H)) { | ||
720 | evpe(vpeflags); | ||
721 | emt(dmt_flag); | ||
722 | local_irq_restore(flags); | ||
723 | |||
724 | printk(KERN_WARNING "VPE loader: TC %d is already active!\n", | ||
725 | t->index); | ||
726 | |||
727 | return -ENOEXEC; | ||
728 | } | ||
729 | |||
730 | /* Write the address we want it to start running from in the TCPC register. */ | ||
731 | write_tc_c0_tcrestart((unsigned long)v->__start); | ||
732 | write_tc_c0_tccontext((unsigned long)0); | ||
733 | |||
734 | /* | ||
735 | * Mark the TC as activated, not interrupt exempt and not dynamically | ||
736 | * allocatable | ||
737 | */ | ||
738 | val = read_tc_c0_tcstatus(); | ||
739 | val = (val & ~(TCSTATUS_DA | TCSTATUS_IXMT)) | TCSTATUS_A; | ||
740 | write_tc_c0_tcstatus(val); | ||
741 | |||
742 | write_tc_c0_tchalt(read_tc_c0_tchalt() & ~TCHALT_H); | ||
743 | |||
744 | /* | ||
745 | * The sde-kit passes 'memsize' to __start in $a3, so set something | ||
746 | * here... Or set $a3 to zero and define DFLT_STACK_SIZE and | ||
747 | * DFLT_HEAP_SIZE when you compile your program | ||
748 | */ | ||
749 | mttgpr(6, v->ntcs); | ||
750 | mttgpr(7, physical_memsize); | ||
751 | |||
752 | /* set up VPE1 */ | ||
753 | /* | ||
754 | * bind the TC to VPE 1 as late as possible so we only have the final | ||
755 | * VPE registers to set up, and so an EJTAG probe can trigger on it | ||
756 | */ | ||
757 | write_tc_c0_tcbind((read_tc_c0_tcbind() & ~TCBIND_CURVPE) | 1); | ||
758 | |||
759 | write_vpe_c0_vpeconf0(read_vpe_c0_vpeconf0() & ~(VPECONF0_VPA)); | ||
760 | |||
761 | back_to_back_c0_hazard(); | ||
762 | |||
763 | /* Set up the XTC bit in vpeconf0 to point at our tc */ | ||
764 | write_vpe_c0_vpeconf0( (read_vpe_c0_vpeconf0() & ~(VPECONF0_XTC)) | ||
765 | | (t->index << VPECONF0_XTC_SHIFT)); | ||
766 | |||
767 | back_to_back_c0_hazard(); | ||
768 | |||
769 | /* enable this VPE */ | ||
770 | write_vpe_c0_vpeconf0(read_vpe_c0_vpeconf0() | VPECONF0_VPA); | ||
771 | |||
772 | /* clear out any left overs from a previous program */ | ||
773 | write_vpe_c0_status(0); | ||
774 | write_vpe_c0_cause(0); | ||
775 | |||
776 | /* take system out of configuration state */ | ||
777 | clear_c0_mvpcontrol(MVPCONTROL_VPC); | ||
778 | |||
779 | /* | ||
780 | * SMTC/SMVP kernels manage VPE enable independently, | ||
781 | * but uniprocessor kernels need to turn it on, even | ||
782 | * if that wasn't the pre-dvpe() state. | ||
783 | */ | ||
784 | #ifdef CONFIG_SMP | ||
785 | evpe(vpeflags); | ||
786 | #else | ||
787 | evpe(EVPE_ENABLE); | ||
788 | #endif | ||
789 | emt(dmt_flag); | ||
790 | local_irq_restore(flags); | ||
791 | |||
792 | list_for_each_entry(n, &v->notify, list) | ||
793 | n->start(minor); | ||
794 | |||
795 | return 0; | ||
796 | } | ||
797 | |||
798 | static int find_vpe_symbols(struct vpe * v, Elf_Shdr * sechdrs, | ||
799 | unsigned int symindex, const char *strtab, | 551 | unsigned int symindex, const char *strtab, |
800 | struct module *mod) | 552 | struct module *mod) |
801 | { | 553 | { |
@@ -803,16 +555,14 @@ static int find_vpe_symbols(struct vpe * v, Elf_Shdr * sechdrs, | |||
803 | unsigned int i, n = sechdrs[symindex].sh_size / sizeof(Elf_Sym); | 555 | unsigned int i, n = sechdrs[symindex].sh_size / sizeof(Elf_Sym); |
804 | 556 | ||
805 | for (i = 1; i < n; i++) { | 557 | for (i = 1; i < n; i++) { |
806 | if (strcmp(strtab + sym[i].st_name, "__start") == 0) { | 558 | if (strcmp(strtab + sym[i].st_name, "__start") == 0) |
807 | v->__start = sym[i].st_value; | 559 | v->__start = sym[i].st_value; |
808 | } | ||
809 | 560 | ||
810 | if (strcmp(strtab + sym[i].st_name, "vpe_shared") == 0) { | 561 | if (strcmp(strtab + sym[i].st_name, "vpe_shared") == 0) |
811 | v->shared_ptr = (void *)sym[i].st_value; | 562 | v->shared_ptr = (void *)sym[i].st_value; |
812 | } | ||
813 | } | 563 | } |
814 | 564 | ||
815 | if ( (v->__start == 0) || (v->shared_ptr == NULL)) | 565 | if ((v->__start == 0) || (v->shared_ptr == NULL)) |
816 | return -1; | 566 | return -1; |
817 | 567 | ||
818 | return 0; | 568 | return 0; |
@@ -823,14 +573,14 @@ static int find_vpe_symbols(struct vpe * v, Elf_Shdr * sechdrs, | |||
823 | * contents of the program (p)buffer performing relocatations/etc, free's it | 573 | * contents of the program (p)buffer performing relocatations/etc, free's it |
824 | * when finished. | 574 | * when finished. |
825 | */ | 575 | */ |
826 | static int vpe_elfload(struct vpe * v) | 576 | static int vpe_elfload(struct vpe *v) |
827 | { | 577 | { |
828 | Elf_Ehdr *hdr; | 578 | Elf_Ehdr *hdr; |
829 | Elf_Shdr *sechdrs; | 579 | Elf_Shdr *sechdrs; |
830 | long err = 0; | 580 | long err = 0; |
831 | char *secstrings, *strtab = NULL; | 581 | char *secstrings, *strtab = NULL; |
832 | unsigned int len, i, symindex = 0, strindex = 0, relocate = 0; | 582 | unsigned int len, i, symindex = 0, strindex = 0, relocate = 0; |
833 | struct module mod; // so we can re-use the relocations code | 583 | struct module mod; /* so we can re-use the relocations code */ |
834 | 584 | ||
835 | memset(&mod, 0, sizeof(struct module)); | 585 | memset(&mod, 0, sizeof(struct module)); |
836 | strcpy(mod.name, "VPE loader"); | 586 | strcpy(mod.name, "VPE loader"); |
@@ -844,8 +594,7 @@ static int vpe_elfload(struct vpe * v) | |||
844 | || (hdr->e_type != ET_REL && hdr->e_type != ET_EXEC) | 594 | || (hdr->e_type != ET_REL && hdr->e_type != ET_EXEC) |
845 | || !elf_check_arch(hdr) | 595 | || !elf_check_arch(hdr) |
846 | || hdr->e_shentsize != sizeof(*sechdrs)) { | 596 | || hdr->e_shentsize != sizeof(*sechdrs)) { |
847 | printk(KERN_WARNING | 597 | pr_warn("VPE loader: program wrong arch or weird elf version\n"); |
848 | "VPE loader: program wrong arch or weird elf version\n"); | ||
849 | 598 | ||
850 | return -ENOEXEC; | 599 | return -ENOEXEC; |
851 | } | 600 | } |
@@ -854,8 +603,7 @@ static int vpe_elfload(struct vpe * v) | |||
854 | relocate = 1; | 603 | relocate = 1; |
855 | 604 | ||
856 | if (len < hdr->e_shoff + hdr->e_shnum * sizeof(Elf_Shdr)) { | 605 | if (len < hdr->e_shoff + hdr->e_shnum * sizeof(Elf_Shdr)) { |
857 | printk(KERN_ERR "VPE loader: program length %u truncated\n", | 606 | pr_err("VPE loader: program length %u truncated\n", len); |
858 | len); | ||
859 | 607 | ||
860 | return -ENOEXEC; | 608 | return -ENOEXEC; |
861 | } | 609 | } |
@@ -870,22 +618,24 @@ static int vpe_elfload(struct vpe * v) | |||
870 | 618 | ||
871 | if (relocate) { | 619 | if (relocate) { |
872 | for (i = 1; i < hdr->e_shnum; i++) { | 620 | for (i = 1; i < hdr->e_shnum; i++) { |
873 | if (sechdrs[i].sh_type != SHT_NOBITS | 621 | if ((sechdrs[i].sh_type != SHT_NOBITS) && |
874 | && len < sechdrs[i].sh_offset + sechdrs[i].sh_size) { | 622 | (len < sechdrs[i].sh_offset + sechdrs[i].sh_size)) { |
875 | printk(KERN_ERR "VPE program length %u truncated\n", | 623 | pr_err("VPE program length %u truncated\n", |
876 | len); | 624 | len); |
877 | return -ENOEXEC; | 625 | return -ENOEXEC; |
878 | } | 626 | } |
879 | 627 | ||
880 | /* Mark all sections sh_addr with their address in the | 628 | /* Mark all sections sh_addr with their address in the |
881 | temporary image. */ | 629 | temporary image. */ |
882 | sechdrs[i].sh_addr = (size_t) hdr + sechdrs[i].sh_offset; | 630 | sechdrs[i].sh_addr = (size_t) hdr + |
631 | sechdrs[i].sh_offset; | ||
883 | 632 | ||
884 | /* Internal symbols and strings. */ | 633 | /* Internal symbols and strings. */ |
885 | if (sechdrs[i].sh_type == SHT_SYMTAB) { | 634 | if (sechdrs[i].sh_type == SHT_SYMTAB) { |
886 | symindex = i; | 635 | symindex = i; |
887 | strindex = sechdrs[i].sh_link; | 636 | strindex = sechdrs[i].sh_link; |
888 | strtab = (char *)hdr + sechdrs[strindex].sh_offset; | 637 | strtab = (char *)hdr + |
638 | sechdrs[strindex].sh_offset; | ||
889 | } | 639 | } |
890 | } | 640 | } |
891 | layout_sections(&mod, hdr, sechdrs, secstrings); | 641 | layout_sections(&mod, hdr, sechdrs, secstrings); |
@@ -912,8 +662,9 @@ static int vpe_elfload(struct vpe * v) | |||
912 | /* Update sh_addr to point to copy in image. */ | 662 | /* Update sh_addr to point to copy in image. */ |
913 | sechdrs[i].sh_addr = (unsigned long)dest; | 663 | sechdrs[i].sh_addr = (unsigned long)dest; |
914 | 664 | ||
915 | printk(KERN_DEBUG " section sh_name %s sh_addr 0x%x\n", | 665 | pr_debug(" section sh_name %s sh_addr 0x%x\n", |
916 | secstrings + sechdrs[i].sh_name, sechdrs[i].sh_addr); | 666 | secstrings + sechdrs[i].sh_name, |
667 | sechdrs[i].sh_addr); | ||
917 | } | 668 | } |
918 | 669 | ||
919 | /* Fix up syms, so that st_value is a pointer to location. */ | 670 | /* Fix up syms, so that st_value is a pointer to location. */ |
@@ -934,17 +685,18 @@ static int vpe_elfload(struct vpe * v) | |||
934 | continue; | 685 | continue; |
935 | 686 | ||
936 | if (sechdrs[i].sh_type == SHT_REL) | 687 | if (sechdrs[i].sh_type == SHT_REL) |
937 | err = apply_relocations(sechdrs, strtab, symindex, i, | 688 | err = apply_relocations(sechdrs, strtab, |
938 | &mod); | 689 | symindex, i, &mod); |
939 | else if (sechdrs[i].sh_type == SHT_RELA) | 690 | else if (sechdrs[i].sh_type == SHT_RELA) |
940 | err = apply_relocate_add(sechdrs, strtab, symindex, i, | 691 | err = apply_relocate_add(sechdrs, strtab, |
941 | &mod); | 692 | symindex, i, &mod); |
942 | if (err < 0) | 693 | if (err < 0) |
943 | return err; | 694 | return err; |
944 | 695 | ||
945 | } | 696 | } |
946 | } else { | 697 | } else { |
947 | struct elf_phdr *phdr = (struct elf_phdr *) ((char *)hdr + hdr->e_phoff); | 698 | struct elf_phdr *phdr = (struct elf_phdr *) |
699 | ((char *)hdr + hdr->e_phoff); | ||
948 | 700 | ||
949 | for (i = 0; i < hdr->e_phnum; i++) { | 701 | for (i = 0; i < hdr->e_phnum; i++) { |
950 | if (phdr->p_type == PT_LOAD) { | 702 | if (phdr->p_type == PT_LOAD) { |
@@ -962,11 +714,15 @@ static int vpe_elfload(struct vpe * v) | |||
962 | if (sechdrs[i].sh_type == SHT_SYMTAB) { | 714 | if (sechdrs[i].sh_type == SHT_SYMTAB) { |
963 | symindex = i; | 715 | symindex = i; |
964 | strindex = sechdrs[i].sh_link; | 716 | strindex = sechdrs[i].sh_link; |
965 | strtab = (char *)hdr + sechdrs[strindex].sh_offset; | 717 | strtab = (char *)hdr + |
718 | sechdrs[strindex].sh_offset; | ||
966 | 719 | ||
967 | /* mark the symtab's address for when we try to find the | 720 | /* |
968 | magic symbols */ | 721 | * mark symtab's address for when we try |
969 | sechdrs[i].sh_addr = (size_t) hdr + sechdrs[i].sh_offset; | 722 | * to find the magic symbols |
723 | */ | ||
724 | sechdrs[i].sh_addr = (size_t) hdr + | ||
725 | sechdrs[i].sh_offset; | ||
970 | } | 726 | } |
971 | } | 727 | } |
972 | } | 728 | } |
@@ -977,53 +733,19 @@ static int vpe_elfload(struct vpe * v) | |||
977 | 733 | ||
978 | if ((find_vpe_symbols(v, sechdrs, symindex, strtab, &mod)) < 0) { | 734 | if ((find_vpe_symbols(v, sechdrs, symindex, strtab, &mod)) < 0) { |
979 | if (v->__start == 0) { | 735 | if (v->__start == 0) { |
980 | printk(KERN_WARNING "VPE loader: program does not contain " | 736 | pr_warn("VPE loader: program does not contain a __start symbol\n"); |
981 | "a __start symbol\n"); | ||
982 | return -ENOEXEC; | 737 | return -ENOEXEC; |
983 | } | 738 | } |
984 | 739 | ||
985 | if (v->shared_ptr == NULL) | 740 | if (v->shared_ptr == NULL) |
986 | printk(KERN_WARNING "VPE loader: " | 741 | pr_warn("VPE loader: program does not contain vpe_shared symbol.\n" |
987 | "program does not contain vpe_shared symbol.\n" | 742 | " Unable to use AMVP (AP/SP) facilities.\n"); |
988 | " Unable to use AMVP (AP/SP) facilities.\n"); | ||
989 | } | 743 | } |
990 | 744 | ||
991 | printk(" elf loaded\n"); | 745 | pr_info(" elf loaded\n"); |
992 | return 0; | 746 | return 0; |
993 | } | 747 | } |
994 | 748 | ||
995 | static void cleanup_tc(struct tc *tc) | ||
996 | { | ||
997 | unsigned long flags; | ||
998 | unsigned int mtflags, vpflags; | ||
999 | int tmp; | ||
1000 | |||
1001 | local_irq_save(flags); | ||
1002 | mtflags = dmt(); | ||
1003 | vpflags = dvpe(); | ||
1004 | /* Put MVPE's into 'configuration state' */ | ||
1005 | set_c0_mvpcontrol(MVPCONTROL_VPC); | ||
1006 | |||
1007 | settc(tc->index); | ||
1008 | tmp = read_tc_c0_tcstatus(); | ||
1009 | |||
1010 | /* mark not allocated and not dynamically allocatable */ | ||
1011 | tmp &= ~(TCSTATUS_A | TCSTATUS_DA); | ||
1012 | tmp |= TCSTATUS_IXMT; /* interrupt exempt */ | ||
1013 | write_tc_c0_tcstatus(tmp); | ||
1014 | |||
1015 | write_tc_c0_tchalt(TCHALT_H); | ||
1016 | mips_ihb(); | ||
1017 | |||
1018 | /* bind it to anything other than VPE1 */ | ||
1019 | // write_tc_c0_tcbind(read_tc_c0_tcbind() & ~TCBIND_CURVPE); // | TCBIND_CURVPE | ||
1020 | |||
1021 | clear_c0_mvpcontrol(MVPCONTROL_VPC); | ||
1022 | evpe(vpflags); | ||
1023 | emt(mtflags); | ||
1024 | local_irq_restore(flags); | ||
1025 | } | ||
1026 | |||
1027 | static int getcwd(char *buff, int size) | 749 | static int getcwd(char *buff, int size) |
1028 | { | 750 | { |
1029 | mm_segment_t old_fs; | 751 | mm_segment_t old_fs; |
@@ -1043,39 +765,39 @@ static int getcwd(char *buff, int size) | |||
1043 | static int vpe_open(struct inode *inode, struct file *filp) | 765 | static int vpe_open(struct inode *inode, struct file *filp) |
1044 | { | 766 | { |
1045 | enum vpe_state state; | 767 | enum vpe_state state; |
1046 | struct vpe_notifications *not; | 768 | struct vpe_notifications *notifier; |
1047 | struct vpe *v; | 769 | struct vpe *v; |
1048 | int ret; | 770 | int ret; |
1049 | 771 | ||
1050 | if (minor != iminor(inode)) { | 772 | if (VPE_MODULE_MINOR != iminor(inode)) { |
1051 | /* assume only 1 device at the moment. */ | 773 | /* assume only 1 device at the moment. */ |
1052 | pr_warning("VPE loader: only vpe1 is supported\n"); | 774 | pr_warn("VPE loader: only vpe1 is supported\n"); |
1053 | 775 | ||
1054 | return -ENODEV; | 776 | return -ENODEV; |
1055 | } | 777 | } |
1056 | 778 | ||
1057 | if ((v = get_vpe(tclimit)) == NULL) { | 779 | v = get_vpe(aprp_cpu_index()); |
1058 | pr_warning("VPE loader: unable to get vpe\n"); | 780 | if (v == NULL) { |
781 | pr_warn("VPE loader: unable to get vpe\n"); | ||
1059 | 782 | ||
1060 | return -ENODEV; | 783 | return -ENODEV; |
1061 | } | 784 | } |
1062 | 785 | ||
1063 | state = xchg(&v->state, VPE_STATE_INUSE); | 786 | state = xchg(&v->state, VPE_STATE_INUSE); |
1064 | if (state != VPE_STATE_UNUSED) { | 787 | if (state != VPE_STATE_UNUSED) { |
1065 | printk(KERN_DEBUG "VPE loader: tc in use dumping regs\n"); | 788 | pr_debug("VPE loader: tc in use dumping regs\n"); |
1066 | 789 | ||
1067 | list_for_each_entry(not, &v->notify, list) { | 790 | list_for_each_entry(notifier, &v->notify, list) |
1068 | not->stop(tclimit); | 791 | notifier->stop(aprp_cpu_index()); |
1069 | } | ||
1070 | 792 | ||
1071 | release_progmem(v->load_addr); | 793 | release_progmem(v->load_addr); |
1072 | cleanup_tc(get_tc(tclimit)); | 794 | cleanup_tc(get_tc(aprp_cpu_index())); |
1073 | } | 795 | } |
1074 | 796 | ||
1075 | /* this of-course trashes what was there before... */ | 797 | /* this of-course trashes what was there before... */ |
1076 | v->pbuffer = vmalloc(P_SIZE); | 798 | v->pbuffer = vmalloc(P_SIZE); |
1077 | if (!v->pbuffer) { | 799 | if (!v->pbuffer) { |
1078 | pr_warning("VPE loader: unable to allocate memory\n"); | 800 | pr_warn("VPE loader: unable to allocate memory\n"); |
1079 | return -ENOMEM; | 801 | return -ENOMEM; |
1080 | } | 802 | } |
1081 | v->plen = P_SIZE; | 803 | v->plen = P_SIZE; |
@@ -1085,7 +807,7 @@ static int vpe_open(struct inode *inode, struct file *filp) | |||
1085 | v->cwd[0] = 0; | 807 | v->cwd[0] = 0; |
1086 | ret = getcwd(v->cwd, VPE_PATH_MAX); | 808 | ret = getcwd(v->cwd, VPE_PATH_MAX); |
1087 | if (ret < 0) | 809 | if (ret < 0) |
1088 | printk(KERN_WARNING "VPE loader: open, getcwd returned %d\n", ret); | 810 | pr_warn("VPE loader: open, getcwd returned %d\n", ret); |
1089 | 811 | ||
1090 | v->shared_ptr = NULL; | 812 | v->shared_ptr = NULL; |
1091 | v->__start = 0; | 813 | v->__start = 0; |
@@ -1099,20 +821,20 @@ static int vpe_release(struct inode *inode, struct file *filp) | |||
1099 | Elf_Ehdr *hdr; | 821 | Elf_Ehdr *hdr; |
1100 | int ret = 0; | 822 | int ret = 0; |
1101 | 823 | ||
1102 | v = get_vpe(tclimit); | 824 | v = get_vpe(aprp_cpu_index()); |
1103 | if (v == NULL) | 825 | if (v == NULL) |
1104 | return -ENODEV; | 826 | return -ENODEV; |
1105 | 827 | ||
1106 | hdr = (Elf_Ehdr *) v->pbuffer; | 828 | hdr = (Elf_Ehdr *) v->pbuffer; |
1107 | if (memcmp(hdr->e_ident, ELFMAG, SELFMAG) == 0) { | 829 | if (memcmp(hdr->e_ident, ELFMAG, SELFMAG) == 0) { |
1108 | if (vpe_elfload(v) >= 0) { | 830 | if ((vpe_elfload(v) >= 0) && vpe_run) { |
1109 | vpe_run(v); | 831 | vpe_run(v); |
1110 | } else { | 832 | } else { |
1111 | printk(KERN_WARNING "VPE loader: ELF load failed.\n"); | 833 | pr_warn("VPE loader: ELF load failed.\n"); |
1112 | ret = -ENOEXEC; | 834 | ret = -ENOEXEC; |
1113 | } | 835 | } |
1114 | } else { | 836 | } else { |
1115 | printk(KERN_WARNING "VPE loader: only elf files are supported\n"); | 837 | pr_warn("VPE loader: only elf files are supported\n"); |
1116 | ret = -ENOEXEC; | 838 | ret = -ENOEXEC; |
1117 | } | 839 | } |
1118 | 840 | ||
@@ -1130,22 +852,22 @@ static int vpe_release(struct inode *inode, struct file *filp) | |||
1130 | return ret; | 852 | return ret; |
1131 | } | 853 | } |
1132 | 854 | ||
1133 | static ssize_t vpe_write(struct file *file, const char __user * buffer, | 855 | static ssize_t vpe_write(struct file *file, const char __user *buffer, |
1134 | size_t count, loff_t * ppos) | 856 | size_t count, loff_t *ppos) |
1135 | { | 857 | { |
1136 | size_t ret = count; | 858 | size_t ret = count; |
1137 | struct vpe *v; | 859 | struct vpe *v; |
1138 | 860 | ||
1139 | if (iminor(file_inode(file)) != minor) | 861 | if (iminor(file_inode(file)) != VPE_MODULE_MINOR) |
1140 | return -ENODEV; | 862 | return -ENODEV; |
1141 | 863 | ||
1142 | v = get_vpe(tclimit); | 864 | v = get_vpe(aprp_cpu_index()); |
865 | |||
1143 | if (v == NULL) | 866 | if (v == NULL) |
1144 | return -ENODEV; | 867 | return -ENODEV; |
1145 | 868 | ||
1146 | if ((count + v->len) > v->plen) { | 869 | if ((count + v->len) > v->plen) { |
1147 | printk(KERN_WARNING | 870 | pr_warn("VPE loader: elf size too big. Perhaps strip uneeded symbols\n"); |
1148 | "VPE loader: elf size too big. Perhaps strip uneeded symbols\n"); | ||
1149 | return -ENOMEM; | 871 | return -ENOMEM; |
1150 | } | 872 | } |
1151 | 873 | ||
@@ -1157,7 +879,7 @@ static ssize_t vpe_write(struct file *file, const char __user * buffer, | |||
1157 | return ret; | 879 | return ret; |
1158 | } | 880 | } |
1159 | 881 | ||
1160 | static const struct file_operations vpe_fops = { | 882 | const struct file_operations vpe_fops = { |
1161 | .owner = THIS_MODULE, | 883 | .owner = THIS_MODULE, |
1162 | .open = vpe_open, | 884 | .open = vpe_open, |
1163 | .release = vpe_release, | 885 | .release = vpe_release, |
@@ -1165,396 +887,40 @@ static const struct file_operations vpe_fops = { | |||
1165 | .llseek = noop_llseek, | 887 | .llseek = noop_llseek, |
1166 | }; | 888 | }; |
1167 | 889 | ||
1168 | /* module wrapper entry points */ | ||
1169 | /* give me a vpe */ | ||
1170 | vpe_handle vpe_alloc(void) | ||
1171 | { | ||
1172 | int i; | ||
1173 | struct vpe *v; | ||
1174 | |||
1175 | /* find a vpe */ | ||
1176 | for (i = 1; i < MAX_VPES; i++) { | ||
1177 | if ((v = get_vpe(i)) != NULL) { | ||
1178 | v->state = VPE_STATE_INUSE; | ||
1179 | return v; | ||
1180 | } | ||
1181 | } | ||
1182 | return NULL; | ||
1183 | } | ||
1184 | |||
1185 | EXPORT_SYMBOL(vpe_alloc); | ||
1186 | |||
1187 | /* start running from here */ | ||
1188 | int vpe_start(vpe_handle vpe, unsigned long start) | ||
1189 | { | ||
1190 | struct vpe *v = vpe; | ||
1191 | |||
1192 | v->__start = start; | ||
1193 | return vpe_run(v); | ||
1194 | } | ||
1195 | |||
1196 | EXPORT_SYMBOL(vpe_start); | ||
1197 | |||
1198 | /* halt it for now */ | ||
1199 | int vpe_stop(vpe_handle vpe) | ||
1200 | { | ||
1201 | struct vpe *v = vpe; | ||
1202 | struct tc *t; | ||
1203 | unsigned int evpe_flags; | ||
1204 | |||
1205 | evpe_flags = dvpe(); | ||
1206 | |||
1207 | if ((t = list_entry(v->tc.next, struct tc, tc)) != NULL) { | ||
1208 | |||
1209 | settc(t->index); | ||
1210 | write_vpe_c0_vpeconf0(read_vpe_c0_vpeconf0() & ~VPECONF0_VPA); | ||
1211 | } | ||
1212 | |||
1213 | evpe(evpe_flags); | ||
1214 | |||
1215 | return 0; | ||
1216 | } | ||
1217 | |||
1218 | EXPORT_SYMBOL(vpe_stop); | ||
1219 | |||
1220 | /* I've done with it thank you */ | ||
1221 | int vpe_free(vpe_handle vpe) | ||
1222 | { | ||
1223 | struct vpe *v = vpe; | ||
1224 | struct tc *t; | ||
1225 | unsigned int evpe_flags; | ||
1226 | |||
1227 | if ((t = list_entry(v->tc.next, struct tc, tc)) == NULL) { | ||
1228 | return -ENOEXEC; | ||
1229 | } | ||
1230 | |||
1231 | evpe_flags = dvpe(); | ||
1232 | |||
1233 | /* Put MVPE's into 'configuration state' */ | ||
1234 | set_c0_mvpcontrol(MVPCONTROL_VPC); | ||
1235 | |||
1236 | settc(t->index); | ||
1237 | write_vpe_c0_vpeconf0(read_vpe_c0_vpeconf0() & ~VPECONF0_VPA); | ||
1238 | |||
1239 | /* halt the TC */ | ||
1240 | write_tc_c0_tchalt(TCHALT_H); | ||
1241 | mips_ihb(); | ||
1242 | |||
1243 | /* mark the TC unallocated */ | ||
1244 | write_tc_c0_tcstatus(read_tc_c0_tcstatus() & ~TCSTATUS_A); | ||
1245 | |||
1246 | v->state = VPE_STATE_UNUSED; | ||
1247 | |||
1248 | clear_c0_mvpcontrol(MVPCONTROL_VPC); | ||
1249 | evpe(evpe_flags); | ||
1250 | |||
1251 | return 0; | ||
1252 | } | ||
1253 | |||
1254 | EXPORT_SYMBOL(vpe_free); | ||
1255 | |||
1256 | void *vpe_get_shared(int index) | 890 | void *vpe_get_shared(int index) |
1257 | { | 891 | { |
1258 | struct vpe *v; | 892 | struct vpe *v = get_vpe(index); |
1259 | 893 | ||
1260 | if ((v = get_vpe(index)) == NULL) | 894 | if (v == NULL) |
1261 | return NULL; | 895 | return NULL; |
1262 | 896 | ||
1263 | return v->shared_ptr; | 897 | return v->shared_ptr; |
1264 | } | 898 | } |
1265 | |||
1266 | EXPORT_SYMBOL(vpe_get_shared); | 899 | EXPORT_SYMBOL(vpe_get_shared); |
1267 | 900 | ||
1268 | int vpe_notify(int index, struct vpe_notifications *notify) | 901 | int vpe_notify(int index, struct vpe_notifications *notify) |
1269 | { | 902 | { |
1270 | struct vpe *v; | 903 | struct vpe *v = get_vpe(index); |
1271 | 904 | ||
1272 | if ((v = get_vpe(index)) == NULL) | 905 | if (v == NULL) |
1273 | return -1; | 906 | return -1; |
1274 | 907 | ||
1275 | list_add(¬ify->list, &v->notify); | 908 | list_add(¬ify->list, &v->notify); |
1276 | return 0; | 909 | return 0; |
1277 | } | 910 | } |
1278 | |||
1279 | EXPORT_SYMBOL(vpe_notify); | 911 | EXPORT_SYMBOL(vpe_notify); |
1280 | 912 | ||
1281 | char *vpe_getcwd(int index) | 913 | char *vpe_getcwd(int index) |
1282 | { | 914 | { |
1283 | struct vpe *v; | 915 | struct vpe *v = get_vpe(index); |
1284 | 916 | ||
1285 | if ((v = get_vpe(index)) == NULL) | 917 | if (v == NULL) |
1286 | return NULL; | 918 | return NULL; |
1287 | 919 | ||
1288 | return v->cwd; | 920 | return v->cwd; |
1289 | } | 921 | } |
1290 | |||
1291 | EXPORT_SYMBOL(vpe_getcwd); | 922 | EXPORT_SYMBOL(vpe_getcwd); |
1292 | 923 | ||
1293 | static ssize_t store_kill(struct device *dev, struct device_attribute *attr, | ||
1294 | const char *buf, size_t len) | ||
1295 | { | ||
1296 | struct vpe *vpe = get_vpe(tclimit); | ||
1297 | struct vpe_notifications *not; | ||
1298 | |||
1299 | list_for_each_entry(not, &vpe->notify, list) { | ||
1300 | not->stop(tclimit); | ||
1301 | } | ||
1302 | |||
1303 | release_progmem(vpe->load_addr); | ||
1304 | cleanup_tc(get_tc(tclimit)); | ||
1305 | vpe_stop(vpe); | ||
1306 | vpe_free(vpe); | ||
1307 | |||
1308 | return len; | ||
1309 | } | ||
1310 | static DEVICE_ATTR(kill, S_IWUSR, NULL, store_kill); | ||
1311 | |||
1312 | static ssize_t ntcs_show(struct device *cd, struct device_attribute *attr, | ||
1313 | char *buf) | ||
1314 | { | ||
1315 | struct vpe *vpe = get_vpe(tclimit); | ||
1316 | |||
1317 | return sprintf(buf, "%d\n", vpe->ntcs); | ||
1318 | } | ||
1319 | |||
1320 | static ssize_t ntcs_store(struct device *dev, struct device_attribute *attr, | ||
1321 | const char *buf, size_t len) | ||
1322 | { | ||
1323 | struct vpe *vpe = get_vpe(tclimit); | ||
1324 | unsigned long new; | ||
1325 | char *endp; | ||
1326 | |||
1327 | new = simple_strtoul(buf, &endp, 0); | ||
1328 | if (endp == buf) | ||
1329 | goto out_einval; | ||
1330 | |||
1331 | if (new == 0 || new > (hw_tcs - tclimit)) | ||
1332 | goto out_einval; | ||
1333 | |||
1334 | vpe->ntcs = new; | ||
1335 | |||
1336 | return len; | ||
1337 | |||
1338 | out_einval: | ||
1339 | return -EINVAL; | ||
1340 | } | ||
1341 | static DEVICE_ATTR_RW(ntcs); | ||
1342 | |||
1343 | static struct attribute *vpe_attrs[] = { | ||
1344 | &dev_attr_kill.attr, | ||
1345 | &dev_attr_ntcs.attr, | ||
1346 | NULL, | ||
1347 | }; | ||
1348 | ATTRIBUTE_GROUPS(vpe); | ||
1349 | |||
1350 | static void vpe_device_release(struct device *cd) | ||
1351 | { | ||
1352 | kfree(cd); | ||
1353 | } | ||
1354 | |||
1355 | struct class vpe_class = { | ||
1356 | .name = "vpe", | ||
1357 | .owner = THIS_MODULE, | ||
1358 | .dev_release = vpe_device_release, | ||
1359 | .dev_groups = vpe_groups, | ||
1360 | }; | ||
1361 | |||
1362 | struct device vpe_device; | ||
1363 | |||
1364 | static int __init vpe_module_init(void) | ||
1365 | { | ||
1366 | unsigned int mtflags, vpflags; | ||
1367 | unsigned long flags, val; | ||
1368 | struct vpe *v = NULL; | ||
1369 | struct tc *t; | ||
1370 | int tc, err; | ||
1371 | |||
1372 | if (!cpu_has_mipsmt) { | ||
1373 | printk("VPE loader: not a MIPS MT capable processor\n"); | ||
1374 | return -ENODEV; | ||
1375 | } | ||
1376 | |||
1377 | if (vpelimit == 0) { | ||
1378 | printk(KERN_WARNING "No VPEs reserved for AP/SP, not " | ||
1379 | "initializing VPE loader.\nPass maxvpes=<n> argument as " | ||
1380 | "kernel argument\n"); | ||
1381 | |||
1382 | return -ENODEV; | ||
1383 | } | ||
1384 | |||
1385 | if (tclimit == 0) { | ||
1386 | printk(KERN_WARNING "No TCs reserved for AP/SP, not " | ||
1387 | "initializing VPE loader.\nPass maxtcs=<n> argument as " | ||
1388 | "kernel argument\n"); | ||
1389 | |||
1390 | return -ENODEV; | ||
1391 | } | ||
1392 | |||
1393 | major = register_chrdev(0, module_name, &vpe_fops); | ||
1394 | if (major < 0) { | ||
1395 | printk("VPE loader: unable to register character device\n"); | ||
1396 | return major; | ||
1397 | } | ||
1398 | |||
1399 | err = class_register(&vpe_class); | ||
1400 | if (err) { | ||
1401 | printk(KERN_ERR "vpe_class registration failed\n"); | ||
1402 | goto out_chrdev; | ||
1403 | } | ||
1404 | |||
1405 | device_initialize(&vpe_device); | ||
1406 | vpe_device.class = &vpe_class, | ||
1407 | vpe_device.parent = NULL, | ||
1408 | dev_set_name(&vpe_device, "vpe1"); | ||
1409 | vpe_device.devt = MKDEV(major, minor); | ||
1410 | err = device_add(&vpe_device); | ||
1411 | if (err) { | ||
1412 | printk(KERN_ERR "Adding vpe_device failed\n"); | ||
1413 | goto out_class; | ||
1414 | } | ||
1415 | |||
1416 | local_irq_save(flags); | ||
1417 | mtflags = dmt(); | ||
1418 | vpflags = dvpe(); | ||
1419 | |||
1420 | /* Put MVPE's into 'configuration state' */ | ||
1421 | set_c0_mvpcontrol(MVPCONTROL_VPC); | ||
1422 | |||
1423 | /* dump_mtregs(); */ | ||
1424 | |||
1425 | val = read_c0_mvpconf0(); | ||
1426 | hw_tcs = (val & MVPCONF0_PTC) + 1; | ||
1427 | hw_vpes = ((val & MVPCONF0_PVPE) >> MVPCONF0_PVPE_SHIFT) + 1; | ||
1428 | |||
1429 | for (tc = tclimit; tc < hw_tcs; tc++) { | ||
1430 | /* | ||
1431 | * Must re-enable multithreading temporarily or in case we | ||
1432 | * reschedule send IPIs or similar we might hang. | ||
1433 | */ | ||
1434 | clear_c0_mvpcontrol(MVPCONTROL_VPC); | ||
1435 | evpe(vpflags); | ||
1436 | emt(mtflags); | ||
1437 | local_irq_restore(flags); | ||
1438 | t = alloc_tc(tc); | ||
1439 | if (!t) { | ||
1440 | err = -ENOMEM; | ||
1441 | goto out; | ||
1442 | } | ||
1443 | |||
1444 | local_irq_save(flags); | ||
1445 | mtflags = dmt(); | ||
1446 | vpflags = dvpe(); | ||
1447 | set_c0_mvpcontrol(MVPCONTROL_VPC); | ||
1448 | |||
1449 | /* VPE's */ | ||
1450 | if (tc < hw_tcs) { | ||
1451 | settc(tc); | ||
1452 | |||
1453 | if ((v = alloc_vpe(tc)) == NULL) { | ||
1454 | printk(KERN_WARNING "VPE: unable to allocate VPE\n"); | ||
1455 | |||
1456 | goto out_reenable; | ||
1457 | } | ||
1458 | |||
1459 | v->ntcs = hw_tcs - tclimit; | ||
1460 | |||
1461 | /* add the tc to the list of this vpe's tc's. */ | ||
1462 | list_add(&t->tc, &v->tc); | ||
1463 | |||
1464 | /* deactivate all but vpe0 */ | ||
1465 | if (tc >= tclimit) { | ||
1466 | unsigned long tmp = read_vpe_c0_vpeconf0(); | ||
1467 | |||
1468 | tmp &= ~VPECONF0_VPA; | ||
1469 | |||
1470 | /* master VPE */ | ||
1471 | tmp |= VPECONF0_MVP; | ||
1472 | write_vpe_c0_vpeconf0(tmp); | ||
1473 | } | ||
1474 | |||
1475 | /* disable multi-threading with TC's */ | ||
1476 | write_vpe_c0_vpecontrol(read_vpe_c0_vpecontrol() & ~VPECONTROL_TE); | ||
1477 | |||
1478 | if (tc >= vpelimit) { | ||
1479 | /* | ||
1480 | * Set config to be the same as vpe0, | ||
1481 | * particularly kseg0 coherency alg | ||
1482 | */ | ||
1483 | write_vpe_c0_config(read_c0_config()); | ||
1484 | } | ||
1485 | } | ||
1486 | |||
1487 | /* TC's */ | ||
1488 | t->pvpe = v; /* set the parent vpe */ | ||
1489 | |||
1490 | if (tc >= tclimit) { | ||
1491 | unsigned long tmp; | ||
1492 | |||
1493 | settc(tc); | ||
1494 | |||
1495 | /* Any TC that is bound to VPE0 gets left as is - in case | ||
1496 | we are running SMTC on VPE0. A TC that is bound to any | ||
1497 | other VPE gets bound to VPE0, ideally I'd like to make | ||
1498 | it homeless but it doesn't appear to let me bind a TC | ||
1499 | to a non-existent VPE. Which is perfectly reasonable. | ||
1500 | |||
1501 | The (un)bound state is visible to an EJTAG probe so may | ||
1502 | notify GDB... | ||
1503 | */ | ||
1504 | |||
1505 | if (((tmp = read_tc_c0_tcbind()) & TCBIND_CURVPE)) { | ||
1506 | /* tc is bound >vpe0 */ | ||
1507 | write_tc_c0_tcbind(tmp & ~TCBIND_CURVPE); | ||
1508 | |||
1509 | t->pvpe = get_vpe(0); /* set the parent vpe */ | ||
1510 | } | ||
1511 | |||
1512 | /* halt the TC */ | ||
1513 | write_tc_c0_tchalt(TCHALT_H); | ||
1514 | mips_ihb(); | ||
1515 | |||
1516 | tmp = read_tc_c0_tcstatus(); | ||
1517 | |||
1518 | /* mark not activated and not dynamically allocatable */ | ||
1519 | tmp &= ~(TCSTATUS_A | TCSTATUS_DA); | ||
1520 | tmp |= TCSTATUS_IXMT; /* interrupt exempt */ | ||
1521 | write_tc_c0_tcstatus(tmp); | ||
1522 | } | ||
1523 | } | ||
1524 | |||
1525 | out_reenable: | ||
1526 | /* release config state */ | ||
1527 | clear_c0_mvpcontrol(MVPCONTROL_VPC); | ||
1528 | |||
1529 | evpe(vpflags); | ||
1530 | emt(mtflags); | ||
1531 | local_irq_restore(flags); | ||
1532 | |||
1533 | return 0; | ||
1534 | |||
1535 | out_class: | ||
1536 | class_unregister(&vpe_class); | ||
1537 | out_chrdev: | ||
1538 | unregister_chrdev(major, module_name); | ||
1539 | |||
1540 | out: | ||
1541 | return err; | ||
1542 | } | ||
1543 | |||
1544 | static void __exit vpe_module_exit(void) | ||
1545 | { | ||
1546 | struct vpe *v, *n; | ||
1547 | |||
1548 | device_del(&vpe_device); | ||
1549 | unregister_chrdev(major, module_name); | ||
1550 | |||
1551 | /* No locking needed here */ | ||
1552 | list_for_each_entry_safe(v, n, &vpecontrol.vpe_list, list) { | ||
1553 | if (v->state != VPE_STATE_UNUSED) | ||
1554 | release_vpe(v); | ||
1555 | } | ||
1556 | } | ||
1557 | |||
1558 | module_init(vpe_module_init); | 924 | module_init(vpe_module_init); |
1559 | module_exit(vpe_module_exit); | 925 | module_exit(vpe_module_exit); |
1560 | MODULE_DESCRIPTION("MIPS VPE Loader"); | 926 | MODULE_DESCRIPTION("MIPS VPE Loader"); |