diff options
author | Franck Bui-Huu <fbuihuu@gmail.com> | 2007-10-18 03:11:14 -0400 |
---|---|---|
committer | Ralf Baechle <ralf@linux-mips.org> | 2008-01-29 05:14:56 -0500 |
commit | a9af6041e9e7ed0278d03333c3a22a8509f33615 (patch) | |
tree | 1d05905c82e0a36c5c4c5d04c1933224fb82ae92 /arch/mips/mm | |
parent | 3f4ac2dc0504962b6cb9367adbc13272393b80b3 (diff) |
[MIPS] tlbex.c: Cleanup __init usages.
Signed-off-by: Franck Bui-Huu <fbuihuu@gmail.com>
Signed-off-by: Ralf Baechle <ralf@linux-mips.org>
Diffstat (limited to 'arch/mips/mm')
-rw-r--r-- | arch/mips/mm/tlbex.c | 98 |
1 files changed, 49 insertions, 49 deletions
diff --git a/arch/mips/mm/tlbex.c b/arch/mips/mm/tlbex.c index f8925ba0b39e..77530fd96f1c 100644 --- a/arch/mips/mm/tlbex.c +++ b/arch/mips/mm/tlbex.c | |||
@@ -67,7 +67,7 @@ static inline int __maybe_unused r10000_llsc_war(void) | |||
67 | * why; it's not an issue caused by the core RTL. | 67 | * why; it's not an issue caused by the core RTL. |
68 | * | 68 | * |
69 | */ | 69 | */ |
70 | static __init int __attribute__((unused)) m4kc_tlbp_war(void) | 70 | static int __init m4kc_tlbp_war(void) |
71 | { | 71 | { |
72 | return (current_cpu_data.processor_id & 0xffff00) == | 72 | return (current_cpu_data.processor_id & 0xffff00) == |
73 | (PRID_COMP_MIPS | PRID_IMP_4KC); | 73 | (PRID_COMP_MIPS | PRID_IMP_4KC); |
@@ -141,7 +141,7 @@ struct insn { | |||
141 | | (e) << RE_SH \ | 141 | | (e) << RE_SH \ |
142 | | (f) << FUNC_SH) | 142 | | (f) << FUNC_SH) |
143 | 143 | ||
144 | static __initdata struct insn insn_table[] = { | 144 | static struct insn insn_table[] __initdata = { |
145 | { insn_addiu, M(addiu_op, 0, 0, 0, 0, 0), RS | RT | SIMM }, | 145 | { insn_addiu, M(addiu_op, 0, 0, 0, 0, 0), RS | RT | SIMM }, |
146 | { insn_addu, M(spec_op, 0, 0, 0, 0, addu_op), RS | RT | RD }, | 146 | { insn_addu, M(spec_op, 0, 0, 0, 0, addu_op), RS | RT | RD }, |
147 | { insn_and, M(spec_op, 0, 0, 0, 0, and_op), RS | RT | RD }, | 147 | { insn_and, M(spec_op, 0, 0, 0, 0, and_op), RS | RT | RD }, |
@@ -194,7 +194,7 @@ static __initdata struct insn insn_table[] = { | |||
194 | 194 | ||
195 | #undef M | 195 | #undef M |
196 | 196 | ||
197 | static __init u32 build_rs(u32 arg) | 197 | static u32 __init build_rs(u32 arg) |
198 | { | 198 | { |
199 | if (arg & ~RS_MASK) | 199 | if (arg & ~RS_MASK) |
200 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 200 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
@@ -202,7 +202,7 @@ static __init u32 build_rs(u32 arg) | |||
202 | return (arg & RS_MASK) << RS_SH; | 202 | return (arg & RS_MASK) << RS_SH; |
203 | } | 203 | } |
204 | 204 | ||
205 | static __init u32 build_rt(u32 arg) | 205 | static u32 __init build_rt(u32 arg) |
206 | { | 206 | { |
207 | if (arg & ~RT_MASK) | 207 | if (arg & ~RT_MASK) |
208 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 208 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
@@ -210,7 +210,7 @@ static __init u32 build_rt(u32 arg) | |||
210 | return (arg & RT_MASK) << RT_SH; | 210 | return (arg & RT_MASK) << RT_SH; |
211 | } | 211 | } |
212 | 212 | ||
213 | static __init u32 build_rd(u32 arg) | 213 | static u32 __init build_rd(u32 arg) |
214 | { | 214 | { |
215 | if (arg & ~RD_MASK) | 215 | if (arg & ~RD_MASK) |
216 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 216 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
@@ -218,7 +218,7 @@ static __init u32 build_rd(u32 arg) | |||
218 | return (arg & RD_MASK) << RD_SH; | 218 | return (arg & RD_MASK) << RD_SH; |
219 | } | 219 | } |
220 | 220 | ||
221 | static __init u32 build_re(u32 arg) | 221 | static u32 __init build_re(u32 arg) |
222 | { | 222 | { |
223 | if (arg & ~RE_MASK) | 223 | if (arg & ~RE_MASK) |
224 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 224 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
@@ -226,7 +226,7 @@ static __init u32 build_re(u32 arg) | |||
226 | return (arg & RE_MASK) << RE_SH; | 226 | return (arg & RE_MASK) << RE_SH; |
227 | } | 227 | } |
228 | 228 | ||
229 | static __init u32 build_simm(s32 arg) | 229 | static u32 __init build_simm(s32 arg) |
230 | { | 230 | { |
231 | if (arg > 0x7fff || arg < -0x8000) | 231 | if (arg > 0x7fff || arg < -0x8000) |
232 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 232 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
@@ -234,7 +234,7 @@ static __init u32 build_simm(s32 arg) | |||
234 | return arg & 0xffff; | 234 | return arg & 0xffff; |
235 | } | 235 | } |
236 | 236 | ||
237 | static __init u32 build_uimm(u32 arg) | 237 | static u32 __init build_uimm(u32 arg) |
238 | { | 238 | { |
239 | if (arg & ~IMM_MASK) | 239 | if (arg & ~IMM_MASK) |
240 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 240 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
@@ -242,7 +242,7 @@ static __init u32 build_uimm(u32 arg) | |||
242 | return arg & IMM_MASK; | 242 | return arg & IMM_MASK; |
243 | } | 243 | } |
244 | 244 | ||
245 | static __init u32 build_bimm(s32 arg) | 245 | static u32 __init build_bimm(s32 arg) |
246 | { | 246 | { |
247 | if (arg > 0x1ffff || arg < -0x20000) | 247 | if (arg > 0x1ffff || arg < -0x20000) |
248 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 248 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
@@ -253,7 +253,7 @@ static __init u32 build_bimm(s32 arg) | |||
253 | return ((arg < 0) ? (1 << 15) : 0) | ((arg >> 2) & 0x7fff); | 253 | return ((arg < 0) ? (1 << 15) : 0) | ((arg >> 2) & 0x7fff); |
254 | } | 254 | } |
255 | 255 | ||
256 | static __init u32 build_jimm(u32 arg) | 256 | static u32 __init build_jimm(u32 arg) |
257 | { | 257 | { |
258 | if (arg & ~((JIMM_MASK) << 2)) | 258 | if (arg & ~((JIMM_MASK) << 2)) |
259 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 259 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
@@ -261,7 +261,7 @@ static __init u32 build_jimm(u32 arg) | |||
261 | return (arg >> 2) & JIMM_MASK; | 261 | return (arg >> 2) & JIMM_MASK; |
262 | } | 262 | } |
263 | 263 | ||
264 | static __init u32 build_func(u32 arg) | 264 | static u32 __init build_func(u32 arg) |
265 | { | 265 | { |
266 | if (arg & ~FUNC_MASK) | 266 | if (arg & ~FUNC_MASK) |
267 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 267 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
@@ -269,7 +269,7 @@ static __init u32 build_func(u32 arg) | |||
269 | return arg & FUNC_MASK; | 269 | return arg & FUNC_MASK; |
270 | } | 270 | } |
271 | 271 | ||
272 | static __init u32 build_set(u32 arg) | 272 | static u32 __init build_set(u32 arg) |
273 | { | 273 | { |
274 | if (arg & ~SET_MASK) | 274 | if (arg & ~SET_MASK) |
275 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 275 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
@@ -316,69 +316,69 @@ static void __init build_insn(u32 **buf, enum opcode opc, ...) | |||
316 | } | 316 | } |
317 | 317 | ||
318 | #define I_u1u2u3(op) \ | 318 | #define I_u1u2u3(op) \ |
319 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 319 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
320 | unsigned int b, unsigned int c) \ | 320 | unsigned int b, unsigned int c) \ |
321 | { \ | 321 | { \ |
322 | build_insn(buf, insn##op, a, b, c); \ | 322 | build_insn(buf, insn##op, a, b, c); \ |
323 | } | 323 | } |
324 | 324 | ||
325 | #define I_u2u1u3(op) \ | 325 | #define I_u2u1u3(op) \ |
326 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 326 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
327 | unsigned int b, unsigned int c) \ | 327 | unsigned int b, unsigned int c) \ |
328 | { \ | 328 | { \ |
329 | build_insn(buf, insn##op, b, a, c); \ | 329 | build_insn(buf, insn##op, b, a, c); \ |
330 | } | 330 | } |
331 | 331 | ||
332 | #define I_u3u1u2(op) \ | 332 | #define I_u3u1u2(op) \ |
333 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 333 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
334 | unsigned int b, unsigned int c) \ | 334 | unsigned int b, unsigned int c) \ |
335 | { \ | 335 | { \ |
336 | build_insn(buf, insn##op, b, c, a); \ | 336 | build_insn(buf, insn##op, b, c, a); \ |
337 | } | 337 | } |
338 | 338 | ||
339 | #define I_u1u2s3(op) \ | 339 | #define I_u1u2s3(op) \ |
340 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 340 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
341 | unsigned int b, signed int c) \ | 341 | unsigned int b, signed int c) \ |
342 | { \ | 342 | { \ |
343 | build_insn(buf, insn##op, a, b, c); \ | 343 | build_insn(buf, insn##op, a, b, c); \ |
344 | } | 344 | } |
345 | 345 | ||
346 | #define I_u2s3u1(op) \ | 346 | #define I_u2s3u1(op) \ |
347 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 347 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
348 | signed int b, unsigned int c) \ | 348 | signed int b, unsigned int c) \ |
349 | { \ | 349 | { \ |
350 | build_insn(buf, insn##op, c, a, b); \ | 350 | build_insn(buf, insn##op, c, a, b); \ |
351 | } | 351 | } |
352 | 352 | ||
353 | #define I_u2u1s3(op) \ | 353 | #define I_u2u1s3(op) \ |
354 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 354 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
355 | unsigned int b, signed int c) \ | 355 | unsigned int b, signed int c) \ |
356 | { \ | 356 | { \ |
357 | build_insn(buf, insn##op, b, a, c); \ | 357 | build_insn(buf, insn##op, b, a, c); \ |
358 | } | 358 | } |
359 | 359 | ||
360 | #define I_u1u2(op) \ | 360 | #define I_u1u2(op) \ |
361 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 361 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
362 | unsigned int b) \ | 362 | unsigned int b) \ |
363 | { \ | 363 | { \ |
364 | build_insn(buf, insn##op, a, b); \ | 364 | build_insn(buf, insn##op, a, b); \ |
365 | } | 365 | } |
366 | 366 | ||
367 | #define I_u1s2(op) \ | 367 | #define I_u1s2(op) \ |
368 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 368 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
369 | signed int b) \ | 369 | signed int b) \ |
370 | { \ | 370 | { \ |
371 | build_insn(buf, insn##op, a, b); \ | 371 | build_insn(buf, insn##op, a, b); \ |
372 | } | 372 | } |
373 | 373 | ||
374 | #define I_u1(op) \ | 374 | #define I_u1(op) \ |
375 | static inline void __init i##op(u32 **buf, unsigned int a) \ | 375 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a) \ |
376 | { \ | 376 | { \ |
377 | build_insn(buf, insn##op, a); \ | 377 | build_insn(buf, insn##op, a); \ |
378 | } | 378 | } |
379 | 379 | ||
380 | #define I_0(op) \ | 380 | #define I_0(op) \ |
381 | static inline void __init i##op(u32 **buf) \ | 381 | static void __init __maybe_unused i##op(u32 **buf) \ |
382 | { \ | 382 | { \ |
383 | build_insn(buf, insn##op); \ | 383 | build_insn(buf, insn##op); \ |
384 | } | 384 | } |
@@ -458,7 +458,7 @@ struct label { | |||
458 | enum label_id lab; | 458 | enum label_id lab; |
459 | }; | 459 | }; |
460 | 460 | ||
461 | static __init void build_label(struct label **lab, u32 *addr, | 461 | static void __init build_label(struct label **lab, u32 *addr, |
462 | enum label_id l) | 462 | enum label_id l) |
463 | { | 463 | { |
464 | (*lab)->addr = addr; | 464 | (*lab)->addr = addr; |
@@ -467,7 +467,7 @@ static __init void build_label(struct label **lab, u32 *addr, | |||
467 | } | 467 | } |
468 | 468 | ||
469 | #define L_LA(lb) \ | 469 | #define L_LA(lb) \ |
470 | static inline void l##lb(struct label **lab, u32 *addr) \ | 470 | static inline void __init l##lb(struct label **lab, u32 *addr) \ |
471 | { \ | 471 | { \ |
472 | build_label(lab, addr, label##lb); \ | 472 | build_label(lab, addr, label##lb); \ |
473 | } | 473 | } |
@@ -526,7 +526,7 @@ L_LA(_r3000_write_probe_fail) | |||
526 | #define i_ssnop(buf) i_sll(buf, 0, 0, 1) | 526 | #define i_ssnop(buf) i_sll(buf, 0, 0, 1) |
527 | #define i_ehb(buf) i_sll(buf, 0, 0, 3) | 527 | #define i_ehb(buf) i_sll(buf, 0, 0, 3) |
528 | 528 | ||
529 | static __init int __maybe_unused in_compat_space_p(long addr) | 529 | static int __init __maybe_unused in_compat_space_p(long addr) |
530 | { | 530 | { |
531 | /* Is this address in 32bit compat space? */ | 531 | /* Is this address in 32bit compat space? */ |
532 | #ifdef CONFIG_64BIT | 532 | #ifdef CONFIG_64BIT |
@@ -536,7 +536,7 @@ static __init int __maybe_unused in_compat_space_p(long addr) | |||
536 | #endif | 536 | #endif |
537 | } | 537 | } |
538 | 538 | ||
539 | static __init int __maybe_unused rel_highest(long val) | 539 | static int __init __maybe_unused rel_highest(long val) |
540 | { | 540 | { |
541 | #ifdef CONFIG_64BIT | 541 | #ifdef CONFIG_64BIT |
542 | return ((((val + 0x800080008000L) >> 48) & 0xffff) ^ 0x8000) - 0x8000; | 542 | return ((((val + 0x800080008000L) >> 48) & 0xffff) ^ 0x8000) - 0x8000; |
@@ -545,7 +545,7 @@ static __init int __maybe_unused rel_highest(long val) | |||
545 | #endif | 545 | #endif |
546 | } | 546 | } |
547 | 547 | ||
548 | static __init int __maybe_unused rel_higher(long val) | 548 | static int __init __maybe_unused rel_higher(long val) |
549 | { | 549 | { |
550 | #ifdef CONFIG_64BIT | 550 | #ifdef CONFIG_64BIT |
551 | return ((((val + 0x80008000L) >> 32) & 0xffff) ^ 0x8000) - 0x8000; | 551 | return ((((val + 0x80008000L) >> 32) & 0xffff) ^ 0x8000) - 0x8000; |
@@ -554,17 +554,17 @@ static __init int __maybe_unused rel_higher(long val) | |||
554 | #endif | 554 | #endif |
555 | } | 555 | } |
556 | 556 | ||
557 | static __init int rel_hi(long val) | 557 | static int __init rel_hi(long val) |
558 | { | 558 | { |
559 | return ((((val + 0x8000L) >> 16) & 0xffff) ^ 0x8000) - 0x8000; | 559 | return ((((val + 0x8000L) >> 16) & 0xffff) ^ 0x8000) - 0x8000; |
560 | } | 560 | } |
561 | 561 | ||
562 | static __init int rel_lo(long val) | 562 | static int __init rel_lo(long val) |
563 | { | 563 | { |
564 | return ((val & 0xffff) ^ 0x8000) - 0x8000; | 564 | return ((val & 0xffff) ^ 0x8000) - 0x8000; |
565 | } | 565 | } |
566 | 566 | ||
567 | static __init void i_LA_mostly(u32 **buf, unsigned int rs, long addr) | 567 | static void __init i_LA_mostly(u32 **buf, unsigned int rs, long addr) |
568 | { | 568 | { |
569 | if (!in_compat_space_p(addr)) { | 569 | if (!in_compat_space_p(addr)) { |
570 | i_lui(buf, rs, rel_highest(addr)); | 570 | i_lui(buf, rs, rel_highest(addr)); |
@@ -580,7 +580,7 @@ static __init void i_LA_mostly(u32 **buf, unsigned int rs, long addr) | |||
580 | i_lui(buf, rs, rel_hi(addr)); | 580 | i_lui(buf, rs, rel_hi(addr)); |
581 | } | 581 | } |
582 | 582 | ||
583 | static __init void __maybe_unused i_LA(u32 **buf, unsigned int rs, long addr) | 583 | static void __init __maybe_unused i_LA(u32 **buf, unsigned int rs, long addr) |
584 | { | 584 | { |
585 | i_LA_mostly(buf, rs, addr); | 585 | i_LA_mostly(buf, rs, addr); |
586 | if (rel_lo(addr)) { | 586 | if (rel_lo(addr)) { |
@@ -601,7 +601,7 @@ struct reloc { | |||
601 | enum label_id lab; | 601 | enum label_id lab; |
602 | }; | 602 | }; |
603 | 603 | ||
604 | static __init void r_mips_pc16(struct reloc **rel, u32 *addr, | 604 | static void __init r_mips_pc16(struct reloc **rel, u32 *addr, |
605 | enum label_id l) | 605 | enum label_id l) |
606 | { | 606 | { |
607 | (*rel)->addr = addr; | 607 | (*rel)->addr = addr; |
@@ -626,7 +626,7 @@ static inline void __resolve_relocs(struct reloc *rel, struct label *lab) | |||
626 | } | 626 | } |
627 | } | 627 | } |
628 | 628 | ||
629 | static __init void resolve_relocs(struct reloc *rel, struct label *lab) | 629 | static void __init resolve_relocs(struct reloc *rel, struct label *lab) |
630 | { | 630 | { |
631 | struct label *l; | 631 | struct label *l; |
632 | 632 | ||
@@ -636,7 +636,7 @@ static __init void resolve_relocs(struct reloc *rel, struct label *lab) | |||
636 | __resolve_relocs(rel, l); | 636 | __resolve_relocs(rel, l); |
637 | } | 637 | } |
638 | 638 | ||
639 | static __init void move_relocs(struct reloc *rel, u32 *first, u32 *end, | 639 | static void __init move_relocs(struct reloc *rel, u32 *first, u32 *end, |
640 | long off) | 640 | long off) |
641 | { | 641 | { |
642 | for (; rel->lab != label_invalid; rel++) | 642 | for (; rel->lab != label_invalid; rel++) |
@@ -644,7 +644,7 @@ static __init void move_relocs(struct reloc *rel, u32 *first, u32 *end, | |||
644 | rel->addr += off; | 644 | rel->addr += off; |
645 | } | 645 | } |
646 | 646 | ||
647 | static __init void move_labels(struct label *lab, u32 *first, u32 *end, | 647 | static void __init move_labels(struct label *lab, u32 *first, u32 *end, |
648 | long off) | 648 | long off) |
649 | { | 649 | { |
650 | for (; lab->lab != label_invalid; lab++) | 650 | for (; lab->lab != label_invalid; lab++) |
@@ -652,7 +652,7 @@ static __init void move_labels(struct label *lab, u32 *first, u32 *end, | |||
652 | lab->addr += off; | 652 | lab->addr += off; |
653 | } | 653 | } |
654 | 654 | ||
655 | static __init void copy_handler(struct reloc *rel, struct label *lab, | 655 | static void __init copy_handler(struct reloc *rel, struct label *lab, |
656 | u32 *first, u32 *end, u32 *target) | 656 | u32 *first, u32 *end, u32 *target) |
657 | { | 657 | { |
658 | long off = (long)(target - first); | 658 | long off = (long)(target - first); |
@@ -663,7 +663,7 @@ static __init void copy_handler(struct reloc *rel, struct label *lab, | |||
663 | move_labels(lab, first, end, off); | 663 | move_labels(lab, first, end, off); |
664 | } | 664 | } |
665 | 665 | ||
666 | static __init int __maybe_unused insn_has_bdelay(struct reloc *rel, | 666 | static int __init __maybe_unused insn_has_bdelay(struct reloc *rel, |
667 | u32 *addr) | 667 | u32 *addr) |
668 | { | 668 | { |
669 | for (; rel->lab != label_invalid; rel++) { | 669 | for (; rel->lab != label_invalid; rel++) { |
@@ -755,11 +755,11 @@ il_bgez(u32 **p, struct reloc **r, unsigned int reg, enum label_id l) | |||
755 | * We deliberately chose a buffer size of 128, so we won't scribble | 755 | * We deliberately chose a buffer size of 128, so we won't scribble |
756 | * over anything important on overflow before we panic. | 756 | * over anything important on overflow before we panic. |
757 | */ | 757 | */ |
758 | static __initdata u32 tlb_handler[128]; | 758 | static u32 tlb_handler[128] __initdata; |
759 | 759 | ||
760 | /* simply assume worst case size for labels and relocs */ | 760 | /* simply assume worst case size for labels and relocs */ |
761 | static __initdata struct label labels[128]; | 761 | static struct label labels[128] __initdata; |
762 | static __initdata struct reloc relocs[128]; | 762 | static struct reloc relocs[128] __initdata; |
763 | 763 | ||
764 | /* | 764 | /* |
765 | * The R3000 TLB handler is simple. | 765 | * The R3000 TLB handler is simple. |
@@ -813,7 +813,7 @@ static void __init build_r3000_tlb_refill_handler(void) | |||
813 | * other one.To keep things simple, we first assume linear space, | 813 | * other one.To keep things simple, we first assume linear space, |
814 | * then we relocate it to the final handler layout as needed. | 814 | * then we relocate it to the final handler layout as needed. |
815 | */ | 815 | */ |
816 | static __initdata u32 final_handler[64]; | 816 | static u32 final_handler[64] __initdata; |
817 | 817 | ||
818 | /* | 818 | /* |
819 | * Hazards | 819 | * Hazards |
@@ -837,7 +837,7 @@ static __initdata u32 final_handler[64]; | |||
837 | * | 837 | * |
838 | * As if we MIPS hackers wouldn't know how to nop pipelines happy ... | 838 | * As if we MIPS hackers wouldn't know how to nop pipelines happy ... |
839 | */ | 839 | */ |
840 | static __init void __maybe_unused build_tlb_probe_entry(u32 **p) | 840 | static void __init __maybe_unused build_tlb_probe_entry(u32 **p) |
841 | { | 841 | { |
842 | switch (current_cpu_type()) { | 842 | switch (current_cpu_type()) { |
843 | /* Found by experiment: R4600 v2.0 needs this, too. */ | 843 | /* Found by experiment: R4600 v2.0 needs this, too. */ |
@@ -861,7 +861,7 @@ static __init void __maybe_unused build_tlb_probe_entry(u32 **p) | |||
861 | */ | 861 | */ |
862 | enum tlb_write_entry { tlb_random, tlb_indexed }; | 862 | enum tlb_write_entry { tlb_random, tlb_indexed }; |
863 | 863 | ||
864 | static __init void build_tlb_write_entry(u32 **p, struct label **l, | 864 | static void __init build_tlb_write_entry(u32 **p, struct label **l, |
865 | struct reloc **r, | 865 | struct reloc **r, |
866 | enum tlb_write_entry wmode) | 866 | enum tlb_write_entry wmode) |
867 | { | 867 | { |
@@ -1003,7 +1003,7 @@ static __init void build_tlb_write_entry(u32 **p, struct label **l, | |||
1003 | * TMP and PTR are scratch. | 1003 | * TMP and PTR are scratch. |
1004 | * TMP will be clobbered, PTR will hold the pmd entry. | 1004 | * TMP will be clobbered, PTR will hold the pmd entry. |
1005 | */ | 1005 | */ |
1006 | static __init void | 1006 | static void __init |
1007 | build_get_pmde64(u32 **p, struct label **l, struct reloc **r, | 1007 | build_get_pmde64(u32 **p, struct label **l, struct reloc **r, |
1008 | unsigned int tmp, unsigned int ptr) | 1008 | unsigned int tmp, unsigned int ptr) |
1009 | { | 1009 | { |
@@ -1064,7 +1064,7 @@ build_get_pmde64(u32 **p, struct label **l, struct reloc **r, | |||
1064 | * BVADDR is the faulting address, PTR is scratch. | 1064 | * BVADDR is the faulting address, PTR is scratch. |
1065 | * PTR will hold the pgd for vmalloc. | 1065 | * PTR will hold the pgd for vmalloc. |
1066 | */ | 1066 | */ |
1067 | static __init void | 1067 | static void __init |
1068 | build_get_pgd_vmalloc64(u32 **p, struct label **l, struct reloc **r, | 1068 | build_get_pgd_vmalloc64(u32 **p, struct label **l, struct reloc **r, |
1069 | unsigned int bvaddr, unsigned int ptr) | 1069 | unsigned int bvaddr, unsigned int ptr) |
1070 | { | 1070 | { |
@@ -1134,7 +1134,7 @@ build_get_pgd_vmalloc64(u32 **p, struct label **l, struct reloc **r, | |||
1134 | * TMP and PTR are scratch. | 1134 | * TMP and PTR are scratch. |
1135 | * TMP will be clobbered, PTR will hold the pgd entry. | 1135 | * TMP will be clobbered, PTR will hold the pgd entry. |
1136 | */ | 1136 | */ |
1137 | static __init void __maybe_unused | 1137 | static void __init __maybe_unused |
1138 | build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr) | 1138 | build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr) |
1139 | { | 1139 | { |
1140 | long pgdc = (long)pgd_current; | 1140 | long pgdc = (long)pgd_current; |
@@ -1169,7 +1169,7 @@ build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr) | |||
1169 | 1169 | ||
1170 | #endif /* !CONFIG_64BIT */ | 1170 | #endif /* !CONFIG_64BIT */ |
1171 | 1171 | ||
1172 | static __init void build_adjust_context(u32 **p, unsigned int ctx) | 1172 | static void __init build_adjust_context(u32 **p, unsigned int ctx) |
1173 | { | 1173 | { |
1174 | unsigned int shift = 4 - (PTE_T_LOG2 + 1) + PAGE_SHIFT - 12; | 1174 | unsigned int shift = 4 - (PTE_T_LOG2 + 1) + PAGE_SHIFT - 12; |
1175 | unsigned int mask = (PTRS_PER_PTE / 2 - 1) << (PTE_T_LOG2 + 1); | 1175 | unsigned int mask = (PTRS_PER_PTE / 2 - 1) << (PTE_T_LOG2 + 1); |
@@ -1195,7 +1195,7 @@ static __init void build_adjust_context(u32 **p, unsigned int ctx) | |||
1195 | i_andi(p, ctx, ctx, mask); | 1195 | i_andi(p, ctx, ctx, mask); |
1196 | } | 1196 | } |
1197 | 1197 | ||
1198 | static __init void build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr) | 1198 | static void __init build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr) |
1199 | { | 1199 | { |
1200 | /* | 1200 | /* |
1201 | * Bug workaround for the Nevada. It seems as if under certain | 1201 | * Bug workaround for the Nevada. It seems as if under certain |
@@ -1220,7 +1220,7 @@ static __init void build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr) | |||
1220 | i_ADDU(p, ptr, ptr, tmp); /* add in offset */ | 1220 | i_ADDU(p, ptr, ptr, tmp); /* add in offset */ |
1221 | } | 1221 | } |
1222 | 1222 | ||
1223 | static __init void build_update_entries(u32 **p, unsigned int tmp, | 1223 | static void __init build_update_entries(u32 **p, unsigned int tmp, |
1224 | unsigned int ptep) | 1224 | unsigned int ptep) |
1225 | { | 1225 | { |
1226 | /* | 1226 | /* |