aboutsummaryrefslogtreecommitdiffstats
path: root/arch/mips/mm
diff options
context:
space:
mode:
authorFranck Bui-Huu <fbuihuu@gmail.com>2007-10-18 03:11:14 -0400
committerRalf Baechle <ralf@linux-mips.org>2008-01-29 05:14:56 -0500
commita9af6041e9e7ed0278d03333c3a22a8509f33615 (patch)
tree1d05905c82e0a36c5c4c5d04c1933224fb82ae92 /arch/mips/mm
parent3f4ac2dc0504962b6cb9367adbc13272393b80b3 (diff)
[MIPS] tlbex.c: Cleanup __init usages.
Signed-off-by: Franck Bui-Huu <fbuihuu@gmail.com> Signed-off-by: Ralf Baechle <ralf@linux-mips.org>
Diffstat (limited to 'arch/mips/mm')
-rw-r--r--arch/mips/mm/tlbex.c98
1 files changed, 49 insertions, 49 deletions
diff --git a/arch/mips/mm/tlbex.c b/arch/mips/mm/tlbex.c
index f8925ba0b39e..77530fd96f1c 100644
--- a/arch/mips/mm/tlbex.c
+++ b/arch/mips/mm/tlbex.c
@@ -67,7 +67,7 @@ static inline int __maybe_unused r10000_llsc_war(void)
67 * why; it's not an issue caused by the core RTL. 67 * why; it's not an issue caused by the core RTL.
68 * 68 *
69 */ 69 */
70static __init int __attribute__((unused)) m4kc_tlbp_war(void) 70static int __init m4kc_tlbp_war(void)
71{ 71{
72 return (current_cpu_data.processor_id & 0xffff00) == 72 return (current_cpu_data.processor_id & 0xffff00) ==
73 (PRID_COMP_MIPS | PRID_IMP_4KC); 73 (PRID_COMP_MIPS | PRID_IMP_4KC);
@@ -141,7 +141,7 @@ struct insn {
141 | (e) << RE_SH \ 141 | (e) << RE_SH \
142 | (f) << FUNC_SH) 142 | (f) << FUNC_SH)
143 143
144static __initdata struct insn insn_table[] = { 144static struct insn insn_table[] __initdata = {
145 { insn_addiu, M(addiu_op, 0, 0, 0, 0, 0), RS | RT | SIMM }, 145 { insn_addiu, M(addiu_op, 0, 0, 0, 0, 0), RS | RT | SIMM },
146 { insn_addu, M(spec_op, 0, 0, 0, 0, addu_op), RS | RT | RD }, 146 { insn_addu, M(spec_op, 0, 0, 0, 0, addu_op), RS | RT | RD },
147 { insn_and, M(spec_op, 0, 0, 0, 0, and_op), RS | RT | RD }, 147 { insn_and, M(spec_op, 0, 0, 0, 0, and_op), RS | RT | RD },
@@ -194,7 +194,7 @@ static __initdata struct insn insn_table[] = {
194 194
195#undef M 195#undef M
196 196
197static __init u32 build_rs(u32 arg) 197static u32 __init build_rs(u32 arg)
198{ 198{
199 if (arg & ~RS_MASK) 199 if (arg & ~RS_MASK)
200 printk(KERN_WARNING "TLB synthesizer field overflow\n"); 200 printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -202,7 +202,7 @@ static __init u32 build_rs(u32 arg)
202 return (arg & RS_MASK) << RS_SH; 202 return (arg & RS_MASK) << RS_SH;
203} 203}
204 204
205static __init u32 build_rt(u32 arg) 205static u32 __init build_rt(u32 arg)
206{ 206{
207 if (arg & ~RT_MASK) 207 if (arg & ~RT_MASK)
208 printk(KERN_WARNING "TLB synthesizer field overflow\n"); 208 printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -210,7 +210,7 @@ static __init u32 build_rt(u32 arg)
210 return (arg & RT_MASK) << RT_SH; 210 return (arg & RT_MASK) << RT_SH;
211} 211}
212 212
213static __init u32 build_rd(u32 arg) 213static u32 __init build_rd(u32 arg)
214{ 214{
215 if (arg & ~RD_MASK) 215 if (arg & ~RD_MASK)
216 printk(KERN_WARNING "TLB synthesizer field overflow\n"); 216 printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -218,7 +218,7 @@ static __init u32 build_rd(u32 arg)
218 return (arg & RD_MASK) << RD_SH; 218 return (arg & RD_MASK) << RD_SH;
219} 219}
220 220
221static __init u32 build_re(u32 arg) 221static u32 __init build_re(u32 arg)
222{ 222{
223 if (arg & ~RE_MASK) 223 if (arg & ~RE_MASK)
224 printk(KERN_WARNING "TLB synthesizer field overflow\n"); 224 printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -226,7 +226,7 @@ static __init u32 build_re(u32 arg)
226 return (arg & RE_MASK) << RE_SH; 226 return (arg & RE_MASK) << RE_SH;
227} 227}
228 228
229static __init u32 build_simm(s32 arg) 229static u32 __init build_simm(s32 arg)
230{ 230{
231 if (arg > 0x7fff || arg < -0x8000) 231 if (arg > 0x7fff || arg < -0x8000)
232 printk(KERN_WARNING "TLB synthesizer field overflow\n"); 232 printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -234,7 +234,7 @@ static __init u32 build_simm(s32 arg)
234 return arg & 0xffff; 234 return arg & 0xffff;
235} 235}
236 236
237static __init u32 build_uimm(u32 arg) 237static u32 __init build_uimm(u32 arg)
238{ 238{
239 if (arg & ~IMM_MASK) 239 if (arg & ~IMM_MASK)
240 printk(KERN_WARNING "TLB synthesizer field overflow\n"); 240 printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -242,7 +242,7 @@ static __init u32 build_uimm(u32 arg)
242 return arg & IMM_MASK; 242 return arg & IMM_MASK;
243} 243}
244 244
245static __init u32 build_bimm(s32 arg) 245static u32 __init build_bimm(s32 arg)
246{ 246{
247 if (arg > 0x1ffff || arg < -0x20000) 247 if (arg > 0x1ffff || arg < -0x20000)
248 printk(KERN_WARNING "TLB synthesizer field overflow\n"); 248 printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -253,7 +253,7 @@ static __init u32 build_bimm(s32 arg)
253 return ((arg < 0) ? (1 << 15) : 0) | ((arg >> 2) & 0x7fff); 253 return ((arg < 0) ? (1 << 15) : 0) | ((arg >> 2) & 0x7fff);
254} 254}
255 255
256static __init u32 build_jimm(u32 arg) 256static u32 __init build_jimm(u32 arg)
257{ 257{
258 if (arg & ~((JIMM_MASK) << 2)) 258 if (arg & ~((JIMM_MASK) << 2))
259 printk(KERN_WARNING "TLB synthesizer field overflow\n"); 259 printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -261,7 +261,7 @@ static __init u32 build_jimm(u32 arg)
261 return (arg >> 2) & JIMM_MASK; 261 return (arg >> 2) & JIMM_MASK;
262} 262}
263 263
264static __init u32 build_func(u32 arg) 264static u32 __init build_func(u32 arg)
265{ 265{
266 if (arg & ~FUNC_MASK) 266 if (arg & ~FUNC_MASK)
267 printk(KERN_WARNING "TLB synthesizer field overflow\n"); 267 printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -269,7 +269,7 @@ static __init u32 build_func(u32 arg)
269 return arg & FUNC_MASK; 269 return arg & FUNC_MASK;
270} 270}
271 271
272static __init u32 build_set(u32 arg) 272static u32 __init build_set(u32 arg)
273{ 273{
274 if (arg & ~SET_MASK) 274 if (arg & ~SET_MASK)
275 printk(KERN_WARNING "TLB synthesizer field overflow\n"); 275 printk(KERN_WARNING "TLB synthesizer field overflow\n");
@@ -316,69 +316,69 @@ static void __init build_insn(u32 **buf, enum opcode opc, ...)
316} 316}
317 317
318#define I_u1u2u3(op) \ 318#define I_u1u2u3(op) \
319 static inline void __init i##op(u32 **buf, unsigned int a, \ 319 static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \
320 unsigned int b, unsigned int c) \ 320 unsigned int b, unsigned int c) \
321 { \ 321 { \
322 build_insn(buf, insn##op, a, b, c); \ 322 build_insn(buf, insn##op, a, b, c); \
323 } 323 }
324 324
325#define I_u2u1u3(op) \ 325#define I_u2u1u3(op) \
326 static inline void __init i##op(u32 **buf, unsigned int a, \ 326 static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \
327 unsigned int b, unsigned int c) \ 327 unsigned int b, unsigned int c) \
328 { \ 328 { \
329 build_insn(buf, insn##op, b, a, c); \ 329 build_insn(buf, insn##op, b, a, c); \
330 } 330 }
331 331
332#define I_u3u1u2(op) \ 332#define I_u3u1u2(op) \
333 static inline void __init i##op(u32 **buf, unsigned int a, \ 333 static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \
334 unsigned int b, unsigned int c) \ 334 unsigned int b, unsigned int c) \
335 { \ 335 { \
336 build_insn(buf, insn##op, b, c, a); \ 336 build_insn(buf, insn##op, b, c, a); \
337 } 337 }
338 338
339#define I_u1u2s3(op) \ 339#define I_u1u2s3(op) \
340 static inline void __init i##op(u32 **buf, unsigned int a, \ 340 static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \
341 unsigned int b, signed int c) \ 341 unsigned int b, signed int c) \
342 { \ 342 { \
343 build_insn(buf, insn##op, a, b, c); \ 343 build_insn(buf, insn##op, a, b, c); \
344 } 344 }
345 345
346#define I_u2s3u1(op) \ 346#define I_u2s3u1(op) \
347 static inline void __init i##op(u32 **buf, unsigned int a, \ 347 static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \
348 signed int b, unsigned int c) \ 348 signed int b, unsigned int c) \
349 { \ 349 { \
350 build_insn(buf, insn##op, c, a, b); \ 350 build_insn(buf, insn##op, c, a, b); \
351 } 351 }
352 352
353#define I_u2u1s3(op) \ 353#define I_u2u1s3(op) \
354 static inline void __init i##op(u32 **buf, unsigned int a, \ 354 static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \
355 unsigned int b, signed int c) \ 355 unsigned int b, signed int c) \
356 { \ 356 { \
357 build_insn(buf, insn##op, b, a, c); \ 357 build_insn(buf, insn##op, b, a, c); \
358 } 358 }
359 359
360#define I_u1u2(op) \ 360#define I_u1u2(op) \
361 static inline void __init i##op(u32 **buf, unsigned int a, \ 361 static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \
362 unsigned int b) \ 362 unsigned int b) \
363 { \ 363 { \
364 build_insn(buf, insn##op, a, b); \ 364 build_insn(buf, insn##op, a, b); \
365 } 365 }
366 366
367#define I_u1s2(op) \ 367#define I_u1s2(op) \
368 static inline void __init i##op(u32 **buf, unsigned int a, \ 368 static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \
369 signed int b) \ 369 signed int b) \
370 { \ 370 { \
371 build_insn(buf, insn##op, a, b); \ 371 build_insn(buf, insn##op, a, b); \
372 } 372 }
373 373
374#define I_u1(op) \ 374#define I_u1(op) \
375 static inline void __init i##op(u32 **buf, unsigned int a) \ 375 static void __init __maybe_unused i##op(u32 **buf, unsigned int a) \
376 { \ 376 { \
377 build_insn(buf, insn##op, a); \ 377 build_insn(buf, insn##op, a); \
378 } 378 }
379 379
380#define I_0(op) \ 380#define I_0(op) \
381 static inline void __init i##op(u32 **buf) \ 381 static void __init __maybe_unused i##op(u32 **buf) \
382 { \ 382 { \
383 build_insn(buf, insn##op); \ 383 build_insn(buf, insn##op); \
384 } 384 }
@@ -458,7 +458,7 @@ struct label {
458 enum label_id lab; 458 enum label_id lab;
459}; 459};
460 460
461static __init void build_label(struct label **lab, u32 *addr, 461static void __init build_label(struct label **lab, u32 *addr,
462 enum label_id l) 462 enum label_id l)
463{ 463{
464 (*lab)->addr = addr; 464 (*lab)->addr = addr;
@@ -467,7 +467,7 @@ static __init void build_label(struct label **lab, u32 *addr,
467} 467}
468 468
469#define L_LA(lb) \ 469#define L_LA(lb) \
470 static inline void l##lb(struct label **lab, u32 *addr) \ 470 static inline void __init l##lb(struct label **lab, u32 *addr) \
471 { \ 471 { \
472 build_label(lab, addr, label##lb); \ 472 build_label(lab, addr, label##lb); \
473 } 473 }
@@ -526,7 +526,7 @@ L_LA(_r3000_write_probe_fail)
526#define i_ssnop(buf) i_sll(buf, 0, 0, 1) 526#define i_ssnop(buf) i_sll(buf, 0, 0, 1)
527#define i_ehb(buf) i_sll(buf, 0, 0, 3) 527#define i_ehb(buf) i_sll(buf, 0, 0, 3)
528 528
529static __init int __maybe_unused in_compat_space_p(long addr) 529static int __init __maybe_unused in_compat_space_p(long addr)
530{ 530{
531 /* Is this address in 32bit compat space? */ 531 /* Is this address in 32bit compat space? */
532#ifdef CONFIG_64BIT 532#ifdef CONFIG_64BIT
@@ -536,7 +536,7 @@ static __init int __maybe_unused in_compat_space_p(long addr)
536#endif 536#endif
537} 537}
538 538
539static __init int __maybe_unused rel_highest(long val) 539static int __init __maybe_unused rel_highest(long val)
540{ 540{
541#ifdef CONFIG_64BIT 541#ifdef CONFIG_64BIT
542 return ((((val + 0x800080008000L) >> 48) & 0xffff) ^ 0x8000) - 0x8000; 542 return ((((val + 0x800080008000L) >> 48) & 0xffff) ^ 0x8000) - 0x8000;
@@ -545,7 +545,7 @@ static __init int __maybe_unused rel_highest(long val)
545#endif 545#endif
546} 546}
547 547
548static __init int __maybe_unused rel_higher(long val) 548static int __init __maybe_unused rel_higher(long val)
549{ 549{
550#ifdef CONFIG_64BIT 550#ifdef CONFIG_64BIT
551 return ((((val + 0x80008000L) >> 32) & 0xffff) ^ 0x8000) - 0x8000; 551 return ((((val + 0x80008000L) >> 32) & 0xffff) ^ 0x8000) - 0x8000;
@@ -554,17 +554,17 @@ static __init int __maybe_unused rel_higher(long val)
554#endif 554#endif
555} 555}
556 556
557static __init int rel_hi(long val) 557static int __init rel_hi(long val)
558{ 558{
559 return ((((val + 0x8000L) >> 16) & 0xffff) ^ 0x8000) - 0x8000; 559 return ((((val + 0x8000L) >> 16) & 0xffff) ^ 0x8000) - 0x8000;
560} 560}
561 561
562static __init int rel_lo(long val) 562static int __init rel_lo(long val)
563{ 563{
564 return ((val & 0xffff) ^ 0x8000) - 0x8000; 564 return ((val & 0xffff) ^ 0x8000) - 0x8000;
565} 565}
566 566
567static __init void i_LA_mostly(u32 **buf, unsigned int rs, long addr) 567static void __init i_LA_mostly(u32 **buf, unsigned int rs, long addr)
568{ 568{
569 if (!in_compat_space_p(addr)) { 569 if (!in_compat_space_p(addr)) {
570 i_lui(buf, rs, rel_highest(addr)); 570 i_lui(buf, rs, rel_highest(addr));
@@ -580,7 +580,7 @@ static __init void i_LA_mostly(u32 **buf, unsigned int rs, long addr)
580 i_lui(buf, rs, rel_hi(addr)); 580 i_lui(buf, rs, rel_hi(addr));
581} 581}
582 582
583static __init void __maybe_unused i_LA(u32 **buf, unsigned int rs, long addr) 583static void __init __maybe_unused i_LA(u32 **buf, unsigned int rs, long addr)
584{ 584{
585 i_LA_mostly(buf, rs, addr); 585 i_LA_mostly(buf, rs, addr);
586 if (rel_lo(addr)) { 586 if (rel_lo(addr)) {
@@ -601,7 +601,7 @@ struct reloc {
601 enum label_id lab; 601 enum label_id lab;
602}; 602};
603 603
604static __init void r_mips_pc16(struct reloc **rel, u32 *addr, 604static void __init r_mips_pc16(struct reloc **rel, u32 *addr,
605 enum label_id l) 605 enum label_id l)
606{ 606{
607 (*rel)->addr = addr; 607 (*rel)->addr = addr;
@@ -626,7 +626,7 @@ static inline void __resolve_relocs(struct reloc *rel, struct label *lab)
626 } 626 }
627} 627}
628 628
629static __init void resolve_relocs(struct reloc *rel, struct label *lab) 629static void __init resolve_relocs(struct reloc *rel, struct label *lab)
630{ 630{
631 struct label *l; 631 struct label *l;
632 632
@@ -636,7 +636,7 @@ static __init void resolve_relocs(struct reloc *rel, struct label *lab)
636 __resolve_relocs(rel, l); 636 __resolve_relocs(rel, l);
637} 637}
638 638
639static __init void move_relocs(struct reloc *rel, u32 *first, u32 *end, 639static void __init move_relocs(struct reloc *rel, u32 *first, u32 *end,
640 long off) 640 long off)
641{ 641{
642 for (; rel->lab != label_invalid; rel++) 642 for (; rel->lab != label_invalid; rel++)
@@ -644,7 +644,7 @@ static __init void move_relocs(struct reloc *rel, u32 *first, u32 *end,
644 rel->addr += off; 644 rel->addr += off;
645} 645}
646 646
647static __init void move_labels(struct label *lab, u32 *first, u32 *end, 647static void __init move_labels(struct label *lab, u32 *first, u32 *end,
648 long off) 648 long off)
649{ 649{
650 for (; lab->lab != label_invalid; lab++) 650 for (; lab->lab != label_invalid; lab++)
@@ -652,7 +652,7 @@ static __init void move_labels(struct label *lab, u32 *first, u32 *end,
652 lab->addr += off; 652 lab->addr += off;
653} 653}
654 654
655static __init void copy_handler(struct reloc *rel, struct label *lab, 655static void __init copy_handler(struct reloc *rel, struct label *lab,
656 u32 *first, u32 *end, u32 *target) 656 u32 *first, u32 *end, u32 *target)
657{ 657{
658 long off = (long)(target - first); 658 long off = (long)(target - first);
@@ -663,7 +663,7 @@ static __init void copy_handler(struct reloc *rel, struct label *lab,
663 move_labels(lab, first, end, off); 663 move_labels(lab, first, end, off);
664} 664}
665 665
666static __init int __maybe_unused insn_has_bdelay(struct reloc *rel, 666static int __init __maybe_unused insn_has_bdelay(struct reloc *rel,
667 u32 *addr) 667 u32 *addr)
668{ 668{
669 for (; rel->lab != label_invalid; rel++) { 669 for (; rel->lab != label_invalid; rel++) {
@@ -755,11 +755,11 @@ il_bgez(u32 **p, struct reloc **r, unsigned int reg, enum label_id l)
755 * We deliberately chose a buffer size of 128, so we won't scribble 755 * We deliberately chose a buffer size of 128, so we won't scribble
756 * over anything important on overflow before we panic. 756 * over anything important on overflow before we panic.
757 */ 757 */
758static __initdata u32 tlb_handler[128]; 758static u32 tlb_handler[128] __initdata;
759 759
760/* simply assume worst case size for labels and relocs */ 760/* simply assume worst case size for labels and relocs */
761static __initdata struct label labels[128]; 761static struct label labels[128] __initdata;
762static __initdata struct reloc relocs[128]; 762static struct reloc relocs[128] __initdata;
763 763
764/* 764/*
765 * The R3000 TLB handler is simple. 765 * The R3000 TLB handler is simple.
@@ -813,7 +813,7 @@ static void __init build_r3000_tlb_refill_handler(void)
813 * other one.To keep things simple, we first assume linear space, 813 * other one.To keep things simple, we first assume linear space,
814 * then we relocate it to the final handler layout as needed. 814 * then we relocate it to the final handler layout as needed.
815 */ 815 */
816static __initdata u32 final_handler[64]; 816static u32 final_handler[64] __initdata;
817 817
818/* 818/*
819 * Hazards 819 * Hazards
@@ -837,7 +837,7 @@ static __initdata u32 final_handler[64];
837 * 837 *
838 * As if we MIPS hackers wouldn't know how to nop pipelines happy ... 838 * As if we MIPS hackers wouldn't know how to nop pipelines happy ...
839 */ 839 */
840static __init void __maybe_unused build_tlb_probe_entry(u32 **p) 840static void __init __maybe_unused build_tlb_probe_entry(u32 **p)
841{ 841{
842 switch (current_cpu_type()) { 842 switch (current_cpu_type()) {
843 /* Found by experiment: R4600 v2.0 needs this, too. */ 843 /* Found by experiment: R4600 v2.0 needs this, too. */
@@ -861,7 +861,7 @@ static __init void __maybe_unused build_tlb_probe_entry(u32 **p)
861 */ 861 */
862enum tlb_write_entry { tlb_random, tlb_indexed }; 862enum tlb_write_entry { tlb_random, tlb_indexed };
863 863
864static __init void build_tlb_write_entry(u32 **p, struct label **l, 864static void __init build_tlb_write_entry(u32 **p, struct label **l,
865 struct reloc **r, 865 struct reloc **r,
866 enum tlb_write_entry wmode) 866 enum tlb_write_entry wmode)
867{ 867{
@@ -1003,7 +1003,7 @@ static __init void build_tlb_write_entry(u32 **p, struct label **l,
1003 * TMP and PTR are scratch. 1003 * TMP and PTR are scratch.
1004 * TMP will be clobbered, PTR will hold the pmd entry. 1004 * TMP will be clobbered, PTR will hold the pmd entry.
1005 */ 1005 */
1006static __init void 1006static void __init
1007build_get_pmde64(u32 **p, struct label **l, struct reloc **r, 1007build_get_pmde64(u32 **p, struct label **l, struct reloc **r,
1008 unsigned int tmp, unsigned int ptr) 1008 unsigned int tmp, unsigned int ptr)
1009{ 1009{
@@ -1064,7 +1064,7 @@ build_get_pmde64(u32 **p, struct label **l, struct reloc **r,
1064 * BVADDR is the faulting address, PTR is scratch. 1064 * BVADDR is the faulting address, PTR is scratch.
1065 * PTR will hold the pgd for vmalloc. 1065 * PTR will hold the pgd for vmalloc.
1066 */ 1066 */
1067static __init void 1067static void __init
1068build_get_pgd_vmalloc64(u32 **p, struct label **l, struct reloc **r, 1068build_get_pgd_vmalloc64(u32 **p, struct label **l, struct reloc **r,
1069 unsigned int bvaddr, unsigned int ptr) 1069 unsigned int bvaddr, unsigned int ptr)
1070{ 1070{
@@ -1134,7 +1134,7 @@ build_get_pgd_vmalloc64(u32 **p, struct label **l, struct reloc **r,
1134 * TMP and PTR are scratch. 1134 * TMP and PTR are scratch.
1135 * TMP will be clobbered, PTR will hold the pgd entry. 1135 * TMP will be clobbered, PTR will hold the pgd entry.
1136 */ 1136 */
1137static __init void __maybe_unused 1137static void __init __maybe_unused
1138build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr) 1138build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr)
1139{ 1139{
1140 long pgdc = (long)pgd_current; 1140 long pgdc = (long)pgd_current;
@@ -1169,7 +1169,7 @@ build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr)
1169 1169
1170#endif /* !CONFIG_64BIT */ 1170#endif /* !CONFIG_64BIT */
1171 1171
1172static __init void build_adjust_context(u32 **p, unsigned int ctx) 1172static void __init build_adjust_context(u32 **p, unsigned int ctx)
1173{ 1173{
1174 unsigned int shift = 4 - (PTE_T_LOG2 + 1) + PAGE_SHIFT - 12; 1174 unsigned int shift = 4 - (PTE_T_LOG2 + 1) + PAGE_SHIFT - 12;
1175 unsigned int mask = (PTRS_PER_PTE / 2 - 1) << (PTE_T_LOG2 + 1); 1175 unsigned int mask = (PTRS_PER_PTE / 2 - 1) << (PTE_T_LOG2 + 1);
@@ -1195,7 +1195,7 @@ static __init void build_adjust_context(u32 **p, unsigned int ctx)
1195 i_andi(p, ctx, ctx, mask); 1195 i_andi(p, ctx, ctx, mask);
1196} 1196}
1197 1197
1198static __init void build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr) 1198static void __init build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr)
1199{ 1199{
1200 /* 1200 /*
1201 * Bug workaround for the Nevada. It seems as if under certain 1201 * Bug workaround for the Nevada. It seems as if under certain
@@ -1220,7 +1220,7 @@ static __init void build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr)
1220 i_ADDU(p, ptr, ptr, tmp); /* add in offset */ 1220 i_ADDU(p, ptr, ptr, tmp); /* add in offset */
1221} 1221}
1222 1222
1223static __init void build_update_entries(u32 **p, unsigned int tmp, 1223static void __init build_update_entries(u32 **p, unsigned int tmp,
1224 unsigned int ptep) 1224 unsigned int ptep)
1225{ 1225{
1226 /* 1226 /*