diff options
| author | Franck Bui-Huu <fbuihuu@gmail.com> | 2007-10-18 03:11:14 -0400 |
|---|---|---|
| committer | Ralf Baechle <ralf@linux-mips.org> | 2008-01-29 05:14:56 -0500 |
| commit | a9af6041e9e7ed0278d03333c3a22a8509f33615 (patch) | |
| tree | 1d05905c82e0a36c5c4c5d04c1933224fb82ae92 | |
| parent | 3f4ac2dc0504962b6cb9367adbc13272393b80b3 (diff) | |
[MIPS] tlbex.c: Cleanup __init usages.
Signed-off-by: Franck Bui-Huu <fbuihuu@gmail.com>
Signed-off-by: Ralf Baechle <ralf@linux-mips.org>
| -rw-r--r-- | arch/mips/mm/tlbex.c | 98 |
1 files changed, 49 insertions, 49 deletions
diff --git a/arch/mips/mm/tlbex.c b/arch/mips/mm/tlbex.c index f8925ba0b39e..77530fd96f1c 100644 --- a/arch/mips/mm/tlbex.c +++ b/arch/mips/mm/tlbex.c | |||
| @@ -67,7 +67,7 @@ static inline int __maybe_unused r10000_llsc_war(void) | |||
| 67 | * why; it's not an issue caused by the core RTL. | 67 | * why; it's not an issue caused by the core RTL. |
| 68 | * | 68 | * |
| 69 | */ | 69 | */ |
| 70 | static __init int __attribute__((unused)) m4kc_tlbp_war(void) | 70 | static int __init m4kc_tlbp_war(void) |
| 71 | { | 71 | { |
| 72 | return (current_cpu_data.processor_id & 0xffff00) == | 72 | return (current_cpu_data.processor_id & 0xffff00) == |
| 73 | (PRID_COMP_MIPS | PRID_IMP_4KC); | 73 | (PRID_COMP_MIPS | PRID_IMP_4KC); |
| @@ -141,7 +141,7 @@ struct insn { | |||
| 141 | | (e) << RE_SH \ | 141 | | (e) << RE_SH \ |
| 142 | | (f) << FUNC_SH) | 142 | | (f) << FUNC_SH) |
| 143 | 143 | ||
| 144 | static __initdata struct insn insn_table[] = { | 144 | static struct insn insn_table[] __initdata = { |
| 145 | { insn_addiu, M(addiu_op, 0, 0, 0, 0, 0), RS | RT | SIMM }, | 145 | { insn_addiu, M(addiu_op, 0, 0, 0, 0, 0), RS | RT | SIMM }, |
| 146 | { insn_addu, M(spec_op, 0, 0, 0, 0, addu_op), RS | RT | RD }, | 146 | { insn_addu, M(spec_op, 0, 0, 0, 0, addu_op), RS | RT | RD }, |
| 147 | { insn_and, M(spec_op, 0, 0, 0, 0, and_op), RS | RT | RD }, | 147 | { insn_and, M(spec_op, 0, 0, 0, 0, and_op), RS | RT | RD }, |
| @@ -194,7 +194,7 @@ static __initdata struct insn insn_table[] = { | |||
| 194 | 194 | ||
| 195 | #undef M | 195 | #undef M |
| 196 | 196 | ||
| 197 | static __init u32 build_rs(u32 arg) | 197 | static u32 __init build_rs(u32 arg) |
| 198 | { | 198 | { |
| 199 | if (arg & ~RS_MASK) | 199 | if (arg & ~RS_MASK) |
| 200 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 200 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
| @@ -202,7 +202,7 @@ static __init u32 build_rs(u32 arg) | |||
| 202 | return (arg & RS_MASK) << RS_SH; | 202 | return (arg & RS_MASK) << RS_SH; |
| 203 | } | 203 | } |
| 204 | 204 | ||
| 205 | static __init u32 build_rt(u32 arg) | 205 | static u32 __init build_rt(u32 arg) |
| 206 | { | 206 | { |
| 207 | if (arg & ~RT_MASK) | 207 | if (arg & ~RT_MASK) |
| 208 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 208 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
| @@ -210,7 +210,7 @@ static __init u32 build_rt(u32 arg) | |||
| 210 | return (arg & RT_MASK) << RT_SH; | 210 | return (arg & RT_MASK) << RT_SH; |
| 211 | } | 211 | } |
| 212 | 212 | ||
| 213 | static __init u32 build_rd(u32 arg) | 213 | static u32 __init build_rd(u32 arg) |
| 214 | { | 214 | { |
| 215 | if (arg & ~RD_MASK) | 215 | if (arg & ~RD_MASK) |
| 216 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 216 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
| @@ -218,7 +218,7 @@ static __init u32 build_rd(u32 arg) | |||
| 218 | return (arg & RD_MASK) << RD_SH; | 218 | return (arg & RD_MASK) << RD_SH; |
| 219 | } | 219 | } |
| 220 | 220 | ||
| 221 | static __init u32 build_re(u32 arg) | 221 | static u32 __init build_re(u32 arg) |
| 222 | { | 222 | { |
| 223 | if (arg & ~RE_MASK) | 223 | if (arg & ~RE_MASK) |
| 224 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 224 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
| @@ -226,7 +226,7 @@ static __init u32 build_re(u32 arg) | |||
| 226 | return (arg & RE_MASK) << RE_SH; | 226 | return (arg & RE_MASK) << RE_SH; |
| 227 | } | 227 | } |
| 228 | 228 | ||
| 229 | static __init u32 build_simm(s32 arg) | 229 | static u32 __init build_simm(s32 arg) |
| 230 | { | 230 | { |
| 231 | if (arg > 0x7fff || arg < -0x8000) | 231 | if (arg > 0x7fff || arg < -0x8000) |
| 232 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 232 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
| @@ -234,7 +234,7 @@ static __init u32 build_simm(s32 arg) | |||
| 234 | return arg & 0xffff; | 234 | return arg & 0xffff; |
| 235 | } | 235 | } |
| 236 | 236 | ||
| 237 | static __init u32 build_uimm(u32 arg) | 237 | static u32 __init build_uimm(u32 arg) |
| 238 | { | 238 | { |
| 239 | if (arg & ~IMM_MASK) | 239 | if (arg & ~IMM_MASK) |
| 240 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 240 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
| @@ -242,7 +242,7 @@ static __init u32 build_uimm(u32 arg) | |||
| 242 | return arg & IMM_MASK; | 242 | return arg & IMM_MASK; |
| 243 | } | 243 | } |
| 244 | 244 | ||
| 245 | static __init u32 build_bimm(s32 arg) | 245 | static u32 __init build_bimm(s32 arg) |
| 246 | { | 246 | { |
| 247 | if (arg > 0x1ffff || arg < -0x20000) | 247 | if (arg > 0x1ffff || arg < -0x20000) |
| 248 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 248 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
| @@ -253,7 +253,7 @@ static __init u32 build_bimm(s32 arg) | |||
| 253 | return ((arg < 0) ? (1 << 15) : 0) | ((arg >> 2) & 0x7fff); | 253 | return ((arg < 0) ? (1 << 15) : 0) | ((arg >> 2) & 0x7fff); |
| 254 | } | 254 | } |
| 255 | 255 | ||
| 256 | static __init u32 build_jimm(u32 arg) | 256 | static u32 __init build_jimm(u32 arg) |
| 257 | { | 257 | { |
| 258 | if (arg & ~((JIMM_MASK) << 2)) | 258 | if (arg & ~((JIMM_MASK) << 2)) |
| 259 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 259 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
| @@ -261,7 +261,7 @@ static __init u32 build_jimm(u32 arg) | |||
| 261 | return (arg >> 2) & JIMM_MASK; | 261 | return (arg >> 2) & JIMM_MASK; |
| 262 | } | 262 | } |
| 263 | 263 | ||
| 264 | static __init u32 build_func(u32 arg) | 264 | static u32 __init build_func(u32 arg) |
| 265 | { | 265 | { |
| 266 | if (arg & ~FUNC_MASK) | 266 | if (arg & ~FUNC_MASK) |
| 267 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 267 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
| @@ -269,7 +269,7 @@ static __init u32 build_func(u32 arg) | |||
| 269 | return arg & FUNC_MASK; | 269 | return arg & FUNC_MASK; |
| 270 | } | 270 | } |
| 271 | 271 | ||
| 272 | static __init u32 build_set(u32 arg) | 272 | static u32 __init build_set(u32 arg) |
| 273 | { | 273 | { |
| 274 | if (arg & ~SET_MASK) | 274 | if (arg & ~SET_MASK) |
| 275 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); | 275 | printk(KERN_WARNING "TLB synthesizer field overflow\n"); |
| @@ -316,69 +316,69 @@ static void __init build_insn(u32 **buf, enum opcode opc, ...) | |||
| 316 | } | 316 | } |
| 317 | 317 | ||
| 318 | #define I_u1u2u3(op) \ | 318 | #define I_u1u2u3(op) \ |
| 319 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 319 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
| 320 | unsigned int b, unsigned int c) \ | 320 | unsigned int b, unsigned int c) \ |
| 321 | { \ | 321 | { \ |
| 322 | build_insn(buf, insn##op, a, b, c); \ | 322 | build_insn(buf, insn##op, a, b, c); \ |
| 323 | } | 323 | } |
| 324 | 324 | ||
| 325 | #define I_u2u1u3(op) \ | 325 | #define I_u2u1u3(op) \ |
| 326 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 326 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
| 327 | unsigned int b, unsigned int c) \ | 327 | unsigned int b, unsigned int c) \ |
| 328 | { \ | 328 | { \ |
| 329 | build_insn(buf, insn##op, b, a, c); \ | 329 | build_insn(buf, insn##op, b, a, c); \ |
| 330 | } | 330 | } |
| 331 | 331 | ||
| 332 | #define I_u3u1u2(op) \ | 332 | #define I_u3u1u2(op) \ |
| 333 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 333 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
| 334 | unsigned int b, unsigned int c) \ | 334 | unsigned int b, unsigned int c) \ |
| 335 | { \ | 335 | { \ |
| 336 | build_insn(buf, insn##op, b, c, a); \ | 336 | build_insn(buf, insn##op, b, c, a); \ |
| 337 | } | 337 | } |
| 338 | 338 | ||
| 339 | #define I_u1u2s3(op) \ | 339 | #define I_u1u2s3(op) \ |
| 340 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 340 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
| 341 | unsigned int b, signed int c) \ | 341 | unsigned int b, signed int c) \ |
| 342 | { \ | 342 | { \ |
| 343 | build_insn(buf, insn##op, a, b, c); \ | 343 | build_insn(buf, insn##op, a, b, c); \ |
| 344 | } | 344 | } |
| 345 | 345 | ||
| 346 | #define I_u2s3u1(op) \ | 346 | #define I_u2s3u1(op) \ |
| 347 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 347 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
| 348 | signed int b, unsigned int c) \ | 348 | signed int b, unsigned int c) \ |
| 349 | { \ | 349 | { \ |
| 350 | build_insn(buf, insn##op, c, a, b); \ | 350 | build_insn(buf, insn##op, c, a, b); \ |
| 351 | } | 351 | } |
| 352 | 352 | ||
| 353 | #define I_u2u1s3(op) \ | 353 | #define I_u2u1s3(op) \ |
| 354 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 354 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
| 355 | unsigned int b, signed int c) \ | 355 | unsigned int b, signed int c) \ |
| 356 | { \ | 356 | { \ |
| 357 | build_insn(buf, insn##op, b, a, c); \ | 357 | build_insn(buf, insn##op, b, a, c); \ |
| 358 | } | 358 | } |
| 359 | 359 | ||
| 360 | #define I_u1u2(op) \ | 360 | #define I_u1u2(op) \ |
| 361 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 361 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
| 362 | unsigned int b) \ | 362 | unsigned int b) \ |
| 363 | { \ | 363 | { \ |
| 364 | build_insn(buf, insn##op, a, b); \ | 364 | build_insn(buf, insn##op, a, b); \ |
| 365 | } | 365 | } |
| 366 | 366 | ||
| 367 | #define I_u1s2(op) \ | 367 | #define I_u1s2(op) \ |
| 368 | static inline void __init i##op(u32 **buf, unsigned int a, \ | 368 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a, \ |
| 369 | signed int b) \ | 369 | signed int b) \ |
| 370 | { \ | 370 | { \ |
| 371 | build_insn(buf, insn##op, a, b); \ | 371 | build_insn(buf, insn##op, a, b); \ |
| 372 | } | 372 | } |
| 373 | 373 | ||
| 374 | #define I_u1(op) \ | 374 | #define I_u1(op) \ |
| 375 | static inline void __init i##op(u32 **buf, unsigned int a) \ | 375 | static void __init __maybe_unused i##op(u32 **buf, unsigned int a) \ |
| 376 | { \ | 376 | { \ |
| 377 | build_insn(buf, insn##op, a); \ | 377 | build_insn(buf, insn##op, a); \ |
| 378 | } | 378 | } |
| 379 | 379 | ||
| 380 | #define I_0(op) \ | 380 | #define I_0(op) \ |
| 381 | static inline void __init i##op(u32 **buf) \ | 381 | static void __init __maybe_unused i##op(u32 **buf) \ |
| 382 | { \ | 382 | { \ |
| 383 | build_insn(buf, insn##op); \ | 383 | build_insn(buf, insn##op); \ |
| 384 | } | 384 | } |
| @@ -458,7 +458,7 @@ struct label { | |||
| 458 | enum label_id lab; | 458 | enum label_id lab; |
| 459 | }; | 459 | }; |
| 460 | 460 | ||
| 461 | static __init void build_label(struct label **lab, u32 *addr, | 461 | static void __init build_label(struct label **lab, u32 *addr, |
| 462 | enum label_id l) | 462 | enum label_id l) |
| 463 | { | 463 | { |
| 464 | (*lab)->addr = addr; | 464 | (*lab)->addr = addr; |
| @@ -467,7 +467,7 @@ static __init void build_label(struct label **lab, u32 *addr, | |||
| 467 | } | 467 | } |
| 468 | 468 | ||
| 469 | #define L_LA(lb) \ | 469 | #define L_LA(lb) \ |
| 470 | static inline void l##lb(struct label **lab, u32 *addr) \ | 470 | static inline void __init l##lb(struct label **lab, u32 *addr) \ |
| 471 | { \ | 471 | { \ |
| 472 | build_label(lab, addr, label##lb); \ | 472 | build_label(lab, addr, label##lb); \ |
| 473 | } | 473 | } |
| @@ -526,7 +526,7 @@ L_LA(_r3000_write_probe_fail) | |||
| 526 | #define i_ssnop(buf) i_sll(buf, 0, 0, 1) | 526 | #define i_ssnop(buf) i_sll(buf, 0, 0, 1) |
| 527 | #define i_ehb(buf) i_sll(buf, 0, 0, 3) | 527 | #define i_ehb(buf) i_sll(buf, 0, 0, 3) |
| 528 | 528 | ||
| 529 | static __init int __maybe_unused in_compat_space_p(long addr) | 529 | static int __init __maybe_unused in_compat_space_p(long addr) |
| 530 | { | 530 | { |
| 531 | /* Is this address in 32bit compat space? */ | 531 | /* Is this address in 32bit compat space? */ |
| 532 | #ifdef CONFIG_64BIT | 532 | #ifdef CONFIG_64BIT |
| @@ -536,7 +536,7 @@ static __init int __maybe_unused in_compat_space_p(long addr) | |||
| 536 | #endif | 536 | #endif |
| 537 | } | 537 | } |
| 538 | 538 | ||
| 539 | static __init int __maybe_unused rel_highest(long val) | 539 | static int __init __maybe_unused rel_highest(long val) |
| 540 | { | 540 | { |
| 541 | #ifdef CONFIG_64BIT | 541 | #ifdef CONFIG_64BIT |
| 542 | return ((((val + 0x800080008000L) >> 48) & 0xffff) ^ 0x8000) - 0x8000; | 542 | return ((((val + 0x800080008000L) >> 48) & 0xffff) ^ 0x8000) - 0x8000; |
| @@ -545,7 +545,7 @@ static __init int __maybe_unused rel_highest(long val) | |||
| 545 | #endif | 545 | #endif |
| 546 | } | 546 | } |
| 547 | 547 | ||
| 548 | static __init int __maybe_unused rel_higher(long val) | 548 | static int __init __maybe_unused rel_higher(long val) |
| 549 | { | 549 | { |
| 550 | #ifdef CONFIG_64BIT | 550 | #ifdef CONFIG_64BIT |
| 551 | return ((((val + 0x80008000L) >> 32) & 0xffff) ^ 0x8000) - 0x8000; | 551 | return ((((val + 0x80008000L) >> 32) & 0xffff) ^ 0x8000) - 0x8000; |
| @@ -554,17 +554,17 @@ static __init int __maybe_unused rel_higher(long val) | |||
| 554 | #endif | 554 | #endif |
| 555 | } | 555 | } |
| 556 | 556 | ||
| 557 | static __init int rel_hi(long val) | 557 | static int __init rel_hi(long val) |
| 558 | { | 558 | { |
| 559 | return ((((val + 0x8000L) >> 16) & 0xffff) ^ 0x8000) - 0x8000; | 559 | return ((((val + 0x8000L) >> 16) & 0xffff) ^ 0x8000) - 0x8000; |
| 560 | } | 560 | } |
| 561 | 561 | ||
| 562 | static __init int rel_lo(long val) | 562 | static int __init rel_lo(long val) |
| 563 | { | 563 | { |
| 564 | return ((val & 0xffff) ^ 0x8000) - 0x8000; | 564 | return ((val & 0xffff) ^ 0x8000) - 0x8000; |
| 565 | } | 565 | } |
| 566 | 566 | ||
| 567 | static __init void i_LA_mostly(u32 **buf, unsigned int rs, long addr) | 567 | static void __init i_LA_mostly(u32 **buf, unsigned int rs, long addr) |
| 568 | { | 568 | { |
| 569 | if (!in_compat_space_p(addr)) { | 569 | if (!in_compat_space_p(addr)) { |
| 570 | i_lui(buf, rs, rel_highest(addr)); | 570 | i_lui(buf, rs, rel_highest(addr)); |
| @@ -580,7 +580,7 @@ static __init void i_LA_mostly(u32 **buf, unsigned int rs, long addr) | |||
| 580 | i_lui(buf, rs, rel_hi(addr)); | 580 | i_lui(buf, rs, rel_hi(addr)); |
| 581 | } | 581 | } |
| 582 | 582 | ||
| 583 | static __init void __maybe_unused i_LA(u32 **buf, unsigned int rs, long addr) | 583 | static void __init __maybe_unused i_LA(u32 **buf, unsigned int rs, long addr) |
| 584 | { | 584 | { |
| 585 | i_LA_mostly(buf, rs, addr); | 585 | i_LA_mostly(buf, rs, addr); |
| 586 | if (rel_lo(addr)) { | 586 | if (rel_lo(addr)) { |
| @@ -601,7 +601,7 @@ struct reloc { | |||
| 601 | enum label_id lab; | 601 | enum label_id lab; |
| 602 | }; | 602 | }; |
| 603 | 603 | ||
| 604 | static __init void r_mips_pc16(struct reloc **rel, u32 *addr, | 604 | static void __init r_mips_pc16(struct reloc **rel, u32 *addr, |
| 605 | enum label_id l) | 605 | enum label_id l) |
| 606 | { | 606 | { |
| 607 | (*rel)->addr = addr; | 607 | (*rel)->addr = addr; |
| @@ -626,7 +626,7 @@ static inline void __resolve_relocs(struct reloc *rel, struct label *lab) | |||
| 626 | } | 626 | } |
| 627 | } | 627 | } |
| 628 | 628 | ||
| 629 | static __init void resolve_relocs(struct reloc *rel, struct label *lab) | 629 | static void __init resolve_relocs(struct reloc *rel, struct label *lab) |
| 630 | { | 630 | { |
| 631 | struct label *l; | 631 | struct label *l; |
| 632 | 632 | ||
| @@ -636,7 +636,7 @@ static __init void resolve_relocs(struct reloc *rel, struct label *lab) | |||
| 636 | __resolve_relocs(rel, l); | 636 | __resolve_relocs(rel, l); |
| 637 | } | 637 | } |
| 638 | 638 | ||
| 639 | static __init void move_relocs(struct reloc *rel, u32 *first, u32 *end, | 639 | static void __init move_relocs(struct reloc *rel, u32 *first, u32 *end, |
| 640 | long off) | 640 | long off) |
| 641 | { | 641 | { |
| 642 | for (; rel->lab != label_invalid; rel++) | 642 | for (; rel->lab != label_invalid; rel++) |
| @@ -644,7 +644,7 @@ static __init void move_relocs(struct reloc *rel, u32 *first, u32 *end, | |||
| 644 | rel->addr += off; | 644 | rel->addr += off; |
| 645 | } | 645 | } |
| 646 | 646 | ||
| 647 | static __init void move_labels(struct label *lab, u32 *first, u32 *end, | 647 | static void __init move_labels(struct label *lab, u32 *first, u32 *end, |
| 648 | long off) | 648 | long off) |
| 649 | { | 649 | { |
| 650 | for (; lab->lab != label_invalid; lab++) | 650 | for (; lab->lab != label_invalid; lab++) |
| @@ -652,7 +652,7 @@ static __init void move_labels(struct label *lab, u32 *first, u32 *end, | |||
| 652 | lab->addr += off; | 652 | lab->addr += off; |
| 653 | } | 653 | } |
| 654 | 654 | ||
| 655 | static __init void copy_handler(struct reloc *rel, struct label *lab, | 655 | static void __init copy_handler(struct reloc *rel, struct label *lab, |
| 656 | u32 *first, u32 *end, u32 *target) | 656 | u32 *first, u32 *end, u32 *target) |
| 657 | { | 657 | { |
| 658 | long off = (long)(target - first); | 658 | long off = (long)(target - first); |
| @@ -663,7 +663,7 @@ static __init void copy_handler(struct reloc *rel, struct label *lab, | |||
| 663 | move_labels(lab, first, end, off); | 663 | move_labels(lab, first, end, off); |
| 664 | } | 664 | } |
| 665 | 665 | ||
| 666 | static __init int __maybe_unused insn_has_bdelay(struct reloc *rel, | 666 | static int __init __maybe_unused insn_has_bdelay(struct reloc *rel, |
| 667 | u32 *addr) | 667 | u32 *addr) |
| 668 | { | 668 | { |
| 669 | for (; rel->lab != label_invalid; rel++) { | 669 | for (; rel->lab != label_invalid; rel++) { |
| @@ -755,11 +755,11 @@ il_bgez(u32 **p, struct reloc **r, unsigned int reg, enum label_id l) | |||
| 755 | * We deliberately chose a buffer size of 128, so we won't scribble | 755 | * We deliberately chose a buffer size of 128, so we won't scribble |
| 756 | * over anything important on overflow before we panic. | 756 | * over anything important on overflow before we panic. |
| 757 | */ | 757 | */ |
| 758 | static __initdata u32 tlb_handler[128]; | 758 | static u32 tlb_handler[128] __initdata; |
| 759 | 759 | ||
| 760 | /* simply assume worst case size for labels and relocs */ | 760 | /* simply assume worst case size for labels and relocs */ |
| 761 | static __initdata struct label labels[128]; | 761 | static struct label labels[128] __initdata; |
| 762 | static __initdata struct reloc relocs[128]; | 762 | static struct reloc relocs[128] __initdata; |
| 763 | 763 | ||
| 764 | /* | 764 | /* |
| 765 | * The R3000 TLB handler is simple. | 765 | * The R3000 TLB handler is simple. |
| @@ -813,7 +813,7 @@ static void __init build_r3000_tlb_refill_handler(void) | |||
| 813 | * other one.To keep things simple, we first assume linear space, | 813 | * other one.To keep things simple, we first assume linear space, |
| 814 | * then we relocate it to the final handler layout as needed. | 814 | * then we relocate it to the final handler layout as needed. |
| 815 | */ | 815 | */ |
| 816 | static __initdata u32 final_handler[64]; | 816 | static u32 final_handler[64] __initdata; |
| 817 | 817 | ||
| 818 | /* | 818 | /* |
| 819 | * Hazards | 819 | * Hazards |
| @@ -837,7 +837,7 @@ static __initdata u32 final_handler[64]; | |||
| 837 | * | 837 | * |
| 838 | * As if we MIPS hackers wouldn't know how to nop pipelines happy ... | 838 | * As if we MIPS hackers wouldn't know how to nop pipelines happy ... |
| 839 | */ | 839 | */ |
| 840 | static __init void __maybe_unused build_tlb_probe_entry(u32 **p) | 840 | static void __init __maybe_unused build_tlb_probe_entry(u32 **p) |
| 841 | { | 841 | { |
| 842 | switch (current_cpu_type()) { | 842 | switch (current_cpu_type()) { |
| 843 | /* Found by experiment: R4600 v2.0 needs this, too. */ | 843 | /* Found by experiment: R4600 v2.0 needs this, too. */ |
| @@ -861,7 +861,7 @@ static __init void __maybe_unused build_tlb_probe_entry(u32 **p) | |||
| 861 | */ | 861 | */ |
| 862 | enum tlb_write_entry { tlb_random, tlb_indexed }; | 862 | enum tlb_write_entry { tlb_random, tlb_indexed }; |
| 863 | 863 | ||
| 864 | static __init void build_tlb_write_entry(u32 **p, struct label **l, | 864 | static void __init build_tlb_write_entry(u32 **p, struct label **l, |
| 865 | struct reloc **r, | 865 | struct reloc **r, |
| 866 | enum tlb_write_entry wmode) | 866 | enum tlb_write_entry wmode) |
| 867 | { | 867 | { |
| @@ -1003,7 +1003,7 @@ static __init void build_tlb_write_entry(u32 **p, struct label **l, | |||
| 1003 | * TMP and PTR are scratch. | 1003 | * TMP and PTR are scratch. |
| 1004 | * TMP will be clobbered, PTR will hold the pmd entry. | 1004 | * TMP will be clobbered, PTR will hold the pmd entry. |
| 1005 | */ | 1005 | */ |
| 1006 | static __init void | 1006 | static void __init |
| 1007 | build_get_pmde64(u32 **p, struct label **l, struct reloc **r, | 1007 | build_get_pmde64(u32 **p, struct label **l, struct reloc **r, |
| 1008 | unsigned int tmp, unsigned int ptr) | 1008 | unsigned int tmp, unsigned int ptr) |
| 1009 | { | 1009 | { |
| @@ -1064,7 +1064,7 @@ build_get_pmde64(u32 **p, struct label **l, struct reloc **r, | |||
| 1064 | * BVADDR is the faulting address, PTR is scratch. | 1064 | * BVADDR is the faulting address, PTR is scratch. |
| 1065 | * PTR will hold the pgd for vmalloc. | 1065 | * PTR will hold the pgd for vmalloc. |
| 1066 | */ | 1066 | */ |
| 1067 | static __init void | 1067 | static void __init |
| 1068 | build_get_pgd_vmalloc64(u32 **p, struct label **l, struct reloc **r, | 1068 | build_get_pgd_vmalloc64(u32 **p, struct label **l, struct reloc **r, |
| 1069 | unsigned int bvaddr, unsigned int ptr) | 1069 | unsigned int bvaddr, unsigned int ptr) |
| 1070 | { | 1070 | { |
| @@ -1134,7 +1134,7 @@ build_get_pgd_vmalloc64(u32 **p, struct label **l, struct reloc **r, | |||
| 1134 | * TMP and PTR are scratch. | 1134 | * TMP and PTR are scratch. |
| 1135 | * TMP will be clobbered, PTR will hold the pgd entry. | 1135 | * TMP will be clobbered, PTR will hold the pgd entry. |
| 1136 | */ | 1136 | */ |
| 1137 | static __init void __maybe_unused | 1137 | static void __init __maybe_unused |
| 1138 | build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr) | 1138 | build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr) |
| 1139 | { | 1139 | { |
| 1140 | long pgdc = (long)pgd_current; | 1140 | long pgdc = (long)pgd_current; |
| @@ -1169,7 +1169,7 @@ build_get_pgde32(u32 **p, unsigned int tmp, unsigned int ptr) | |||
| 1169 | 1169 | ||
| 1170 | #endif /* !CONFIG_64BIT */ | 1170 | #endif /* !CONFIG_64BIT */ |
| 1171 | 1171 | ||
| 1172 | static __init void build_adjust_context(u32 **p, unsigned int ctx) | 1172 | static void __init build_adjust_context(u32 **p, unsigned int ctx) |
| 1173 | { | 1173 | { |
| 1174 | unsigned int shift = 4 - (PTE_T_LOG2 + 1) + PAGE_SHIFT - 12; | 1174 | unsigned int shift = 4 - (PTE_T_LOG2 + 1) + PAGE_SHIFT - 12; |
| 1175 | unsigned int mask = (PTRS_PER_PTE / 2 - 1) << (PTE_T_LOG2 + 1); | 1175 | unsigned int mask = (PTRS_PER_PTE / 2 - 1) << (PTE_T_LOG2 + 1); |
| @@ -1195,7 +1195,7 @@ static __init void build_adjust_context(u32 **p, unsigned int ctx) | |||
| 1195 | i_andi(p, ctx, ctx, mask); | 1195 | i_andi(p, ctx, ctx, mask); |
| 1196 | } | 1196 | } |
| 1197 | 1197 | ||
| 1198 | static __init void build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr) | 1198 | static void __init build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr) |
| 1199 | { | 1199 | { |
| 1200 | /* | 1200 | /* |
| 1201 | * Bug workaround for the Nevada. It seems as if under certain | 1201 | * Bug workaround for the Nevada. It seems as if under certain |
| @@ -1220,7 +1220,7 @@ static __init void build_get_ptep(u32 **p, unsigned int tmp, unsigned int ptr) | |||
| 1220 | i_ADDU(p, ptr, ptr, tmp); /* add in offset */ | 1220 | i_ADDU(p, ptr, ptr, tmp); /* add in offset */ |
| 1221 | } | 1221 | } |
| 1222 | 1222 | ||
| 1223 | static __init void build_update_entries(u32 **p, unsigned int tmp, | 1223 | static void __init build_update_entries(u32 **p, unsigned int tmp, |
| 1224 | unsigned int ptep) | 1224 | unsigned int ptep) |
| 1225 | { | 1225 | { |
| 1226 | /* | 1226 | /* |
