diff options
Diffstat (limited to 'arch/m32r/kernel/smp.c')
-rw-r--r-- | arch/m32r/kernel/smp.c | 68 |
1 files changed, 28 insertions, 40 deletions
diff --git a/arch/m32r/kernel/smp.c b/arch/m32r/kernel/smp.c index 31cef20b2996..092d40a6708e 100644 --- a/arch/m32r/kernel/smp.c +++ b/arch/m32r/kernel/smp.c | |||
@@ -30,6 +30,7 @@ | |||
30 | #include <asm/io.h> | 30 | #include <asm/io.h> |
31 | #include <asm/mmu_context.h> | 31 | #include <asm/mmu_context.h> |
32 | #include <asm/m32r.h> | 32 | #include <asm/m32r.h> |
33 | #include <asm/tlbflush.h> | ||
33 | 34 | ||
34 | /*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*/ | 35 | /*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*/ |
35 | /* Data structures and variables */ | 36 | /* Data structures and variables */ |
@@ -61,33 +62,22 @@ extern spinlock_t ipi_lock[]; | |||
61 | /* Function Prototypes */ | 62 | /* Function Prototypes */ |
62 | /*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*/ | 63 | /*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*/ |
63 | 64 | ||
64 | void smp_send_reschedule(int); | ||
65 | void smp_reschedule_interrupt(void); | 65 | void smp_reschedule_interrupt(void); |
66 | |||
67 | void smp_flush_cache_all(void); | ||
68 | void smp_flush_cache_all_interrupt(void); | 66 | void smp_flush_cache_all_interrupt(void); |
69 | 67 | ||
70 | void smp_flush_tlb_all(void); | ||
71 | static void flush_tlb_all_ipi(void *); | 68 | static void flush_tlb_all_ipi(void *); |
72 | |||
73 | void smp_flush_tlb_mm(struct mm_struct *); | ||
74 | void smp_flush_tlb_range(struct vm_area_struct *, unsigned long, \ | ||
75 | unsigned long); | ||
76 | void smp_flush_tlb_page(struct vm_area_struct *, unsigned long); | ||
77 | static void flush_tlb_others(cpumask_t, struct mm_struct *, | 69 | static void flush_tlb_others(cpumask_t, struct mm_struct *, |
78 | struct vm_area_struct *, unsigned long); | 70 | struct vm_area_struct *, unsigned long); |
71 | |||
79 | void smp_invalidate_interrupt(void); | 72 | void smp_invalidate_interrupt(void); |
80 | 73 | ||
81 | void smp_send_stop(void); | ||
82 | static void stop_this_cpu(void *); | 74 | static void stop_this_cpu(void *); |
83 | 75 | ||
84 | void smp_send_timer(void); | ||
85 | void smp_ipi_timer_interrupt(struct pt_regs *); | 76 | void smp_ipi_timer_interrupt(struct pt_regs *); |
86 | void smp_local_timer_interrupt(void); | 77 | void smp_local_timer_interrupt(void); |
87 | 78 | ||
88 | static void send_IPI_allbutself(int, int); | 79 | static void send_IPI_allbutself(int, int); |
89 | static void send_IPI_mask(const struct cpumask *, int, int); | 80 | static void send_IPI_mask(const struct cpumask *, int, int); |
90 | unsigned long send_IPI_mask_phys(cpumask_t, int, int); | ||
91 | 81 | ||
92 | /*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*/ | 82 | /*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*/ |
93 | /* Rescheduling request Routines */ | 83 | /* Rescheduling request Routines */ |
@@ -122,8 +112,6 @@ void smp_send_reschedule(int cpu_id) | |||
122 | * | 112 | * |
123 | * Description: This routine executes on CPU which received | 113 | * Description: This routine executes on CPU which received |
124 | * 'RESCHEDULE_IPI'. | 114 | * 'RESCHEDULE_IPI'. |
125 | * Rescheduling is processed at the exit of interrupt | ||
126 | * operation. | ||
127 | * | 115 | * |
128 | * Born on Date: 2002.02.05 | 116 | * Born on Date: 2002.02.05 |
129 | * | 117 | * |
@@ -138,7 +126,7 @@ void smp_send_reschedule(int cpu_id) | |||
138 | *==========================================================================*/ | 126 | *==========================================================================*/ |
139 | void smp_reschedule_interrupt(void) | 127 | void smp_reschedule_interrupt(void) |
140 | { | 128 | { |
141 | /* nothing to do */ | 129 | scheduler_ipi(); |
142 | } | 130 | } |
143 | 131 | ||
144 | /*==========================================================================* | 132 | /*==========================================================================* |
@@ -164,10 +152,10 @@ void smp_flush_cache_all(void) | |||
164 | unsigned long *mask; | 152 | unsigned long *mask; |
165 | 153 | ||
166 | preempt_disable(); | 154 | preempt_disable(); |
167 | cpumask = cpu_online_map; | 155 | cpumask_copy(&cpumask, cpu_online_mask); |
168 | cpu_clear(smp_processor_id(), cpumask); | 156 | cpumask_clear_cpu(smp_processor_id(), &cpumask); |
169 | spin_lock(&flushcache_lock); | 157 | spin_lock(&flushcache_lock); |
170 | mask=cpus_addr(cpumask); | 158 | mask=cpumask_bits(&cpumask); |
171 | atomic_set_mask(*mask, (atomic_t *)&flushcache_cpumask); | 159 | atomic_set_mask(*mask, (atomic_t *)&flushcache_cpumask); |
172 | send_IPI_mask(&cpumask, INVALIDATE_CACHE_IPI, 0); | 160 | send_IPI_mask(&cpumask, INVALIDATE_CACHE_IPI, 0); |
173 | _flush_cache_copyback_all(); | 161 | _flush_cache_copyback_all(); |
@@ -265,8 +253,8 @@ void smp_flush_tlb_mm(struct mm_struct *mm) | |||
265 | preempt_disable(); | 253 | preempt_disable(); |
266 | cpu_id = smp_processor_id(); | 254 | cpu_id = smp_processor_id(); |
267 | mmc = &mm->context[cpu_id]; | 255 | mmc = &mm->context[cpu_id]; |
268 | cpu_mask = *mm_cpumask(mm); | 256 | cpumask_copy(&cpu_mask, mm_cpumask(mm)); |
269 | cpu_clear(cpu_id, cpu_mask); | 257 | cpumask_clear_cpu(cpu_id, &cpu_mask); |
270 | 258 | ||
271 | if (*mmc != NO_CONTEXT) { | 259 | if (*mmc != NO_CONTEXT) { |
272 | local_irq_save(flags); | 260 | local_irq_save(flags); |
@@ -277,7 +265,7 @@ void smp_flush_tlb_mm(struct mm_struct *mm) | |||
277 | cpumask_clear_cpu(cpu_id, mm_cpumask(mm)); | 265 | cpumask_clear_cpu(cpu_id, mm_cpumask(mm)); |
278 | local_irq_restore(flags); | 266 | local_irq_restore(flags); |
279 | } | 267 | } |
280 | if (!cpus_empty(cpu_mask)) | 268 | if (!cpumask_empty(&cpu_mask)) |
281 | flush_tlb_others(cpu_mask, mm, NULL, FLUSH_ALL); | 269 | flush_tlb_others(cpu_mask, mm, NULL, FLUSH_ALL); |
282 | 270 | ||
283 | preempt_enable(); | 271 | preempt_enable(); |
@@ -335,8 +323,8 @@ void smp_flush_tlb_page(struct vm_area_struct *vma, unsigned long va) | |||
335 | preempt_disable(); | 323 | preempt_disable(); |
336 | cpu_id = smp_processor_id(); | 324 | cpu_id = smp_processor_id(); |
337 | mmc = &mm->context[cpu_id]; | 325 | mmc = &mm->context[cpu_id]; |
338 | cpu_mask = *mm_cpumask(mm); | 326 | cpumask_copy(&cpu_mask, mm_cpumask(mm)); |
339 | cpu_clear(cpu_id, cpu_mask); | 327 | cpumask_clear_cpu(cpu_id, &cpu_mask); |
340 | 328 | ||
341 | #ifdef DEBUG_SMP | 329 | #ifdef DEBUG_SMP |
342 | if (!mm) | 330 | if (!mm) |
@@ -350,7 +338,7 @@ void smp_flush_tlb_page(struct vm_area_struct *vma, unsigned long va) | |||
350 | __flush_tlb_page(va); | 338 | __flush_tlb_page(va); |
351 | local_irq_restore(flags); | 339 | local_irq_restore(flags); |
352 | } | 340 | } |
353 | if (!cpus_empty(cpu_mask)) | 341 | if (!cpumask_empty(&cpu_mask)) |
354 | flush_tlb_others(cpu_mask, mm, vma, va); | 342 | flush_tlb_others(cpu_mask, mm, vma, va); |
355 | 343 | ||
356 | preempt_enable(); | 344 | preempt_enable(); |
@@ -397,14 +385,14 @@ static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, | |||
397 | * - current CPU must not be in mask | 385 | * - current CPU must not be in mask |
398 | * - mask must exist :) | 386 | * - mask must exist :) |
399 | */ | 387 | */ |
400 | BUG_ON(cpus_empty(cpumask)); | 388 | BUG_ON(cpumask_empty(&cpumask)); |
401 | 389 | ||
402 | BUG_ON(cpu_isset(smp_processor_id(), cpumask)); | 390 | BUG_ON(cpumask_test_cpu(smp_processor_id(), &cpumask)); |
403 | BUG_ON(!mm); | 391 | BUG_ON(!mm); |
404 | 392 | ||
405 | /* If a CPU which we ran on has gone down, OK. */ | 393 | /* If a CPU which we ran on has gone down, OK. */ |
406 | cpus_and(cpumask, cpumask, cpu_online_map); | 394 | cpumask_and(&cpumask, &cpumask, cpu_online_mask); |
407 | if (cpus_empty(cpumask)) | 395 | if (cpumask_empty(&cpumask)) |
408 | return; | 396 | return; |
409 | 397 | ||
410 | /* | 398 | /* |
@@ -418,7 +406,7 @@ static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, | |||
418 | flush_mm = mm; | 406 | flush_mm = mm; |
419 | flush_vma = vma; | 407 | flush_vma = vma; |
420 | flush_va = va; | 408 | flush_va = va; |
421 | mask=cpus_addr(cpumask); | 409 | mask=cpumask_bits(&cpumask); |
422 | atomic_set_mask(*mask, (atomic_t *)&flush_cpumask); | 410 | atomic_set_mask(*mask, (atomic_t *)&flush_cpumask); |
423 | 411 | ||
424 | /* | 412 | /* |
@@ -427,7 +415,7 @@ static void flush_tlb_others(cpumask_t cpumask, struct mm_struct *mm, | |||
427 | */ | 415 | */ |
428 | send_IPI_mask(&cpumask, INVALIDATE_TLB_IPI, 0); | 416 | send_IPI_mask(&cpumask, INVALIDATE_TLB_IPI, 0); |
429 | 417 | ||
430 | while (!cpus_empty(flush_cpumask)) { | 418 | while (!cpumask_empty((cpumask_t*)&flush_cpumask)) { |
431 | /* nothing. lockup detection does not belong here */ | 419 | /* nothing. lockup detection does not belong here */ |
432 | mb(); | 420 | mb(); |
433 | } | 421 | } |
@@ -462,7 +450,7 @@ void smp_invalidate_interrupt(void) | |||
462 | int cpu_id = smp_processor_id(); | 450 | int cpu_id = smp_processor_id(); |
463 | unsigned long *mmc = &flush_mm->context[cpu_id]; | 451 | unsigned long *mmc = &flush_mm->context[cpu_id]; |
464 | 452 | ||
465 | if (!cpu_isset(cpu_id, flush_cpumask)) | 453 | if (!cpumask_test_cpu(cpu_id, &flush_cpumask)) |
466 | return; | 454 | return; |
467 | 455 | ||
468 | if (flush_va == FLUSH_ALL) { | 456 | if (flush_va == FLUSH_ALL) { |
@@ -480,7 +468,7 @@ void smp_invalidate_interrupt(void) | |||
480 | __flush_tlb_page(va); | 468 | __flush_tlb_page(va); |
481 | } | 469 | } |
482 | } | 470 | } |
483 | cpu_clear(cpu_id, flush_cpumask); | 471 | cpumask_clear_cpu(cpu_id, (cpumask_t*)&flush_cpumask); |
484 | } | 472 | } |
485 | 473 | ||
486 | /*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*/ | 474 | /*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*=*/ |
@@ -532,7 +520,7 @@ static void stop_this_cpu(void *dummy) | |||
532 | /* | 520 | /* |
533 | * Remove this CPU: | 521 | * Remove this CPU: |
534 | */ | 522 | */ |
535 | cpu_clear(cpu_id, cpu_online_map); | 523 | set_cpu_online(cpu_id, false); |
536 | 524 | ||
537 | /* | 525 | /* |
538 | * PSW IE = 1; | 526 | * PSW IE = 1; |
@@ -727,8 +715,8 @@ static void send_IPI_allbutself(int ipi_num, int try) | |||
727 | { | 715 | { |
728 | cpumask_t cpumask; | 716 | cpumask_t cpumask; |
729 | 717 | ||
730 | cpumask = cpu_online_map; | 718 | cpumask_copy(&cpumask, cpu_online_mask); |
731 | cpu_clear(smp_processor_id(), cpumask); | 719 | cpumask_clear_cpu(smp_processor_id(), &cpumask); |
732 | 720 | ||
733 | send_IPI_mask(&cpumask, ipi_num, try); | 721 | send_IPI_mask(&cpumask, ipi_num, try); |
734 | } | 722 | } |
@@ -765,13 +753,13 @@ static void send_IPI_mask(const struct cpumask *cpumask, int ipi_num, int try) | |||
765 | cpumask_and(&tmp, cpumask, cpu_online_mask); | 753 | cpumask_and(&tmp, cpumask, cpu_online_mask); |
766 | BUG_ON(!cpumask_equal(cpumask, &tmp)); | 754 | BUG_ON(!cpumask_equal(cpumask, &tmp)); |
767 | 755 | ||
768 | physid_mask = CPU_MASK_NONE; | 756 | cpumask_clear(&physid_mask); |
769 | for_each_cpu(cpu_id, cpumask) { | 757 | for_each_cpu(cpu_id, cpumask) { |
770 | if ((phys_id = cpu_to_physid(cpu_id)) != -1) | 758 | if ((phys_id = cpu_to_physid(cpu_id)) != -1) |
771 | cpu_set(phys_id, physid_mask); | 759 | cpumask_set_cpu(phys_id, &physid_mask); |
772 | } | 760 | } |
773 | 761 | ||
774 | send_IPI_mask_phys(physid_mask, ipi_num, try); | 762 | send_IPI_mask_phys(&physid_mask, ipi_num, try); |
775 | } | 763 | } |
776 | 764 | ||
777 | /*==========================================================================* | 765 | /*==========================================================================* |
@@ -794,14 +782,14 @@ static void send_IPI_mask(const struct cpumask *cpumask, int ipi_num, int try) | |||
794 | * ---------- --- -------------------------------------------------------- | 782 | * ---------- --- -------------------------------------------------------- |
795 | * | 783 | * |
796 | *==========================================================================*/ | 784 | *==========================================================================*/ |
797 | unsigned long send_IPI_mask_phys(cpumask_t physid_mask, int ipi_num, | 785 | unsigned long send_IPI_mask_phys(const cpumask_t *physid_mask, int ipi_num, |
798 | int try) | 786 | int try) |
799 | { | 787 | { |
800 | spinlock_t *ipilock; | 788 | spinlock_t *ipilock; |
801 | volatile unsigned long *ipicr_addr; | 789 | volatile unsigned long *ipicr_addr; |
802 | unsigned long ipicr_val; | 790 | unsigned long ipicr_val; |
803 | unsigned long my_physid_mask; | 791 | unsigned long my_physid_mask; |
804 | unsigned long mask = cpus_addr(physid_mask)[0]; | 792 | unsigned long mask = cpumask_bits(physid_mask)[0]; |
805 | 793 | ||
806 | 794 | ||
807 | if (mask & ~physids_coerce(phys_cpu_present_map)) | 795 | if (mask & ~physids_coerce(phys_cpu_present_map)) |