diff options
| -rw-r--r-- | arch/sparc64/kernel/smp.c | 31 | ||||
| -rw-r--r-- | include/asm-sparc64/mmu_context.h | 46 | ||||
| -rw-r--r-- | kernel/fork.c | 7 |
3 files changed, 29 insertions, 55 deletions
diff --git a/arch/sparc64/kernel/smp.c b/arch/sparc64/kernel/smp.c index b137fd63f5e1..a9089e2140e9 100644 --- a/arch/sparc64/kernel/smp.c +++ b/arch/sparc64/kernel/smp.c | |||
| @@ -883,34 +883,13 @@ void smp_flush_tlb_pending(struct mm_struct *mm, unsigned long nr, unsigned long | |||
| 883 | u32 ctx = CTX_HWBITS(mm->context); | 883 | u32 ctx = CTX_HWBITS(mm->context); |
| 884 | int cpu = get_cpu(); | 884 | int cpu = get_cpu(); |
| 885 | 885 | ||
| 886 | if (mm == current->active_mm && atomic_read(&mm->mm_users) == 1) { | 886 | if (mm == current->active_mm && atomic_read(&mm->mm_users) == 1) |
| 887 | mm->cpu_vm_mask = cpumask_of_cpu(cpu); | 887 | mm->cpu_vm_mask = cpumask_of_cpu(cpu); |
| 888 | goto local_flush_and_out; | 888 | else |
| 889 | } else { | 889 | smp_cross_call_masked(&xcall_flush_tlb_pending, |
| 890 | /* This optimization is not valid. Normally | 890 | ctx, nr, (unsigned long) vaddrs, |
| 891 | * we will be holding the page_table_lock, but | 891 | mm->cpu_vm_mask); |
| 892 | * there is an exception which is copy_page_range() | ||
| 893 | * when forking. The lock is held during the individual | ||
| 894 | * page table updates in the parent, but not at the | ||
| 895 | * top level, which is where we are invoked. | ||
| 896 | */ | ||
| 897 | if (0) { | ||
| 898 | cpumask_t this_cpu_mask = cpumask_of_cpu(cpu); | ||
| 899 | |||
| 900 | /* By virtue of running under the mm->page_table_lock, | ||
| 901 | * and mmu_context.h:switch_mm doing the same, the | ||
| 902 | * following operation is safe. | ||
| 903 | */ | ||
| 904 | if (cpus_equal(mm->cpu_vm_mask, this_cpu_mask)) | ||
| 905 | goto local_flush_and_out; | ||
| 906 | } | ||
| 907 | } | ||
| 908 | |||
| 909 | smp_cross_call_masked(&xcall_flush_tlb_pending, | ||
| 910 | ctx, nr, (unsigned long) vaddrs, | ||
| 911 | mm->cpu_vm_mask); | ||
| 912 | 892 | ||
| 913 | local_flush_and_out: | ||
| 914 | __flush_tlb_pending(ctx, nr, vaddrs); | 893 | __flush_tlb_pending(ctx, nr, vaddrs); |
| 915 | 894 | ||
| 916 | put_cpu(); | 895 | put_cpu(); |
diff --git a/include/asm-sparc64/mmu_context.h b/include/asm-sparc64/mmu_context.h index 87c43c67866e..08ba72d7722c 100644 --- a/include/asm-sparc64/mmu_context.h +++ b/include/asm-sparc64/mmu_context.h | |||
| @@ -87,37 +87,35 @@ extern void __flush_tlb_mm(unsigned long, unsigned long); | |||
| 87 | static inline void switch_mm(struct mm_struct *old_mm, struct mm_struct *mm, struct task_struct *tsk) | 87 | static inline void switch_mm(struct mm_struct *old_mm, struct mm_struct *mm, struct task_struct *tsk) |
| 88 | { | 88 | { |
| 89 | unsigned long ctx_valid; | 89 | unsigned long ctx_valid; |
| 90 | int cpu; | ||
| 90 | 91 | ||
| 92 | /* Note: page_table_lock is used here to serialize switch_mm | ||
| 93 | * and activate_mm, and their calls to get_new_mmu_context. | ||
| 94 | * This use of page_table_lock is unrelated to its other uses. | ||
| 95 | */ | ||
| 91 | spin_lock(&mm->page_table_lock); | 96 | spin_lock(&mm->page_table_lock); |
| 92 | if (CTX_VALID(mm->context)) | 97 | ctx_valid = CTX_VALID(mm->context); |
| 93 | ctx_valid = 1; | 98 | if (!ctx_valid) |
| 94 | else | 99 | get_new_mmu_context(mm); |
| 95 | ctx_valid = 0; | 100 | spin_unlock(&mm->page_table_lock); |
| 96 | 101 | ||
| 97 | if (!ctx_valid || (old_mm != mm)) { | 102 | if (!ctx_valid || (old_mm != mm)) { |
| 98 | if (!ctx_valid) | ||
| 99 | get_new_mmu_context(mm); | ||
| 100 | |||
| 101 | load_secondary_context(mm); | 103 | load_secondary_context(mm); |
| 102 | reload_tlbmiss_state(tsk, mm); | 104 | reload_tlbmiss_state(tsk, mm); |
| 103 | } | 105 | } |
| 104 | 106 | ||
| 105 | { | 107 | /* Even if (mm == old_mm) we _must_ check |
| 106 | int cpu = smp_processor_id(); | 108 | * the cpu_vm_mask. If we do not we could |
| 107 | 109 | * corrupt the TLB state because of how | |
| 108 | /* Even if (mm == old_mm) we _must_ check | 110 | * smp_flush_tlb_{page,range,mm} on sparc64 |
| 109 | * the cpu_vm_mask. If we do not we could | 111 | * and lazy tlb switches work. -DaveM |
| 110 | * corrupt the TLB state because of how | 112 | */ |
| 111 | * smp_flush_tlb_{page,range,mm} on sparc64 | 113 | cpu = smp_processor_id(); |
| 112 | * and lazy tlb switches work. -DaveM | 114 | if (!ctx_valid || !cpu_isset(cpu, mm->cpu_vm_mask)) { |
| 113 | */ | 115 | cpu_set(cpu, mm->cpu_vm_mask); |
| 114 | if (!ctx_valid || !cpu_isset(cpu, mm->cpu_vm_mask)) { | 116 | __flush_tlb_mm(CTX_HWBITS(mm->context), |
| 115 | cpu_set(cpu, mm->cpu_vm_mask); | 117 | SECONDARY_CONTEXT); |
| 116 | __flush_tlb_mm(CTX_HWBITS(mm->context), | ||
| 117 | SECONDARY_CONTEXT); | ||
| 118 | } | ||
| 119 | } | 118 | } |
| 120 | spin_unlock(&mm->page_table_lock); | ||
| 121 | } | 119 | } |
| 122 | 120 | ||
| 123 | #define deactivate_mm(tsk,mm) do { } while (0) | 121 | #define deactivate_mm(tsk,mm) do { } while (0) |
| @@ -127,6 +125,10 @@ static inline void activate_mm(struct mm_struct *active_mm, struct mm_struct *mm | |||
| 127 | { | 125 | { |
| 128 | int cpu; | 126 | int cpu; |
| 129 | 127 | ||
| 128 | /* Note: page_table_lock is used here to serialize switch_mm | ||
| 129 | * and activate_mm, and their calls to get_new_mmu_context. | ||
| 130 | * This use of page_table_lock is unrelated to its other uses. | ||
| 131 | */ | ||
| 130 | spin_lock(&mm->page_table_lock); | 132 | spin_lock(&mm->page_table_lock); |
| 131 | if (!CTX_VALID(mm->context)) | 133 | if (!CTX_VALID(mm->context)) |
| 132 | get_new_mmu_context(mm); | 134 | get_new_mmu_context(mm); |
diff --git a/kernel/fork.c b/kernel/fork.c index efac2c58ec7d..158710d22566 100644 --- a/kernel/fork.c +++ b/kernel/fork.c | |||
| @@ -470,13 +470,6 @@ static int copy_mm(unsigned long clone_flags, struct task_struct * tsk) | |||
| 470 | if (clone_flags & CLONE_VM) { | 470 | if (clone_flags & CLONE_VM) { |
| 471 | atomic_inc(&oldmm->mm_users); | 471 | atomic_inc(&oldmm->mm_users); |
| 472 | mm = oldmm; | 472 | mm = oldmm; |
| 473 | /* | ||
| 474 | * There are cases where the PTL is held to ensure no | ||
| 475 | * new threads start up in user mode using an mm, which | ||
| 476 | * allows optimizing out ipis; the tlb_gather_mmu code | ||
| 477 | * is an example. | ||
| 478 | */ | ||
| 479 | spin_unlock_wait(&oldmm->page_table_lock); | ||
| 480 | goto good_mm; | 473 | goto good_mm; |
| 481 | } | 474 | } |
| 482 | 475 | ||
