diff options
author | Rusty Russell <rusty@rustcorp.com.au> | 2009-09-24 11:34:49 -0400 |
---|---|---|
committer | Rusty Russell <rusty@rustcorp.com.au> | 2009-09-23 20:04:49 -0400 |
commit | 56f8ba83a52b9f9e3711eff8e54168ac14aa288f (patch) | |
tree | e030f7f3a191384268d86863ca43237a137e8f51 /arch/arm/kernel | |
parent | a6a01063de6298c60f2506dc7659403e02b4b224 (diff) |
cpumask: use mm_cpumask() wrapper: arm
Makes code futureproof against the impending change to mm->cpu_vm_mask.
It's also a chance to use the new cpumask_ ops which take a pointer
(the older ones are deprecated, but there's no hurry for arch code).
Signed-off-by: Rusty Russell <rusty@rustcorp.com.au>
Diffstat (limited to 'arch/arm/kernel')
-rw-r--r-- | arch/arm/kernel/smp.c | 10 |
1 files changed, 5 insertions, 5 deletions
diff --git a/arch/arm/kernel/smp.c b/arch/arm/kernel/smp.c index de885fd256c5..e0d32770bb3d 100644 --- a/arch/arm/kernel/smp.c +++ b/arch/arm/kernel/smp.c | |||
@@ -189,7 +189,7 @@ int __cpuexit __cpu_disable(void) | |||
189 | read_lock(&tasklist_lock); | 189 | read_lock(&tasklist_lock); |
190 | for_each_process(p) { | 190 | for_each_process(p) { |
191 | if (p->mm) | 191 | if (p->mm) |
192 | cpu_clear(cpu, p->mm->cpu_vm_mask); | 192 | cpumask_clear_cpu(cpu, mm_cpumask(p->mm)); |
193 | } | 193 | } |
194 | read_unlock(&tasklist_lock); | 194 | read_unlock(&tasklist_lock); |
195 | 195 | ||
@@ -257,7 +257,7 @@ asmlinkage void __cpuinit secondary_start_kernel(void) | |||
257 | atomic_inc(&mm->mm_users); | 257 | atomic_inc(&mm->mm_users); |
258 | atomic_inc(&mm->mm_count); | 258 | atomic_inc(&mm->mm_count); |
259 | current->active_mm = mm; | 259 | current->active_mm = mm; |
260 | cpu_set(cpu, mm->cpu_vm_mask); | 260 | cpumask_set_cpu(cpu, mm_cpumask(mm)); |
261 | cpu_switch_mm(mm->pgd, mm); | 261 | cpu_switch_mm(mm->pgd, mm); |
262 | enter_lazy_tlb(mm, current); | 262 | enter_lazy_tlb(mm, current); |
263 | local_flush_tlb_all(); | 263 | local_flush_tlb_all(); |
@@ -643,7 +643,7 @@ void flush_tlb_all(void) | |||
643 | void flush_tlb_mm(struct mm_struct *mm) | 643 | void flush_tlb_mm(struct mm_struct *mm) |
644 | { | 644 | { |
645 | if (tlb_ops_need_broadcast()) | 645 | if (tlb_ops_need_broadcast()) |
646 | on_each_cpu_mask(ipi_flush_tlb_mm, mm, 1, &mm->cpu_vm_mask); | 646 | on_each_cpu_mask(ipi_flush_tlb_mm, mm, 1, mm_cpumask(mm)); |
647 | else | 647 | else |
648 | local_flush_tlb_mm(mm); | 648 | local_flush_tlb_mm(mm); |
649 | } | 649 | } |
@@ -654,7 +654,7 @@ void flush_tlb_page(struct vm_area_struct *vma, unsigned long uaddr) | |||
654 | struct tlb_args ta; | 654 | struct tlb_args ta; |
655 | ta.ta_vma = vma; | 655 | ta.ta_vma = vma; |
656 | ta.ta_start = uaddr; | 656 | ta.ta_start = uaddr; |
657 | on_each_cpu_mask(ipi_flush_tlb_page, &ta, 1, &vma->vm_mm->cpu_vm_mask); | 657 | on_each_cpu_mask(ipi_flush_tlb_page, &ta, 1, mm_cpumask(vma->vm_mm)); |
658 | } else | 658 | } else |
659 | local_flush_tlb_page(vma, uaddr); | 659 | local_flush_tlb_page(vma, uaddr); |
660 | } | 660 | } |
@@ -677,7 +677,7 @@ void flush_tlb_range(struct vm_area_struct *vma, | |||
677 | ta.ta_vma = vma; | 677 | ta.ta_vma = vma; |
678 | ta.ta_start = start; | 678 | ta.ta_start = start; |
679 | ta.ta_end = end; | 679 | ta.ta_end = end; |
680 | on_each_cpu_mask(ipi_flush_tlb_range, &ta, 1, &vma->vm_mm->cpu_vm_mask); | 680 | on_each_cpu_mask(ipi_flush_tlb_range, &ta, 1, mm_cpumask(vma->vm_mm)); |
681 | } else | 681 | } else |
682 | local_flush_tlb_range(vma, start, end); | 682 | local_flush_tlb_range(vma, start, end); |
683 | } | 683 | } |