diff options
Diffstat (limited to 'arch/m32r')
| -rw-r--r-- | arch/m32r/include/asm/local.h | 25 | ||||
| -rw-r--r-- | arch/m32r/include/asm/tlbflush.h | 2 | ||||
| -rw-r--r-- | arch/m32r/mm/fault-nommu.c | 2 | ||||
| -rw-r--r-- | arch/m32r/mm/fault.c | 6 |
4 files changed, 5 insertions, 30 deletions
diff --git a/arch/m32r/include/asm/local.h b/arch/m32r/include/asm/local.h index 22256d138630..734bca87018a 100644 --- a/arch/m32r/include/asm/local.h +++ b/arch/m32r/include/asm/local.h | |||
| @@ -338,29 +338,4 @@ static inline void local_set_mask(unsigned long mask, local_t *addr) | |||
| 338 | * a variable, not an address. | 338 | * a variable, not an address. |
| 339 | */ | 339 | */ |
| 340 | 340 | ||
| 341 | /* Need to disable preemption for the cpu local counters otherwise we could | ||
| 342 | still access a variable of a previous CPU in a non local way. */ | ||
| 343 | #define cpu_local_wrap_v(l) \ | ||
| 344 | ({ local_t res__; \ | ||
| 345 | preempt_disable(); \ | ||
| 346 | res__ = (l); \ | ||
| 347 | preempt_enable(); \ | ||
| 348 | res__; }) | ||
| 349 | #define cpu_local_wrap(l) \ | ||
| 350 | ({ preempt_disable(); \ | ||
| 351 | l; \ | ||
| 352 | preempt_enable(); }) \ | ||
| 353 | |||
| 354 | #define cpu_local_read(l) cpu_local_wrap_v(local_read(&__get_cpu_var(l))) | ||
| 355 | #define cpu_local_set(l, i) cpu_local_wrap(local_set(&__get_cpu_var(l), (i))) | ||
| 356 | #define cpu_local_inc(l) cpu_local_wrap(local_inc(&__get_cpu_var(l))) | ||
| 357 | #define cpu_local_dec(l) cpu_local_wrap(local_dec(&__get_cpu_var(l))) | ||
| 358 | #define cpu_local_add(i, l) cpu_local_wrap(local_add((i), &__get_cpu_var(l))) | ||
| 359 | #define cpu_local_sub(i, l) cpu_local_wrap(local_sub((i), &__get_cpu_var(l))) | ||
| 360 | |||
| 361 | #define __cpu_local_inc(l) cpu_local_inc(l) | ||
| 362 | #define __cpu_local_dec(l) cpu_local_dec(l) | ||
| 363 | #define __cpu_local_add(i, l) cpu_local_add((i), (l)) | ||
| 364 | #define __cpu_local_sub(i, l) cpu_local_sub((i), (l)) | ||
| 365 | |||
| 366 | #endif /* __M32R_LOCAL_H */ | 341 | #endif /* __M32R_LOCAL_H */ |
diff --git a/arch/m32r/include/asm/tlbflush.h b/arch/m32r/include/asm/tlbflush.h index 0ef95307784e..92614b0ccf17 100644 --- a/arch/m32r/include/asm/tlbflush.h +++ b/arch/m32r/include/asm/tlbflush.h | |||
| @@ -92,6 +92,6 @@ static __inline__ void __flush_tlb_all(void) | |||
| 92 | ); | 92 | ); |
| 93 | } | 93 | } |
| 94 | 94 | ||
| 95 | extern void update_mmu_cache(struct vm_area_struct *, unsigned long, pte_t); | 95 | extern void update_mmu_cache(struct vm_area_struct *, unsigned long, pte_t *); |
| 96 | 96 | ||
| 97 | #endif /* _ASM_M32R_TLBFLUSH_H */ | 97 | #endif /* _ASM_M32R_TLBFLUSH_H */ |
diff --git a/arch/m32r/mm/fault-nommu.c b/arch/m32r/mm/fault-nommu.c index 88469178ea6b..888aab1157ed 100644 --- a/arch/m32r/mm/fault-nommu.c +++ b/arch/m32r/mm/fault-nommu.c | |||
| @@ -95,7 +95,7 @@ asmlinkage void do_page_fault(struct pt_regs *regs, unsigned long error_code, | |||
| 95 | * update_mmu_cache() | 95 | * update_mmu_cache() |
| 96 | *======================================================================*/ | 96 | *======================================================================*/ |
| 97 | void update_mmu_cache(struct vm_area_struct *vma, unsigned long addr, | 97 | void update_mmu_cache(struct vm_area_struct *vma, unsigned long addr, |
| 98 | pte_t pte) | 98 | pte_t *ptep) |
| 99 | { | 99 | { |
| 100 | BUG(); | 100 | BUG(); |
| 101 | } | 101 | } |
diff --git a/arch/m32r/mm/fault.c b/arch/m32r/mm/fault.c index 7274b47f4c22..28ee389e5f5a 100644 --- a/arch/m32r/mm/fault.c +++ b/arch/m32r/mm/fault.c | |||
| @@ -336,7 +336,7 @@ vmalloc_fault: | |||
| 336 | 336 | ||
| 337 | addr = (address & PAGE_MASK); | 337 | addr = (address & PAGE_MASK); |
| 338 | set_thread_fault_code(error_code); | 338 | set_thread_fault_code(error_code); |
| 339 | update_mmu_cache(NULL, addr, *pte_k); | 339 | update_mmu_cache(NULL, addr, pte_k); |
| 340 | set_thread_fault_code(0); | 340 | set_thread_fault_code(0); |
| 341 | return; | 341 | return; |
| 342 | } | 342 | } |
| @@ -349,7 +349,7 @@ vmalloc_fault: | |||
| 349 | #define ITLB_END (unsigned long *)(ITLB_BASE + (NR_TLB_ENTRIES * 8)) | 349 | #define ITLB_END (unsigned long *)(ITLB_BASE + (NR_TLB_ENTRIES * 8)) |
| 350 | #define DTLB_END (unsigned long *)(DTLB_BASE + (NR_TLB_ENTRIES * 8)) | 350 | #define DTLB_END (unsigned long *)(DTLB_BASE + (NR_TLB_ENTRIES * 8)) |
| 351 | void update_mmu_cache(struct vm_area_struct *vma, unsigned long vaddr, | 351 | void update_mmu_cache(struct vm_area_struct *vma, unsigned long vaddr, |
| 352 | pte_t pte) | 352 | pte_t *ptep) |
| 353 | { | 353 | { |
| 354 | volatile unsigned long *entry1, *entry2; | 354 | volatile unsigned long *entry1, *entry2; |
| 355 | unsigned long pte_data, flags; | 355 | unsigned long pte_data, flags; |
| @@ -365,7 +365,7 @@ void update_mmu_cache(struct vm_area_struct *vma, unsigned long vaddr, | |||
| 365 | 365 | ||
| 366 | vaddr = (vaddr & PAGE_MASK) | get_asid(); | 366 | vaddr = (vaddr & PAGE_MASK) | get_asid(); |
| 367 | 367 | ||
| 368 | pte_data = pte_val(pte); | 368 | pte_data = pte_val(*ptep); |
| 369 | 369 | ||
| 370 | #ifdef CONFIG_CHIP_OPSP | 370 | #ifdef CONFIG_CHIP_OPSP |
| 371 | entry1 = (unsigned long *)ITLB_BASE; | 371 | entry1 = (unsigned long *)ITLB_BASE; |
