diff options
Diffstat (limited to 'arch/sh/mm/fault.c')
-rw-r--r-- | arch/sh/mm/fault.c | 86 |
1 files changed, 0 insertions, 86 deletions
diff --git a/arch/sh/mm/fault.c b/arch/sh/mm/fault.c index 128907ef7fcd..123fb80c859d 100644 --- a/arch/sh/mm/fault.c +++ b/arch/sh/mm/fault.c | |||
@@ -223,89 +223,3 @@ do_sigbus: | |||
223 | if (!user_mode(regs)) | 223 | if (!user_mode(regs)) |
224 | goto no_context; | 224 | goto no_context; |
225 | } | 225 | } |
226 | |||
227 | #ifdef CONFIG_SH_STORE_QUEUES | ||
228 | /* | ||
229 | * This is a special case for the SH-4 store queues, as pages for this | ||
230 | * space still need to be faulted in before it's possible to flush the | ||
231 | * store queue cache for writeout to the remapped region. | ||
232 | */ | ||
233 | #define P3_ADDR_MAX (P4SEG_STORE_QUE + 0x04000000) | ||
234 | #else | ||
235 | #define P3_ADDR_MAX P4SEG | ||
236 | #endif | ||
237 | |||
238 | /* | ||
239 | * Called with interrupts disabled. | ||
240 | */ | ||
241 | asmlinkage int __kprobes __do_page_fault(struct pt_regs *regs, | ||
242 | unsigned long writeaccess, | ||
243 | unsigned long address) | ||
244 | { | ||
245 | pgd_t *pgd; | ||
246 | pud_t *pud; | ||
247 | pmd_t *pmd; | ||
248 | pte_t *pte; | ||
249 | pte_t entry; | ||
250 | struct mm_struct *mm = current->mm; | ||
251 | spinlock_t *ptl; | ||
252 | int ret = 1; | ||
253 | |||
254 | #ifdef CONFIG_SH_KGDB | ||
255 | if (kgdb_nofault && kgdb_bus_err_hook) | ||
256 | kgdb_bus_err_hook(); | ||
257 | #endif | ||
258 | |||
259 | /* | ||
260 | * We don't take page faults for P1, P2, and parts of P4, these | ||
261 | * are always mapped, whether it be due to legacy behaviour in | ||
262 | * 29-bit mode, or due to PMB configuration in 32-bit mode. | ||
263 | */ | ||
264 | if (address >= P3SEG && address < P3_ADDR_MAX) { | ||
265 | pgd = pgd_offset_k(address); | ||
266 | mm = NULL; | ||
267 | } else { | ||
268 | if (unlikely(address >= TASK_SIZE || !mm)) | ||
269 | return 1; | ||
270 | |||
271 | pgd = pgd_offset(mm, address); | ||
272 | } | ||
273 | |||
274 | pud = pud_offset(pgd, address); | ||
275 | if (pud_none_or_clear_bad(pud)) | ||
276 | return 1; | ||
277 | pmd = pmd_offset(pud, address); | ||
278 | if (pmd_none_or_clear_bad(pmd)) | ||
279 | return 1; | ||
280 | |||
281 | if (mm) | ||
282 | pte = pte_offset_map_lock(mm, pmd, address, &ptl); | ||
283 | else | ||
284 | pte = pte_offset_kernel(pmd, address); | ||
285 | |||
286 | entry = *pte; | ||
287 | if (unlikely(pte_none(entry) || pte_not_present(entry))) | ||
288 | goto unlock; | ||
289 | if (unlikely(writeaccess && !pte_write(entry))) | ||
290 | goto unlock; | ||
291 | |||
292 | if (writeaccess) | ||
293 | entry = pte_mkdirty(entry); | ||
294 | entry = pte_mkyoung(entry); | ||
295 | |||
296 | #ifdef CONFIG_CPU_SH4 | ||
297 | /* | ||
298 | * ITLB is not affected by "ldtlb" instruction. | ||
299 | * So, we need to flush the entry by ourselves. | ||
300 | */ | ||
301 | __flush_tlb_page(get_asid(), address & PAGE_MASK); | ||
302 | #endif | ||
303 | |||
304 | set_pte(pte, entry); | ||
305 | update_mmu_cache(NULL, address, entry); | ||
306 | ret = 0; | ||
307 | unlock: | ||
308 | if (mm) | ||
309 | pte_unmap_unlock(pte, ptl); | ||
310 | return ret; | ||
311 | } | ||