diff options
author | Ingo Molnar <mingo@elte.hu> | 2009-01-28 17:12:55 -0500 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2009-01-28 17:12:55 -0500 |
commit | 6a385db5ce7f1fd2c68ec511e44587b67dab8fca (patch) | |
tree | 9324c8ae6f7be54b9fdbd6b60f759292aa727b1f /arch/x86/mm/fault.c | |
parent | 18e352e4a73465349711a9324767e1b2453383e2 (diff) | |
parent | 4369f1fb7cd4cf777312f43e1cb9aa5504fc4125 (diff) |
Merge branch 'core/percpu' into x86/core
Conflicts:
kernel/irq/handle.c
Diffstat (limited to 'arch/x86/mm/fault.c')
-rw-r--r-- | arch/x86/mm/fault.c | 445 |
1 files changed, 263 insertions, 182 deletions
diff --git a/arch/x86/mm/fault.c b/arch/x86/mm/fault.c index 90dfae511a41..65709a6aa6ee 100644 --- a/arch/x86/mm/fault.c +++ b/arch/x86/mm/fault.c | |||
@@ -26,6 +26,7 @@ | |||
26 | #include <linux/kprobes.h> | 26 | #include <linux/kprobes.h> |
27 | #include <linux/uaccess.h> | 27 | #include <linux/uaccess.h> |
28 | #include <linux/kdebug.h> | 28 | #include <linux/kdebug.h> |
29 | #include <linux/magic.h> | ||
29 | 30 | ||
30 | #include <asm/system.h> | 31 | #include <asm/system.h> |
31 | #include <asm/desc.h> | 32 | #include <asm/desc.h> |
@@ -91,8 +92,8 @@ static inline int notify_page_fault(struct pt_regs *regs) | |||
91 | * | 92 | * |
92 | * Opcode checker based on code by Richard Brunner | 93 | * Opcode checker based on code by Richard Brunner |
93 | */ | 94 | */ |
94 | static int is_prefetch(struct pt_regs *regs, unsigned long addr, | 95 | static int is_prefetch(struct pt_regs *regs, unsigned long error_code, |
95 | unsigned long error_code) | 96 | unsigned long addr) |
96 | { | 97 | { |
97 | unsigned char *instr; | 98 | unsigned char *instr; |
98 | int scan_more = 1; | 99 | int scan_more = 1; |
@@ -409,15 +410,15 @@ static void show_fault_oops(struct pt_regs *regs, unsigned long error_code, | |||
409 | } | 410 | } |
410 | 411 | ||
411 | #ifdef CONFIG_X86_64 | 412 | #ifdef CONFIG_X86_64 |
412 | static noinline void pgtable_bad(unsigned long address, struct pt_regs *regs, | 413 | static noinline void pgtable_bad(struct pt_regs *regs, |
413 | unsigned long error_code) | 414 | unsigned long error_code, unsigned long address) |
414 | { | 415 | { |
415 | unsigned long flags = oops_begin(); | 416 | unsigned long flags = oops_begin(); |
416 | int sig = SIGKILL; | 417 | int sig = SIGKILL; |
417 | struct task_struct *tsk; | 418 | struct task_struct *tsk = current; |
418 | 419 | ||
419 | printk(KERN_ALERT "%s: Corrupted page table at address %lx\n", | 420 | printk(KERN_ALERT "%s: Corrupted page table at address %lx\n", |
420 | current->comm, address); | 421 | tsk->comm, address); |
421 | dump_pagetable(address); | 422 | dump_pagetable(address); |
422 | tsk = current; | 423 | tsk = current; |
423 | tsk->thread.cr2 = address; | 424 | tsk->thread.cr2 = address; |
@@ -429,6 +430,196 @@ static noinline void pgtable_bad(unsigned long address, struct pt_regs *regs, | |||
429 | } | 430 | } |
430 | #endif | 431 | #endif |
431 | 432 | ||
433 | static noinline void no_context(struct pt_regs *regs, | ||
434 | unsigned long error_code, unsigned long address) | ||
435 | { | ||
436 | struct task_struct *tsk = current; | ||
437 | unsigned long *stackend; | ||
438 | |||
439 | #ifdef CONFIG_X86_64 | ||
440 | unsigned long flags; | ||
441 | int sig; | ||
442 | #endif | ||
443 | |||
444 | /* Are we prepared to handle this kernel fault? */ | ||
445 | if (fixup_exception(regs)) | ||
446 | return; | ||
447 | |||
448 | /* | ||
449 | * X86_32 | ||
450 | * Valid to do another page fault here, because if this fault | ||
451 | * had been triggered by is_prefetch fixup_exception would have | ||
452 | * handled it. | ||
453 | * | ||
454 | * X86_64 | ||
455 | * Hall of shame of CPU/BIOS bugs. | ||
456 | */ | ||
457 | if (is_prefetch(regs, error_code, address)) | ||
458 | return; | ||
459 | |||
460 | if (is_errata93(regs, address)) | ||
461 | return; | ||
462 | |||
463 | /* | ||
464 | * Oops. The kernel tried to access some bad page. We'll have to | ||
465 | * terminate things with extreme prejudice. | ||
466 | */ | ||
467 | #ifdef CONFIG_X86_32 | ||
468 | bust_spinlocks(1); | ||
469 | #else | ||
470 | flags = oops_begin(); | ||
471 | #endif | ||
472 | |||
473 | show_fault_oops(regs, error_code, address); | ||
474 | |||
475 | stackend = end_of_stack(tsk); | ||
476 | if (*stackend != STACK_END_MAGIC) | ||
477 | printk(KERN_ALERT "Thread overran stack, or stack corrupted\n"); | ||
478 | |||
479 | tsk->thread.cr2 = address; | ||
480 | tsk->thread.trap_no = 14; | ||
481 | tsk->thread.error_code = error_code; | ||
482 | |||
483 | #ifdef CONFIG_X86_32 | ||
484 | die("Oops", regs, error_code); | ||
485 | bust_spinlocks(0); | ||
486 | do_exit(SIGKILL); | ||
487 | #else | ||
488 | sig = SIGKILL; | ||
489 | if (__die("Oops", regs, error_code)) | ||
490 | sig = 0; | ||
491 | /* Executive summary in case the body of the oops scrolled away */ | ||
492 | printk(KERN_EMERG "CR2: %016lx\n", address); | ||
493 | oops_end(flags, regs, sig); | ||
494 | #endif | ||
495 | } | ||
496 | |||
497 | static void __bad_area_nosemaphore(struct pt_regs *regs, | ||
498 | unsigned long error_code, unsigned long address, | ||
499 | int si_code) | ||
500 | { | ||
501 | struct task_struct *tsk = current; | ||
502 | |||
503 | /* User mode accesses just cause a SIGSEGV */ | ||
504 | if (error_code & PF_USER) { | ||
505 | /* | ||
506 | * It's possible to have interrupts off here. | ||
507 | */ | ||
508 | local_irq_enable(); | ||
509 | |||
510 | /* | ||
511 | * Valid to do another page fault here because this one came | ||
512 | * from user space. | ||
513 | */ | ||
514 | if (is_prefetch(regs, error_code, address)) | ||
515 | return; | ||
516 | |||
517 | if (is_errata100(regs, address)) | ||
518 | return; | ||
519 | |||
520 | if (show_unhandled_signals && unhandled_signal(tsk, SIGSEGV) && | ||
521 | printk_ratelimit()) { | ||
522 | printk( | ||
523 | "%s%s[%d]: segfault at %lx ip %p sp %p error %lx", | ||
524 | task_pid_nr(tsk) > 1 ? KERN_INFO : KERN_EMERG, | ||
525 | tsk->comm, task_pid_nr(tsk), address, | ||
526 | (void *) regs->ip, (void *) regs->sp, error_code); | ||
527 | print_vma_addr(" in ", regs->ip); | ||
528 | printk("\n"); | ||
529 | } | ||
530 | |||
531 | tsk->thread.cr2 = address; | ||
532 | /* Kernel addresses are always protection faults */ | ||
533 | tsk->thread.error_code = error_code | (address >= TASK_SIZE); | ||
534 | tsk->thread.trap_no = 14; | ||
535 | force_sig_info_fault(SIGSEGV, si_code, address, tsk); | ||
536 | return; | ||
537 | } | ||
538 | |||
539 | if (is_f00f_bug(regs, address)) | ||
540 | return; | ||
541 | |||
542 | no_context(regs, error_code, address); | ||
543 | } | ||
544 | |||
545 | static noinline void bad_area_nosemaphore(struct pt_regs *regs, | ||
546 | unsigned long error_code, unsigned long address) | ||
547 | { | ||
548 | __bad_area_nosemaphore(regs, error_code, address, SEGV_MAPERR); | ||
549 | } | ||
550 | |||
551 | static void __bad_area(struct pt_regs *regs, | ||
552 | unsigned long error_code, unsigned long address, | ||
553 | int si_code) | ||
554 | { | ||
555 | struct mm_struct *mm = current->mm; | ||
556 | |||
557 | /* | ||
558 | * Something tried to access memory that isn't in our memory map.. | ||
559 | * Fix it, but check if it's kernel or user first.. | ||
560 | */ | ||
561 | up_read(&mm->mmap_sem); | ||
562 | |||
563 | __bad_area_nosemaphore(regs, error_code, address, si_code); | ||
564 | } | ||
565 | |||
566 | static noinline void bad_area(struct pt_regs *regs, | ||
567 | unsigned long error_code, unsigned long address) | ||
568 | { | ||
569 | __bad_area(regs, error_code, address, SEGV_MAPERR); | ||
570 | } | ||
571 | |||
572 | static noinline void bad_area_access_error(struct pt_regs *regs, | ||
573 | unsigned long error_code, unsigned long address) | ||
574 | { | ||
575 | __bad_area(regs, error_code, address, SEGV_ACCERR); | ||
576 | } | ||
577 | |||
578 | /* TODO: fixup for "mm-invoke-oom-killer-from-page-fault.patch" */ | ||
579 | static void out_of_memory(struct pt_regs *regs, | ||
580 | unsigned long error_code, unsigned long address) | ||
581 | { | ||
582 | /* | ||
583 | * We ran out of memory, call the OOM killer, and return the userspace | ||
584 | * (which will retry the fault, or kill us if we got oom-killed). | ||
585 | */ | ||
586 | up_read(¤t->mm->mmap_sem); | ||
587 | pagefault_out_of_memory(); | ||
588 | } | ||
589 | |||
590 | static void do_sigbus(struct pt_regs *regs, | ||
591 | unsigned long error_code, unsigned long address) | ||
592 | { | ||
593 | struct task_struct *tsk = current; | ||
594 | struct mm_struct *mm = tsk->mm; | ||
595 | |||
596 | up_read(&mm->mmap_sem); | ||
597 | |||
598 | /* Kernel mode? Handle exceptions or die */ | ||
599 | if (!(error_code & PF_USER)) | ||
600 | no_context(regs, error_code, address); | ||
601 | #ifdef CONFIG_X86_32 | ||
602 | /* User space => ok to do another page fault */ | ||
603 | if (is_prefetch(regs, error_code, address)) | ||
604 | return; | ||
605 | #endif | ||
606 | tsk->thread.cr2 = address; | ||
607 | tsk->thread.error_code = error_code; | ||
608 | tsk->thread.trap_no = 14; | ||
609 | force_sig_info_fault(SIGBUS, BUS_ADRERR, address, tsk); | ||
610 | } | ||
611 | |||
612 | static noinline void mm_fault_error(struct pt_regs *regs, | ||
613 | unsigned long error_code, unsigned long address, unsigned int fault) | ||
614 | { | ||
615 | if (fault & VM_FAULT_OOM) | ||
616 | out_of_memory(regs, error_code, address); | ||
617 | else if (fault & VM_FAULT_SIGBUS) | ||
618 | do_sigbus(regs, error_code, address); | ||
619 | else | ||
620 | BUG(); | ||
621 | } | ||
622 | |||
432 | static int spurious_fault_check(unsigned long error_code, pte_t *pte) | 623 | static int spurious_fault_check(unsigned long error_code, pte_t *pte) |
433 | { | 624 | { |
434 | if ((error_code & PF_WRITE) && !pte_write(*pte)) | 625 | if ((error_code & PF_WRITE) && !pte_write(*pte)) |
@@ -448,8 +639,8 @@ static int spurious_fault_check(unsigned long error_code, pte_t *pte) | |||
448 | * There are no security implications to leaving a stale TLB when | 639 | * There are no security implications to leaving a stale TLB when |
449 | * increasing the permissions on a page. | 640 | * increasing the permissions on a page. |
450 | */ | 641 | */ |
451 | static int spurious_fault(unsigned long address, | 642 | static noinline int spurious_fault(unsigned long error_code, |
452 | unsigned long error_code) | 643 | unsigned long address) |
453 | { | 644 | { |
454 | pgd_t *pgd; | 645 | pgd_t *pgd; |
455 | pud_t *pud; | 646 | pud_t *pud; |
@@ -494,7 +685,7 @@ static int spurious_fault(unsigned long address, | |||
494 | * | 685 | * |
495 | * This assumes no large pages in there. | 686 | * This assumes no large pages in there. |
496 | */ | 687 | */ |
497 | static int vmalloc_fault(unsigned long address) | 688 | static noinline int vmalloc_fault(unsigned long address) |
498 | { | 689 | { |
499 | #ifdef CONFIG_X86_32 | 690 | #ifdef CONFIG_X86_32 |
500 | unsigned long pgd_paddr; | 691 | unsigned long pgd_paddr; |
@@ -573,6 +764,25 @@ static int vmalloc_fault(unsigned long address) | |||
573 | 764 | ||
574 | int show_unhandled_signals = 1; | 765 | int show_unhandled_signals = 1; |
575 | 766 | ||
767 | static inline int access_error(unsigned long error_code, int write, | ||
768 | struct vm_area_struct *vma) | ||
769 | { | ||
770 | if (write) { | ||
771 | /* write, present and write, not present */ | ||
772 | if (unlikely(!(vma->vm_flags & VM_WRITE))) | ||
773 | return 1; | ||
774 | } else if (unlikely(error_code & PF_PROT)) { | ||
775 | /* read, present */ | ||
776 | return 1; | ||
777 | } else { | ||
778 | /* read, not present */ | ||
779 | if (unlikely(!(vma->vm_flags & (VM_READ | VM_EXEC | VM_WRITE)))) | ||
780 | return 1; | ||
781 | } | ||
782 | |||
783 | return 0; | ||
784 | } | ||
785 | |||
576 | /* | 786 | /* |
577 | * This routine handles page faults. It determines the address, | 787 | * This routine handles page faults. It determines the address, |
578 | * and the problem, and then passes it off to one of the appropriate | 788 | * and the problem, and then passes it off to one of the appropriate |
@@ -583,16 +793,12 @@ asmlinkage | |||
583 | #endif | 793 | #endif |
584 | void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code) | 794 | void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code) |
585 | { | 795 | { |
796 | unsigned long address; | ||
586 | struct task_struct *tsk; | 797 | struct task_struct *tsk; |
587 | struct mm_struct *mm; | 798 | struct mm_struct *mm; |
588 | struct vm_area_struct *vma; | 799 | struct vm_area_struct *vma; |
589 | unsigned long address; | 800 | int write; |
590 | int write, si_code; | ||
591 | int fault; | 801 | int fault; |
592 | #ifdef CONFIG_X86_64 | ||
593 | unsigned long flags; | ||
594 | int sig; | ||
595 | #endif | ||
596 | 802 | ||
597 | tsk = current; | 803 | tsk = current; |
598 | mm = tsk->mm; | 804 | mm = tsk->mm; |
@@ -601,9 +807,7 @@ void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code) | |||
601 | /* get the address */ | 807 | /* get the address */ |
602 | address = read_cr2(); | 808 | address = read_cr2(); |
603 | 809 | ||
604 | si_code = SEGV_MAPERR; | 810 | if (unlikely(notify_page_fault(regs))) |
605 | |||
606 | if (notify_page_fault(regs)) | ||
607 | return; | 811 | return; |
608 | if (unlikely(kmmio_fault(regs, address))) | 812 | if (unlikely(kmmio_fault(regs, address))) |
609 | return; | 813 | return; |
@@ -631,17 +835,17 @@ void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code) | |||
631 | return; | 835 | return; |
632 | 836 | ||
633 | /* Can handle a stale RO->RW TLB */ | 837 | /* Can handle a stale RO->RW TLB */ |
634 | if (spurious_fault(address, error_code)) | 838 | if (spurious_fault(error_code, address)) |
635 | return; | 839 | return; |
636 | 840 | ||
637 | /* | 841 | /* |
638 | * Don't take the mm semaphore here. If we fixup a prefetch | 842 | * Don't take the mm semaphore here. If we fixup a prefetch |
639 | * fault we could otherwise deadlock. | 843 | * fault we could otherwise deadlock. |
640 | */ | 844 | */ |
641 | goto bad_area_nosemaphore; | 845 | bad_area_nosemaphore(regs, error_code, address); |
846 | return; | ||
642 | } | 847 | } |
643 | 848 | ||
644 | |||
645 | /* | 849 | /* |
646 | * It's safe to allow irq's after cr2 has been saved and the | 850 | * It's safe to allow irq's after cr2 has been saved and the |
647 | * vmalloc fault has been handled. | 851 | * vmalloc fault has been handled. |
@@ -657,15 +861,17 @@ void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code) | |||
657 | 861 | ||
658 | #ifdef CONFIG_X86_64 | 862 | #ifdef CONFIG_X86_64 |
659 | if (unlikely(error_code & PF_RSVD)) | 863 | if (unlikely(error_code & PF_RSVD)) |
660 | pgtable_bad(address, regs, error_code); | 864 | pgtable_bad(regs, error_code, address); |
661 | #endif | 865 | #endif |
662 | 866 | ||
663 | /* | 867 | /* |
664 | * If we're in an interrupt, have no user context or are running in an | 868 | * If we're in an interrupt, have no user context or are running in an |
665 | * atomic region then we must not take the fault. | 869 | * atomic region then we must not take the fault. |
666 | */ | 870 | */ |
667 | if (unlikely(in_atomic() || !mm)) | 871 | if (unlikely(in_atomic() || !mm)) { |
668 | goto bad_area_nosemaphore; | 872 | bad_area_nosemaphore(regs, error_code, address); |
873 | return; | ||
874 | } | ||
669 | 875 | ||
670 | /* | 876 | /* |
671 | * When running in the kernel we expect faults to occur only to | 877 | * When running in the kernel we expect faults to occur only to |
@@ -683,20 +889,26 @@ void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code) | |||
683 | * source. If this is invalid we can skip the address space check, | 889 | * source. If this is invalid we can skip the address space check, |
684 | * thus avoiding the deadlock. | 890 | * thus avoiding the deadlock. |
685 | */ | 891 | */ |
686 | if (!down_read_trylock(&mm->mmap_sem)) { | 892 | if (unlikely(!down_read_trylock(&mm->mmap_sem))) { |
687 | if ((error_code & PF_USER) == 0 && | 893 | if ((error_code & PF_USER) == 0 && |
688 | !search_exception_tables(regs->ip)) | 894 | !search_exception_tables(regs->ip)) { |
689 | goto bad_area_nosemaphore; | 895 | bad_area_nosemaphore(regs, error_code, address); |
896 | return; | ||
897 | } | ||
690 | down_read(&mm->mmap_sem); | 898 | down_read(&mm->mmap_sem); |
691 | } | 899 | } |
692 | 900 | ||
693 | vma = find_vma(mm, address); | 901 | vma = find_vma(mm, address); |
694 | if (!vma) | 902 | if (unlikely(!vma)) { |
695 | goto bad_area; | 903 | bad_area(regs, error_code, address); |
696 | if (vma->vm_start <= address) | 904 | return; |
905 | } | ||
906 | if (likely(vma->vm_start <= address)) | ||
697 | goto good_area; | 907 | goto good_area; |
698 | if (!(vma->vm_flags & VM_GROWSDOWN)) | 908 | if (unlikely(!(vma->vm_flags & VM_GROWSDOWN))) { |
699 | goto bad_area; | 909 | bad_area(regs, error_code, address); |
910 | return; | ||
911 | } | ||
700 | if (error_code & PF_USER) { | 912 | if (error_code & PF_USER) { |
701 | /* | 913 | /* |
702 | * Accessing the stack below %sp is always a bug. | 914 | * Accessing the stack below %sp is always a bug. |
@@ -704,31 +916,25 @@ void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code) | |||
704 | * and pusha to work. ("enter $65535,$31" pushes | 916 | * and pusha to work. ("enter $65535,$31" pushes |
705 | * 32 pointers and then decrements %sp by 65535.) | 917 | * 32 pointers and then decrements %sp by 65535.) |
706 | */ | 918 | */ |
707 | if (address + 65536 + 32 * sizeof(unsigned long) < regs->sp) | 919 | if (unlikely(address + 65536 + 32 * sizeof(unsigned long) < regs->sp)) { |
708 | goto bad_area; | 920 | bad_area(regs, error_code, address); |
921 | return; | ||
922 | } | ||
709 | } | 923 | } |
710 | if (expand_stack(vma, address)) | 924 | if (unlikely(expand_stack(vma, address))) { |
711 | goto bad_area; | 925 | bad_area(regs, error_code, address); |
712 | /* | 926 | return; |
713 | * Ok, we have a good vm_area for this memory access, so | 927 | } |
714 | * we can handle it.. | 928 | |
715 | */ | 929 | /* |
930 | * Ok, we have a good vm_area for this memory access, so | ||
931 | * we can handle it.. | ||
932 | */ | ||
716 | good_area: | 933 | good_area: |
717 | si_code = SEGV_ACCERR; | 934 | write = error_code & PF_WRITE; |
718 | write = 0; | 935 | if (unlikely(access_error(error_code, write, vma))) { |
719 | switch (error_code & (PF_PROT|PF_WRITE)) { | 936 | bad_area_access_error(regs, error_code, address); |
720 | default: /* 3: write, present */ | 937 | return; |
721 | /* fall through */ | ||
722 | case PF_WRITE: /* write, not present */ | ||
723 | if (!(vma->vm_flags & VM_WRITE)) | ||
724 | goto bad_area; | ||
725 | write++; | ||
726 | break; | ||
727 | case PF_PROT: /* read, present */ | ||
728 | goto bad_area; | ||
729 | case 0: /* read, not present */ | ||
730 | if (!(vma->vm_flags & (VM_READ | VM_EXEC | VM_WRITE))) | ||
731 | goto bad_area; | ||
732 | } | 938 | } |
733 | 939 | ||
734 | /* | 940 | /* |
@@ -738,11 +944,8 @@ good_area: | |||
738 | */ | 944 | */ |
739 | fault = handle_mm_fault(mm, vma, address, write); | 945 | fault = handle_mm_fault(mm, vma, address, write); |
740 | if (unlikely(fault & VM_FAULT_ERROR)) { | 946 | if (unlikely(fault & VM_FAULT_ERROR)) { |
741 | if (fault & VM_FAULT_OOM) | 947 | mm_fault_error(regs, error_code, address, fault); |
742 | goto out_of_memory; | 948 | return; |
743 | else if (fault & VM_FAULT_SIGBUS) | ||
744 | goto do_sigbus; | ||
745 | BUG(); | ||
746 | } | 949 | } |
747 | if (fault & VM_FAULT_MAJOR) | 950 | if (fault & VM_FAULT_MAJOR) |
748 | tsk->maj_flt++; | 951 | tsk->maj_flt++; |
@@ -760,128 +963,6 @@ good_area: | |||
760 | } | 963 | } |
761 | #endif | 964 | #endif |
762 | up_read(&mm->mmap_sem); | 965 | up_read(&mm->mmap_sem); |
763 | return; | ||
764 | |||
765 | /* | ||
766 | * Something tried to access memory that isn't in our memory map.. | ||
767 | * Fix it, but check if it's kernel or user first.. | ||
768 | */ | ||
769 | bad_area: | ||
770 | up_read(&mm->mmap_sem); | ||
771 | |||
772 | bad_area_nosemaphore: | ||
773 | /* User mode accesses just cause a SIGSEGV */ | ||
774 | if (error_code & PF_USER) { | ||
775 | /* | ||
776 | * It's possible to have interrupts off here. | ||
777 | */ | ||
778 | local_irq_enable(); | ||
779 | |||
780 | /* | ||
781 | * Valid to do another page fault here because this one came | ||
782 | * from user space. | ||
783 | */ | ||
784 | if (is_prefetch(regs, address, error_code)) | ||
785 | return; | ||
786 | |||
787 | if (is_errata100(regs, address)) | ||
788 | return; | ||
789 | |||
790 | if (show_unhandled_signals && unhandled_signal(tsk, SIGSEGV) && | ||
791 | printk_ratelimit()) { | ||
792 | printk( | ||
793 | "%s%s[%d]: segfault at %lx ip %p sp %p error %lx", | ||
794 | task_pid_nr(tsk) > 1 ? KERN_INFO : KERN_EMERG, | ||
795 | tsk->comm, task_pid_nr(tsk), address, | ||
796 | (void *) regs->ip, (void *) regs->sp, error_code); | ||
797 | print_vma_addr(" in ", regs->ip); | ||
798 | printk("\n"); | ||
799 | } | ||
800 | |||
801 | tsk->thread.cr2 = address; | ||
802 | /* Kernel addresses are always protection faults */ | ||
803 | tsk->thread.error_code = error_code | (address >= TASK_SIZE); | ||
804 | tsk->thread.trap_no = 14; | ||
805 | force_sig_info_fault(SIGSEGV, si_code, address, tsk); | ||
806 | return; | ||
807 | } | ||
808 | |||
809 | if (is_f00f_bug(regs, address)) | ||
810 | return; | ||
811 | |||
812 | no_context: | ||
813 | /* Are we prepared to handle this kernel fault? */ | ||
814 | if (fixup_exception(regs)) | ||
815 | return; | ||
816 | |||
817 | /* | ||
818 | * X86_32 | ||
819 | * Valid to do another page fault here, because if this fault | ||
820 | * had been triggered by is_prefetch fixup_exception would have | ||
821 | * handled it. | ||
822 | * | ||
823 | * X86_64 | ||
824 | * Hall of shame of CPU/BIOS bugs. | ||
825 | */ | ||
826 | if (is_prefetch(regs, address, error_code)) | ||
827 | return; | ||
828 | |||
829 | if (is_errata93(regs, address)) | ||
830 | return; | ||
831 | |||
832 | /* | ||
833 | * Oops. The kernel tried to access some bad page. We'll have to | ||
834 | * terminate things with extreme prejudice. | ||
835 | */ | ||
836 | #ifdef CONFIG_X86_32 | ||
837 | bust_spinlocks(1); | ||
838 | #else | ||
839 | flags = oops_begin(); | ||
840 | #endif | ||
841 | |||
842 | show_fault_oops(regs, error_code, address); | ||
843 | |||
844 | tsk->thread.cr2 = address; | ||
845 | tsk->thread.trap_no = 14; | ||
846 | tsk->thread.error_code = error_code; | ||
847 | |||
848 | #ifdef CONFIG_X86_32 | ||
849 | die("Oops", regs, error_code); | ||
850 | bust_spinlocks(0); | ||
851 | do_exit(SIGKILL); | ||
852 | #else | ||
853 | sig = SIGKILL; | ||
854 | if (__die("Oops", regs, error_code)) | ||
855 | sig = 0; | ||
856 | /* Executive summary in case the body of the oops scrolled away */ | ||
857 | printk(KERN_EMERG "CR2: %016lx\n", address); | ||
858 | oops_end(flags, regs, sig); | ||
859 | #endif | ||
860 | |||
861 | out_of_memory: | ||
862 | /* | ||
863 | * We ran out of memory, call the OOM killer, and return the userspace | ||
864 | * (which will retry the fault, or kill us if we got oom-killed). | ||
865 | */ | ||
866 | up_read(&mm->mmap_sem); | ||
867 | pagefault_out_of_memory(); | ||
868 | return; | ||
869 | |||
870 | do_sigbus: | ||
871 | up_read(&mm->mmap_sem); | ||
872 | |||
873 | /* Kernel mode? Handle exceptions or die */ | ||
874 | if (!(error_code & PF_USER)) | ||
875 | goto no_context; | ||
876 | #ifdef CONFIG_X86_32 | ||
877 | /* User space => ok to do another page fault */ | ||
878 | if (is_prefetch(regs, address, error_code)) | ||
879 | return; | ||
880 | #endif | ||
881 | tsk->thread.cr2 = address; | ||
882 | tsk->thread.error_code = error_code; | ||
883 | tsk->thread.trap_no = 14; | ||
884 | force_sig_info_fault(SIGBUS, BUS_ADRERR, address, tsk); | ||
885 | } | 966 | } |
886 | 967 | ||
887 | DEFINE_SPINLOCK(pgd_lock); | 968 | DEFINE_SPINLOCK(pgd_lock); |