aboutsummaryrefslogtreecommitdiffstats
path: root/arch/x86/mm/fault.c
diff options
context:
space:
mode:
Diffstat (limited to 'arch/x86/mm/fault.c')
-rw-r--r--arch/x86/mm/fault.c450
1 files changed, 262 insertions, 188 deletions
diff --git a/arch/x86/mm/fault.c b/arch/x86/mm/fault.c
index 37242c405f16..65709a6aa6ee 100644
--- a/arch/x86/mm/fault.c
+++ b/arch/x86/mm/fault.c
@@ -92,8 +92,8 @@ static inline int notify_page_fault(struct pt_regs *regs)
92 * 92 *
93 * Opcode checker based on code by Richard Brunner 93 * Opcode checker based on code by Richard Brunner
94 */ 94 */
95static int is_prefetch(struct pt_regs *regs, unsigned long addr, 95static int is_prefetch(struct pt_regs *regs, unsigned long error_code,
96 unsigned long error_code) 96 unsigned long addr)
97{ 97{
98 unsigned char *instr; 98 unsigned char *instr;
99 int scan_more = 1; 99 int scan_more = 1;
@@ -410,15 +410,15 @@ static void show_fault_oops(struct pt_regs *regs, unsigned long error_code,
410} 410}
411 411
412#ifdef CONFIG_X86_64 412#ifdef CONFIG_X86_64
413static noinline void pgtable_bad(unsigned long address, struct pt_regs *regs, 413static noinline void pgtable_bad(struct pt_regs *regs,
414 unsigned long error_code) 414 unsigned long error_code, unsigned long address)
415{ 415{
416 unsigned long flags = oops_begin(); 416 unsigned long flags = oops_begin();
417 int sig = SIGKILL; 417 int sig = SIGKILL;
418 struct task_struct *tsk; 418 struct task_struct *tsk = current;
419 419
420 printk(KERN_ALERT "%s: Corrupted page table at address %lx\n", 420 printk(KERN_ALERT "%s: Corrupted page table at address %lx\n",
421 current->comm, address); 421 tsk->comm, address);
422 dump_pagetable(address); 422 dump_pagetable(address);
423 tsk = current; 423 tsk = current;
424 tsk->thread.cr2 = address; 424 tsk->thread.cr2 = address;
@@ -430,6 +430,196 @@ static noinline void pgtable_bad(unsigned long address, struct pt_regs *regs,
430} 430}
431#endif 431#endif
432 432
433static noinline void no_context(struct pt_regs *regs,
434 unsigned long error_code, unsigned long address)
435{
436 struct task_struct *tsk = current;
437 unsigned long *stackend;
438
439#ifdef CONFIG_X86_64
440 unsigned long flags;
441 int sig;
442#endif
443
444 /* Are we prepared to handle this kernel fault? */
445 if (fixup_exception(regs))
446 return;
447
448 /*
449 * X86_32
450 * Valid to do another page fault here, because if this fault
451 * had been triggered by is_prefetch fixup_exception would have
452 * handled it.
453 *
454 * X86_64
455 * Hall of shame of CPU/BIOS bugs.
456 */
457 if (is_prefetch(regs, error_code, address))
458 return;
459
460 if (is_errata93(regs, address))
461 return;
462
463 /*
464 * Oops. The kernel tried to access some bad page. We'll have to
465 * terminate things with extreme prejudice.
466 */
467#ifdef CONFIG_X86_32
468 bust_spinlocks(1);
469#else
470 flags = oops_begin();
471#endif
472
473 show_fault_oops(regs, error_code, address);
474
475 stackend = end_of_stack(tsk);
476 if (*stackend != STACK_END_MAGIC)
477 printk(KERN_ALERT "Thread overran stack, or stack corrupted\n");
478
479 tsk->thread.cr2 = address;
480 tsk->thread.trap_no = 14;
481 tsk->thread.error_code = error_code;
482
483#ifdef CONFIG_X86_32
484 die("Oops", regs, error_code);
485 bust_spinlocks(0);
486 do_exit(SIGKILL);
487#else
488 sig = SIGKILL;
489 if (__die("Oops", regs, error_code))
490 sig = 0;
491 /* Executive summary in case the body of the oops scrolled away */
492 printk(KERN_EMERG "CR2: %016lx\n", address);
493 oops_end(flags, regs, sig);
494#endif
495}
496
497static void __bad_area_nosemaphore(struct pt_regs *regs,
498 unsigned long error_code, unsigned long address,
499 int si_code)
500{
501 struct task_struct *tsk = current;
502
503 /* User mode accesses just cause a SIGSEGV */
504 if (error_code & PF_USER) {
505 /*
506 * It's possible to have interrupts off here.
507 */
508 local_irq_enable();
509
510 /*
511 * Valid to do another page fault here because this one came
512 * from user space.
513 */
514 if (is_prefetch(regs, error_code, address))
515 return;
516
517 if (is_errata100(regs, address))
518 return;
519
520 if (show_unhandled_signals && unhandled_signal(tsk, SIGSEGV) &&
521 printk_ratelimit()) {
522 printk(
523 "%s%s[%d]: segfault at %lx ip %p sp %p error %lx",
524 task_pid_nr(tsk) > 1 ? KERN_INFO : KERN_EMERG,
525 tsk->comm, task_pid_nr(tsk), address,
526 (void *) regs->ip, (void *) regs->sp, error_code);
527 print_vma_addr(" in ", regs->ip);
528 printk("\n");
529 }
530
531 tsk->thread.cr2 = address;
532 /* Kernel addresses are always protection faults */
533 tsk->thread.error_code = error_code | (address >= TASK_SIZE);
534 tsk->thread.trap_no = 14;
535 force_sig_info_fault(SIGSEGV, si_code, address, tsk);
536 return;
537 }
538
539 if (is_f00f_bug(regs, address))
540 return;
541
542 no_context(regs, error_code, address);
543}
544
545static noinline void bad_area_nosemaphore(struct pt_regs *regs,
546 unsigned long error_code, unsigned long address)
547{
548 __bad_area_nosemaphore(regs, error_code, address, SEGV_MAPERR);
549}
550
551static void __bad_area(struct pt_regs *regs,
552 unsigned long error_code, unsigned long address,
553 int si_code)
554{
555 struct mm_struct *mm = current->mm;
556
557 /*
558 * Something tried to access memory that isn't in our memory map..
559 * Fix it, but check if it's kernel or user first..
560 */
561 up_read(&mm->mmap_sem);
562
563 __bad_area_nosemaphore(regs, error_code, address, si_code);
564}
565
566static noinline void bad_area(struct pt_regs *regs,
567 unsigned long error_code, unsigned long address)
568{
569 __bad_area(regs, error_code, address, SEGV_MAPERR);
570}
571
572static noinline void bad_area_access_error(struct pt_regs *regs,
573 unsigned long error_code, unsigned long address)
574{
575 __bad_area(regs, error_code, address, SEGV_ACCERR);
576}
577
578/* TODO: fixup for "mm-invoke-oom-killer-from-page-fault.patch" */
579static void out_of_memory(struct pt_regs *regs,
580 unsigned long error_code, unsigned long address)
581{
582 /*
583 * We ran out of memory, call the OOM killer, and return the userspace
584 * (which will retry the fault, or kill us if we got oom-killed).
585 */
586 up_read(&current->mm->mmap_sem);
587 pagefault_out_of_memory();
588}
589
590static void do_sigbus(struct pt_regs *regs,
591 unsigned long error_code, unsigned long address)
592{
593 struct task_struct *tsk = current;
594 struct mm_struct *mm = tsk->mm;
595
596 up_read(&mm->mmap_sem);
597
598 /* Kernel mode? Handle exceptions or die */
599 if (!(error_code & PF_USER))
600 no_context(regs, error_code, address);
601#ifdef CONFIG_X86_32
602 /* User space => ok to do another page fault */
603 if (is_prefetch(regs, error_code, address))
604 return;
605#endif
606 tsk->thread.cr2 = address;
607 tsk->thread.error_code = error_code;
608 tsk->thread.trap_no = 14;
609 force_sig_info_fault(SIGBUS, BUS_ADRERR, address, tsk);
610}
611
612static noinline void mm_fault_error(struct pt_regs *regs,
613 unsigned long error_code, unsigned long address, unsigned int fault)
614{
615 if (fault & VM_FAULT_OOM)
616 out_of_memory(regs, error_code, address);
617 else if (fault & VM_FAULT_SIGBUS)
618 do_sigbus(regs, error_code, address);
619 else
620 BUG();
621}
622
433static int spurious_fault_check(unsigned long error_code, pte_t *pte) 623static int spurious_fault_check(unsigned long error_code, pte_t *pte)
434{ 624{
435 if ((error_code & PF_WRITE) && !pte_write(*pte)) 625 if ((error_code & PF_WRITE) && !pte_write(*pte))
@@ -449,8 +639,8 @@ static int spurious_fault_check(unsigned long error_code, pte_t *pte)
449 * There are no security implications to leaving a stale TLB when 639 * There are no security implications to leaving a stale TLB when
450 * increasing the permissions on a page. 640 * increasing the permissions on a page.
451 */ 641 */
452static int spurious_fault(unsigned long address, 642static noinline int spurious_fault(unsigned long error_code,
453 unsigned long error_code) 643 unsigned long address)
454{ 644{
455 pgd_t *pgd; 645 pgd_t *pgd;
456 pud_t *pud; 646 pud_t *pud;
@@ -495,7 +685,7 @@ static int spurious_fault(unsigned long address,
495 * 685 *
496 * This assumes no large pages in there. 686 * This assumes no large pages in there.
497 */ 687 */
498static int vmalloc_fault(unsigned long address) 688static noinline int vmalloc_fault(unsigned long address)
499{ 689{
500#ifdef CONFIG_X86_32 690#ifdef CONFIG_X86_32
501 unsigned long pgd_paddr; 691 unsigned long pgd_paddr;
@@ -574,6 +764,25 @@ static int vmalloc_fault(unsigned long address)
574 764
575int show_unhandled_signals = 1; 765int show_unhandled_signals = 1;
576 766
767static inline int access_error(unsigned long error_code, int write,
768 struct vm_area_struct *vma)
769{
770 if (write) {
771 /* write, present and write, not present */
772 if (unlikely(!(vma->vm_flags & VM_WRITE)))
773 return 1;
774 } else if (unlikely(error_code & PF_PROT)) {
775 /* read, present */
776 return 1;
777 } else {
778 /* read, not present */
779 if (unlikely(!(vma->vm_flags & (VM_READ | VM_EXEC | VM_WRITE))))
780 return 1;
781 }
782
783 return 0;
784}
785
577/* 786/*
578 * This routine handles page faults. It determines the address, 787 * This routine handles page faults. It determines the address,
579 * and the problem, and then passes it off to one of the appropriate 788 * and the problem, and then passes it off to one of the appropriate
@@ -584,18 +793,12 @@ asmlinkage
584#endif 793#endif
585void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code) 794void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code)
586{ 795{
796 unsigned long address;
587 struct task_struct *tsk; 797 struct task_struct *tsk;
588 struct mm_struct *mm; 798 struct mm_struct *mm;
589 struct vm_area_struct *vma; 799 struct vm_area_struct *vma;
590 unsigned long address; 800 int write;
591 int write, si_code;
592 int fault; 801 int fault;
593 unsigned long *stackend;
594
595#ifdef CONFIG_X86_64
596 unsigned long flags;
597 int sig;
598#endif
599 802
600 tsk = current; 803 tsk = current;
601 mm = tsk->mm; 804 mm = tsk->mm;
@@ -604,9 +807,7 @@ void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code)
604 /* get the address */ 807 /* get the address */
605 address = read_cr2(); 808 address = read_cr2();
606 809
607 si_code = SEGV_MAPERR; 810 if (unlikely(notify_page_fault(regs)))
608
609 if (notify_page_fault(regs))
610 return; 811 return;
611 if (unlikely(kmmio_fault(regs, address))) 812 if (unlikely(kmmio_fault(regs, address)))
612 return; 813 return;
@@ -634,17 +835,17 @@ void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code)
634 return; 835 return;
635 836
636 /* Can handle a stale RO->RW TLB */ 837 /* Can handle a stale RO->RW TLB */
637 if (spurious_fault(address, error_code)) 838 if (spurious_fault(error_code, address))
638 return; 839 return;
639 840
640 /* 841 /*
641 * Don't take the mm semaphore here. If we fixup a prefetch 842 * Don't take the mm semaphore here. If we fixup a prefetch
642 * fault we could otherwise deadlock. 843 * fault we could otherwise deadlock.
643 */ 844 */
644 goto bad_area_nosemaphore; 845 bad_area_nosemaphore(regs, error_code, address);
846 return;
645 } 847 }
646 848
647
648 /* 849 /*
649 * It's safe to allow irq's after cr2 has been saved and the 850 * It's safe to allow irq's after cr2 has been saved and the
650 * vmalloc fault has been handled. 851 * vmalloc fault has been handled.
@@ -660,15 +861,17 @@ void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code)
660 861
661#ifdef CONFIG_X86_64 862#ifdef CONFIG_X86_64
662 if (unlikely(error_code & PF_RSVD)) 863 if (unlikely(error_code & PF_RSVD))
663 pgtable_bad(address, regs, error_code); 864 pgtable_bad(regs, error_code, address);
664#endif 865#endif
665 866
666 /* 867 /*
667 * If we're in an interrupt, have no user context or are running in an 868 * If we're in an interrupt, have no user context or are running in an
668 * atomic region then we must not take the fault. 869 * atomic region then we must not take the fault.
669 */ 870 */
670 if (unlikely(in_atomic() || !mm)) 871 if (unlikely(in_atomic() || !mm)) {
671 goto bad_area_nosemaphore; 872 bad_area_nosemaphore(regs, error_code, address);
873 return;
874 }
672 875
673 /* 876 /*
674 * When running in the kernel we expect faults to occur only to 877 * When running in the kernel we expect faults to occur only to
@@ -686,20 +889,26 @@ void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code)
686 * source. If this is invalid we can skip the address space check, 889 * source. If this is invalid we can skip the address space check,
687 * thus avoiding the deadlock. 890 * thus avoiding the deadlock.
688 */ 891 */
689 if (!down_read_trylock(&mm->mmap_sem)) { 892 if (unlikely(!down_read_trylock(&mm->mmap_sem))) {
690 if ((error_code & PF_USER) == 0 && 893 if ((error_code & PF_USER) == 0 &&
691 !search_exception_tables(regs->ip)) 894 !search_exception_tables(regs->ip)) {
692 goto bad_area_nosemaphore; 895 bad_area_nosemaphore(regs, error_code, address);
896 return;
897 }
693 down_read(&mm->mmap_sem); 898 down_read(&mm->mmap_sem);
694 } 899 }
695 900
696 vma = find_vma(mm, address); 901 vma = find_vma(mm, address);
697 if (!vma) 902 if (unlikely(!vma)) {
698 goto bad_area; 903 bad_area(regs, error_code, address);
699 if (vma->vm_start <= address) 904 return;
905 }
906 if (likely(vma->vm_start <= address))
700 goto good_area; 907 goto good_area;
701 if (!(vma->vm_flags & VM_GROWSDOWN)) 908 if (unlikely(!(vma->vm_flags & VM_GROWSDOWN))) {
702 goto bad_area; 909 bad_area(regs, error_code, address);
910 return;
911 }
703 if (error_code & PF_USER) { 912 if (error_code & PF_USER) {
704 /* 913 /*
705 * Accessing the stack below %sp is always a bug. 914 * Accessing the stack below %sp is always a bug.
@@ -707,31 +916,25 @@ void __kprobes do_page_fault(struct pt_regs *regs, unsigned long error_code)
707 * and pusha to work. ("enter $65535,$31" pushes 916 * and pusha to work. ("enter $65535,$31" pushes
708 * 32 pointers and then decrements %sp by 65535.) 917 * 32 pointers and then decrements %sp by 65535.)
709 */ 918 */
710 if (address + 65536 + 32 * sizeof(unsigned long) < regs->sp) 919 if (unlikely(address + 65536 + 32 * sizeof(unsigned long) < regs->sp)) {
711 goto bad_area; 920 bad_area(regs, error_code, address);
921 return;
922 }
712 } 923 }
713 if (expand_stack(vma, address)) 924 if (unlikely(expand_stack(vma, address))) {
714 goto bad_area; 925 bad_area(regs, error_code, address);
715/* 926 return;
716 * Ok, we have a good vm_area for this memory access, so 927 }
717 * we can handle it.. 928
718 */ 929 /*
930 * Ok, we have a good vm_area for this memory access, so
931 * we can handle it..
932 */
719good_area: 933good_area:
720 si_code = SEGV_ACCERR; 934 write = error_code & PF_WRITE;
721 write = 0; 935 if (unlikely(access_error(error_code, write, vma))) {
722 switch (error_code & (PF_PROT|PF_WRITE)) { 936 bad_area_access_error(regs, error_code, address);
723 default: /* 3: write, present */ 937 return;
724 /* fall through */
725 case PF_WRITE: /* write, not present */
726 if (!(vma->vm_flags & VM_WRITE))
727 goto bad_area;
728 write++;
729 break;
730 case PF_PROT: /* read, present */
731 goto bad_area;
732 case 0: /* read, not present */
733 if (!(vma->vm_flags & (VM_READ | VM_EXEC | VM_WRITE)))
734 goto bad_area;
735 } 938 }
736 939
737 /* 940 /*
@@ -741,11 +944,8 @@ good_area:
741 */ 944 */
742 fault = handle_mm_fault(mm, vma, address, write); 945 fault = handle_mm_fault(mm, vma, address, write);
743 if (unlikely(fault & VM_FAULT_ERROR)) { 946 if (unlikely(fault & VM_FAULT_ERROR)) {
744 if (fault & VM_FAULT_OOM) 947 mm_fault_error(regs, error_code, address, fault);
745 goto out_of_memory; 948 return;
746 else if (fault & VM_FAULT_SIGBUS)
747 goto do_sigbus;
748 BUG();
749 } 949 }
750 if (fault & VM_FAULT_MAJOR) 950 if (fault & VM_FAULT_MAJOR)
751 tsk->maj_flt++; 951 tsk->maj_flt++;
@@ -763,132 +963,6 @@ good_area:
763 } 963 }
764#endif 964#endif
765 up_read(&mm->mmap_sem); 965 up_read(&mm->mmap_sem);
766 return;
767
768/*
769 * Something tried to access memory that isn't in our memory map..
770 * Fix it, but check if it's kernel or user first..
771 */
772bad_area:
773 up_read(&mm->mmap_sem);
774
775bad_area_nosemaphore:
776 /* User mode accesses just cause a SIGSEGV */
777 if (error_code & PF_USER) {
778 /*
779 * It's possible to have interrupts off here.
780 */
781 local_irq_enable();
782
783 /*
784 * Valid to do another page fault here because this one came
785 * from user space.
786 */
787 if (is_prefetch(regs, address, error_code))
788 return;
789
790 if (is_errata100(regs, address))
791 return;
792
793 if (show_unhandled_signals && unhandled_signal(tsk, SIGSEGV) &&
794 printk_ratelimit()) {
795 printk(
796 "%s%s[%d]: segfault at %lx ip %p sp %p error %lx",
797 task_pid_nr(tsk) > 1 ? KERN_INFO : KERN_EMERG,
798 tsk->comm, task_pid_nr(tsk), address,
799 (void *) regs->ip, (void *) regs->sp, error_code);
800 print_vma_addr(" in ", regs->ip);
801 printk("\n");
802 }
803
804 tsk->thread.cr2 = address;
805 /* Kernel addresses are always protection faults */
806 tsk->thread.error_code = error_code | (address >= TASK_SIZE);
807 tsk->thread.trap_no = 14;
808 force_sig_info_fault(SIGSEGV, si_code, address, tsk);
809 return;
810 }
811
812 if (is_f00f_bug(regs, address))
813 return;
814
815no_context:
816 /* Are we prepared to handle this kernel fault? */
817 if (fixup_exception(regs))
818 return;
819
820 /*
821 * X86_32
822 * Valid to do another page fault here, because if this fault
823 * had been triggered by is_prefetch fixup_exception would have
824 * handled it.
825 *
826 * X86_64
827 * Hall of shame of CPU/BIOS bugs.
828 */
829 if (is_prefetch(regs, address, error_code))
830 return;
831
832 if (is_errata93(regs, address))
833 return;
834
835/*
836 * Oops. The kernel tried to access some bad page. We'll have to
837 * terminate things with extreme prejudice.
838 */
839#ifdef CONFIG_X86_32
840 bust_spinlocks(1);
841#else
842 flags = oops_begin();
843#endif
844
845 show_fault_oops(regs, error_code, address);
846
847 stackend = end_of_stack(tsk);
848 if (*stackend != STACK_END_MAGIC)
849 printk(KERN_ALERT "Thread overran stack, or stack corrupted\n");
850
851 tsk->thread.cr2 = address;
852 tsk->thread.trap_no = 14;
853 tsk->thread.error_code = error_code;
854
855#ifdef CONFIG_X86_32
856 die("Oops", regs, error_code);
857 bust_spinlocks(0);
858 do_exit(SIGKILL);
859#else
860 sig = SIGKILL;
861 if (__die("Oops", regs, error_code))
862 sig = 0;
863 /* Executive summary in case the body of the oops scrolled away */
864 printk(KERN_EMERG "CR2: %016lx\n", address);
865 oops_end(flags, regs, sig);
866#endif
867
868out_of_memory:
869 /*
870 * We ran out of memory, call the OOM killer, and return the userspace
871 * (which will retry the fault, or kill us if we got oom-killed).
872 */
873 up_read(&mm->mmap_sem);
874 pagefault_out_of_memory();
875 return;
876
877do_sigbus:
878 up_read(&mm->mmap_sem);
879
880 /* Kernel mode? Handle exceptions or die */
881 if (!(error_code & PF_USER))
882 goto no_context;
883#ifdef CONFIG_X86_32
884 /* User space => ok to do another page fault */
885 if (is_prefetch(regs, address, error_code))
886 return;
887#endif
888 tsk->thread.cr2 = address;
889 tsk->thread.error_code = error_code;
890 tsk->thread.trap_no = 14;
891 force_sig_info_fault(SIGBUS, BUS_ADRERR, address, tsk);
892} 966}
893 967
894DEFINE_SPINLOCK(pgd_lock); 968DEFINE_SPINLOCK(pgd_lock);