aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/kprobes.c
diff options
context:
space:
mode:
Diffstat (limited to 'kernel/kprobes.c')
-rw-r--r--kernel/kprobes.c124
1 files changed, 87 insertions, 37 deletions
diff --git a/kernel/kprobes.c b/kernel/kprobes.c
index f02a4311768b..76275fc025a5 100644
--- a/kernel/kprobes.c
+++ b/kernel/kprobes.c
@@ -580,6 +580,7 @@ static int __kprobes __register_kprobe(struct kprobe *p,
580 } 580 }
581 581
582 p->nmissed = 0; 582 p->nmissed = 0;
583 INIT_LIST_HEAD(&p->list);
583 mutex_lock(&kprobe_mutex); 584 mutex_lock(&kprobe_mutex);
584 old_p = get_kprobe(p->addr); 585 old_p = get_kprobe(p->addr);
585 if (old_p) { 586 if (old_p) {
@@ -606,35 +607,28 @@ out:
606 return ret; 607 return ret;
607} 608}
608 609
609int __kprobes register_kprobe(struct kprobe *p) 610/*
610{ 611 * Unregister a kprobe without a scheduler synchronization.
611 return __register_kprobe(p, (unsigned long)__builtin_return_address(0)); 612 */
612} 613static int __kprobes __unregister_kprobe_top(struct kprobe *p)
613
614void __kprobes unregister_kprobe(struct kprobe *p)
615{ 614{
616 struct module *mod;
617 struct kprobe *old_p, *list_p; 615 struct kprobe *old_p, *list_p;
618 int cleanup_p;
619 616
620 mutex_lock(&kprobe_mutex);
621 old_p = get_kprobe(p->addr); 617 old_p = get_kprobe(p->addr);
622 if (unlikely(!old_p)) { 618 if (unlikely(!old_p))
623 mutex_unlock(&kprobe_mutex); 619 return -EINVAL;
624 return; 620
625 }
626 if (p != old_p) { 621 if (p != old_p) {
627 list_for_each_entry_rcu(list_p, &old_p->list, list) 622 list_for_each_entry_rcu(list_p, &old_p->list, list)
628 if (list_p == p) 623 if (list_p == p)
629 /* kprobe p is a valid probe */ 624 /* kprobe p is a valid probe */
630 goto valid_p; 625 goto valid_p;
631 mutex_unlock(&kprobe_mutex); 626 return -EINVAL;
632 return;
633 } 627 }
634valid_p: 628valid_p:
635 if (old_p == p || 629 if (old_p == p ||
636 (old_p->pre_handler == aggr_pre_handler && 630 (old_p->pre_handler == aggr_pre_handler &&
637 p->list.next == &old_p->list && p->list.prev == &old_p->list)) { 631 list_is_singular(&old_p->list))) {
638 /* 632 /*
639 * Only probe on the hash list. Disarm only if kprobes are 633 * Only probe on the hash list. Disarm only if kprobes are
640 * enabled - otherwise, the breakpoint would already have 634 * enabled - otherwise, the breakpoint would already have
@@ -643,43 +637,97 @@ valid_p:
643 if (kprobe_enabled) 637 if (kprobe_enabled)
644 arch_disarm_kprobe(p); 638 arch_disarm_kprobe(p);
645 hlist_del_rcu(&old_p->hlist); 639 hlist_del_rcu(&old_p->hlist);
646 cleanup_p = 1;
647 } else { 640 } else {
641 if (p->break_handler)
642 old_p->break_handler = NULL;
643 if (p->post_handler) {
644 list_for_each_entry_rcu(list_p, &old_p->list, list) {
645 if ((list_p != p) && (list_p->post_handler))
646 goto noclean;
647 }
648 old_p->post_handler = NULL;
649 }
650noclean:
648 list_del_rcu(&p->list); 651 list_del_rcu(&p->list);
649 cleanup_p = 0;
650 } 652 }
653 return 0;
654}
651 655
652 mutex_unlock(&kprobe_mutex); 656static void __kprobes __unregister_kprobe_bottom(struct kprobe *p)
657{
658 struct module *mod;
659 struct kprobe *old_p;
653 660
654 synchronize_sched();
655 if (p->mod_refcounted) { 661 if (p->mod_refcounted) {
656 mod = module_text_address((unsigned long)p->addr); 662 mod = module_text_address((unsigned long)p->addr);
657 if (mod) 663 if (mod)
658 module_put(mod); 664 module_put(mod);
659 } 665 }
660 666
661 if (cleanup_p) { 667 if (list_empty(&p->list) || list_is_singular(&p->list)) {
662 if (p != old_p) { 668 if (!list_empty(&p->list)) {
663 list_del_rcu(&p->list); 669 /* "p" is the last child of an aggr_kprobe */
670 old_p = list_entry(p->list.next, struct kprobe, list);
671 list_del(&p->list);
664 kfree(old_p); 672 kfree(old_p);
665 } 673 }
666 arch_remove_kprobe(p); 674 arch_remove_kprobe(p);
667 } else { 675 }
668 mutex_lock(&kprobe_mutex); 676}
669 if (p->break_handler) 677
670 old_p->break_handler = NULL; 678static int __register_kprobes(struct kprobe **kps, int num,
671 if (p->post_handler){ 679 unsigned long called_from)
672 list_for_each_entry_rcu(list_p, &old_p->list, list){ 680{
673 if (list_p->post_handler){ 681 int i, ret = 0;
674 cleanup_p = 2; 682
675 break; 683 if (num <= 0)
676 } 684 return -EINVAL;
677 } 685 for (i = 0; i < num; i++) {
678 if (cleanup_p == 0) 686 ret = __register_kprobe(kps[i], called_from);
679 old_p->post_handler = NULL; 687 if (ret < 0 && i > 0) {
688 unregister_kprobes(kps, i);
689 break;
680 } 690 }
681 mutex_unlock(&kprobe_mutex);
682 } 691 }
692 return ret;
693}
694
695/*
696 * Registration and unregistration functions for kprobe.
697 */
698int __kprobes register_kprobe(struct kprobe *p)
699{
700 return __register_kprobes(&p, 1,
701 (unsigned long)__builtin_return_address(0));
702}
703
704void __kprobes unregister_kprobe(struct kprobe *p)
705{
706 unregister_kprobes(&p, 1);
707}
708
709int __kprobes register_kprobes(struct kprobe **kps, int num)
710{
711 return __register_kprobes(kps, num,
712 (unsigned long)__builtin_return_address(0));
713}
714
715void __kprobes unregister_kprobes(struct kprobe **kps, int num)
716{
717 int i;
718
719 if (num <= 0)
720 return;
721 mutex_lock(&kprobe_mutex);
722 for (i = 0; i < num; i++)
723 if (__unregister_kprobe_top(kps[i]) < 0)
724 kps[i]->addr = NULL;
725 mutex_unlock(&kprobe_mutex);
726
727 synchronize_sched();
728 for (i = 0; i < num; i++)
729 if (kps[i]->addr)
730 __unregister_kprobe_bottom(kps[i]);
683} 731}
684 732
685static struct notifier_block kprobe_exceptions_nb = { 733static struct notifier_block kprobe_exceptions_nb = {
@@ -1118,6 +1166,8 @@ module_init(init_kprobes);
1118 1166
1119EXPORT_SYMBOL_GPL(register_kprobe); 1167EXPORT_SYMBOL_GPL(register_kprobe);
1120EXPORT_SYMBOL_GPL(unregister_kprobe); 1168EXPORT_SYMBOL_GPL(unregister_kprobe);
1169EXPORT_SYMBOL_GPL(register_kprobes);
1170EXPORT_SYMBOL_GPL(unregister_kprobes);
1121EXPORT_SYMBOL_GPL(register_jprobe); 1171EXPORT_SYMBOL_GPL(register_jprobe);
1122EXPORT_SYMBOL_GPL(unregister_jprobe); 1172EXPORT_SYMBOL_GPL(unregister_jprobe);
1123#ifdef CONFIG_KPROBES 1173#ifdef CONFIG_KPROBES