aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorMasami Hiramatsu <masami.hiramatsu.pt@hitachi.com>2010-12-03 04:54:09 -0500
committerIngo Molnar <mingo@elte.hu>2010-12-06 11:59:30 -0500
commit6274de4984a630b45c6934b3ee62e5692c745328 (patch)
treedfe66f06596b0165c87a75d800fa83acb6201d1b
parent61f4e13ffd85c037a942c5b7fd13f2b0c3162862 (diff)
kprobes: Support delayed unoptimizing
Unoptimization occurs when a probe is unregistered or disabled, and is heavy because it recovers instructions by using stop_machine(). This patch delays unoptimization operations and unoptimize several probes at once by using text_poke_smp_batch(). This can avoid unexpected system slowdown coming from stop_machine(). Changes in v5: - Split this patch into several cleanup patches and this patch. - Fix some text_mutex lock miss. - Use bool instead of int for behavior flags. - Add additional comment for (un)optimizing path. Changes in v2: - Use dynamic allocated buffers and params. Signed-off-by: Masami Hiramatsu <masami.hiramatsu.pt@hitachi.com> Cc: Rusty Russell <rusty@rustcorp.com.au> Cc: Frederic Weisbecker <fweisbec@gmail.com> Cc: Ananth N Mavinakayanahalli <ananth@in.ibm.com> Cc: Jason Baron <jbaron@redhat.com> Cc: Mathieu Desnoyers <mathieu.desnoyers@efficios.com> Cc: 2nddept-manager@sdl.hitachi.co.jp LKML-Reference: <20101203095409.2961.82733.stgit@ltc236.sdl.hitachi.co.jp> Signed-off-by: Ingo Molnar <mingo@elte.hu>
-rw-r--r--arch/x86/kernel/kprobes.c4
-rw-r--r--kernel/kprobes.c310
2 files changed, 237 insertions, 77 deletions
diff --git a/arch/x86/kernel/kprobes.c b/arch/x86/kernel/kprobes.c
index 1cbd54c0df99..da51dc8e77cb 100644
--- a/arch/x86/kernel/kprobes.c
+++ b/arch/x86/kernel/kprobes.c
@@ -1184,6 +1184,10 @@ static void __kprobes optimized_callback(struct optimized_kprobe *op,
1184{ 1184{
1185 struct kprobe_ctlblk *kcb = get_kprobe_ctlblk(); 1185 struct kprobe_ctlblk *kcb = get_kprobe_ctlblk();
1186 1186
1187 /* This is possible if op is under delayed unoptimizing */
1188 if (kprobe_disabled(&op->kp))
1189 return;
1190
1187 preempt_disable(); 1191 preempt_disable();
1188 if (kprobe_running()) { 1192 if (kprobe_running()) {
1189 kprobes_inc_nmissed_count(&op->kp); 1193 kprobes_inc_nmissed_count(&op->kp);
diff --git a/kernel/kprobes.c b/kernel/kprobes.c
index 1f4f9b9d5c89..ba4d4c0740cf 100644
--- a/kernel/kprobes.c
+++ b/kernel/kprobes.c
@@ -354,6 +354,13 @@ static inline int kprobe_aggrprobe(struct kprobe *p)
354 return p->pre_handler == aggr_pre_handler; 354 return p->pre_handler == aggr_pre_handler;
355} 355}
356 356
357/* Return true(!0) if the kprobe is unused */
358static inline int kprobe_unused(struct kprobe *p)
359{
360 return kprobe_aggrprobe(p) && kprobe_disabled(p) &&
361 list_empty(&p->list);
362}
363
357/* 364/*
358 * Keep all fields in the kprobe consistent 365 * Keep all fields in the kprobe consistent
359 */ 366 */
@@ -384,6 +391,17 @@ void __kprobes opt_pre_handler(struct kprobe *p, struct pt_regs *regs)
384 } 391 }
385} 392}
386 393
394/* Free optimized instructions and optimized_kprobe */
395static __kprobes void free_aggr_kprobe(struct kprobe *p)
396{
397 struct optimized_kprobe *op;
398
399 op = container_of(p, struct optimized_kprobe, kp);
400 arch_remove_optimized_kprobe(op);
401 arch_remove_kprobe(p);
402 kfree(op);
403}
404
387/* Return true(!0) if the kprobe is ready for optimization. */ 405/* Return true(!0) if the kprobe is ready for optimization. */
388static inline int kprobe_optready(struct kprobe *p) 406static inline int kprobe_optready(struct kprobe *p)
389{ 407{
@@ -397,6 +415,33 @@ static inline int kprobe_optready(struct kprobe *p)
397 return 0; 415 return 0;
398} 416}
399 417
418/* Return true(!0) if the kprobe is disarmed. Note: p must be on hash list */
419static inline int kprobe_disarmed(struct kprobe *p)
420{
421 struct optimized_kprobe *op;
422
423 /* If kprobe is not aggr/opt probe, just return kprobe is disabled */
424 if (!kprobe_aggrprobe(p))
425 return kprobe_disabled(p);
426
427 op = container_of(p, struct optimized_kprobe, kp);
428
429 return kprobe_disabled(p) && list_empty(&op->list);
430}
431
432/* Return true(!0) if the probe is queued on (un)optimizing lists */
433static int __kprobes kprobe_queued(struct kprobe *p)
434{
435 struct optimized_kprobe *op;
436
437 if (kprobe_aggrprobe(p)) {
438 op = container_of(p, struct optimized_kprobe, kp);
439 if (!list_empty(&op->list))
440 return 1;
441 }
442 return 0;
443}
444
400/* 445/*
401 * Return an optimized kprobe whose optimizing code replaces 446 * Return an optimized kprobe whose optimizing code replaces
402 * instructions including addr (exclude breakpoint). 447 * instructions including addr (exclude breakpoint).
@@ -422,9 +467,11 @@ static struct kprobe *__kprobes get_optimized_kprobe(unsigned long addr)
422 467
423/* Optimization staging list, protected by kprobe_mutex */ 468/* Optimization staging list, protected by kprobe_mutex */
424static LIST_HEAD(optimizing_list); 469static LIST_HEAD(optimizing_list);
470static LIST_HEAD(unoptimizing_list);
425 471
426static void kprobe_optimizer(struct work_struct *work); 472static void kprobe_optimizer(struct work_struct *work);
427static DECLARE_DELAYED_WORK(optimizing_work, kprobe_optimizer); 473static DECLARE_DELAYED_WORK(optimizing_work, kprobe_optimizer);
474static DECLARE_COMPLETION(optimizer_comp);
428#define OPTIMIZE_DELAY 5 475#define OPTIMIZE_DELAY 5
429 476
430/* 477/*
@@ -435,6 +482,11 @@ static __kprobes void do_optimize_kprobes(void)
435{ 482{
436 struct optimized_kprobe *op, *tmp; 483 struct optimized_kprobe *op, *tmp;
437 484
485 /* Optimization never be done when disarmed */
486 if (kprobes_all_disarmed || !kprobes_allow_optimization ||
487 list_empty(&optimizing_list))
488 return;
489
438 /* 490 /*
439 * The optimization/unoptimization refers online_cpus via 491 * The optimization/unoptimization refers online_cpus via
440 * stop_machine() and cpu-hotplug modifies online_cpus. 492 * stop_machine() and cpu-hotplug modifies online_cpus.
@@ -457,17 +509,79 @@ static __kprobes void do_optimize_kprobes(void)
457 put_online_cpus(); 509 put_online_cpus();
458} 510}
459 511
512/*
513 * Unoptimize (replace a jump with a breakpoint and remove the breakpoint
514 * if need) kprobes listed on unoptimizing_list.
515 */
516static __kprobes void do_unoptimize_kprobes(struct list_head *free_list)
517{
518 struct optimized_kprobe *op, *tmp;
519
520 /* Unoptimization must be done anytime */
521 if (list_empty(&unoptimizing_list))
522 return;
523
524 /* Ditto to do_optimize_kprobes */
525 get_online_cpus();
526 mutex_lock(&text_mutex);
527 list_for_each_entry_safe(op, tmp, &unoptimizing_list, list) {
528 /* Unoptimize kprobes */
529 arch_unoptimize_kprobe(op);
530 /* Disarm probes if marked disabled */
531 if (kprobe_disabled(&op->kp))
532 arch_disarm_kprobe(&op->kp);
533 if (kprobe_unused(&op->kp)) {
534 /*
535 * Remove unused probes from hash list. After waiting
536 * for synchronization, these probes are reclaimed.
537 * (reclaiming is done by do_free_cleaned_kprobes.)
538 */
539 hlist_del_rcu(&op->kp.hlist);
540 /* Move only unused probes on free_list */
541 list_move(&op->list, free_list);
542 } else
543 list_del_init(&op->list);
544 }
545 mutex_unlock(&text_mutex);
546 put_online_cpus();
547}
548
549/* Reclaim all kprobes on the free_list */
550static __kprobes void do_free_cleaned_kprobes(struct list_head *free_list)
551{
552 struct optimized_kprobe *op, *tmp;
553
554 list_for_each_entry_safe(op, tmp, free_list, list) {
555 BUG_ON(!kprobe_unused(&op->kp));
556 list_del_init(&op->list);
557 free_aggr_kprobe(&op->kp);
558 }
559}
560
561/* Start optimizer after OPTIMIZE_DELAY passed */
562static __kprobes void kick_kprobe_optimizer(void)
563{
564 if (!delayed_work_pending(&optimizing_work))
565 schedule_delayed_work(&optimizing_work, OPTIMIZE_DELAY);
566}
567
460/* Kprobe jump optimizer */ 568/* Kprobe jump optimizer */
461static __kprobes void kprobe_optimizer(struct work_struct *work) 569static __kprobes void kprobe_optimizer(struct work_struct *work)
462{ 570{
571 LIST_HEAD(free_list);
572
463 /* Lock modules while optimizing kprobes */ 573 /* Lock modules while optimizing kprobes */
464 mutex_lock(&module_mutex); 574 mutex_lock(&module_mutex);
465 mutex_lock(&kprobe_mutex); 575 mutex_lock(&kprobe_mutex);
466 if (kprobes_all_disarmed || !kprobes_allow_optimization)
467 goto end;
468 576
469 /* 577 /*
470 * Wait for quiesence period to ensure all running interrupts 578 * Step 1: Unoptimize kprobes and collect cleaned (unused and disarmed)
579 * kprobes before waiting for quiesence period.
580 */
581 do_unoptimize_kprobes(&free_list);
582
583 /*
584 * Step 2: Wait for quiesence period to ensure all running interrupts
471 * are done. Because optprobe may modify multiple instructions 585 * are done. Because optprobe may modify multiple instructions
472 * there is a chance that Nth instruction is interrupted. In that 586 * there is a chance that Nth instruction is interrupted. In that
473 * case, running interrupt can return to 2nd-Nth byte of jump 587 * case, running interrupt can return to 2nd-Nth byte of jump
@@ -475,10 +589,24 @@ static __kprobes void kprobe_optimizer(struct work_struct *work)
475 */ 589 */
476 synchronize_sched(); 590 synchronize_sched();
477 591
592 /* Step 3: Optimize kprobes after quiesence period */
478 do_optimize_kprobes(); 593 do_optimize_kprobes();
479end: 594
595 /* Step 4: Free cleaned kprobes after quiesence period */
596 do_free_cleaned_kprobes(&free_list);
597
480 mutex_unlock(&kprobe_mutex); 598 mutex_unlock(&kprobe_mutex);
481 mutex_unlock(&module_mutex); 599 mutex_unlock(&module_mutex);
600
601 /* Wake up all waiters */
602 complete_all(&optimizer_comp);
603}
604
605/* Wait for completing optimization and unoptimization */
606static __kprobes void wait_for_kprobe_optimizer(void)
607{
608 if (delayed_work_pending(&optimizing_work))
609 wait_for_completion(&optimizer_comp);
482} 610}
483 611
484/* Optimize kprobe if p is ready to be optimized */ 612/* Optimize kprobe if p is ready to be optimized */
@@ -504,27 +632,63 @@ static __kprobes void optimize_kprobe(struct kprobe *p)
504 /* Check if it is already optimized. */ 632 /* Check if it is already optimized. */
505 if (op->kp.flags & KPROBE_FLAG_OPTIMIZED) 633 if (op->kp.flags & KPROBE_FLAG_OPTIMIZED)
506 return; 634 return;
507
508 op->kp.flags |= KPROBE_FLAG_OPTIMIZED; 635 op->kp.flags |= KPROBE_FLAG_OPTIMIZED;
509 list_add(&op->list, &optimizing_list); 636
510 if (!delayed_work_pending(&optimizing_work)) 637 if (!list_empty(&op->list))
511 schedule_delayed_work(&optimizing_work, OPTIMIZE_DELAY); 638 /* This is under unoptimizing. Just dequeue the probe */
639 list_del_init(&op->list);
640 else {
641 list_add(&op->list, &optimizing_list);
642 kick_kprobe_optimizer();
643 }
644}
645
646/* Short cut to direct unoptimizing */
647static __kprobes void force_unoptimize_kprobe(struct optimized_kprobe *op)
648{
649 get_online_cpus();
650 arch_unoptimize_kprobe(op);
651 put_online_cpus();
652 if (kprobe_disabled(&op->kp))
653 arch_disarm_kprobe(&op->kp);
512} 654}
513 655
514/* Unoptimize a kprobe if p is optimized */ 656/* Unoptimize a kprobe if p is optimized */
515static __kprobes void unoptimize_kprobe(struct kprobe *p) 657static __kprobes void unoptimize_kprobe(struct kprobe *p, bool force)
516{ 658{
517 struct optimized_kprobe *op; 659 struct optimized_kprobe *op;
518 660
519 if ((p->flags & KPROBE_FLAG_OPTIMIZED) && kprobe_aggrprobe(p)) { 661 if (!kprobe_aggrprobe(p) || kprobe_disarmed(p))
520 op = container_of(p, struct optimized_kprobe, kp); 662 return; /* This is not an optprobe nor optimized */
521 if (!list_empty(&op->list)) 663
522 /* Dequeue from the optimization queue */ 664 op = container_of(p, struct optimized_kprobe, kp);
665 if (!kprobe_optimized(p)) {
666 /* Unoptimized or unoptimizing case */
667 if (force && !list_empty(&op->list)) {
668 /*
669 * Only if this is unoptimizing kprobe and forced,
670 * forcibly unoptimize it. (No need to unoptimize
671 * unoptimized kprobe again :)
672 */
523 list_del_init(&op->list); 673 list_del_init(&op->list);
524 else 674 force_unoptimize_kprobe(op);
525 /* Replace jump with break */ 675 }
526 arch_unoptimize_kprobe(op); 676 return;
527 op->kp.flags &= ~KPROBE_FLAG_OPTIMIZED; 677 }
678
679 op->kp.flags &= ~KPROBE_FLAG_OPTIMIZED;
680 if (!list_empty(&op->list)) {
681 /* Dequeue from the optimization queue */
682 list_del_init(&op->list);
683 return;
684 }
685 /* Optimized kprobe case */
686 if (force)
687 /* Forcibly update the code: this is a special case */
688 force_unoptimize_kprobe(op);
689 else {
690 list_add(&op->list, &unoptimizing_list);
691 kick_kprobe_optimizer();
528 } 692 }
529} 693}
530 694
@@ -534,12 +698,12 @@ static void __kprobes kill_optimized_kprobe(struct kprobe *p)
534 struct optimized_kprobe *op; 698 struct optimized_kprobe *op;
535 699
536 op = container_of(p, struct optimized_kprobe, kp); 700 op = container_of(p, struct optimized_kprobe, kp);
537 if (!list_empty(&op->list)) { 701 if (!list_empty(&op->list))
538 /* Dequeue from the optimization queue */ 702 /* Dequeue from the (un)optimization queue */
539 list_del_init(&op->list); 703 list_del_init(&op->list);
540 op->kp.flags &= ~KPROBE_FLAG_OPTIMIZED; 704
541 } 705 op->kp.flags &= ~KPROBE_FLAG_OPTIMIZED;
542 /* Don't unoptimize, because the target code will be freed. */ 706 /* Don't touch the code, because it is already freed. */
543 arch_remove_optimized_kprobe(op); 707 arch_remove_optimized_kprobe(op);
544} 708}
545 709
@@ -552,16 +716,6 @@ static __kprobes void prepare_optimized_kprobe(struct kprobe *p)
552 arch_prepare_optimized_kprobe(op); 716 arch_prepare_optimized_kprobe(op);
553} 717}
554 718
555/* Free optimized instructions and optimized_kprobe */
556static __kprobes void free_aggr_kprobe(struct kprobe *p)
557{
558 struct optimized_kprobe *op;
559
560 op = container_of(p, struct optimized_kprobe, kp);
561 arch_remove_optimized_kprobe(op);
562 kfree(op);
563}
564
565/* Allocate new optimized_kprobe and try to prepare optimized instructions */ 719/* Allocate new optimized_kprobe and try to prepare optimized instructions */
566static __kprobes struct kprobe *alloc_aggr_kprobe(struct kprobe *p) 720static __kprobes struct kprobe *alloc_aggr_kprobe(struct kprobe *p)
567{ 721{
@@ -596,7 +750,8 @@ static __kprobes void try_to_optimize_kprobe(struct kprobe *p)
596 op = container_of(ap, struct optimized_kprobe, kp); 750 op = container_of(ap, struct optimized_kprobe, kp);
597 if (!arch_prepared_optinsn(&op->optinsn)) { 751 if (!arch_prepared_optinsn(&op->optinsn)) {
598 /* If failed to setup optimizing, fallback to kprobe */ 752 /* If failed to setup optimizing, fallback to kprobe */
599 free_aggr_kprobe(ap); 753 arch_remove_optimized_kprobe(op);
754 kfree(op);
600 return; 755 return;
601 } 756 }
602 757
@@ -640,21 +795,16 @@ static void __kprobes unoptimize_all_kprobes(void)
640 return; 795 return;
641 796
642 kprobes_allow_optimization = false; 797 kprobes_allow_optimization = false;
643 printk(KERN_INFO "Kprobes globally unoptimized\n");
644 get_online_cpus(); /* For avoiding text_mutex deadlock */
645 mutex_lock(&text_mutex);
646 for (i = 0; i < KPROBE_TABLE_SIZE; i++) { 798 for (i = 0; i < KPROBE_TABLE_SIZE; i++) {
647 head = &kprobe_table[i]; 799 head = &kprobe_table[i];
648 hlist_for_each_entry_rcu(p, node, head, hlist) { 800 hlist_for_each_entry_rcu(p, node, head, hlist) {
649 if (!kprobe_disabled(p)) 801 if (!kprobe_disabled(p))
650 unoptimize_kprobe(p); 802 unoptimize_kprobe(p, false);
651 } 803 }
652 } 804 }
653 805 /* Wait for unoptimizing completion */
654 mutex_unlock(&text_mutex); 806 wait_for_kprobe_optimizer();
655 put_online_cpus(); 807 printk(KERN_INFO "Kprobes globally unoptimized\n");
656 /* Allow all currently running kprobes to complete */
657 synchronize_sched();
658} 808}
659 809
660int sysctl_kprobes_optimization; 810int sysctl_kprobes_optimization;
@@ -678,6 +828,7 @@ int proc_kprobes_optimization_handler(struct ctl_table *table, int write,
678} 828}
679#endif /* CONFIG_SYSCTL */ 829#endif /* CONFIG_SYSCTL */
680 830
831/* Put a breakpoint for a probe. Must be called with text_mutex locked */
681static void __kprobes __arm_kprobe(struct kprobe *p) 832static void __kprobes __arm_kprobe(struct kprobe *p)
682{ 833{
683 struct kprobe *_p; 834 struct kprobe *_p;
@@ -685,37 +836,45 @@ static void __kprobes __arm_kprobe(struct kprobe *p)
685 /* Check collision with other optimized kprobes */ 836 /* Check collision with other optimized kprobes */
686 _p = get_optimized_kprobe((unsigned long)p->addr); 837 _p = get_optimized_kprobe((unsigned long)p->addr);
687 if (unlikely(_p)) 838 if (unlikely(_p))
688 unoptimize_kprobe(_p); /* Fallback to unoptimized kprobe */ 839 /* Fallback to unoptimized kprobe */
840 unoptimize_kprobe(_p, true);
689 841
690 arch_arm_kprobe(p); 842 arch_arm_kprobe(p);
691 optimize_kprobe(p); /* Try to optimize (add kprobe to a list) */ 843 optimize_kprobe(p); /* Try to optimize (add kprobe to a list) */
692} 844}
693 845
694static void __kprobes __disarm_kprobe(struct kprobe *p) 846/* Remove the breakpoint of a probe. Must be called with text_mutex locked */
847static void __kprobes __disarm_kprobe(struct kprobe *p, bool reopt)
695{ 848{
696 struct kprobe *_p; 849 struct kprobe *_p;
697 850
698 unoptimize_kprobe(p); /* Try to unoptimize */ 851 unoptimize_kprobe(p, false); /* Try to unoptimize */
699 arch_disarm_kprobe(p);
700 852
701 /* If another kprobe was blocked, optimize it. */ 853 if (!kprobe_queued(p)) {
702 _p = get_optimized_kprobe((unsigned long)p->addr); 854 arch_disarm_kprobe(p);
703 if (unlikely(_p)) 855 /* If another kprobe was blocked, optimize it. */
704 optimize_kprobe(_p); 856 _p = get_optimized_kprobe((unsigned long)p->addr);
857 if (unlikely(_p) && reopt)
858 optimize_kprobe(_p);
859 }
860 /* TODO: reoptimize others after unoptimized this probe */
705} 861}
706 862
707#else /* !CONFIG_OPTPROBES */ 863#else /* !CONFIG_OPTPROBES */
708 864
709#define optimize_kprobe(p) do {} while (0) 865#define optimize_kprobe(p) do {} while (0)
710#define unoptimize_kprobe(p) do {} while (0) 866#define unoptimize_kprobe(p, f) do {} while (0)
711#define kill_optimized_kprobe(p) do {} while (0) 867#define kill_optimized_kprobe(p) do {} while (0)
712#define prepare_optimized_kprobe(p) do {} while (0) 868#define prepare_optimized_kprobe(p) do {} while (0)
713#define try_to_optimize_kprobe(p) do {} while (0) 869#define try_to_optimize_kprobe(p) do {} while (0)
714#define __arm_kprobe(p) arch_arm_kprobe(p) 870#define __arm_kprobe(p) arch_arm_kprobe(p)
715#define __disarm_kprobe(p) arch_disarm_kprobe(p) 871#define __disarm_kprobe(p, o) arch_disarm_kprobe(p)
872#define kprobe_disarmed(p) kprobe_disabled(p)
873#define wait_for_kprobe_optimizer() do {} while (0)
716 874
717static __kprobes void free_aggr_kprobe(struct kprobe *p) 875static __kprobes void free_aggr_kprobe(struct kprobe *p)
718{ 876{
877 arch_remove_kprobe(p);
719 kfree(p); 878 kfree(p);
720} 879}
721 880
@@ -741,11 +900,10 @@ static void __kprobes arm_kprobe(struct kprobe *kp)
741/* Disarm a kprobe with text_mutex */ 900/* Disarm a kprobe with text_mutex */
742static void __kprobes disarm_kprobe(struct kprobe *kp) 901static void __kprobes disarm_kprobe(struct kprobe *kp)
743{ 902{
744 get_online_cpus(); /* For avoiding text_mutex deadlock */ 903 /* Ditto */
745 mutex_lock(&text_mutex); 904 mutex_lock(&text_mutex);
746 __disarm_kprobe(kp); 905 __disarm_kprobe(kp, true);
747 mutex_unlock(&text_mutex); 906 mutex_unlock(&text_mutex);
748 put_online_cpus();
749} 907}
750 908
751/* 909/*
@@ -951,7 +1109,7 @@ static int __kprobes add_new_kprobe(struct kprobe *ap, struct kprobe *p)
951 BUG_ON(kprobe_gone(ap) || kprobe_gone(p)); 1109 BUG_ON(kprobe_gone(ap) || kprobe_gone(p));
952 1110
953 if (p->break_handler || p->post_handler) 1111 if (p->break_handler || p->post_handler)
954 unoptimize_kprobe(ap); /* Fall back to normal kprobe */ 1112 unoptimize_kprobe(ap, true); /* Fall back to normal kprobe */
955 1113
956 if (p->break_handler) { 1114 if (p->break_handler) {
957 if (ap->break_handler) 1115 if (ap->break_handler)
@@ -1014,7 +1172,9 @@ static int __kprobes register_aggr_kprobe(struct kprobe *orig_p,
1014 if (!ap) 1172 if (!ap)
1015 return -ENOMEM; 1173 return -ENOMEM;
1016 init_aggr_kprobe(ap, orig_p); 1174 init_aggr_kprobe(ap, orig_p);
1017 } 1175 } else if (kprobe_unused(ap))
1176 /* Busy to die */
1177 return -EBUSY;
1018 1178
1019 if (kprobe_gone(ap)) { 1179 if (kprobe_gone(ap)) {
1020 /* 1180 /*
@@ -1283,8 +1443,11 @@ static int __kprobes __unregister_kprobe_top(struct kprobe *p)
1283 /* Following process expects this probe is an aggrprobe */ 1443 /* Following process expects this probe is an aggrprobe */
1284 WARN_ON(!kprobe_aggrprobe(ap)); 1444 WARN_ON(!kprobe_aggrprobe(ap));
1285 1445
1286 if (list_is_singular(&ap->list)) 1446 if (list_is_singular(&ap->list) && kprobe_disarmed(ap))
1287 /* This probe is the last child of aggrprobe */ 1447 /*
1448 * !disarmed could be happen if the probe is under delayed
1449 * unoptimizing.
1450 */
1288 goto disarmed; 1451 goto disarmed;
1289 else { 1452 else {
1290 /* If disabling probe has special handlers, update aggrprobe */ 1453 /* If disabling probe has special handlers, update aggrprobe */
@@ -1313,6 +1476,7 @@ noclean:
1313 return 0; 1476 return 0;
1314 1477
1315disarmed: 1478disarmed:
1479 BUG_ON(!kprobe_disarmed(ap));
1316 hlist_del_rcu(&ap->hlist); 1480 hlist_del_rcu(&ap->hlist);
1317 return 0; 1481 return 0;
1318} 1482}
@@ -1322,14 +1486,15 @@ static void __kprobes __unregister_kprobe_bottom(struct kprobe *p)
1322 struct kprobe *ap; 1486 struct kprobe *ap;
1323 1487
1324 if (list_empty(&p->list)) 1488 if (list_empty(&p->list))
1489 /* This is an independent kprobe */
1325 arch_remove_kprobe(p); 1490 arch_remove_kprobe(p);
1326 else if (list_is_singular(&p->list)) { 1491 else if (list_is_singular(&p->list)) {
1327 /* "p" is the last child of an aggr_kprobe */ 1492 /* This is the last child of an aggrprobe */
1328 ap = list_entry(p->list.next, struct kprobe, list); 1493 ap = list_entry(p->list.next, struct kprobe, list);
1329 list_del(&p->list); 1494 list_del(&p->list);
1330 arch_remove_kprobe(ap);
1331 free_aggr_kprobe(ap); 1495 free_aggr_kprobe(ap);
1332 } 1496 }
1497 /* Otherwise, do nothing. */
1333} 1498}
1334 1499
1335int __kprobes register_kprobes(struct kprobe **kps, int num) 1500int __kprobes register_kprobes(struct kprobe **kps, int num)
@@ -1951,36 +2116,27 @@ static void __kprobes disarm_all_kprobes(void)
1951 mutex_lock(&kprobe_mutex); 2116 mutex_lock(&kprobe_mutex);
1952 2117
1953 /* If kprobes are already disarmed, just return */ 2118 /* If kprobes are already disarmed, just return */
1954 if (kprobes_all_disarmed) 2119 if (kprobes_all_disarmed) {
1955 goto already_disabled; 2120 mutex_unlock(&kprobe_mutex);
2121 return;
2122 }
1956 2123
1957 kprobes_all_disarmed = true; 2124 kprobes_all_disarmed = true;
1958 printk(KERN_INFO "Kprobes globally disabled\n"); 2125 printk(KERN_INFO "Kprobes globally disabled\n");
1959 2126
1960 /*
1961 * Here we call get_online_cpus() for avoiding text_mutex deadlock,
1962 * because disarming may also unoptimize kprobes.
1963 */
1964 get_online_cpus();
1965 mutex_lock(&text_mutex); 2127 mutex_lock(&text_mutex);
1966 for (i = 0; i < KPROBE_TABLE_SIZE; i++) { 2128 for (i = 0; i < KPROBE_TABLE_SIZE; i++) {
1967 head = &kprobe_table[i]; 2129 head = &kprobe_table[i];
1968 hlist_for_each_entry_rcu(p, node, head, hlist) { 2130 hlist_for_each_entry_rcu(p, node, head, hlist) {
1969 if (!arch_trampoline_kprobe(p) && !kprobe_disabled(p)) 2131 if (!arch_trampoline_kprobe(p) && !kprobe_disabled(p))
1970 __disarm_kprobe(p); 2132 __disarm_kprobe(p, false);
1971 } 2133 }
1972 } 2134 }
1973
1974 mutex_unlock(&text_mutex); 2135 mutex_unlock(&text_mutex);
1975 put_online_cpus();
1976 mutex_unlock(&kprobe_mutex); 2136 mutex_unlock(&kprobe_mutex);
1977 /* Allow all currently running kprobes to complete */
1978 synchronize_sched();
1979 return;
1980 2137
1981already_disabled: 2138 /* Wait for disarming all kprobes by optimizer */
1982 mutex_unlock(&kprobe_mutex); 2139 wait_for_kprobe_optimizer();
1983 return;
1984} 2140}
1985 2141
1986/* 2142/*