diff options
Diffstat (limited to 'arch/x86')
-rw-r--r-- | arch/x86/kernel/ftrace.c | 69 | ||||
-rw-r--r-- | arch/x86/kernel/kprobes.c | 17 |
2 files changed, 50 insertions, 36 deletions
diff --git a/arch/x86/kernel/ftrace.c b/arch/x86/kernel/ftrace.c index 1d0d7f42efe3..61df77532120 100644 --- a/arch/x86/kernel/ftrace.c +++ b/arch/x86/kernel/ftrace.c | |||
@@ -79,11 +79,11 @@ static unsigned char *ftrace_call_replace(unsigned long ip, unsigned long addr) | |||
79 | * | 79 | * |
80 | * 1) Put the instruction pointer into the IP buffer | 80 | * 1) Put the instruction pointer into the IP buffer |
81 | * and the new code into the "code" buffer. | 81 | * and the new code into the "code" buffer. |
82 | * 2) Set a flag that says we are modifying code | 82 | * 2) Wait for any running NMIs to finish and set a flag that says |
83 | * 3) Wait for any running NMIs to finish. | 83 | * we are modifying code, it is done in an atomic operation. |
84 | * 4) Write the code | 84 | * 3) Write the code |
85 | * 5) clear the flag. | 85 | * 4) clear the flag. |
86 | * 6) Wait for any running NMIs to finish. | 86 | * 5) Wait for any running NMIs to finish. |
87 | * | 87 | * |
88 | * If an NMI is executed, the first thing it does is to call | 88 | * If an NMI is executed, the first thing it does is to call |
89 | * "ftrace_nmi_enter". This will check if the flag is set to write | 89 | * "ftrace_nmi_enter". This will check if the flag is set to write |
@@ -95,9 +95,9 @@ static unsigned char *ftrace_call_replace(unsigned long ip, unsigned long addr) | |||
95 | * are the same as what exists. | 95 | * are the same as what exists. |
96 | */ | 96 | */ |
97 | 97 | ||
98 | #define MOD_CODE_WRITE_FLAG (1 << 31) /* set when NMI should do the write */ | ||
98 | static atomic_t nmi_running = ATOMIC_INIT(0); | 99 | static atomic_t nmi_running = ATOMIC_INIT(0); |
99 | static int mod_code_status; /* holds return value of text write */ | 100 | static int mod_code_status; /* holds return value of text write */ |
100 | static int mod_code_write; /* set when NMI should do the write */ | ||
101 | static void *mod_code_ip; /* holds the IP to write to */ | 101 | static void *mod_code_ip; /* holds the IP to write to */ |
102 | static void *mod_code_newcode; /* holds the text to write to the IP */ | 102 | static void *mod_code_newcode; /* holds the text to write to the IP */ |
103 | 103 | ||
@@ -114,6 +114,20 @@ int ftrace_arch_read_dyn_info(char *buf, int size) | |||
114 | return r; | 114 | return r; |
115 | } | 115 | } |
116 | 116 | ||
117 | static void clear_mod_flag(void) | ||
118 | { | ||
119 | int old = atomic_read(&nmi_running); | ||
120 | |||
121 | for (;;) { | ||
122 | int new = old & ~MOD_CODE_WRITE_FLAG; | ||
123 | |||
124 | if (old == new) | ||
125 | break; | ||
126 | |||
127 | old = atomic_cmpxchg(&nmi_running, old, new); | ||
128 | } | ||
129 | } | ||
130 | |||
117 | static void ftrace_mod_code(void) | 131 | static void ftrace_mod_code(void) |
118 | { | 132 | { |
119 | /* | 133 | /* |
@@ -127,27 +141,39 @@ static void ftrace_mod_code(void) | |||
127 | 141 | ||
128 | /* if we fail, then kill any new writers */ | 142 | /* if we fail, then kill any new writers */ |
129 | if (mod_code_status) | 143 | if (mod_code_status) |
130 | mod_code_write = 0; | 144 | clear_mod_flag(); |
131 | } | 145 | } |
132 | 146 | ||
133 | void ftrace_nmi_enter(void) | 147 | void ftrace_nmi_enter(void) |
134 | { | 148 | { |
135 | atomic_inc(&nmi_running); | 149 | if (atomic_inc_return(&nmi_running) & MOD_CODE_WRITE_FLAG) { |
136 | /* Must have nmi_running seen before reading write flag */ | 150 | smp_rmb(); |
137 | smp_mb(); | ||
138 | if (mod_code_write) { | ||
139 | ftrace_mod_code(); | 151 | ftrace_mod_code(); |
140 | atomic_inc(&nmi_update_count); | 152 | atomic_inc(&nmi_update_count); |
141 | } | 153 | } |
154 | /* Must have previous changes seen before executions */ | ||
155 | smp_mb(); | ||
142 | } | 156 | } |
143 | 157 | ||
144 | void ftrace_nmi_exit(void) | 158 | void ftrace_nmi_exit(void) |
145 | { | 159 | { |
146 | /* Finish all executions before clearing nmi_running */ | 160 | /* Finish all executions before clearing nmi_running */ |
147 | smp_wmb(); | 161 | smp_mb(); |
148 | atomic_dec(&nmi_running); | 162 | atomic_dec(&nmi_running); |
149 | } | 163 | } |
150 | 164 | ||
165 | static void wait_for_nmi_and_set_mod_flag(void) | ||
166 | { | ||
167 | if (!atomic_cmpxchg(&nmi_running, 0, MOD_CODE_WRITE_FLAG)) | ||
168 | return; | ||
169 | |||
170 | do { | ||
171 | cpu_relax(); | ||
172 | } while (atomic_cmpxchg(&nmi_running, 0, MOD_CODE_WRITE_FLAG)); | ||
173 | |||
174 | nmi_wait_count++; | ||
175 | } | ||
176 | |||
151 | static void wait_for_nmi(void) | 177 | static void wait_for_nmi(void) |
152 | { | 178 | { |
153 | if (!atomic_read(&nmi_running)) | 179 | if (!atomic_read(&nmi_running)) |
@@ -167,14 +193,9 @@ do_ftrace_mod_code(unsigned long ip, void *new_code) | |||
167 | mod_code_newcode = new_code; | 193 | mod_code_newcode = new_code; |
168 | 194 | ||
169 | /* The buffers need to be visible before we let NMIs write them */ | 195 | /* The buffers need to be visible before we let NMIs write them */ |
170 | smp_wmb(); | ||
171 | |||
172 | mod_code_write = 1; | ||
173 | |||
174 | /* Make sure write bit is visible before we wait on NMIs */ | ||
175 | smp_mb(); | 196 | smp_mb(); |
176 | 197 | ||
177 | wait_for_nmi(); | 198 | wait_for_nmi_and_set_mod_flag(); |
178 | 199 | ||
179 | /* Make sure all running NMIs have finished before we write the code */ | 200 | /* Make sure all running NMIs have finished before we write the code */ |
180 | smp_mb(); | 201 | smp_mb(); |
@@ -182,13 +203,9 @@ do_ftrace_mod_code(unsigned long ip, void *new_code) | |||
182 | ftrace_mod_code(); | 203 | ftrace_mod_code(); |
183 | 204 | ||
184 | /* Make sure the write happens before clearing the bit */ | 205 | /* Make sure the write happens before clearing the bit */ |
185 | smp_wmb(); | ||
186 | |||
187 | mod_code_write = 0; | ||
188 | |||
189 | /* make sure NMIs see the cleared bit */ | ||
190 | smp_mb(); | 206 | smp_mb(); |
191 | 207 | ||
208 | clear_mod_flag(); | ||
192 | wait_for_nmi(); | 209 | wait_for_nmi(); |
193 | 210 | ||
194 | return mod_code_status; | 211 | return mod_code_status; |
@@ -393,7 +410,6 @@ int ftrace_disable_ftrace_graph_caller(void) | |||
393 | void prepare_ftrace_return(unsigned long *parent, unsigned long self_addr) | 410 | void prepare_ftrace_return(unsigned long *parent, unsigned long self_addr) |
394 | { | 411 | { |
395 | unsigned long old; | 412 | unsigned long old; |
396 | unsigned long long calltime; | ||
397 | int faulted; | 413 | int faulted; |
398 | struct ftrace_graph_ent trace; | 414 | struct ftrace_graph_ent trace; |
399 | unsigned long return_hooker = (unsigned long) | 415 | unsigned long return_hooker = (unsigned long) |
@@ -436,10 +452,7 @@ void prepare_ftrace_return(unsigned long *parent, unsigned long self_addr) | |||
436 | return; | 452 | return; |
437 | } | 453 | } |
438 | 454 | ||
439 | calltime = trace_clock_local(); | 455 | if (ftrace_push_return_trace(old, self_addr, &trace.depth) == -EBUSY) { |
440 | |||
441 | if (ftrace_push_return_trace(old, calltime, | ||
442 | self_addr, &trace.depth) == -EBUSY) { | ||
443 | *parent = old; | 456 | *parent = old; |
444 | return; | 457 | return; |
445 | } | 458 | } |
diff --git a/arch/x86/kernel/kprobes.c b/arch/x86/kernel/kprobes.c index 4558dd3918cf..759095d53a06 100644 --- a/arch/x86/kernel/kprobes.c +++ b/arch/x86/kernel/kprobes.c | |||
@@ -638,13 +638,13 @@ static void __used __kprobes kretprobe_trampoline_holder(void) | |||
638 | #else | 638 | #else |
639 | " pushf\n" | 639 | " pushf\n" |
640 | /* | 640 | /* |
641 | * Skip cs, ip, orig_ax. | 641 | * Skip cs, ip, orig_ax and gs. |
642 | * trampoline_handler() will plug in these values | 642 | * trampoline_handler() will plug in these values |
643 | */ | 643 | */ |
644 | " subl $12, %esp\n" | 644 | " subl $16, %esp\n" |
645 | " pushl %fs\n" | 645 | " pushl %fs\n" |
646 | " pushl %ds\n" | ||
647 | " pushl %es\n" | 646 | " pushl %es\n" |
647 | " pushl %ds\n" | ||
648 | " pushl %eax\n" | 648 | " pushl %eax\n" |
649 | " pushl %ebp\n" | 649 | " pushl %ebp\n" |
650 | " pushl %edi\n" | 650 | " pushl %edi\n" |
@@ -655,10 +655,10 @@ static void __used __kprobes kretprobe_trampoline_holder(void) | |||
655 | " movl %esp, %eax\n" | 655 | " movl %esp, %eax\n" |
656 | " call trampoline_handler\n" | 656 | " call trampoline_handler\n" |
657 | /* Move flags to cs */ | 657 | /* Move flags to cs */ |
658 | " movl 52(%esp), %edx\n" | 658 | " movl 56(%esp), %edx\n" |
659 | " movl %edx, 48(%esp)\n" | 659 | " movl %edx, 52(%esp)\n" |
660 | /* Replace saved flags with true return address. */ | 660 | /* Replace saved flags with true return address. */ |
661 | " movl %eax, 52(%esp)\n" | 661 | " movl %eax, 56(%esp)\n" |
662 | " popl %ebx\n" | 662 | " popl %ebx\n" |
663 | " popl %ecx\n" | 663 | " popl %ecx\n" |
664 | " popl %edx\n" | 664 | " popl %edx\n" |
@@ -666,8 +666,8 @@ static void __used __kprobes kretprobe_trampoline_holder(void) | |||
666 | " popl %edi\n" | 666 | " popl %edi\n" |
667 | " popl %ebp\n" | 667 | " popl %ebp\n" |
668 | " popl %eax\n" | 668 | " popl %eax\n" |
669 | /* Skip ip, orig_ax, es, ds, fs */ | 669 | /* Skip ds, es, fs, gs, orig_ax and ip */ |
670 | " addl $20, %esp\n" | 670 | " addl $24, %esp\n" |
671 | " popf\n" | 671 | " popf\n" |
672 | #endif | 672 | #endif |
673 | " ret\n"); | 673 | " ret\n"); |
@@ -691,6 +691,7 @@ static __used __kprobes void *trampoline_handler(struct pt_regs *regs) | |||
691 | regs->cs = __KERNEL_CS; | 691 | regs->cs = __KERNEL_CS; |
692 | #else | 692 | #else |
693 | regs->cs = __KERNEL_CS | get_kernel_rpl(); | 693 | regs->cs = __KERNEL_CS | get_kernel_rpl(); |
694 | regs->gs = 0; | ||
694 | #endif | 695 | #endif |
695 | regs->ip = trampoline_address; | 696 | regs->ip = trampoline_address; |
696 | regs->orig_ax = ~0UL; | 697 | regs->orig_ax = ~0UL; |