diff options
Diffstat (limited to 'kernel/trace/trace_syscalls.c')
-rw-r--r-- | kernel/trace/trace_syscalls.c | 146 |
1 files changed, 120 insertions, 26 deletions
diff --git a/kernel/trace/trace_syscalls.c b/kernel/trace/trace_syscalls.c index 4d6d711717f2..d2c859cec9ea 100644 --- a/kernel/trace/trace_syscalls.c +++ b/kernel/trace/trace_syscalls.c | |||
@@ -15,6 +15,54 @@ static int sys_refcount_exit; | |||
15 | static DECLARE_BITMAP(enabled_enter_syscalls, NR_syscalls); | 15 | static DECLARE_BITMAP(enabled_enter_syscalls, NR_syscalls); |
16 | static DECLARE_BITMAP(enabled_exit_syscalls, NR_syscalls); | 16 | static DECLARE_BITMAP(enabled_exit_syscalls, NR_syscalls); |
17 | 17 | ||
18 | static int syscall_enter_register(struct ftrace_event_call *event, | ||
19 | enum trace_reg type); | ||
20 | static int syscall_exit_register(struct ftrace_event_call *event, | ||
21 | enum trace_reg type); | ||
22 | |||
23 | static int syscall_enter_define_fields(struct ftrace_event_call *call); | ||
24 | static int syscall_exit_define_fields(struct ftrace_event_call *call); | ||
25 | |||
26 | static struct list_head * | ||
27 | syscall_get_enter_fields(struct ftrace_event_call *call) | ||
28 | { | ||
29 | struct syscall_metadata *entry = call->data; | ||
30 | |||
31 | return &entry->enter_fields; | ||
32 | } | ||
33 | |||
34 | static struct list_head * | ||
35 | syscall_get_exit_fields(struct ftrace_event_call *call) | ||
36 | { | ||
37 | struct syscall_metadata *entry = call->data; | ||
38 | |||
39 | return &entry->exit_fields; | ||
40 | } | ||
41 | |||
42 | struct trace_event_functions enter_syscall_print_funcs = { | ||
43 | .trace = print_syscall_enter, | ||
44 | }; | ||
45 | |||
46 | struct trace_event_functions exit_syscall_print_funcs = { | ||
47 | .trace = print_syscall_exit, | ||
48 | }; | ||
49 | |||
50 | struct ftrace_event_class event_class_syscall_enter = { | ||
51 | .system = "syscalls", | ||
52 | .reg = syscall_enter_register, | ||
53 | .define_fields = syscall_enter_define_fields, | ||
54 | .get_fields = syscall_get_enter_fields, | ||
55 | .raw_init = init_syscall_trace, | ||
56 | }; | ||
57 | |||
58 | struct ftrace_event_class event_class_syscall_exit = { | ||
59 | .system = "syscalls", | ||
60 | .reg = syscall_exit_register, | ||
61 | .define_fields = syscall_exit_define_fields, | ||
62 | .get_fields = syscall_get_exit_fields, | ||
63 | .raw_init = init_syscall_trace, | ||
64 | }; | ||
65 | |||
18 | extern unsigned long __start_syscalls_metadata[]; | 66 | extern unsigned long __start_syscalls_metadata[]; |
19 | extern unsigned long __stop_syscalls_metadata[]; | 67 | extern unsigned long __stop_syscalls_metadata[]; |
20 | 68 | ||
@@ -53,7 +101,8 @@ static struct syscall_metadata *syscall_nr_to_meta(int nr) | |||
53 | } | 101 | } |
54 | 102 | ||
55 | enum print_line_t | 103 | enum print_line_t |
56 | print_syscall_enter(struct trace_iterator *iter, int flags) | 104 | print_syscall_enter(struct trace_iterator *iter, int flags, |
105 | struct trace_event *event) | ||
57 | { | 106 | { |
58 | struct trace_seq *s = &iter->seq; | 107 | struct trace_seq *s = &iter->seq; |
59 | struct trace_entry *ent = iter->ent; | 108 | struct trace_entry *ent = iter->ent; |
@@ -68,7 +117,7 @@ print_syscall_enter(struct trace_iterator *iter, int flags) | |||
68 | if (!entry) | 117 | if (!entry) |
69 | goto end; | 118 | goto end; |
70 | 119 | ||
71 | if (entry->enter_event->id != ent->type) { | 120 | if (entry->enter_event->event.type != ent->type) { |
72 | WARN_ON_ONCE(1); | 121 | WARN_ON_ONCE(1); |
73 | goto end; | 122 | goto end; |
74 | } | 123 | } |
@@ -105,7 +154,8 @@ end: | |||
105 | } | 154 | } |
106 | 155 | ||
107 | enum print_line_t | 156 | enum print_line_t |
108 | print_syscall_exit(struct trace_iterator *iter, int flags) | 157 | print_syscall_exit(struct trace_iterator *iter, int flags, |
158 | struct trace_event *event) | ||
109 | { | 159 | { |
110 | struct trace_seq *s = &iter->seq; | 160 | struct trace_seq *s = &iter->seq; |
111 | struct trace_entry *ent = iter->ent; | 161 | struct trace_entry *ent = iter->ent; |
@@ -123,7 +173,7 @@ print_syscall_exit(struct trace_iterator *iter, int flags) | |||
123 | return TRACE_TYPE_HANDLED; | 173 | return TRACE_TYPE_HANDLED; |
124 | } | 174 | } |
125 | 175 | ||
126 | if (entry->exit_event->id != ent->type) { | 176 | if (entry->exit_event->event.type != ent->type) { |
127 | WARN_ON_ONCE(1); | 177 | WARN_ON_ONCE(1); |
128 | return TRACE_TYPE_UNHANDLED; | 178 | return TRACE_TYPE_UNHANDLED; |
129 | } | 179 | } |
@@ -205,7 +255,7 @@ static void free_syscall_print_fmt(struct ftrace_event_call *call) | |||
205 | kfree(call->print_fmt); | 255 | kfree(call->print_fmt); |
206 | } | 256 | } |
207 | 257 | ||
208 | int syscall_enter_define_fields(struct ftrace_event_call *call) | 258 | static int syscall_enter_define_fields(struct ftrace_event_call *call) |
209 | { | 259 | { |
210 | struct syscall_trace_enter trace; | 260 | struct syscall_trace_enter trace; |
211 | struct syscall_metadata *meta = call->data; | 261 | struct syscall_metadata *meta = call->data; |
@@ -228,7 +278,7 @@ int syscall_enter_define_fields(struct ftrace_event_call *call) | |||
228 | return ret; | 278 | return ret; |
229 | } | 279 | } |
230 | 280 | ||
231 | int syscall_exit_define_fields(struct ftrace_event_call *call) | 281 | static int syscall_exit_define_fields(struct ftrace_event_call *call) |
232 | { | 282 | { |
233 | struct syscall_trace_exit trace; | 283 | struct syscall_trace_exit trace; |
234 | int ret; | 284 | int ret; |
@@ -243,7 +293,7 @@ int syscall_exit_define_fields(struct ftrace_event_call *call) | |||
243 | return ret; | 293 | return ret; |
244 | } | 294 | } |
245 | 295 | ||
246 | void ftrace_syscall_enter(struct pt_regs *regs, long id) | 296 | void ftrace_syscall_enter(void *ignore, struct pt_regs *regs, long id) |
247 | { | 297 | { |
248 | struct syscall_trace_enter *entry; | 298 | struct syscall_trace_enter *entry; |
249 | struct syscall_metadata *sys_data; | 299 | struct syscall_metadata *sys_data; |
@@ -265,7 +315,7 @@ void ftrace_syscall_enter(struct pt_regs *regs, long id) | |||
265 | size = sizeof(*entry) + sizeof(unsigned long) * sys_data->nb_args; | 315 | size = sizeof(*entry) + sizeof(unsigned long) * sys_data->nb_args; |
266 | 316 | ||
267 | event = trace_current_buffer_lock_reserve(&buffer, | 317 | event = trace_current_buffer_lock_reserve(&buffer, |
268 | sys_data->enter_event->id, size, 0, 0); | 318 | sys_data->enter_event->event.type, size, 0, 0); |
269 | if (!event) | 319 | if (!event) |
270 | return; | 320 | return; |
271 | 321 | ||
@@ -278,7 +328,7 @@ void ftrace_syscall_enter(struct pt_regs *regs, long id) | |||
278 | trace_current_buffer_unlock_commit(buffer, event, 0, 0); | 328 | trace_current_buffer_unlock_commit(buffer, event, 0, 0); |
279 | } | 329 | } |
280 | 330 | ||
281 | void ftrace_syscall_exit(struct pt_regs *regs, long ret) | 331 | void ftrace_syscall_exit(void *ignore, struct pt_regs *regs, long ret) |
282 | { | 332 | { |
283 | struct syscall_trace_exit *entry; | 333 | struct syscall_trace_exit *entry; |
284 | struct syscall_metadata *sys_data; | 334 | struct syscall_metadata *sys_data; |
@@ -297,7 +347,7 @@ void ftrace_syscall_exit(struct pt_regs *regs, long ret) | |||
297 | return; | 347 | return; |
298 | 348 | ||
299 | event = trace_current_buffer_lock_reserve(&buffer, | 349 | event = trace_current_buffer_lock_reserve(&buffer, |
300 | sys_data->exit_event->id, sizeof(*entry), 0, 0); | 350 | sys_data->exit_event->event.type, sizeof(*entry), 0, 0); |
301 | if (!event) | 351 | if (!event) |
302 | return; | 352 | return; |
303 | 353 | ||
@@ -320,7 +370,7 @@ int reg_event_syscall_enter(struct ftrace_event_call *call) | |||
320 | return -ENOSYS; | 370 | return -ENOSYS; |
321 | mutex_lock(&syscall_trace_lock); | 371 | mutex_lock(&syscall_trace_lock); |
322 | if (!sys_refcount_enter) | 372 | if (!sys_refcount_enter) |
323 | ret = register_trace_sys_enter(ftrace_syscall_enter); | 373 | ret = register_trace_sys_enter(ftrace_syscall_enter, NULL); |
324 | if (!ret) { | 374 | if (!ret) { |
325 | set_bit(num, enabled_enter_syscalls); | 375 | set_bit(num, enabled_enter_syscalls); |
326 | sys_refcount_enter++; | 376 | sys_refcount_enter++; |
@@ -340,7 +390,7 @@ void unreg_event_syscall_enter(struct ftrace_event_call *call) | |||
340 | sys_refcount_enter--; | 390 | sys_refcount_enter--; |
341 | clear_bit(num, enabled_enter_syscalls); | 391 | clear_bit(num, enabled_enter_syscalls); |
342 | if (!sys_refcount_enter) | 392 | if (!sys_refcount_enter) |
343 | unregister_trace_sys_enter(ftrace_syscall_enter); | 393 | unregister_trace_sys_enter(ftrace_syscall_enter, NULL); |
344 | mutex_unlock(&syscall_trace_lock); | 394 | mutex_unlock(&syscall_trace_lock); |
345 | } | 395 | } |
346 | 396 | ||
@@ -354,7 +404,7 @@ int reg_event_syscall_exit(struct ftrace_event_call *call) | |||
354 | return -ENOSYS; | 404 | return -ENOSYS; |
355 | mutex_lock(&syscall_trace_lock); | 405 | mutex_lock(&syscall_trace_lock); |
356 | if (!sys_refcount_exit) | 406 | if (!sys_refcount_exit) |
357 | ret = register_trace_sys_exit(ftrace_syscall_exit); | 407 | ret = register_trace_sys_exit(ftrace_syscall_exit, NULL); |
358 | if (!ret) { | 408 | if (!ret) { |
359 | set_bit(num, enabled_exit_syscalls); | 409 | set_bit(num, enabled_exit_syscalls); |
360 | sys_refcount_exit++; | 410 | sys_refcount_exit++; |
@@ -374,7 +424,7 @@ void unreg_event_syscall_exit(struct ftrace_event_call *call) | |||
374 | sys_refcount_exit--; | 424 | sys_refcount_exit--; |
375 | clear_bit(num, enabled_exit_syscalls); | 425 | clear_bit(num, enabled_exit_syscalls); |
376 | if (!sys_refcount_exit) | 426 | if (!sys_refcount_exit) |
377 | unregister_trace_sys_exit(ftrace_syscall_exit); | 427 | unregister_trace_sys_exit(ftrace_syscall_exit, NULL); |
378 | mutex_unlock(&syscall_trace_lock); | 428 | mutex_unlock(&syscall_trace_lock); |
379 | } | 429 | } |
380 | 430 | ||
@@ -434,11 +484,11 @@ static DECLARE_BITMAP(enabled_perf_exit_syscalls, NR_syscalls); | |||
434 | static int sys_perf_refcount_enter; | 484 | static int sys_perf_refcount_enter; |
435 | static int sys_perf_refcount_exit; | 485 | static int sys_perf_refcount_exit; |
436 | 486 | ||
437 | static void perf_syscall_enter(struct pt_regs *regs, long id) | 487 | static void perf_syscall_enter(void *ignore, struct pt_regs *regs, long id) |
438 | { | 488 | { |
439 | struct syscall_metadata *sys_data; | 489 | struct syscall_metadata *sys_data; |
440 | struct syscall_trace_enter *rec; | 490 | struct syscall_trace_enter *rec; |
441 | unsigned long flags; | 491 | struct hlist_head *head; |
442 | int syscall_nr; | 492 | int syscall_nr; |
443 | int rctx; | 493 | int rctx; |
444 | int size; | 494 | int size; |
@@ -461,14 +511,16 @@ static void perf_syscall_enter(struct pt_regs *regs, long id) | |||
461 | return; | 511 | return; |
462 | 512 | ||
463 | rec = (struct syscall_trace_enter *)perf_trace_buf_prepare(size, | 513 | rec = (struct syscall_trace_enter *)perf_trace_buf_prepare(size, |
464 | sys_data->enter_event->id, &rctx, &flags); | 514 | sys_data->enter_event->event.type, regs, &rctx); |
465 | if (!rec) | 515 | if (!rec) |
466 | return; | 516 | return; |
467 | 517 | ||
468 | rec->nr = syscall_nr; | 518 | rec->nr = syscall_nr; |
469 | syscall_get_arguments(current, regs, 0, sys_data->nb_args, | 519 | syscall_get_arguments(current, regs, 0, sys_data->nb_args, |
470 | (unsigned long *)&rec->args); | 520 | (unsigned long *)&rec->args); |
471 | perf_trace_buf_submit(rec, size, rctx, 0, 1, flags, regs); | 521 | |
522 | head = per_cpu_ptr(sys_data->enter_event->perf_events, smp_processor_id()); | ||
523 | perf_trace_buf_submit(rec, size, rctx, 0, 1, regs, head); | ||
472 | } | 524 | } |
473 | 525 | ||
474 | int perf_sysenter_enable(struct ftrace_event_call *call) | 526 | int perf_sysenter_enable(struct ftrace_event_call *call) |
@@ -480,7 +532,7 @@ int perf_sysenter_enable(struct ftrace_event_call *call) | |||
480 | 532 | ||
481 | mutex_lock(&syscall_trace_lock); | 533 | mutex_lock(&syscall_trace_lock); |
482 | if (!sys_perf_refcount_enter) | 534 | if (!sys_perf_refcount_enter) |
483 | ret = register_trace_sys_enter(perf_syscall_enter); | 535 | ret = register_trace_sys_enter(perf_syscall_enter, NULL); |
484 | if (ret) { | 536 | if (ret) { |
485 | pr_info("event trace: Could not activate" | 537 | pr_info("event trace: Could not activate" |
486 | "syscall entry trace point"); | 538 | "syscall entry trace point"); |
@@ -502,15 +554,15 @@ void perf_sysenter_disable(struct ftrace_event_call *call) | |||
502 | sys_perf_refcount_enter--; | 554 | sys_perf_refcount_enter--; |
503 | clear_bit(num, enabled_perf_enter_syscalls); | 555 | clear_bit(num, enabled_perf_enter_syscalls); |
504 | if (!sys_perf_refcount_enter) | 556 | if (!sys_perf_refcount_enter) |
505 | unregister_trace_sys_enter(perf_syscall_enter); | 557 | unregister_trace_sys_enter(perf_syscall_enter, NULL); |
506 | mutex_unlock(&syscall_trace_lock); | 558 | mutex_unlock(&syscall_trace_lock); |
507 | } | 559 | } |
508 | 560 | ||
509 | static void perf_syscall_exit(struct pt_regs *regs, long ret) | 561 | static void perf_syscall_exit(void *ignore, struct pt_regs *regs, long ret) |
510 | { | 562 | { |
511 | struct syscall_metadata *sys_data; | 563 | struct syscall_metadata *sys_data; |
512 | struct syscall_trace_exit *rec; | 564 | struct syscall_trace_exit *rec; |
513 | unsigned long flags; | 565 | struct hlist_head *head; |
514 | int syscall_nr; | 566 | int syscall_nr; |
515 | int rctx; | 567 | int rctx; |
516 | int size; | 568 | int size; |
@@ -536,14 +588,15 @@ static void perf_syscall_exit(struct pt_regs *regs, long ret) | |||
536 | return; | 588 | return; |
537 | 589 | ||
538 | rec = (struct syscall_trace_exit *)perf_trace_buf_prepare(size, | 590 | rec = (struct syscall_trace_exit *)perf_trace_buf_prepare(size, |
539 | sys_data->exit_event->id, &rctx, &flags); | 591 | sys_data->exit_event->event.type, regs, &rctx); |
540 | if (!rec) | 592 | if (!rec) |
541 | return; | 593 | return; |
542 | 594 | ||
543 | rec->nr = syscall_nr; | 595 | rec->nr = syscall_nr; |
544 | rec->ret = syscall_get_return_value(current, regs); | 596 | rec->ret = syscall_get_return_value(current, regs); |
545 | 597 | ||
546 | perf_trace_buf_submit(rec, size, rctx, 0, 1, flags, regs); | 598 | head = per_cpu_ptr(sys_data->exit_event->perf_events, smp_processor_id()); |
599 | perf_trace_buf_submit(rec, size, rctx, 0, 1, regs, head); | ||
547 | } | 600 | } |
548 | 601 | ||
549 | int perf_sysexit_enable(struct ftrace_event_call *call) | 602 | int perf_sysexit_enable(struct ftrace_event_call *call) |
@@ -555,7 +608,7 @@ int perf_sysexit_enable(struct ftrace_event_call *call) | |||
555 | 608 | ||
556 | mutex_lock(&syscall_trace_lock); | 609 | mutex_lock(&syscall_trace_lock); |
557 | if (!sys_perf_refcount_exit) | 610 | if (!sys_perf_refcount_exit) |
558 | ret = register_trace_sys_exit(perf_syscall_exit); | 611 | ret = register_trace_sys_exit(perf_syscall_exit, NULL); |
559 | if (ret) { | 612 | if (ret) { |
560 | pr_info("event trace: Could not activate" | 613 | pr_info("event trace: Could not activate" |
561 | "syscall exit trace point"); | 614 | "syscall exit trace point"); |
@@ -577,9 +630,50 @@ void perf_sysexit_disable(struct ftrace_event_call *call) | |||
577 | sys_perf_refcount_exit--; | 630 | sys_perf_refcount_exit--; |
578 | clear_bit(num, enabled_perf_exit_syscalls); | 631 | clear_bit(num, enabled_perf_exit_syscalls); |
579 | if (!sys_perf_refcount_exit) | 632 | if (!sys_perf_refcount_exit) |
580 | unregister_trace_sys_exit(perf_syscall_exit); | 633 | unregister_trace_sys_exit(perf_syscall_exit, NULL); |
581 | mutex_unlock(&syscall_trace_lock); | 634 | mutex_unlock(&syscall_trace_lock); |
582 | } | 635 | } |
583 | 636 | ||
584 | #endif /* CONFIG_PERF_EVENTS */ | 637 | #endif /* CONFIG_PERF_EVENTS */ |
585 | 638 | ||
639 | static int syscall_enter_register(struct ftrace_event_call *event, | ||
640 | enum trace_reg type) | ||
641 | { | ||
642 | switch (type) { | ||
643 | case TRACE_REG_REGISTER: | ||
644 | return reg_event_syscall_enter(event); | ||
645 | case TRACE_REG_UNREGISTER: | ||
646 | unreg_event_syscall_enter(event); | ||
647 | return 0; | ||
648 | |||
649 | #ifdef CONFIG_PERF_EVENTS | ||
650 | case TRACE_REG_PERF_REGISTER: | ||
651 | return perf_sysenter_enable(event); | ||
652 | case TRACE_REG_PERF_UNREGISTER: | ||
653 | perf_sysenter_disable(event); | ||
654 | return 0; | ||
655 | #endif | ||
656 | } | ||
657 | return 0; | ||
658 | } | ||
659 | |||
660 | static int syscall_exit_register(struct ftrace_event_call *event, | ||
661 | enum trace_reg type) | ||
662 | { | ||
663 | switch (type) { | ||
664 | case TRACE_REG_REGISTER: | ||
665 | return reg_event_syscall_exit(event); | ||
666 | case TRACE_REG_UNREGISTER: | ||
667 | unreg_event_syscall_exit(event); | ||
668 | return 0; | ||
669 | |||
670 | #ifdef CONFIG_PERF_EVENTS | ||
671 | case TRACE_REG_PERF_REGISTER: | ||
672 | return perf_sysexit_enable(event); | ||
673 | case TRACE_REG_PERF_UNREGISTER: | ||
674 | perf_sysexit_disable(event); | ||
675 | return 0; | ||
676 | #endif | ||
677 | } | ||
678 | return 0; | ||
679 | } | ||