diff options
Diffstat (limited to 'tools/perf/util/session.c')
-rw-r--r-- | tools/perf/util/session.c | 579 |
1 files changed, 391 insertions, 188 deletions
diff --git a/tools/perf/util/session.c b/tools/perf/util/session.c index fa9d652c2dc3..6fb4694d05fa 100644 --- a/tools/perf/util/session.c +++ b/tools/perf/util/session.c | |||
@@ -65,9 +65,49 @@ out_close: | |||
65 | return -1; | 65 | return -1; |
66 | } | 66 | } |
67 | 67 | ||
68 | static void perf_session__id_header_size(struct perf_session *session) | ||
69 | { | ||
70 | struct sample_data *data; | ||
71 | u64 sample_type = session->sample_type; | ||
72 | u16 size = 0; | ||
73 | |||
74 | if (!session->sample_id_all) | ||
75 | goto out; | ||
76 | |||
77 | if (sample_type & PERF_SAMPLE_TID) | ||
78 | size += sizeof(data->tid) * 2; | ||
79 | |||
80 | if (sample_type & PERF_SAMPLE_TIME) | ||
81 | size += sizeof(data->time); | ||
82 | |||
83 | if (sample_type & PERF_SAMPLE_ID) | ||
84 | size += sizeof(data->id); | ||
85 | |||
86 | if (sample_type & PERF_SAMPLE_STREAM_ID) | ||
87 | size += sizeof(data->stream_id); | ||
88 | |||
89 | if (sample_type & PERF_SAMPLE_CPU) | ||
90 | size += sizeof(data->cpu) * 2; | ||
91 | out: | ||
92 | session->id_hdr_size = size; | ||
93 | } | ||
94 | |||
95 | void perf_session__set_sample_id_all(struct perf_session *session, bool value) | ||
96 | { | ||
97 | session->sample_id_all = value; | ||
98 | perf_session__id_header_size(session); | ||
99 | } | ||
100 | |||
101 | void perf_session__set_sample_type(struct perf_session *session, u64 type) | ||
102 | { | ||
103 | session->sample_type = type; | ||
104 | } | ||
105 | |||
68 | void perf_session__update_sample_type(struct perf_session *self) | 106 | void perf_session__update_sample_type(struct perf_session *self) |
69 | { | 107 | { |
70 | self->sample_type = perf_header__sample_type(&self->header); | 108 | self->sample_type = perf_header__sample_type(&self->header); |
109 | self->sample_id_all = perf_header__sample_id_all(&self->header); | ||
110 | perf_session__id_header_size(self); | ||
71 | } | 111 | } |
72 | 112 | ||
73 | int perf_session__create_kernel_maps(struct perf_session *self) | 113 | int perf_session__create_kernel_maps(struct perf_session *self) |
@@ -85,7 +125,9 @@ static void perf_session__destroy_kernel_maps(struct perf_session *self) | |||
85 | machines__destroy_guest_kernel_maps(&self->machines); | 125 | machines__destroy_guest_kernel_maps(&self->machines); |
86 | } | 126 | } |
87 | 127 | ||
88 | struct perf_session *perf_session__new(const char *filename, int mode, bool force, bool repipe) | 128 | struct perf_session *perf_session__new(const char *filename, int mode, |
129 | bool force, bool repipe, | ||
130 | struct perf_event_ops *ops) | ||
89 | { | 131 | { |
90 | size_t len = filename ? strlen(filename) + 1 : 0; | 132 | size_t len = filename ? strlen(filename) + 1 : 0; |
91 | struct perf_session *self = zalloc(sizeof(*self) + len); | 133 | struct perf_session *self = zalloc(sizeof(*self) + len); |
@@ -101,10 +143,20 @@ struct perf_session *perf_session__new(const char *filename, int mode, bool forc | |||
101 | INIT_LIST_HEAD(&self->dead_threads); | 143 | INIT_LIST_HEAD(&self->dead_threads); |
102 | self->hists_tree = RB_ROOT; | 144 | self->hists_tree = RB_ROOT; |
103 | self->last_match = NULL; | 145 | self->last_match = NULL; |
104 | self->mmap_window = 32; | 146 | /* |
147 | * On 64bit we can mmap the data file in one go. No need for tiny mmap | ||
148 | * slices. On 32bit we use 32MB. | ||
149 | */ | ||
150 | #if BITS_PER_LONG == 64 | ||
151 | self->mmap_window = ULLONG_MAX; | ||
152 | #else | ||
153 | self->mmap_window = 32 * 1024 * 1024ULL; | ||
154 | #endif | ||
105 | self->machines = RB_ROOT; | 155 | self->machines = RB_ROOT; |
106 | self->repipe = repipe; | 156 | self->repipe = repipe; |
107 | INIT_LIST_HEAD(&self->ordered_samples.samples_head); | 157 | INIT_LIST_HEAD(&self->ordered_samples.samples); |
158 | INIT_LIST_HEAD(&self->ordered_samples.sample_cache); | ||
159 | INIT_LIST_HEAD(&self->ordered_samples.to_free); | ||
108 | machine__init(&self->host_machine, "", HOST_KERNEL_ID); | 160 | machine__init(&self->host_machine, "", HOST_KERNEL_ID); |
109 | 161 | ||
110 | if (mode == O_RDONLY) { | 162 | if (mode == O_RDONLY) { |
@@ -120,6 +172,13 @@ struct perf_session *perf_session__new(const char *filename, int mode, bool forc | |||
120 | } | 172 | } |
121 | 173 | ||
122 | perf_session__update_sample_type(self); | 174 | perf_session__update_sample_type(self); |
175 | |||
176 | if (ops && ops->ordering_requires_timestamps && | ||
177 | ops->ordered_samples && !self->sample_id_all) { | ||
178 | dump_printf("WARNING: No sample_id_all support, falling back to unordered processing\n"); | ||
179 | ops->ordered_samples = false; | ||
180 | } | ||
181 | |||
123 | out: | 182 | out: |
124 | return self; | 183 | return self; |
125 | out_free: | 184 | out_free: |
@@ -230,7 +289,15 @@ struct map_symbol *perf_session__resolve_callchain(struct perf_session *self, | |||
230 | return syms; | 289 | return syms; |
231 | } | 290 | } |
232 | 291 | ||
292 | static int process_event_synth_stub(event_t *event __used, | ||
293 | struct perf_session *session __used) | ||
294 | { | ||
295 | dump_printf(": unhandled!\n"); | ||
296 | return 0; | ||
297 | } | ||
298 | |||
233 | static int process_event_stub(event_t *event __used, | 299 | static int process_event_stub(event_t *event __used, |
300 | struct sample_data *sample __used, | ||
234 | struct perf_session *session __used) | 301 | struct perf_session *session __used) |
235 | { | 302 | { |
236 | dump_printf(": unhandled!\n"); | 303 | dump_printf(": unhandled!\n"); |
@@ -262,7 +329,7 @@ static void perf_event_ops__fill_defaults(struct perf_event_ops *handler) | |||
262 | if (handler->exit == NULL) | 329 | if (handler->exit == NULL) |
263 | handler->exit = process_event_stub; | 330 | handler->exit = process_event_stub; |
264 | if (handler->lost == NULL) | 331 | if (handler->lost == NULL) |
265 | handler->lost = process_event_stub; | 332 | handler->lost = event__process_lost; |
266 | if (handler->read == NULL) | 333 | if (handler->read == NULL) |
267 | handler->read = process_event_stub; | 334 | handler->read = process_event_stub; |
268 | if (handler->throttle == NULL) | 335 | if (handler->throttle == NULL) |
@@ -270,13 +337,13 @@ static void perf_event_ops__fill_defaults(struct perf_event_ops *handler) | |||
270 | if (handler->unthrottle == NULL) | 337 | if (handler->unthrottle == NULL) |
271 | handler->unthrottle = process_event_stub; | 338 | handler->unthrottle = process_event_stub; |
272 | if (handler->attr == NULL) | 339 | if (handler->attr == NULL) |
273 | handler->attr = process_event_stub; | 340 | handler->attr = process_event_synth_stub; |
274 | if (handler->event_type == NULL) | 341 | if (handler->event_type == NULL) |
275 | handler->event_type = process_event_stub; | 342 | handler->event_type = process_event_synth_stub; |
276 | if (handler->tracing_data == NULL) | 343 | if (handler->tracing_data == NULL) |
277 | handler->tracing_data = process_event_stub; | 344 | handler->tracing_data = process_event_synth_stub; |
278 | if (handler->build_id == NULL) | 345 | if (handler->build_id == NULL) |
279 | handler->build_id = process_event_stub; | 346 | handler->build_id = process_event_synth_stub; |
280 | if (handler->finished_round == NULL) { | 347 | if (handler->finished_round == NULL) { |
281 | if (handler->ordered_samples) | 348 | if (handler->ordered_samples) |
282 | handler->finished_round = process_finished_round; | 349 | handler->finished_round = process_finished_round; |
@@ -386,33 +453,61 @@ static event__swap_op event__swap_ops[] = { | |||
386 | 453 | ||
387 | struct sample_queue { | 454 | struct sample_queue { |
388 | u64 timestamp; | 455 | u64 timestamp; |
389 | struct sample_event *event; | 456 | u64 file_offset; |
457 | event_t *event; | ||
390 | struct list_head list; | 458 | struct list_head list; |
391 | }; | 459 | }; |
392 | 460 | ||
461 | static void perf_session_free_sample_buffers(struct perf_session *session) | ||
462 | { | ||
463 | struct ordered_samples *os = &session->ordered_samples; | ||
464 | |||
465 | while (!list_empty(&os->to_free)) { | ||
466 | struct sample_queue *sq; | ||
467 | |||
468 | sq = list_entry(os->to_free.next, struct sample_queue, list); | ||
469 | list_del(&sq->list); | ||
470 | free(sq); | ||
471 | } | ||
472 | } | ||
473 | |||
474 | static int perf_session_deliver_event(struct perf_session *session, | ||
475 | event_t *event, | ||
476 | struct sample_data *sample, | ||
477 | struct perf_event_ops *ops, | ||
478 | u64 file_offset); | ||
479 | |||
393 | static void flush_sample_queue(struct perf_session *s, | 480 | static void flush_sample_queue(struct perf_session *s, |
394 | struct perf_event_ops *ops) | 481 | struct perf_event_ops *ops) |
395 | { | 482 | { |
396 | struct list_head *head = &s->ordered_samples.samples_head; | 483 | struct ordered_samples *os = &s->ordered_samples; |
397 | u64 limit = s->ordered_samples.next_flush; | 484 | struct list_head *head = &os->samples; |
398 | struct sample_queue *tmp, *iter; | 485 | struct sample_queue *tmp, *iter; |
486 | struct sample_data sample; | ||
487 | u64 limit = os->next_flush; | ||
488 | u64 last_ts = os->last_sample ? os->last_sample->timestamp : 0ULL; | ||
399 | 489 | ||
400 | if (!ops->ordered_samples || !limit) | 490 | if (!ops->ordered_samples || !limit) |
401 | return; | 491 | return; |
402 | 492 | ||
403 | list_for_each_entry_safe(iter, tmp, head, list) { | 493 | list_for_each_entry_safe(iter, tmp, head, list) { |
404 | if (iter->timestamp > limit) | 494 | if (iter->timestamp > limit) |
405 | return; | 495 | break; |
406 | 496 | ||
407 | if (iter == s->ordered_samples.last_inserted) | 497 | event__parse_sample(iter->event, s, &sample); |
408 | s->ordered_samples.last_inserted = NULL; | 498 | perf_session_deliver_event(s, iter->event, &sample, ops, |
499 | iter->file_offset); | ||
409 | 500 | ||
410 | ops->sample((event_t *)iter->event, s); | 501 | os->last_flush = iter->timestamp; |
411 | |||
412 | s->ordered_samples.last_flush = iter->timestamp; | ||
413 | list_del(&iter->list); | 502 | list_del(&iter->list); |
414 | free(iter->event); | 503 | list_add(&iter->list, &os->sample_cache); |
415 | free(iter); | 504 | } |
505 | |||
506 | if (list_empty(head)) { | ||
507 | os->last_sample = NULL; | ||
508 | } else if (last_ts <= limit) { | ||
509 | os->last_sample = | ||
510 | list_entry(head->prev, struct sample_queue, list); | ||
416 | } | 511 | } |
417 | } | 512 | } |
418 | 513 | ||
@@ -465,178 +560,265 @@ static int process_finished_round(event_t *event __used, | |||
465 | return 0; | 560 | return 0; |
466 | } | 561 | } |
467 | 562 | ||
468 | static void __queue_sample_end(struct sample_queue *new, struct list_head *head) | ||
469 | { | ||
470 | struct sample_queue *iter; | ||
471 | |||
472 | list_for_each_entry_reverse(iter, head, list) { | ||
473 | if (iter->timestamp < new->timestamp) { | ||
474 | list_add(&new->list, &iter->list); | ||
475 | return; | ||
476 | } | ||
477 | } | ||
478 | |||
479 | list_add(&new->list, head); | ||
480 | } | ||
481 | |||
482 | static void __queue_sample_before(struct sample_queue *new, | ||
483 | struct sample_queue *iter, | ||
484 | struct list_head *head) | ||
485 | { | ||
486 | list_for_each_entry_continue_reverse(iter, head, list) { | ||
487 | if (iter->timestamp < new->timestamp) { | ||
488 | list_add(&new->list, &iter->list); | ||
489 | return; | ||
490 | } | ||
491 | } | ||
492 | |||
493 | list_add(&new->list, head); | ||
494 | } | ||
495 | |||
496 | static void __queue_sample_after(struct sample_queue *new, | ||
497 | struct sample_queue *iter, | ||
498 | struct list_head *head) | ||
499 | { | ||
500 | list_for_each_entry_continue(iter, head, list) { | ||
501 | if (iter->timestamp > new->timestamp) { | ||
502 | list_add_tail(&new->list, &iter->list); | ||
503 | return; | ||
504 | } | ||
505 | } | ||
506 | list_add_tail(&new->list, head); | ||
507 | } | ||
508 | |||
509 | /* The queue is ordered by time */ | 563 | /* The queue is ordered by time */ |
510 | static void __queue_sample_event(struct sample_queue *new, | 564 | static void __queue_event(struct sample_queue *new, struct perf_session *s) |
511 | struct perf_session *s) | ||
512 | { | 565 | { |
513 | struct sample_queue *last_inserted = s->ordered_samples.last_inserted; | 566 | struct ordered_samples *os = &s->ordered_samples; |
514 | struct list_head *head = &s->ordered_samples.samples_head; | 567 | struct sample_queue *sample = os->last_sample; |
568 | u64 timestamp = new->timestamp; | ||
569 | struct list_head *p; | ||
515 | 570 | ||
571 | os->last_sample = new; | ||
516 | 572 | ||
517 | if (!last_inserted) { | 573 | if (!sample) { |
518 | __queue_sample_end(new, head); | 574 | list_add(&new->list, &os->samples); |
575 | os->max_timestamp = timestamp; | ||
519 | return; | 576 | return; |
520 | } | 577 | } |
521 | 578 | ||
522 | /* | 579 | /* |
523 | * Most of the time the current event has a timestamp | 580 | * last_sample might point to some random place in the list as it's |
524 | * very close to the last event inserted, unless we just switched | 581 | * the last queued event. We expect that the new event is close to |
525 | * to another event buffer. Having a sorting based on a list and | 582 | * this. |
526 | * on the last inserted event that is close to the current one is | ||
527 | * probably more efficient than an rbtree based sorting. | ||
528 | */ | 583 | */ |
529 | if (last_inserted->timestamp >= new->timestamp) | 584 | if (sample->timestamp <= timestamp) { |
530 | __queue_sample_before(new, last_inserted, head); | 585 | while (sample->timestamp <= timestamp) { |
531 | else | 586 | p = sample->list.next; |
532 | __queue_sample_after(new, last_inserted, head); | 587 | if (p == &os->samples) { |
588 | list_add_tail(&new->list, &os->samples); | ||
589 | os->max_timestamp = timestamp; | ||
590 | return; | ||
591 | } | ||
592 | sample = list_entry(p, struct sample_queue, list); | ||
593 | } | ||
594 | list_add_tail(&new->list, &sample->list); | ||
595 | } else { | ||
596 | while (sample->timestamp > timestamp) { | ||
597 | p = sample->list.prev; | ||
598 | if (p == &os->samples) { | ||
599 | list_add(&new->list, &os->samples); | ||
600 | return; | ||
601 | } | ||
602 | sample = list_entry(p, struct sample_queue, list); | ||
603 | } | ||
604 | list_add(&new->list, &sample->list); | ||
605 | } | ||
533 | } | 606 | } |
534 | 607 | ||
535 | static int queue_sample_event(event_t *event, struct sample_data *data, | 608 | #define MAX_SAMPLE_BUFFER (64 * 1024 / sizeof(struct sample_queue)) |
536 | struct perf_session *s) | 609 | |
610 | static int perf_session_queue_event(struct perf_session *s, event_t *event, | ||
611 | struct sample_data *data, u64 file_offset) | ||
537 | { | 612 | { |
613 | struct ordered_samples *os = &s->ordered_samples; | ||
614 | struct list_head *sc = &os->sample_cache; | ||
538 | u64 timestamp = data->time; | 615 | u64 timestamp = data->time; |
539 | struct sample_queue *new; | 616 | struct sample_queue *new; |
540 | 617 | ||
618 | if (!timestamp || timestamp == ~0ULL) | ||
619 | return -ETIME; | ||
541 | 620 | ||
542 | if (timestamp < s->ordered_samples.last_flush) { | 621 | if (timestamp < s->ordered_samples.last_flush) { |
543 | printf("Warning: Timestamp below last timeslice flush\n"); | 622 | printf("Warning: Timestamp below last timeslice flush\n"); |
544 | return -EINVAL; | 623 | return -EINVAL; |
545 | } | 624 | } |
546 | 625 | ||
547 | new = malloc(sizeof(*new)); | 626 | if (!list_empty(sc)) { |
548 | if (!new) | 627 | new = list_entry(sc->next, struct sample_queue, list); |
549 | return -ENOMEM; | 628 | list_del(&new->list); |
629 | } else if (os->sample_buffer) { | ||
630 | new = os->sample_buffer + os->sample_buffer_idx; | ||
631 | if (++os->sample_buffer_idx == MAX_SAMPLE_BUFFER) | ||
632 | os->sample_buffer = NULL; | ||
633 | } else { | ||
634 | os->sample_buffer = malloc(MAX_SAMPLE_BUFFER * sizeof(*new)); | ||
635 | if (!os->sample_buffer) | ||
636 | return -ENOMEM; | ||
637 | list_add(&os->sample_buffer->list, &os->to_free); | ||
638 | os->sample_buffer_idx = 2; | ||
639 | new = os->sample_buffer + 1; | ||
640 | } | ||
550 | 641 | ||
551 | new->timestamp = timestamp; | 642 | new->timestamp = timestamp; |
643 | new->file_offset = file_offset; | ||
644 | new->event = event; | ||
552 | 645 | ||
553 | new->event = malloc(event->header.size); | 646 | __queue_event(new, s); |
554 | if (!new->event) { | ||
555 | free(new); | ||
556 | return -ENOMEM; | ||
557 | } | ||
558 | 647 | ||
559 | memcpy(new->event, event, event->header.size); | 648 | return 0; |
649 | } | ||
560 | 650 | ||
561 | __queue_sample_event(new, s); | 651 | static void callchain__printf(struct sample_data *sample) |
562 | s->ordered_samples.last_inserted = new; | 652 | { |
653 | unsigned int i; | ||
563 | 654 | ||
564 | if (new->timestamp > s->ordered_samples.max_timestamp) | 655 | printf("... chain: nr:%Lu\n", sample->callchain->nr); |
565 | s->ordered_samples.max_timestamp = new->timestamp; | ||
566 | 656 | ||
567 | return 0; | 657 | for (i = 0; i < sample->callchain->nr; i++) |
658 | printf("..... %2d: %016Lx\n", i, sample->callchain->ips[i]); | ||
568 | } | 659 | } |
569 | 660 | ||
570 | static int perf_session__process_sample(event_t *event, struct perf_session *s, | 661 | static void perf_session__print_tstamp(struct perf_session *session, |
571 | struct perf_event_ops *ops) | 662 | event_t *event, |
663 | struct sample_data *sample) | ||
572 | { | 664 | { |
573 | struct sample_data data; | 665 | if (event->header.type != PERF_RECORD_SAMPLE && |
666 | !session->sample_id_all) { | ||
667 | fputs("-1 -1 ", stdout); | ||
668 | return; | ||
669 | } | ||
574 | 670 | ||
575 | if (!ops->ordered_samples) | 671 | if ((session->sample_type & PERF_SAMPLE_CPU)) |
576 | return ops->sample(event, s); | 672 | printf("%u ", sample->cpu); |
577 | 673 | ||
578 | bzero(&data, sizeof(struct sample_data)); | 674 | if (session->sample_type & PERF_SAMPLE_TIME) |
579 | event__parse_sample(event, s->sample_type, &data); | 675 | printf("%Lu ", sample->time); |
676 | } | ||
580 | 677 | ||
581 | queue_sample_event(event, &data, s); | 678 | static void dump_event(struct perf_session *session, event_t *event, |
679 | u64 file_offset, struct sample_data *sample) | ||
680 | { | ||
681 | if (!dump_trace) | ||
682 | return; | ||
582 | 683 | ||
583 | return 0; | 684 | printf("\n%#Lx [%#x]: event: %d\n", file_offset, event->header.size, |
685 | event->header.type); | ||
686 | |||
687 | trace_event(event); | ||
688 | |||
689 | if (sample) | ||
690 | perf_session__print_tstamp(session, event, sample); | ||
691 | |||
692 | printf("%#Lx [%#x]: PERF_RECORD_%s", file_offset, event->header.size, | ||
693 | event__get_event_name(event->header.type)); | ||
584 | } | 694 | } |
585 | 695 | ||
586 | static int perf_session__process_event(struct perf_session *self, | 696 | static void dump_sample(struct perf_session *session, event_t *event, |
587 | event_t *event, | 697 | struct sample_data *sample) |
588 | struct perf_event_ops *ops, | ||
589 | u64 offset, u64 head) | ||
590 | { | 698 | { |
591 | trace_event(event); | 699 | if (!dump_trace) |
700 | return; | ||
592 | 701 | ||
593 | if (event->header.type < PERF_RECORD_HEADER_MAX) { | 702 | printf("(IP, %d): %d/%d: %#Lx period: %Ld\n", event->header.misc, |
594 | dump_printf("%#Lx [%#x]: PERF_RECORD_%s", | 703 | sample->pid, sample->tid, sample->ip, sample->period); |
595 | offset + head, event->header.size, | ||
596 | event__name[event->header.type]); | ||
597 | hists__inc_nr_events(&self->hists, event->header.type); | ||
598 | } | ||
599 | 704 | ||
600 | if (self->header.needs_swap && event__swap_ops[event->header.type]) | 705 | if (session->sample_type & PERF_SAMPLE_CALLCHAIN) |
601 | event__swap_ops[event->header.type](event); | 706 | callchain__printf(sample); |
707 | } | ||
708 | |||
709 | static int perf_session_deliver_event(struct perf_session *session, | ||
710 | event_t *event, | ||
711 | struct sample_data *sample, | ||
712 | struct perf_event_ops *ops, | ||
713 | u64 file_offset) | ||
714 | { | ||
715 | dump_event(session, event, file_offset, sample); | ||
602 | 716 | ||
603 | switch (event->header.type) { | 717 | switch (event->header.type) { |
604 | case PERF_RECORD_SAMPLE: | 718 | case PERF_RECORD_SAMPLE: |
605 | return perf_session__process_sample(event, self, ops); | 719 | dump_sample(session, event, sample); |
720 | return ops->sample(event, sample, session); | ||
606 | case PERF_RECORD_MMAP: | 721 | case PERF_RECORD_MMAP: |
607 | return ops->mmap(event, self); | 722 | return ops->mmap(event, sample, session); |
608 | case PERF_RECORD_COMM: | 723 | case PERF_RECORD_COMM: |
609 | return ops->comm(event, self); | 724 | return ops->comm(event, sample, session); |
610 | case PERF_RECORD_FORK: | 725 | case PERF_RECORD_FORK: |
611 | return ops->fork(event, self); | 726 | return ops->fork(event, sample, session); |
612 | case PERF_RECORD_EXIT: | 727 | case PERF_RECORD_EXIT: |
613 | return ops->exit(event, self); | 728 | return ops->exit(event, sample, session); |
614 | case PERF_RECORD_LOST: | 729 | case PERF_RECORD_LOST: |
615 | return ops->lost(event, self); | 730 | return ops->lost(event, sample, session); |
616 | case PERF_RECORD_READ: | 731 | case PERF_RECORD_READ: |
617 | return ops->read(event, self); | 732 | return ops->read(event, sample, session); |
618 | case PERF_RECORD_THROTTLE: | 733 | case PERF_RECORD_THROTTLE: |
619 | return ops->throttle(event, self); | 734 | return ops->throttle(event, sample, session); |
620 | case PERF_RECORD_UNTHROTTLE: | 735 | case PERF_RECORD_UNTHROTTLE: |
621 | return ops->unthrottle(event, self); | 736 | return ops->unthrottle(event, sample, session); |
737 | default: | ||
738 | ++session->hists.stats.nr_unknown_events; | ||
739 | return -1; | ||
740 | } | ||
741 | } | ||
742 | |||
743 | static int perf_session__preprocess_sample(struct perf_session *session, | ||
744 | event_t *event, struct sample_data *sample) | ||
745 | { | ||
746 | if (event->header.type != PERF_RECORD_SAMPLE || | ||
747 | !(session->sample_type & PERF_SAMPLE_CALLCHAIN)) | ||
748 | return 0; | ||
749 | |||
750 | if (!ip_callchain__valid(sample->callchain, event)) { | ||
751 | pr_debug("call-chain problem with event, skipping it.\n"); | ||
752 | ++session->hists.stats.nr_invalid_chains; | ||
753 | session->hists.stats.total_invalid_chains += sample->period; | ||
754 | return -EINVAL; | ||
755 | } | ||
756 | return 0; | ||
757 | } | ||
758 | |||
759 | static int perf_session__process_user_event(struct perf_session *session, event_t *event, | ||
760 | struct perf_event_ops *ops, u64 file_offset) | ||
761 | { | ||
762 | dump_event(session, event, file_offset, NULL); | ||
763 | |||
764 | /* These events are processed right away */ | ||
765 | switch (event->header.type) { | ||
622 | case PERF_RECORD_HEADER_ATTR: | 766 | case PERF_RECORD_HEADER_ATTR: |
623 | return ops->attr(event, self); | 767 | return ops->attr(event, session); |
624 | case PERF_RECORD_HEADER_EVENT_TYPE: | 768 | case PERF_RECORD_HEADER_EVENT_TYPE: |
625 | return ops->event_type(event, self); | 769 | return ops->event_type(event, session); |
626 | case PERF_RECORD_HEADER_TRACING_DATA: | 770 | case PERF_RECORD_HEADER_TRACING_DATA: |
627 | /* setup for reading amidst mmap */ | 771 | /* setup for reading amidst mmap */ |
628 | lseek(self->fd, offset + head, SEEK_SET); | 772 | lseek(session->fd, file_offset, SEEK_SET); |
629 | return ops->tracing_data(event, self); | 773 | return ops->tracing_data(event, session); |
630 | case PERF_RECORD_HEADER_BUILD_ID: | 774 | case PERF_RECORD_HEADER_BUILD_ID: |
631 | return ops->build_id(event, self); | 775 | return ops->build_id(event, session); |
632 | case PERF_RECORD_FINISHED_ROUND: | 776 | case PERF_RECORD_FINISHED_ROUND: |
633 | return ops->finished_round(event, self, ops); | 777 | return ops->finished_round(event, session, ops); |
634 | default: | 778 | default: |
635 | ++self->hists.stats.nr_unknown_events; | 779 | return -EINVAL; |
636 | return -1; | ||
637 | } | 780 | } |
638 | } | 781 | } |
639 | 782 | ||
783 | static int perf_session__process_event(struct perf_session *session, | ||
784 | event_t *event, | ||
785 | struct perf_event_ops *ops, | ||
786 | u64 file_offset) | ||
787 | { | ||
788 | struct sample_data sample; | ||
789 | int ret; | ||
790 | |||
791 | if (session->header.needs_swap && event__swap_ops[event->header.type]) | ||
792 | event__swap_ops[event->header.type](event); | ||
793 | |||
794 | if (event->header.type >= PERF_RECORD_HEADER_MAX) | ||
795 | return -EINVAL; | ||
796 | |||
797 | hists__inc_nr_events(&session->hists, event->header.type); | ||
798 | |||
799 | if (event->header.type >= PERF_RECORD_USER_TYPE_START) | ||
800 | return perf_session__process_user_event(session, event, ops, file_offset); | ||
801 | |||
802 | /* | ||
803 | * For all kernel events we get the sample data | ||
804 | */ | ||
805 | event__parse_sample(event, session, &sample); | ||
806 | |||
807 | /* Preprocess sample records - precheck callchains */ | ||
808 | if (perf_session__preprocess_sample(session, event, &sample)) | ||
809 | return 0; | ||
810 | |||
811 | if (ops->ordered_samples) { | ||
812 | ret = perf_session_queue_event(session, event, &sample, | ||
813 | file_offset); | ||
814 | if (ret != -ETIME) | ||
815 | return ret; | ||
816 | } | ||
817 | |||
818 | return perf_session_deliver_event(session, event, &sample, ops, | ||
819 | file_offset); | ||
820 | } | ||
821 | |||
640 | void perf_event_header__bswap(struct perf_event_header *self) | 822 | void perf_event_header__bswap(struct perf_event_header *self) |
641 | { | 823 | { |
642 | self->type = bswap_32(self->type); | 824 | self->type = bswap_32(self->type); |
@@ -656,21 +838,33 @@ static struct thread *perf_session__register_idle_thread(struct perf_session *se | |||
656 | return thread; | 838 | return thread; |
657 | } | 839 | } |
658 | 840 | ||
659 | int do_read(int fd, void *buf, size_t size) | 841 | static void perf_session__warn_about_errors(const struct perf_session *session, |
842 | const struct perf_event_ops *ops) | ||
660 | { | 843 | { |
661 | void *buf_start = buf; | 844 | if (ops->lost == event__process_lost && |
662 | 845 | session->hists.stats.total_lost != 0) { | |
663 | while (size) { | 846 | ui__warning("Processed %Lu events and LOST %Lu!\n\n" |
664 | int ret = read(fd, buf, size); | 847 | "Check IO/CPU overload!\n\n", |
665 | 848 | session->hists.stats.total_period, | |
666 | if (ret <= 0) | 849 | session->hists.stats.total_lost); |
667 | return ret; | 850 | } |
668 | 851 | ||
669 | size -= ret; | 852 | if (session->hists.stats.nr_unknown_events != 0) { |
670 | buf += ret; | 853 | ui__warning("Found %u unknown events!\n\n" |
854 | "Is this an older tool processing a perf.data " | ||
855 | "file generated by a more recent tool?\n\n" | ||
856 | "If that is not the case, consider " | ||
857 | "reporting to linux-kernel@vger.kernel.org.\n\n", | ||
858 | session->hists.stats.nr_unknown_events); | ||
671 | } | 859 | } |
672 | 860 | ||
673 | return buf - buf_start; | 861 | if (session->hists.stats.nr_invalid_chains != 0) { |
862 | ui__warning("Found invalid callchains!\n\n" | ||
863 | "%u out of %u events were discarded for this reason.\n\n" | ||
864 | "Consider reporting to linux-kernel@vger.kernel.org.\n\n", | ||
865 | session->hists.stats.nr_invalid_chains, | ||
866 | session->hists.stats.nr_events[PERF_RECORD_SAMPLE]); | ||
867 | } | ||
674 | } | 868 | } |
675 | 869 | ||
676 | #define session_done() (*(volatile int *)(&session_done)) | 870 | #define session_done() (*(volatile int *)(&session_done)) |
@@ -690,7 +884,7 @@ static int __perf_session__process_pipe_events(struct perf_session *self, | |||
690 | 884 | ||
691 | head = 0; | 885 | head = 0; |
692 | more: | 886 | more: |
693 | err = do_read(self->fd, &event, sizeof(struct perf_event_header)); | 887 | err = readn(self->fd, &event, sizeof(struct perf_event_header)); |
694 | if (err <= 0) { | 888 | if (err <= 0) { |
695 | if (err == 0) | 889 | if (err == 0) |
696 | goto done; | 890 | goto done; |
@@ -710,8 +904,7 @@ more: | |||
710 | p += sizeof(struct perf_event_header); | 904 | p += sizeof(struct perf_event_header); |
711 | 905 | ||
712 | if (size - sizeof(struct perf_event_header)) { | 906 | if (size - sizeof(struct perf_event_header)) { |
713 | err = do_read(self->fd, p, | 907 | err = readn(self->fd, p, size - sizeof(struct perf_event_header)); |
714 | size - sizeof(struct perf_event_header)); | ||
715 | if (err <= 0) { | 908 | if (err <= 0) { |
716 | if (err == 0) { | 909 | if (err == 0) { |
717 | pr_err("unexpected end of event stream\n"); | 910 | pr_err("unexpected end of event stream\n"); |
@@ -724,8 +917,7 @@ more: | |||
724 | } | 917 | } |
725 | 918 | ||
726 | if (size == 0 || | 919 | if (size == 0 || |
727 | (skip = perf_session__process_event(self, &event, ops, | 920 | (skip = perf_session__process_event(self, &event, ops, head)) < 0) { |
728 | 0, head)) < 0) { | ||
729 | dump_printf("%#Lx [%#x]: skipping unknown header type: %d\n", | 921 | dump_printf("%#Lx [%#x]: skipping unknown header type: %d\n", |
730 | head, event.header.size, event.header.type); | 922 | head, event.header.size, event.header.type); |
731 | /* | 923 | /* |
@@ -740,9 +932,6 @@ more: | |||
740 | 932 | ||
741 | head += size; | 933 | head += size; |
742 | 934 | ||
743 | dump_printf("\n%#Lx [%#x]: event: %d\n", | ||
744 | head, event.header.size, event.header.type); | ||
745 | |||
746 | if (skip > 0) | 935 | if (skip > 0) |
747 | head += skip; | 936 | head += skip; |
748 | 937 | ||
@@ -751,82 +940,91 @@ more: | |||
751 | done: | 940 | done: |
752 | err = 0; | 941 | err = 0; |
753 | out_err: | 942 | out_err: |
943 | perf_session__warn_about_errors(self, ops); | ||
944 | perf_session_free_sample_buffers(self); | ||
754 | return err; | 945 | return err; |
755 | } | 946 | } |
756 | 947 | ||
757 | int __perf_session__process_events(struct perf_session *self, | 948 | int __perf_session__process_events(struct perf_session *session, |
758 | u64 data_offset, u64 data_size, | 949 | u64 data_offset, u64 data_size, |
759 | u64 file_size, struct perf_event_ops *ops) | 950 | u64 file_size, struct perf_event_ops *ops) |
760 | { | 951 | { |
761 | int err, mmap_prot, mmap_flags; | 952 | u64 head, page_offset, file_offset, file_pos, progress_next; |
762 | u64 head, shift; | 953 | int err, mmap_prot, mmap_flags, map_idx = 0; |
763 | u64 offset = 0; | 954 | struct ui_progress *progress; |
764 | size_t page_size; | 955 | size_t page_size, mmap_size; |
956 | char *buf, *mmaps[8]; | ||
765 | event_t *event; | 957 | event_t *event; |
766 | uint32_t size; | 958 | uint32_t size; |
767 | char *buf; | ||
768 | struct ui_progress *progress = ui_progress__new("Processing events...", | ||
769 | self->size); | ||
770 | if (progress == NULL) | ||
771 | return -1; | ||
772 | 959 | ||
773 | perf_event_ops__fill_defaults(ops); | 960 | perf_event_ops__fill_defaults(ops); |
774 | 961 | ||
775 | page_size = sysconf(_SC_PAGESIZE); | 962 | page_size = sysconf(_SC_PAGESIZE); |
776 | 963 | ||
777 | head = data_offset; | 964 | page_offset = page_size * (data_offset / page_size); |
778 | shift = page_size * (head / page_size); | 965 | file_offset = page_offset; |
779 | offset += shift; | 966 | head = data_offset - page_offset; |
780 | head -= shift; | 967 | |
968 | if (data_offset + data_size < file_size) | ||
969 | file_size = data_offset + data_size; | ||
970 | |||
971 | progress_next = file_size / 16; | ||
972 | progress = ui_progress__new("Processing events...", file_size); | ||
973 | if (progress == NULL) | ||
974 | return -1; | ||
975 | |||
976 | mmap_size = session->mmap_window; | ||
977 | if (mmap_size > file_size) | ||
978 | mmap_size = file_size; | ||
979 | |||
980 | memset(mmaps, 0, sizeof(mmaps)); | ||
781 | 981 | ||
782 | mmap_prot = PROT_READ; | 982 | mmap_prot = PROT_READ; |
783 | mmap_flags = MAP_SHARED; | 983 | mmap_flags = MAP_SHARED; |
784 | 984 | ||
785 | if (self->header.needs_swap) { | 985 | if (session->header.needs_swap) { |
786 | mmap_prot |= PROT_WRITE; | 986 | mmap_prot |= PROT_WRITE; |
787 | mmap_flags = MAP_PRIVATE; | 987 | mmap_flags = MAP_PRIVATE; |
788 | } | 988 | } |
789 | remap: | 989 | remap: |
790 | buf = mmap(NULL, page_size * self->mmap_window, mmap_prot, | 990 | buf = mmap(NULL, mmap_size, mmap_prot, mmap_flags, session->fd, |
791 | mmap_flags, self->fd, offset); | 991 | file_offset); |
792 | if (buf == MAP_FAILED) { | 992 | if (buf == MAP_FAILED) { |
793 | pr_err("failed to mmap file\n"); | 993 | pr_err("failed to mmap file\n"); |
794 | err = -errno; | 994 | err = -errno; |
795 | goto out_err; | 995 | goto out_err; |
796 | } | 996 | } |
997 | mmaps[map_idx] = buf; | ||
998 | map_idx = (map_idx + 1) & (ARRAY_SIZE(mmaps) - 1); | ||
999 | file_pos = file_offset + head; | ||
797 | 1000 | ||
798 | more: | 1001 | more: |
799 | event = (event_t *)(buf + head); | 1002 | event = (event_t *)(buf + head); |
800 | ui_progress__update(progress, offset); | ||
801 | 1003 | ||
802 | if (self->header.needs_swap) | 1004 | if (session->header.needs_swap) |
803 | perf_event_header__bswap(&event->header); | 1005 | perf_event_header__bswap(&event->header); |
804 | size = event->header.size; | 1006 | size = event->header.size; |
805 | if (size == 0) | 1007 | if (size == 0) |
806 | size = 8; | 1008 | size = 8; |
807 | 1009 | ||
808 | if (head + event->header.size >= page_size * self->mmap_window) { | 1010 | if (head + event->header.size >= mmap_size) { |
809 | int munmap_ret; | 1011 | if (mmaps[map_idx]) { |
810 | 1012 | munmap(mmaps[map_idx], mmap_size); | |
811 | shift = page_size * (head / page_size); | 1013 | mmaps[map_idx] = NULL; |
812 | 1014 | } | |
813 | munmap_ret = munmap(buf, page_size * self->mmap_window); | ||
814 | assert(munmap_ret == 0); | ||
815 | 1015 | ||
816 | offset += shift; | 1016 | page_offset = page_size * (head / page_size); |
817 | head -= shift; | 1017 | file_offset += page_offset; |
1018 | head -= page_offset; | ||
818 | goto remap; | 1019 | goto remap; |
819 | } | 1020 | } |
820 | 1021 | ||
821 | size = event->header.size; | 1022 | size = event->header.size; |
822 | 1023 | ||
823 | dump_printf("\n%#Lx [%#x]: event: %d\n", | ||
824 | offset + head, event->header.size, event->header.type); | ||
825 | |||
826 | if (size == 0 || | 1024 | if (size == 0 || |
827 | perf_session__process_event(self, event, ops, offset, head) < 0) { | 1025 | perf_session__process_event(session, event, ops, file_pos) < 0) { |
828 | dump_printf("%#Lx [%#x]: skipping unknown header type: %d\n", | 1026 | dump_printf("%#Lx [%#x]: skipping unknown header type: %d\n", |
829 | offset + head, event->header.size, | 1027 | file_offset + head, event->header.size, |
830 | event->header.type); | 1028 | event->header.type); |
831 | /* | 1029 | /* |
832 | * assume we lost track of the stream, check alignment, and | 1030 | * assume we lost track of the stream, check alignment, and |
@@ -839,19 +1037,24 @@ more: | |||
839 | } | 1037 | } |
840 | 1038 | ||
841 | head += size; | 1039 | head += size; |
1040 | file_pos += size; | ||
842 | 1041 | ||
843 | if (offset + head >= data_offset + data_size) | 1042 | if (file_pos >= progress_next) { |
844 | goto done; | 1043 | progress_next += file_size / 16; |
1044 | ui_progress__update(progress, file_pos); | ||
1045 | } | ||
845 | 1046 | ||
846 | if (offset + head < file_size) | 1047 | if (file_pos < file_size) |
847 | goto more; | 1048 | goto more; |
848 | done: | 1049 | |
849 | err = 0; | 1050 | err = 0; |
850 | /* do the final flush for ordered samples */ | 1051 | /* do the final flush for ordered samples */ |
851 | self->ordered_samples.next_flush = ULLONG_MAX; | 1052 | session->ordered_samples.next_flush = ULLONG_MAX; |
852 | flush_sample_queue(self, ops); | 1053 | flush_sample_queue(session, ops); |
853 | out_err: | 1054 | out_err: |
854 | ui_progress__delete(progress); | 1055 | ui_progress__delete(progress); |
1056 | perf_session__warn_about_errors(session, ops); | ||
1057 | perf_session_free_sample_buffers(session); | ||
855 | return err; | 1058 | return err; |
856 | } | 1059 | } |
857 | 1060 | ||