diff options
author | Mel Gorman <mel@csn.ul.ie> | 2009-09-21 20:02:44 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2009-09-22 10:17:34 -0400 |
commit | 0d3d062a6e289e065bd0aa537a6806a1806bf8aa (patch) | |
tree | 9895e9cb48674d072885af3424e1ef145ec81f28 | |
parent | e0fff1bd12469c45dab088e353d8882761387bb6 (diff) |
tracing, page-allocator: add trace event for page traffic related to the buddy lists
The page allocation trace event reports that a page was successfully
allocated but it does not specify where it came from. When analysing
performance, it can be important to distinguish between pages coming from
the per-cpu allocator and pages coming from the buddy lists as the latter
requires the zone lock to the taken and more data structures to be
examined.
This patch adds a trace event for __rmqueue reporting when a page is being
allocated from the buddy lists. It distinguishes between being called to
refill the per-cpu lists or whether it is a high-order allocation.
Similarly, this patch adds an event to catch when the PCP lists are being
drained a little and pages are going back to the buddy lists.
This is trickier to draw conclusions from but high activity on those
events could explain why there were a large number of cache misses on a
page-allocator-intensive workload. The coalescing and splitting of
buddies involves a lot of writing of page metadata and cache line bounces
not to mention the acquisition of an interrupt-safe lock necessary to
enter this path.
[akpm@linux-foundation.org: fix build]
Signed-off-by: Mel Gorman <mel@csn.ul.ie>
Acked-by: Rik van Riel <riel@redhat.com>
Reviewed-by: Ingo Molnar <mingo@elte.hu>
Cc: Larry Woodman <lwoodman@redhat.com>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Li Ming Chun <macli@brc.ubc.ca>
Reviewed-by: KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-rw-r--r-- | include/trace/events/kmem.h | 51 | ||||
-rw-r--r-- | mm/page_alloc.c | 3 |
2 files changed, 54 insertions, 0 deletions
diff --git a/include/trace/events/kmem.h b/include/trace/events/kmem.h index aae16ee17601..eaf46bdd18a5 100644 --- a/include/trace/events/kmem.h +++ b/include/trace/events/kmem.h | |||
@@ -299,6 +299,57 @@ TRACE_EVENT(mm_page_alloc, | |||
299 | show_gfp_flags(__entry->gfp_flags)) | 299 | show_gfp_flags(__entry->gfp_flags)) |
300 | ); | 300 | ); |
301 | 301 | ||
302 | TRACE_EVENT(mm_page_alloc_zone_locked, | ||
303 | |||
304 | TP_PROTO(struct page *page, unsigned int order, int migratetype), | ||
305 | |||
306 | TP_ARGS(page, order, migratetype), | ||
307 | |||
308 | TP_STRUCT__entry( | ||
309 | __field( struct page *, page ) | ||
310 | __field( unsigned int, order ) | ||
311 | __field( int, migratetype ) | ||
312 | ), | ||
313 | |||
314 | TP_fast_assign( | ||
315 | __entry->page = page; | ||
316 | __entry->order = order; | ||
317 | __entry->migratetype = migratetype; | ||
318 | ), | ||
319 | |||
320 | TP_printk("page=%p pfn=%lu order=%u migratetype=%d percpu_refill=%d", | ||
321 | __entry->page, | ||
322 | page_to_pfn(__entry->page), | ||
323 | __entry->order, | ||
324 | __entry->migratetype, | ||
325 | __entry->order == 0) | ||
326 | ); | ||
327 | |||
328 | TRACE_EVENT(mm_page_pcpu_drain, | ||
329 | |||
330 | TP_PROTO(struct page *page, int order, int migratetype), | ||
331 | |||
332 | TP_ARGS(page, order, migratetype), | ||
333 | |||
334 | TP_STRUCT__entry( | ||
335 | __field( struct page *, page ) | ||
336 | __field( int, order ) | ||
337 | __field( int, migratetype ) | ||
338 | ), | ||
339 | |||
340 | TP_fast_assign( | ||
341 | __entry->page = page; | ||
342 | __entry->order = order; | ||
343 | __entry->migratetype = migratetype; | ||
344 | ), | ||
345 | |||
346 | TP_printk("page=%p pfn=%lu order=%d migratetype=%d", | ||
347 | __entry->page, | ||
348 | page_to_pfn(__entry->page), | ||
349 | __entry->order, | ||
350 | __entry->migratetype) | ||
351 | ); | ||
352 | |||
302 | TRACE_EVENT(mm_page_alloc_extfrag, | 353 | TRACE_EVENT(mm_page_alloc_extfrag, |
303 | 354 | ||
304 | TP_PROTO(struct page *page, | 355 | TP_PROTO(struct page *page, |
diff --git a/mm/page_alloc.c b/mm/page_alloc.c index 77f517c18b37..4c847cc57caf 100644 --- a/mm/page_alloc.c +++ b/mm/page_alloc.c | |||
@@ -48,6 +48,7 @@ | |||
48 | #include <linux/page_cgroup.h> | 48 | #include <linux/page_cgroup.h> |
49 | #include <linux/debugobjects.h> | 49 | #include <linux/debugobjects.h> |
50 | #include <linux/kmemleak.h> | 50 | #include <linux/kmemleak.h> |
51 | #include <trace/events/kmem.h> | ||
51 | 52 | ||
52 | #include <asm/tlbflush.h> | 53 | #include <asm/tlbflush.h> |
53 | #include <asm/div64.h> | 54 | #include <asm/div64.h> |
@@ -535,6 +536,7 @@ static void free_pages_bulk(struct zone *zone, int count, | |||
535 | page = list_entry(list->prev, struct page, lru); | 536 | page = list_entry(list->prev, struct page, lru); |
536 | /* have to delete it as __free_one_page list manipulates */ | 537 | /* have to delete it as __free_one_page list manipulates */ |
537 | list_del(&page->lru); | 538 | list_del(&page->lru); |
539 | trace_mm_page_pcpu_drain(page, order, page_private(page)); | ||
538 | __free_one_page(page, zone, order, page_private(page)); | 540 | __free_one_page(page, zone, order, page_private(page)); |
539 | } | 541 | } |
540 | spin_unlock(&zone->lock); | 542 | spin_unlock(&zone->lock); |
@@ -890,6 +892,7 @@ retry_reserve: | |||
890 | } | 892 | } |
891 | } | 893 | } |
892 | 894 | ||
895 | trace_mm_page_alloc_zone_locked(page, order, migratetype); | ||
893 | return page; | 896 | return page; |
894 | } | 897 | } |
895 | 898 | ||