diff options
author | Steven Rostedt <srostedt@redhat.com> | 2009-03-03 19:51:40 -0500 |
---|---|---|
committer | Steven Rostedt <srostedt@redhat.com> | 2009-03-03 20:52:27 -0500 |
commit | 474d32b68d6d842f3e710e9ae9fe2568c53339f8 (patch) | |
tree | 674376b2f7a45cf6558879a4985398397dc96e79 | |
parent | e3d6bf0a0781a269f34250fd41e0d3dbfe540cf1 (diff) |
ring-buffer: make ring_buffer_read_page read from start on partial page
Impact: dont leave holes in read buffer page
The ring_buffer_read_page swaps a given page with the reader page
of the ring buffer, if certain conditions are set:
1) requested length is big enough to hold entire page data
2) a writer is not currently on the page
3) the page is not partially consumed.
Instead of swapping with the supplied page. It copies the data to
the supplied page instead. But currently the data is copied in the
same offset as the source page. This causes a hole at the start
of the reader page. This complicates the use of this function.
Instead, it should copy the data at the beginning of the function
and update the index fields accordingly.
Other small clean ups are also done in this patch.
Signed-off-by: Steven Rostedt <srostedt@redhat.com>
-rw-r--r-- | kernel/trace/ring_buffer.c | 43 |
1 files changed, 33 insertions, 10 deletions
diff --git a/kernel/trace/ring_buffer.c b/kernel/trace/ring_buffer.c index 27cf834d8b4e..f2a163db52f9 100644 --- a/kernel/trace/ring_buffer.c +++ b/kernel/trace/ring_buffer.c | |||
@@ -61,6 +61,8 @@ enum { | |||
61 | 61 | ||
62 | static unsigned long ring_buffer_flags __read_mostly = RB_BUFFERS_ON; | 62 | static unsigned long ring_buffer_flags __read_mostly = RB_BUFFERS_ON; |
63 | 63 | ||
64 | #define BUF_PAGE_HDR_SIZE offsetof(struct buffer_data_page, data) | ||
65 | |||
64 | /** | 66 | /** |
65 | * tracing_on - enable all tracing buffers | 67 | * tracing_on - enable all tracing buffers |
66 | * | 68 | * |
@@ -234,9 +236,16 @@ static void rb_init_page(struct buffer_data_page *bpage) | |||
234 | local_set(&bpage->commit, 0); | 236 | local_set(&bpage->commit, 0); |
235 | } | 237 | } |
236 | 238 | ||
239 | /** | ||
240 | * ring_buffer_page_len - the size of data on the page. | ||
241 | * @page: The page to read | ||
242 | * | ||
243 | * Returns the amount of data on the page, including buffer page header. | ||
244 | */ | ||
237 | size_t ring_buffer_page_len(void *page) | 245 | size_t ring_buffer_page_len(void *page) |
238 | { | 246 | { |
239 | return local_read(&((struct buffer_data_page *)page)->commit); | 247 | return local_read(&((struct buffer_data_page *)page)->commit) |
248 | + BUF_PAGE_HDR_SIZE; | ||
240 | } | 249 | } |
241 | 250 | ||
242 | /* | 251 | /* |
@@ -259,7 +268,7 @@ static inline int test_time_stamp(u64 delta) | |||
259 | return 0; | 268 | return 0; |
260 | } | 269 | } |
261 | 270 | ||
262 | #define BUF_PAGE_SIZE (PAGE_SIZE - offsetof(struct buffer_data_page, data)) | 271 | #define BUF_PAGE_SIZE (PAGE_SIZE - BUF_PAGE_HDR_SIZE) |
263 | 272 | ||
264 | /* | 273 | /* |
265 | * head_page == tail_page && head == tail then buffer is empty. | 274 | * head_page == tail_page && head == tail then buffer is empty. |
@@ -2454,6 +2463,15 @@ int ring_buffer_read_page(struct ring_buffer *buffer, | |||
2454 | unsigned int read; | 2463 | unsigned int read; |
2455 | int ret = -1; | 2464 | int ret = -1; |
2456 | 2465 | ||
2466 | /* | ||
2467 | * If len is not big enough to hold the page header, then | ||
2468 | * we can not copy anything. | ||
2469 | */ | ||
2470 | if (len <= BUF_PAGE_HDR_SIZE) | ||
2471 | return -1; | ||
2472 | |||
2473 | len -= BUF_PAGE_HDR_SIZE; | ||
2474 | |||
2457 | if (!data_page) | 2475 | if (!data_page) |
2458 | return -1; | 2476 | return -1; |
2459 | 2477 | ||
@@ -2473,15 +2491,17 @@ int ring_buffer_read_page(struct ring_buffer *buffer, | |||
2473 | commit = rb_page_commit(reader); | 2491 | commit = rb_page_commit(reader); |
2474 | 2492 | ||
2475 | /* | 2493 | /* |
2476 | * If len > what's left on the page, and the writer is also off of | 2494 | * If this page has been partially read or |
2477 | * the read page, then simply switch the read page with the given | 2495 | * if len is not big enough to read the rest of the page or |
2478 | * page. Otherwise we need to copy the data from the reader to the | 2496 | * a writer is still on the page, then |
2479 | * writer. | 2497 | * we must copy the data from the page to the buffer. |
2498 | * Otherwise, we can simply swap the page with the one passed in. | ||
2480 | */ | 2499 | */ |
2481 | if ((len < (commit - read)) || | 2500 | if (read || (len < (commit - read)) || |
2482 | cpu_buffer->reader_page == cpu_buffer->commit_page) { | 2501 | cpu_buffer->reader_page == cpu_buffer->commit_page) { |
2483 | struct buffer_data_page *rpage = cpu_buffer->reader_page->page; | 2502 | struct buffer_data_page *rpage = cpu_buffer->reader_page->page; |
2484 | unsigned int pos = read; | 2503 | unsigned int rpos = read; |
2504 | unsigned int pos = 0; | ||
2485 | unsigned int size; | 2505 | unsigned int size; |
2486 | 2506 | ||
2487 | if (full) | 2507 | if (full) |
@@ -2497,12 +2517,13 @@ int ring_buffer_read_page(struct ring_buffer *buffer, | |||
2497 | 2517 | ||
2498 | /* Need to copy one event at a time */ | 2518 | /* Need to copy one event at a time */ |
2499 | do { | 2519 | do { |
2500 | memcpy(bpage->data + pos, rpage->data + pos, size); | 2520 | memcpy(bpage->data + pos, rpage->data + rpos, size); |
2501 | 2521 | ||
2502 | len -= size; | 2522 | len -= size; |
2503 | 2523 | ||
2504 | rb_advance_reader(cpu_buffer); | 2524 | rb_advance_reader(cpu_buffer); |
2505 | pos = reader->read; | 2525 | rpos = reader->read; |
2526 | pos += size; | ||
2506 | 2527 | ||
2507 | event = rb_reader_event(cpu_buffer); | 2528 | event = rb_reader_event(cpu_buffer); |
2508 | size = rb_event_length(event); | 2529 | size = rb_event_length(event); |
@@ -2512,6 +2533,8 @@ int ring_buffer_read_page(struct ring_buffer *buffer, | |||
2512 | local_set(&bpage->commit, pos); | 2533 | local_set(&bpage->commit, pos); |
2513 | bpage->time_stamp = rpage->time_stamp; | 2534 | bpage->time_stamp = rpage->time_stamp; |
2514 | 2535 | ||
2536 | /* we copied everything to the beginning */ | ||
2537 | read = 0; | ||
2515 | } else { | 2538 | } else { |
2516 | /* swap the pages */ | 2539 | /* swap the pages */ |
2517 | rb_init_page(bpage); | 2540 | rb_init_page(bpage); |