diff options
Diffstat (limited to 'block/blk-core.c')
-rw-r--r-- | block/blk-core.c | 46 |
1 files changed, 27 insertions, 19 deletions
diff --git a/block/blk-core.c b/block/blk-core.c index c36aa98fafa3..561e8a1b43a4 100644 --- a/block/blk-core.c +++ b/block/blk-core.c | |||
@@ -28,9 +28,23 @@ | |||
28 | #include <linux/task_io_accounting_ops.h> | 28 | #include <linux/task_io_accounting_ops.h> |
29 | #include <linux/blktrace_api.h> | 29 | #include <linux/blktrace_api.h> |
30 | #include <linux/fault-inject.h> | 30 | #include <linux/fault-inject.h> |
31 | #include <trace/block.h> | ||
31 | 32 | ||
32 | #include "blk.h" | 33 | #include "blk.h" |
33 | 34 | ||
35 | DEFINE_TRACE(block_plug); | ||
36 | DEFINE_TRACE(block_unplug_io); | ||
37 | DEFINE_TRACE(block_unplug_timer); | ||
38 | DEFINE_TRACE(block_getrq); | ||
39 | DEFINE_TRACE(block_sleeprq); | ||
40 | DEFINE_TRACE(block_rq_requeue); | ||
41 | DEFINE_TRACE(block_bio_backmerge); | ||
42 | DEFINE_TRACE(block_bio_frontmerge); | ||
43 | DEFINE_TRACE(block_bio_queue); | ||
44 | DEFINE_TRACE(block_rq_complete); | ||
45 | DEFINE_TRACE(block_remap); /* Also used in drivers/md/dm.c */ | ||
46 | EXPORT_TRACEPOINT_SYMBOL_GPL(block_remap); | ||
47 | |||
34 | static int __make_request(struct request_queue *q, struct bio *bio); | 48 | static int __make_request(struct request_queue *q, struct bio *bio); |
35 | 49 | ||
36 | /* | 50 | /* |
@@ -205,7 +219,7 @@ void blk_plug_device(struct request_queue *q) | |||
205 | 219 | ||
206 | if (!queue_flag_test_and_set(QUEUE_FLAG_PLUGGED, q)) { | 220 | if (!queue_flag_test_and_set(QUEUE_FLAG_PLUGGED, q)) { |
207 | mod_timer(&q->unplug_timer, jiffies + q->unplug_delay); | 221 | mod_timer(&q->unplug_timer, jiffies + q->unplug_delay); |
208 | blk_add_trace_generic(q, NULL, 0, BLK_TA_PLUG); | 222 | trace_block_plug(q); |
209 | } | 223 | } |
210 | } | 224 | } |
211 | EXPORT_SYMBOL(blk_plug_device); | 225 | EXPORT_SYMBOL(blk_plug_device); |
@@ -292,9 +306,7 @@ void blk_unplug_work(struct work_struct *work) | |||
292 | struct request_queue *q = | 306 | struct request_queue *q = |
293 | container_of(work, struct request_queue, unplug_work); | 307 | container_of(work, struct request_queue, unplug_work); |
294 | 308 | ||
295 | blk_add_trace_pdu_int(q, BLK_TA_UNPLUG_IO, NULL, | 309 | trace_block_unplug_io(q); |
296 | q->rq.count[READ] + q->rq.count[WRITE]); | ||
297 | |||
298 | q->unplug_fn(q); | 310 | q->unplug_fn(q); |
299 | } | 311 | } |
300 | 312 | ||
@@ -302,9 +314,7 @@ void blk_unplug_timeout(unsigned long data) | |||
302 | { | 314 | { |
303 | struct request_queue *q = (struct request_queue *)data; | 315 | struct request_queue *q = (struct request_queue *)data; |
304 | 316 | ||
305 | blk_add_trace_pdu_int(q, BLK_TA_UNPLUG_TIMER, NULL, | 317 | trace_block_unplug_timer(q); |
306 | q->rq.count[READ] + q->rq.count[WRITE]); | ||
307 | |||
308 | kblockd_schedule_work(q, &q->unplug_work); | 318 | kblockd_schedule_work(q, &q->unplug_work); |
309 | } | 319 | } |
310 | 320 | ||
@@ -314,9 +324,7 @@ void blk_unplug(struct request_queue *q) | |||
314 | * devices don't necessarily have an ->unplug_fn defined | 324 | * devices don't necessarily have an ->unplug_fn defined |
315 | */ | 325 | */ |
316 | if (q->unplug_fn) { | 326 | if (q->unplug_fn) { |
317 | blk_add_trace_pdu_int(q, BLK_TA_UNPLUG_IO, NULL, | 327 | trace_block_unplug_io(q); |
318 | q->rq.count[READ] + q->rq.count[WRITE]); | ||
319 | |||
320 | q->unplug_fn(q); | 328 | q->unplug_fn(q); |
321 | } | 329 | } |
322 | } | 330 | } |
@@ -822,7 +830,7 @@ rq_starved: | |||
822 | if (ioc_batching(q, ioc)) | 830 | if (ioc_batching(q, ioc)) |
823 | ioc->nr_batch_requests--; | 831 | ioc->nr_batch_requests--; |
824 | 832 | ||
825 | blk_add_trace_generic(q, bio, rw, BLK_TA_GETRQ); | 833 | trace_block_getrq(q, bio, rw); |
826 | out: | 834 | out: |
827 | return rq; | 835 | return rq; |
828 | } | 836 | } |
@@ -848,7 +856,7 @@ static struct request *get_request_wait(struct request_queue *q, int rw_flags, | |||
848 | prepare_to_wait_exclusive(&rl->wait[rw], &wait, | 856 | prepare_to_wait_exclusive(&rl->wait[rw], &wait, |
849 | TASK_UNINTERRUPTIBLE); | 857 | TASK_UNINTERRUPTIBLE); |
850 | 858 | ||
851 | blk_add_trace_generic(q, bio, rw, BLK_TA_SLEEPRQ); | 859 | trace_block_sleeprq(q, bio, rw); |
852 | 860 | ||
853 | __generic_unplug_device(q); | 861 | __generic_unplug_device(q); |
854 | spin_unlock_irq(q->queue_lock); | 862 | spin_unlock_irq(q->queue_lock); |
@@ -928,7 +936,7 @@ void blk_requeue_request(struct request_queue *q, struct request *rq) | |||
928 | { | 936 | { |
929 | blk_delete_timer(rq); | 937 | blk_delete_timer(rq); |
930 | blk_clear_rq_complete(rq); | 938 | blk_clear_rq_complete(rq); |
931 | blk_add_trace_rq(q, rq, BLK_TA_REQUEUE); | 939 | trace_block_rq_requeue(q, rq); |
932 | 940 | ||
933 | if (blk_rq_tagged(rq)) | 941 | if (blk_rq_tagged(rq)) |
934 | blk_queue_end_tag(q, rq); | 942 | blk_queue_end_tag(q, rq); |
@@ -1167,7 +1175,7 @@ static int __make_request(struct request_queue *q, struct bio *bio) | |||
1167 | if (!ll_back_merge_fn(q, req, bio)) | 1175 | if (!ll_back_merge_fn(q, req, bio)) |
1168 | break; | 1176 | break; |
1169 | 1177 | ||
1170 | blk_add_trace_bio(q, bio, BLK_TA_BACKMERGE); | 1178 | trace_block_bio_backmerge(q, bio); |
1171 | 1179 | ||
1172 | req->biotail->bi_next = bio; | 1180 | req->biotail->bi_next = bio; |
1173 | req->biotail = bio; | 1181 | req->biotail = bio; |
@@ -1186,7 +1194,7 @@ static int __make_request(struct request_queue *q, struct bio *bio) | |||
1186 | if (!ll_front_merge_fn(q, req, bio)) | 1194 | if (!ll_front_merge_fn(q, req, bio)) |
1187 | break; | 1195 | break; |
1188 | 1196 | ||
1189 | blk_add_trace_bio(q, bio, BLK_TA_FRONTMERGE); | 1197 | trace_block_bio_frontmerge(q, bio); |
1190 | 1198 | ||
1191 | bio->bi_next = req->bio; | 1199 | bio->bi_next = req->bio; |
1192 | req->bio = bio; | 1200 | req->bio = bio; |
@@ -1269,7 +1277,7 @@ static inline void blk_partition_remap(struct bio *bio) | |||
1269 | bio->bi_sector += p->start_sect; | 1277 | bio->bi_sector += p->start_sect; |
1270 | bio->bi_bdev = bdev->bd_contains; | 1278 | bio->bi_bdev = bdev->bd_contains; |
1271 | 1279 | ||
1272 | blk_add_trace_remap(bdev_get_queue(bio->bi_bdev), bio, | 1280 | trace_block_remap(bdev_get_queue(bio->bi_bdev), bio, |
1273 | bdev->bd_dev, bio->bi_sector, | 1281 | bdev->bd_dev, bio->bi_sector, |
1274 | bio->bi_sector - p->start_sect); | 1282 | bio->bi_sector - p->start_sect); |
1275 | } | 1283 | } |
@@ -1441,10 +1449,10 @@ end_io: | |||
1441 | goto end_io; | 1449 | goto end_io; |
1442 | 1450 | ||
1443 | if (old_sector != -1) | 1451 | if (old_sector != -1) |
1444 | blk_add_trace_remap(q, bio, old_dev, bio->bi_sector, | 1452 | trace_block_remap(q, bio, old_dev, bio->bi_sector, |
1445 | old_sector); | 1453 | old_sector); |
1446 | 1454 | ||
1447 | blk_add_trace_bio(q, bio, BLK_TA_QUEUE); | 1455 | trace_block_bio_queue(q, bio); |
1448 | 1456 | ||
1449 | old_sector = bio->bi_sector; | 1457 | old_sector = bio->bi_sector; |
1450 | old_dev = bio->bi_bdev->bd_dev; | 1458 | old_dev = bio->bi_bdev->bd_dev; |
@@ -1678,7 +1686,7 @@ static int __end_that_request_first(struct request *req, int error, | |||
1678 | int total_bytes, bio_nbytes, next_idx = 0; | 1686 | int total_bytes, bio_nbytes, next_idx = 0; |
1679 | struct bio *bio; | 1687 | struct bio *bio; |
1680 | 1688 | ||
1681 | blk_add_trace_rq(req->q, req, BLK_TA_COMPLETE); | 1689 | trace_block_rq_complete(req->q, req); |
1682 | 1690 | ||
1683 | /* | 1691 | /* |
1684 | * for a REQ_TYPE_BLOCK_PC request, we want to carry any eventual | 1692 | * for a REQ_TYPE_BLOCK_PC request, we want to carry any eventual |