diff options
author | Christoph Hellwig <hch@infradead.org> | 2014-03-21 10:57:37 -0400 |
---|---|---|
committer | Jens Axboe <axboe@fb.com> | 2014-03-21 10:57:37 -0400 |
commit | eeabc850b79336575da7be3dbe186a2da4de8293 (patch) | |
tree | 4cc10d227c84bbe254bc3ce1b1513c9894ab364d /block | |
parent | 081241e592c47f4ed2999a0b576ae85e765c6da4 (diff) |
blk-mq: merge blk_mq_insert_request and blk_mq_run_request
It's almost identical to blk_mq_insert_request, so fold the two into one
slightly more generic function by making the flush special case a bit
smarted.
Signed-off-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Jens Axboe <axboe@fb.com>
Diffstat (limited to 'block')
-rw-r--r-- | block/blk-exec.c | 2 | ||||
-rw-r--r-- | block/blk-flush.c | 4 | ||||
-rw-r--r-- | block/blk-mq.c | 53 | ||||
-rw-r--r-- | block/blk-mq.h | 1 |
4 files changed, 13 insertions, 47 deletions
diff --git a/block/blk-exec.c b/block/blk-exec.c index c68613bb4c79..dbf4502b1d67 100644 --- a/block/blk-exec.c +++ b/block/blk-exec.c | |||
@@ -65,7 +65,7 @@ void blk_execute_rq_nowait(struct request_queue *q, struct gendisk *bd_disk, | |||
65 | * be resued after dying flag is set | 65 | * be resued after dying flag is set |
66 | */ | 66 | */ |
67 | if (q->mq_ops) { | 67 | if (q->mq_ops) { |
68 | blk_mq_insert_request(q, rq, at_head, true); | 68 | blk_mq_insert_request(rq, at_head, true, false); |
69 | return; | 69 | return; |
70 | } | 70 | } |
71 | 71 | ||
diff --git a/block/blk-flush.c b/block/blk-flush.c index 66e2b697f5db..f598f794c3c6 100644 --- a/block/blk-flush.c +++ b/block/blk-flush.c | |||
@@ -137,7 +137,7 @@ static void mq_flush_run(struct work_struct *work) | |||
137 | rq = container_of(work, struct request, mq_flush_work); | 137 | rq = container_of(work, struct request, mq_flush_work); |
138 | 138 | ||
139 | memset(&rq->csd, 0, sizeof(rq->csd)); | 139 | memset(&rq->csd, 0, sizeof(rq->csd)); |
140 | blk_mq_run_request(rq, true, false); | 140 | blk_mq_insert_request(rq, false, true, false); |
141 | } | 141 | } |
142 | 142 | ||
143 | static bool blk_flush_queue_rq(struct request *rq) | 143 | static bool blk_flush_queue_rq(struct request *rq) |
@@ -411,7 +411,7 @@ void blk_insert_flush(struct request *rq) | |||
411 | if ((policy & REQ_FSEQ_DATA) && | 411 | if ((policy & REQ_FSEQ_DATA) && |
412 | !(policy & (REQ_FSEQ_PREFLUSH | REQ_FSEQ_POSTFLUSH))) { | 412 | !(policy & (REQ_FSEQ_PREFLUSH | REQ_FSEQ_POSTFLUSH))) { |
413 | if (q->mq_ops) { | 413 | if (q->mq_ops) { |
414 | blk_mq_run_request(rq, false, true); | 414 | blk_mq_insert_request(rq, false, false, true); |
415 | } else | 415 | } else |
416 | list_add_tail(&rq->queuelist, &q->queue_head); | 416 | list_add_tail(&rq->queuelist, &q->queue_head); |
417 | return; | 417 | return; |
diff --git a/block/blk-mq.c b/block/blk-mq.c index a56e77383738..81ff7879bac8 100644 --- a/block/blk-mq.c +++ b/block/blk-mq.c | |||
@@ -724,61 +724,28 @@ static void __blk_mq_insert_request(struct blk_mq_hw_ctx *hctx, | |||
724 | blk_mq_add_timer(rq); | 724 | blk_mq_add_timer(rq); |
725 | } | 725 | } |
726 | 726 | ||
727 | void blk_mq_insert_request(struct request_queue *q, struct request *rq, | 727 | void blk_mq_insert_request(struct request *rq, bool at_head, bool run_queue, |
728 | bool at_head, bool run_queue) | 728 | bool async) |
729 | { | 729 | { |
730 | struct request_queue *q = rq->q; | ||
730 | struct blk_mq_hw_ctx *hctx; | 731 | struct blk_mq_hw_ctx *hctx; |
731 | struct blk_mq_ctx *ctx, *current_ctx; | 732 | struct blk_mq_ctx *ctx = rq->mq_ctx, *current_ctx; |
733 | |||
734 | current_ctx = blk_mq_get_ctx(q); | ||
735 | if (!cpu_online(ctx->cpu)) | ||
736 | rq->mq_ctx = ctx = current_ctx; | ||
732 | 737 | ||
733 | ctx = rq->mq_ctx; | ||
734 | hctx = q->mq_ops->map_queue(q, ctx->cpu); | 738 | hctx = q->mq_ops->map_queue(q, ctx->cpu); |
735 | 739 | ||
736 | if (rq->cmd_flags & (REQ_FLUSH | REQ_FUA)) { | 740 | if (rq->cmd_flags & (REQ_FLUSH | REQ_FUA) && |
741 | !(rq->cmd_flags & (REQ_FLUSH_SEQ))) { | ||
737 | blk_insert_flush(rq); | 742 | blk_insert_flush(rq); |
738 | } else { | 743 | } else { |
739 | current_ctx = blk_mq_get_ctx(q); | ||
740 | |||
741 | if (!cpu_online(ctx->cpu)) { | ||
742 | ctx = current_ctx; | ||
743 | hctx = q->mq_ops->map_queue(q, ctx->cpu); | ||
744 | rq->mq_ctx = ctx; | ||
745 | } | ||
746 | spin_lock(&ctx->lock); | 744 | spin_lock(&ctx->lock); |
747 | __blk_mq_insert_request(hctx, rq, at_head); | 745 | __blk_mq_insert_request(hctx, rq, at_head); |
748 | spin_unlock(&ctx->lock); | 746 | spin_unlock(&ctx->lock); |
749 | |||
750 | blk_mq_put_ctx(current_ctx); | ||
751 | } | 747 | } |
752 | 748 | ||
753 | if (run_queue) | ||
754 | __blk_mq_run_hw_queue(hctx); | ||
755 | } | ||
756 | EXPORT_SYMBOL(blk_mq_insert_request); | ||
757 | |||
758 | /* | ||
759 | * This is a special version of blk_mq_insert_request to bypass FLUSH request | ||
760 | * check. Should only be used internally. | ||
761 | */ | ||
762 | void blk_mq_run_request(struct request *rq, bool run_queue, bool async) | ||
763 | { | ||
764 | struct request_queue *q = rq->q; | ||
765 | struct blk_mq_hw_ctx *hctx; | ||
766 | struct blk_mq_ctx *ctx, *current_ctx; | ||
767 | |||
768 | current_ctx = blk_mq_get_ctx(q); | ||
769 | |||
770 | ctx = rq->mq_ctx; | ||
771 | if (!cpu_online(ctx->cpu)) { | ||
772 | ctx = current_ctx; | ||
773 | rq->mq_ctx = ctx; | ||
774 | } | ||
775 | hctx = q->mq_ops->map_queue(q, ctx->cpu); | ||
776 | |||
777 | /* ctx->cpu might be offline */ | ||
778 | spin_lock(&ctx->lock); | ||
779 | __blk_mq_insert_request(hctx, rq, false); | ||
780 | spin_unlock(&ctx->lock); | ||
781 | |||
782 | blk_mq_put_ctx(current_ctx); | 749 | blk_mq_put_ctx(current_ctx); |
783 | 750 | ||
784 | if (run_queue) | 751 | if (run_queue) |
diff --git a/block/blk-mq.h b/block/blk-mq.h index 361f9343dab1..ebbe6bac9d61 100644 --- a/block/blk-mq.h +++ b/block/blk-mq.h | |||
@@ -23,7 +23,6 @@ struct blk_mq_ctx { | |||
23 | }; | 23 | }; |
24 | 24 | ||
25 | void __blk_mq_complete_request(struct request *rq); | 25 | void __blk_mq_complete_request(struct request *rq); |
26 | void blk_mq_run_request(struct request *rq, bool run_queue, bool async); | ||
27 | void blk_mq_run_hw_queue(struct blk_mq_hw_ctx *hctx, bool async); | 26 | void blk_mq_run_hw_queue(struct blk_mq_hw_ctx *hctx, bool async); |
28 | void blk_mq_init_flush(struct request_queue *q); | 27 | void blk_mq_init_flush(struct request_queue *q); |
29 | void blk_mq_drain_queue(struct request_queue *q); | 28 | void blk_mq_drain_queue(struct request_queue *q); |