diff options
author | Tejun Heo <tj@kernel.org> | 2012-06-04 23:40:59 -0400 |
---|---|---|
committer | Jens Axboe <axboe@kernel.dk> | 2012-06-25 05:53:52 -0400 |
commit | 5b788ce3e2acac9bf109743b1281d77347cf2101 (patch) | |
tree | 907477e42d27bec9a2060fcc709402b7636390c9 /block/blk-core.c | |
parent | 8a5ecdd42862bf87ceab00bf2a15d7eabf58c02d (diff) |
block: prepare for multiple request_lists
Request allocation is about to be made per-blkg meaning that there'll
be multiple request lists.
* Make queue full state per request_list. blk_*queue_full() functions
are renamed to blk_*rl_full() and takes @rl instead of @q.
* Rename blk_init_free_list() to blk_init_rl() and make it take @rl
instead of @q. Also add @gfp_mask parameter.
* Add blk_exit_rl() instead of destroying rl directly from
blk_release_queue().
* Add request_list->q and make request alloc/free functions -
blk_free_request(), [__]freed_request(), __get_request() - take @rl
instead of @q.
This patch doesn't introduce any functional difference.
Signed-off-by: Tejun Heo <tj@kernel.org>
Acked-by: Vivek Goyal <vgoyal@redhat.com>
Signed-off-by: Jens Axboe <axboe@kernel.dk>
Diffstat (limited to 'block/blk-core.c')
-rw-r--r-- | block/blk-core.c | 56 |
1 files changed, 31 insertions, 25 deletions
diff --git a/block/blk-core.c b/block/blk-core.c index a2648153691f..f392a2edf462 100644 --- a/block/blk-core.c +++ b/block/blk-core.c | |||
@@ -517,13 +517,13 @@ void blk_cleanup_queue(struct request_queue *q) | |||
517 | } | 517 | } |
518 | EXPORT_SYMBOL(blk_cleanup_queue); | 518 | EXPORT_SYMBOL(blk_cleanup_queue); |
519 | 519 | ||
520 | static int blk_init_free_list(struct request_queue *q) | 520 | int blk_init_rl(struct request_list *rl, struct request_queue *q, |
521 | gfp_t gfp_mask) | ||
521 | { | 522 | { |
522 | struct request_list *rl = &q->rq; | ||
523 | |||
524 | if (unlikely(rl->rq_pool)) | 523 | if (unlikely(rl->rq_pool)) |
525 | return 0; | 524 | return 0; |
526 | 525 | ||
526 | rl->q = q; | ||
527 | rl->count[BLK_RW_SYNC] = rl->count[BLK_RW_ASYNC] = 0; | 527 | rl->count[BLK_RW_SYNC] = rl->count[BLK_RW_ASYNC] = 0; |
528 | rl->starved[BLK_RW_SYNC] = rl->starved[BLK_RW_ASYNC] = 0; | 528 | rl->starved[BLK_RW_SYNC] = rl->starved[BLK_RW_ASYNC] = 0; |
529 | init_waitqueue_head(&rl->wait[BLK_RW_SYNC]); | 529 | init_waitqueue_head(&rl->wait[BLK_RW_SYNC]); |
@@ -531,13 +531,19 @@ static int blk_init_free_list(struct request_queue *q) | |||
531 | 531 | ||
532 | rl->rq_pool = mempool_create_node(BLKDEV_MIN_RQ, mempool_alloc_slab, | 532 | rl->rq_pool = mempool_create_node(BLKDEV_MIN_RQ, mempool_alloc_slab, |
533 | mempool_free_slab, request_cachep, | 533 | mempool_free_slab, request_cachep, |
534 | GFP_KERNEL, q->node); | 534 | gfp_mask, q->node); |
535 | if (!rl->rq_pool) | 535 | if (!rl->rq_pool) |
536 | return -ENOMEM; | 536 | return -ENOMEM; |
537 | 537 | ||
538 | return 0; | 538 | return 0; |
539 | } | 539 | } |
540 | 540 | ||
541 | void blk_exit_rl(struct request_list *rl) | ||
542 | { | ||
543 | if (rl->rq_pool) | ||
544 | mempool_destroy(rl->rq_pool); | ||
545 | } | ||
546 | |||
541 | struct request_queue *blk_alloc_queue(gfp_t gfp_mask) | 547 | struct request_queue *blk_alloc_queue(gfp_t gfp_mask) |
542 | { | 548 | { |
543 | return blk_alloc_queue_node(gfp_mask, -1); | 549 | return blk_alloc_queue_node(gfp_mask, -1); |
@@ -679,7 +685,7 @@ blk_init_allocated_queue(struct request_queue *q, request_fn_proc *rfn, | |||
679 | if (!q) | 685 | if (!q) |
680 | return NULL; | 686 | return NULL; |
681 | 687 | ||
682 | if (blk_init_free_list(q)) | 688 | if (blk_init_rl(&q->rq, q, GFP_KERNEL)) |
683 | return NULL; | 689 | return NULL; |
684 | 690 | ||
685 | q->request_fn = rfn; | 691 | q->request_fn = rfn; |
@@ -721,15 +727,15 @@ bool blk_get_queue(struct request_queue *q) | |||
721 | } | 727 | } |
722 | EXPORT_SYMBOL(blk_get_queue); | 728 | EXPORT_SYMBOL(blk_get_queue); |
723 | 729 | ||
724 | static inline void blk_free_request(struct request_queue *q, struct request *rq) | 730 | static inline void blk_free_request(struct request_list *rl, struct request *rq) |
725 | { | 731 | { |
726 | if (rq->cmd_flags & REQ_ELVPRIV) { | 732 | if (rq->cmd_flags & REQ_ELVPRIV) { |
727 | elv_put_request(q, rq); | 733 | elv_put_request(rl->q, rq); |
728 | if (rq->elv.icq) | 734 | if (rq->elv.icq) |
729 | put_io_context(rq->elv.icq->ioc); | 735 | put_io_context(rq->elv.icq->ioc); |
730 | } | 736 | } |
731 | 737 | ||
732 | mempool_free(rq, q->rq.rq_pool); | 738 | mempool_free(rq, rl->rq_pool); |
733 | } | 739 | } |
734 | 740 | ||
735 | /* | 741 | /* |
@@ -766,9 +772,9 @@ static void ioc_set_batching(struct request_queue *q, struct io_context *ioc) | |||
766 | ioc->last_waited = jiffies; | 772 | ioc->last_waited = jiffies; |
767 | } | 773 | } |
768 | 774 | ||
769 | static void __freed_request(struct request_queue *q, int sync) | 775 | static void __freed_request(struct request_list *rl, int sync) |
770 | { | 776 | { |
771 | struct request_list *rl = &q->rq; | 777 | struct request_queue *q = rl->q; |
772 | 778 | ||
773 | if (rl->count[sync] < queue_congestion_off_threshold(q)) | 779 | if (rl->count[sync] < queue_congestion_off_threshold(q)) |
774 | blk_clear_queue_congested(q, sync); | 780 | blk_clear_queue_congested(q, sync); |
@@ -777,7 +783,7 @@ static void __freed_request(struct request_queue *q, int sync) | |||
777 | if (waitqueue_active(&rl->wait[sync])) | 783 | if (waitqueue_active(&rl->wait[sync])) |
778 | wake_up(&rl->wait[sync]); | 784 | wake_up(&rl->wait[sync]); |
779 | 785 | ||
780 | blk_clear_queue_full(q, sync); | 786 | blk_clear_rl_full(rl, sync); |
781 | } | 787 | } |
782 | } | 788 | } |
783 | 789 | ||
@@ -785,9 +791,9 @@ static void __freed_request(struct request_queue *q, int sync) | |||
785 | * A request has just been released. Account for it, update the full and | 791 | * A request has just been released. Account for it, update the full and |
786 | * congestion status, wake up any waiters. Called under q->queue_lock. | 792 | * congestion status, wake up any waiters. Called under q->queue_lock. |
787 | */ | 793 | */ |
788 | static void freed_request(struct request_queue *q, unsigned int flags) | 794 | static void freed_request(struct request_list *rl, unsigned int flags) |
789 | { | 795 | { |
790 | struct request_list *rl = &q->rq; | 796 | struct request_queue *q = rl->q; |
791 | int sync = rw_is_sync(flags); | 797 | int sync = rw_is_sync(flags); |
792 | 798 | ||
793 | q->nr_rqs[sync]--; | 799 | q->nr_rqs[sync]--; |
@@ -795,10 +801,10 @@ static void freed_request(struct request_queue *q, unsigned int flags) | |||
795 | if (flags & REQ_ELVPRIV) | 801 | if (flags & REQ_ELVPRIV) |
796 | q->nr_rqs_elvpriv--; | 802 | q->nr_rqs_elvpriv--; |
797 | 803 | ||
798 | __freed_request(q, sync); | 804 | __freed_request(rl, sync); |
799 | 805 | ||
800 | if (unlikely(rl->starved[sync ^ 1])) | 806 | if (unlikely(rl->starved[sync ^ 1])) |
801 | __freed_request(q, sync ^ 1); | 807 | __freed_request(rl, sync ^ 1); |
802 | } | 808 | } |
803 | 809 | ||
804 | /* | 810 | /* |
@@ -838,7 +844,7 @@ static struct io_context *rq_ioc(struct bio *bio) | |||
838 | 844 | ||
839 | /** | 845 | /** |
840 | * __get_request - get a free request | 846 | * __get_request - get a free request |
841 | * @q: request_queue to allocate request from | 847 | * @rl: request list to allocate from |
842 | * @rw_flags: RW and SYNC flags | 848 | * @rw_flags: RW and SYNC flags |
843 | * @bio: bio to allocate request for (can be %NULL) | 849 | * @bio: bio to allocate request for (can be %NULL) |
844 | * @gfp_mask: allocation mask | 850 | * @gfp_mask: allocation mask |
@@ -850,11 +856,11 @@ static struct io_context *rq_ioc(struct bio *bio) | |||
850 | * Returns %NULL on failure, with @q->queue_lock held. | 856 | * Returns %NULL on failure, with @q->queue_lock held. |
851 | * Returns !%NULL on success, with @q->queue_lock *not held*. | 857 | * Returns !%NULL on success, with @q->queue_lock *not held*. |
852 | */ | 858 | */ |
853 | static struct request *__get_request(struct request_queue *q, int rw_flags, | 859 | static struct request *__get_request(struct request_list *rl, int rw_flags, |
854 | struct bio *bio, gfp_t gfp_mask) | 860 | struct bio *bio, gfp_t gfp_mask) |
855 | { | 861 | { |
862 | struct request_queue *q = rl->q; | ||
856 | struct request *rq; | 863 | struct request *rq; |
857 | struct request_list *rl = &q->rq; | ||
858 | struct elevator_type *et = q->elevator->type; | 864 | struct elevator_type *et = q->elevator->type; |
859 | struct io_context *ioc = rq_ioc(bio); | 865 | struct io_context *ioc = rq_ioc(bio); |
860 | struct io_cq *icq = NULL; | 866 | struct io_cq *icq = NULL; |
@@ -876,9 +882,9 @@ static struct request *__get_request(struct request_queue *q, int rw_flags, | |||
876 | * This process will be allowed to complete a batch of | 882 | * This process will be allowed to complete a batch of |
877 | * requests, others will be blocked. | 883 | * requests, others will be blocked. |
878 | */ | 884 | */ |
879 | if (!blk_queue_full(q, is_sync)) { | 885 | if (!blk_rl_full(rl, is_sync)) { |
880 | ioc_set_batching(q, ioc); | 886 | ioc_set_batching(q, ioc); |
881 | blk_set_queue_full(q, is_sync); | 887 | blk_set_rl_full(rl, is_sync); |
882 | } else { | 888 | } else { |
883 | if (may_queue != ELV_MQUEUE_MUST | 889 | if (may_queue != ELV_MQUEUE_MUST |
884 | && !ioc_batching(q, ioc)) { | 890 | && !ioc_batching(q, ioc)) { |
@@ -928,7 +934,7 @@ static struct request *__get_request(struct request_queue *q, int rw_flags, | |||
928 | spin_unlock_irq(q->queue_lock); | 934 | spin_unlock_irq(q->queue_lock); |
929 | 935 | ||
930 | /* allocate and init request */ | 936 | /* allocate and init request */ |
931 | rq = mempool_alloc(q->rq.rq_pool, gfp_mask); | 937 | rq = mempool_alloc(rl->rq_pool, gfp_mask); |
932 | if (!rq) | 938 | if (!rq) |
933 | goto fail_alloc; | 939 | goto fail_alloc; |
934 | 940 | ||
@@ -992,7 +998,7 @@ fail_alloc: | |||
992 | * queue, but this is pretty rare. | 998 | * queue, but this is pretty rare. |
993 | */ | 999 | */ |
994 | spin_lock_irq(q->queue_lock); | 1000 | spin_lock_irq(q->queue_lock); |
995 | freed_request(q, rw_flags); | 1001 | freed_request(rl, rw_flags); |
996 | 1002 | ||
997 | /* | 1003 | /* |
998 | * in the very unlikely event that allocation failed and no | 1004 | * in the very unlikely event that allocation failed and no |
@@ -1029,7 +1035,7 @@ static struct request *get_request(struct request_queue *q, int rw_flags, | |||
1029 | struct request_list *rl = &q->rq; | 1035 | struct request_list *rl = &q->rq; |
1030 | struct request *rq; | 1036 | struct request *rq; |
1031 | retry: | 1037 | retry: |
1032 | rq = __get_request(q, rw_flags, bio, gfp_mask); | 1038 | rq = __get_request(&q->rq, rw_flags, bio, gfp_mask); |
1033 | if (rq) | 1039 | if (rq) |
1034 | return rq; | 1040 | return rq; |
1035 | 1041 | ||
@@ -1229,8 +1235,8 @@ void __blk_put_request(struct request_queue *q, struct request *req) | |||
1229 | BUG_ON(!list_empty(&req->queuelist)); | 1235 | BUG_ON(!list_empty(&req->queuelist)); |
1230 | BUG_ON(!hlist_unhashed(&req->hash)); | 1236 | BUG_ON(!hlist_unhashed(&req->hash)); |
1231 | 1237 | ||
1232 | blk_free_request(q, req); | 1238 | blk_free_request(&q->rq, req); |
1233 | freed_request(q, flags); | 1239 | freed_request(&q->rq, flags); |
1234 | } | 1240 | } |
1235 | } | 1241 | } |
1236 | EXPORT_SYMBOL_GPL(__blk_put_request); | 1242 | EXPORT_SYMBOL_GPL(__blk_put_request); |