summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorMing Lei <ming.lei@redhat.com>2019-06-04 09:08:02 -0400
committerJens Axboe <axboe@kernel.dk>2019-06-07 00:39:39 -0400
commitc3e2219216c92919a6bd1711f340f5faa98695e6 (patch)
tree392cb3a4e9f9de822cd0ef9bfcfca787adcca366
parentcf1db7fc8c2d31222701bd5c01b9cbaf89d8e7ce (diff)
block: free sched's request pool in blk_cleanup_queue
In theory, IO scheduler belongs to request queue, and the request pool of sched tags belongs to the request queue too. However, the current tags allocation interfaces are re-used for both driver tags and sched tags, and driver tags is definitely host wide, and doesn't belong to any request queue, same with its request pool. So we need tagset instance for freeing request of sched tags. Meantime, blk_mq_free_tag_set() often follows blk_cleanup_queue() in case of non-BLK_MQ_F_TAG_SHARED, this way requires that request pool of sched tags to be freed before calling blk_mq_free_tag_set(). Commit 47cdee29ef9d94e ("block: move blk_exit_queue into __blk_release_queue") moves blk_exit_queue into __blk_release_queue for simplying the fast path in generic_make_request(), then causes oops during freeing requests of sched tags in __blk_release_queue(). Fix the above issue by move freeing request pool of sched tags into blk_cleanup_queue(), this way is safe becasue queue has been frozen and no any in-queue requests at that time. Freeing sched tags has to be kept in queue's release handler becasue there might be un-completed dispatch activity which might refer to sched tags. Cc: Bart Van Assche <bvanassche@acm.org> Cc: Christoph Hellwig <hch@lst.de> Fixes: 47cdee29ef9d94e485eb08f962c74943023a5271 ("block: move blk_exit_queue into __blk_release_queue") Tested-by: Yi Zhang <yi.zhang@redhat.com> Reported-by: kernel test robot <rong.a.chen@intel.com> Signed-off-by: Ming Lei <ming.lei@redhat.com> Signed-off-by: Jens Axboe <axboe@kernel.dk>
-rw-r--r--block/blk-core.c13
-rw-r--r--block/blk-mq-sched.c30
-rw-r--r--block/blk-mq-sched.h1
-rw-r--r--block/blk-sysfs.c2
-rw-r--r--block/blk.h10
-rw-r--r--block/elevator.c2
6 files changed, 52 insertions, 6 deletions
diff --git a/block/blk-core.c b/block/blk-core.c
index ee1b35fe8572..8340f69670d8 100644
--- a/block/blk-core.c
+++ b/block/blk-core.c
@@ -320,6 +320,19 @@ void blk_cleanup_queue(struct request_queue *q)
320 if (queue_is_mq(q)) 320 if (queue_is_mq(q))
321 blk_mq_exit_queue(q); 321 blk_mq_exit_queue(q);
322 322
323 /*
324 * In theory, request pool of sched_tags belongs to request queue.
325 * However, the current implementation requires tag_set for freeing
326 * requests, so free the pool now.
327 *
328 * Queue has become frozen, there can't be any in-queue requests, so
329 * it is safe to free requests now.
330 */
331 mutex_lock(&q->sysfs_lock);
332 if (q->elevator)
333 blk_mq_sched_free_requests(q);
334 mutex_unlock(&q->sysfs_lock);
335
323 percpu_ref_exit(&q->q_usage_counter); 336 percpu_ref_exit(&q->q_usage_counter);
324 337
325 /* @q is and will stay empty, shutdown and put */ 338 /* @q is and will stay empty, shutdown and put */
diff --git a/block/blk-mq-sched.c b/block/blk-mq-sched.c
index 74c6bb871f7e..500cb04901cc 100644
--- a/block/blk-mq-sched.c
+++ b/block/blk-mq-sched.c
@@ -475,14 +475,18 @@ static int blk_mq_sched_alloc_tags(struct request_queue *q,
475 return ret; 475 return ret;
476} 476}
477 477
478/* called in queue's release handler, tagset has gone away */
478static void blk_mq_sched_tags_teardown(struct request_queue *q) 479static void blk_mq_sched_tags_teardown(struct request_queue *q)
479{ 480{
480 struct blk_mq_tag_set *set = q->tag_set;
481 struct blk_mq_hw_ctx *hctx; 481 struct blk_mq_hw_ctx *hctx;
482 int i; 482 int i;
483 483
484 queue_for_each_hw_ctx(q, hctx, i) 484 queue_for_each_hw_ctx(q, hctx, i) {
485 blk_mq_sched_free_tags(set, hctx, i); 485 if (hctx->sched_tags) {
486 blk_mq_free_rq_map(hctx->sched_tags);
487 hctx->sched_tags = NULL;
488 }
489 }
486} 490}
487 491
488int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e) 492int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e)
@@ -523,6 +527,7 @@ int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e)
523 ret = e->ops.init_hctx(hctx, i); 527 ret = e->ops.init_hctx(hctx, i);
524 if (ret) { 528 if (ret) {
525 eq = q->elevator; 529 eq = q->elevator;
530 blk_mq_sched_free_requests(q);
526 blk_mq_exit_sched(q, eq); 531 blk_mq_exit_sched(q, eq);
527 kobject_put(&eq->kobj); 532 kobject_put(&eq->kobj);
528 return ret; 533 return ret;
@@ -534,11 +539,30 @@ int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e)
534 return 0; 539 return 0;
535 540
536err: 541err:
542 blk_mq_sched_free_requests(q);
537 blk_mq_sched_tags_teardown(q); 543 blk_mq_sched_tags_teardown(q);
538 q->elevator = NULL; 544 q->elevator = NULL;
539 return ret; 545 return ret;
540} 546}
541 547
548/*
549 * called in either blk_queue_cleanup or elevator_switch, tagset
550 * is required for freeing requests
551 */
552void blk_mq_sched_free_requests(struct request_queue *q)
553{
554 struct blk_mq_hw_ctx *hctx;
555 int i;
556
557 lockdep_assert_held(&q->sysfs_lock);
558 WARN_ON(!q->elevator);
559
560 queue_for_each_hw_ctx(q, hctx, i) {
561 if (hctx->sched_tags)
562 blk_mq_free_rqs(q->tag_set, hctx->sched_tags, i);
563 }
564}
565
542void blk_mq_exit_sched(struct request_queue *q, struct elevator_queue *e) 566void blk_mq_exit_sched(struct request_queue *q, struct elevator_queue *e)
543{ 567{
544 struct blk_mq_hw_ctx *hctx; 568 struct blk_mq_hw_ctx *hctx;
diff --git a/block/blk-mq-sched.h b/block/blk-mq-sched.h
index c7bdb52367ac..3cf92cbbd8ac 100644
--- a/block/blk-mq-sched.h
+++ b/block/blk-mq-sched.h
@@ -28,6 +28,7 @@ void blk_mq_sched_dispatch_requests(struct blk_mq_hw_ctx *hctx);
28 28
29int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e); 29int blk_mq_init_sched(struct request_queue *q, struct elevator_type *e);
30void blk_mq_exit_sched(struct request_queue *q, struct elevator_queue *e); 30void blk_mq_exit_sched(struct request_queue *q, struct elevator_queue *e);
31void blk_mq_sched_free_requests(struct request_queue *q);
31 32
32static inline bool 33static inline bool
33blk_mq_sched_bio_merge(struct request_queue *q, struct bio *bio) 34blk_mq_sched_bio_merge(struct request_queue *q, struct bio *bio)
diff --git a/block/blk-sysfs.c b/block/blk-sysfs.c
index 75b5281cc577..977c659dcd18 100644
--- a/block/blk-sysfs.c
+++ b/block/blk-sysfs.c
@@ -850,7 +850,7 @@ static void blk_exit_queue(struct request_queue *q)
850 */ 850 */
851 if (q->elevator) { 851 if (q->elevator) {
852 ioc_clear_queue(q); 852 ioc_clear_queue(q);
853 elevator_exit(q, q->elevator); 853 __elevator_exit(q, q->elevator);
854 q->elevator = NULL; 854 q->elevator = NULL;
855 } 855 }
856 856
diff --git a/block/blk.h b/block/blk.h
index 91b3581b7c7a..7814aa207153 100644
--- a/block/blk.h
+++ b/block/blk.h
@@ -6,6 +6,7 @@
6#include <linux/blk-mq.h> 6#include <linux/blk-mq.h>
7#include <xen/xen.h> 7#include <xen/xen.h>
8#include "blk-mq.h" 8#include "blk-mq.h"
9#include "blk-mq-sched.h"
9 10
10/* Max future timer expiry for timeouts */ 11/* Max future timer expiry for timeouts */
11#define BLK_MAX_TIMEOUT (5 * HZ) 12#define BLK_MAX_TIMEOUT (5 * HZ)
@@ -176,10 +177,17 @@ void blk_insert_flush(struct request *rq);
176int elevator_init_mq(struct request_queue *q); 177int elevator_init_mq(struct request_queue *q);
177int elevator_switch_mq(struct request_queue *q, 178int elevator_switch_mq(struct request_queue *q,
178 struct elevator_type *new_e); 179 struct elevator_type *new_e);
179void elevator_exit(struct request_queue *, struct elevator_queue *); 180void __elevator_exit(struct request_queue *, struct elevator_queue *);
180int elv_register_queue(struct request_queue *q); 181int elv_register_queue(struct request_queue *q);
181void elv_unregister_queue(struct request_queue *q); 182void elv_unregister_queue(struct request_queue *q);
182 183
184static inline void elevator_exit(struct request_queue *q,
185 struct elevator_queue *e)
186{
187 blk_mq_sched_free_requests(q);
188 __elevator_exit(q, e);
189}
190
183struct hd_struct *__disk_get_part(struct gendisk *disk, int partno); 191struct hd_struct *__disk_get_part(struct gendisk *disk, int partno);
184 192
185#ifdef CONFIG_FAIL_IO_TIMEOUT 193#ifdef CONFIG_FAIL_IO_TIMEOUT
diff --git a/block/elevator.c b/block/elevator.c
index ec55d5fc0b3e..2f17d66d0e61 100644
--- a/block/elevator.c
+++ b/block/elevator.c
@@ -178,7 +178,7 @@ static void elevator_release(struct kobject *kobj)
178 kfree(e); 178 kfree(e);
179} 179}
180 180
181void elevator_exit(struct request_queue *q, struct elevator_queue *e) 181void __elevator_exit(struct request_queue *q, struct elevator_queue *e)
182{ 182{
183 mutex_lock(&e->sysfs_lock); 183 mutex_lock(&e->sysfs_lock);
184 if (e->type->ops.exit_sched) 184 if (e->type->ops.exit_sched)