diff options
author | Christoph Hellwig <hch@lst.de> | 2015-11-26 03:13:05 -0500 |
---|---|---|
committer | Jens Axboe <axboe@fb.com> | 2015-12-01 12:53:59 -0500 |
commit | 6f3b0e8bcf3cbb87a7459b3ed018d31d918df3f8 (patch) | |
tree | 78520f9313db6743cb1ba6feb805120e80113a8d | |
parent | d7cf931dd9f18ce8ee7a0a9b7813a19fb2c8f5e9 (diff) |
blk-mq: add a flags parameter to blk_mq_alloc_request
We already have the reserved flag, and a nowait flag awkwardly encoded as
a gfp_t. Add a real flags argument to make the scheme more extensible and
allow for a nicer calling convention.
Signed-off-by: Christoph Hellwig <hch@lst.de>
Signed-off-by: Jens Axboe <axboe@fb.com>
-rw-r--r-- | block/blk-core.c | 11 | ||||
-rw-r--r-- | block/blk-mq-tag.c | 11 | ||||
-rw-r--r-- | block/blk-mq.c | 20 | ||||
-rw-r--r-- | block/blk-mq.h | 11 | ||||
-rw-r--r-- | drivers/block/mtip32xx/mtip32xx.c | 2 | ||||
-rw-r--r-- | drivers/block/null_blk.c | 2 | ||||
-rw-r--r-- | drivers/nvme/host/lightnvm.c | 2 | ||||
-rw-r--r-- | drivers/nvme/host/pci.c | 11 | ||||
-rw-r--r-- | fs/block_dev.c | 4 | ||||
-rw-r--r-- | include/linux/blk-mq.h | 8 | ||||
-rw-r--r-- | include/linux/blkdev.h | 2 |
11 files changed, 42 insertions, 42 deletions
diff --git a/block/blk-core.c b/block/blk-core.c index c88a946eca49..5ec996036e16 100644 --- a/block/blk-core.c +++ b/block/blk-core.c | |||
@@ -630,7 +630,7 @@ struct request_queue *blk_alloc_queue(gfp_t gfp_mask) | |||
630 | } | 630 | } |
631 | EXPORT_SYMBOL(blk_alloc_queue); | 631 | EXPORT_SYMBOL(blk_alloc_queue); |
632 | 632 | ||
633 | int blk_queue_enter(struct request_queue *q, gfp_t gfp) | 633 | int blk_queue_enter(struct request_queue *q, bool nowait) |
634 | { | 634 | { |
635 | while (true) { | 635 | while (true) { |
636 | int ret; | 636 | int ret; |
@@ -638,7 +638,7 @@ int blk_queue_enter(struct request_queue *q, gfp_t gfp) | |||
638 | if (percpu_ref_tryget_live(&q->q_usage_counter)) | 638 | if (percpu_ref_tryget_live(&q->q_usage_counter)) |
639 | return 0; | 639 | return 0; |
640 | 640 | ||
641 | if (!gfpflags_allow_blocking(gfp)) | 641 | if (nowait) |
642 | return -EBUSY; | 642 | return -EBUSY; |
643 | 643 | ||
644 | ret = wait_event_interruptible(q->mq_freeze_wq, | 644 | ret = wait_event_interruptible(q->mq_freeze_wq, |
@@ -1276,7 +1276,9 @@ static struct request *blk_old_get_request(struct request_queue *q, int rw, | |||
1276 | struct request *blk_get_request(struct request_queue *q, int rw, gfp_t gfp_mask) | 1276 | struct request *blk_get_request(struct request_queue *q, int rw, gfp_t gfp_mask) |
1277 | { | 1277 | { |
1278 | if (q->mq_ops) | 1278 | if (q->mq_ops) |
1279 | return blk_mq_alloc_request(q, rw, gfp_mask, false); | 1279 | return blk_mq_alloc_request(q, rw, |
1280 | (gfp_mask & __GFP_DIRECT_RECLAIM) ? | ||
1281 | 0 : BLK_MQ_REQ_NOWAIT); | ||
1280 | else | 1282 | else |
1281 | return blk_old_get_request(q, rw, gfp_mask); | 1283 | return blk_old_get_request(q, rw, gfp_mask); |
1282 | } | 1284 | } |
@@ -2044,8 +2046,7 @@ blk_qc_t generic_make_request(struct bio *bio) | |||
2044 | do { | 2046 | do { |
2045 | struct request_queue *q = bdev_get_queue(bio->bi_bdev); | 2047 | struct request_queue *q = bdev_get_queue(bio->bi_bdev); |
2046 | 2048 | ||
2047 | if (likely(blk_queue_enter(q, __GFP_DIRECT_RECLAIM) == 0)) { | 2049 | if (likely(blk_queue_enter(q, false) == 0)) { |
2048 | |||
2049 | ret = q->make_request_fn(q, bio); | 2050 | ret = q->make_request_fn(q, bio); |
2050 | 2051 | ||
2051 | blk_queue_exit(q); | 2052 | blk_queue_exit(q); |
diff --git a/block/blk-mq-tag.c b/block/blk-mq-tag.c index a07ca3488d96..abdbb47405cb 100644 --- a/block/blk-mq-tag.c +++ b/block/blk-mq-tag.c | |||
@@ -268,7 +268,7 @@ static int bt_get(struct blk_mq_alloc_data *data, | |||
268 | if (tag != -1) | 268 | if (tag != -1) |
269 | return tag; | 269 | return tag; |
270 | 270 | ||
271 | if (!gfpflags_allow_blocking(data->gfp)) | 271 | if (data->flags & BLK_MQ_REQ_NOWAIT) |
272 | return -1; | 272 | return -1; |
273 | 273 | ||
274 | bs = bt_wait_ptr(bt, hctx); | 274 | bs = bt_wait_ptr(bt, hctx); |
@@ -303,7 +303,7 @@ static int bt_get(struct blk_mq_alloc_data *data, | |||
303 | data->ctx = blk_mq_get_ctx(data->q); | 303 | data->ctx = blk_mq_get_ctx(data->q); |
304 | data->hctx = data->q->mq_ops->map_queue(data->q, | 304 | data->hctx = data->q->mq_ops->map_queue(data->q, |
305 | data->ctx->cpu); | 305 | data->ctx->cpu); |
306 | if (data->reserved) { | 306 | if (data->flags & BLK_MQ_REQ_RESERVED) { |
307 | bt = &data->hctx->tags->breserved_tags; | 307 | bt = &data->hctx->tags->breserved_tags; |
308 | } else { | 308 | } else { |
309 | last_tag = &data->ctx->last_tag; | 309 | last_tag = &data->ctx->last_tag; |
@@ -349,10 +349,9 @@ static unsigned int __blk_mq_get_reserved_tag(struct blk_mq_alloc_data *data) | |||
349 | 349 | ||
350 | unsigned int blk_mq_get_tag(struct blk_mq_alloc_data *data) | 350 | unsigned int blk_mq_get_tag(struct blk_mq_alloc_data *data) |
351 | { | 351 | { |
352 | if (!data->reserved) | 352 | if (data->flags & BLK_MQ_REQ_RESERVED) |
353 | return __blk_mq_get_tag(data); | 353 | return __blk_mq_get_reserved_tag(data); |
354 | 354 | return __blk_mq_get_tag(data); | |
355 | return __blk_mq_get_reserved_tag(data); | ||
356 | } | 355 | } |
357 | 356 | ||
358 | static struct bt_wait_state *bt_wake_ptr(struct blk_mq_bitmap_tags *bt) | 357 | static struct bt_wait_state *bt_wake_ptr(struct blk_mq_bitmap_tags *bt) |
diff --git a/block/blk-mq.c b/block/blk-mq.c index 6d6f8feb48c0..93a4e1956915 100644 --- a/block/blk-mq.c +++ b/block/blk-mq.c | |||
@@ -229,8 +229,8 @@ __blk_mq_alloc_request(struct blk_mq_alloc_data *data, int rw) | |||
229 | return NULL; | 229 | return NULL; |
230 | } | 230 | } |
231 | 231 | ||
232 | struct request *blk_mq_alloc_request(struct request_queue *q, int rw, gfp_t gfp, | 232 | struct request *blk_mq_alloc_request(struct request_queue *q, int rw, |
233 | bool reserved) | 233 | unsigned int flags) |
234 | { | 234 | { |
235 | struct blk_mq_ctx *ctx; | 235 | struct blk_mq_ctx *ctx; |
236 | struct blk_mq_hw_ctx *hctx; | 236 | struct blk_mq_hw_ctx *hctx; |
@@ -238,24 +238,22 @@ struct request *blk_mq_alloc_request(struct request_queue *q, int rw, gfp_t gfp, | |||
238 | struct blk_mq_alloc_data alloc_data; | 238 | struct blk_mq_alloc_data alloc_data; |
239 | int ret; | 239 | int ret; |
240 | 240 | ||
241 | ret = blk_queue_enter(q, gfp); | 241 | ret = blk_queue_enter(q, flags & BLK_MQ_REQ_NOWAIT); |
242 | if (ret) | 242 | if (ret) |
243 | return ERR_PTR(ret); | 243 | return ERR_PTR(ret); |
244 | 244 | ||
245 | ctx = blk_mq_get_ctx(q); | 245 | ctx = blk_mq_get_ctx(q); |
246 | hctx = q->mq_ops->map_queue(q, ctx->cpu); | 246 | hctx = q->mq_ops->map_queue(q, ctx->cpu); |
247 | blk_mq_set_alloc_data(&alloc_data, q, gfp & ~__GFP_DIRECT_RECLAIM, | 247 | blk_mq_set_alloc_data(&alloc_data, q, flags, ctx, hctx); |
248 | reserved, ctx, hctx); | ||
249 | 248 | ||
250 | rq = __blk_mq_alloc_request(&alloc_data, rw); | 249 | rq = __blk_mq_alloc_request(&alloc_data, rw); |
251 | if (!rq && (gfp & __GFP_DIRECT_RECLAIM)) { | 250 | if (!rq && !(flags & BLK_MQ_REQ_NOWAIT)) { |
252 | __blk_mq_run_hw_queue(hctx); | 251 | __blk_mq_run_hw_queue(hctx); |
253 | blk_mq_put_ctx(ctx); | 252 | blk_mq_put_ctx(ctx); |
254 | 253 | ||
255 | ctx = blk_mq_get_ctx(q); | 254 | ctx = blk_mq_get_ctx(q); |
256 | hctx = q->mq_ops->map_queue(q, ctx->cpu); | 255 | hctx = q->mq_ops->map_queue(q, ctx->cpu); |
257 | blk_mq_set_alloc_data(&alloc_data, q, gfp, reserved, ctx, | 256 | blk_mq_set_alloc_data(&alloc_data, q, flags, ctx, hctx); |
258 | hctx); | ||
259 | rq = __blk_mq_alloc_request(&alloc_data, rw); | 257 | rq = __blk_mq_alloc_request(&alloc_data, rw); |
260 | ctx = alloc_data.ctx; | 258 | ctx = alloc_data.ctx; |
261 | } | 259 | } |
@@ -1175,8 +1173,7 @@ static struct request *blk_mq_map_request(struct request_queue *q, | |||
1175 | rw |= REQ_SYNC; | 1173 | rw |= REQ_SYNC; |
1176 | 1174 | ||
1177 | trace_block_getrq(q, bio, rw); | 1175 | trace_block_getrq(q, bio, rw); |
1178 | blk_mq_set_alloc_data(&alloc_data, q, GFP_ATOMIC, false, ctx, | 1176 | blk_mq_set_alloc_data(&alloc_data, q, BLK_MQ_REQ_NOWAIT, ctx, hctx); |
1179 | hctx); | ||
1180 | rq = __blk_mq_alloc_request(&alloc_data, rw); | 1177 | rq = __blk_mq_alloc_request(&alloc_data, rw); |
1181 | if (unlikely(!rq)) { | 1178 | if (unlikely(!rq)) { |
1182 | __blk_mq_run_hw_queue(hctx); | 1179 | __blk_mq_run_hw_queue(hctx); |
@@ -1185,8 +1182,7 @@ static struct request *blk_mq_map_request(struct request_queue *q, | |||
1185 | 1182 | ||
1186 | ctx = blk_mq_get_ctx(q); | 1183 | ctx = blk_mq_get_ctx(q); |
1187 | hctx = q->mq_ops->map_queue(q, ctx->cpu); | 1184 | hctx = q->mq_ops->map_queue(q, ctx->cpu); |
1188 | blk_mq_set_alloc_data(&alloc_data, q, | 1185 | blk_mq_set_alloc_data(&alloc_data, q, 0, ctx, hctx); |
1189 | __GFP_RECLAIM|__GFP_HIGH, false, ctx, hctx); | ||
1190 | rq = __blk_mq_alloc_request(&alloc_data, rw); | 1186 | rq = __blk_mq_alloc_request(&alloc_data, rw); |
1191 | ctx = alloc_data.ctx; | 1187 | ctx = alloc_data.ctx; |
1192 | hctx = alloc_data.hctx; | 1188 | hctx = alloc_data.hctx; |
diff --git a/block/blk-mq.h b/block/blk-mq.h index 713820b47b31..eaede8e45c9c 100644 --- a/block/blk-mq.h +++ b/block/blk-mq.h | |||
@@ -96,8 +96,7 @@ static inline void blk_mq_put_ctx(struct blk_mq_ctx *ctx) | |||
96 | struct blk_mq_alloc_data { | 96 | struct blk_mq_alloc_data { |
97 | /* input parameter */ | 97 | /* input parameter */ |
98 | struct request_queue *q; | 98 | struct request_queue *q; |
99 | gfp_t gfp; | 99 | unsigned int flags; |
100 | bool reserved; | ||
101 | 100 | ||
102 | /* input & output parameter */ | 101 | /* input & output parameter */ |
103 | struct blk_mq_ctx *ctx; | 102 | struct blk_mq_ctx *ctx; |
@@ -105,13 +104,11 @@ struct blk_mq_alloc_data { | |||
105 | }; | 104 | }; |
106 | 105 | ||
107 | static inline void blk_mq_set_alloc_data(struct blk_mq_alloc_data *data, | 106 | static inline void blk_mq_set_alloc_data(struct blk_mq_alloc_data *data, |
108 | struct request_queue *q, gfp_t gfp, bool reserved, | 107 | struct request_queue *q, unsigned int flags, |
109 | struct blk_mq_ctx *ctx, | 108 | struct blk_mq_ctx *ctx, struct blk_mq_hw_ctx *hctx) |
110 | struct blk_mq_hw_ctx *hctx) | ||
111 | { | 109 | { |
112 | data->q = q; | 110 | data->q = q; |
113 | data->gfp = gfp; | 111 | data->flags = flags; |
114 | data->reserved = reserved; | ||
115 | data->ctx = ctx; | 112 | data->ctx = ctx; |
116 | data->hctx = hctx; | 113 | data->hctx = hctx; |
117 | } | 114 | } |
diff --git a/drivers/block/mtip32xx/mtip32xx.c b/drivers/block/mtip32xx/mtip32xx.c index 3457ac8c03e2..10bd8d0a9d9c 100644 --- a/drivers/block/mtip32xx/mtip32xx.c +++ b/drivers/block/mtip32xx/mtip32xx.c | |||
@@ -173,7 +173,7 @@ static struct mtip_cmd *mtip_get_int_command(struct driver_data *dd) | |||
173 | { | 173 | { |
174 | struct request *rq; | 174 | struct request *rq; |
175 | 175 | ||
176 | rq = blk_mq_alloc_request(dd->queue, 0, __GFP_RECLAIM, true); | 176 | rq = blk_mq_alloc_request(dd->queue, 0, BLK_MQ_REQ_RESERVED); |
177 | return blk_mq_rq_to_pdu(rq); | 177 | return blk_mq_rq_to_pdu(rq); |
178 | } | 178 | } |
179 | 179 | ||
diff --git a/drivers/block/null_blk.c b/drivers/block/null_blk.c index 5c8ba5484d86..fa742dddf3f8 100644 --- a/drivers/block/null_blk.c +++ b/drivers/block/null_blk.c | |||
@@ -464,7 +464,7 @@ static int null_lnvm_submit_io(struct request_queue *q, struct nvm_rq *rqd) | |||
464 | struct request *rq; | 464 | struct request *rq; |
465 | struct bio *bio = rqd->bio; | 465 | struct bio *bio = rqd->bio; |
466 | 466 | ||
467 | rq = blk_mq_alloc_request(q, bio_rw(bio), GFP_KERNEL, 0); | 467 | rq = blk_mq_alloc_request(q, bio_rw(bio), 0); |
468 | if (IS_ERR(rq)) | 468 | if (IS_ERR(rq)) |
469 | return -ENOMEM; | 469 | return -ENOMEM; |
470 | 470 | ||
diff --git a/drivers/nvme/host/lightnvm.c b/drivers/nvme/host/lightnvm.c index 9202d1a468d0..d5622f9164ad 100644 --- a/drivers/nvme/host/lightnvm.c +++ b/drivers/nvme/host/lightnvm.c | |||
@@ -470,7 +470,7 @@ static int nvme_nvm_submit_io(struct request_queue *q, struct nvm_rq *rqd) | |||
470 | struct bio *bio = rqd->bio; | 470 | struct bio *bio = rqd->bio; |
471 | struct nvme_nvm_command *cmd; | 471 | struct nvme_nvm_command *cmd; |
472 | 472 | ||
473 | rq = blk_mq_alloc_request(q, bio_rw(bio), GFP_KERNEL, 0); | 473 | rq = blk_mq_alloc_request(q, bio_rw(bio), 0); |
474 | if (IS_ERR(rq)) | 474 | if (IS_ERR(rq)) |
475 | return -ENOMEM; | 475 | return -ENOMEM; |
476 | 476 | ||
diff --git a/drivers/nvme/host/pci.c b/drivers/nvme/host/pci.c index f3b53af789ef..b8a02221233c 100644 --- a/drivers/nvme/host/pci.c +++ b/drivers/nvme/host/pci.c | |||
@@ -1041,7 +1041,7 @@ int __nvme_submit_sync_cmd(struct request_queue *q, struct nvme_command *cmd, | |||
1041 | struct request *req; | 1041 | struct request *req; |
1042 | int ret; | 1042 | int ret; |
1043 | 1043 | ||
1044 | req = blk_mq_alloc_request(q, write, GFP_KERNEL, false); | 1044 | req = blk_mq_alloc_request(q, write, 0); |
1045 | if (IS_ERR(req)) | 1045 | if (IS_ERR(req)) |
1046 | return PTR_ERR(req); | 1046 | return PTR_ERR(req); |
1047 | 1047 | ||
@@ -1094,7 +1094,8 @@ static int nvme_submit_async_admin_req(struct nvme_dev *dev) | |||
1094 | struct nvme_cmd_info *cmd_info; | 1094 | struct nvme_cmd_info *cmd_info; |
1095 | struct request *req; | 1095 | struct request *req; |
1096 | 1096 | ||
1097 | req = blk_mq_alloc_request(dev->admin_q, WRITE, GFP_ATOMIC, true); | 1097 | req = blk_mq_alloc_request(dev->admin_q, WRITE, |
1098 | BLK_MQ_REQ_NOWAIT | BLK_MQ_REQ_RESERVED); | ||
1098 | if (IS_ERR(req)) | 1099 | if (IS_ERR(req)) |
1099 | return PTR_ERR(req); | 1100 | return PTR_ERR(req); |
1100 | 1101 | ||
@@ -1119,7 +1120,7 @@ static int nvme_submit_admin_async_cmd(struct nvme_dev *dev, | |||
1119 | struct request *req; | 1120 | struct request *req; |
1120 | struct nvme_cmd_info *cmd_rq; | 1121 | struct nvme_cmd_info *cmd_rq; |
1121 | 1122 | ||
1122 | req = blk_mq_alloc_request(dev->admin_q, WRITE, GFP_KERNEL, false); | 1123 | req = blk_mq_alloc_request(dev->admin_q, WRITE, 0); |
1123 | if (IS_ERR(req)) | 1124 | if (IS_ERR(req)) |
1124 | return PTR_ERR(req); | 1125 | return PTR_ERR(req); |
1125 | 1126 | ||
@@ -1320,8 +1321,8 @@ static void nvme_abort_req(struct request *req) | |||
1320 | if (!dev->abort_limit) | 1321 | if (!dev->abort_limit) |
1321 | return; | 1322 | return; |
1322 | 1323 | ||
1323 | abort_req = blk_mq_alloc_request(dev->admin_q, WRITE, GFP_ATOMIC, | 1324 | abort_req = blk_mq_alloc_request(dev->admin_q, WRITE, |
1324 | false); | 1325 | BLK_MQ_REQ_NOWAIT); |
1325 | if (IS_ERR(abort_req)) | 1326 | if (IS_ERR(abort_req)) |
1326 | return; | 1327 | return; |
1327 | 1328 | ||
diff --git a/fs/block_dev.c b/fs/block_dev.c index c25639e907bd..aa1a45985889 100644 --- a/fs/block_dev.c +++ b/fs/block_dev.c | |||
@@ -395,7 +395,7 @@ int bdev_read_page(struct block_device *bdev, sector_t sector, | |||
395 | if (!ops->rw_page || bdev_get_integrity(bdev)) | 395 | if (!ops->rw_page || bdev_get_integrity(bdev)) |
396 | return result; | 396 | return result; |
397 | 397 | ||
398 | result = blk_queue_enter(bdev->bd_queue, GFP_KERNEL); | 398 | result = blk_queue_enter(bdev->bd_queue, false); |
399 | if (result) | 399 | if (result) |
400 | return result; | 400 | return result; |
401 | result = ops->rw_page(bdev, sector + get_start_sect(bdev), page, READ); | 401 | result = ops->rw_page(bdev, sector + get_start_sect(bdev), page, READ); |
@@ -432,7 +432,7 @@ int bdev_write_page(struct block_device *bdev, sector_t sector, | |||
432 | 432 | ||
433 | if (!ops->rw_page || bdev_get_integrity(bdev)) | 433 | if (!ops->rw_page || bdev_get_integrity(bdev)) |
434 | return -EOPNOTSUPP; | 434 | return -EOPNOTSUPP; |
435 | result = blk_queue_enter(bdev->bd_queue, GFP_KERNEL); | 435 | result = blk_queue_enter(bdev->bd_queue, false); |
436 | if (result) | 436 | if (result) |
437 | return result; | 437 | return result; |
438 | 438 | ||
diff --git a/include/linux/blk-mq.h b/include/linux/blk-mq.h index daf17d70aeca..7fc9296b5742 100644 --- a/include/linux/blk-mq.h +++ b/include/linux/blk-mq.h | |||
@@ -188,8 +188,14 @@ void blk_mq_insert_request(struct request *, bool, bool, bool); | |||
188 | void blk_mq_free_request(struct request *rq); | 188 | void blk_mq_free_request(struct request *rq); |
189 | void blk_mq_free_hctx_request(struct blk_mq_hw_ctx *, struct request *rq); | 189 | void blk_mq_free_hctx_request(struct blk_mq_hw_ctx *, struct request *rq); |
190 | bool blk_mq_can_queue(struct blk_mq_hw_ctx *); | 190 | bool blk_mq_can_queue(struct blk_mq_hw_ctx *); |
191 | |||
192 | enum { | ||
193 | BLK_MQ_REQ_NOWAIT = (1 << 0), /* return when out of requests */ | ||
194 | BLK_MQ_REQ_RESERVED = (1 << 1), /* allocate from reserved pool */ | ||
195 | }; | ||
196 | |||
191 | struct request *blk_mq_alloc_request(struct request_queue *q, int rw, | 197 | struct request *blk_mq_alloc_request(struct request_queue *q, int rw, |
192 | gfp_t gfp, bool reserved); | 198 | unsigned int flags); |
193 | struct request *blk_mq_tag_to_rq(struct blk_mq_tags *tags, unsigned int tag); | 199 | struct request *blk_mq_tag_to_rq(struct blk_mq_tags *tags, unsigned int tag); |
194 | struct cpumask *blk_mq_tags_cpumask(struct blk_mq_tags *tags); | 200 | struct cpumask *blk_mq_tags_cpumask(struct blk_mq_tags *tags); |
195 | 201 | ||
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h index c0d2b7927c1f..e711f294934c 100644 --- a/include/linux/blkdev.h +++ b/include/linux/blkdev.h | |||
@@ -794,7 +794,7 @@ extern int scsi_cmd_ioctl(struct request_queue *, struct gendisk *, fmode_t, | |||
794 | extern int sg_scsi_ioctl(struct request_queue *, struct gendisk *, fmode_t, | 794 | extern int sg_scsi_ioctl(struct request_queue *, struct gendisk *, fmode_t, |
795 | struct scsi_ioctl_command __user *); | 795 | struct scsi_ioctl_command __user *); |
796 | 796 | ||
797 | extern int blk_queue_enter(struct request_queue *q, gfp_t gfp); | 797 | extern int blk_queue_enter(struct request_queue *q, bool nowait); |
798 | extern void blk_queue_exit(struct request_queue *q); | 798 | extern void blk_queue_exit(struct request_queue *q); |
799 | extern void blk_start_queue(struct request_queue *q); | 799 | extern void blk_start_queue(struct request_queue *q); |
800 | extern void blk_stop_queue(struct request_queue *q); | 800 | extern void blk_stop_queue(struct request_queue *q); |