aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorChristoph Hellwig <hch@lst.de>2015-11-26 03:13:05 -0500
committerJens Axboe <axboe@fb.com>2015-12-01 12:53:59 -0500
commit6f3b0e8bcf3cbb87a7459b3ed018d31d918df3f8 (patch)
tree78520f9313db6743cb1ba6feb805120e80113a8d
parentd7cf931dd9f18ce8ee7a0a9b7813a19fb2c8f5e9 (diff)
blk-mq: add a flags parameter to blk_mq_alloc_request
We already have the reserved flag, and a nowait flag awkwardly encoded as a gfp_t. Add a real flags argument to make the scheme more extensible and allow for a nicer calling convention. Signed-off-by: Christoph Hellwig <hch@lst.de> Signed-off-by: Jens Axboe <axboe@fb.com>
-rw-r--r--block/blk-core.c11
-rw-r--r--block/blk-mq-tag.c11
-rw-r--r--block/blk-mq.c20
-rw-r--r--block/blk-mq.h11
-rw-r--r--drivers/block/mtip32xx/mtip32xx.c2
-rw-r--r--drivers/block/null_blk.c2
-rw-r--r--drivers/nvme/host/lightnvm.c2
-rw-r--r--drivers/nvme/host/pci.c11
-rw-r--r--fs/block_dev.c4
-rw-r--r--include/linux/blk-mq.h8
-rw-r--r--include/linux/blkdev.h2
11 files changed, 42 insertions, 42 deletions
diff --git a/block/blk-core.c b/block/blk-core.c
index c88a946eca49..5ec996036e16 100644
--- a/block/blk-core.c
+++ b/block/blk-core.c
@@ -630,7 +630,7 @@ struct request_queue *blk_alloc_queue(gfp_t gfp_mask)
630} 630}
631EXPORT_SYMBOL(blk_alloc_queue); 631EXPORT_SYMBOL(blk_alloc_queue);
632 632
633int blk_queue_enter(struct request_queue *q, gfp_t gfp) 633int blk_queue_enter(struct request_queue *q, bool nowait)
634{ 634{
635 while (true) { 635 while (true) {
636 int ret; 636 int ret;
@@ -638,7 +638,7 @@ int blk_queue_enter(struct request_queue *q, gfp_t gfp)
638 if (percpu_ref_tryget_live(&q->q_usage_counter)) 638 if (percpu_ref_tryget_live(&q->q_usage_counter))
639 return 0; 639 return 0;
640 640
641 if (!gfpflags_allow_blocking(gfp)) 641 if (nowait)
642 return -EBUSY; 642 return -EBUSY;
643 643
644 ret = wait_event_interruptible(q->mq_freeze_wq, 644 ret = wait_event_interruptible(q->mq_freeze_wq,
@@ -1276,7 +1276,9 @@ static struct request *blk_old_get_request(struct request_queue *q, int rw,
1276struct request *blk_get_request(struct request_queue *q, int rw, gfp_t gfp_mask) 1276struct request *blk_get_request(struct request_queue *q, int rw, gfp_t gfp_mask)
1277{ 1277{
1278 if (q->mq_ops) 1278 if (q->mq_ops)
1279 return blk_mq_alloc_request(q, rw, gfp_mask, false); 1279 return blk_mq_alloc_request(q, rw,
1280 (gfp_mask & __GFP_DIRECT_RECLAIM) ?
1281 0 : BLK_MQ_REQ_NOWAIT);
1280 else 1282 else
1281 return blk_old_get_request(q, rw, gfp_mask); 1283 return blk_old_get_request(q, rw, gfp_mask);
1282} 1284}
@@ -2044,8 +2046,7 @@ blk_qc_t generic_make_request(struct bio *bio)
2044 do { 2046 do {
2045 struct request_queue *q = bdev_get_queue(bio->bi_bdev); 2047 struct request_queue *q = bdev_get_queue(bio->bi_bdev);
2046 2048
2047 if (likely(blk_queue_enter(q, __GFP_DIRECT_RECLAIM) == 0)) { 2049 if (likely(blk_queue_enter(q, false) == 0)) {
2048
2049 ret = q->make_request_fn(q, bio); 2050 ret = q->make_request_fn(q, bio);
2050 2051
2051 blk_queue_exit(q); 2052 blk_queue_exit(q);
diff --git a/block/blk-mq-tag.c b/block/blk-mq-tag.c
index a07ca3488d96..abdbb47405cb 100644
--- a/block/blk-mq-tag.c
+++ b/block/blk-mq-tag.c
@@ -268,7 +268,7 @@ static int bt_get(struct blk_mq_alloc_data *data,
268 if (tag != -1) 268 if (tag != -1)
269 return tag; 269 return tag;
270 270
271 if (!gfpflags_allow_blocking(data->gfp)) 271 if (data->flags & BLK_MQ_REQ_NOWAIT)
272 return -1; 272 return -1;
273 273
274 bs = bt_wait_ptr(bt, hctx); 274 bs = bt_wait_ptr(bt, hctx);
@@ -303,7 +303,7 @@ static int bt_get(struct blk_mq_alloc_data *data,
303 data->ctx = blk_mq_get_ctx(data->q); 303 data->ctx = blk_mq_get_ctx(data->q);
304 data->hctx = data->q->mq_ops->map_queue(data->q, 304 data->hctx = data->q->mq_ops->map_queue(data->q,
305 data->ctx->cpu); 305 data->ctx->cpu);
306 if (data->reserved) { 306 if (data->flags & BLK_MQ_REQ_RESERVED) {
307 bt = &data->hctx->tags->breserved_tags; 307 bt = &data->hctx->tags->breserved_tags;
308 } else { 308 } else {
309 last_tag = &data->ctx->last_tag; 309 last_tag = &data->ctx->last_tag;
@@ -349,10 +349,9 @@ static unsigned int __blk_mq_get_reserved_tag(struct blk_mq_alloc_data *data)
349 349
350unsigned int blk_mq_get_tag(struct blk_mq_alloc_data *data) 350unsigned int blk_mq_get_tag(struct blk_mq_alloc_data *data)
351{ 351{
352 if (!data->reserved) 352 if (data->flags & BLK_MQ_REQ_RESERVED)
353 return __blk_mq_get_tag(data); 353 return __blk_mq_get_reserved_tag(data);
354 354 return __blk_mq_get_tag(data);
355 return __blk_mq_get_reserved_tag(data);
356} 355}
357 356
358static struct bt_wait_state *bt_wake_ptr(struct blk_mq_bitmap_tags *bt) 357static struct bt_wait_state *bt_wake_ptr(struct blk_mq_bitmap_tags *bt)
diff --git a/block/blk-mq.c b/block/blk-mq.c
index 6d6f8feb48c0..93a4e1956915 100644
--- a/block/blk-mq.c
+++ b/block/blk-mq.c
@@ -229,8 +229,8 @@ __blk_mq_alloc_request(struct blk_mq_alloc_data *data, int rw)
229 return NULL; 229 return NULL;
230} 230}
231 231
232struct request *blk_mq_alloc_request(struct request_queue *q, int rw, gfp_t gfp, 232struct request *blk_mq_alloc_request(struct request_queue *q, int rw,
233 bool reserved) 233 unsigned int flags)
234{ 234{
235 struct blk_mq_ctx *ctx; 235 struct blk_mq_ctx *ctx;
236 struct blk_mq_hw_ctx *hctx; 236 struct blk_mq_hw_ctx *hctx;
@@ -238,24 +238,22 @@ struct request *blk_mq_alloc_request(struct request_queue *q, int rw, gfp_t gfp,
238 struct blk_mq_alloc_data alloc_data; 238 struct blk_mq_alloc_data alloc_data;
239 int ret; 239 int ret;
240 240
241 ret = blk_queue_enter(q, gfp); 241 ret = blk_queue_enter(q, flags & BLK_MQ_REQ_NOWAIT);
242 if (ret) 242 if (ret)
243 return ERR_PTR(ret); 243 return ERR_PTR(ret);
244 244
245 ctx = blk_mq_get_ctx(q); 245 ctx = blk_mq_get_ctx(q);
246 hctx = q->mq_ops->map_queue(q, ctx->cpu); 246 hctx = q->mq_ops->map_queue(q, ctx->cpu);
247 blk_mq_set_alloc_data(&alloc_data, q, gfp & ~__GFP_DIRECT_RECLAIM, 247 blk_mq_set_alloc_data(&alloc_data, q, flags, ctx, hctx);
248 reserved, ctx, hctx);
249 248
250 rq = __blk_mq_alloc_request(&alloc_data, rw); 249 rq = __blk_mq_alloc_request(&alloc_data, rw);
251 if (!rq && (gfp & __GFP_DIRECT_RECLAIM)) { 250 if (!rq && !(flags & BLK_MQ_REQ_NOWAIT)) {
252 __blk_mq_run_hw_queue(hctx); 251 __blk_mq_run_hw_queue(hctx);
253 blk_mq_put_ctx(ctx); 252 blk_mq_put_ctx(ctx);
254 253
255 ctx = blk_mq_get_ctx(q); 254 ctx = blk_mq_get_ctx(q);
256 hctx = q->mq_ops->map_queue(q, ctx->cpu); 255 hctx = q->mq_ops->map_queue(q, ctx->cpu);
257 blk_mq_set_alloc_data(&alloc_data, q, gfp, reserved, ctx, 256 blk_mq_set_alloc_data(&alloc_data, q, flags, ctx, hctx);
258 hctx);
259 rq = __blk_mq_alloc_request(&alloc_data, rw); 257 rq = __blk_mq_alloc_request(&alloc_data, rw);
260 ctx = alloc_data.ctx; 258 ctx = alloc_data.ctx;
261 } 259 }
@@ -1175,8 +1173,7 @@ static struct request *blk_mq_map_request(struct request_queue *q,
1175 rw |= REQ_SYNC; 1173 rw |= REQ_SYNC;
1176 1174
1177 trace_block_getrq(q, bio, rw); 1175 trace_block_getrq(q, bio, rw);
1178 blk_mq_set_alloc_data(&alloc_data, q, GFP_ATOMIC, false, ctx, 1176 blk_mq_set_alloc_data(&alloc_data, q, BLK_MQ_REQ_NOWAIT, ctx, hctx);
1179 hctx);
1180 rq = __blk_mq_alloc_request(&alloc_data, rw); 1177 rq = __blk_mq_alloc_request(&alloc_data, rw);
1181 if (unlikely(!rq)) { 1178 if (unlikely(!rq)) {
1182 __blk_mq_run_hw_queue(hctx); 1179 __blk_mq_run_hw_queue(hctx);
@@ -1185,8 +1182,7 @@ static struct request *blk_mq_map_request(struct request_queue *q,
1185 1182
1186 ctx = blk_mq_get_ctx(q); 1183 ctx = blk_mq_get_ctx(q);
1187 hctx = q->mq_ops->map_queue(q, ctx->cpu); 1184 hctx = q->mq_ops->map_queue(q, ctx->cpu);
1188 blk_mq_set_alloc_data(&alloc_data, q, 1185 blk_mq_set_alloc_data(&alloc_data, q, 0, ctx, hctx);
1189 __GFP_RECLAIM|__GFP_HIGH, false, ctx, hctx);
1190 rq = __blk_mq_alloc_request(&alloc_data, rw); 1186 rq = __blk_mq_alloc_request(&alloc_data, rw);
1191 ctx = alloc_data.ctx; 1187 ctx = alloc_data.ctx;
1192 hctx = alloc_data.hctx; 1188 hctx = alloc_data.hctx;
diff --git a/block/blk-mq.h b/block/blk-mq.h
index 713820b47b31..eaede8e45c9c 100644
--- a/block/blk-mq.h
+++ b/block/blk-mq.h
@@ -96,8 +96,7 @@ static inline void blk_mq_put_ctx(struct blk_mq_ctx *ctx)
96struct blk_mq_alloc_data { 96struct blk_mq_alloc_data {
97 /* input parameter */ 97 /* input parameter */
98 struct request_queue *q; 98 struct request_queue *q;
99 gfp_t gfp; 99 unsigned int flags;
100 bool reserved;
101 100
102 /* input & output parameter */ 101 /* input & output parameter */
103 struct blk_mq_ctx *ctx; 102 struct blk_mq_ctx *ctx;
@@ -105,13 +104,11 @@ struct blk_mq_alloc_data {
105}; 104};
106 105
107static inline void blk_mq_set_alloc_data(struct blk_mq_alloc_data *data, 106static inline void blk_mq_set_alloc_data(struct blk_mq_alloc_data *data,
108 struct request_queue *q, gfp_t gfp, bool reserved, 107 struct request_queue *q, unsigned int flags,
109 struct blk_mq_ctx *ctx, 108 struct blk_mq_ctx *ctx, struct blk_mq_hw_ctx *hctx)
110 struct blk_mq_hw_ctx *hctx)
111{ 109{
112 data->q = q; 110 data->q = q;
113 data->gfp = gfp; 111 data->flags = flags;
114 data->reserved = reserved;
115 data->ctx = ctx; 112 data->ctx = ctx;
116 data->hctx = hctx; 113 data->hctx = hctx;
117} 114}
diff --git a/drivers/block/mtip32xx/mtip32xx.c b/drivers/block/mtip32xx/mtip32xx.c
index 3457ac8c03e2..10bd8d0a9d9c 100644
--- a/drivers/block/mtip32xx/mtip32xx.c
+++ b/drivers/block/mtip32xx/mtip32xx.c
@@ -173,7 +173,7 @@ static struct mtip_cmd *mtip_get_int_command(struct driver_data *dd)
173{ 173{
174 struct request *rq; 174 struct request *rq;
175 175
176 rq = blk_mq_alloc_request(dd->queue, 0, __GFP_RECLAIM, true); 176 rq = blk_mq_alloc_request(dd->queue, 0, BLK_MQ_REQ_RESERVED);
177 return blk_mq_rq_to_pdu(rq); 177 return blk_mq_rq_to_pdu(rq);
178} 178}
179 179
diff --git a/drivers/block/null_blk.c b/drivers/block/null_blk.c
index 5c8ba5484d86..fa742dddf3f8 100644
--- a/drivers/block/null_blk.c
+++ b/drivers/block/null_blk.c
@@ -464,7 +464,7 @@ static int null_lnvm_submit_io(struct request_queue *q, struct nvm_rq *rqd)
464 struct request *rq; 464 struct request *rq;
465 struct bio *bio = rqd->bio; 465 struct bio *bio = rqd->bio;
466 466
467 rq = blk_mq_alloc_request(q, bio_rw(bio), GFP_KERNEL, 0); 467 rq = blk_mq_alloc_request(q, bio_rw(bio), 0);
468 if (IS_ERR(rq)) 468 if (IS_ERR(rq))
469 return -ENOMEM; 469 return -ENOMEM;
470 470
diff --git a/drivers/nvme/host/lightnvm.c b/drivers/nvme/host/lightnvm.c
index 9202d1a468d0..d5622f9164ad 100644
--- a/drivers/nvme/host/lightnvm.c
+++ b/drivers/nvme/host/lightnvm.c
@@ -470,7 +470,7 @@ static int nvme_nvm_submit_io(struct request_queue *q, struct nvm_rq *rqd)
470 struct bio *bio = rqd->bio; 470 struct bio *bio = rqd->bio;
471 struct nvme_nvm_command *cmd; 471 struct nvme_nvm_command *cmd;
472 472
473 rq = blk_mq_alloc_request(q, bio_rw(bio), GFP_KERNEL, 0); 473 rq = blk_mq_alloc_request(q, bio_rw(bio), 0);
474 if (IS_ERR(rq)) 474 if (IS_ERR(rq))
475 return -ENOMEM; 475 return -ENOMEM;
476 476
diff --git a/drivers/nvme/host/pci.c b/drivers/nvme/host/pci.c
index f3b53af789ef..b8a02221233c 100644
--- a/drivers/nvme/host/pci.c
+++ b/drivers/nvme/host/pci.c
@@ -1041,7 +1041,7 @@ int __nvme_submit_sync_cmd(struct request_queue *q, struct nvme_command *cmd,
1041 struct request *req; 1041 struct request *req;
1042 int ret; 1042 int ret;
1043 1043
1044 req = blk_mq_alloc_request(q, write, GFP_KERNEL, false); 1044 req = blk_mq_alloc_request(q, write, 0);
1045 if (IS_ERR(req)) 1045 if (IS_ERR(req))
1046 return PTR_ERR(req); 1046 return PTR_ERR(req);
1047 1047
@@ -1094,7 +1094,8 @@ static int nvme_submit_async_admin_req(struct nvme_dev *dev)
1094 struct nvme_cmd_info *cmd_info; 1094 struct nvme_cmd_info *cmd_info;
1095 struct request *req; 1095 struct request *req;
1096 1096
1097 req = blk_mq_alloc_request(dev->admin_q, WRITE, GFP_ATOMIC, true); 1097 req = blk_mq_alloc_request(dev->admin_q, WRITE,
1098 BLK_MQ_REQ_NOWAIT | BLK_MQ_REQ_RESERVED);
1098 if (IS_ERR(req)) 1099 if (IS_ERR(req))
1099 return PTR_ERR(req); 1100 return PTR_ERR(req);
1100 1101
@@ -1119,7 +1120,7 @@ static int nvme_submit_admin_async_cmd(struct nvme_dev *dev,
1119 struct request *req; 1120 struct request *req;
1120 struct nvme_cmd_info *cmd_rq; 1121 struct nvme_cmd_info *cmd_rq;
1121 1122
1122 req = blk_mq_alloc_request(dev->admin_q, WRITE, GFP_KERNEL, false); 1123 req = blk_mq_alloc_request(dev->admin_q, WRITE, 0);
1123 if (IS_ERR(req)) 1124 if (IS_ERR(req))
1124 return PTR_ERR(req); 1125 return PTR_ERR(req);
1125 1126
@@ -1320,8 +1321,8 @@ static void nvme_abort_req(struct request *req)
1320 if (!dev->abort_limit) 1321 if (!dev->abort_limit)
1321 return; 1322 return;
1322 1323
1323 abort_req = blk_mq_alloc_request(dev->admin_q, WRITE, GFP_ATOMIC, 1324 abort_req = blk_mq_alloc_request(dev->admin_q, WRITE,
1324 false); 1325 BLK_MQ_REQ_NOWAIT);
1325 if (IS_ERR(abort_req)) 1326 if (IS_ERR(abort_req))
1326 return; 1327 return;
1327 1328
diff --git a/fs/block_dev.c b/fs/block_dev.c
index c25639e907bd..aa1a45985889 100644
--- a/fs/block_dev.c
+++ b/fs/block_dev.c
@@ -395,7 +395,7 @@ int bdev_read_page(struct block_device *bdev, sector_t sector,
395 if (!ops->rw_page || bdev_get_integrity(bdev)) 395 if (!ops->rw_page || bdev_get_integrity(bdev))
396 return result; 396 return result;
397 397
398 result = blk_queue_enter(bdev->bd_queue, GFP_KERNEL); 398 result = blk_queue_enter(bdev->bd_queue, false);
399 if (result) 399 if (result)
400 return result; 400 return result;
401 result = ops->rw_page(bdev, sector + get_start_sect(bdev), page, READ); 401 result = ops->rw_page(bdev, sector + get_start_sect(bdev), page, READ);
@@ -432,7 +432,7 @@ int bdev_write_page(struct block_device *bdev, sector_t sector,
432 432
433 if (!ops->rw_page || bdev_get_integrity(bdev)) 433 if (!ops->rw_page || bdev_get_integrity(bdev))
434 return -EOPNOTSUPP; 434 return -EOPNOTSUPP;
435 result = blk_queue_enter(bdev->bd_queue, GFP_KERNEL); 435 result = blk_queue_enter(bdev->bd_queue, false);
436 if (result) 436 if (result)
437 return result; 437 return result;
438 438
diff --git a/include/linux/blk-mq.h b/include/linux/blk-mq.h
index daf17d70aeca..7fc9296b5742 100644
--- a/include/linux/blk-mq.h
+++ b/include/linux/blk-mq.h
@@ -188,8 +188,14 @@ void blk_mq_insert_request(struct request *, bool, bool, bool);
188void blk_mq_free_request(struct request *rq); 188void blk_mq_free_request(struct request *rq);
189void blk_mq_free_hctx_request(struct blk_mq_hw_ctx *, struct request *rq); 189void blk_mq_free_hctx_request(struct blk_mq_hw_ctx *, struct request *rq);
190bool blk_mq_can_queue(struct blk_mq_hw_ctx *); 190bool blk_mq_can_queue(struct blk_mq_hw_ctx *);
191
192enum {
193 BLK_MQ_REQ_NOWAIT = (1 << 0), /* return when out of requests */
194 BLK_MQ_REQ_RESERVED = (1 << 1), /* allocate from reserved pool */
195};
196
191struct request *blk_mq_alloc_request(struct request_queue *q, int rw, 197struct request *blk_mq_alloc_request(struct request_queue *q, int rw,
192 gfp_t gfp, bool reserved); 198 unsigned int flags);
193struct request *blk_mq_tag_to_rq(struct blk_mq_tags *tags, unsigned int tag); 199struct request *blk_mq_tag_to_rq(struct blk_mq_tags *tags, unsigned int tag);
194struct cpumask *blk_mq_tags_cpumask(struct blk_mq_tags *tags); 200struct cpumask *blk_mq_tags_cpumask(struct blk_mq_tags *tags);
195 201
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h
index c0d2b7927c1f..e711f294934c 100644
--- a/include/linux/blkdev.h
+++ b/include/linux/blkdev.h
@@ -794,7 +794,7 @@ extern int scsi_cmd_ioctl(struct request_queue *, struct gendisk *, fmode_t,
794extern int sg_scsi_ioctl(struct request_queue *, struct gendisk *, fmode_t, 794extern int sg_scsi_ioctl(struct request_queue *, struct gendisk *, fmode_t,
795 struct scsi_ioctl_command __user *); 795 struct scsi_ioctl_command __user *);
796 796
797extern int blk_queue_enter(struct request_queue *q, gfp_t gfp); 797extern int blk_queue_enter(struct request_queue *q, bool nowait);
798extern void blk_queue_exit(struct request_queue *q); 798extern void blk_queue_exit(struct request_queue *q);
799extern void blk_start_queue(struct request_queue *q); 799extern void blk_start_queue(struct request_queue *q);
800extern void blk_stop_queue(struct request_queue *q); 800extern void blk_stop_queue(struct request_queue *q);