diff options
Diffstat (limited to 'block')
-rw-r--r-- | block/bsg.c | 84 |
1 files changed, 23 insertions, 61 deletions
diff --git a/block/bsg.c b/block/bsg.c index a333c9337093..2f78d7d34b9d 100644 --- a/block/bsg.c +++ b/block/bsg.c | |||
@@ -115,9 +115,9 @@ static void bsg_free_command(struct bsg_command *bc) | |||
115 | wake_up(&bd->wq_free); | 115 | wake_up(&bd->wq_free); |
116 | } | 116 | } |
117 | 117 | ||
118 | static struct bsg_command *__bsg_alloc_command(struct bsg_device *bd) | 118 | static struct bsg_command *bsg_alloc_command(struct bsg_device *bd) |
119 | { | 119 | { |
120 | struct bsg_command *bc = NULL; | 120 | struct bsg_command *bc = ERR_PTR(-EINVAL); |
121 | 121 | ||
122 | spin_lock_irq(&bd->lock); | 122 | spin_lock_irq(&bd->lock); |
123 | 123 | ||
@@ -131,6 +131,7 @@ static struct bsg_command *__bsg_alloc_command(struct bsg_device *bd) | |||
131 | if (unlikely(!bc)) { | 131 | if (unlikely(!bc)) { |
132 | spin_lock_irq(&bd->lock); | 132 | spin_lock_irq(&bd->lock); |
133 | bd->queued_cmds--; | 133 | bd->queued_cmds--; |
134 | bc = ERR_PTR(-ENOMEM); | ||
134 | goto out; | 135 | goto out; |
135 | } | 136 | } |
136 | 137 | ||
@@ -198,30 +199,6 @@ unlock: | |||
198 | return ret; | 199 | return ret; |
199 | } | 200 | } |
200 | 201 | ||
201 | /* | ||
202 | * get a new free command, blocking if needed and specified | ||
203 | */ | ||
204 | static struct bsg_command *bsg_get_command(struct bsg_device *bd) | ||
205 | { | ||
206 | struct bsg_command *bc; | ||
207 | int ret; | ||
208 | |||
209 | do { | ||
210 | bc = __bsg_alloc_command(bd); | ||
211 | if (bc) | ||
212 | break; | ||
213 | |||
214 | ret = bsg_io_schedule(bd, TASK_INTERRUPTIBLE); | ||
215 | if (ret) { | ||
216 | bc = ERR_PTR(ret); | ||
217 | break; | ||
218 | } | ||
219 | |||
220 | } while (1); | ||
221 | |||
222 | return bc; | ||
223 | } | ||
224 | |||
225 | static int blk_fill_sgv4_hdr_rq(request_queue_t *q, struct request *rq, | 202 | static int blk_fill_sgv4_hdr_rq(request_queue_t *q, struct request *rq, |
226 | struct sg_io_v4 *hdr, int has_write_perm) | 203 | struct sg_io_v4 *hdr, int has_write_perm) |
227 | { | 204 | { |
@@ -397,7 +374,7 @@ static inline struct bsg_command *bsg_next_done_cmd(struct bsg_device *bd) | |||
397 | /* | 374 | /* |
398 | * Get a finished command from the done list | 375 | * Get a finished command from the done list |
399 | */ | 376 | */ |
400 | static struct bsg_command *__bsg_get_done_cmd(struct bsg_device *bd, int state) | 377 | static struct bsg_command *bsg_get_done_cmd(struct bsg_device *bd) |
401 | { | 378 | { |
402 | struct bsg_command *bc; | 379 | struct bsg_command *bc; |
403 | int ret; | 380 | int ret; |
@@ -407,9 +384,14 @@ static struct bsg_command *__bsg_get_done_cmd(struct bsg_device *bd, int state) | |||
407 | if (bc) | 384 | if (bc) |
408 | break; | 385 | break; |
409 | 386 | ||
410 | ret = bsg_io_schedule(bd, state); | 387 | if (!test_bit(BSG_F_BLOCK, &bd->flags)) { |
388 | bc = ERR_PTR(-EAGAIN); | ||
389 | break; | ||
390 | } | ||
391 | |||
392 | ret = wait_event_interruptible(bd->wq_done, bd->done_cmds); | ||
411 | if (ret) { | 393 | if (ret) { |
412 | bc = ERR_PTR(ret); | 394 | bc = ERR_PTR(-ERESTARTSYS); |
413 | break; | 395 | break; |
414 | } | 396 | } |
415 | } while (1); | 397 | } while (1); |
@@ -419,18 +401,6 @@ static struct bsg_command *__bsg_get_done_cmd(struct bsg_device *bd, int state) | |||
419 | return bc; | 401 | return bc; |
420 | } | 402 | } |
421 | 403 | ||
422 | static struct bsg_command * | ||
423 | bsg_get_done_cmd(struct bsg_device *bd, const struct iovec *iov) | ||
424 | { | ||
425 | return __bsg_get_done_cmd(bd, TASK_INTERRUPTIBLE); | ||
426 | } | ||
427 | |||
428 | static struct bsg_command * | ||
429 | bsg_get_done_cmd_nosignals(struct bsg_device *bd) | ||
430 | { | ||
431 | return __bsg_get_done_cmd(bd, TASK_UNINTERRUPTIBLE); | ||
432 | } | ||
433 | |||
434 | static int blk_complete_sgv4_hdr_rq(struct request *rq, struct sg_io_v4 *hdr, | 404 | static int blk_complete_sgv4_hdr_rq(struct request *rq, struct sg_io_v4 *hdr, |
435 | struct bio *bio) | 405 | struct bio *bio) |
436 | { | 406 | { |
@@ -496,19 +466,16 @@ static int bsg_complete_all_commands(struct bsg_device *bd) | |||
496 | */ | 466 | */ |
497 | ret = 0; | 467 | ret = 0; |
498 | do { | 468 | do { |
499 | bc = bsg_get_done_cmd_nosignals(bd); | 469 | spin_lock_irq(&bd->lock); |
500 | 470 | if (!bd->queued_cmds) { | |
501 | /* | 471 | spin_unlock_irq(&bd->lock); |
502 | * we _must_ complete before restarting, because | ||
503 | * bsg_release can't handle this failing. | ||
504 | */ | ||
505 | if (PTR_ERR(bc) == -ERESTARTSYS) | ||
506 | continue; | ||
507 | if (IS_ERR(bc)) { | ||
508 | ret = PTR_ERR(bc); | ||
509 | break; | 472 | break; |
510 | } | 473 | } |
511 | 474 | ||
475 | bc = bsg_get_done_cmd(bd); | ||
476 | if (IS_ERR(bc)) | ||
477 | break; | ||
478 | |||
512 | tret = blk_complete_sgv4_hdr_rq(bc->rq, &bc->hdr, bc->bio); | 479 | tret = blk_complete_sgv4_hdr_rq(bc->rq, &bc->hdr, bc->bio); |
513 | if (!ret) | 480 | if (!ret) |
514 | ret = tret; | 481 | ret = tret; |
@@ -519,11 +486,9 @@ static int bsg_complete_all_commands(struct bsg_device *bd) | |||
519 | return ret; | 486 | return ret; |
520 | } | 487 | } |
521 | 488 | ||
522 | typedef struct bsg_command *(*bsg_command_callback)(struct bsg_device *bd, const struct iovec *iov); | ||
523 | |||
524 | static ssize_t | 489 | static ssize_t |
525 | __bsg_read(char __user *buf, size_t count, bsg_command_callback get_bc, | 490 | __bsg_read(char __user *buf, size_t count, struct bsg_device *bd, |
526 | struct bsg_device *bd, const struct iovec *iov, ssize_t *bytes_read) | 491 | const struct iovec *iov, ssize_t *bytes_read) |
527 | { | 492 | { |
528 | struct bsg_command *bc; | 493 | struct bsg_command *bc; |
529 | int nr_commands, ret; | 494 | int nr_commands, ret; |
@@ -534,7 +499,7 @@ __bsg_read(char __user *buf, size_t count, bsg_command_callback get_bc, | |||
534 | ret = 0; | 499 | ret = 0; |
535 | nr_commands = count / sizeof(struct sg_io_v4); | 500 | nr_commands = count / sizeof(struct sg_io_v4); |
536 | while (nr_commands) { | 501 | while (nr_commands) { |
537 | bc = get_bc(bd, iov); | 502 | bc = bsg_get_done_cmd(bd); |
538 | if (IS_ERR(bc)) { | 503 | if (IS_ERR(bc)) { |
539 | ret = PTR_ERR(bc); | 504 | ret = PTR_ERR(bc); |
540 | break; | 505 | break; |
@@ -598,8 +563,7 @@ bsg_read(struct file *file, char __user *buf, size_t count, loff_t *ppos) | |||
598 | 563 | ||
599 | bsg_set_block(bd, file); | 564 | bsg_set_block(bd, file); |
600 | bytes_read = 0; | 565 | bytes_read = 0; |
601 | ret = __bsg_read(buf, count, bsg_get_done_cmd, | 566 | ret = __bsg_read(buf, count, bd, NULL, &bytes_read); |
602 | bd, NULL, &bytes_read); | ||
603 | *ppos = bytes_read; | 567 | *ppos = bytes_read; |
604 | 568 | ||
605 | if (!bytes_read || (bytes_read && err_block_err(ret))) | 569 | if (!bytes_read || (bytes_read && err_block_err(ret))) |
@@ -625,9 +589,7 @@ static ssize_t __bsg_write(struct bsg_device *bd, const char __user *buf, | |||
625 | while (nr_commands) { | 589 | while (nr_commands) { |
626 | request_queue_t *q = bd->queue; | 590 | request_queue_t *q = bd->queue; |
627 | 591 | ||
628 | bc = bsg_get_command(bd); | 592 | bc = bsg_alloc_command(bd); |
629 | if (!bc) | ||
630 | break; | ||
631 | if (IS_ERR(bc)) { | 593 | if (IS_ERR(bc)) { |
632 | ret = PTR_ERR(bc); | 594 | ret = PTR_ERR(bc); |
633 | bc = NULL; | 595 | bc = NULL; |