diff options
Diffstat (limited to 'block')
-rw-r--r-- | block/blktrace.c | 24 | ||||
-rw-r--r-- | block/compat_ioctl.c | 5 | ||||
-rw-r--r-- | block/elevator.c | 26 | ||||
-rw-r--r-- | block/ll_rw_blk.c | 49 |
4 files changed, 92 insertions, 12 deletions
diff --git a/block/blktrace.c b/block/blktrace.c index 9b4da4ae3c7d..568588cd16b2 100644 --- a/block/blktrace.c +++ b/block/blktrace.c | |||
@@ -235,7 +235,7 @@ static void blk_trace_cleanup(struct blk_trace *bt) | |||
235 | kfree(bt); | 235 | kfree(bt); |
236 | } | 236 | } |
237 | 237 | ||
238 | static int blk_trace_remove(struct request_queue *q) | 238 | int blk_trace_remove(struct request_queue *q) |
239 | { | 239 | { |
240 | struct blk_trace *bt; | 240 | struct blk_trace *bt; |
241 | 241 | ||
@@ -249,6 +249,7 @@ static int blk_trace_remove(struct request_queue *q) | |||
249 | 249 | ||
250 | return 0; | 250 | return 0; |
251 | } | 251 | } |
252 | EXPORT_SYMBOL_GPL(blk_trace_remove); | ||
252 | 253 | ||
253 | static int blk_dropped_open(struct inode *inode, struct file *filp) | 254 | static int blk_dropped_open(struct inode *inode, struct file *filp) |
254 | { | 255 | { |
@@ -316,18 +317,17 @@ static struct rchan_callbacks blk_relay_callbacks = { | |||
316 | /* | 317 | /* |
317 | * Setup everything required to start tracing | 318 | * Setup everything required to start tracing |
318 | */ | 319 | */ |
319 | int do_blk_trace_setup(struct request_queue *q, struct block_device *bdev, | 320 | int do_blk_trace_setup(struct request_queue *q, char *name, dev_t dev, |
320 | struct blk_user_trace_setup *buts) | 321 | struct blk_user_trace_setup *buts) |
321 | { | 322 | { |
322 | struct blk_trace *old_bt, *bt = NULL; | 323 | struct blk_trace *old_bt, *bt = NULL; |
323 | struct dentry *dir = NULL; | 324 | struct dentry *dir = NULL; |
324 | char b[BDEVNAME_SIZE]; | ||
325 | int ret, i; | 325 | int ret, i; |
326 | 326 | ||
327 | if (!buts->buf_size || !buts->buf_nr) | 327 | if (!buts->buf_size || !buts->buf_nr) |
328 | return -EINVAL; | 328 | return -EINVAL; |
329 | 329 | ||
330 | strcpy(buts->name, bdevname(bdev, b)); | 330 | strcpy(buts->name, name); |
331 | 331 | ||
332 | /* | 332 | /* |
333 | * some device names have larger paths - convert the slashes | 333 | * some device names have larger paths - convert the slashes |
@@ -352,7 +352,7 @@ int do_blk_trace_setup(struct request_queue *q, struct block_device *bdev, | |||
352 | goto err; | 352 | goto err; |
353 | 353 | ||
354 | bt->dir = dir; | 354 | bt->dir = dir; |
355 | bt->dev = bdev->bd_dev; | 355 | bt->dev = dev; |
356 | atomic_set(&bt->dropped, 0); | 356 | atomic_set(&bt->dropped, 0); |
357 | 357 | ||
358 | ret = -EIO; | 358 | ret = -EIO; |
@@ -399,8 +399,8 @@ err: | |||
399 | return ret; | 399 | return ret; |
400 | } | 400 | } |
401 | 401 | ||
402 | static int blk_trace_setup(struct request_queue *q, struct block_device *bdev, | 402 | int blk_trace_setup(struct request_queue *q, char *name, dev_t dev, |
403 | char __user *arg) | 403 | char __user *arg) |
404 | { | 404 | { |
405 | struct blk_user_trace_setup buts; | 405 | struct blk_user_trace_setup buts; |
406 | int ret; | 406 | int ret; |
@@ -409,7 +409,7 @@ static int blk_trace_setup(struct request_queue *q, struct block_device *bdev, | |||
409 | if (ret) | 409 | if (ret) |
410 | return -EFAULT; | 410 | return -EFAULT; |
411 | 411 | ||
412 | ret = do_blk_trace_setup(q, bdev, &buts); | 412 | ret = do_blk_trace_setup(q, name, dev, &buts); |
413 | if (ret) | 413 | if (ret) |
414 | return ret; | 414 | return ret; |
415 | 415 | ||
@@ -418,8 +418,9 @@ static int blk_trace_setup(struct request_queue *q, struct block_device *bdev, | |||
418 | 418 | ||
419 | return 0; | 419 | return 0; |
420 | } | 420 | } |
421 | EXPORT_SYMBOL_GPL(blk_trace_setup); | ||
421 | 422 | ||
422 | static int blk_trace_startstop(struct request_queue *q, int start) | 423 | int blk_trace_startstop(struct request_queue *q, int start) |
423 | { | 424 | { |
424 | struct blk_trace *bt; | 425 | struct blk_trace *bt; |
425 | int ret; | 426 | int ret; |
@@ -452,6 +453,7 @@ static int blk_trace_startstop(struct request_queue *q, int start) | |||
452 | 453 | ||
453 | return ret; | 454 | return ret; |
454 | } | 455 | } |
456 | EXPORT_SYMBOL_GPL(blk_trace_startstop); | ||
455 | 457 | ||
456 | /** | 458 | /** |
457 | * blk_trace_ioctl: - handle the ioctls associated with tracing | 459 | * blk_trace_ioctl: - handle the ioctls associated with tracing |
@@ -464,6 +466,7 @@ int blk_trace_ioctl(struct block_device *bdev, unsigned cmd, char __user *arg) | |||
464 | { | 466 | { |
465 | struct request_queue *q; | 467 | struct request_queue *q; |
466 | int ret, start = 0; | 468 | int ret, start = 0; |
469 | char b[BDEVNAME_SIZE]; | ||
467 | 470 | ||
468 | q = bdev_get_queue(bdev); | 471 | q = bdev_get_queue(bdev); |
469 | if (!q) | 472 | if (!q) |
@@ -473,7 +476,8 @@ int blk_trace_ioctl(struct block_device *bdev, unsigned cmd, char __user *arg) | |||
473 | 476 | ||
474 | switch (cmd) { | 477 | switch (cmd) { |
475 | case BLKTRACESETUP: | 478 | case BLKTRACESETUP: |
476 | ret = blk_trace_setup(q, bdev, arg); | 479 | strcpy(b, bdevname(bdev, b)); |
480 | ret = blk_trace_setup(q, b, bdev->bd_dev, arg); | ||
477 | break; | 481 | break; |
478 | case BLKTRACESTART: | 482 | case BLKTRACESTART: |
479 | start = 1; | 483 | start = 1; |
diff --git a/block/compat_ioctl.c b/block/compat_ioctl.c index cae0a852619e..b73373216b0e 100644 --- a/block/compat_ioctl.c +++ b/block/compat_ioctl.c | |||
@@ -545,6 +545,7 @@ static int compat_blk_trace_setup(struct block_device *bdev, char __user *arg) | |||
545 | struct blk_user_trace_setup buts; | 545 | struct blk_user_trace_setup buts; |
546 | struct compat_blk_user_trace_setup cbuts; | 546 | struct compat_blk_user_trace_setup cbuts; |
547 | struct request_queue *q; | 547 | struct request_queue *q; |
548 | char b[BDEVNAME_SIZE]; | ||
548 | int ret; | 549 | int ret; |
549 | 550 | ||
550 | q = bdev_get_queue(bdev); | 551 | q = bdev_get_queue(bdev); |
@@ -554,6 +555,8 @@ static int compat_blk_trace_setup(struct block_device *bdev, char __user *arg) | |||
554 | if (copy_from_user(&cbuts, arg, sizeof(cbuts))) | 555 | if (copy_from_user(&cbuts, arg, sizeof(cbuts))) |
555 | return -EFAULT; | 556 | return -EFAULT; |
556 | 557 | ||
558 | strcpy(b, bdevname(bdev, b)); | ||
559 | |||
557 | buts = (struct blk_user_trace_setup) { | 560 | buts = (struct blk_user_trace_setup) { |
558 | .act_mask = cbuts.act_mask, | 561 | .act_mask = cbuts.act_mask, |
559 | .buf_size = cbuts.buf_size, | 562 | .buf_size = cbuts.buf_size, |
@@ -565,7 +568,7 @@ static int compat_blk_trace_setup(struct block_device *bdev, char __user *arg) | |||
565 | memcpy(&buts.name, &cbuts.name, 32); | 568 | memcpy(&buts.name, &cbuts.name, 32); |
566 | 569 | ||
567 | mutex_lock(&bdev->bd_mutex); | 570 | mutex_lock(&bdev->bd_mutex); |
568 | ret = do_blk_trace_setup(q, bdev, &buts); | 571 | ret = do_blk_trace_setup(q, b, bdev->bd_dev, &buts); |
569 | mutex_unlock(&bdev->bd_mutex); | 572 | mutex_unlock(&bdev->bd_mutex); |
570 | if (ret) | 573 | if (ret) |
571 | return ret; | 574 | return ret; |
diff --git a/block/elevator.c b/block/elevator.c index f9736fbdab03..8cd5775acd7a 100644 --- a/block/elevator.c +++ b/block/elevator.c | |||
@@ -741,7 +741,21 @@ struct request *elv_next_request(struct request_queue *q) | |||
741 | q->boundary_rq = NULL; | 741 | q->boundary_rq = NULL; |
742 | } | 742 | } |
743 | 743 | ||
744 | if ((rq->cmd_flags & REQ_DONTPREP) || !q->prep_rq_fn) | 744 | if (rq->cmd_flags & REQ_DONTPREP) |
745 | break; | ||
746 | |||
747 | if (q->dma_drain_size && rq->data_len) { | ||
748 | /* | ||
749 | * make sure space for the drain appears we | ||
750 | * know we can do this because max_hw_segments | ||
751 | * has been adjusted to be one fewer than the | ||
752 | * device can handle | ||
753 | */ | ||
754 | rq->nr_phys_segments++; | ||
755 | rq->nr_hw_segments++; | ||
756 | } | ||
757 | |||
758 | if (!q->prep_rq_fn) | ||
745 | break; | 759 | break; |
746 | 760 | ||
747 | ret = q->prep_rq_fn(q, rq); | 761 | ret = q->prep_rq_fn(q, rq); |
@@ -754,6 +768,16 @@ struct request *elv_next_request(struct request_queue *q) | |||
754 | * avoid resource deadlock. REQ_STARTED will | 768 | * avoid resource deadlock. REQ_STARTED will |
755 | * prevent other fs requests from passing this one. | 769 | * prevent other fs requests from passing this one. |
756 | */ | 770 | */ |
771 | if (q->dma_drain_size && rq->data_len && | ||
772 | !(rq->cmd_flags & REQ_DONTPREP)) { | ||
773 | /* | ||
774 | * remove the space for the drain we added | ||
775 | * so that we don't add it again | ||
776 | */ | ||
777 | --rq->nr_phys_segments; | ||
778 | --rq->nr_hw_segments; | ||
779 | } | ||
780 | |||
757 | rq = NULL; | 781 | rq = NULL; |
758 | break; | 782 | break; |
759 | } else if (ret == BLKPREP_KILL) { | 783 | } else if (ret == BLKPREP_KILL) { |
diff --git a/block/ll_rw_blk.c b/block/ll_rw_blk.c index c16fdfed8c62..1932a56f5e4b 100644 --- a/block/ll_rw_blk.c +++ b/block/ll_rw_blk.c | |||
@@ -721,6 +721,45 @@ void blk_queue_stack_limits(struct request_queue *t, struct request_queue *b) | |||
721 | EXPORT_SYMBOL(blk_queue_stack_limits); | 721 | EXPORT_SYMBOL(blk_queue_stack_limits); |
722 | 722 | ||
723 | /** | 723 | /** |
724 | * blk_queue_dma_drain - Set up a drain buffer for excess dma. | ||
725 | * | ||
726 | * @q: the request queue for the device | ||
727 | * @buf: physically contiguous buffer | ||
728 | * @size: size of the buffer in bytes | ||
729 | * | ||
730 | * Some devices have excess DMA problems and can't simply discard (or | ||
731 | * zero fill) the unwanted piece of the transfer. They have to have a | ||
732 | * real area of memory to transfer it into. The use case for this is | ||
733 | * ATAPI devices in DMA mode. If the packet command causes a transfer | ||
734 | * bigger than the transfer size some HBAs will lock up if there | ||
735 | * aren't DMA elements to contain the excess transfer. What this API | ||
736 | * does is adjust the queue so that the buf is always appended | ||
737 | * silently to the scatterlist. | ||
738 | * | ||
739 | * Note: This routine adjusts max_hw_segments to make room for | ||
740 | * appending the drain buffer. If you call | ||
741 | * blk_queue_max_hw_segments() or blk_queue_max_phys_segments() after | ||
742 | * calling this routine, you must set the limit to one fewer than your | ||
743 | * device can support otherwise there won't be room for the drain | ||
744 | * buffer. | ||
745 | */ | ||
746 | int blk_queue_dma_drain(struct request_queue *q, void *buf, | ||
747 | unsigned int size) | ||
748 | { | ||
749 | if (q->max_hw_segments < 2 || q->max_phys_segments < 2) | ||
750 | return -EINVAL; | ||
751 | /* make room for appending the drain */ | ||
752 | --q->max_hw_segments; | ||
753 | --q->max_phys_segments; | ||
754 | q->dma_drain_buffer = buf; | ||
755 | q->dma_drain_size = size; | ||
756 | |||
757 | return 0; | ||
758 | } | ||
759 | |||
760 | EXPORT_SYMBOL_GPL(blk_queue_dma_drain); | ||
761 | |||
762 | /** | ||
724 | * blk_queue_segment_boundary - set boundary rules for segment merging | 763 | * blk_queue_segment_boundary - set boundary rules for segment merging |
725 | * @q: the request queue for the device | 764 | * @q: the request queue for the device |
726 | * @mask: the memory boundary mask | 765 | * @mask: the memory boundary mask |
@@ -1374,6 +1413,16 @@ new_segment: | |||
1374 | bvprv = bvec; | 1413 | bvprv = bvec; |
1375 | } /* segments in rq */ | 1414 | } /* segments in rq */ |
1376 | 1415 | ||
1416 | if (q->dma_drain_size) { | ||
1417 | sg->page_link &= ~0x02; | ||
1418 | sg = sg_next(sg); | ||
1419 | sg_set_page(sg, virt_to_page(q->dma_drain_buffer), | ||
1420 | q->dma_drain_size, | ||
1421 | ((unsigned long)q->dma_drain_buffer) & | ||
1422 | (PAGE_SIZE - 1)); | ||
1423 | nsegs++; | ||
1424 | } | ||
1425 | |||
1377 | if (sg) | 1426 | if (sg) |
1378 | sg_mark_end(sg); | 1427 | sg_mark_end(sg); |
1379 | 1428 | ||