diff options
author | Jens Axboe <jens.axboe@oracle.com> | 2007-07-24 03:28:11 -0400 |
---|---|---|
committer | Jens Axboe <jens.axboe@oracle.com> | 2007-07-24 03:28:11 -0400 |
commit | 165125e1e480f9510a5ffcfbfee4e3ee38c05f23 (patch) | |
tree | 8009c8a5ff09e26dc2418d42f66ecafb055c52a2 /Documentation | |
parent | f695baf2df9e0413d3521661070103711545207a (diff) |
[BLOCK] Get rid of request_queue_t typedef
Some of the code has been gradually transitioned to using the proper
struct request_queue, but there's lots left. So do a full sweet of
the kernel and get rid of this typedef and replace its uses with
the proper type.
Signed-off-by: Jens Axboe <jens.axboe@oracle.com>
Diffstat (limited to 'Documentation')
-rw-r--r-- | Documentation/block/barrier.txt | 6 | ||||
-rw-r--r-- | Documentation/block/biodoc.txt | 10 | ||||
-rw-r--r-- | Documentation/block/request.txt | 2 | ||||
-rw-r--r-- | Documentation/iostats.txt | 2 |
4 files changed, 10 insertions, 10 deletions
diff --git a/Documentation/block/barrier.txt b/Documentation/block/barrier.txt index 7d279f2f5bb2..2c2f24f634e4 100644 --- a/Documentation/block/barrier.txt +++ b/Documentation/block/barrier.txt | |||
@@ -79,9 +79,9 @@ and how to prepare flush requests. Note that the term 'ordered' is | |||
79 | used to indicate the whole sequence of performing barrier requests | 79 | used to indicate the whole sequence of performing barrier requests |
80 | including draining and flushing. | 80 | including draining and flushing. |
81 | 81 | ||
82 | typedef void (prepare_flush_fn)(request_queue_t *q, struct request *rq); | 82 | typedef void (prepare_flush_fn)(struct request_queue *q, struct request *rq); |
83 | 83 | ||
84 | int blk_queue_ordered(request_queue_t *q, unsigned ordered, | 84 | int blk_queue_ordered(struct request_queue *q, unsigned ordered, |
85 | prepare_flush_fn *prepare_flush_fn); | 85 | prepare_flush_fn *prepare_flush_fn); |
86 | 86 | ||
87 | @q : the queue in question | 87 | @q : the queue in question |
@@ -92,7 +92,7 @@ int blk_queue_ordered(request_queue_t *q, unsigned ordered, | |||
92 | For example, SCSI disk driver's prepare_flush_fn looks like the | 92 | For example, SCSI disk driver's prepare_flush_fn looks like the |
93 | following. | 93 | following. |
94 | 94 | ||
95 | static void sd_prepare_flush(request_queue_t *q, struct request *rq) | 95 | static void sd_prepare_flush(struct request_queue *q, struct request *rq) |
96 | { | 96 | { |
97 | memset(rq->cmd, 0, sizeof(rq->cmd)); | 97 | memset(rq->cmd, 0, sizeof(rq->cmd)); |
98 | rq->cmd_type = REQ_TYPE_BLOCK_PC; | 98 | rq->cmd_type = REQ_TYPE_BLOCK_PC; |
diff --git a/Documentation/block/biodoc.txt b/Documentation/block/biodoc.txt index 3adaace328a6..8af392fc6ef0 100644 --- a/Documentation/block/biodoc.txt +++ b/Documentation/block/biodoc.txt | |||
@@ -740,12 +740,12 @@ Block now offers some simple generic functionality to help support command | |||
740 | queueing (typically known as tagged command queueing), ie manage more than | 740 | queueing (typically known as tagged command queueing), ie manage more than |
741 | one outstanding command on a queue at any given time. | 741 | one outstanding command on a queue at any given time. |
742 | 742 | ||
743 | blk_queue_init_tags(request_queue_t *q, int depth) | 743 | blk_queue_init_tags(struct request_queue *q, int depth) |
744 | 744 | ||
745 | Initialize internal command tagging structures for a maximum | 745 | Initialize internal command tagging structures for a maximum |
746 | depth of 'depth'. | 746 | depth of 'depth'. |
747 | 747 | ||
748 | blk_queue_free_tags((request_queue_t *q) | 748 | blk_queue_free_tags((struct request_queue *q) |
749 | 749 | ||
750 | Teardown tag info associated with the queue. This will be done | 750 | Teardown tag info associated with the queue. This will be done |
751 | automatically by block if blk_queue_cleanup() is called on a queue | 751 | automatically by block if blk_queue_cleanup() is called on a queue |
@@ -754,7 +754,7 @@ one outstanding command on a queue at any given time. | |||
754 | The above are initialization and exit management, the main helpers during | 754 | The above are initialization and exit management, the main helpers during |
755 | normal operations are: | 755 | normal operations are: |
756 | 756 | ||
757 | blk_queue_start_tag(request_queue_t *q, struct request *rq) | 757 | blk_queue_start_tag(struct request_queue *q, struct request *rq) |
758 | 758 | ||
759 | Start tagged operation for this request. A free tag number between | 759 | Start tagged operation for this request. A free tag number between |
760 | 0 and 'depth' is assigned to the request (rq->tag holds this number), | 760 | 0 and 'depth' is assigned to the request (rq->tag holds this number), |
@@ -762,7 +762,7 @@ normal operations are: | |||
762 | for this queue is already achieved (or if the tag wasn't started for | 762 | for this queue is already achieved (or if the tag wasn't started for |
763 | some other reason), 1 is returned. Otherwise 0 is returned. | 763 | some other reason), 1 is returned. Otherwise 0 is returned. |
764 | 764 | ||
765 | blk_queue_end_tag(request_queue_t *q, struct request *rq) | 765 | blk_queue_end_tag(struct request_queue *q, struct request *rq) |
766 | 766 | ||
767 | End tagged operation on this request. 'rq' is removed from the internal | 767 | End tagged operation on this request. 'rq' is removed from the internal |
768 | book keeping structures. | 768 | book keeping structures. |
@@ -781,7 +781,7 @@ queue. For instance, on IDE any tagged request error needs to clear both | |||
781 | the hardware and software block queue and enable the driver to sanely restart | 781 | the hardware and software block queue and enable the driver to sanely restart |
782 | all the outstanding requests. There's a third helper to do that: | 782 | all the outstanding requests. There's a third helper to do that: |
783 | 783 | ||
784 | blk_queue_invalidate_tags(request_queue_t *q) | 784 | blk_queue_invalidate_tags(struct request_queue *q) |
785 | 785 | ||
786 | Clear the internal block tag queue and re-add all the pending requests | 786 | Clear the internal block tag queue and re-add all the pending requests |
787 | to the request queue. The driver will receive them again on the | 787 | to the request queue. The driver will receive them again on the |
diff --git a/Documentation/block/request.txt b/Documentation/block/request.txt index 75924e2a6975..fff58acb40a3 100644 --- a/Documentation/block/request.txt +++ b/Documentation/block/request.txt | |||
@@ -83,6 +83,6 @@ struct bio *bio DBI First bio in request | |||
83 | 83 | ||
84 | struct bio *biotail DBI Last bio in request | 84 | struct bio *biotail DBI Last bio in request |
85 | 85 | ||
86 | request_queue_t *q DB Request queue this request belongs to | 86 | struct request_queue *q DB Request queue this request belongs to |
87 | 87 | ||
88 | struct request_list *rl B Request list this request came from | 88 | struct request_list *rl B Request list this request came from |
diff --git a/Documentation/iostats.txt b/Documentation/iostats.txt index 09a1bafe2528..b963c3b4afa5 100644 --- a/Documentation/iostats.txt +++ b/Documentation/iostats.txt | |||
@@ -79,7 +79,7 @@ Field 8 -- # of milliseconds spent writing | |||
79 | measured from __make_request() to end_that_request_last()). | 79 | measured from __make_request() to end_that_request_last()). |
80 | Field 9 -- # of I/Os currently in progress | 80 | Field 9 -- # of I/Os currently in progress |
81 | The only field that should go to zero. Incremented as requests are | 81 | The only field that should go to zero. Incremented as requests are |
82 | given to appropriate request_queue_t and decremented as they finish. | 82 | given to appropriate struct request_queue and decremented as they finish. |
83 | Field 10 -- # of milliseconds spent doing I/Os | 83 | Field 10 -- # of milliseconds spent doing I/Os |
84 | This field is increases so long as field 9 is nonzero. | 84 | This field is increases so long as field 9 is nonzero. |
85 | Field 11 -- weighted # of milliseconds spent doing I/Os | 85 | Field 11 -- weighted # of milliseconds spent doing I/Os |