diff options
Diffstat (limited to 'Documentation/block')
-rw-r--r-- | Documentation/block/as-iosched.txt | 4 | ||||
-rw-r--r-- | Documentation/block/barrier.txt | 6 | ||||
-rw-r--r-- | Documentation/block/biodoc.txt | 10 | ||||
-rw-r--r-- | Documentation/block/deadline-iosched.txt | 4 |
4 files changed, 12 insertions, 12 deletions
diff --git a/Documentation/block/as-iosched.txt b/Documentation/block/as-iosched.txt index 6f47332c883d..e2a66f8143c5 100644 --- a/Documentation/block/as-iosched.txt +++ b/Documentation/block/as-iosched.txt | |||
@@ -99,8 +99,8 @@ contrast, many write requests may be dispatched to the disk controller | |||
99 | at a time during a write batch. It is this characteristic that can make | 99 | at a time during a write batch. It is this characteristic that can make |
100 | the anticipatory scheduler perform anomalously with controllers supporting | 100 | the anticipatory scheduler perform anomalously with controllers supporting |
101 | TCQ, or with hardware striped RAID devices. Setting the antic_expire | 101 | TCQ, or with hardware striped RAID devices. Setting the antic_expire |
102 | queue paramter (see below) to zero disables this behavior, and the anticipatory | 102 | queue parameter (see below) to zero disables this behavior, and the |
103 | scheduler behaves essentially like the deadline scheduler. | 103 | anticipatory scheduler behaves essentially like the deadline scheduler. |
104 | 104 | ||
105 | When read anticipation is enabled (antic_expire is not zero), reads | 105 | When read anticipation is enabled (antic_expire is not zero), reads |
106 | are dispatched to the disk controller one at a time. | 106 | are dispatched to the disk controller one at a time. |
diff --git a/Documentation/block/barrier.txt b/Documentation/block/barrier.txt index 03971518b222..a272c3db8094 100644 --- a/Documentation/block/barrier.txt +++ b/Documentation/block/barrier.txt | |||
@@ -25,7 +25,7 @@ of the following three ways. | |||
25 | i. For devices which have queue depth greater than 1 (TCQ devices) and | 25 | i. For devices which have queue depth greater than 1 (TCQ devices) and |
26 | support ordered tags, block layer can just issue the barrier as an | 26 | support ordered tags, block layer can just issue the barrier as an |
27 | ordered request and the lower level driver, controller and drive | 27 | ordered request and the lower level driver, controller and drive |
28 | itself are responsible for making sure that the ordering contraint is | 28 | itself are responsible for making sure that the ordering constraint is |
29 | met. Most modern SCSI controllers/drives should support this. | 29 | met. Most modern SCSI controllers/drives should support this. |
30 | 30 | ||
31 | NOTE: SCSI ordered tag isn't currently used due to limitation in the | 31 | NOTE: SCSI ordered tag isn't currently used due to limitation in the |
@@ -42,7 +42,7 @@ iii. Devices which have queue depth of 1. This is a degenerate case | |||
42 | of ii. Just keeping issue order suffices. Ancient SCSI | 42 | of ii. Just keeping issue order suffices. Ancient SCSI |
43 | controllers/drives and IDE drives are in this category. | 43 | controllers/drives and IDE drives are in this category. |
44 | 44 | ||
45 | 2. Forced flushing to physcial medium | 45 | 2. Forced flushing to physical medium |
46 | 46 | ||
47 | Again, if you're not gonna do synchronization with disk drives (dang, | 47 | Again, if you're not gonna do synchronization with disk drives (dang, |
48 | it sounds even more appealing now!), the reason you use I/O barriers | 48 | it sounds even more appealing now!), the reason you use I/O barriers |
@@ -56,7 +56,7 @@ There are four cases, | |||
56 | i. No write-back cache. Keeping requests ordered is enough. | 56 | i. No write-back cache. Keeping requests ordered is enough. |
57 | 57 | ||
58 | ii. Write-back cache but no flush operation. There's no way to | 58 | ii. Write-back cache but no flush operation. There's no way to |
59 | gurantee physical-medium commit order. This kind of devices can't to | 59 | guarantee physical-medium commit order. This kind of devices can't to |
60 | I/O barriers. | 60 | I/O barriers. |
61 | 61 | ||
62 | iii. Write-back cache and flush operation but no FUA (forced unit | 62 | iii. Write-back cache and flush operation but no FUA (forced unit |
diff --git a/Documentation/block/biodoc.txt b/Documentation/block/biodoc.txt index f989a9e839b4..34bf8f60d8f8 100644 --- a/Documentation/block/biodoc.txt +++ b/Documentation/block/biodoc.txt | |||
@@ -135,7 +135,7 @@ Some new queue property settings: | |||
135 | Sets two variables that limit the size of the request. | 135 | Sets two variables that limit the size of the request. |
136 | 136 | ||
137 | - The request queue's max_sectors, which is a soft size in | 137 | - The request queue's max_sectors, which is a soft size in |
138 | in units of 512 byte sectors, and could be dynamically varied | 138 | units of 512 byte sectors, and could be dynamically varied |
139 | by the core kernel. | 139 | by the core kernel. |
140 | 140 | ||
141 | - The request queue's max_hw_sectors, which is a hard limit | 141 | - The request queue's max_hw_sectors, which is a hard limit |
@@ -783,7 +783,7 @@ all the outstanding requests. There's a third helper to do that: | |||
783 | 783 | ||
784 | blk_queue_invalidate_tags(request_queue_t *q) | 784 | blk_queue_invalidate_tags(request_queue_t *q) |
785 | 785 | ||
786 | Clear the internal block tag queue and readd all the pending requests | 786 | Clear the internal block tag queue and re-add all the pending requests |
787 | to the request queue. The driver will receive them again on the | 787 | to the request queue. The driver will receive them again on the |
788 | next request_fn run, just like it did the first time it encountered | 788 | next request_fn run, just like it did the first time it encountered |
789 | them. | 789 | them. |
@@ -890,7 +890,7 @@ Aside: | |||
890 | 890 | ||
891 | Kvec i/o: | 891 | Kvec i/o: |
892 | 892 | ||
893 | Ben LaHaise's aio code uses a slighly different structure instead | 893 | Ben LaHaise's aio code uses a slightly different structure instead |
894 | of kiobufs, called a kvec_cb. This contains an array of <page, offset, len> | 894 | of kiobufs, called a kvec_cb. This contains an array of <page, offset, len> |
895 | tuples (very much like the networking code), together with a callback function | 895 | tuples (very much like the networking code), together with a callback function |
896 | and data pointer. This is embedded into a brw_cb structure when passed | 896 | and data pointer. This is embedded into a brw_cb structure when passed |
@@ -988,7 +988,7 @@ elevator_exit_fn Allocate and free any elevator specific storage | |||
988 | for a queue. | 988 | for a queue. |
989 | 989 | ||
990 | 4.2 Request flows seen by I/O schedulers | 990 | 4.2 Request flows seen by I/O schedulers |
991 | All requests seens by I/O schedulers strictly follow one of the following three | 991 | All requests seen by I/O schedulers strictly follow one of the following three |
992 | flows. | 992 | flows. |
993 | 993 | ||
994 | set_req_fn -> | 994 | set_req_fn -> |
@@ -1203,6 +1203,6 @@ temporarily map a bio into the virtual address space. | |||
1203 | and Linus' comments - Jan 2001) | 1203 | and Linus' comments - Jan 2001) |
1204 | 9.2 Discussions about kiobuf and bh design on lkml between sct, linus, alan | 1204 | 9.2 Discussions about kiobuf and bh design on lkml between sct, linus, alan |
1205 | et al - Feb-March 2001 (many of the initial thoughts that led to bio were | 1205 | et al - Feb-March 2001 (many of the initial thoughts that led to bio were |
1206 | brought up in this discusion thread) | 1206 | brought up in this discussion thread) |
1207 | 9.3 Discussions on mempool on lkml - Dec 2001. | 1207 | 9.3 Discussions on mempool on lkml - Dec 2001. |
1208 | 1208 | ||
diff --git a/Documentation/block/deadline-iosched.txt b/Documentation/block/deadline-iosched.txt index c918b3a6022d..be08ffd1e9b8 100644 --- a/Documentation/block/deadline-iosched.txt +++ b/Documentation/block/deadline-iosched.txt | |||
@@ -23,11 +23,11 @@ you can do so by typing: | |||
23 | read_expire (in ms) | 23 | read_expire (in ms) |
24 | ----------- | 24 | ----------- |
25 | 25 | ||
26 | The goal of the deadline io scheduler is to attempt to guarentee a start | 26 | The goal of the deadline io scheduler is to attempt to guarantee a start |
27 | service time for a request. As we focus mainly on read latencies, this is | 27 | service time for a request. As we focus mainly on read latencies, this is |
28 | tunable. When a read request first enters the io scheduler, it is assigned | 28 | tunable. When a read request first enters the io scheduler, it is assigned |
29 | a deadline that is the current time + the read_expire value in units of | 29 | a deadline that is the current time + the read_expire value in units of |
30 | miliseconds. | 30 | milliseconds. |
31 | 31 | ||
32 | 32 | ||
33 | write_expire (in ms) | 33 | write_expire (in ms) |