aboutsummaryrefslogtreecommitdiffstats
path: root/Documentation/block
diff options
context:
space:
mode:
Diffstat (limited to 'Documentation/block')
-rw-r--r--Documentation/block/as-iosched.txt4
-rw-r--r--Documentation/block/barrier.txt6
-rw-r--r--Documentation/block/biodoc.txt10
-rw-r--r--Documentation/block/deadline-iosched.txt4
4 files changed, 12 insertions, 12 deletions
diff --git a/Documentation/block/as-iosched.txt b/Documentation/block/as-iosched.txt
index 6f47332c883d..e2a66f8143c5 100644
--- a/Documentation/block/as-iosched.txt
+++ b/Documentation/block/as-iosched.txt
@@ -99,8 +99,8 @@ contrast, many write requests may be dispatched to the disk controller
99at a time during a write batch. It is this characteristic that can make 99at a time during a write batch. It is this characteristic that can make
100the anticipatory scheduler perform anomalously with controllers supporting 100the anticipatory scheduler perform anomalously with controllers supporting
101TCQ, or with hardware striped RAID devices. Setting the antic_expire 101TCQ, or with hardware striped RAID devices. Setting the antic_expire
102queue paramter (see below) to zero disables this behavior, and the anticipatory 102queue parameter (see below) to zero disables this behavior, and the
103scheduler behaves essentially like the deadline scheduler. 103anticipatory scheduler behaves essentially like the deadline scheduler.
104 104
105When read anticipation is enabled (antic_expire is not zero), reads 105When read anticipation is enabled (antic_expire is not zero), reads
106are dispatched to the disk controller one at a time. 106are dispatched to the disk controller one at a time.
diff --git a/Documentation/block/barrier.txt b/Documentation/block/barrier.txt
index 03971518b222..a272c3db8094 100644
--- a/Documentation/block/barrier.txt
+++ b/Documentation/block/barrier.txt
@@ -25,7 +25,7 @@ of the following three ways.
25i. For devices which have queue depth greater than 1 (TCQ devices) and 25i. For devices which have queue depth greater than 1 (TCQ devices) and
26support ordered tags, block layer can just issue the barrier as an 26support ordered tags, block layer can just issue the barrier as an
27ordered request and the lower level driver, controller and drive 27ordered request and the lower level driver, controller and drive
28itself are responsible for making sure that the ordering contraint is 28itself are responsible for making sure that the ordering constraint is
29met. Most modern SCSI controllers/drives should support this. 29met. Most modern SCSI controllers/drives should support this.
30 30
31NOTE: SCSI ordered tag isn't currently used due to limitation in the 31NOTE: SCSI ordered tag isn't currently used due to limitation in the
@@ -42,7 +42,7 @@ iii. Devices which have queue depth of 1. This is a degenerate case
42of ii. Just keeping issue order suffices. Ancient SCSI 42of ii. Just keeping issue order suffices. Ancient SCSI
43controllers/drives and IDE drives are in this category. 43controllers/drives and IDE drives are in this category.
44 44
452. Forced flushing to physcial medium 452. Forced flushing to physical medium
46 46
47Again, if you're not gonna do synchronization with disk drives (dang, 47Again, if you're not gonna do synchronization with disk drives (dang,
48it sounds even more appealing now!), the reason you use I/O barriers 48it sounds even more appealing now!), the reason you use I/O barriers
@@ -56,7 +56,7 @@ There are four cases,
56i. No write-back cache. Keeping requests ordered is enough. 56i. No write-back cache. Keeping requests ordered is enough.
57 57
58ii. Write-back cache but no flush operation. There's no way to 58ii. Write-back cache but no flush operation. There's no way to
59gurantee physical-medium commit order. This kind of devices can't to 59guarantee physical-medium commit order. This kind of devices can't to
60I/O barriers. 60I/O barriers.
61 61
62iii. Write-back cache and flush operation but no FUA (forced unit 62iii. Write-back cache and flush operation but no FUA (forced unit
diff --git a/Documentation/block/biodoc.txt b/Documentation/block/biodoc.txt
index f989a9e839b4..34bf8f60d8f8 100644
--- a/Documentation/block/biodoc.txt
+++ b/Documentation/block/biodoc.txt
@@ -135,7 +135,7 @@ Some new queue property settings:
135 Sets two variables that limit the size of the request. 135 Sets two variables that limit the size of the request.
136 136
137 - The request queue's max_sectors, which is a soft size in 137 - The request queue's max_sectors, which is a soft size in
138 in units of 512 byte sectors, and could be dynamically varied 138 units of 512 byte sectors, and could be dynamically varied
139 by the core kernel. 139 by the core kernel.
140 140
141 - The request queue's max_hw_sectors, which is a hard limit 141 - The request queue's max_hw_sectors, which is a hard limit
@@ -783,7 +783,7 @@ all the outstanding requests. There's a third helper to do that:
783 783
784 blk_queue_invalidate_tags(request_queue_t *q) 784 blk_queue_invalidate_tags(request_queue_t *q)
785 785
786 Clear the internal block tag queue and readd all the pending requests 786 Clear the internal block tag queue and re-add all the pending requests
787 to the request queue. The driver will receive them again on the 787 to the request queue. The driver will receive them again on the
788 next request_fn run, just like it did the first time it encountered 788 next request_fn run, just like it did the first time it encountered
789 them. 789 them.
@@ -890,7 +890,7 @@ Aside:
890 890
891 Kvec i/o: 891 Kvec i/o:
892 892
893 Ben LaHaise's aio code uses a slighly different structure instead 893 Ben LaHaise's aio code uses a slightly different structure instead
894 of kiobufs, called a kvec_cb. This contains an array of <page, offset, len> 894 of kiobufs, called a kvec_cb. This contains an array of <page, offset, len>
895 tuples (very much like the networking code), together with a callback function 895 tuples (very much like the networking code), together with a callback function
896 and data pointer. This is embedded into a brw_cb structure when passed 896 and data pointer. This is embedded into a brw_cb structure when passed
@@ -988,7 +988,7 @@ elevator_exit_fn Allocate and free any elevator specific storage
988 for a queue. 988 for a queue.
989 989
9904.2 Request flows seen by I/O schedulers 9904.2 Request flows seen by I/O schedulers
991All requests seens by I/O schedulers strictly follow one of the following three 991All requests seen by I/O schedulers strictly follow one of the following three
992flows. 992flows.
993 993
994 set_req_fn -> 994 set_req_fn ->
@@ -1203,6 +1203,6 @@ temporarily map a bio into the virtual address space.
1203and Linus' comments - Jan 2001) 1203and Linus' comments - Jan 2001)
12049.2 Discussions about kiobuf and bh design on lkml between sct, linus, alan 12049.2 Discussions about kiobuf and bh design on lkml between sct, linus, alan
1205et al - Feb-March 2001 (many of the initial thoughts that led to bio were 1205et al - Feb-March 2001 (many of the initial thoughts that led to bio were
1206brought up in this discusion thread) 1206brought up in this discussion thread)
12079.3 Discussions on mempool on lkml - Dec 2001. 12079.3 Discussions on mempool on lkml - Dec 2001.
1208 1208
diff --git a/Documentation/block/deadline-iosched.txt b/Documentation/block/deadline-iosched.txt
index c918b3a6022d..be08ffd1e9b8 100644
--- a/Documentation/block/deadline-iosched.txt
+++ b/Documentation/block/deadline-iosched.txt
@@ -23,11 +23,11 @@ you can do so by typing:
23read_expire (in ms) 23read_expire (in ms)
24----------- 24-----------
25 25
26The goal of the deadline io scheduler is to attempt to guarentee a start 26The goal of the deadline io scheduler is to attempt to guarantee a start
27service time for a request. As we focus mainly on read latencies, this is 27service time for a request. As we focus mainly on read latencies, this is
28tunable. When a read request first enters the io scheduler, it is assigned 28tunable. When a read request first enters the io scheduler, it is assigned
29a deadline that is the current time + the read_expire value in units of 29a deadline that is the current time + the read_expire value in units of
30miliseconds. 30milliseconds.
31 31
32 32
33write_expire (in ms) 33write_expire (in ms)