diff options
author | Tejun Heo <tj@kernel.org> | 2008-11-27 23:32:05 -0500 |
---|---|---|
committer | Jens Axboe <jens.axboe@oracle.com> | 2008-12-29 02:28:45 -0500 |
commit | 8f11b3e99a1136fcbb67316c3260f085299c0bff (patch) | |
tree | bb9b12fe23aceac19e24f674786612d0fcad2142 /block/blk-barrier.c | |
parent | f671620e7d895af221bdfeda751d54fa55ed9546 (diff) |
block: make barrier completion more robust
Barrier completion had the following assumptions.
* start_ordered() couldn't finish the whole sequence properly. If all
actions are to be skipped, q->ordseq is set correctly but the actual
completion was never triggered thus hanging the barrier request.
* Drain completion in elv_complete_request() assumed that there's
always at least one request in the queue when drain completes.
Both assumptions are true but these assumptions need to be removed to
improve empty barrier implementation. This patch makes the following
changes.
* Make start_ordered() use blk_ordered_complete_seq() to mark skipped
steps complete and notify __elv_next_request() that it should fetch
the next request if the whole barrier has completed inside
start_ordered().
* Make drain completion path in elv_complete_request() check whether
the queue is empty. Empty queue also indicates drain completion.
* While at it, convert 0/1 return from blk_do_ordered() to false/true.
Signed-off-by: Tejun Heo <tj@kernel.org>
Signed-off-by: Jens Axboe <jens.axboe@oracle.com>
Diffstat (limited to 'block/blk-barrier.c')
-rw-r--r-- | block/blk-barrier.c | 45 |
1 files changed, 27 insertions, 18 deletions
diff --git a/block/blk-barrier.c b/block/blk-barrier.c index 1efabf829c53..b03d88013e1e 100644 --- a/block/blk-barrier.c +++ b/block/blk-barrier.c | |||
@@ -88,7 +88,7 @@ unsigned blk_ordered_req_seq(struct request *rq) | |||
88 | return QUEUE_ORDSEQ_DONE; | 88 | return QUEUE_ORDSEQ_DONE; |
89 | } | 89 | } |
90 | 90 | ||
91 | void blk_ordered_complete_seq(struct request_queue *q, unsigned seq, int error) | 91 | bool blk_ordered_complete_seq(struct request_queue *q, unsigned seq, int error) |
92 | { | 92 | { |
93 | struct request *rq; | 93 | struct request *rq; |
94 | 94 | ||
@@ -99,7 +99,7 @@ void blk_ordered_complete_seq(struct request_queue *q, unsigned seq, int error) | |||
99 | q->ordseq |= seq; | 99 | q->ordseq |= seq; |
100 | 100 | ||
101 | if (blk_ordered_cur_seq(q) != QUEUE_ORDSEQ_DONE) | 101 | if (blk_ordered_cur_seq(q) != QUEUE_ORDSEQ_DONE) |
102 | return; | 102 | return false; |
103 | 103 | ||
104 | /* | 104 | /* |
105 | * Okay, sequence complete. | 105 | * Okay, sequence complete. |
@@ -109,6 +109,8 @@ void blk_ordered_complete_seq(struct request_queue *q, unsigned seq, int error) | |||
109 | 109 | ||
110 | if (__blk_end_request(rq, q->orderr, blk_rq_bytes(rq))) | 110 | if (__blk_end_request(rq, q->orderr, blk_rq_bytes(rq))) |
111 | BUG(); | 111 | BUG(); |
112 | |||
113 | return true; | ||
112 | } | 114 | } |
113 | 115 | ||
114 | static void pre_flush_end_io(struct request *rq, int error) | 116 | static void pre_flush_end_io(struct request *rq, int error) |
@@ -151,9 +153,11 @@ static void queue_flush(struct request_queue *q, unsigned which) | |||
151 | elv_insert(q, rq, ELEVATOR_INSERT_FRONT); | 153 | elv_insert(q, rq, ELEVATOR_INSERT_FRONT); |
152 | } | 154 | } |
153 | 155 | ||
154 | static inline struct request *start_ordered(struct request_queue *q, | 156 | static inline bool start_ordered(struct request_queue *q, struct request **rqp) |
155 | struct request *rq) | ||
156 | { | 157 | { |
158 | struct request *rq = *rqp; | ||
159 | unsigned skip = 0; | ||
160 | |||
157 | q->orderr = 0; | 161 | q->orderr = 0; |
158 | q->ordered = q->next_ordered; | 162 | q->ordered = q->next_ordered; |
159 | q->ordseq |= QUEUE_ORDSEQ_STARTED; | 163 | q->ordseq |= QUEUE_ORDSEQ_STARTED; |
@@ -177,7 +181,7 @@ static inline struct request *start_ordered(struct request_queue *q, | |||
177 | queue_flush(q, QUEUE_ORDERED_DO_POSTFLUSH); | 181 | queue_flush(q, QUEUE_ORDERED_DO_POSTFLUSH); |
178 | rq = &q->post_flush_rq; | 182 | rq = &q->post_flush_rq; |
179 | } else | 183 | } else |
180 | q->ordseq |= QUEUE_ORDSEQ_POSTFLUSH; | 184 | skip |= QUEUE_ORDSEQ_POSTFLUSH; |
181 | 185 | ||
182 | if (q->ordered & QUEUE_ORDERED_DO_BAR) { | 186 | if (q->ordered & QUEUE_ORDERED_DO_BAR) { |
183 | rq = &q->bar_rq; | 187 | rq = &q->bar_rq; |
@@ -193,35 +197,40 @@ static inline struct request *start_ordered(struct request_queue *q, | |||
193 | 197 | ||
194 | elv_insert(q, rq, ELEVATOR_INSERT_FRONT); | 198 | elv_insert(q, rq, ELEVATOR_INSERT_FRONT); |
195 | } else | 199 | } else |
196 | q->ordseq |= QUEUE_ORDSEQ_BAR; | 200 | skip |= QUEUE_ORDSEQ_BAR; |
197 | 201 | ||
198 | if (q->ordered & QUEUE_ORDERED_DO_PREFLUSH) { | 202 | if (q->ordered & QUEUE_ORDERED_DO_PREFLUSH) { |
199 | queue_flush(q, QUEUE_ORDERED_DO_PREFLUSH); | 203 | queue_flush(q, QUEUE_ORDERED_DO_PREFLUSH); |
200 | rq = &q->pre_flush_rq; | 204 | rq = &q->pre_flush_rq; |
201 | } else | 205 | } else |
202 | q->ordseq |= QUEUE_ORDSEQ_PREFLUSH; | 206 | skip |= QUEUE_ORDSEQ_PREFLUSH; |
203 | 207 | ||
204 | if ((q->ordered & QUEUE_ORDERED_BY_DRAIN) && q->in_flight) | 208 | if ((q->ordered & QUEUE_ORDERED_BY_DRAIN) && q->in_flight) |
205 | rq = NULL; | 209 | rq = NULL; |
206 | else | 210 | else |
207 | q->ordseq |= QUEUE_ORDSEQ_DRAIN; | 211 | skip |= QUEUE_ORDSEQ_DRAIN; |
212 | |||
213 | *rqp = rq; | ||
208 | 214 | ||
209 | return rq; | 215 | /* |
216 | * Complete skipped sequences. If whole sequence is complete, | ||
217 | * return false to tell elevator that this request is gone. | ||
218 | */ | ||
219 | return !blk_ordered_complete_seq(q, skip, 0); | ||
210 | } | 220 | } |
211 | 221 | ||
212 | int blk_do_ordered(struct request_queue *q, struct request **rqp) | 222 | bool blk_do_ordered(struct request_queue *q, struct request **rqp) |
213 | { | 223 | { |
214 | struct request *rq = *rqp; | 224 | struct request *rq = *rqp; |
215 | const int is_barrier = blk_fs_request(rq) && blk_barrier_rq(rq); | 225 | const int is_barrier = blk_fs_request(rq) && blk_barrier_rq(rq); |
216 | 226 | ||
217 | if (!q->ordseq) { | 227 | if (!q->ordseq) { |
218 | if (!is_barrier) | 228 | if (!is_barrier) |
219 | return 1; | 229 | return true; |
220 | 230 | ||
221 | if (q->next_ordered != QUEUE_ORDERED_NONE) { | 231 | if (q->next_ordered != QUEUE_ORDERED_NONE) |
222 | *rqp = start_ordered(q, rq); | 232 | return start_ordered(q, rqp); |
223 | return 1; | 233 | else { |
224 | } else { | ||
225 | /* | 234 | /* |
226 | * Queue ordering not supported. Terminate | 235 | * Queue ordering not supported. Terminate |
227 | * with prejudice. | 236 | * with prejudice. |
@@ -231,7 +240,7 @@ int blk_do_ordered(struct request_queue *q, struct request **rqp) | |||
231 | blk_rq_bytes(rq))) | 240 | blk_rq_bytes(rq))) |
232 | BUG(); | 241 | BUG(); |
233 | *rqp = NULL; | 242 | *rqp = NULL; |
234 | return 0; | 243 | return false; |
235 | } | 244 | } |
236 | } | 245 | } |
237 | 246 | ||
@@ -242,7 +251,7 @@ int blk_do_ordered(struct request_queue *q, struct request **rqp) | |||
242 | /* Special requests are not subject to ordering rules. */ | 251 | /* Special requests are not subject to ordering rules. */ |
243 | if (!blk_fs_request(rq) && | 252 | if (!blk_fs_request(rq) && |
244 | rq != &q->pre_flush_rq && rq != &q->post_flush_rq) | 253 | rq != &q->pre_flush_rq && rq != &q->post_flush_rq) |
245 | return 1; | 254 | return true; |
246 | 255 | ||
247 | if (q->ordered & QUEUE_ORDERED_BY_TAG) { | 256 | if (q->ordered & QUEUE_ORDERED_BY_TAG) { |
248 | /* Ordered by tag. Blocking the next barrier is enough. */ | 257 | /* Ordered by tag. Blocking the next barrier is enough. */ |
@@ -255,7 +264,7 @@ int blk_do_ordered(struct request_queue *q, struct request **rqp) | |||
255 | *rqp = NULL; | 264 | *rqp = NULL; |
256 | } | 265 | } |
257 | 266 | ||
258 | return 1; | 267 | return true; |
259 | } | 268 | } |
260 | 269 | ||
261 | static void bio_end_empty_barrier(struct bio *bio, int err) | 270 | static void bio_end_empty_barrier(struct bio *bio, int err) |