diff options
Diffstat (limited to 'block/blk.h')
-rw-r--r-- | block/blk.h | 45 |
1 files changed, 30 insertions, 15 deletions
diff --git a/block/blk.h b/block/blk.h index d6b911ac002c..d6586287adc9 100644 --- a/block/blk.h +++ b/block/blk.h | |||
@@ -18,8 +18,6 @@ int blk_rq_append_bio(struct request_queue *q, struct request *rq, | |||
18 | void blk_dequeue_request(struct request *rq); | 18 | void blk_dequeue_request(struct request *rq); |
19 | void __blk_queue_free_tags(struct request_queue *q); | 19 | void __blk_queue_free_tags(struct request_queue *q); |
20 | 20 | ||
21 | void blk_unplug_work(struct work_struct *work); | ||
22 | void blk_unplug_timeout(unsigned long data); | ||
23 | void blk_rq_timed_out_timer(unsigned long data); | 21 | void blk_rq_timed_out_timer(unsigned long data); |
24 | void blk_delete_timer(struct request *); | 22 | void blk_delete_timer(struct request *); |
25 | void blk_add_timer(struct request *); | 23 | void blk_add_timer(struct request *); |
@@ -34,7 +32,7 @@ enum rq_atomic_flags { | |||
34 | 32 | ||
35 | /* | 33 | /* |
36 | * EH timer and IO completion will both attempt to 'grab' the request, make | 34 | * EH timer and IO completion will both attempt to 'grab' the request, make |
37 | * sure that only one of them suceeds | 35 | * sure that only one of them succeeds |
38 | */ | 36 | */ |
39 | static inline int blk_mark_rq_complete(struct request *rq) | 37 | static inline int blk_mark_rq_complete(struct request *rq) |
40 | { | 38 | { |
@@ -51,18 +49,41 @@ static inline void blk_clear_rq_complete(struct request *rq) | |||
51 | */ | 49 | */ |
52 | #define ELV_ON_HASH(rq) (!hlist_unhashed(&(rq)->hash)) | 50 | #define ELV_ON_HASH(rq) (!hlist_unhashed(&(rq)->hash)) |
53 | 51 | ||
52 | void blk_insert_flush(struct request *rq); | ||
53 | void blk_abort_flushes(struct request_queue *q); | ||
54 | |||
54 | static inline struct request *__elv_next_request(struct request_queue *q) | 55 | static inline struct request *__elv_next_request(struct request_queue *q) |
55 | { | 56 | { |
56 | struct request *rq; | 57 | struct request *rq; |
57 | 58 | ||
58 | while (1) { | 59 | while (1) { |
59 | while (!list_empty(&q->queue_head)) { | 60 | if (!list_empty(&q->queue_head)) { |
60 | rq = list_entry_rq(q->queue_head.next); | 61 | rq = list_entry_rq(q->queue_head.next); |
61 | if (blk_do_ordered(q, &rq)) | 62 | return rq; |
62 | return rq; | ||
63 | } | 63 | } |
64 | 64 | ||
65 | if (!q->elevator->ops->elevator_dispatch_fn(q, 0)) | 65 | /* |
66 | * Flush request is running and flush request isn't queueable | ||
67 | * in the drive, we can hold the queue till flush request is | ||
68 | * finished. Even we don't do this, driver can't dispatch next | ||
69 | * requests and will requeue them. And this can improve | ||
70 | * throughput too. For example, we have request flush1, write1, | ||
71 | * flush 2. flush1 is dispatched, then queue is hold, write1 | ||
72 | * isn't inserted to queue. After flush1 is finished, flush2 | ||
73 | * will be dispatched. Since disk cache is already clean, | ||
74 | * flush2 will be finished very soon, so looks like flush2 is | ||
75 | * folded to flush1. | ||
76 | * Since the queue is hold, a flag is set to indicate the queue | ||
77 | * should be restarted later. Please see flush_end_io() for | ||
78 | * details. | ||
79 | */ | ||
80 | if (q->flush_pending_idx != q->flush_running_idx && | ||
81 | !queue_flush_queueable(q)) { | ||
82 | q->flush_queue_delayed = 1; | ||
83 | return NULL; | ||
84 | } | ||
85 | if (test_bit(QUEUE_FLAG_DEAD, &q->queue_flags) || | ||
86 | !q->elevator->ops->elevator_dispatch_fn(q, 0)) | ||
66 | return NULL; | 87 | return NULL; |
67 | } | 88 | } |
68 | } | 89 | } |
@@ -103,6 +124,8 @@ int ll_front_merge_fn(struct request_queue *q, struct request *req, | |||
103 | struct bio *bio); | 124 | struct bio *bio); |
104 | int attempt_back_merge(struct request_queue *q, struct request *rq); | 125 | int attempt_back_merge(struct request_queue *q, struct request *rq); |
105 | int attempt_front_merge(struct request_queue *q, struct request *rq); | 126 | int attempt_front_merge(struct request_queue *q, struct request *rq); |
127 | int blk_attempt_req_merge(struct request_queue *q, struct request *rq, | ||
128 | struct request *next); | ||
106 | void blk_recalc_rq_segments(struct request *rq); | 129 | void blk_recalc_rq_segments(struct request *rq); |
107 | void blk_rq_set_mixed_merge(struct request *rq); | 130 | void blk_rq_set_mixed_merge(struct request *rq); |
108 | 131 | ||
@@ -132,14 +155,6 @@ static inline int queue_congestion_off_threshold(struct request_queue *q) | |||
132 | return q->nr_congestion_off; | 155 | return q->nr_congestion_off; |
133 | } | 156 | } |
134 | 157 | ||
135 | #if defined(CONFIG_BLK_DEV_INTEGRITY) | ||
136 | |||
137 | #define rq_for_each_integrity_segment(bvl, _rq, _iter) \ | ||
138 | __rq_for_each_bio(_iter.bio, _rq) \ | ||
139 | bip_for_each_vec(bvl, _iter.bio->bi_integrity, _iter.i) | ||
140 | |||
141 | #endif /* BLK_DEV_INTEGRITY */ | ||
142 | |||
143 | static inline int blk_cpu_to_group(int cpu) | 158 | static inline int blk_cpu_to_group(int cpu) |
144 | { | 159 | { |
145 | int group = NR_CPUS; | 160 | int group = NR_CPUS; |