diff options
author | Jens Axboe <jens.axboe@oracle.com> | 2008-01-29 07:54:41 -0500 |
---|---|---|
committer | Jens Axboe <jens.axboe@oracle.com> | 2008-01-29 15:55:09 -0500 |
commit | 26b8256e2bb930a8e4d4d10aa74950d8921376b8 (patch) | |
tree | 36fc1011aa68526dc1fb5b237e330ca2c27c9939 | |
parent | 86db1e29772372155db08ff48a9ceb76e11a2ad1 (diff) |
block: get rid of unnecessary forward declarations in blk-core.c
Signed-off-by: Jens Axboe <jens.axboe@oracle.com>
-rw-r--r-- | block/blk-core.c | 62 |
1 files changed, 30 insertions, 32 deletions
diff --git a/block/blk-core.c b/block/blk-core.c index 2c73ed1a8131..3d415ec10fb8 100644 --- a/block/blk-core.c +++ b/block/blk-core.c | |||
@@ -33,9 +33,7 @@ | |||
33 | 33 | ||
34 | #include "blk.h" | 34 | #include "blk.h" |
35 | 35 | ||
36 | static void drive_stat_acct(struct request *rq, int new_io); | ||
37 | static int __make_request(struct request_queue *q, struct bio *bio); | 36 | static int __make_request(struct request_queue *q, struct bio *bio); |
38 | static void blk_recalc_rq_segments(struct request *rq); | ||
39 | 37 | ||
40 | /* | 38 | /* |
41 | * For the allocated request tables | 39 | * For the allocated request tables |
@@ -54,6 +52,21 @@ static struct workqueue_struct *kblockd_workqueue; | |||
54 | 52 | ||
55 | static DEFINE_PER_CPU(struct list_head, blk_cpu_done); | 53 | static DEFINE_PER_CPU(struct list_head, blk_cpu_done); |
56 | 54 | ||
55 | static void drive_stat_acct(struct request *rq, int new_io) | ||
56 | { | ||
57 | int rw = rq_data_dir(rq); | ||
58 | |||
59 | if (!blk_fs_request(rq) || !rq->rq_disk) | ||
60 | return; | ||
61 | |||
62 | if (!new_io) { | ||
63 | __disk_stat_inc(rq->rq_disk, merges[rw]); | ||
64 | } else { | ||
65 | disk_round_stats(rq->rq_disk); | ||
66 | rq->rq_disk->in_flight++; | ||
67 | } | ||
68 | } | ||
69 | |||
57 | void blk_queue_congestion_threshold(struct request_queue *q) | 70 | void blk_queue_congestion_threshold(struct request_queue *q) |
58 | { | 71 | { |
59 | int nr; | 72 | int nr; |
@@ -168,21 +181,6 @@ void blk_dump_rq_flags(struct request *rq, char *msg) | |||
168 | 181 | ||
169 | EXPORT_SYMBOL(blk_dump_rq_flags); | 182 | EXPORT_SYMBOL(blk_dump_rq_flags); |
170 | 183 | ||
171 | void blk_recount_segments(struct request_queue *q, struct bio *bio) | ||
172 | { | ||
173 | struct request rq; | ||
174 | struct bio *nxt = bio->bi_next; | ||
175 | rq.q = q; | ||
176 | rq.bio = rq.biotail = bio; | ||
177 | bio->bi_next = NULL; | ||
178 | blk_recalc_rq_segments(&rq); | ||
179 | bio->bi_next = nxt; | ||
180 | bio->bi_phys_segments = rq.nr_phys_segments; | ||
181 | bio->bi_hw_segments = rq.nr_hw_segments; | ||
182 | bio->bi_flags |= (1 << BIO_SEG_VALID); | ||
183 | } | ||
184 | EXPORT_SYMBOL(blk_recount_segments); | ||
185 | |||
186 | static void blk_recalc_rq_segments(struct request *rq) | 184 | static void blk_recalc_rq_segments(struct request *rq) |
187 | { | 185 | { |
188 | int nr_phys_segs; | 186 | int nr_phys_segs; |
@@ -255,6 +253,21 @@ new_hw_segment: | |||
255 | rq->nr_hw_segments = nr_hw_segs; | 253 | rq->nr_hw_segments = nr_hw_segs; |
256 | } | 254 | } |
257 | 255 | ||
256 | void blk_recount_segments(struct request_queue *q, struct bio *bio) | ||
257 | { | ||
258 | struct request rq; | ||
259 | struct bio *nxt = bio->bi_next; | ||
260 | rq.q = q; | ||
261 | rq.bio = rq.biotail = bio; | ||
262 | bio->bi_next = NULL; | ||
263 | blk_recalc_rq_segments(&rq); | ||
264 | bio->bi_next = nxt; | ||
265 | bio->bi_phys_segments = rq.nr_phys_segments; | ||
266 | bio->bi_hw_segments = rq.nr_hw_segments; | ||
267 | bio->bi_flags |= (1 << BIO_SEG_VALID); | ||
268 | } | ||
269 | EXPORT_SYMBOL(blk_recount_segments); | ||
270 | |||
258 | static int blk_phys_contig_segment(struct request_queue *q, struct bio *bio, | 271 | static int blk_phys_contig_segment(struct request_queue *q, struct bio *bio, |
259 | struct bio *nxt) | 272 | struct bio *nxt) |
260 | { | 273 | { |
@@ -1305,21 +1318,6 @@ void blk_insert_request(struct request_queue *q, struct request *rq, | |||
1305 | 1318 | ||
1306 | EXPORT_SYMBOL(blk_insert_request); | 1319 | EXPORT_SYMBOL(blk_insert_request); |
1307 | 1320 | ||
1308 | static void drive_stat_acct(struct request *rq, int new_io) | ||
1309 | { | ||
1310 | int rw = rq_data_dir(rq); | ||
1311 | |||
1312 | if (!blk_fs_request(rq) || !rq->rq_disk) | ||
1313 | return; | ||
1314 | |||
1315 | if (!new_io) { | ||
1316 | __disk_stat_inc(rq->rq_disk, merges[rw]); | ||
1317 | } else { | ||
1318 | disk_round_stats(rq->rq_disk); | ||
1319 | rq->rq_disk->in_flight++; | ||
1320 | } | ||
1321 | } | ||
1322 | |||
1323 | /* | 1321 | /* |
1324 | * add-request adds a request to the linked list. | 1322 | * add-request adds a request to the linked list. |
1325 | * queue lock is held and interrupts disabled, as we muck with the | 1323 | * queue lock is held and interrupts disabled, as we muck with the |