diff options
author | NeilBrown <neilb@suse.de> | 2011-10-11 01:48:43 -0400 |
---|---|---|
committer | NeilBrown <neilb@suse.de> | 2011-10-11 01:48:43 -0400 |
commit | 9f2c9d12bcc53fcb3b787023723754e84d1aef8b (patch) | |
tree | 85fbc63c6e480bdebe06529e1af4d3d483d70452 /drivers/md/raid10.c | |
parent | 2b8bf3451d1e3133ebc3998721d14013a6c27114 (diff) |
md: remove typedefs: r10bio_t -> struct r10bio and r1bio_t -> struct r1bio
Signed-off-by: NeilBrown <neilb@suse.de>
Diffstat (limited to 'drivers/md/raid10.c')
-rw-r--r-- | drivers/md/raid10.c | 68 |
1 files changed, 34 insertions, 34 deletions
diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c index 8c3bd6061f96..6927998e253a 100644 --- a/drivers/md/raid10.c +++ b/drivers/md/raid10.c | |||
@@ -64,7 +64,7 @@ static void lower_barrier(conf_t *conf); | |||
64 | static void * r10bio_pool_alloc(gfp_t gfp_flags, void *data) | 64 | static void * r10bio_pool_alloc(gfp_t gfp_flags, void *data) |
65 | { | 65 | { |
66 | conf_t *conf = data; | 66 | conf_t *conf = data; |
67 | int size = offsetof(struct r10bio_s, devs[conf->copies]); | 67 | int size = offsetof(struct r10bio, devs[conf->copies]); |
68 | 68 | ||
69 | /* allocate a r10bio with room for raid_disks entries in the bios array */ | 69 | /* allocate a r10bio with room for raid_disks entries in the bios array */ |
70 | return kzalloc(size, gfp_flags); | 70 | return kzalloc(size, gfp_flags); |
@@ -94,7 +94,7 @@ static void * r10buf_pool_alloc(gfp_t gfp_flags, void *data) | |||
94 | { | 94 | { |
95 | conf_t *conf = data; | 95 | conf_t *conf = data; |
96 | struct page *page; | 96 | struct page *page; |
97 | r10bio_t *r10_bio; | 97 | struct r10bio *r10_bio; |
98 | struct bio *bio; | 98 | struct bio *bio; |
99 | int i, j; | 99 | int i, j; |
100 | int nalloc; | 100 | int nalloc; |
@@ -159,7 +159,7 @@ static void r10buf_pool_free(void *__r10_bio, void *data) | |||
159 | { | 159 | { |
160 | int i; | 160 | int i; |
161 | conf_t *conf = data; | 161 | conf_t *conf = data; |
162 | r10bio_t *r10bio = __r10_bio; | 162 | struct r10bio *r10bio = __r10_bio; |
163 | int j; | 163 | int j; |
164 | 164 | ||
165 | for (j=0; j < conf->copies; j++) { | 165 | for (j=0; j < conf->copies; j++) { |
@@ -175,7 +175,7 @@ static void r10buf_pool_free(void *__r10_bio, void *data) | |||
175 | r10bio_pool_free(r10bio, conf); | 175 | r10bio_pool_free(r10bio, conf); |
176 | } | 176 | } |
177 | 177 | ||
178 | static void put_all_bios(conf_t *conf, r10bio_t *r10_bio) | 178 | static void put_all_bios(conf_t *conf, struct r10bio *r10_bio) |
179 | { | 179 | { |
180 | int i; | 180 | int i; |
181 | 181 | ||
@@ -187,7 +187,7 @@ static void put_all_bios(conf_t *conf, r10bio_t *r10_bio) | |||
187 | } | 187 | } |
188 | } | 188 | } |
189 | 189 | ||
190 | static void free_r10bio(r10bio_t *r10_bio) | 190 | static void free_r10bio(struct r10bio *r10_bio) |
191 | { | 191 | { |
192 | conf_t *conf = r10_bio->mddev->private; | 192 | conf_t *conf = r10_bio->mddev->private; |
193 | 193 | ||
@@ -195,7 +195,7 @@ static void free_r10bio(r10bio_t *r10_bio) | |||
195 | mempool_free(r10_bio, conf->r10bio_pool); | 195 | mempool_free(r10_bio, conf->r10bio_pool); |
196 | } | 196 | } |
197 | 197 | ||
198 | static void put_buf(r10bio_t *r10_bio) | 198 | static void put_buf(struct r10bio *r10_bio) |
199 | { | 199 | { |
200 | conf_t *conf = r10_bio->mddev->private; | 200 | conf_t *conf = r10_bio->mddev->private; |
201 | 201 | ||
@@ -204,7 +204,7 @@ static void put_buf(r10bio_t *r10_bio) | |||
204 | lower_barrier(conf); | 204 | lower_barrier(conf); |
205 | } | 205 | } |
206 | 206 | ||
207 | static void reschedule_retry(r10bio_t *r10_bio) | 207 | static void reschedule_retry(struct r10bio *r10_bio) |
208 | { | 208 | { |
209 | unsigned long flags; | 209 | unsigned long flags; |
210 | struct mddev *mddev = r10_bio->mddev; | 210 | struct mddev *mddev = r10_bio->mddev; |
@@ -226,7 +226,7 @@ static void reschedule_retry(r10bio_t *r10_bio) | |||
226 | * operation and are ready to return a success/failure code to the buffer | 226 | * operation and are ready to return a success/failure code to the buffer |
227 | * cache layer. | 227 | * cache layer. |
228 | */ | 228 | */ |
229 | static void raid_end_bio_io(r10bio_t *r10_bio) | 229 | static void raid_end_bio_io(struct r10bio *r10_bio) |
230 | { | 230 | { |
231 | struct bio *bio = r10_bio->master_bio; | 231 | struct bio *bio = r10_bio->master_bio; |
232 | int done; | 232 | int done; |
@@ -256,7 +256,7 @@ static void raid_end_bio_io(r10bio_t *r10_bio) | |||
256 | /* | 256 | /* |
257 | * Update disk head position estimator based on IRQ completion info. | 257 | * Update disk head position estimator based on IRQ completion info. |
258 | */ | 258 | */ |
259 | static inline void update_head_pos(int slot, r10bio_t *r10_bio) | 259 | static inline void update_head_pos(int slot, struct r10bio *r10_bio) |
260 | { | 260 | { |
261 | conf_t *conf = r10_bio->mddev->private; | 261 | conf_t *conf = r10_bio->mddev->private; |
262 | 262 | ||
@@ -267,7 +267,7 @@ static inline void update_head_pos(int slot, r10bio_t *r10_bio) | |||
267 | /* | 267 | /* |
268 | * Find the disk number which triggered given bio | 268 | * Find the disk number which triggered given bio |
269 | */ | 269 | */ |
270 | static int find_bio_disk(conf_t *conf, r10bio_t *r10_bio, | 270 | static int find_bio_disk(conf_t *conf, struct r10bio *r10_bio, |
271 | struct bio *bio, int *slotp) | 271 | struct bio *bio, int *slotp) |
272 | { | 272 | { |
273 | int slot; | 273 | int slot; |
@@ -287,7 +287,7 @@ static int find_bio_disk(conf_t *conf, r10bio_t *r10_bio, | |||
287 | static void raid10_end_read_request(struct bio *bio, int error) | 287 | static void raid10_end_read_request(struct bio *bio, int error) |
288 | { | 288 | { |
289 | int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags); | 289 | int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags); |
290 | r10bio_t *r10_bio = bio->bi_private; | 290 | struct r10bio *r10_bio = bio->bi_private; |
291 | int slot, dev; | 291 | int slot, dev; |
292 | conf_t *conf = r10_bio->mddev->private; | 292 | conf_t *conf = r10_bio->mddev->private; |
293 | 293 | ||
@@ -327,7 +327,7 @@ static void raid10_end_read_request(struct bio *bio, int error) | |||
327 | } | 327 | } |
328 | } | 328 | } |
329 | 329 | ||
330 | static void close_write(r10bio_t *r10_bio) | 330 | static void close_write(struct r10bio *r10_bio) |
331 | { | 331 | { |
332 | /* clear the bitmap if all writes complete successfully */ | 332 | /* clear the bitmap if all writes complete successfully */ |
333 | bitmap_endwrite(r10_bio->mddev->bitmap, r10_bio->sector, | 333 | bitmap_endwrite(r10_bio->mddev->bitmap, r10_bio->sector, |
@@ -337,7 +337,7 @@ static void close_write(r10bio_t *r10_bio) | |||
337 | md_write_end(r10_bio->mddev); | 337 | md_write_end(r10_bio->mddev); |
338 | } | 338 | } |
339 | 339 | ||
340 | static void one_write_done(r10bio_t *r10_bio) | 340 | static void one_write_done(struct r10bio *r10_bio) |
341 | { | 341 | { |
342 | if (atomic_dec_and_test(&r10_bio->remaining)) { | 342 | if (atomic_dec_and_test(&r10_bio->remaining)) { |
343 | if (test_bit(R10BIO_WriteError, &r10_bio->state)) | 343 | if (test_bit(R10BIO_WriteError, &r10_bio->state)) |
@@ -355,7 +355,7 @@ static void one_write_done(r10bio_t *r10_bio) | |||
355 | static void raid10_end_write_request(struct bio *bio, int error) | 355 | static void raid10_end_write_request(struct bio *bio, int error) |
356 | { | 356 | { |
357 | int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags); | 357 | int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags); |
358 | r10bio_t *r10_bio = bio->bi_private; | 358 | struct r10bio *r10_bio = bio->bi_private; |
359 | int dev; | 359 | int dev; |
360 | int dec_rdev = 1; | 360 | int dec_rdev = 1; |
361 | conf_t *conf = r10_bio->mddev->private; | 361 | conf_t *conf = r10_bio->mddev->private; |
@@ -433,7 +433,7 @@ static void raid10_end_write_request(struct bio *bio, int error) | |||
433 | * sector offset to a virtual address | 433 | * sector offset to a virtual address |
434 | */ | 434 | */ |
435 | 435 | ||
436 | static void raid10_find_phys(conf_t *conf, r10bio_t *r10bio) | 436 | static void raid10_find_phys(conf_t *conf, struct r10bio *r10bio) |
437 | { | 437 | { |
438 | int n,f; | 438 | int n,f; |
439 | sector_t sector; | 439 | sector_t sector; |
@@ -555,7 +555,7 @@ static int raid10_mergeable_bvec(struct request_queue *q, | |||
555 | * FIXME: possibly should rethink readbalancing and do it differently | 555 | * FIXME: possibly should rethink readbalancing and do it differently |
556 | * depending on near_copies / far_copies geometry. | 556 | * depending on near_copies / far_copies geometry. |
557 | */ | 557 | */ |
558 | static int read_balance(conf_t *conf, r10bio_t *r10_bio, int *max_sectors) | 558 | static int read_balance(conf_t *conf, struct r10bio *r10_bio, int *max_sectors) |
559 | { | 559 | { |
560 | const sector_t this_sector = r10_bio->sector; | 560 | const sector_t this_sector = r10_bio->sector; |
561 | int disk, slot; | 561 | int disk, slot; |
@@ -834,7 +834,7 @@ static int make_request(struct mddev *mddev, struct bio * bio) | |||
834 | { | 834 | { |
835 | conf_t *conf = mddev->private; | 835 | conf_t *conf = mddev->private; |
836 | mirror_info_t *mirror; | 836 | mirror_info_t *mirror; |
837 | r10bio_t *r10_bio; | 837 | struct r10bio *r10_bio; |
838 | struct bio *read_bio; | 838 | struct bio *read_bio; |
839 | int i; | 839 | int i; |
840 | int chunk_sects = conf->chunk_mask + 1; | 840 | int chunk_sects = conf->chunk_mask + 1; |
@@ -1411,7 +1411,7 @@ abort: | |||
1411 | 1411 | ||
1412 | static void end_sync_read(struct bio *bio, int error) | 1412 | static void end_sync_read(struct bio *bio, int error) |
1413 | { | 1413 | { |
1414 | r10bio_t *r10_bio = bio->bi_private; | 1414 | struct r10bio *r10_bio = bio->bi_private; |
1415 | conf_t *conf = r10_bio->mddev->private; | 1415 | conf_t *conf = r10_bio->mddev->private; |
1416 | int d; | 1416 | int d; |
1417 | 1417 | ||
@@ -1439,7 +1439,7 @@ static void end_sync_read(struct bio *bio, int error) | |||
1439 | } | 1439 | } |
1440 | } | 1440 | } |
1441 | 1441 | ||
1442 | static void end_sync_request(r10bio_t *r10_bio) | 1442 | static void end_sync_request(struct r10bio *r10_bio) |
1443 | { | 1443 | { |
1444 | struct mddev *mddev = r10_bio->mddev; | 1444 | struct mddev *mddev = r10_bio->mddev; |
1445 | 1445 | ||
@@ -1455,7 +1455,7 @@ static void end_sync_request(r10bio_t *r10_bio) | |||
1455 | md_done_sync(mddev, s, 1); | 1455 | md_done_sync(mddev, s, 1); |
1456 | break; | 1456 | break; |
1457 | } else { | 1457 | } else { |
1458 | r10bio_t *r10_bio2 = (r10bio_t *)r10_bio->master_bio; | 1458 | struct r10bio *r10_bio2 = (struct r10bio *)r10_bio->master_bio; |
1459 | if (test_bit(R10BIO_MadeGood, &r10_bio->state) || | 1459 | if (test_bit(R10BIO_MadeGood, &r10_bio->state) || |
1460 | test_bit(R10BIO_WriteError, &r10_bio->state)) | 1460 | test_bit(R10BIO_WriteError, &r10_bio->state)) |
1461 | reschedule_retry(r10_bio); | 1461 | reschedule_retry(r10_bio); |
@@ -1469,7 +1469,7 @@ static void end_sync_request(r10bio_t *r10_bio) | |||
1469 | static void end_sync_write(struct bio *bio, int error) | 1469 | static void end_sync_write(struct bio *bio, int error) |
1470 | { | 1470 | { |
1471 | int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags); | 1471 | int uptodate = test_bit(BIO_UPTODATE, &bio->bi_flags); |
1472 | r10bio_t *r10_bio = bio->bi_private; | 1472 | struct r10bio *r10_bio = bio->bi_private; |
1473 | struct mddev *mddev = r10_bio->mddev; | 1473 | struct mddev *mddev = r10_bio->mddev; |
1474 | conf_t *conf = mddev->private; | 1474 | conf_t *conf = mddev->private; |
1475 | int d; | 1475 | int d; |
@@ -1509,7 +1509,7 @@ static void end_sync_write(struct bio *bio, int error) | |||
1509 | * We check if all blocks are in-sync and only write to blocks that | 1509 | * We check if all blocks are in-sync and only write to blocks that |
1510 | * aren't in sync | 1510 | * aren't in sync |
1511 | */ | 1511 | */ |
1512 | static void sync_request_write(struct mddev *mddev, r10bio_t *r10_bio) | 1512 | static void sync_request_write(struct mddev *mddev, struct r10bio *r10_bio) |
1513 | { | 1513 | { |
1514 | conf_t *conf = mddev->private; | 1514 | conf_t *conf = mddev->private; |
1515 | int i, first; | 1515 | int i, first; |
@@ -1609,7 +1609,7 @@ done: | |||
1609 | * The second for writing. | 1609 | * The second for writing. |
1610 | * | 1610 | * |
1611 | */ | 1611 | */ |
1612 | static void fix_recovery_read_error(r10bio_t *r10_bio) | 1612 | static void fix_recovery_read_error(struct r10bio *r10_bio) |
1613 | { | 1613 | { |
1614 | /* We got a read error during recovery. | 1614 | /* We got a read error during recovery. |
1615 | * We repeat the read in smaller page-sized sections. | 1615 | * We repeat the read in smaller page-sized sections. |
@@ -1688,7 +1688,7 @@ static void fix_recovery_read_error(r10bio_t *r10_bio) | |||
1688 | } | 1688 | } |
1689 | } | 1689 | } |
1690 | 1690 | ||
1691 | static void recovery_request_write(struct mddev *mddev, r10bio_t *r10_bio) | 1691 | static void recovery_request_write(struct mddev *mddev, struct r10bio *r10_bio) |
1692 | { | 1692 | { |
1693 | conf_t *conf = mddev->private; | 1693 | conf_t *conf = mddev->private; |
1694 | int d; | 1694 | int d; |
@@ -1778,7 +1778,7 @@ static int r10_sync_page_io(struct md_rdev *rdev, sector_t sector, | |||
1778 | * 3. Performs writes following reads for array synchronising. | 1778 | * 3. Performs writes following reads for array synchronising. |
1779 | */ | 1779 | */ |
1780 | 1780 | ||
1781 | static void fix_read_error(conf_t *conf, struct mddev *mddev, r10bio_t *r10_bio) | 1781 | static void fix_read_error(conf_t *conf, struct mddev *mddev, struct r10bio *r10_bio) |
1782 | { | 1782 | { |
1783 | int sect = 0; /* Offset from r10_bio->sector */ | 1783 | int sect = 0; /* Offset from r10_bio->sector */ |
1784 | int sectors = r10_bio->sectors; | 1784 | int sectors = r10_bio->sectors; |
@@ -1983,7 +1983,7 @@ static int submit_bio_wait(int rw, struct bio *bio) | |||
1983 | return test_bit(BIO_UPTODATE, &bio->bi_flags); | 1983 | return test_bit(BIO_UPTODATE, &bio->bi_flags); |
1984 | } | 1984 | } |
1985 | 1985 | ||
1986 | static int narrow_write_error(r10bio_t *r10_bio, int i) | 1986 | static int narrow_write_error(struct r10bio *r10_bio, int i) |
1987 | { | 1987 | { |
1988 | struct bio *bio = r10_bio->master_bio; | 1988 | struct bio *bio = r10_bio->master_bio; |
1989 | struct mddev *mddev = r10_bio->mddev; | 1989 | struct mddev *mddev = r10_bio->mddev; |
@@ -2040,7 +2040,7 @@ static int narrow_write_error(r10bio_t *r10_bio, int i) | |||
2040 | return ok; | 2040 | return ok; |
2041 | } | 2041 | } |
2042 | 2042 | ||
2043 | static void handle_read_error(struct mddev *mddev, r10bio_t *r10_bio) | 2043 | static void handle_read_error(struct mddev *mddev, struct r10bio *r10_bio) |
2044 | { | 2044 | { |
2045 | int slot = r10_bio->read_slot; | 2045 | int slot = r10_bio->read_slot; |
2046 | int mirror = r10_bio->devs[slot].devnum; | 2046 | int mirror = r10_bio->devs[slot].devnum; |
@@ -2139,7 +2139,7 @@ read_more: | |||
2139 | generic_make_request(bio); | 2139 | generic_make_request(bio); |
2140 | } | 2140 | } |
2141 | 2141 | ||
2142 | static void handle_write_completed(conf_t *conf, r10bio_t *r10_bio) | 2142 | static void handle_write_completed(conf_t *conf, struct r10bio *r10_bio) |
2143 | { | 2143 | { |
2144 | /* Some sort of write request has finished and it | 2144 | /* Some sort of write request has finished and it |
2145 | * succeeded in writing where we thought there was a | 2145 | * succeeded in writing where we thought there was a |
@@ -2202,7 +2202,7 @@ static void handle_write_completed(conf_t *conf, r10bio_t *r10_bio) | |||
2202 | 2202 | ||
2203 | static void raid10d(struct mddev *mddev) | 2203 | static void raid10d(struct mddev *mddev) |
2204 | { | 2204 | { |
2205 | r10bio_t *r10_bio; | 2205 | struct r10bio *r10_bio; |
2206 | unsigned long flags; | 2206 | unsigned long flags; |
2207 | conf_t *conf = mddev->private; | 2207 | conf_t *conf = mddev->private; |
2208 | struct list_head *head = &conf->retry_list; | 2208 | struct list_head *head = &conf->retry_list; |
@@ -2220,7 +2220,7 @@ static void raid10d(struct mddev *mddev) | |||
2220 | spin_unlock_irqrestore(&conf->device_lock, flags); | 2220 | spin_unlock_irqrestore(&conf->device_lock, flags); |
2221 | break; | 2221 | break; |
2222 | } | 2222 | } |
2223 | r10_bio = list_entry(head->prev, r10bio_t, retry_list); | 2223 | r10_bio = list_entry(head->prev, struct r10bio, retry_list); |
2224 | list_del(head->prev); | 2224 | list_del(head->prev); |
2225 | conf->nr_queued--; | 2225 | conf->nr_queued--; |
2226 | spin_unlock_irqrestore(&conf->device_lock, flags); | 2226 | spin_unlock_irqrestore(&conf->device_lock, flags); |
@@ -2301,7 +2301,7 @@ static sector_t sync_request(struct mddev *mddev, sector_t sector_nr, | |||
2301 | int *skipped, int go_faster) | 2301 | int *skipped, int go_faster) |
2302 | { | 2302 | { |
2303 | conf_t *conf = mddev->private; | 2303 | conf_t *conf = mddev->private; |
2304 | r10bio_t *r10_bio; | 2304 | struct r10bio *r10_bio; |
2305 | struct bio *biolist = NULL, *bio; | 2305 | struct bio *biolist = NULL, *bio; |
2306 | sector_t max_sector, nr_sectors; | 2306 | sector_t max_sector, nr_sectors; |
2307 | int i; | 2307 | int i; |
@@ -2393,7 +2393,7 @@ static sector_t sync_request(struct mddev *mddev, sector_t sector_nr, | |||
2393 | 2393 | ||
2394 | for (i=0 ; i<conf->raid_disks; i++) { | 2394 | for (i=0 ; i<conf->raid_disks; i++) { |
2395 | int still_degraded; | 2395 | int still_degraded; |
2396 | r10bio_t *rb2; | 2396 | struct r10bio *rb2; |
2397 | sector_t sect; | 2397 | sector_t sect; |
2398 | int must_sync; | 2398 | int must_sync; |
2399 | int any_working; | 2399 | int any_working; |
@@ -2547,8 +2547,8 @@ static sector_t sync_request(struct mddev *mddev, sector_t sector_nr, | |||
2547 | } | 2547 | } |
2548 | if (biolist == NULL) { | 2548 | if (biolist == NULL) { |
2549 | while (r10_bio) { | 2549 | while (r10_bio) { |
2550 | r10bio_t *rb2 = r10_bio; | 2550 | struct r10bio *rb2 = r10_bio; |
2551 | r10_bio = (r10bio_t*) rb2->master_bio; | 2551 | r10_bio = (struct r10bio*) rb2->master_bio; |
2552 | rb2->master_bio = NULL; | 2552 | rb2->master_bio = NULL; |
2553 | put_buf(rb2); | 2553 | put_buf(rb2); |
2554 | } | 2554 | } |