diff options
Diffstat (limited to 'drivers/md/raid5.c')
| -rw-r--r-- | drivers/md/raid5.c | 18 |
1 files changed, 9 insertions, 9 deletions
diff --git a/drivers/md/raid5.c b/drivers/md/raid5.c index d90ee145effe..2aff4be35dc4 100644 --- a/drivers/md/raid5.c +++ b/drivers/md/raid5.c | |||
| @@ -289,7 +289,7 @@ static struct stripe_head *__find_stripe(raid5_conf_t *conf, sector_t sector, in | |||
| 289 | } | 289 | } |
| 290 | 290 | ||
| 291 | static void unplug_slaves(mddev_t *mddev); | 291 | static void unplug_slaves(mddev_t *mddev); |
| 292 | static void raid5_unplug_device(request_queue_t *q); | 292 | static void raid5_unplug_device(struct request_queue *q); |
| 293 | 293 | ||
| 294 | static struct stripe_head *get_active_stripe(raid5_conf_t *conf, sector_t sector, int disks, | 294 | static struct stripe_head *get_active_stripe(raid5_conf_t *conf, sector_t sector, int disks, |
| 295 | int pd_idx, int noblock) | 295 | int pd_idx, int noblock) |
| @@ -3182,7 +3182,7 @@ static void unplug_slaves(mddev_t *mddev) | |||
| 3182 | for (i=0; i<mddev->raid_disks; i++) { | 3182 | for (i=0; i<mddev->raid_disks; i++) { |
| 3183 | mdk_rdev_t *rdev = rcu_dereference(conf->disks[i].rdev); | 3183 | mdk_rdev_t *rdev = rcu_dereference(conf->disks[i].rdev); |
| 3184 | if (rdev && !test_bit(Faulty, &rdev->flags) && atomic_read(&rdev->nr_pending)) { | 3184 | if (rdev && !test_bit(Faulty, &rdev->flags) && atomic_read(&rdev->nr_pending)) { |
| 3185 | request_queue_t *r_queue = bdev_get_queue(rdev->bdev); | 3185 | struct request_queue *r_queue = bdev_get_queue(rdev->bdev); |
| 3186 | 3186 | ||
| 3187 | atomic_inc(&rdev->nr_pending); | 3187 | atomic_inc(&rdev->nr_pending); |
| 3188 | rcu_read_unlock(); | 3188 | rcu_read_unlock(); |
| @@ -3197,7 +3197,7 @@ static void unplug_slaves(mddev_t *mddev) | |||
| 3197 | rcu_read_unlock(); | 3197 | rcu_read_unlock(); |
| 3198 | } | 3198 | } |
| 3199 | 3199 | ||
| 3200 | static void raid5_unplug_device(request_queue_t *q) | 3200 | static void raid5_unplug_device(struct request_queue *q) |
| 3201 | { | 3201 | { |
| 3202 | mddev_t *mddev = q->queuedata; | 3202 | mddev_t *mddev = q->queuedata; |
| 3203 | raid5_conf_t *conf = mddev_to_conf(mddev); | 3203 | raid5_conf_t *conf = mddev_to_conf(mddev); |
| @@ -3216,7 +3216,7 @@ static void raid5_unplug_device(request_queue_t *q) | |||
| 3216 | unplug_slaves(mddev); | 3216 | unplug_slaves(mddev); |
| 3217 | } | 3217 | } |
| 3218 | 3218 | ||
| 3219 | static int raid5_issue_flush(request_queue_t *q, struct gendisk *disk, | 3219 | static int raid5_issue_flush(struct request_queue *q, struct gendisk *disk, |
| 3220 | sector_t *error_sector) | 3220 | sector_t *error_sector) |
| 3221 | { | 3221 | { |
| 3222 | mddev_t *mddev = q->queuedata; | 3222 | mddev_t *mddev = q->queuedata; |
| @@ -3228,7 +3228,7 @@ static int raid5_issue_flush(request_queue_t *q, struct gendisk *disk, | |||
| 3228 | mdk_rdev_t *rdev = rcu_dereference(conf->disks[i].rdev); | 3228 | mdk_rdev_t *rdev = rcu_dereference(conf->disks[i].rdev); |
| 3229 | if (rdev && !test_bit(Faulty, &rdev->flags)) { | 3229 | if (rdev && !test_bit(Faulty, &rdev->flags)) { |
| 3230 | struct block_device *bdev = rdev->bdev; | 3230 | struct block_device *bdev = rdev->bdev; |
| 3231 | request_queue_t *r_queue = bdev_get_queue(bdev); | 3231 | struct request_queue *r_queue = bdev_get_queue(bdev); |
| 3232 | 3232 | ||
| 3233 | if (!r_queue->issue_flush_fn) | 3233 | if (!r_queue->issue_flush_fn) |
| 3234 | ret = -EOPNOTSUPP; | 3234 | ret = -EOPNOTSUPP; |
| @@ -3267,7 +3267,7 @@ static int raid5_congested(void *data, int bits) | |||
| 3267 | /* We want read requests to align with chunks where possible, | 3267 | /* We want read requests to align with chunks where possible, |
| 3268 | * but write requests don't need to. | 3268 | * but write requests don't need to. |
| 3269 | */ | 3269 | */ |
| 3270 | static int raid5_mergeable_bvec(request_queue_t *q, struct bio *bio, struct bio_vec *biovec) | 3270 | static int raid5_mergeable_bvec(struct request_queue *q, struct bio *bio, struct bio_vec *biovec) |
| 3271 | { | 3271 | { |
| 3272 | mddev_t *mddev = q->queuedata; | 3272 | mddev_t *mddev = q->queuedata; |
| 3273 | sector_t sector = bio->bi_sector + get_start_sect(bio->bi_bdev); | 3273 | sector_t sector = bio->bi_sector + get_start_sect(bio->bi_bdev); |
| @@ -3377,7 +3377,7 @@ static int raid5_align_endio(struct bio *bi, unsigned int bytes, int error) | |||
| 3377 | 3377 | ||
| 3378 | static int bio_fits_rdev(struct bio *bi) | 3378 | static int bio_fits_rdev(struct bio *bi) |
| 3379 | { | 3379 | { |
| 3380 | request_queue_t *q = bdev_get_queue(bi->bi_bdev); | 3380 | struct request_queue *q = bdev_get_queue(bi->bi_bdev); |
| 3381 | 3381 | ||
| 3382 | if ((bi->bi_size>>9) > q->max_sectors) | 3382 | if ((bi->bi_size>>9) > q->max_sectors) |
| 3383 | return 0; | 3383 | return 0; |
| @@ -3396,7 +3396,7 @@ static int bio_fits_rdev(struct bio *bi) | |||
| 3396 | } | 3396 | } |
| 3397 | 3397 | ||
| 3398 | 3398 | ||
| 3399 | static int chunk_aligned_read(request_queue_t *q, struct bio * raid_bio) | 3399 | static int chunk_aligned_read(struct request_queue *q, struct bio * raid_bio) |
| 3400 | { | 3400 | { |
| 3401 | mddev_t *mddev = q->queuedata; | 3401 | mddev_t *mddev = q->queuedata; |
| 3402 | raid5_conf_t *conf = mddev_to_conf(mddev); | 3402 | raid5_conf_t *conf = mddev_to_conf(mddev); |
| @@ -3466,7 +3466,7 @@ static int chunk_aligned_read(request_queue_t *q, struct bio * raid_bio) | |||
| 3466 | } | 3466 | } |
| 3467 | 3467 | ||
| 3468 | 3468 | ||
| 3469 | static int make_request(request_queue_t *q, struct bio * bi) | 3469 | static int make_request(struct request_queue *q, struct bio * bi) |
| 3470 | { | 3470 | { |
| 3471 | mddev_t *mddev = q->queuedata; | 3471 | mddev_t *mddev = q->queuedata; |
| 3472 | raid5_conf_t *conf = mddev_to_conf(mddev); | 3472 | raid5_conf_t *conf = mddev_to_conf(mddev); |
