aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/md/raid1.c
diff options
context:
space:
mode:
Diffstat (limited to 'drivers/md/raid1.c')
-rw-r--r--drivers/md/raid1.c110
1 files changed, 56 insertions, 54 deletions
diff --git a/drivers/md/raid1.c b/drivers/md/raid1.c
index b3856db8d6c2..ea1f1eb93c77 100644
--- a/drivers/md/raid1.c
+++ b/drivers/md/raid1.c
@@ -177,6 +177,13 @@ static inline void free_r1bio(r1bio_t *r1_bio)
177static inline void put_buf(r1bio_t *r1_bio) 177static inline void put_buf(r1bio_t *r1_bio)
178{ 178{
179 conf_t *conf = mddev_to_conf(r1_bio->mddev); 179 conf_t *conf = mddev_to_conf(r1_bio->mddev);
180 int i;
181
182 for (i=0; i<conf->raid_disks; i++) {
183 struct bio *bio = r1_bio->bios[i];
184 if (bio->bi_end_io)
185 rdev_dec_pending(conf->mirrors[i].rdev, r1_bio->mddev);
186 }
180 187
181 mempool_free(r1_bio, conf->r1buf_pool); 188 mempool_free(r1_bio, conf->r1buf_pool);
182 189
@@ -1085,7 +1092,6 @@ static int end_sync_read(struct bio *bio, unsigned int bytes_done, int error)
1085 conf->mirrors[r1_bio->read_disk].rdev); 1092 conf->mirrors[r1_bio->read_disk].rdev);
1086 } else 1093 } else
1087 set_bit(R1BIO_Uptodate, &r1_bio->state); 1094 set_bit(R1BIO_Uptodate, &r1_bio->state);
1088 rdev_dec_pending(conf->mirrors[r1_bio->read_disk].rdev, conf->mddev);
1089 reschedule_retry(r1_bio); 1095 reschedule_retry(r1_bio);
1090 return 0; 1096 return 0;
1091} 1097}
@@ -1116,7 +1122,6 @@ static int end_sync_write(struct bio *bio, unsigned int bytes_done, int error)
1116 md_done_sync(mddev, r1_bio->sectors, uptodate); 1122 md_done_sync(mddev, r1_bio->sectors, uptodate);
1117 put_buf(r1_bio); 1123 put_buf(r1_bio);
1118 } 1124 }
1119 rdev_dec_pending(conf->mirrors[mirror].rdev, mddev);
1120 return 0; 1125 return 0;
1121} 1126}
1122 1127
@@ -1153,10 +1158,14 @@ static void sync_request_write(mddev_t *mddev, r1bio_t *r1_bio)
1153 atomic_set(&r1_bio->remaining, 1); 1158 atomic_set(&r1_bio->remaining, 1);
1154 for (i = 0; i < disks ; i++) { 1159 for (i = 0; i < disks ; i++) {
1155 wbio = r1_bio->bios[i]; 1160 wbio = r1_bio->bios[i];
1156 if (wbio->bi_end_io != end_sync_write) 1161 if (wbio->bi_end_io == NULL ||
1162 (wbio->bi_end_io == end_sync_read &&
1163 (i == r1_bio->read_disk ||
1164 !test_bit(MD_RECOVERY_SYNC, &mddev->recovery))))
1157 continue; 1165 continue;
1158 1166
1159 atomic_inc(&conf->mirrors[i].rdev->nr_pending); 1167 wbio->bi_rw = WRITE;
1168 wbio->bi_end_io = end_sync_write;
1160 atomic_inc(&r1_bio->remaining); 1169 atomic_inc(&r1_bio->remaining);
1161 md_sync_acct(conf->mirrors[i].rdev->bdev, wbio->bi_size >> 9); 1170 md_sync_acct(conf->mirrors[i].rdev->bdev, wbio->bi_size >> 9);
1162 1171
@@ -1388,14 +1397,13 @@ static int init_resync(conf_t *conf)
1388static sector_t sync_request(mddev_t *mddev, sector_t sector_nr, int *skipped, int go_faster) 1397static sector_t sync_request(mddev_t *mddev, sector_t sector_nr, int *skipped, int go_faster)
1389{ 1398{
1390 conf_t *conf = mddev_to_conf(mddev); 1399 conf_t *conf = mddev_to_conf(mddev);
1391 mirror_info_t *mirror;
1392 r1bio_t *r1_bio; 1400 r1bio_t *r1_bio;
1393 struct bio *bio; 1401 struct bio *bio;
1394 sector_t max_sector, nr_sectors; 1402 sector_t max_sector, nr_sectors;
1395 int disk; 1403 int disk = -1;
1396 int i; 1404 int i;
1397 int wonly; 1405 int wonly = -1;
1398 int write_targets = 0; 1406 int write_targets = 0, read_targets = 0;
1399 int sync_blocks; 1407 int sync_blocks;
1400 int still_degraded = 0; 1408 int still_degraded = 0;
1401 1409
@@ -1447,44 +1455,24 @@ static sector_t sync_request(mddev_t *mddev, sector_t sector_nr, int *skipped, i
1447 1455
1448 conf->next_resync = sector_nr; 1456 conf->next_resync = sector_nr;
1449 1457
1458 r1_bio = mempool_alloc(conf->r1buf_pool, GFP_NOIO);
1459 rcu_read_lock();
1450 /* 1460 /*
1451 * If reconstructing, and >1 working disc, 1461 * If we get a correctably read error during resync or recovery,
1452 * could dedicate one to rebuild and others to 1462 * we might want to read from a different device. So we
1453 * service read requests .. 1463 * flag all drives that could conceivably be read from for READ,
1464 * and any others (which will be non-In_sync devices) for WRITE.
1465 * If a read fails, we try reading from something else for which READ
1466 * is OK.
1454 */ 1467 */
1455 disk = conf->last_used;
1456 /* make sure disk is operational */
1457 wonly = disk;
1458 while (conf->mirrors[disk].rdev == NULL ||
1459 !test_bit(In_sync, &conf->mirrors[disk].rdev->flags) ||
1460 test_bit(WriteMostly, &conf->mirrors[disk].rdev->flags)
1461 ) {
1462 if (conf->mirrors[disk].rdev &&
1463 test_bit(In_sync, &conf->mirrors[disk].rdev->flags))
1464 wonly = disk;
1465 if (disk <= 0)
1466 disk = conf->raid_disks;
1467 disk--;
1468 if (disk == conf->last_used) {
1469 disk = wonly;
1470 break;
1471 }
1472 }
1473 conf->last_used = disk;
1474 atomic_inc(&conf->mirrors[disk].rdev->nr_pending);
1475
1476
1477 mirror = conf->mirrors + disk;
1478
1479 r1_bio = mempool_alloc(conf->r1buf_pool, GFP_NOIO);
1480 1468
1481 r1_bio->mddev = mddev; 1469 r1_bio->mddev = mddev;
1482 r1_bio->sector = sector_nr; 1470 r1_bio->sector = sector_nr;
1483 r1_bio->state = 0; 1471 r1_bio->state = 0;
1484 set_bit(R1BIO_IsSync, &r1_bio->state); 1472 set_bit(R1BIO_IsSync, &r1_bio->state);
1485 r1_bio->read_disk = disk;
1486 1473
1487 for (i=0; i < conf->raid_disks; i++) { 1474 for (i=0; i < conf->raid_disks; i++) {
1475 mdk_rdev_t *rdev;
1488 bio = r1_bio->bios[i]; 1476 bio = r1_bio->bios[i];
1489 1477
1490 /* take from bio_init */ 1478 /* take from bio_init */
@@ -1499,35 +1487,49 @@ static sector_t sync_request(mddev_t *mddev, sector_t sector_nr, int *skipped, i
1499 bio->bi_end_io = NULL; 1487 bio->bi_end_io = NULL;
1500 bio->bi_private = NULL; 1488 bio->bi_private = NULL;
1501 1489
1502 if (i == disk) { 1490 rdev = rcu_dereference(conf->mirrors[i].rdev);
1503 bio->bi_rw = READ; 1491 if (rdev == NULL ||
1504 bio->bi_end_io = end_sync_read; 1492 test_bit(Faulty, &rdev->flags)) {
1505 } else if (conf->mirrors[i].rdev == NULL ||
1506 test_bit(Faulty, &conf->mirrors[i].rdev->flags)) {
1507 still_degraded = 1; 1493 still_degraded = 1;
1508 continue; 1494 continue;
1509 } else if (!test_bit(In_sync, &conf->mirrors[i].rdev->flags) || 1495 } else if (!test_bit(In_sync, &rdev->flags)) {
1510 sector_nr + RESYNC_SECTORS > mddev->recovery_cp ||
1511 test_bit(MD_RECOVERY_REQUESTED, &mddev->recovery)) {
1512 bio->bi_rw = WRITE; 1496 bio->bi_rw = WRITE;
1513 bio->bi_end_io = end_sync_write; 1497 bio->bi_end_io = end_sync_write;
1514 write_targets ++; 1498 write_targets ++;
1515 } else 1499 } else {
1516 /* no need to read or write here */ 1500 /* may need to read from here */
1517 continue; 1501 bio->bi_rw = READ;
1518 bio->bi_sector = sector_nr + conf->mirrors[i].rdev->data_offset; 1502 bio->bi_end_io = end_sync_read;
1519 bio->bi_bdev = conf->mirrors[i].rdev->bdev; 1503 if (test_bit(WriteMostly, &rdev->flags)) {
1504 if (wonly < 0)
1505 wonly = i;
1506 } else {
1507 if (disk < 0)
1508 disk = i;
1509 }
1510 read_targets++;
1511 }
1512 atomic_inc(&rdev->nr_pending);
1513 bio->bi_sector = sector_nr + rdev->data_offset;
1514 bio->bi_bdev = rdev->bdev;
1520 bio->bi_private = r1_bio; 1515 bio->bi_private = r1_bio;
1521 } 1516 }
1517 rcu_read_unlock();
1518 if (disk < 0)
1519 disk = wonly;
1520 r1_bio->read_disk = disk;
1522 1521
1523 if (write_targets == 0) { 1522 if (test_bit(MD_RECOVERY_SYNC, &mddev->recovery) && read_targets > 0)
1523 /* extra read targets are also write targets */
1524 write_targets += read_targets-1;
1525
1526 if (write_targets == 0 || read_targets == 0) {
1524 /* There is nowhere to write, so all non-sync 1527 /* There is nowhere to write, so all non-sync
1525 * drives must be failed - so we are finished 1528 * drives must be failed - so we are finished
1526 */ 1529 */
1527 sector_t rv = max_sector - sector_nr; 1530 sector_t rv = max_sector - sector_nr;
1528 *skipped = 1; 1531 *skipped = 1;
1529 put_buf(r1_bio); 1532 put_buf(r1_bio);
1530 rdev_dec_pending(conf->mirrors[disk].rdev, mddev);
1531 return rv; 1533 return rv;
1532 } 1534 }
1533 1535
@@ -1578,10 +1580,10 @@ static sector_t sync_request(mddev_t *mddev, sector_t sector_nr, int *skipped, i
1578 sync_blocks -= (len>>9); 1580 sync_blocks -= (len>>9);
1579 } while (r1_bio->bios[disk]->bi_vcnt < RESYNC_PAGES); 1581 } while (r1_bio->bios[disk]->bi_vcnt < RESYNC_PAGES);
1580 bio_full: 1582 bio_full:
1581 bio = r1_bio->bios[disk]; 1583 bio = r1_bio->bios[r1_bio->read_disk];
1582 r1_bio->sectors = nr_sectors; 1584 r1_bio->sectors = nr_sectors;
1583 1585
1584 md_sync_acct(mirror->rdev->bdev, nr_sectors); 1586 md_sync_acct(conf->mirrors[r1_bio->read_disk].rdev->bdev, nr_sectors);
1585 1587
1586 generic_make_request(bio); 1588 generic_make_request(bio);
1587 1589