aboutsummaryrefslogtreecommitdiffstats
path: root/drivers
diff options
context:
space:
mode:
Diffstat (limited to 'drivers')
-rw-r--r--drivers/md/md.c61
1 files changed, 49 insertions, 12 deletions
diff --git a/drivers/md/md.c b/drivers/md/md.c
index 970a8c42ba92..38697283aaf4 100644
--- a/drivers/md/md.c
+++ b/drivers/md/md.c
@@ -214,16 +214,33 @@ static inline mddev_t *mddev_get(mddev_t *mddev)
214 return mddev; 214 return mddev;
215} 215}
216 216
217static void mddev_delayed_delete(struct work_struct *ws)
218{
219 mddev_t *mddev = container_of(ws, mddev_t, del_work);
220 kobject_del(&mddev->kobj);
221 kobject_put(&mddev->kobj);
222}
223
217static void mddev_put(mddev_t *mddev) 224static void mddev_put(mddev_t *mddev)
218{ 225{
219 if (!atomic_dec_and_lock(&mddev->active, &all_mddevs_lock)) 226 if (!atomic_dec_and_lock(&mddev->active, &all_mddevs_lock))
220 return; 227 return;
221 if (!mddev->raid_disks && list_empty(&mddev->disks)) { 228 if (!mddev->raid_disks && list_empty(&mddev->disks) &&
229 !mddev->hold_active) {
222 list_del(&mddev->all_mddevs); 230 list_del(&mddev->all_mddevs);
223 spin_unlock(&all_mddevs_lock); 231 if (mddev->gendisk) {
224 kobject_put(&mddev->kobj); 232 /* we did a probe so need to clean up.
225 } else 233 * Call schedule_work inside the spinlock
226 spin_unlock(&all_mddevs_lock); 234 * so that flush_scheduled_work() after
235 * mddev_find will succeed in waiting for the
236 * work to be done.
237 */
238 INIT_WORK(&mddev->del_work, mddev_delayed_delete);
239 schedule_work(&mddev->del_work);
240 } else
241 kfree(mddev);
242 }
243 spin_unlock(&all_mddevs_lock);
227} 244}
228 245
229static mddev_t * mddev_find(dev_t unit) 246static mddev_t * mddev_find(dev_t unit)
@@ -242,6 +259,7 @@ static mddev_t * mddev_find(dev_t unit)
242 259
243 if (new) { 260 if (new) {
244 list_add(&new->all_mddevs, &all_mddevs); 261 list_add(&new->all_mddevs, &all_mddevs);
262 mddev->hold_active = UNTIL_IOCTL;
245 spin_unlock(&all_mddevs_lock); 263 spin_unlock(&all_mddevs_lock);
246 return new; 264 return new;
247 } 265 }
@@ -3435,6 +3453,8 @@ md_attr_store(struct kobject *kobj, struct attribute *attr,
3435 if (!capable(CAP_SYS_ADMIN)) 3453 if (!capable(CAP_SYS_ADMIN))
3436 return -EACCES; 3454 return -EACCES;
3437 rv = mddev_lock(mddev); 3455 rv = mddev_lock(mddev);
3456 if (mddev->hold_active == UNTIL_IOCTL)
3457 mddev->hold_active = 0;
3438 if (!rv) { 3458 if (!rv) {
3439 rv = entry->store(mddev, page, length); 3459 rv = entry->store(mddev, page, length);
3440 mddev_unlock(mddev); 3460 mddev_unlock(mddev);
@@ -3484,6 +3504,11 @@ static struct kobject *md_probe(dev_t dev, int *part, void *data)
3484 if (!mddev) 3504 if (!mddev)
3485 return NULL; 3505 return NULL;
3486 3506
3507 /* wait for any previous instance if this device
3508 * to be completed removed (mddev_delayed_delete).
3509 */
3510 flush_scheduled_work();
3511
3487 mutex_lock(&disks_mutex); 3512 mutex_lock(&disks_mutex);
3488 if (mddev->gendisk) { 3513 if (mddev->gendisk) {
3489 mutex_unlock(&disks_mutex); 3514 mutex_unlock(&disks_mutex);
@@ -3520,7 +3545,7 @@ static struct kobject *md_probe(dev_t dev, int *part, void *data)
3520 disk->private_data = mddev; 3545 disk->private_data = mddev;
3521 disk->queue = mddev->queue; 3546 disk->queue = mddev->queue;
3522 /* Allow extended partitions. This makes the 3547 /* Allow extended partitions. This makes the
3523 * 'mdp' device redundant, but we can really 3548 * 'mdp' device redundant, but we can't really
3524 * remove it now. 3549 * remove it now.
3525 */ 3550 */
3526 disk->flags |= GENHD_FL_EXT_DEVT; 3551 disk->flags |= GENHD_FL_EXT_DEVT;
@@ -3536,6 +3561,7 @@ static struct kobject *md_probe(dev_t dev, int *part, void *data)
3536 kobject_uevent(&mddev->kobj, KOBJ_ADD); 3561 kobject_uevent(&mddev->kobj, KOBJ_ADD);
3537 mddev->sysfs_state = sysfs_get_dirent(mddev->kobj.sd, "array_state"); 3562 mddev->sysfs_state = sysfs_get_dirent(mddev->kobj.sd, "array_state");
3538 } 3563 }
3564 mddev_put(mddev);
3539 return NULL; 3565 return NULL;
3540} 3566}
3541 3567
@@ -5054,6 +5080,9 @@ static int md_ioctl(struct block_device *bdev, fmode_t mode,
5054 5080
5055done_unlock: 5081done_unlock:
5056abort_unlock: 5082abort_unlock:
5083 if (mddev->hold_active == UNTIL_IOCTL &&
5084 err != -EINVAL)
5085 mddev->hold_active = 0;
5057 mddev_unlock(mddev); 5086 mddev_unlock(mddev);
5058 5087
5059 return err; 5088 return err;
@@ -5070,14 +5099,25 @@ static int md_open(struct block_device *bdev, fmode_t mode)
5070 * Succeed if we can lock the mddev, which confirms that 5099 * Succeed if we can lock the mddev, which confirms that
5071 * it isn't being stopped right now. 5100 * it isn't being stopped right now.
5072 */ 5101 */
5073 mddev_t *mddev = bdev->bd_disk->private_data; 5102 mddev_t *mddev = mddev_find(bdev->bd_dev);
5074 int err; 5103 int err;
5075 5104
5105 if (mddev->gendisk != bdev->bd_disk) {
5106 /* we are racing with mddev_put which is discarding this
5107 * bd_disk.
5108 */
5109 mddev_put(mddev);
5110 /* Wait until bdev->bd_disk is definitely gone */
5111 flush_scheduled_work();
5112 /* Then retry the open from the top */
5113 return -ERESTARTSYS;
5114 }
5115 BUG_ON(mddev != bdev->bd_disk->private_data);
5116
5076 if ((err = mutex_lock_interruptible_nested(&mddev->reconfig_mutex, 1))) 5117 if ((err = mutex_lock_interruptible_nested(&mddev->reconfig_mutex, 1)))
5077 goto out; 5118 goto out;
5078 5119
5079 err = 0; 5120 err = 0;
5080 mddev_get(mddev);
5081 atomic_inc(&mddev->openers); 5121 atomic_inc(&mddev->openers);
5082 mddev_unlock(mddev); 5122 mddev_unlock(mddev);
5083 5123
@@ -6436,11 +6476,8 @@ static __exit void md_exit(void)
6436 unregister_sysctl_table(raid_table_header); 6476 unregister_sysctl_table(raid_table_header);
6437 remove_proc_entry("mdstat", NULL); 6477 remove_proc_entry("mdstat", NULL);
6438 for_each_mddev(mddev, tmp) { 6478 for_each_mddev(mddev, tmp) {
6439 struct gendisk *disk = mddev->gendisk;
6440 if (!disk)
6441 continue;
6442 export_array(mddev); 6479 export_array(mddev);
6443 mddev_put(mddev); 6480 mddev->hold_active = 0;
6444 } 6481 }
6445} 6482}
6446 6483