diff options
author | Neil Brown <neilb@suse.de> | 2008-05-14 19:05:54 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2008-05-14 22:11:15 -0400 |
commit | e7e72bf641b1fc7b9df6f40bd2c36dfccd8d647c (patch) | |
tree | 81b1db5434c9635bf23fb40415056e10390cd692 | |
parent | 4920916f728fe3c51f54c25ab7b3d271254aab5a (diff) |
Remove blkdev warning triggered by using md
As setting and clearing queue flags now requires that we hold a spinlock
on the queue, and as blk_queue_stack_limits is called without that lock,
get the lock inside blk_queue_stack_limits.
For blk_queue_stack_limits to be able to find the right lock, each md
personality needs to set q->queue_lock to point to the appropriate lock.
Those personalities which didn't previously use a spin_lock, us
q->__queue_lock. So always initialise that lock when allocated.
With this in place, setting/clearing of the QUEUE_FLAG_PLUGGED bit will no
longer cause warnings as it will be clear that the proper lock is held.
Thanks to Dan Williams for review and fixing the silly bugs.
Signed-off-by: NeilBrown <neilb@suse.de>
Cc: Dan Williams <dan.j.williams@intel.com>
Cc: Jens Axboe <jens.axboe@oracle.com>
Cc: Alistair John Strachan <alistair@devzero.co.uk>
Cc: Nick Piggin <npiggin@suse.de>
Cc: "Rafael J. Wysocki" <rjw@sisk.pl>
Cc: Jacek Luczak <difrost.kernel@gmail.com>
Cc: Prakash Punnoor <prakash@punnoor.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
-rw-r--r-- | block/blk-core.c | 5 | ||||
-rw-r--r-- | block/blk-settings.c | 8 | ||||
-rw-r--r-- | drivers/md/linear.c | 1 | ||||
-rw-r--r-- | drivers/md/multipath.c | 1 | ||||
-rw-r--r-- | drivers/md/raid0.c | 1 | ||||
-rw-r--r-- | drivers/md/raid1.c | 4 | ||||
-rw-r--r-- | drivers/md/raid10.c | 4 | ||||
-rw-r--r-- | drivers/md/raid5.c | 1 |
8 files changed, 19 insertions, 6 deletions
diff --git a/block/blk-core.c b/block/blk-core.c index 2987fe47b5ee..6a9cc0d22a61 100644 --- a/block/blk-core.c +++ b/block/blk-core.c | |||
@@ -482,6 +482,7 @@ struct request_queue *blk_alloc_queue_node(gfp_t gfp_mask, int node_id) | |||
482 | kobject_init(&q->kobj, &blk_queue_ktype); | 482 | kobject_init(&q->kobj, &blk_queue_ktype); |
483 | 483 | ||
484 | mutex_init(&q->sysfs_lock); | 484 | mutex_init(&q->sysfs_lock); |
485 | spin_lock_init(&q->__queue_lock); | ||
485 | 486 | ||
486 | return q; | 487 | return q; |
487 | } | 488 | } |
@@ -544,10 +545,8 @@ blk_init_queue_node(request_fn_proc *rfn, spinlock_t *lock, int node_id) | |||
544 | * if caller didn't supply a lock, they get per-queue locking with | 545 | * if caller didn't supply a lock, they get per-queue locking with |
545 | * our embedded lock | 546 | * our embedded lock |
546 | */ | 547 | */ |
547 | if (!lock) { | 548 | if (!lock) |
548 | spin_lock_init(&q->__queue_lock); | ||
549 | lock = &q->__queue_lock; | 549 | lock = &q->__queue_lock; |
550 | } | ||
551 | 550 | ||
552 | q->request_fn = rfn; | 551 | q->request_fn = rfn; |
553 | q->prep_rq_fn = NULL; | 552 | q->prep_rq_fn = NULL; |
diff --git a/block/blk-settings.c b/block/blk-settings.c index bb93d4c32775..8dd86418f35d 100644 --- a/block/blk-settings.c +++ b/block/blk-settings.c | |||
@@ -286,8 +286,14 @@ void blk_queue_stack_limits(struct request_queue *t, struct request_queue *b) | |||
286 | t->max_hw_segments = min(t->max_hw_segments, b->max_hw_segments); | 286 | t->max_hw_segments = min(t->max_hw_segments, b->max_hw_segments); |
287 | t->max_segment_size = min(t->max_segment_size, b->max_segment_size); | 287 | t->max_segment_size = min(t->max_segment_size, b->max_segment_size); |
288 | t->hardsect_size = max(t->hardsect_size, b->hardsect_size); | 288 | t->hardsect_size = max(t->hardsect_size, b->hardsect_size); |
289 | if (!test_bit(QUEUE_FLAG_CLUSTER, &b->queue_flags)) | 289 | if (!t->queue_lock) |
290 | WARN_ON_ONCE(1); | ||
291 | else if (!test_bit(QUEUE_FLAG_CLUSTER, &b->queue_flags)) { | ||
292 | unsigned long flags; | ||
293 | spin_lock_irqsave(t->queue_lock, flags); | ||
290 | queue_flag_clear(QUEUE_FLAG_CLUSTER, t); | 294 | queue_flag_clear(QUEUE_FLAG_CLUSTER, t); |
295 | spin_unlock_irqrestore(t->queue_lock, flags); | ||
296 | } | ||
291 | } | 297 | } |
292 | EXPORT_SYMBOL(blk_queue_stack_limits); | 298 | EXPORT_SYMBOL(blk_queue_stack_limits); |
293 | 299 | ||
diff --git a/drivers/md/linear.c b/drivers/md/linear.c index 0b8511776b3e..10748240cb2f 100644 --- a/drivers/md/linear.c +++ b/drivers/md/linear.c | |||
@@ -250,6 +250,7 @@ static int linear_run (mddev_t *mddev) | |||
250 | { | 250 | { |
251 | linear_conf_t *conf; | 251 | linear_conf_t *conf; |
252 | 252 | ||
253 | mddev->queue->queue_lock = &mddev->queue->__queue_lock; | ||
253 | conf = linear_conf(mddev, mddev->raid_disks); | 254 | conf = linear_conf(mddev, mddev->raid_disks); |
254 | 255 | ||
255 | if (!conf) | 256 | if (!conf) |
diff --git a/drivers/md/multipath.c b/drivers/md/multipath.c index 42ee1a2dc144..4f4d1f383842 100644 --- a/drivers/md/multipath.c +++ b/drivers/md/multipath.c | |||
@@ -417,6 +417,7 @@ static int multipath_run (mddev_t *mddev) | |||
417 | * bookkeeping area. [whatever we allocate in multipath_run(), | 417 | * bookkeeping area. [whatever we allocate in multipath_run(), |
418 | * should be freed in multipath_stop()] | 418 | * should be freed in multipath_stop()] |
419 | */ | 419 | */ |
420 | mddev->queue->queue_lock = &mddev->queue->__queue_lock; | ||
420 | 421 | ||
421 | conf = kzalloc(sizeof(multipath_conf_t), GFP_KERNEL); | 422 | conf = kzalloc(sizeof(multipath_conf_t), GFP_KERNEL); |
422 | mddev->private = conf; | 423 | mddev->private = conf; |
diff --git a/drivers/md/raid0.c b/drivers/md/raid0.c index 818b48284096..914c04ddec7c 100644 --- a/drivers/md/raid0.c +++ b/drivers/md/raid0.c | |||
@@ -280,6 +280,7 @@ static int raid0_run (mddev_t *mddev) | |||
280 | (mddev->chunk_size>>1)-1); | 280 | (mddev->chunk_size>>1)-1); |
281 | blk_queue_max_sectors(mddev->queue, mddev->chunk_size >> 9); | 281 | blk_queue_max_sectors(mddev->queue, mddev->chunk_size >> 9); |
282 | blk_queue_segment_boundary(mddev->queue, (mddev->chunk_size>>1) - 1); | 282 | blk_queue_segment_boundary(mddev->queue, (mddev->chunk_size>>1) - 1); |
283 | mddev->queue->queue_lock = &mddev->queue->__queue_lock; | ||
283 | 284 | ||
284 | conf = kmalloc(sizeof (raid0_conf_t), GFP_KERNEL); | 285 | conf = kmalloc(sizeof (raid0_conf_t), GFP_KERNEL); |
285 | if (!conf) | 286 | if (!conf) |
diff --git a/drivers/md/raid1.c b/drivers/md/raid1.c index 6778b7cb39bd..ac409b7d83f5 100644 --- a/drivers/md/raid1.c +++ b/drivers/md/raid1.c | |||
@@ -1935,6 +1935,9 @@ static int run(mddev_t *mddev) | |||
1935 | if (!conf->r1bio_pool) | 1935 | if (!conf->r1bio_pool) |
1936 | goto out_no_mem; | 1936 | goto out_no_mem; |
1937 | 1937 | ||
1938 | spin_lock_init(&conf->device_lock); | ||
1939 | mddev->queue->queue_lock = &conf->device_lock; | ||
1940 | |||
1938 | rdev_for_each(rdev, tmp, mddev) { | 1941 | rdev_for_each(rdev, tmp, mddev) { |
1939 | disk_idx = rdev->raid_disk; | 1942 | disk_idx = rdev->raid_disk; |
1940 | if (disk_idx >= mddev->raid_disks | 1943 | if (disk_idx >= mddev->raid_disks |
@@ -1958,7 +1961,6 @@ static int run(mddev_t *mddev) | |||
1958 | } | 1961 | } |
1959 | conf->raid_disks = mddev->raid_disks; | 1962 | conf->raid_disks = mddev->raid_disks; |
1960 | conf->mddev = mddev; | 1963 | conf->mddev = mddev; |
1961 | spin_lock_init(&conf->device_lock); | ||
1962 | INIT_LIST_HEAD(&conf->retry_list); | 1964 | INIT_LIST_HEAD(&conf->retry_list); |
1963 | 1965 | ||
1964 | spin_lock_init(&conf->resync_lock); | 1966 | spin_lock_init(&conf->resync_lock); |
diff --git a/drivers/md/raid10.c b/drivers/md/raid10.c index faf3d8912979..8536ede1e712 100644 --- a/drivers/md/raid10.c +++ b/drivers/md/raid10.c | |||
@@ -2082,6 +2082,9 @@ static int run(mddev_t *mddev) | |||
2082 | goto out_free_conf; | 2082 | goto out_free_conf; |
2083 | } | 2083 | } |
2084 | 2084 | ||
2085 | spin_lock_init(&conf->device_lock); | ||
2086 | mddev->queue->queue_lock = &conf->device_lock; | ||
2087 | |||
2085 | rdev_for_each(rdev, tmp, mddev) { | 2088 | rdev_for_each(rdev, tmp, mddev) { |
2086 | disk_idx = rdev->raid_disk; | 2089 | disk_idx = rdev->raid_disk; |
2087 | if (disk_idx >= mddev->raid_disks | 2090 | if (disk_idx >= mddev->raid_disks |
@@ -2103,7 +2106,6 @@ static int run(mddev_t *mddev) | |||
2103 | 2106 | ||
2104 | disk->head_position = 0; | 2107 | disk->head_position = 0; |
2105 | } | 2108 | } |
2106 | spin_lock_init(&conf->device_lock); | ||
2107 | INIT_LIST_HEAD(&conf->retry_list); | 2109 | INIT_LIST_HEAD(&conf->retry_list); |
2108 | 2110 | ||
2109 | spin_lock_init(&conf->resync_lock); | 2111 | spin_lock_init(&conf->resync_lock); |
diff --git a/drivers/md/raid5.c b/drivers/md/raid5.c index ee0ea9183080..93fde48c0f42 100644 --- a/drivers/md/raid5.c +++ b/drivers/md/raid5.c | |||
@@ -4257,6 +4257,7 @@ static int run(mddev_t *mddev) | |||
4257 | goto abort; | 4257 | goto abort; |
4258 | } | 4258 | } |
4259 | spin_lock_init(&conf->device_lock); | 4259 | spin_lock_init(&conf->device_lock); |
4260 | mddev->queue->queue_lock = &conf->device_lock; | ||
4260 | init_waitqueue_head(&conf->wait_for_stripe); | 4261 | init_waitqueue_head(&conf->wait_for_stripe); |
4261 | init_waitqueue_head(&conf->wait_for_overlap); | 4262 | init_waitqueue_head(&conf->wait_for_overlap); |
4262 | INIT_LIST_HEAD(&conf->handle_list); | 4263 | INIT_LIST_HEAD(&conf->handle_list); |