diff options
Diffstat (limited to 'block')
-rw-r--r-- | block/blk-cgroup.c | 2 | ||||
-rw-r--r-- | block/blk-core.c | 26 | ||||
-rw-r--r-- | block/blk-exec.c | 2 | ||||
-rw-r--r-- | block/blk-sysfs.c | 4 | ||||
-rw-r--r-- | block/blk-throttle.c | 2 | ||||
-rw-r--r-- | block/blk.h | 2 |
6 files changed, 19 insertions, 19 deletions
diff --git a/block/blk-cgroup.c b/block/blk-cgroup.c index d0b770391ad4..5dea4e8dbc55 100644 --- a/block/blk-cgroup.c +++ b/block/blk-cgroup.c | |||
@@ -231,7 +231,7 @@ struct blkcg_gq *blkg_lookup_create(struct blkcg *blkcg, | |||
231 | * we shouldn't allow anything to go through for a bypassing queue. | 231 | * we shouldn't allow anything to go through for a bypassing queue. |
232 | */ | 232 | */ |
233 | if (unlikely(blk_queue_bypass(q))) | 233 | if (unlikely(blk_queue_bypass(q))) |
234 | return ERR_PTR(blk_queue_dead(q) ? -EINVAL : -EBUSY); | 234 | return ERR_PTR(blk_queue_dying(q) ? -EINVAL : -EBUSY); |
235 | return __blkg_lookup_create(blkcg, q, NULL); | 235 | return __blkg_lookup_create(blkcg, q, NULL); |
236 | } | 236 | } |
237 | EXPORT_SYMBOL_GPL(blkg_lookup_create); | 237 | EXPORT_SYMBOL_GPL(blkg_lookup_create); |
diff --git a/block/blk-core.c b/block/blk-core.c index ee0e5cafa859..1a95272cca50 100644 --- a/block/blk-core.c +++ b/block/blk-core.c | |||
@@ -473,20 +473,20 @@ EXPORT_SYMBOL_GPL(blk_queue_bypass_end); | |||
473 | * blk_cleanup_queue - shutdown a request queue | 473 | * blk_cleanup_queue - shutdown a request queue |
474 | * @q: request queue to shutdown | 474 | * @q: request queue to shutdown |
475 | * | 475 | * |
476 | * Mark @q DEAD, drain all pending requests, destroy and put it. All | 476 | * Mark @q DYING, drain all pending requests, destroy and put it. All |
477 | * future requests will be failed immediately with -ENODEV. | 477 | * future requests will be failed immediately with -ENODEV. |
478 | */ | 478 | */ |
479 | void blk_cleanup_queue(struct request_queue *q) | 479 | void blk_cleanup_queue(struct request_queue *q) |
480 | { | 480 | { |
481 | spinlock_t *lock = q->queue_lock; | 481 | spinlock_t *lock = q->queue_lock; |
482 | 482 | ||
483 | /* mark @q DEAD, no new request or merges will be allowed afterwards */ | 483 | /* mark @q DYING, no new request or merges will be allowed afterwards */ |
484 | mutex_lock(&q->sysfs_lock); | 484 | mutex_lock(&q->sysfs_lock); |
485 | queue_flag_set_unlocked(QUEUE_FLAG_DEAD, q); | 485 | queue_flag_set_unlocked(QUEUE_FLAG_DYING, q); |
486 | spin_lock_irq(lock); | 486 | spin_lock_irq(lock); |
487 | 487 | ||
488 | /* | 488 | /* |
489 | * Dead queue is permanently in bypass mode till released. Note | 489 | * A dying queue is permanently in bypass mode till released. Note |
490 | * that, unlike blk_queue_bypass_start(), we aren't performing | 490 | * that, unlike blk_queue_bypass_start(), we aren't performing |
491 | * synchronize_rcu() after entering bypass mode to avoid the delay | 491 | * synchronize_rcu() after entering bypass mode to avoid the delay |
492 | * as some drivers create and destroy a lot of queues while | 492 | * as some drivers create and destroy a lot of queues while |
@@ -499,11 +499,11 @@ void blk_cleanup_queue(struct request_queue *q) | |||
499 | 499 | ||
500 | queue_flag_set(QUEUE_FLAG_NOMERGES, q); | 500 | queue_flag_set(QUEUE_FLAG_NOMERGES, q); |
501 | queue_flag_set(QUEUE_FLAG_NOXMERGES, q); | 501 | queue_flag_set(QUEUE_FLAG_NOXMERGES, q); |
502 | queue_flag_set(QUEUE_FLAG_DEAD, q); | 502 | queue_flag_set(QUEUE_FLAG_DYING, q); |
503 | spin_unlock_irq(lock); | 503 | spin_unlock_irq(lock); |
504 | mutex_unlock(&q->sysfs_lock); | 504 | mutex_unlock(&q->sysfs_lock); |
505 | 505 | ||
506 | /* drain all requests queued before DEAD marking */ | 506 | /* drain all requests queued before DYING marking */ |
507 | blk_drain_queue(q, true); | 507 | blk_drain_queue(q, true); |
508 | 508 | ||
509 | /* @q won't process any more request, flush async actions */ | 509 | /* @q won't process any more request, flush async actions */ |
@@ -716,7 +716,7 @@ EXPORT_SYMBOL(blk_init_allocated_queue); | |||
716 | 716 | ||
717 | bool blk_get_queue(struct request_queue *q) | 717 | bool blk_get_queue(struct request_queue *q) |
718 | { | 718 | { |
719 | if (likely(!blk_queue_dead(q))) { | 719 | if (likely(!blk_queue_dying(q))) { |
720 | __blk_get_queue(q); | 720 | __blk_get_queue(q); |
721 | return true; | 721 | return true; |
722 | } | 722 | } |
@@ -870,7 +870,7 @@ static struct request *__get_request(struct request_list *rl, int rw_flags, | |||
870 | const bool is_sync = rw_is_sync(rw_flags) != 0; | 870 | const bool is_sync = rw_is_sync(rw_flags) != 0; |
871 | int may_queue; | 871 | int may_queue; |
872 | 872 | ||
873 | if (unlikely(blk_queue_dead(q))) | 873 | if (unlikely(blk_queue_dying(q))) |
874 | return NULL; | 874 | return NULL; |
875 | 875 | ||
876 | may_queue = elv_may_queue(q, rw_flags); | 876 | may_queue = elv_may_queue(q, rw_flags); |
@@ -1050,7 +1050,7 @@ retry: | |||
1050 | if (rq) | 1050 | if (rq) |
1051 | return rq; | 1051 | return rq; |
1052 | 1052 | ||
1053 | if (!(gfp_mask & __GFP_WAIT) || unlikely(blk_queue_dead(q))) { | 1053 | if (!(gfp_mask & __GFP_WAIT) || unlikely(blk_queue_dying(q))) { |
1054 | blk_put_rl(rl); | 1054 | blk_put_rl(rl); |
1055 | return NULL; | 1055 | return NULL; |
1056 | } | 1056 | } |
@@ -1910,7 +1910,7 @@ int blk_insert_cloned_request(struct request_queue *q, struct request *rq) | |||
1910 | return -EIO; | 1910 | return -EIO; |
1911 | 1911 | ||
1912 | spin_lock_irqsave(q->queue_lock, flags); | 1912 | spin_lock_irqsave(q->queue_lock, flags); |
1913 | if (unlikely(blk_queue_dead(q))) { | 1913 | if (unlikely(blk_queue_dying(q))) { |
1914 | spin_unlock_irqrestore(q->queue_lock, flags); | 1914 | spin_unlock_irqrestore(q->queue_lock, flags); |
1915 | return -ENODEV; | 1915 | return -ENODEV; |
1916 | } | 1916 | } |
@@ -2885,9 +2885,9 @@ static void queue_unplugged(struct request_queue *q, unsigned int depth, | |||
2885 | trace_block_unplug(q, depth, !from_schedule); | 2885 | trace_block_unplug(q, depth, !from_schedule); |
2886 | 2886 | ||
2887 | /* | 2887 | /* |
2888 | * Don't mess with dead queue. | 2888 | * Don't mess with a dying queue. |
2889 | */ | 2889 | */ |
2890 | if (unlikely(blk_queue_dead(q))) { | 2890 | if (unlikely(blk_queue_dying(q))) { |
2891 | spin_unlock(q->queue_lock); | 2891 | spin_unlock(q->queue_lock); |
2892 | return; | 2892 | return; |
2893 | } | 2893 | } |
@@ -2996,7 +2996,7 @@ void blk_flush_plug_list(struct blk_plug *plug, bool from_schedule) | |||
2996 | /* | 2996 | /* |
2997 | * Short-circuit if @q is dead | 2997 | * Short-circuit if @q is dead |
2998 | */ | 2998 | */ |
2999 | if (unlikely(blk_queue_dead(q))) { | 2999 | if (unlikely(blk_queue_dying(q))) { |
3000 | __blk_end_request_all(rq, -ENODEV); | 3000 | __blk_end_request_all(rq, -ENODEV); |
3001 | continue; | 3001 | continue; |
3002 | } | 3002 | } |
diff --git a/block/blk-exec.c b/block/blk-exec.c index 8b6dc5bd4dd0..4aec98df7ba5 100644 --- a/block/blk-exec.c +++ b/block/blk-exec.c | |||
@@ -60,7 +60,7 @@ void blk_execute_rq_nowait(struct request_queue *q, struct gendisk *bd_disk, | |||
60 | 60 | ||
61 | spin_lock_irq(q->queue_lock); | 61 | spin_lock_irq(q->queue_lock); |
62 | 62 | ||
63 | if (unlikely(blk_queue_dead(q))) { | 63 | if (unlikely(blk_queue_dying(q))) { |
64 | rq->errors = -ENXIO; | 64 | rq->errors = -ENXIO; |
65 | if (rq->end_io) | 65 | if (rq->end_io) |
66 | rq->end_io(rq, rq->errors); | 66 | rq->end_io(rq, rq->errors); |
diff --git a/block/blk-sysfs.c b/block/blk-sysfs.c index ce6204608822..788147797a79 100644 --- a/block/blk-sysfs.c +++ b/block/blk-sysfs.c | |||
@@ -466,7 +466,7 @@ queue_attr_show(struct kobject *kobj, struct attribute *attr, char *page) | |||
466 | if (!entry->show) | 466 | if (!entry->show) |
467 | return -EIO; | 467 | return -EIO; |
468 | mutex_lock(&q->sysfs_lock); | 468 | mutex_lock(&q->sysfs_lock); |
469 | if (blk_queue_dead(q)) { | 469 | if (blk_queue_dying(q)) { |
470 | mutex_unlock(&q->sysfs_lock); | 470 | mutex_unlock(&q->sysfs_lock); |
471 | return -ENOENT; | 471 | return -ENOENT; |
472 | } | 472 | } |
@@ -488,7 +488,7 @@ queue_attr_store(struct kobject *kobj, struct attribute *attr, | |||
488 | 488 | ||
489 | q = container_of(kobj, struct request_queue, kobj); | 489 | q = container_of(kobj, struct request_queue, kobj); |
490 | mutex_lock(&q->sysfs_lock); | 490 | mutex_lock(&q->sysfs_lock); |
491 | if (blk_queue_dead(q)) { | 491 | if (blk_queue_dying(q)) { |
492 | mutex_unlock(&q->sysfs_lock); | 492 | mutex_unlock(&q->sysfs_lock); |
493 | return -ENOENT; | 493 | return -ENOENT; |
494 | } | 494 | } |
diff --git a/block/blk-throttle.c b/block/blk-throttle.c index a9664fa0b609..31146225f3d0 100644 --- a/block/blk-throttle.c +++ b/block/blk-throttle.c | |||
@@ -302,7 +302,7 @@ static struct throtl_grp *throtl_lookup_create_tg(struct throtl_data *td, | |||
302 | /* if %NULL and @q is alive, fall back to root_tg */ | 302 | /* if %NULL and @q is alive, fall back to root_tg */ |
303 | if (!IS_ERR(blkg)) | 303 | if (!IS_ERR(blkg)) |
304 | tg = blkg_to_tg(blkg); | 304 | tg = blkg_to_tg(blkg); |
305 | else if (!blk_queue_dead(q)) | 305 | else if (!blk_queue_dying(q)) |
306 | tg = td_root_tg(td); | 306 | tg = td_root_tg(td); |
307 | } | 307 | } |
308 | 308 | ||
diff --git a/block/blk.h b/block/blk.h index ca51543b248c..2218a8a78292 100644 --- a/block/blk.h +++ b/block/blk.h | |||
@@ -96,7 +96,7 @@ static inline struct request *__elv_next_request(struct request_queue *q) | |||
96 | q->flush_queue_delayed = 1; | 96 | q->flush_queue_delayed = 1; |
97 | return NULL; | 97 | return NULL; |
98 | } | 98 | } |
99 | if (unlikely(blk_queue_dead(q)) || | 99 | if (unlikely(blk_queue_dying(q)) || |
100 | !q->elevator->type->ops.elevator_dispatch_fn(q, 0)) | 100 | !q->elevator->type->ops.elevator_dispatch_fn(q, 0)) |
101 | return NULL; | 101 | return NULL; |
102 | } | 102 | } |