summaryrefslogtreecommitdiffstats
path: root/block/bfq-iosched.c
diff options
context:
space:
mode:
Diffstat (limited to 'block/bfq-iosched.c')
-rw-r--r--block/bfq-iosched.c16
1 files changed, 8 insertions, 8 deletions
diff --git a/block/bfq-iosched.c b/block/bfq-iosched.c
index c7636cbefc85..67b22c924aee 100644
--- a/block/bfq-iosched.c
+++ b/block/bfq-iosched.c
@@ -399,9 +399,9 @@ static struct bfq_io_cq *bfq_bic_lookup(struct bfq_data *bfqd,
399 unsigned long flags; 399 unsigned long flags;
400 struct bfq_io_cq *icq; 400 struct bfq_io_cq *icq;
401 401
402 spin_lock_irqsave(q->queue_lock, flags); 402 spin_lock_irqsave(&q->queue_lock, flags);
403 icq = icq_to_bic(ioc_lookup_icq(ioc, q)); 403 icq = icq_to_bic(ioc_lookup_icq(ioc, q));
404 spin_unlock_irqrestore(q->queue_lock, flags); 404 spin_unlock_irqrestore(&q->queue_lock, flags);
405 405
406 return icq; 406 return icq;
407 } 407 }
@@ -4034,7 +4034,7 @@ static void bfq_update_dispatch_stats(struct request_queue *q,
4034 * In addition, the following queue lock guarantees that 4034 * In addition, the following queue lock guarantees that
4035 * bfqq_group(bfqq) exists as well. 4035 * bfqq_group(bfqq) exists as well.
4036 */ 4036 */
4037 spin_lock_irq(q->queue_lock); 4037 spin_lock_irq(&q->queue_lock);
4038 if (idle_timer_disabled) 4038 if (idle_timer_disabled)
4039 /* 4039 /*
4040 * Since the idle timer has been disabled, 4040 * Since the idle timer has been disabled,
@@ -4053,7 +4053,7 @@ static void bfq_update_dispatch_stats(struct request_queue *q,
4053 bfqg_stats_set_start_empty_time(bfqg); 4053 bfqg_stats_set_start_empty_time(bfqg);
4054 bfqg_stats_update_io_remove(bfqg, rq->cmd_flags); 4054 bfqg_stats_update_io_remove(bfqg, rq->cmd_flags);
4055 } 4055 }
4056 spin_unlock_irq(q->queue_lock); 4056 spin_unlock_irq(&q->queue_lock);
4057} 4057}
4058#else 4058#else
4059static inline void bfq_update_dispatch_stats(struct request_queue *q, 4059static inline void bfq_update_dispatch_stats(struct request_queue *q,
@@ -4637,11 +4637,11 @@ static void bfq_update_insert_stats(struct request_queue *q,
4637 * In addition, the following queue lock guarantees that 4637 * In addition, the following queue lock guarantees that
4638 * bfqq_group(bfqq) exists as well. 4638 * bfqq_group(bfqq) exists as well.
4639 */ 4639 */
4640 spin_lock_irq(q->queue_lock); 4640 spin_lock_irq(&q->queue_lock);
4641 bfqg_stats_update_io_add(bfqq_group(bfqq), bfqq, cmd_flags); 4641 bfqg_stats_update_io_add(bfqq_group(bfqq), bfqq, cmd_flags);
4642 if (idle_timer_disabled) 4642 if (idle_timer_disabled)
4643 bfqg_stats_update_idle_time(bfqq_group(bfqq)); 4643 bfqg_stats_update_idle_time(bfqq_group(bfqq));
4644 spin_unlock_irq(q->queue_lock); 4644 spin_unlock_irq(&q->queue_lock);
4645} 4645}
4646#else 4646#else
4647static inline void bfq_update_insert_stats(struct request_queue *q, 4647static inline void bfq_update_insert_stats(struct request_queue *q,
@@ -5382,9 +5382,9 @@ static int bfq_init_queue(struct request_queue *q, struct elevator_type *e)
5382 } 5382 }
5383 eq->elevator_data = bfqd; 5383 eq->elevator_data = bfqd;
5384 5384
5385 spin_lock_irq(q->queue_lock); 5385 spin_lock_irq(&q->queue_lock);
5386 q->elevator = eq; 5386 q->elevator = eq;
5387 spin_unlock_irq(q->queue_lock); 5387 spin_unlock_irq(&q->queue_lock);
5388 5388
5389 /* 5389 /*
5390 * Our fallback bfqq if bfq_find_alloc_queue() runs into OOM issues. 5390 * Our fallback bfqq if bfq_find_alloc_queue() runs into OOM issues.