diff options
-rw-r--r-- | block/cfq-iosched.c | 54 | ||||
-rw-r--r-- | block/ll_rw_blk.c | 2 | ||||
-rw-r--r-- | fs/ioprio.c | 4 | ||||
-rw-r--r-- | include/linux/blkdev.h | 2 |
4 files changed, 21 insertions, 41 deletions
diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c index ec24284e9d39..33e0b0c5e31d 100644 --- a/block/cfq-iosched.c +++ b/block/cfq-iosched.c | |||
@@ -31,8 +31,6 @@ static int cfq_slice_idle = HZ / 125; | |||
31 | 31 | ||
32 | #define CFQ_KEY_ASYNC (0) | 32 | #define CFQ_KEY_ASYNC (0) |
33 | 33 | ||
34 | static DEFINE_SPINLOCK(cfq_exit_lock); | ||
35 | |||
36 | /* | 34 | /* |
37 | * for the hash of cfqq inside the cfqd | 35 | * for the hash of cfqq inside the cfqd |
38 | */ | 36 | */ |
@@ -1084,12 +1082,6 @@ static void cfq_free_io_context(struct io_context *ioc) | |||
1084 | complete(ioc_gone); | 1082 | complete(ioc_gone); |
1085 | } | 1083 | } |
1086 | 1084 | ||
1087 | static void cfq_trim(struct io_context *ioc) | ||
1088 | { | ||
1089 | ioc->set_ioprio = NULL; | ||
1090 | cfq_free_io_context(ioc); | ||
1091 | } | ||
1092 | |||
1093 | static void cfq_exit_cfqq(struct cfq_data *cfqd, struct cfq_queue *cfqq) | 1085 | static void cfq_exit_cfqq(struct cfq_data *cfqd, struct cfq_queue *cfqq) |
1094 | { | 1086 | { |
1095 | if (unlikely(cfqq == cfqd->active_queue)) | 1087 | if (unlikely(cfqq == cfqd->active_queue)) |
@@ -1101,6 +1093,10 @@ static void cfq_exit_cfqq(struct cfq_data *cfqd, struct cfq_queue *cfqq) | |||
1101 | static void __cfq_exit_single_io_context(struct cfq_data *cfqd, | 1093 | static void __cfq_exit_single_io_context(struct cfq_data *cfqd, |
1102 | struct cfq_io_context *cic) | 1094 | struct cfq_io_context *cic) |
1103 | { | 1095 | { |
1096 | list_del_init(&cic->queue_list); | ||
1097 | smp_wmb(); | ||
1098 | cic->key = NULL; | ||
1099 | |||
1104 | if (cic->cfqq[ASYNC]) { | 1100 | if (cic->cfqq[ASYNC]) { |
1105 | cfq_exit_cfqq(cfqd, cic->cfqq[ASYNC]); | 1101 | cfq_exit_cfqq(cfqd, cic->cfqq[ASYNC]); |
1106 | cic->cfqq[ASYNC] = NULL; | 1102 | cic->cfqq[ASYNC] = NULL; |
@@ -1110,9 +1106,6 @@ static void __cfq_exit_single_io_context(struct cfq_data *cfqd, | |||
1110 | cfq_exit_cfqq(cfqd, cic->cfqq[SYNC]); | 1106 | cfq_exit_cfqq(cfqd, cic->cfqq[SYNC]); |
1111 | cic->cfqq[SYNC] = NULL; | 1107 | cic->cfqq[SYNC] = NULL; |
1112 | } | 1108 | } |
1113 | |||
1114 | cic->key = NULL; | ||
1115 | list_del_init(&cic->queue_list); | ||
1116 | } | 1109 | } |
1117 | 1110 | ||
1118 | 1111 | ||
@@ -1123,27 +1116,23 @@ static void cfq_exit_single_io_context(struct cfq_io_context *cic) | |||
1123 | { | 1116 | { |
1124 | struct cfq_data *cfqd = cic->key; | 1117 | struct cfq_data *cfqd = cic->key; |
1125 | 1118 | ||
1126 | WARN_ON(!irqs_disabled()); | ||
1127 | |||
1128 | if (cfqd) { | 1119 | if (cfqd) { |
1129 | request_queue_t *q = cfqd->queue; | 1120 | request_queue_t *q = cfqd->queue; |
1130 | 1121 | ||
1131 | spin_lock(q->queue_lock); | 1122 | spin_lock_irq(q->queue_lock); |
1132 | __cfq_exit_single_io_context(cfqd, cic); | 1123 | __cfq_exit_single_io_context(cfqd, cic); |
1133 | spin_unlock(q->queue_lock); | 1124 | spin_unlock_irq(q->queue_lock); |
1134 | } | 1125 | } |
1135 | } | 1126 | } |
1136 | 1127 | ||
1137 | static void cfq_exit_io_context(struct io_context *ioc) | 1128 | static void cfq_exit_io_context(struct io_context *ioc) |
1138 | { | 1129 | { |
1139 | struct cfq_io_context *__cic; | 1130 | struct cfq_io_context *__cic; |
1140 | unsigned long flags; | ||
1141 | struct rb_node *n; | 1131 | struct rb_node *n; |
1142 | 1132 | ||
1143 | /* | 1133 | /* |
1144 | * put the reference this task is holding to the various queues | 1134 | * put the reference this task is holding to the various queues |
1145 | */ | 1135 | */ |
1146 | spin_lock_irqsave(&cfq_exit_lock, flags); | ||
1147 | 1136 | ||
1148 | n = rb_first(&ioc->cic_root); | 1137 | n = rb_first(&ioc->cic_root); |
1149 | while (n != NULL) { | 1138 | while (n != NULL) { |
@@ -1152,8 +1141,6 @@ static void cfq_exit_io_context(struct io_context *ioc) | |||
1152 | cfq_exit_single_io_context(__cic); | 1141 | cfq_exit_single_io_context(__cic); |
1153 | n = rb_next(n); | 1142 | n = rb_next(n); |
1154 | } | 1143 | } |
1155 | |||
1156 | spin_unlock_irqrestore(&cfq_exit_lock, flags); | ||
1157 | } | 1144 | } |
1158 | 1145 | ||
1159 | static struct cfq_io_context * | 1146 | static struct cfq_io_context * |
@@ -1248,15 +1235,12 @@ static inline void changed_ioprio(struct cfq_io_context *cic) | |||
1248 | spin_unlock(cfqd->queue->queue_lock); | 1235 | spin_unlock(cfqd->queue->queue_lock); |
1249 | } | 1236 | } |
1250 | 1237 | ||
1251 | /* | 1238 | static void cfq_ioc_set_ioprio(struct io_context *ioc) |
1252 | * callback from sys_ioprio_set, irqs are disabled | ||
1253 | */ | ||
1254 | static int cfq_ioc_set_ioprio(struct io_context *ioc, unsigned int ioprio) | ||
1255 | { | 1239 | { |
1256 | struct cfq_io_context *cic; | 1240 | struct cfq_io_context *cic; |
1257 | struct rb_node *n; | 1241 | struct rb_node *n; |
1258 | 1242 | ||
1259 | spin_lock(&cfq_exit_lock); | 1243 | ioc->ioprio_changed = 0; |
1260 | 1244 | ||
1261 | n = rb_first(&ioc->cic_root); | 1245 | n = rb_first(&ioc->cic_root); |
1262 | while (n != NULL) { | 1246 | while (n != NULL) { |
@@ -1265,10 +1249,6 @@ static int cfq_ioc_set_ioprio(struct io_context *ioc, unsigned int ioprio) | |||
1265 | changed_ioprio(cic); | 1249 | changed_ioprio(cic); |
1266 | n = rb_next(n); | 1250 | n = rb_next(n); |
1267 | } | 1251 | } |
1268 | |||
1269 | spin_unlock(&cfq_exit_lock); | ||
1270 | |||
1271 | return 0; | ||
1272 | } | 1252 | } |
1273 | 1253 | ||
1274 | static struct cfq_queue * | 1254 | static struct cfq_queue * |
@@ -1336,10 +1316,8 @@ out: | |||
1336 | static void | 1316 | static void |
1337 | cfq_drop_dead_cic(struct io_context *ioc, struct cfq_io_context *cic) | 1317 | cfq_drop_dead_cic(struct io_context *ioc, struct cfq_io_context *cic) |
1338 | { | 1318 | { |
1339 | spin_lock(&cfq_exit_lock); | 1319 | WARN_ON(!list_empty(&cic->queue_list)); |
1340 | rb_erase(&cic->rb_node, &ioc->cic_root); | 1320 | rb_erase(&cic->rb_node, &ioc->cic_root); |
1341 | list_del_init(&cic->queue_list); | ||
1342 | spin_unlock(&cfq_exit_lock); | ||
1343 | kmem_cache_free(cfq_ioc_pool, cic); | 1321 | kmem_cache_free(cfq_ioc_pool, cic); |
1344 | atomic_dec(&ioc_count); | 1322 | atomic_dec(&ioc_count); |
1345 | } | 1323 | } |
@@ -1385,7 +1363,6 @@ cfq_cic_link(struct cfq_data *cfqd, struct io_context *ioc, | |||
1385 | cic->ioc = ioc; | 1363 | cic->ioc = ioc; |
1386 | cic->key = cfqd; | 1364 | cic->key = cfqd; |
1387 | 1365 | ||
1388 | ioc->set_ioprio = cfq_ioc_set_ioprio; | ||
1389 | restart: | 1366 | restart: |
1390 | parent = NULL; | 1367 | parent = NULL; |
1391 | p = &ioc->cic_root.rb_node; | 1368 | p = &ioc->cic_root.rb_node; |
@@ -1407,11 +1384,12 @@ restart: | |||
1407 | BUG(); | 1384 | BUG(); |
1408 | } | 1385 | } |
1409 | 1386 | ||
1410 | spin_lock(&cfq_exit_lock); | ||
1411 | rb_link_node(&cic->rb_node, parent, p); | 1387 | rb_link_node(&cic->rb_node, parent, p); |
1412 | rb_insert_color(&cic->rb_node, &ioc->cic_root); | 1388 | rb_insert_color(&cic->rb_node, &ioc->cic_root); |
1389 | |||
1390 | spin_lock_irq(cfqd->queue->queue_lock); | ||
1413 | list_add(&cic->queue_list, &cfqd->cic_list); | 1391 | list_add(&cic->queue_list, &cfqd->cic_list); |
1414 | spin_unlock(&cfq_exit_lock); | 1392 | spin_unlock_irq(cfqd->queue->queue_lock); |
1415 | } | 1393 | } |
1416 | 1394 | ||
1417 | /* | 1395 | /* |
@@ -1441,6 +1419,10 @@ cfq_get_io_context(struct cfq_data *cfqd, gfp_t gfp_mask) | |||
1441 | 1419 | ||
1442 | cfq_cic_link(cfqd, ioc, cic); | 1420 | cfq_cic_link(cfqd, ioc, cic); |
1443 | out: | 1421 | out: |
1422 | smp_read_barrier_depends(); | ||
1423 | if (unlikely(ioc->ioprio_changed)) | ||
1424 | cfq_ioc_set_ioprio(ioc); | ||
1425 | |||
1444 | return cic; | 1426 | return cic; |
1445 | err: | 1427 | err: |
1446 | put_io_context(ioc); | 1428 | put_io_context(ioc); |
@@ -1945,7 +1927,6 @@ static void cfq_exit_queue(elevator_t *e) | |||
1945 | 1927 | ||
1946 | cfq_shutdown_timer_wq(cfqd); | 1928 | cfq_shutdown_timer_wq(cfqd); |
1947 | 1929 | ||
1948 | spin_lock(&cfq_exit_lock); | ||
1949 | spin_lock_irq(q->queue_lock); | 1930 | spin_lock_irq(q->queue_lock); |
1950 | 1931 | ||
1951 | if (cfqd->active_queue) | 1932 | if (cfqd->active_queue) |
@@ -1960,7 +1941,6 @@ static void cfq_exit_queue(elevator_t *e) | |||
1960 | } | 1941 | } |
1961 | 1942 | ||
1962 | spin_unlock_irq(q->queue_lock); | 1943 | spin_unlock_irq(q->queue_lock); |
1963 | spin_unlock(&cfq_exit_lock); | ||
1964 | 1944 | ||
1965 | cfq_shutdown_timer_wq(cfqd); | 1945 | cfq_shutdown_timer_wq(cfqd); |
1966 | 1946 | ||
@@ -2149,7 +2129,7 @@ static struct elevator_type iosched_cfq = { | |||
2149 | .elevator_may_queue_fn = cfq_may_queue, | 2129 | .elevator_may_queue_fn = cfq_may_queue, |
2150 | .elevator_init_fn = cfq_init_queue, | 2130 | .elevator_init_fn = cfq_init_queue, |
2151 | .elevator_exit_fn = cfq_exit_queue, | 2131 | .elevator_exit_fn = cfq_exit_queue, |
2152 | .trim = cfq_trim, | 2132 | .trim = cfq_free_io_context, |
2153 | }, | 2133 | }, |
2154 | .elevator_attrs = cfq_attrs, | 2134 | .elevator_attrs = cfq_attrs, |
2155 | .elevator_name = "cfq", | 2135 | .elevator_name = "cfq", |
diff --git a/block/ll_rw_blk.c b/block/ll_rw_blk.c index e25b4cd2dcd1..508548b834f1 100644 --- a/block/ll_rw_blk.c +++ b/block/ll_rw_blk.c | |||
@@ -3654,7 +3654,7 @@ struct io_context *current_io_context(gfp_t gfp_flags) | |||
3654 | if (ret) { | 3654 | if (ret) { |
3655 | atomic_set(&ret->refcount, 1); | 3655 | atomic_set(&ret->refcount, 1); |
3656 | ret->task = current; | 3656 | ret->task = current; |
3657 | ret->set_ioprio = NULL; | 3657 | ret->ioprio_changed = 0; |
3658 | ret->last_waited = jiffies; /* doesn't matter... */ | 3658 | ret->last_waited = jiffies; /* doesn't matter... */ |
3659 | ret->nr_batch_requests = 0; /* because this is 0 */ | 3659 | ret->nr_batch_requests = 0; /* because this is 0 */ |
3660 | ret->aic = NULL; | 3660 | ret->aic = NULL; |
diff --git a/fs/ioprio.c b/fs/ioprio.c index 78b1deae3fa2..0fd1089d7bf6 100644 --- a/fs/ioprio.c +++ b/fs/ioprio.c | |||
@@ -47,8 +47,8 @@ static int set_task_ioprio(struct task_struct *task, int ioprio) | |||
47 | /* see wmb() in current_io_context() */ | 47 | /* see wmb() in current_io_context() */ |
48 | smp_read_barrier_depends(); | 48 | smp_read_barrier_depends(); |
49 | 49 | ||
50 | if (ioc && ioc->set_ioprio) | 50 | if (ioc) |
51 | ioc->set_ioprio(ioc, ioprio); | 51 | ioc->ioprio_changed = 1; |
52 | 52 | ||
53 | task_unlock(task); | 53 | task_unlock(task); |
54 | return 0; | 54 | return 0; |
diff --git a/include/linux/blkdev.h b/include/linux/blkdev.h index a1e288069e2e..79cb9fa8034a 100644 --- a/include/linux/blkdev.h +++ b/include/linux/blkdev.h | |||
@@ -90,7 +90,7 @@ struct io_context { | |||
90 | atomic_t refcount; | 90 | atomic_t refcount; |
91 | struct task_struct *task; | 91 | struct task_struct *task; |
92 | 92 | ||
93 | int (*set_ioprio)(struct io_context *, unsigned int); | 93 | unsigned int ioprio_changed; |
94 | 94 | ||
95 | /* | 95 | /* |
96 | * For request batching | 96 | * For request batching |