diff options
| author | Jeff Moyer <jmoyer@redhat.com> | 2009-10-23 17:14:52 -0400 |
|---|---|---|
| committer | Jens Axboe <jens.axboe@oracle.com> | 2009-10-26 09:34:47 -0400 |
| commit | e6c5bc737ab71e4af6025ef7d150f5a26ae5f146 (patch) | |
| tree | 01127225a83bdcae30b261b9f21bb89faa7db8ce | |
| parent | b3b6d0408c953524f979468562e7e210d8634150 (diff) | |
cfq: break apart merged cfqqs if they stop cooperating
cfq_queues are merged if they are issuing requests within the mean seek
distance of one another. This patch detects when the coopearting stops and
breaks the queues back up.
Signed-off-by: Jeff Moyer <jmoyer@redhat.com>
Signed-off-by: Jens Axboe <jens.axboe@oracle.com>
| -rw-r--r-- | block/cfq-iosched.c | 79 |
1 files changed, 76 insertions, 3 deletions
diff --git a/block/cfq-iosched.c b/block/cfq-iosched.c index 5e01a0a92c02..47d6aaca0c51 100644 --- a/block/cfq-iosched.c +++ b/block/cfq-iosched.c | |||
| @@ -38,6 +38,12 @@ static int cfq_slice_idle = HZ / 125; | |||
| 38 | */ | 38 | */ |
| 39 | #define CFQ_MIN_TT (2) | 39 | #define CFQ_MIN_TT (2) |
| 40 | 40 | ||
| 41 | /* | ||
| 42 | * Allow merged cfqqs to perform this amount of seeky I/O before | ||
| 43 | * deciding to break the queues up again. | ||
| 44 | */ | ||
| 45 | #define CFQQ_COOP_TOUT (HZ) | ||
| 46 | |||
| 41 | #define CFQ_SLICE_SCALE (5) | 47 | #define CFQ_SLICE_SCALE (5) |
| 42 | #define CFQ_HW_QUEUE_MIN (5) | 48 | #define CFQ_HW_QUEUE_MIN (5) |
| 43 | 49 | ||
| @@ -116,6 +122,7 @@ struct cfq_queue { | |||
| 116 | u64 seek_total; | 122 | u64 seek_total; |
| 117 | sector_t seek_mean; | 123 | sector_t seek_mean; |
| 118 | sector_t last_request_pos; | 124 | sector_t last_request_pos; |
| 125 | unsigned long seeky_start; | ||
| 119 | 126 | ||
| 120 | pid_t pid; | 127 | pid_t pid; |
| 121 | 128 | ||
| @@ -1036,6 +1043,11 @@ static struct cfq_queue *cfq_close_cooperator(struct cfq_data *cfqd, | |||
| 1036 | { | 1043 | { |
| 1037 | struct cfq_queue *cfqq; | 1044 | struct cfq_queue *cfqq; |
| 1038 | 1045 | ||
| 1046 | if (!cfq_cfqq_sync(cur_cfqq)) | ||
| 1047 | return NULL; | ||
| 1048 | if (CFQQ_SEEKY(cur_cfqq)) | ||
| 1049 | return NULL; | ||
| 1050 | |||
| 1039 | /* | 1051 | /* |
| 1040 | * We should notice if some of the queues are cooperating, eg | 1052 | * We should notice if some of the queues are cooperating, eg |
| 1041 | * working closely on the same area of the disk. In that case, | 1053 | * working closely on the same area of the disk. In that case, |
| @@ -1050,6 +1062,8 @@ static struct cfq_queue *cfq_close_cooperator(struct cfq_data *cfqd, | |||
| 1050 | */ | 1062 | */ |
| 1051 | if (!cfq_cfqq_sync(cfqq)) | 1063 | if (!cfq_cfqq_sync(cfqq)) |
| 1052 | return NULL; | 1064 | return NULL; |
| 1065 | if (CFQQ_SEEKY(cfqq)) | ||
| 1066 | return NULL; | ||
| 1053 | 1067 | ||
| 1054 | return cfqq; | 1068 | return cfqq; |
| 1055 | } | 1069 | } |
| @@ -1181,7 +1195,7 @@ static int cfqq_process_refs(struct cfq_queue *cfqq) | |||
| 1181 | 1195 | ||
| 1182 | static void cfq_setup_merge(struct cfq_queue *cfqq, struct cfq_queue *new_cfqq) | 1196 | static void cfq_setup_merge(struct cfq_queue *cfqq, struct cfq_queue *new_cfqq) |
| 1183 | { | 1197 | { |
| 1184 | int process_refs; | 1198 | int process_refs, new_process_refs; |
| 1185 | struct cfq_queue *__cfqq; | 1199 | struct cfq_queue *__cfqq; |
| 1186 | 1200 | ||
| 1187 | /* Avoid a circular list and skip interim queue merges */ | 1201 | /* Avoid a circular list and skip interim queue merges */ |
| @@ -1199,8 +1213,17 @@ static void cfq_setup_merge(struct cfq_queue *cfqq, struct cfq_queue *new_cfqq) | |||
| 1199 | if (process_refs == 0) | 1213 | if (process_refs == 0) |
| 1200 | return; | 1214 | return; |
| 1201 | 1215 | ||
| 1202 | cfqq->new_cfqq = new_cfqq; | 1216 | /* |
| 1203 | atomic_add(process_refs, &new_cfqq->ref); | 1217 | * Merge in the direction of the lesser amount of work. |
| 1218 | */ | ||
| 1219 | new_process_refs = cfqq_process_refs(new_cfqq); | ||
| 1220 | if (new_process_refs >= process_refs) { | ||
| 1221 | cfqq->new_cfqq = new_cfqq; | ||
| 1222 | atomic_add(process_refs, &new_cfqq->ref); | ||
| 1223 | } else { | ||
| 1224 | new_cfqq->new_cfqq = cfqq; | ||
| 1225 | atomic_add(new_process_refs, &cfqq->ref); | ||
| 1226 | } | ||
| 1204 | } | 1227 | } |
| 1205 | 1228 | ||
| 1206 | /* | 1229 | /* |
| @@ -2029,6 +2052,19 @@ cfq_update_io_seektime(struct cfq_data *cfqd, struct cfq_queue *cfqq, | |||
| 2029 | total = cfqq->seek_total + (cfqq->seek_samples/2); | 2052 | total = cfqq->seek_total + (cfqq->seek_samples/2); |
| 2030 | do_div(total, cfqq->seek_samples); | 2053 | do_div(total, cfqq->seek_samples); |
| 2031 | cfqq->seek_mean = (sector_t)total; | 2054 | cfqq->seek_mean = (sector_t)total; |
| 2055 | |||
| 2056 | /* | ||
| 2057 | * If this cfqq is shared between multiple processes, check to | ||
| 2058 | * make sure that those processes are still issuing I/Os within | ||
| 2059 | * the mean seek distance. If not, it may be time to break the | ||
| 2060 | * queues apart again. | ||
| 2061 | */ | ||
| 2062 | if (cfq_cfqq_coop(cfqq)) { | ||
| 2063 | if (CFQQ_SEEKY(cfqq) && !cfqq->seeky_start) | ||
| 2064 | cfqq->seeky_start = jiffies; | ||
| 2065 | else if (!CFQQ_SEEKY(cfqq)) | ||
| 2066 | cfqq->seeky_start = 0; | ||
| 2067 | } | ||
| 2032 | } | 2068 | } |
| 2033 | 2069 | ||
| 2034 | /* | 2070 | /* |
| @@ -2391,6 +2427,32 @@ cfq_merge_cfqqs(struct cfq_data *cfqd, struct cfq_io_context *cic, | |||
| 2391 | return cic_to_cfqq(cic, 1); | 2427 | return cic_to_cfqq(cic, 1); |
| 2392 | } | 2428 | } |
| 2393 | 2429 | ||
| 2430 | static int should_split_cfqq(struct cfq_queue *cfqq) | ||
| 2431 | { | ||
| 2432 | if (cfqq->seeky_start && | ||
| 2433 | time_after(jiffies, cfqq->seeky_start + CFQQ_COOP_TOUT)) | ||
| 2434 | return 1; | ||
| 2435 | return 0; | ||
| 2436 | } | ||
| 2437 | |||
| 2438 | /* | ||
| 2439 | * Returns NULL if a new cfqq should be allocated, or the old cfqq if this | ||
| 2440 | * was the last process referring to said cfqq. | ||
| 2441 | */ | ||
| 2442 | static struct cfq_queue * | ||
| 2443 | split_cfqq(struct cfq_io_context *cic, struct cfq_queue *cfqq) | ||
| 2444 | { | ||
| 2445 | if (cfqq_process_refs(cfqq) == 1) { | ||
| 2446 | cfqq->seeky_start = 0; | ||
| 2447 | cfqq->pid = current->pid; | ||
| 2448 | cfq_clear_cfqq_coop(cfqq); | ||
| 2449 | return cfqq; | ||
| 2450 | } | ||
| 2451 | |||
| 2452 | cic_set_cfqq(cic, NULL, 1); | ||
| 2453 | cfq_put_queue(cfqq); | ||
| 2454 | return NULL; | ||
| 2455 | } | ||
| 2394 | /* | 2456 | /* |
| 2395 | * Allocate cfq data structures associated with this request. | 2457 | * Allocate cfq data structures associated with this request. |
| 2396 | */ | 2458 | */ |
| @@ -2413,12 +2475,23 @@ cfq_set_request(struct request_queue *q, struct request *rq, gfp_t gfp_mask) | |||
| 2413 | if (!cic) | 2475 | if (!cic) |
| 2414 | goto queue_fail; | 2476 | goto queue_fail; |
| 2415 | 2477 | ||
| 2478 | new_queue: | ||
| 2416 | cfqq = cic_to_cfqq(cic, is_sync); | 2479 | cfqq = cic_to_cfqq(cic, is_sync); |
| 2417 | if (!cfqq || cfqq == &cfqd->oom_cfqq) { | 2480 | if (!cfqq || cfqq == &cfqd->oom_cfqq) { |
| 2418 | cfqq = cfq_get_queue(cfqd, is_sync, cic->ioc, gfp_mask); | 2481 | cfqq = cfq_get_queue(cfqd, is_sync, cic->ioc, gfp_mask); |
| 2419 | cic_set_cfqq(cic, cfqq, is_sync); | 2482 | cic_set_cfqq(cic, cfqq, is_sync); |
| 2420 | } else { | 2483 | } else { |
| 2421 | /* | 2484 | /* |
| 2485 | * If the queue was seeky for too long, break it apart. | ||
| 2486 | */ | ||
| 2487 | if (cfq_cfqq_coop(cfqq) && should_split_cfqq(cfqq)) { | ||
| 2488 | cfq_log_cfqq(cfqd, cfqq, "breaking apart cfqq"); | ||
| 2489 | cfqq = split_cfqq(cic, cfqq); | ||
| 2490 | if (!cfqq) | ||
| 2491 | goto new_queue; | ||
| 2492 | } | ||
| 2493 | |||
| 2494 | /* | ||
| 2422 | * Check to see if this queue is scheduled to merge with | 2495 | * Check to see if this queue is scheduled to merge with |
| 2423 | * another, closely cooperating queue. The merging of | 2496 | * another, closely cooperating queue. The merging of |
| 2424 | * queues happens here as it must be done in process context. | 2497 | * queues happens here as it must be done in process context. |
