diff options
| -rw-r--r-- | include/linux/sched.h | 3 | ||||
| -rw-r--r-- | kernel/sched.c | 33 | ||||
| -rw-r--r-- | kernel/sched_debug.c | 1 | ||||
| -rw-r--r-- | kernel/sched_fair.c | 18 | ||||
| -rw-r--r-- | kernel/sched_features.h | 16 |
5 files changed, 0 insertions, 71 deletions
diff --git a/include/linux/sched.h b/include/linux/sched.h index 70c560f5ada..8604884cee8 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h | |||
| @@ -1180,9 +1180,6 @@ struct sched_entity { | |||
| 1180 | u64 vruntime; | 1180 | u64 vruntime; |
| 1181 | u64 prev_sum_exec_runtime; | 1181 | u64 prev_sum_exec_runtime; |
| 1182 | 1182 | ||
| 1183 | u64 last_wakeup; | ||
| 1184 | u64 avg_overlap; | ||
| 1185 | |||
| 1186 | u64 nr_migrations; | 1183 | u64 nr_migrations; |
| 1187 | 1184 | ||
| 1188 | #ifdef CONFIG_SCHEDSTATS | 1185 | #ifdef CONFIG_SCHEDSTATS |
diff --git a/kernel/sched.c b/kernel/sched.c index 35a8626ace7..68ed6f4f3c1 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
| @@ -1887,11 +1887,6 @@ enqueue_task(struct rq *rq, struct task_struct *p, int wakeup, bool head) | |||
| 1887 | 1887 | ||
| 1888 | static void dequeue_task(struct rq *rq, struct task_struct *p, int sleep) | 1888 | static void dequeue_task(struct rq *rq, struct task_struct *p, int sleep) |
| 1889 | { | 1889 | { |
| 1890 | if (sleep && p->se.last_wakeup) { | ||
| 1891 | update_avg(&p->se.avg_overlap, | ||
| 1892 | p->se.sum_exec_runtime - p->se.last_wakeup); | ||
| 1893 | p->se.last_wakeup = 0; | ||
| 1894 | } | ||
| 1895 | sched_info_dequeued(p); | 1890 | sched_info_dequeued(p); |
| 1896 | p->sched_class->dequeue_task(rq, p, sleep); | 1891 | p->sched_class->dequeue_task(rq, p, sleep); |
| 1897 | p->se.on_rq = 0; | 1892 | p->se.on_rq = 0; |
| @@ -2452,15 +2447,6 @@ out_activate: | |||
| 2452 | activate_task(rq, p, 1); | 2447 | activate_task(rq, p, 1); |
| 2453 | success = 1; | 2448 | success = 1; |
| 2454 | 2449 | ||
| 2455 | /* | ||
| 2456 | * Only attribute actual wakeups done by this task. | ||
| 2457 | */ | ||
| 2458 | if (!in_interrupt()) { | ||
| 2459 | struct sched_entity *se = ¤t->se; | ||
| 2460 | |||
| 2461 | se->last_wakeup = se->sum_exec_runtime; | ||
| 2462 | } | ||
| 2463 | |||
| 2464 | out_running: | 2450 | out_running: |
| 2465 | trace_sched_wakeup(rq, p, success); | 2451 | trace_sched_wakeup(rq, p, success); |
| 2466 | check_preempt_curr(rq, p, wake_flags); | 2452 | check_preempt_curr(rq, p, wake_flags); |
| @@ -2522,8 +2508,6 @@ static void __sched_fork(struct task_struct *p) | |||
| 2522 | p->se.sum_exec_runtime = 0; | 2508 | p->se.sum_exec_runtime = 0; |
| 2523 | p->se.prev_sum_exec_runtime = 0; | 2509 | p->se.prev_sum_exec_runtime = 0; |
| 2524 | p->se.nr_migrations = 0; | 2510 | p->se.nr_migrations = 0; |
| 2525 | p->se.last_wakeup = 0; | ||
| 2526 | p->se.avg_overlap = 0; | ||
| 2527 | 2511 | ||
| 2528 | #ifdef CONFIG_SCHEDSTATS | 2512 | #ifdef CONFIG_SCHEDSTATS |
| 2529 | memset(&p->se.statistics, 0, sizeof(p->se.statistics)); | 2513 | memset(&p->se.statistics, 0, sizeof(p->se.statistics)); |
| @@ -3594,23 +3578,6 @@ static inline void schedule_debug(struct task_struct *prev) | |||
| 3594 | 3578 | ||
| 3595 | static void put_prev_task(struct rq *rq, struct task_struct *prev) | 3579 | static void put_prev_task(struct rq *rq, struct task_struct *prev) |
| 3596 | { | 3580 | { |
| 3597 | if (prev->state == TASK_RUNNING) { | ||
| 3598 | u64 runtime = prev->se.sum_exec_runtime; | ||
| 3599 | |||
| 3600 | runtime -= prev->se.prev_sum_exec_runtime; | ||
| 3601 | runtime = min_t(u64, runtime, 2*sysctl_sched_migration_cost); | ||
| 3602 | |||
| 3603 | /* | ||
| 3604 | * In order to avoid avg_overlap growing stale when we are | ||
| 3605 | * indeed overlapping and hence not getting put to sleep, grow | ||
| 3606 | * the avg_overlap on preemption. | ||
| 3607 | * | ||
| 3608 | * We use the average preemption runtime because that | ||
| 3609 | * correlates to the amount of cache footprint a task can | ||
| 3610 | * build up. | ||
| 3611 | */ | ||
| 3612 | update_avg(&prev->se.avg_overlap, runtime); | ||
| 3613 | } | ||
| 3614 | prev->sched_class->put_prev_task(rq, prev); | 3581 | prev->sched_class->put_prev_task(rq, prev); |
| 3615 | } | 3582 | } |
| 3616 | 3583 | ||
diff --git a/kernel/sched_debug.c b/kernel/sched_debug.c index 20b95a420fe..8a46a719f36 100644 --- a/kernel/sched_debug.c +++ b/kernel/sched_debug.c | |||
| @@ -407,7 +407,6 @@ void proc_sched_show_task(struct task_struct *p, struct seq_file *m) | |||
| 407 | PN(se.exec_start); | 407 | PN(se.exec_start); |
| 408 | PN(se.vruntime); | 408 | PN(se.vruntime); |
| 409 | PN(se.sum_exec_runtime); | 409 | PN(se.sum_exec_runtime); |
| 410 | PN(se.avg_overlap); | ||
| 411 | 410 | ||
| 412 | nr_switches = p->nvcsw + p->nivcsw; | 411 | nr_switches = p->nvcsw + p->nivcsw; |
| 413 | 412 | ||
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c index 6fc62854422..c3b69d4b5d6 100644 --- a/kernel/sched_fair.c +++ b/kernel/sched_fair.c | |||
| @@ -1241,7 +1241,6 @@ static inline unsigned long effective_load(struct task_group *tg, int cpu, | |||
| 1241 | 1241 | ||
| 1242 | static int wake_affine(struct sched_domain *sd, struct task_struct *p, int sync) | 1242 | static int wake_affine(struct sched_domain *sd, struct task_struct *p, int sync) |
| 1243 | { | 1243 | { |
| 1244 | struct task_struct *curr = current; | ||
| 1245 | unsigned long this_load, load; | 1244 | unsigned long this_load, load; |
| 1246 | int idx, this_cpu, prev_cpu; | 1245 | int idx, this_cpu, prev_cpu; |
| 1247 | unsigned long tl_per_task; | 1246 | unsigned long tl_per_task; |
| @@ -1256,18 +1255,6 @@ static int wake_affine(struct sched_domain *sd, struct task_struct *p, int sync) | |||
| 1256 | load = source_load(prev_cpu, idx); | 1255 | load = source_load(prev_cpu, idx); |
| 1257 | this_load = target_load(this_cpu, idx); | 1256 | this_load = target_load(this_cpu, idx); |
| 1258 | 1257 | ||
| 1259 | if (sync) { | ||
| 1260 | if (sched_feat(SYNC_LESS) && | ||
| 1261 | (curr->se.avg_overlap > sysctl_sched_migration_cost || | ||
| 1262 | p->se.avg_overlap > sysctl_sched_migration_cost)) | ||
| 1263 | sync = 0; | ||
| 1264 | } else { | ||
| 1265 | if (sched_feat(SYNC_MORE) && | ||
| 1266 | (curr->se.avg_overlap < sysctl_sched_migration_cost && | ||
| 1267 | p->se.avg_overlap < sysctl_sched_migration_cost)) | ||
| 1268 | sync = 1; | ||
| 1269 | } | ||
| 1270 | |||
| 1271 | /* | 1258 | /* |
| 1272 | * If sync wakeup then subtract the (maximum possible) | 1259 | * If sync wakeup then subtract the (maximum possible) |
| 1273 | * effect of the currently running task from the load | 1260 | * effect of the currently running task from the load |
| @@ -1711,11 +1698,6 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p, int wake_ | |||
| 1711 | if (sched_feat(WAKEUP_SYNC) && sync) | 1698 | if (sched_feat(WAKEUP_SYNC) && sync) |
| 1712 | goto preempt; | 1699 | goto preempt; |
| 1713 | 1700 | ||
| 1714 | if (sched_feat(WAKEUP_OVERLAP) && | ||
| 1715 | se->avg_overlap < sysctl_sched_migration_cost && | ||
| 1716 | pse->avg_overlap < sysctl_sched_migration_cost) | ||
| 1717 | goto preempt; | ||
| 1718 | |||
| 1719 | if (!sched_feat(WAKEUP_PREEMPT)) | 1701 | if (!sched_feat(WAKEUP_PREEMPT)) |
| 1720 | return; | 1702 | return; |
| 1721 | 1703 | ||
diff --git a/kernel/sched_features.h b/kernel/sched_features.h index 96ef5dbc66e..c545e048dfe 100644 --- a/kernel/sched_features.h +++ b/kernel/sched_features.h | |||
| @@ -42,12 +42,6 @@ SCHED_FEAT(ASYM_GRAN, 1) | |||
| 42 | SCHED_FEAT(WAKEUP_SYNC, 0) | 42 | SCHED_FEAT(WAKEUP_SYNC, 0) |
| 43 | 43 | ||
| 44 | /* | 44 | /* |
| 45 | * Wakeup preempt based on task behaviour. Tasks that do not overlap | ||
| 46 | * don't get preempted. | ||
| 47 | */ | ||
| 48 | SCHED_FEAT(WAKEUP_OVERLAP, 0) | ||
| 49 | |||
| 50 | /* | ||
| 51 | * Use the SYNC wakeup hint, pipes and the likes use this to indicate | 45 | * Use the SYNC wakeup hint, pipes and the likes use this to indicate |
| 52 | * the remote end is likely to consume the data we just wrote, and | 46 | * the remote end is likely to consume the data we just wrote, and |
| 53 | * therefore has cache benefit from being placed on the same cpu, see | 47 | * therefore has cache benefit from being placed on the same cpu, see |
| @@ -64,16 +58,6 @@ SCHED_FEAT(SYNC_WAKEUPS, 1) | |||
| 64 | SCHED_FEAT(AFFINE_WAKEUPS, 1) | 58 | SCHED_FEAT(AFFINE_WAKEUPS, 1) |
| 65 | 59 | ||
| 66 | /* | 60 | /* |
| 67 | * Weaken SYNC hint based on overlap | ||
| 68 | */ | ||
| 69 | SCHED_FEAT(SYNC_LESS, 1) | ||
| 70 | |||
| 71 | /* | ||
| 72 | * Add SYNC hint based on overlap | ||
| 73 | */ | ||
| 74 | SCHED_FEAT(SYNC_MORE, 0) | ||
| 75 | |||
| 76 | /* | ||
| 77 | * Prefer to schedule the task we woke last (assuming it failed | 61 | * Prefer to schedule the task we woke last (assuming it failed |
| 78 | * wakeup-preemption), since its likely going to consume data we | 62 | * wakeup-preemption), since its likely going to consume data we |
| 79 | * touched, increases cache locality. | 63 | * touched, increases cache locality. |
