diff options
Diffstat (limited to 'kernel/sched.c')
-rw-r--r-- | kernel/sched.c | 42 |
1 files changed, 7 insertions, 35 deletions
diff --git a/kernel/sched.c b/kernel/sched.c index d4dabfcc776c..25cc9b2a8c15 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -1972,42 +1972,11 @@ unsigned long nr_active(void) | |||
1972 | */ | 1972 | */ |
1973 | static void update_cpu_load(struct rq *this_rq) | 1973 | static void update_cpu_load(struct rq *this_rq) |
1974 | { | 1974 | { |
1975 | u64 fair_delta64, exec_delta64, idle_delta64, sample_interval64, tmp64; | ||
1976 | unsigned long total_load = this_rq->ls.load.weight; | 1975 | unsigned long total_load = this_rq->ls.load.weight; |
1977 | unsigned long this_load = total_load; | 1976 | unsigned long this_load = total_load; |
1978 | struct load_stat *ls = &this_rq->ls; | ||
1979 | int i, scale; | 1977 | int i, scale; |
1980 | 1978 | ||
1981 | this_rq->nr_load_updates++; | 1979 | this_rq->nr_load_updates++; |
1982 | if (unlikely(!(sysctl_sched_features & SCHED_FEAT_PRECISE_CPU_LOAD))) | ||
1983 | goto do_avg; | ||
1984 | |||
1985 | /* Update delta_fair/delta_exec fields first */ | ||
1986 | update_curr_load(this_rq); | ||
1987 | |||
1988 | fair_delta64 = ls->delta_fair + 1; | ||
1989 | ls->delta_fair = 0; | ||
1990 | |||
1991 | exec_delta64 = ls->delta_exec + 1; | ||
1992 | ls->delta_exec = 0; | ||
1993 | |||
1994 | sample_interval64 = this_rq->clock - ls->load_update_last; | ||
1995 | ls->load_update_last = this_rq->clock; | ||
1996 | |||
1997 | if ((s64)sample_interval64 < (s64)TICK_NSEC) | ||
1998 | sample_interval64 = TICK_NSEC; | ||
1999 | |||
2000 | if (exec_delta64 > sample_interval64) | ||
2001 | exec_delta64 = sample_interval64; | ||
2002 | |||
2003 | idle_delta64 = sample_interval64 - exec_delta64; | ||
2004 | |||
2005 | tmp64 = div64_64(SCHED_LOAD_SCALE * exec_delta64, fair_delta64); | ||
2006 | tmp64 = div64_64(tmp64 * exec_delta64, sample_interval64); | ||
2007 | |||
2008 | this_load = (unsigned long)tmp64; | ||
2009 | |||
2010 | do_avg: | ||
2011 | 1980 | ||
2012 | /* Update our load: */ | 1981 | /* Update our load: */ |
2013 | for (i = 0, scale = 1; i < CPU_LOAD_IDX_MAX; i++, scale += scale) { | 1982 | for (i = 0, scale = 1; i < CPU_LOAD_IDX_MAX; i++, scale += scale) { |
@@ -2017,7 +1986,13 @@ do_avg: | |||
2017 | 1986 | ||
2018 | old_load = this_rq->cpu_load[i]; | 1987 | old_load = this_rq->cpu_load[i]; |
2019 | new_load = this_load; | 1988 | new_load = this_load; |
2020 | 1989 | /* | |
1990 | * Round up the averaging division if load is increasing. This | ||
1991 | * prevents us from getting stuck on 9 if the load is 10, for | ||
1992 | * example. | ||
1993 | */ | ||
1994 | if (new_load > old_load) | ||
1995 | new_load += scale-1; | ||
2021 | this_rq->cpu_load[i] = (old_load*(scale-1) + new_load) >> i; | 1996 | this_rq->cpu_load[i] = (old_load*(scale-1) + new_load) >> i; |
2022 | } | 1997 | } |
2023 | } | 1998 | } |
@@ -6484,7 +6459,6 @@ static inline void init_cfs_rq(struct cfs_rq *cfs_rq, struct rq *rq) | |||
6484 | 6459 | ||
6485 | void __init sched_init(void) | 6460 | void __init sched_init(void) |
6486 | { | 6461 | { |
6487 | u64 now = sched_clock(); | ||
6488 | int highest_cpu = 0; | 6462 | int highest_cpu = 0; |
6489 | int i, j; | 6463 | int i, j; |
6490 | 6464 | ||
@@ -6509,8 +6483,6 @@ void __init sched_init(void) | |||
6509 | INIT_LIST_HEAD(&rq->leaf_cfs_rq_list); | 6483 | INIT_LIST_HEAD(&rq->leaf_cfs_rq_list); |
6510 | list_add(&rq->cfs.leaf_cfs_rq_list, &rq->leaf_cfs_rq_list); | 6484 | list_add(&rq->cfs.leaf_cfs_rq_list, &rq->leaf_cfs_rq_list); |
6511 | #endif | 6485 | #endif |
6512 | rq->ls.load_update_last = now; | ||
6513 | rq->ls.load_update_start = now; | ||
6514 | 6486 | ||
6515 | for (j = 0; j < CPU_LOAD_IDX_MAX; j++) | 6487 | for (j = 0; j < CPU_LOAD_IDX_MAX; j++) |
6516 | rq->cpu_load[j] = 0; | 6488 | rq->cpu_load[j] = 0; |