diff options
-rw-r--r-- | kernel/sched.c | 29 | ||||
-rw-r--r-- | kernel/sched_fair.c | 16 |
2 files changed, 32 insertions, 13 deletions
diff --git a/kernel/sched.c b/kernel/sched.c index 3878f5018007..b54ecf84b6be 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -814,6 +814,7 @@ const_debug unsigned int sysctl_sched_nr_migrate = 32; | |||
814 | * default: 0.25ms | 814 | * default: 0.25ms |
815 | */ | 815 | */ |
816 | unsigned int sysctl_sched_shares_ratelimit = 250000; | 816 | unsigned int sysctl_sched_shares_ratelimit = 250000; |
817 | unsigned int normalized_sysctl_sched_shares_ratelimit = 250000; | ||
817 | 818 | ||
818 | /* | 819 | /* |
819 | * Inject some fuzzyness into changing the per-cpu group shares | 820 | * Inject some fuzzyness into changing the per-cpu group shares |
@@ -1814,6 +1815,7 @@ static void cfs_rq_set_shares(struct cfs_rq *cfs_rq, unsigned long shares) | |||
1814 | #endif | 1815 | #endif |
1815 | 1816 | ||
1816 | static void calc_load_account_active(struct rq *this_rq); | 1817 | static void calc_load_account_active(struct rq *this_rq); |
1818 | static void update_sysctl(void); | ||
1817 | 1819 | ||
1818 | static inline void __set_task_cpu(struct task_struct *p, unsigned int cpu) | 1820 | static inline void __set_task_cpu(struct task_struct *p, unsigned int cpu) |
1819 | { | 1821 | { |
@@ -7028,22 +7030,23 @@ cpumask_var_t nohz_cpu_mask; | |||
7028 | * | 7030 | * |
7029 | * This idea comes from the SD scheduler of Con Kolivas: | 7031 | * This idea comes from the SD scheduler of Con Kolivas: |
7030 | */ | 7032 | */ |
7031 | static inline void sched_init_granularity(void) | 7033 | static void update_sysctl(void) |
7032 | { | 7034 | { |
7033 | unsigned int factor = 1 + ilog2(num_online_cpus()); | 7035 | unsigned int cpus = min(num_online_cpus(), 8U); |
7034 | const unsigned long limit = 200000000; | 7036 | unsigned int factor = 1 + ilog2(cpus); |
7035 | |||
7036 | sysctl_sched_min_granularity *= factor; | ||
7037 | if (sysctl_sched_min_granularity > limit) | ||
7038 | sysctl_sched_min_granularity = limit; | ||
7039 | |||
7040 | sysctl_sched_latency *= factor; | ||
7041 | if (sysctl_sched_latency > limit) | ||
7042 | sysctl_sched_latency = limit; | ||
7043 | 7037 | ||
7044 | sysctl_sched_wakeup_granularity *= factor; | 7038 | #define SET_SYSCTL(name) \ |
7039 | (sysctl_##name = (factor) * normalized_sysctl_##name) | ||
7040 | SET_SYSCTL(sched_min_granularity); | ||
7041 | SET_SYSCTL(sched_latency); | ||
7042 | SET_SYSCTL(sched_wakeup_granularity); | ||
7043 | SET_SYSCTL(sched_shares_ratelimit); | ||
7044 | #undef SET_SYSCTL | ||
7045 | } | ||
7045 | 7046 | ||
7046 | sysctl_sched_shares_ratelimit *= factor; | 7047 | static inline void sched_init_granularity(void) |
7048 | { | ||
7049 | update_sysctl(); | ||
7047 | } | 7050 | } |
7048 | 7051 | ||
7049 | #ifdef CONFIG_SMP | 7052 | #ifdef CONFIG_SMP |
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c index c163a285bf05..71b3458245e5 100644 --- a/kernel/sched_fair.c +++ b/kernel/sched_fair.c | |||
@@ -35,12 +35,14 @@ | |||
35 | * run vmstat and monitor the context-switches (cs) field) | 35 | * run vmstat and monitor the context-switches (cs) field) |
36 | */ | 36 | */ |
37 | unsigned int sysctl_sched_latency = 5000000ULL; | 37 | unsigned int sysctl_sched_latency = 5000000ULL; |
38 | unsigned int normalized_sysctl_sched_latency = 5000000ULL; | ||
38 | 39 | ||
39 | /* | 40 | /* |
40 | * Minimal preemption granularity for CPU-bound tasks: | 41 | * Minimal preemption granularity for CPU-bound tasks: |
41 | * (default: 1 msec * (1 + ilog(ncpus)), units: nanoseconds) | 42 | * (default: 1 msec * (1 + ilog(ncpus)), units: nanoseconds) |
42 | */ | 43 | */ |
43 | unsigned int sysctl_sched_min_granularity = 1000000ULL; | 44 | unsigned int sysctl_sched_min_granularity = 1000000ULL; |
45 | unsigned int normalized_sysctl_sched_min_granularity = 1000000ULL; | ||
44 | 46 | ||
45 | /* | 47 | /* |
46 | * is kept at sysctl_sched_latency / sysctl_sched_min_granularity | 48 | * is kept at sysctl_sched_latency / sysctl_sched_min_granularity |
@@ -70,6 +72,7 @@ unsigned int __read_mostly sysctl_sched_compat_yield; | |||
70 | * have immediate wakeup/sleep latencies. | 72 | * have immediate wakeup/sleep latencies. |
71 | */ | 73 | */ |
72 | unsigned int sysctl_sched_wakeup_granularity = 1000000UL; | 74 | unsigned int sysctl_sched_wakeup_granularity = 1000000UL; |
75 | unsigned int normalized_sysctl_sched_wakeup_granularity = 1000000UL; | ||
73 | 76 | ||
74 | const_debug unsigned int sysctl_sched_migration_cost = 500000UL; | 77 | const_debug unsigned int sysctl_sched_migration_cost = 500000UL; |
75 | 78 | ||
@@ -1890,6 +1893,17 @@ move_one_task_fair(struct rq *this_rq, int this_cpu, struct rq *busiest, | |||
1890 | 1893 | ||
1891 | return 0; | 1894 | return 0; |
1892 | } | 1895 | } |
1896 | |||
1897 | static void rq_online_fair(struct rq *rq) | ||
1898 | { | ||
1899 | update_sysctl(); | ||
1900 | } | ||
1901 | |||
1902 | static void rq_offline_fair(struct rq *rq) | ||
1903 | { | ||
1904 | update_sysctl(); | ||
1905 | } | ||
1906 | |||
1893 | #endif /* CONFIG_SMP */ | 1907 | #endif /* CONFIG_SMP */ |
1894 | 1908 | ||
1895 | /* | 1909 | /* |
@@ -2035,6 +2049,8 @@ static const struct sched_class fair_sched_class = { | |||
2035 | 2049 | ||
2036 | .load_balance = load_balance_fair, | 2050 | .load_balance = load_balance_fair, |
2037 | .move_one_task = move_one_task_fair, | 2051 | .move_one_task = move_one_task_fair, |
2052 | .rq_online = rq_online_fair, | ||
2053 | .rq_offline = rq_offline_fair, | ||
2038 | #endif | 2054 | #endif |
2039 | 2055 | ||
2040 | .set_curr_task = set_curr_task_fair, | 2056 | .set_curr_task = set_curr_task_fair, |