diff options
Diffstat (limited to 'kernel/sched.c')
-rw-r--r-- | kernel/sched.c | 66 |
1 files changed, 43 insertions, 23 deletions
diff --git a/kernel/sched.c b/kernel/sched.c index deb5ac8c12f3..242d0d47a70d 100644 --- a/kernel/sched.c +++ b/kernel/sched.c | |||
@@ -125,6 +125,9 @@ DEFINE_TRACE(sched_switch); | |||
125 | DEFINE_TRACE(sched_migrate_task); | 125 | DEFINE_TRACE(sched_migrate_task); |
126 | 126 | ||
127 | #ifdef CONFIG_SMP | 127 | #ifdef CONFIG_SMP |
128 | |||
129 | static void double_rq_lock(struct rq *rq1, struct rq *rq2); | ||
130 | |||
128 | /* | 131 | /* |
129 | * Divide a load by a sched group cpu_power : (load / sg->__cpu_power) | 132 | * Divide a load by a sched group cpu_power : (load / sg->__cpu_power) |
130 | * Since cpu_power is a 'constant', we can use a reciprocal divide. | 133 | * Since cpu_power is a 'constant', we can use a reciprocal divide. |
@@ -1320,8 +1323,8 @@ static inline void update_load_sub(struct load_weight *lw, unsigned long dec) | |||
1320 | * slice expiry etc. | 1323 | * slice expiry etc. |
1321 | */ | 1324 | */ |
1322 | 1325 | ||
1323 | #define WEIGHT_IDLEPRIO 2 | 1326 | #define WEIGHT_IDLEPRIO 3 |
1324 | #define WMULT_IDLEPRIO (1 << 31) | 1327 | #define WMULT_IDLEPRIO 1431655765 |
1325 | 1328 | ||
1326 | /* | 1329 | /* |
1327 | * Nice levels are multiplicative, with a gentle 10% change for every | 1330 | * Nice levels are multiplicative, with a gentle 10% change for every |
@@ -2263,6 +2266,16 @@ static int try_to_wake_up(struct task_struct *p, unsigned int state, int sync) | |||
2263 | if (!sched_feat(SYNC_WAKEUPS)) | 2266 | if (!sched_feat(SYNC_WAKEUPS)) |
2264 | sync = 0; | 2267 | sync = 0; |
2265 | 2268 | ||
2269 | if (!sync) { | ||
2270 | if (current->se.avg_overlap < sysctl_sched_migration_cost && | ||
2271 | p->se.avg_overlap < sysctl_sched_migration_cost) | ||
2272 | sync = 1; | ||
2273 | } else { | ||
2274 | if (current->se.avg_overlap >= sysctl_sched_migration_cost || | ||
2275 | p->se.avg_overlap >= sysctl_sched_migration_cost) | ||
2276 | sync = 0; | ||
2277 | } | ||
2278 | |||
2266 | #ifdef CONFIG_SMP | 2279 | #ifdef CONFIG_SMP |
2267 | if (sched_feat(LB_WAKEUP_UPDATE)) { | 2280 | if (sched_feat(LB_WAKEUP_UPDATE)) { |
2268 | struct sched_domain *sd; | 2281 | struct sched_domain *sd; |
@@ -4437,7 +4450,7 @@ void __kprobes sub_preempt_count(int val) | |||
4437 | /* | 4450 | /* |
4438 | * Underflow? | 4451 | * Underflow? |
4439 | */ | 4452 | */ |
4440 | if (DEBUG_LOCKS_WARN_ON(val > preempt_count() - (!!kernel_locked()))) | 4453 | if (DEBUG_LOCKS_WARN_ON(val > preempt_count())) |
4441 | return; | 4454 | return; |
4442 | /* | 4455 | /* |
4443 | * Is the spinlock portion underflowing? | 4456 | * Is the spinlock portion underflowing? |
@@ -5123,7 +5136,7 @@ int can_nice(const struct task_struct *p, const int nice) | |||
5123 | * sys_setpriority is a more generic, but much slower function that | 5136 | * sys_setpriority is a more generic, but much slower function that |
5124 | * does similar things. | 5137 | * does similar things. |
5125 | */ | 5138 | */ |
5126 | asmlinkage long sys_nice(int increment) | 5139 | SYSCALL_DEFINE1(nice, int, increment) |
5127 | { | 5140 | { |
5128 | long nice, retval; | 5141 | long nice, retval; |
5129 | 5142 | ||
@@ -5430,8 +5443,8 @@ do_sched_setscheduler(pid_t pid, int policy, struct sched_param __user *param) | |||
5430 | * @policy: new policy. | 5443 | * @policy: new policy. |
5431 | * @param: structure containing the new RT priority. | 5444 | * @param: structure containing the new RT priority. |
5432 | */ | 5445 | */ |
5433 | asmlinkage long | 5446 | SYSCALL_DEFINE3(sched_setscheduler, pid_t, pid, int, policy, |
5434 | sys_sched_setscheduler(pid_t pid, int policy, struct sched_param __user *param) | 5447 | struct sched_param __user *, param) |
5435 | { | 5448 | { |
5436 | /* negative values for policy are not valid */ | 5449 | /* negative values for policy are not valid */ |
5437 | if (policy < 0) | 5450 | if (policy < 0) |
@@ -5445,7 +5458,7 @@ sys_sched_setscheduler(pid_t pid, int policy, struct sched_param __user *param) | |||
5445 | * @pid: the pid in question. | 5458 | * @pid: the pid in question. |
5446 | * @param: structure containing the new RT priority. | 5459 | * @param: structure containing the new RT priority. |
5447 | */ | 5460 | */ |
5448 | asmlinkage long sys_sched_setparam(pid_t pid, struct sched_param __user *param) | 5461 | SYSCALL_DEFINE2(sched_setparam, pid_t, pid, struct sched_param __user *, param) |
5449 | { | 5462 | { |
5450 | return do_sched_setscheduler(pid, -1, param); | 5463 | return do_sched_setscheduler(pid, -1, param); |
5451 | } | 5464 | } |
@@ -5454,7 +5467,7 @@ asmlinkage long sys_sched_setparam(pid_t pid, struct sched_param __user *param) | |||
5454 | * sys_sched_getscheduler - get the policy (scheduling class) of a thread | 5467 | * sys_sched_getscheduler - get the policy (scheduling class) of a thread |
5455 | * @pid: the pid in question. | 5468 | * @pid: the pid in question. |
5456 | */ | 5469 | */ |
5457 | asmlinkage long sys_sched_getscheduler(pid_t pid) | 5470 | SYSCALL_DEFINE1(sched_getscheduler, pid_t, pid) |
5458 | { | 5471 | { |
5459 | struct task_struct *p; | 5472 | struct task_struct *p; |
5460 | int retval; | 5473 | int retval; |
@@ -5479,7 +5492,7 @@ asmlinkage long sys_sched_getscheduler(pid_t pid) | |||
5479 | * @pid: the pid in question. | 5492 | * @pid: the pid in question. |
5480 | * @param: structure containing the RT priority. | 5493 | * @param: structure containing the RT priority. |
5481 | */ | 5494 | */ |
5482 | asmlinkage long sys_sched_getparam(pid_t pid, struct sched_param __user *param) | 5495 | SYSCALL_DEFINE2(sched_getparam, pid_t, pid, struct sched_param __user *, param) |
5483 | { | 5496 | { |
5484 | struct sched_param lp; | 5497 | struct sched_param lp; |
5485 | struct task_struct *p; | 5498 | struct task_struct *p; |
@@ -5597,8 +5610,8 @@ static int get_user_cpu_mask(unsigned long __user *user_mask_ptr, unsigned len, | |||
5597 | * @len: length in bytes of the bitmask pointed to by user_mask_ptr | 5610 | * @len: length in bytes of the bitmask pointed to by user_mask_ptr |
5598 | * @user_mask_ptr: user-space pointer to the new cpu mask | 5611 | * @user_mask_ptr: user-space pointer to the new cpu mask |
5599 | */ | 5612 | */ |
5600 | asmlinkage long sys_sched_setaffinity(pid_t pid, unsigned int len, | 5613 | SYSCALL_DEFINE3(sched_setaffinity, pid_t, pid, unsigned int, len, |
5601 | unsigned long __user *user_mask_ptr) | 5614 | unsigned long __user *, user_mask_ptr) |
5602 | { | 5615 | { |
5603 | cpumask_var_t new_mask; | 5616 | cpumask_var_t new_mask; |
5604 | int retval; | 5617 | int retval; |
@@ -5645,8 +5658,8 @@ out_unlock: | |||
5645 | * @len: length in bytes of the bitmask pointed to by user_mask_ptr | 5658 | * @len: length in bytes of the bitmask pointed to by user_mask_ptr |
5646 | * @user_mask_ptr: user-space pointer to hold the current cpu mask | 5659 | * @user_mask_ptr: user-space pointer to hold the current cpu mask |
5647 | */ | 5660 | */ |
5648 | asmlinkage long sys_sched_getaffinity(pid_t pid, unsigned int len, | 5661 | SYSCALL_DEFINE3(sched_getaffinity, pid_t, pid, unsigned int, len, |
5649 | unsigned long __user *user_mask_ptr) | 5662 | unsigned long __user *, user_mask_ptr) |
5650 | { | 5663 | { |
5651 | int ret; | 5664 | int ret; |
5652 | cpumask_var_t mask; | 5665 | cpumask_var_t mask; |
@@ -5675,7 +5688,7 @@ asmlinkage long sys_sched_getaffinity(pid_t pid, unsigned int len, | |||
5675 | * This function yields the current CPU to other tasks. If there are no | 5688 | * This function yields the current CPU to other tasks. If there are no |
5676 | * other threads running on this CPU then this function will return. | 5689 | * other threads running on this CPU then this function will return. |
5677 | */ | 5690 | */ |
5678 | asmlinkage long sys_sched_yield(void) | 5691 | SYSCALL_DEFINE0(sched_yield) |
5679 | { | 5692 | { |
5680 | struct rq *rq = this_rq_lock(); | 5693 | struct rq *rq = this_rq_lock(); |
5681 | 5694 | ||
@@ -5816,7 +5829,7 @@ long __sched io_schedule_timeout(long timeout) | |||
5816 | * this syscall returns the maximum rt_priority that can be used | 5829 | * this syscall returns the maximum rt_priority that can be used |
5817 | * by a given scheduling class. | 5830 | * by a given scheduling class. |
5818 | */ | 5831 | */ |
5819 | asmlinkage long sys_sched_get_priority_max(int policy) | 5832 | SYSCALL_DEFINE1(sched_get_priority_max, int, policy) |
5820 | { | 5833 | { |
5821 | int ret = -EINVAL; | 5834 | int ret = -EINVAL; |
5822 | 5835 | ||
@@ -5841,7 +5854,7 @@ asmlinkage long sys_sched_get_priority_max(int policy) | |||
5841 | * this syscall returns the minimum rt_priority that can be used | 5854 | * this syscall returns the minimum rt_priority that can be used |
5842 | * by a given scheduling class. | 5855 | * by a given scheduling class. |
5843 | */ | 5856 | */ |
5844 | asmlinkage long sys_sched_get_priority_min(int policy) | 5857 | SYSCALL_DEFINE1(sched_get_priority_min, int, policy) |
5845 | { | 5858 | { |
5846 | int ret = -EINVAL; | 5859 | int ret = -EINVAL; |
5847 | 5860 | ||
@@ -5866,8 +5879,8 @@ asmlinkage long sys_sched_get_priority_min(int policy) | |||
5866 | * this syscall writes the default timeslice value of a given process | 5879 | * this syscall writes the default timeslice value of a given process |
5867 | * into the user-space timespec buffer. A value of '0' means infinity. | 5880 | * into the user-space timespec buffer. A value of '0' means infinity. |
5868 | */ | 5881 | */ |
5869 | asmlinkage | 5882 | SYSCALL_DEFINE2(sched_rr_get_interval, pid_t, pid, |
5870 | long sys_sched_rr_get_interval(pid_t pid, struct timespec __user *interval) | 5883 | struct timespec __user *, interval) |
5871 | { | 5884 | { |
5872 | struct task_struct *p; | 5885 | struct task_struct *p; |
5873 | unsigned int time_slice; | 5886 | unsigned int time_slice; |
@@ -7282,10 +7295,10 @@ cpu_to_phys_group(int cpu, const struct cpumask *cpu_map, | |||
7282 | * groups, so roll our own. Now each node has its own list of groups which | 7295 | * groups, so roll our own. Now each node has its own list of groups which |
7283 | * gets dynamically allocated. | 7296 | * gets dynamically allocated. |
7284 | */ | 7297 | */ |
7285 | static DEFINE_PER_CPU(struct sched_domain, node_domains); | 7298 | static DEFINE_PER_CPU(struct static_sched_domain, node_domains); |
7286 | static struct sched_group ***sched_group_nodes_bycpu; | 7299 | static struct sched_group ***sched_group_nodes_bycpu; |
7287 | 7300 | ||
7288 | static DEFINE_PER_CPU(struct sched_domain, allnodes_domains); | 7301 | static DEFINE_PER_CPU(struct static_sched_domain, allnodes_domains); |
7289 | static DEFINE_PER_CPU(struct static_sched_group, sched_group_allnodes); | 7302 | static DEFINE_PER_CPU(struct static_sched_group, sched_group_allnodes); |
7290 | 7303 | ||
7291 | static int cpu_to_allnodes_group(int cpu, const struct cpumask *cpu_map, | 7304 | static int cpu_to_allnodes_group(int cpu, const struct cpumask *cpu_map, |
@@ -7560,7 +7573,7 @@ static int __build_sched_domains(const struct cpumask *cpu_map, | |||
7560 | #ifdef CONFIG_NUMA | 7573 | #ifdef CONFIG_NUMA |
7561 | if (cpumask_weight(cpu_map) > | 7574 | if (cpumask_weight(cpu_map) > |
7562 | SD_NODES_PER_DOMAIN*cpumask_weight(nodemask)) { | 7575 | SD_NODES_PER_DOMAIN*cpumask_weight(nodemask)) { |
7563 | sd = &per_cpu(allnodes_domains, i); | 7576 | sd = &per_cpu(allnodes_domains, i).sd; |
7564 | SD_INIT(sd, ALLNODES); | 7577 | SD_INIT(sd, ALLNODES); |
7565 | set_domain_attribute(sd, attr); | 7578 | set_domain_attribute(sd, attr); |
7566 | cpumask_copy(sched_domain_span(sd), cpu_map); | 7579 | cpumask_copy(sched_domain_span(sd), cpu_map); |
@@ -7570,7 +7583,7 @@ static int __build_sched_domains(const struct cpumask *cpu_map, | |||
7570 | } else | 7583 | } else |
7571 | p = NULL; | 7584 | p = NULL; |
7572 | 7585 | ||
7573 | sd = &per_cpu(node_domains, i); | 7586 | sd = &per_cpu(node_domains, i).sd; |
7574 | SD_INIT(sd, NODE); | 7587 | SD_INIT(sd, NODE); |
7575 | set_domain_attribute(sd, attr); | 7588 | set_domain_attribute(sd, attr); |
7576 | sched_domain_node_span(cpu_to_node(i), sched_domain_span(sd)); | 7589 | sched_domain_node_span(cpu_to_node(i), sched_domain_span(sd)); |
@@ -7688,7 +7701,7 @@ static int __build_sched_domains(const struct cpumask *cpu_map, | |||
7688 | for_each_cpu(j, nodemask) { | 7701 | for_each_cpu(j, nodemask) { |
7689 | struct sched_domain *sd; | 7702 | struct sched_domain *sd; |
7690 | 7703 | ||
7691 | sd = &per_cpu(node_domains, j); | 7704 | sd = &per_cpu(node_domains, j).sd; |
7692 | sd->groups = sg; | 7705 | sd->groups = sg; |
7693 | } | 7706 | } |
7694 | sg->__cpu_power = 0; | 7707 | sg->__cpu_power = 0; |
@@ -9047,6 +9060,13 @@ static int tg_schedulable(struct task_group *tg, void *data) | |||
9047 | runtime = d->rt_runtime; | 9060 | runtime = d->rt_runtime; |
9048 | } | 9061 | } |
9049 | 9062 | ||
9063 | #ifdef CONFIG_USER_SCHED | ||
9064 | if (tg == &root_task_group) { | ||
9065 | period = global_rt_period(); | ||
9066 | runtime = global_rt_runtime(); | ||
9067 | } | ||
9068 | #endif | ||
9069 | |||
9050 | /* | 9070 | /* |
9051 | * Cannot have more runtime than the period. | 9071 | * Cannot have more runtime than the period. |
9052 | */ | 9072 | */ |