diff options
Diffstat (limited to 'kernel/sched/sched.h')
| -rw-r--r-- | kernel/sched/sched.h | 25 |
1 files changed, 10 insertions, 15 deletions
diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h index 6d52cea7f33d..c35a1a7dd4d6 100644 --- a/kernel/sched/sched.h +++ b/kernel/sched/sched.h | |||
| @@ -538,22 +538,19 @@ extern int group_balance_cpu(struct sched_group *sg); | |||
| 538 | /* | 538 | /* |
| 539 | * Return the group to which this tasks belongs. | 539 | * Return the group to which this tasks belongs. |
| 540 | * | 540 | * |
| 541 | * We use task_subsys_state_check() and extend the RCU verification with | 541 | * We cannot use task_subsys_state() and friends because the cgroup |
| 542 | * pi->lock and rq->lock because cpu_cgroup_attach() holds those locks for each | 542 | * subsystem changes that value before the cgroup_subsys::attach() method |
| 543 | * task it moves into the cgroup. Therefore by holding either of those locks, | 543 | * is called, therefore we cannot pin it and might observe the wrong value. |
| 544 | * we pin the task to the current cgroup. | 544 | * |
| 545 | * The same is true for autogroup's p->signal->autogroup->tg, the autogroup | ||
| 546 | * core changes this before calling sched_move_task(). | ||
| 547 | * | ||
| 548 | * Instead we use a 'copy' which is updated from sched_move_task() while | ||
| 549 | * holding both task_struct::pi_lock and rq::lock. | ||
| 545 | */ | 550 | */ |
| 546 | static inline struct task_group *task_group(struct task_struct *p) | 551 | static inline struct task_group *task_group(struct task_struct *p) |
| 547 | { | 552 | { |
| 548 | struct task_group *tg; | 553 | return p->sched_task_group; |
| 549 | struct cgroup_subsys_state *css; | ||
| 550 | |||
| 551 | css = task_subsys_state_check(p, cpu_cgroup_subsys_id, | ||
| 552 | lockdep_is_held(&p->pi_lock) || | ||
| 553 | lockdep_is_held(&task_rq(p)->lock)); | ||
| 554 | tg = container_of(css, struct task_group, css); | ||
| 555 | |||
| 556 | return autogroup_task_group(p, tg); | ||
| 557 | } | 554 | } |
| 558 | 555 | ||
| 559 | /* Change a task's cfs_rq and parent entity if it moves across CPUs/groups */ | 556 | /* Change a task's cfs_rq and parent entity if it moves across CPUs/groups */ |
| @@ -942,8 +939,6 @@ static inline u64 sched_avg_period(void) | |||
| 942 | return (u64)sysctl_sched_time_avg * NSEC_PER_MSEC / 2; | 939 | return (u64)sysctl_sched_time_avg * NSEC_PER_MSEC / 2; |
| 943 | } | 940 | } |
| 944 | 941 | ||
| 945 | void calc_load_account_idle(struct rq *this_rq); | ||
| 946 | |||
| 947 | #ifdef CONFIG_SCHED_HRTICK | 942 | #ifdef CONFIG_SCHED_HRTICK |
| 948 | 943 | ||
| 949 | /* | 944 | /* |
