aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/sched/core.c
diff options
context:
space:
mode:
authorYao Dongdong <yaodongdong@huawei.com>2014-12-29 01:41:43 -0500
committerIngo Molnar <mingo@kernel.org>2015-01-14 07:34:17 -0500
commit1b537c7d1e58c761212a193085f9049b58f672e6 (patch)
tree948a7395f03e9c16cfbba7f2f99ecdbd93cac0b6 /kernel/sched/core.c
parentbb04159df99fa353d0fb524574aca03ce2c6515b (diff)
sched/core: Remove check of p->sched_class
Search all usage of p->sched_class in sched/core.c, no one check it before use, so it seems that every task must belong to one sched_class. Signed-off-by: Yao Dongdong <yaodongdong@huawei.com> [ Moved the early class assignment to make it boot. ] Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org> Cc: Linus Torvalds <torvalds@linux-foundation.org> Link: http://lkml.kernel.org/r/1419835303-28958-1-git-send-email-yaodongdong@huawei.com Signed-off-by: Ingo Molnar <mingo@kernel.org>
Diffstat (limited to 'kernel/sched/core.c')
-rw-r--r--kernel/sched/core.c12
1 files changed, 6 insertions, 6 deletions
diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index 95ac795ab3d3..46a2345f9f45 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -4744,7 +4744,7 @@ static struct rq *move_queued_task(struct task_struct *p, int new_cpu)
4744 4744
4745void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask) 4745void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask)
4746{ 4746{
4747 if (p->sched_class && p->sched_class->set_cpus_allowed) 4747 if (p->sched_class->set_cpus_allowed)
4748 p->sched_class->set_cpus_allowed(p, new_mask); 4748 p->sched_class->set_cpus_allowed(p, new_mask);
4749 4749
4750 cpumask_copy(&p->cpus_allowed, new_mask); 4750 cpumask_copy(&p->cpus_allowed, new_mask);
@@ -7254,6 +7254,11 @@ void __init sched_init(void)
7254 enter_lazy_tlb(&init_mm, current); 7254 enter_lazy_tlb(&init_mm, current);
7255 7255
7256 /* 7256 /*
7257 * During early bootup we pretend to be a normal task:
7258 */
7259 current->sched_class = &fair_sched_class;
7260
7261 /*
7257 * Make us the idle thread. Technically, schedule() should not be 7262 * Make us the idle thread. Technically, schedule() should not be
7258 * called from this thread, however somewhere below it might be, 7263 * called from this thread, however somewhere below it might be,
7259 * but because we are the idle thread, we just pick up running again 7264 * but because we are the idle thread, we just pick up running again
@@ -7263,11 +7268,6 @@ void __init sched_init(void)
7263 7268
7264 calc_load_update = jiffies + LOAD_FREQ; 7269 calc_load_update = jiffies + LOAD_FREQ;
7265 7270
7266 /*
7267 * During early bootup we pretend to be a normal task:
7268 */
7269 current->sched_class = &fair_sched_class;
7270
7271#ifdef CONFIG_SMP 7271#ifdef CONFIG_SMP
7272 zalloc_cpumask_var(&sched_domains_tmpmask, GFP_NOWAIT); 7272 zalloc_cpumask_var(&sched_domains_tmpmask, GFP_NOWAIT);
7273 /* May be allocated at isolcpus cmdline parse time */ 7273 /* May be allocated at isolcpus cmdline parse time */