aboutsummaryrefslogtreecommitdiffstats
path: root/kernel
diff options
context:
space:
mode:
authorNick Piggin <nickpiggin@yahoo.com.au>2005-06-25 17:57:27 -0400
committerLinus Torvalds <torvalds@ppc970.osdl.org>2005-06-25 19:24:44 -0400
commit674311d5b411e9042df4fdf7aef0b3c8217b6240 (patch)
tree4ee6b739629e7fc33b519fd087a116b4fa33a217 /kernel
parent3dbd5342074a1e570ec84edf859deb9be588006d (diff)
[PATCH] sched: RCU domains
One of the problems with the multilevel balance-on-fork/exec is that it needs to jump through hoops to satisfy sched-domain's locking semantics (that is, you may traverse your own domain when not preemptable, and you may traverse others' domains when holding their runqueue lock). balance-on-exec had to potentially migrate between more than one CPU before finding a final CPU to migrate to, and balance-on-fork needed to potentially take multiple runqueue locks. So bite the bullet and make sched-domains go completely RCU. This actually simplifies the code quite a bit. From: Ingo Molnar <mingo@elte.hu> schedstats RCU fix, and a nice comment on for_each_domain, from Ingo. Signed-off-by: Ingo Molnar <mingo@elte.hu> Signed-off-by: Nick Piggin <nickpiggin@yahoo.com.au> Acked-by: Ingo Molnar <mingo@elte.hu> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
Diffstat (limited to 'kernel')
-rw-r--r--kernel/sched.c60
1 files changed, 15 insertions, 45 deletions
diff --git a/kernel/sched.c b/kernel/sched.c
index ef32389ee768..54ce787b6207 100644
--- a/kernel/sched.c
+++ b/kernel/sched.c
@@ -260,8 +260,15 @@ struct runqueue {
260 260
261static DEFINE_PER_CPU(struct runqueue, runqueues); 261static DEFINE_PER_CPU(struct runqueue, runqueues);
262 262
263/*
264 * The domain tree (rq->sd) is protected by RCU's quiescent state transition.
265 * See update_sched_domains: synchronize_kernel for details.
266 *
267 * The domain tree of any CPU may only be accessed from within
268 * preempt-disabled sections.
269 */
263#define for_each_domain(cpu, domain) \ 270#define for_each_domain(cpu, domain) \
264 for (domain = cpu_rq(cpu)->sd; domain; domain = domain->parent) 271for (domain = rcu_dereference(cpu_rq(cpu)->sd); domain; domain = domain->parent)
265 272
266#define cpu_rq(cpu) (&per_cpu(runqueues, (cpu))) 273#define cpu_rq(cpu) (&per_cpu(runqueues, (cpu)))
267#define this_rq() (&__get_cpu_var(runqueues)) 274#define this_rq() (&__get_cpu_var(runqueues))
@@ -395,6 +402,7 @@ static int show_schedstat(struct seq_file *seq, void *v)
395 402
396#ifdef CONFIG_SMP 403#ifdef CONFIG_SMP
397 /* domain-specific stats */ 404 /* domain-specific stats */
405 preempt_disable();
398 for_each_domain(cpu, sd) { 406 for_each_domain(cpu, sd) {
399 enum idle_type itype; 407 enum idle_type itype;
400 char mask_str[NR_CPUS]; 408 char mask_str[NR_CPUS];
@@ -419,6 +427,7 @@ static int show_schedstat(struct seq_file *seq, void *v)
419 sd->sbf_cnt, sd->sbf_balanced, sd->sbf_pushed, 427 sd->sbf_cnt, sd->sbf_balanced, sd->sbf_pushed,
420 sd->ttwu_wake_remote, sd->ttwu_move_affine, sd->ttwu_move_balance); 428 sd->ttwu_wake_remote, sd->ttwu_move_affine, sd->ttwu_move_balance);
421 } 429 }
430 preempt_enable();
422#endif 431#endif
423 } 432 }
424 return 0; 433 return 0;
@@ -824,22 +833,12 @@ inline int task_curr(const task_t *p)
824} 833}
825 834
826#ifdef CONFIG_SMP 835#ifdef CONFIG_SMP
827enum request_type {
828 REQ_MOVE_TASK,
829 REQ_SET_DOMAIN,
830};
831
832typedef struct { 836typedef struct {
833 struct list_head list; 837 struct list_head list;
834 enum request_type type;
835 838
836 /* For REQ_MOVE_TASK */
837 task_t *task; 839 task_t *task;
838 int dest_cpu; 840 int dest_cpu;
839 841
840 /* For REQ_SET_DOMAIN */
841 struct sched_domain *sd;
842
843 struct completion done; 842 struct completion done;
844} migration_req_t; 843} migration_req_t;
845 844
@@ -861,7 +860,6 @@ static int migrate_task(task_t *p, int dest_cpu, migration_req_t *req)
861 } 860 }
862 861
863 init_completion(&req->done); 862 init_completion(&req->done);
864 req->type = REQ_MOVE_TASK;
865 req->task = p; 863 req->task = p;
866 req->dest_cpu = dest_cpu; 864 req->dest_cpu = dest_cpu;
867 list_add(&req->list, &rq->migration_queue); 865 list_add(&req->list, &rq->migration_queue);
@@ -4378,17 +4376,9 @@ static int migration_thread(void * data)
4378 req = list_entry(head->next, migration_req_t, list); 4376 req = list_entry(head->next, migration_req_t, list);
4379 list_del_init(head->next); 4377 list_del_init(head->next);
4380 4378
4381 if (req->type == REQ_MOVE_TASK) { 4379 spin_unlock(&rq->lock);
4382 spin_unlock(&rq->lock); 4380 __migrate_task(req->task, cpu, req->dest_cpu);
4383 __migrate_task(req->task, cpu, req->dest_cpu); 4381 local_irq_enable();
4384 local_irq_enable();
4385 } else if (req->type == REQ_SET_DOMAIN) {
4386 rq->sd = req->sd;
4387 spin_unlock_irq(&rq->lock);
4388 } else {
4389 spin_unlock_irq(&rq->lock);
4390 WARN_ON(1);
4391 }
4392 4382
4393 complete(&req->done); 4383 complete(&req->done);
4394 } 4384 }
@@ -4619,7 +4609,6 @@ static int migration_call(struct notifier_block *nfb, unsigned long action,
4619 migration_req_t *req; 4609 migration_req_t *req;
4620 req = list_entry(rq->migration_queue.next, 4610 req = list_entry(rq->migration_queue.next,
4621 migration_req_t, list); 4611 migration_req_t, list);
4622 BUG_ON(req->type != REQ_MOVE_TASK);
4623 list_del_init(&req->list); 4612 list_del_init(&req->list);
4624 complete(&req->done); 4613 complete(&req->done);
4625 } 4614 }
@@ -4800,10 +4789,7 @@ static int __devinit sd_parent_degenerate(struct sched_domain *sd,
4800 */ 4789 */
4801void __devinit cpu_attach_domain(struct sched_domain *sd, int cpu) 4790void __devinit cpu_attach_domain(struct sched_domain *sd, int cpu)
4802{ 4791{
4803 migration_req_t req;
4804 unsigned long flags;
4805 runqueue_t *rq = cpu_rq(cpu); 4792 runqueue_t *rq = cpu_rq(cpu);
4806 int local = 1;
4807 struct sched_domain *tmp; 4793 struct sched_domain *tmp;
4808 4794
4809 /* Remove the sched domains which do not contribute to scheduling. */ 4795 /* Remove the sched domains which do not contribute to scheduling. */
@@ -4820,24 +4806,7 @@ void __devinit cpu_attach_domain(struct sched_domain *sd, int cpu)
4820 4806
4821 sched_domain_debug(sd, cpu); 4807 sched_domain_debug(sd, cpu);
4822 4808
4823 spin_lock_irqsave(&rq->lock, flags); 4809 rcu_assign_pointer(rq->sd, sd);
4824
4825 if (cpu == smp_processor_id() || !cpu_online(cpu)) {
4826 rq->sd = sd;
4827 } else {
4828 init_completion(&req.done);
4829 req.type = REQ_SET_DOMAIN;
4830 req.sd = sd;
4831 list_add(&req.list, &rq->migration_queue);
4832 local = 0;
4833 }
4834
4835 spin_unlock_irqrestore(&rq->lock, flags);
4836
4837 if (!local) {
4838 wake_up_process(rq->migration_thread);
4839 wait_for_completion(&req.done);
4840 }
4841} 4810}
4842 4811
4843/* cpus with isolated domains */ 4812/* cpus with isolated domains */
@@ -5112,6 +5081,7 @@ static int update_sched_domains(struct notifier_block *nfb,
5112 case CPU_DOWN_PREPARE: 5081 case CPU_DOWN_PREPARE:
5113 for_each_online_cpu(i) 5082 for_each_online_cpu(i)
5114 cpu_attach_domain(NULL, i); 5083 cpu_attach_domain(NULL, i);
5084 synchronize_kernel();
5115 arch_destroy_sched_domains(); 5085 arch_destroy_sched_domains();
5116 return NOTIFY_OK; 5086 return NOTIFY_OK;
5117 5087