diff options
author | Paul E. McKenney <paulmck@linux.vnet.ibm.com> | 2010-04-01 20:37:01 -0400 |
---|---|---|
committer | Paul E. McKenney <paulmck@linux.vnet.ibm.com> | 2010-05-10 14:08:33 -0400 |
commit | 25502a6c13745f4650cc59322bd198194f55e796 (patch) | |
tree | d76cc659d3ea797c5da4630e219ac363d17c44a6 /kernel/rcutree.c | |
parent | 99652b54de1ee094236f7171485214071af4ef31 (diff) |
rcu: refactor RCU's context-switch handling
The addition of preemptible RCU to treercu resulted in a bit of
confusion and inefficiency surrounding the handling of context switches
for RCU-sched and for RCU-preempt. For RCU-sched, a context switch
is a quiescent state, pure and simple, just like it always has been.
For RCU-preempt, a context switch is in no way a quiescent state, but
special handling is required when a task blocks in an RCU read-side
critical section.
However, the callout from the scheduler and the outer loop in ksoftirqd
still calls something named rcu_sched_qs(), whose name is no longer
accurate. Furthermore, when rcu_check_callbacks() notes an RCU-sched
quiescent state, it ends up unnecessarily (though harmlessly, aside
from the performance hit) enqueuing the current task if it happens to
be running in an RCU-preempt read-side critical section. This not only
increases the maximum latency of scheduler_tick(), it also needlessly
increases the overhead of the next outermost rcu_read_unlock() invocation.
This patch addresses this situation by separating the notion of RCU's
context-switch handling from that of RCU-sched's quiescent states.
The context-switch handling is covered by rcu_note_context_switch() in
general and by rcu_preempt_note_context_switch() for preemptible RCU.
This permits rcu_sched_qs() to handle quiescent states and only quiescent
states. It also reduces the maximum latency of scheduler_tick(), though
probably by much less than a microsecond. Finally, it means that tasks
within preemptible-RCU read-side critical sections avoid incurring the
overhead of queuing unless there really is a context switch.
Suggested-by: Lai Jiangshan <laijs@cn.fujitsu.com>
Acked-by: Lai Jiangshan <laijs@cn.fujitsu.com>
Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Cc: Ingo Molnar <mingo@elte.hu>
Cc: Peter Zijlstra <peterz@infradead.org>
Diffstat (limited to 'kernel/rcutree.c')
-rw-r--r-- | kernel/rcutree.c | 17 |
1 files changed, 12 insertions, 5 deletions
diff --git a/kernel/rcutree.c b/kernel/rcutree.c index 86bb9499aae6..e33631354b69 100644 --- a/kernel/rcutree.c +++ b/kernel/rcutree.c | |||
@@ -97,25 +97,32 @@ static int rcu_gp_in_progress(struct rcu_state *rsp) | |||
97 | */ | 97 | */ |
98 | void rcu_sched_qs(int cpu) | 98 | void rcu_sched_qs(int cpu) |
99 | { | 99 | { |
100 | struct rcu_data *rdp; | 100 | struct rcu_data *rdp = &per_cpu(rcu_sched_data, cpu); |
101 | 101 | ||
102 | rdp = &per_cpu(rcu_sched_data, cpu); | ||
103 | rdp->passed_quiesc_completed = rdp->gpnum - 1; | 102 | rdp->passed_quiesc_completed = rdp->gpnum - 1; |
104 | barrier(); | 103 | barrier(); |
105 | rdp->passed_quiesc = 1; | 104 | rdp->passed_quiesc = 1; |
106 | rcu_preempt_note_context_switch(cpu); | ||
107 | } | 105 | } |
108 | 106 | ||
109 | void rcu_bh_qs(int cpu) | 107 | void rcu_bh_qs(int cpu) |
110 | { | 108 | { |
111 | struct rcu_data *rdp; | 109 | struct rcu_data *rdp = &per_cpu(rcu_bh_data, cpu); |
112 | 110 | ||
113 | rdp = &per_cpu(rcu_bh_data, cpu); | ||
114 | rdp->passed_quiesc_completed = rdp->gpnum - 1; | 111 | rdp->passed_quiesc_completed = rdp->gpnum - 1; |
115 | barrier(); | 112 | barrier(); |
116 | rdp->passed_quiesc = 1; | 113 | rdp->passed_quiesc = 1; |
117 | } | 114 | } |
118 | 115 | ||
116 | /* | ||
117 | * Note a context switch. This is a quiescent state for RCU-sched, | ||
118 | * and requires special handling for preemptible RCU. | ||
119 | */ | ||
120 | void rcu_note_context_switch(int cpu) | ||
121 | { | ||
122 | rcu_sched_qs(cpu); | ||
123 | rcu_preempt_note_context_switch(cpu); | ||
124 | } | ||
125 | |||
119 | #ifdef CONFIG_NO_HZ | 126 | #ifdef CONFIG_NO_HZ |
120 | DEFINE_PER_CPU(struct rcu_dynticks, rcu_dynticks) = { | 127 | DEFINE_PER_CPU(struct rcu_dynticks, rcu_dynticks) = { |
121 | .dynticks_nesting = 1, | 128 | .dynticks_nesting = 1, |