diff options
author | Paul E. McKenney <paulmck@linux.vnet.ibm.com> | 2010-04-14 20:39:26 -0400 |
---|---|---|
committer | Paul E. McKenney <paulmck@linux.vnet.ibm.com> | 2010-05-10 14:08:35 -0400 |
commit | d21670acab9fcb4bc74a40b68a6941059234c55c (patch) | |
tree | 6a4c054bc4dbadf0524b4e221889a8da558dbdaf /kernel | |
parent | 4a90a0681cf6cd21cd444184302aa045156486b3 (diff) |
rcu: reduce the number of spurious RCU_SOFTIRQ invocations
Lai Jiangshan noted that up to 10% of the RCU_SOFTIRQ are spurious, and
traced this down to the fact that the current grace-period machinery
will uselessly raise RCU_SOFTIRQ when a given CPU needs to go through
a quiescent state, but has not yet done so. In this situation, there
might well be nothing that RCU_SOFTIRQ can do, and the overhead can be
worth worrying about in the ksoftirqd case. This patch therefore avoids
raising RCU_SOFTIRQ in this situation.
Changes since v1 (http://lkml.org/lkml/2010/3/30/122 from Lai Jiangshan):
o Omit the rcu_qs_pending() prechecks, as they aren't that
much less expensive than the quiescent-state checks.
o Merge with the set_need_resched() patch that reduces IPIs.
o Add the new n_rp_report_qs field to the rcu_pending tracing output.
o Update the tracing documentation accordingly.
Signed-off-by: Lai Jiangshan <laijs@cn.fujitsu.com>
Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/rcutree.c | 11 | ||||
-rw-r--r-- | kernel/rcutree.h | 1 | ||||
-rw-r--r-- | kernel/rcutree_trace.c | 4 |
3 files changed, 10 insertions, 6 deletions
diff --git a/kernel/rcutree.c b/kernel/rcutree.c index c60fd74e7ec9..ba6996943e28 100644 --- a/kernel/rcutree.c +++ b/kernel/rcutree.c | |||
@@ -1161,8 +1161,6 @@ static void rcu_do_batch(struct rcu_state *rsp, struct rcu_data *rdp) | |||
1161 | */ | 1161 | */ |
1162 | void rcu_check_callbacks(int cpu, int user) | 1162 | void rcu_check_callbacks(int cpu, int user) |
1163 | { | 1163 | { |
1164 | if (!rcu_pending(cpu)) | ||
1165 | return; /* if nothing for RCU to do. */ | ||
1166 | if (user || | 1164 | if (user || |
1167 | (idle_cpu(cpu) && rcu_scheduler_active && | 1165 | (idle_cpu(cpu) && rcu_scheduler_active && |
1168 | !in_softirq() && hardirq_count() <= (1 << HARDIRQ_SHIFT))) { | 1166 | !in_softirq() && hardirq_count() <= (1 << HARDIRQ_SHIFT))) { |
@@ -1194,7 +1192,8 @@ void rcu_check_callbacks(int cpu, int user) | |||
1194 | rcu_bh_qs(cpu); | 1192 | rcu_bh_qs(cpu); |
1195 | } | 1193 | } |
1196 | rcu_preempt_check_callbacks(cpu); | 1194 | rcu_preempt_check_callbacks(cpu); |
1197 | raise_softirq(RCU_SOFTIRQ); | 1195 | if (rcu_pending(cpu)) |
1196 | raise_softirq(RCU_SOFTIRQ); | ||
1198 | } | 1197 | } |
1199 | 1198 | ||
1200 | #ifdef CONFIG_SMP | 1199 | #ifdef CONFIG_SMP |
@@ -1534,18 +1533,20 @@ static int __rcu_pending(struct rcu_state *rsp, struct rcu_data *rdp) | |||
1534 | check_cpu_stall(rsp, rdp); | 1533 | check_cpu_stall(rsp, rdp); |
1535 | 1534 | ||
1536 | /* Is the RCU core waiting for a quiescent state from this CPU? */ | 1535 | /* Is the RCU core waiting for a quiescent state from this CPU? */ |
1537 | if (rdp->qs_pending) { | 1536 | if (rdp->qs_pending && !rdp->passed_quiesc) { |
1538 | 1537 | ||
1539 | /* | 1538 | /* |
1540 | * If force_quiescent_state() coming soon and this CPU | 1539 | * If force_quiescent_state() coming soon and this CPU |
1541 | * needs a quiescent state, and this is either RCU-sched | 1540 | * needs a quiescent state, and this is either RCU-sched |
1542 | * or RCU-bh, force a local reschedule. | 1541 | * or RCU-bh, force a local reschedule. |
1543 | */ | 1542 | */ |
1543 | rdp->n_rp_qs_pending++; | ||
1544 | if (!rdp->preemptable && | 1544 | if (!rdp->preemptable && |
1545 | ULONG_CMP_LT(ACCESS_ONCE(rsp->jiffies_force_qs) - 1, | 1545 | ULONG_CMP_LT(ACCESS_ONCE(rsp->jiffies_force_qs) - 1, |
1546 | jiffies)) | 1546 | jiffies)) |
1547 | set_need_resched(); | 1547 | set_need_resched(); |
1548 | rdp->n_rp_qs_pending++; | 1548 | } else if (rdp->qs_pending && rdp->passed_quiesc) { |
1549 | rdp->n_rp_report_qs++; | ||
1549 | return 1; | 1550 | return 1; |
1550 | } | 1551 | } |
1551 | 1552 | ||
diff --git a/kernel/rcutree.h b/kernel/rcutree.h index 11f171121ad9..14c040b18ed0 100644 --- a/kernel/rcutree.h +++ b/kernel/rcutree.h | |||
@@ -223,6 +223,7 @@ struct rcu_data { | |||
223 | /* 5) __rcu_pending() statistics. */ | 223 | /* 5) __rcu_pending() statistics. */ |
224 | unsigned long n_rcu_pending; /* rcu_pending() calls since boot. */ | 224 | unsigned long n_rcu_pending; /* rcu_pending() calls since boot. */ |
225 | unsigned long n_rp_qs_pending; | 225 | unsigned long n_rp_qs_pending; |
226 | unsigned long n_rp_report_qs; | ||
226 | unsigned long n_rp_cb_ready; | 227 | unsigned long n_rp_cb_ready; |
227 | unsigned long n_rp_cpu_needs_gp; | 228 | unsigned long n_rp_cpu_needs_gp; |
228 | unsigned long n_rp_gp_completed; | 229 | unsigned long n_rp_gp_completed; |
diff --git a/kernel/rcutree_trace.c b/kernel/rcutree_trace.c index d45db2e35d27..36c95b45738e 100644 --- a/kernel/rcutree_trace.c +++ b/kernel/rcutree_trace.c | |||
@@ -241,11 +241,13 @@ static const struct file_operations rcugp_fops = { | |||
241 | static void print_one_rcu_pending(struct seq_file *m, struct rcu_data *rdp) | 241 | static void print_one_rcu_pending(struct seq_file *m, struct rcu_data *rdp) |
242 | { | 242 | { |
243 | seq_printf(m, "%3d%cnp=%ld " | 243 | seq_printf(m, "%3d%cnp=%ld " |
244 | "qsp=%ld cbr=%ld cng=%ld gpc=%ld gps=%ld nf=%ld nn=%ld\n", | 244 | "qsp=%ld rpq=%ld cbr=%ld cng=%ld " |
245 | "gpc=%ld gps=%ld nf=%ld nn=%ld\n", | ||
245 | rdp->cpu, | 246 | rdp->cpu, |
246 | cpu_is_offline(rdp->cpu) ? '!' : ' ', | 247 | cpu_is_offline(rdp->cpu) ? '!' : ' ', |
247 | rdp->n_rcu_pending, | 248 | rdp->n_rcu_pending, |
248 | rdp->n_rp_qs_pending, | 249 | rdp->n_rp_qs_pending, |
250 | rdp->n_rp_report_qs, | ||
249 | rdp->n_rp_cb_ready, | 251 | rdp->n_rp_cb_ready, |
250 | rdp->n_rp_cpu_needs_gp, | 252 | rdp->n_rp_cpu_needs_gp, |
251 | rdp->n_rp_gp_completed, | 253 | rdp->n_rp_gp_completed, |