diff options
| author | Paul E. McKenney <paulmck@linux.vnet.ibm.com> | 2011-05-21 08:57:18 -0400 |
|---|---|---|
| committer | Paul E. McKenney <paulmck@linux.vnet.ibm.com> | 2011-07-20 00:38:53 -0400 |
| commit | be0e1e21ef707be4d16ea6a96ac9997463e4b8d2 (patch) | |
| tree | fe53c0731f5c3a6e56e2358eee9b8f17c3af83f4 | |
| parent | 7765be2fec0f476fcd61812d5f9406b04c765020 (diff) | |
rcu: Streamline code produced by __rcu_read_unlock()
Given some common flag combinations, particularly -Os, gcc will inline
rcu_read_unlock_special() despite its being in an unlikely() clause.
Use noinline to prohibit this misoptimization.
In addition, move the second barrier() in __rcu_read_unlock() so that
it is not on the common-case code path. This will allow the compiler to
generate better code for the common-case path through __rcu_read_unlock().
Suggested-by: Linus Torvalds <torvalds@linux-foundation.org>
Signed-off-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
Acked-by: Mathieu Desnoyers <mathieu.desnoyers@efficios.com>
| -rw-r--r-- | kernel/rcutree_plugin.h | 12 |
1 files changed, 6 insertions, 6 deletions
diff --git a/kernel/rcutree_plugin.h b/kernel/rcutree_plugin.h index 3a0ae035522..4d2c068ba13 100644 --- a/kernel/rcutree_plugin.h +++ b/kernel/rcutree_plugin.h | |||
| @@ -284,7 +284,7 @@ static struct list_head *rcu_next_node_entry(struct task_struct *t, | |||
| 284 | * notify RCU core processing or task having blocked during the RCU | 284 | * notify RCU core processing or task having blocked during the RCU |
| 285 | * read-side critical section. | 285 | * read-side critical section. |
| 286 | */ | 286 | */ |
| 287 | static void rcu_read_unlock_special(struct task_struct *t) | 287 | static noinline void rcu_read_unlock_special(struct task_struct *t) |
| 288 | { | 288 | { |
| 289 | int empty; | 289 | int empty; |
| 290 | int empty_exp; | 290 | int empty_exp; |
| @@ -391,11 +391,11 @@ void __rcu_read_unlock(void) | |||
| 391 | struct task_struct *t = current; | 391 | struct task_struct *t = current; |
| 392 | 392 | ||
| 393 | barrier(); /* needed if we ever invoke rcu_read_unlock in rcutree.c */ | 393 | barrier(); /* needed if we ever invoke rcu_read_unlock in rcutree.c */ |
| 394 | --t->rcu_read_lock_nesting; | 394 | if (--t->rcu_read_lock_nesting == 0) { |
| 395 | barrier(); /* decrement before load of ->rcu_read_unlock_special */ | 395 | barrier(); /* decr before ->rcu_read_unlock_special load */ |
| 396 | if (t->rcu_read_lock_nesting == 0 && | 396 | if (unlikely(ACCESS_ONCE(t->rcu_read_unlock_special))) |
| 397 | unlikely(ACCESS_ONCE(t->rcu_read_unlock_special))) | 397 | rcu_read_unlock_special(t); |
| 398 | rcu_read_unlock_special(t); | 398 | } |
| 399 | #ifdef CONFIG_PROVE_LOCKING | 399 | #ifdef CONFIG_PROVE_LOCKING |
| 400 | WARN_ON_ONCE(ACCESS_ONCE(t->rcu_read_lock_nesting) < 0); | 400 | WARN_ON_ONCE(ACCESS_ONCE(t->rcu_read_lock_nesting) < 0); |
| 401 | #endif /* #ifdef CONFIG_PROVE_LOCKING */ | 401 | #endif /* #ifdef CONFIG_PROVE_LOCKING */ |
