aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/rcupdate.c
diff options
context:
space:
mode:
authorAndrea Bastoni <bastoni@cs.unc.edu>2010-05-30 19:16:45 -0400
committerAndrea Bastoni <bastoni@cs.unc.edu>2010-05-30 19:16:45 -0400
commitada47b5fe13d89735805b566185f4885f5a3f750 (patch)
tree644b88f8a71896307d71438e9b3af49126ffb22b /kernel/rcupdate.c
parent43e98717ad40a4ae64545b5ba047c7b86aa44f4f (diff)
parent3280f21d43ee541f97f8cda5792150d2dbec20d5 (diff)
Merge branch 'wip-2.6.34' into old-private-masterarchived-private-master
Diffstat (limited to 'kernel/rcupdate.c')
-rw-r--r--kernel/rcupdate.c166
1 files changed, 57 insertions, 109 deletions
diff --git a/kernel/rcupdate.c b/kernel/rcupdate.c
index 400183346ad2..49d808e833b0 100644
--- a/kernel/rcupdate.c
+++ b/kernel/rcupdate.c
@@ -45,143 +45,91 @@
45#include <linux/mutex.h> 45#include <linux/mutex.h>
46#include <linux/module.h> 46#include <linux/module.h>
47#include <linux/kernel_stat.h> 47#include <linux/kernel_stat.h>
48#include <linux/hardirq.h>
48 49
49#ifdef CONFIG_DEBUG_LOCK_ALLOC 50#ifdef CONFIG_DEBUG_LOCK_ALLOC
50static struct lock_class_key rcu_lock_key; 51static struct lock_class_key rcu_lock_key;
51struct lockdep_map rcu_lock_map = 52struct lockdep_map rcu_lock_map =
52 STATIC_LOCKDEP_MAP_INIT("rcu_read_lock", &rcu_lock_key); 53 STATIC_LOCKDEP_MAP_INIT("rcu_read_lock", &rcu_lock_key);
53EXPORT_SYMBOL_GPL(rcu_lock_map); 54EXPORT_SYMBOL_GPL(rcu_lock_map);
54#endif
55 55
56int rcu_scheduler_active __read_mostly; 56static struct lock_class_key rcu_bh_lock_key;
57struct lockdep_map rcu_bh_lock_map =
58 STATIC_LOCKDEP_MAP_INIT("rcu_read_lock_bh", &rcu_bh_lock_key);
59EXPORT_SYMBOL_GPL(rcu_bh_lock_map);
57 60
58/* 61static struct lock_class_key rcu_sched_lock_key;
59 * Awaken the corresponding synchronize_rcu() instance now that a 62struct lockdep_map rcu_sched_lock_map =
60 * grace period has elapsed. 63 STATIC_LOCKDEP_MAP_INIT("rcu_read_lock_sched", &rcu_sched_lock_key);
61 */ 64EXPORT_SYMBOL_GPL(rcu_sched_lock_map);
62void wakeme_after_rcu(struct rcu_head *head) 65#endif
63{
64 struct rcu_synchronize *rcu;
65 66
66 rcu = container_of(head, struct rcu_synchronize, head); 67int rcu_scheduler_active __read_mostly;
67 complete(&rcu->completion); 68EXPORT_SYMBOL_GPL(rcu_scheduler_active);
68}
69 69
70#ifdef CONFIG_TREE_PREEMPT_RCU 70#ifdef CONFIG_DEBUG_LOCK_ALLOC
71 71
72/** 72int debug_lockdep_rcu_enabled(void)
73 * synchronize_rcu - wait until a grace period has elapsed.
74 *
75 * Control will return to the caller some time after a full grace
76 * period has elapsed, in other words after all currently executing RCU
77 * read-side critical sections have completed. RCU read-side critical
78 * sections are delimited by rcu_read_lock() and rcu_read_unlock(),
79 * and may be nested.
80 */
81void synchronize_rcu(void)
82{ 73{
83 struct rcu_synchronize rcu; 74 return rcu_scheduler_active && debug_locks &&
84 75 current->lockdep_recursion == 0;
85 if (!rcu_scheduler_active)
86 return;
87
88 init_completion(&rcu.completion);
89 /* Will wake me after RCU finished. */
90 call_rcu(&rcu.head, wakeme_after_rcu);
91 /* Wait for it. */
92 wait_for_completion(&rcu.completion);
93} 76}
94EXPORT_SYMBOL_GPL(synchronize_rcu); 77EXPORT_SYMBOL_GPL(debug_lockdep_rcu_enabled);
95
96#endif /* #ifdef CONFIG_TREE_PREEMPT_RCU */
97 78
98/** 79/**
99 * synchronize_sched - wait until an rcu-sched grace period has elapsed. 80 * rcu_read_lock_bh_held - might we be in RCU-bh read-side critical section?
100 *
101 * Control will return to the caller some time after a full rcu-sched
102 * grace period has elapsed, in other words after all currently executing
103 * rcu-sched read-side critical sections have completed. These read-side
104 * critical sections are delimited by rcu_read_lock_sched() and
105 * rcu_read_unlock_sched(), and may be nested. Note that preempt_disable(),
106 * local_irq_disable(), and so on may be used in place of
107 * rcu_read_lock_sched().
108 * 81 *
109 * This means that all preempt_disable code sequences, including NMI and 82 * Check for bottom half being disabled, which covers both the
110 * hardware-interrupt handlers, in progress on entry will have completed 83 * CONFIG_PROVE_RCU and not cases. Note that if someone uses
111 * before this primitive returns. However, this does not guarantee that 84 * rcu_read_lock_bh(), but then later enables BH, lockdep (if enabled)
112 * softirq handlers will have completed, since in some kernels, these 85 * will show the situation.
113 * handlers can run in process context, and can block.
114 * 86 *
115 * This primitive provides the guarantees made by the (now removed) 87 * Check debug_lockdep_rcu_enabled() to prevent false positives during boot.
116 * synchronize_kernel() API. In contrast, synchronize_rcu() only
117 * guarantees that rcu_read_lock() sections will have completed.
118 * In "classic RCU", these two guarantees happen to be one and
119 * the same, but can differ in realtime RCU implementations.
120 */ 88 */
121void synchronize_sched(void) 89int rcu_read_lock_bh_held(void)
122{ 90{
123 struct rcu_synchronize rcu; 91 if (!debug_lockdep_rcu_enabled())
124 92 return 1;
125 if (rcu_blocking_is_gp()) 93 return in_softirq();
126 return;
127
128 init_completion(&rcu.completion);
129 /* Will wake me after RCU finished. */
130 call_rcu_sched(&rcu.head, wakeme_after_rcu);
131 /* Wait for it. */
132 wait_for_completion(&rcu.completion);
133} 94}
134EXPORT_SYMBOL_GPL(synchronize_sched); 95EXPORT_SYMBOL_GPL(rcu_read_lock_bh_held);
135
136/**
137 * synchronize_rcu_bh - wait until an rcu_bh grace period has elapsed.
138 *
139 * Control will return to the caller some time after a full rcu_bh grace
140 * period has elapsed, in other words after all currently executing rcu_bh
141 * read-side critical sections have completed. RCU read-side critical
142 * sections are delimited by rcu_read_lock_bh() and rcu_read_unlock_bh(),
143 * and may be nested.
144 */
145void synchronize_rcu_bh(void)
146{
147 struct rcu_synchronize rcu;
148 96
149 if (rcu_blocking_is_gp()) 97#endif /* #ifdef CONFIG_DEBUG_LOCK_ALLOC */
150 return;
151 98
152 init_completion(&rcu.completion); 99/*
153 /* Will wake me after RCU finished. */ 100 * This function is invoked towards the end of the scheduler's initialization
154 call_rcu_bh(&rcu.head, wakeme_after_rcu); 101 * process. Before this is called, the idle task might contain
155 /* Wait for it. */ 102 * RCU read-side critical sections (during which time, this idle
156 wait_for_completion(&rcu.completion); 103 * task is booting the system). After this function is called, the
157} 104 * idle tasks are prohibited from containing RCU read-side critical
158EXPORT_SYMBOL_GPL(synchronize_rcu_bh); 105 * sections.
159 106 */
160static int __cpuinit rcu_barrier_cpu_hotplug(struct notifier_block *self, 107void rcu_scheduler_starting(void)
161 unsigned long action, void *hcpu)
162{ 108{
163 return rcu_cpu_notify(self, action, hcpu); 109 WARN_ON(num_online_cpus() != 1);
110 WARN_ON(nr_context_switches() > 0);
111 rcu_scheduler_active = 1;
164} 112}
165 113
166void __init rcu_init(void) 114/*
115 * Awaken the corresponding synchronize_rcu() instance now that a
116 * grace period has elapsed.
117 */
118void wakeme_after_rcu(struct rcu_head *head)
167{ 119{
168 int i; 120 struct rcu_synchronize *rcu;
169
170 __rcu_init();
171 cpu_notifier(rcu_barrier_cpu_hotplug, 0);
172 121
173 /* 122 rcu = container_of(head, struct rcu_synchronize, head);
174 * We don't need protection against CPU-hotplug here because 123 complete(&rcu->completion);
175 * this is called early in boot, before either interrupts
176 * or the scheduler are operational.
177 */
178 for_each_online_cpu(i)
179 rcu_barrier_cpu_hotplug(NULL, CPU_UP_PREPARE, (void *)(long)i);
180} 124}
181 125
182void rcu_scheduler_starting(void) 126#ifdef CONFIG_PROVE_RCU
127/*
128 * wrapper function to avoid #include problems.
129 */
130int rcu_my_thread_group_empty(void)
183{ 131{
184 WARN_ON(num_online_cpus() != 1); 132 return thread_group_empty(current);
185 WARN_ON(nr_context_switches() > 0);
186 rcu_scheduler_active = 1;
187} 133}
134EXPORT_SYMBOL_GPL(rcu_my_thread_group_empty);
135#endif /* #ifdef CONFIG_PROVE_RCU */