diff options
author | Andrea Bastoni <bastoni@cs.unc.edu> | 2010-05-30 19:16:45 -0400 |
---|---|---|
committer | Andrea Bastoni <bastoni@cs.unc.edu> | 2010-05-30 19:16:45 -0400 |
commit | ada47b5fe13d89735805b566185f4885f5a3f750 (patch) | |
tree | 644b88f8a71896307d71438e9b3af49126ffb22b /kernel/rcupdate.c | |
parent | 43e98717ad40a4ae64545b5ba047c7b86aa44f4f (diff) | |
parent | 3280f21d43ee541f97f8cda5792150d2dbec20d5 (diff) |
Merge branch 'wip-2.6.34' into old-private-masterarchived-private-master
Diffstat (limited to 'kernel/rcupdate.c')
-rw-r--r-- | kernel/rcupdate.c | 166 |
1 files changed, 57 insertions, 109 deletions
diff --git a/kernel/rcupdate.c b/kernel/rcupdate.c index 400183346ad2..49d808e833b0 100644 --- a/kernel/rcupdate.c +++ b/kernel/rcupdate.c | |||
@@ -45,143 +45,91 @@ | |||
45 | #include <linux/mutex.h> | 45 | #include <linux/mutex.h> |
46 | #include <linux/module.h> | 46 | #include <linux/module.h> |
47 | #include <linux/kernel_stat.h> | 47 | #include <linux/kernel_stat.h> |
48 | #include <linux/hardirq.h> | ||
48 | 49 | ||
49 | #ifdef CONFIG_DEBUG_LOCK_ALLOC | 50 | #ifdef CONFIG_DEBUG_LOCK_ALLOC |
50 | static struct lock_class_key rcu_lock_key; | 51 | static struct lock_class_key rcu_lock_key; |
51 | struct lockdep_map rcu_lock_map = | 52 | struct lockdep_map rcu_lock_map = |
52 | STATIC_LOCKDEP_MAP_INIT("rcu_read_lock", &rcu_lock_key); | 53 | STATIC_LOCKDEP_MAP_INIT("rcu_read_lock", &rcu_lock_key); |
53 | EXPORT_SYMBOL_GPL(rcu_lock_map); | 54 | EXPORT_SYMBOL_GPL(rcu_lock_map); |
54 | #endif | ||
55 | 55 | ||
56 | int rcu_scheduler_active __read_mostly; | 56 | static struct lock_class_key rcu_bh_lock_key; |
57 | struct lockdep_map rcu_bh_lock_map = | ||
58 | STATIC_LOCKDEP_MAP_INIT("rcu_read_lock_bh", &rcu_bh_lock_key); | ||
59 | EXPORT_SYMBOL_GPL(rcu_bh_lock_map); | ||
57 | 60 | ||
58 | /* | 61 | static struct lock_class_key rcu_sched_lock_key; |
59 | * Awaken the corresponding synchronize_rcu() instance now that a | 62 | struct lockdep_map rcu_sched_lock_map = |
60 | * grace period has elapsed. | 63 | STATIC_LOCKDEP_MAP_INIT("rcu_read_lock_sched", &rcu_sched_lock_key); |
61 | */ | 64 | EXPORT_SYMBOL_GPL(rcu_sched_lock_map); |
62 | void wakeme_after_rcu(struct rcu_head *head) | 65 | #endif |
63 | { | ||
64 | struct rcu_synchronize *rcu; | ||
65 | 66 | ||
66 | rcu = container_of(head, struct rcu_synchronize, head); | 67 | int rcu_scheduler_active __read_mostly; |
67 | complete(&rcu->completion); | 68 | EXPORT_SYMBOL_GPL(rcu_scheduler_active); |
68 | } | ||
69 | 69 | ||
70 | #ifdef CONFIG_TREE_PREEMPT_RCU | 70 | #ifdef CONFIG_DEBUG_LOCK_ALLOC |
71 | 71 | ||
72 | /** | 72 | int debug_lockdep_rcu_enabled(void) |
73 | * synchronize_rcu - wait until a grace period has elapsed. | ||
74 | * | ||
75 | * Control will return to the caller some time after a full grace | ||
76 | * period has elapsed, in other words after all currently executing RCU | ||
77 | * read-side critical sections have completed. RCU read-side critical | ||
78 | * sections are delimited by rcu_read_lock() and rcu_read_unlock(), | ||
79 | * and may be nested. | ||
80 | */ | ||
81 | void synchronize_rcu(void) | ||
82 | { | 73 | { |
83 | struct rcu_synchronize rcu; | 74 | return rcu_scheduler_active && debug_locks && |
84 | 75 | current->lockdep_recursion == 0; | |
85 | if (!rcu_scheduler_active) | ||
86 | return; | ||
87 | |||
88 | init_completion(&rcu.completion); | ||
89 | /* Will wake me after RCU finished. */ | ||
90 | call_rcu(&rcu.head, wakeme_after_rcu); | ||
91 | /* Wait for it. */ | ||
92 | wait_for_completion(&rcu.completion); | ||
93 | } | 76 | } |
94 | EXPORT_SYMBOL_GPL(synchronize_rcu); | 77 | EXPORT_SYMBOL_GPL(debug_lockdep_rcu_enabled); |
95 | |||
96 | #endif /* #ifdef CONFIG_TREE_PREEMPT_RCU */ | ||
97 | 78 | ||
98 | /** | 79 | /** |
99 | * synchronize_sched - wait until an rcu-sched grace period has elapsed. | 80 | * rcu_read_lock_bh_held - might we be in RCU-bh read-side critical section? |
100 | * | ||
101 | * Control will return to the caller some time after a full rcu-sched | ||
102 | * grace period has elapsed, in other words after all currently executing | ||
103 | * rcu-sched read-side critical sections have completed. These read-side | ||
104 | * critical sections are delimited by rcu_read_lock_sched() and | ||
105 | * rcu_read_unlock_sched(), and may be nested. Note that preempt_disable(), | ||
106 | * local_irq_disable(), and so on may be used in place of | ||
107 | * rcu_read_lock_sched(). | ||
108 | * | 81 | * |
109 | * This means that all preempt_disable code sequences, including NMI and | 82 | * Check for bottom half being disabled, which covers both the |
110 | * hardware-interrupt handlers, in progress on entry will have completed | 83 | * CONFIG_PROVE_RCU and not cases. Note that if someone uses |
111 | * before this primitive returns. However, this does not guarantee that | 84 | * rcu_read_lock_bh(), but then later enables BH, lockdep (if enabled) |
112 | * softirq handlers will have completed, since in some kernels, these | 85 | * will show the situation. |
113 | * handlers can run in process context, and can block. | ||
114 | * | 86 | * |
115 | * This primitive provides the guarantees made by the (now removed) | 87 | * Check debug_lockdep_rcu_enabled() to prevent false positives during boot. |
116 | * synchronize_kernel() API. In contrast, synchronize_rcu() only | ||
117 | * guarantees that rcu_read_lock() sections will have completed. | ||
118 | * In "classic RCU", these two guarantees happen to be one and | ||
119 | * the same, but can differ in realtime RCU implementations. | ||
120 | */ | 88 | */ |
121 | void synchronize_sched(void) | 89 | int rcu_read_lock_bh_held(void) |
122 | { | 90 | { |
123 | struct rcu_synchronize rcu; | 91 | if (!debug_lockdep_rcu_enabled()) |
124 | 92 | return 1; | |
125 | if (rcu_blocking_is_gp()) | 93 | return in_softirq(); |
126 | return; | ||
127 | |||
128 | init_completion(&rcu.completion); | ||
129 | /* Will wake me after RCU finished. */ | ||
130 | call_rcu_sched(&rcu.head, wakeme_after_rcu); | ||
131 | /* Wait for it. */ | ||
132 | wait_for_completion(&rcu.completion); | ||
133 | } | 94 | } |
134 | EXPORT_SYMBOL_GPL(synchronize_sched); | 95 | EXPORT_SYMBOL_GPL(rcu_read_lock_bh_held); |
135 | |||
136 | /** | ||
137 | * synchronize_rcu_bh - wait until an rcu_bh grace period has elapsed. | ||
138 | * | ||
139 | * Control will return to the caller some time after a full rcu_bh grace | ||
140 | * period has elapsed, in other words after all currently executing rcu_bh | ||
141 | * read-side critical sections have completed. RCU read-side critical | ||
142 | * sections are delimited by rcu_read_lock_bh() and rcu_read_unlock_bh(), | ||
143 | * and may be nested. | ||
144 | */ | ||
145 | void synchronize_rcu_bh(void) | ||
146 | { | ||
147 | struct rcu_synchronize rcu; | ||
148 | 96 | ||
149 | if (rcu_blocking_is_gp()) | 97 | #endif /* #ifdef CONFIG_DEBUG_LOCK_ALLOC */ |
150 | return; | ||
151 | 98 | ||
152 | init_completion(&rcu.completion); | 99 | /* |
153 | /* Will wake me after RCU finished. */ | 100 | * This function is invoked towards the end of the scheduler's initialization |
154 | call_rcu_bh(&rcu.head, wakeme_after_rcu); | 101 | * process. Before this is called, the idle task might contain |
155 | /* Wait for it. */ | 102 | * RCU read-side critical sections (during which time, this idle |
156 | wait_for_completion(&rcu.completion); | 103 | * task is booting the system). After this function is called, the |
157 | } | 104 | * idle tasks are prohibited from containing RCU read-side critical |
158 | EXPORT_SYMBOL_GPL(synchronize_rcu_bh); | 105 | * sections. |
159 | 106 | */ | |
160 | static int __cpuinit rcu_barrier_cpu_hotplug(struct notifier_block *self, | 107 | void rcu_scheduler_starting(void) |
161 | unsigned long action, void *hcpu) | ||
162 | { | 108 | { |
163 | return rcu_cpu_notify(self, action, hcpu); | 109 | WARN_ON(num_online_cpus() != 1); |
110 | WARN_ON(nr_context_switches() > 0); | ||
111 | rcu_scheduler_active = 1; | ||
164 | } | 112 | } |
165 | 113 | ||
166 | void __init rcu_init(void) | 114 | /* |
115 | * Awaken the corresponding synchronize_rcu() instance now that a | ||
116 | * grace period has elapsed. | ||
117 | */ | ||
118 | void wakeme_after_rcu(struct rcu_head *head) | ||
167 | { | 119 | { |
168 | int i; | 120 | struct rcu_synchronize *rcu; |
169 | |||
170 | __rcu_init(); | ||
171 | cpu_notifier(rcu_barrier_cpu_hotplug, 0); | ||
172 | 121 | ||
173 | /* | 122 | rcu = container_of(head, struct rcu_synchronize, head); |
174 | * We don't need protection against CPU-hotplug here because | 123 | complete(&rcu->completion); |
175 | * this is called early in boot, before either interrupts | ||
176 | * or the scheduler are operational. | ||
177 | */ | ||
178 | for_each_online_cpu(i) | ||
179 | rcu_barrier_cpu_hotplug(NULL, CPU_UP_PREPARE, (void *)(long)i); | ||
180 | } | 124 | } |
181 | 125 | ||
182 | void rcu_scheduler_starting(void) | 126 | #ifdef CONFIG_PROVE_RCU |
127 | /* | ||
128 | * wrapper function to avoid #include problems. | ||
129 | */ | ||
130 | int rcu_my_thread_group_empty(void) | ||
183 | { | 131 | { |
184 | WARN_ON(num_online_cpus() != 1); | 132 | return thread_group_empty(current); |
185 | WARN_ON(nr_context_switches() > 0); | ||
186 | rcu_scheduler_active = 1; | ||
187 | } | 133 | } |
134 | EXPORT_SYMBOL_GPL(rcu_my_thread_group_empty); | ||
135 | #endif /* #ifdef CONFIG_PROVE_RCU */ | ||