diff options
Diffstat (limited to 'include/linux/wait.h')
-rw-r--r-- | include/linux/wait.h | 393 |
1 files changed, 163 insertions, 230 deletions
diff --git a/include/linux/wait.h b/include/linux/wait.h index a67fc1635592..eaa00b10abaa 100644 --- a/include/linux/wait.h +++ b/include/linux/wait.h | |||
@@ -1,7 +1,8 @@ | |||
1 | #ifndef _LINUX_WAIT_H | 1 | #ifndef _LINUX_WAIT_H |
2 | #define _LINUX_WAIT_H | 2 | #define _LINUX_WAIT_H |
3 | 3 | /* | |
4 | 4 | * Linux wait queue related types and methods | |
5 | */ | ||
5 | #include <linux/list.h> | 6 | #include <linux/list.h> |
6 | #include <linux/stddef.h> | 7 | #include <linux/stddef.h> |
7 | #include <linux/spinlock.h> | 8 | #include <linux/spinlock.h> |
@@ -13,27 +14,27 @@ typedef int (*wait_queue_func_t)(wait_queue_t *wait, unsigned mode, int flags, v | |||
13 | int default_wake_function(wait_queue_t *wait, unsigned mode, int flags, void *key); | 14 | int default_wake_function(wait_queue_t *wait, unsigned mode, int flags, void *key); |
14 | 15 | ||
15 | struct __wait_queue { | 16 | struct __wait_queue { |
16 | unsigned int flags; | 17 | unsigned int flags; |
17 | #define WQ_FLAG_EXCLUSIVE 0x01 | 18 | #define WQ_FLAG_EXCLUSIVE 0x01 |
18 | void *private; | 19 | void *private; |
19 | wait_queue_func_t func; | 20 | wait_queue_func_t func; |
20 | struct list_head task_list; | 21 | struct list_head task_list; |
21 | }; | 22 | }; |
22 | 23 | ||
23 | struct wait_bit_key { | 24 | struct wait_bit_key { |
24 | void *flags; | 25 | void *flags; |
25 | int bit_nr; | 26 | int bit_nr; |
26 | #define WAIT_ATOMIC_T_BIT_NR -1 | 27 | #define WAIT_ATOMIC_T_BIT_NR -1 |
27 | }; | 28 | }; |
28 | 29 | ||
29 | struct wait_bit_queue { | 30 | struct wait_bit_queue { |
30 | struct wait_bit_key key; | 31 | struct wait_bit_key key; |
31 | wait_queue_t wait; | 32 | wait_queue_t wait; |
32 | }; | 33 | }; |
33 | 34 | ||
34 | struct __wait_queue_head { | 35 | struct __wait_queue_head { |
35 | spinlock_t lock; | 36 | spinlock_t lock; |
36 | struct list_head task_list; | 37 | struct list_head task_list; |
37 | }; | 38 | }; |
38 | typedef struct __wait_queue_head wait_queue_head_t; | 39 | typedef struct __wait_queue_head wait_queue_head_t; |
39 | 40 | ||
@@ -84,17 +85,17 @@ extern void __init_waitqueue_head(wait_queue_head_t *q, const char *name, struct | |||
84 | 85 | ||
85 | static inline void init_waitqueue_entry(wait_queue_t *q, struct task_struct *p) | 86 | static inline void init_waitqueue_entry(wait_queue_t *q, struct task_struct *p) |
86 | { | 87 | { |
87 | q->flags = 0; | 88 | q->flags = 0; |
88 | q->private = p; | 89 | q->private = p; |
89 | q->func = default_wake_function; | 90 | q->func = default_wake_function; |
90 | } | 91 | } |
91 | 92 | ||
92 | static inline void init_waitqueue_func_entry(wait_queue_t *q, | 93 | static inline void |
93 | wait_queue_func_t func) | 94 | init_waitqueue_func_entry(wait_queue_t *q, wait_queue_func_t func) |
94 | { | 95 | { |
95 | q->flags = 0; | 96 | q->flags = 0; |
96 | q->private = NULL; | 97 | q->private = NULL; |
97 | q->func = func; | 98 | q->func = func; |
98 | } | 99 | } |
99 | 100 | ||
100 | static inline int waitqueue_active(wait_queue_head_t *q) | 101 | static inline int waitqueue_active(wait_queue_head_t *q) |
@@ -114,8 +115,8 @@ static inline void __add_wait_queue(wait_queue_head_t *head, wait_queue_t *new) | |||
114 | /* | 115 | /* |
115 | * Used for wake-one threads: | 116 | * Used for wake-one threads: |
116 | */ | 117 | */ |
117 | static inline void __add_wait_queue_exclusive(wait_queue_head_t *q, | 118 | static inline void |
118 | wait_queue_t *wait) | 119 | __add_wait_queue_exclusive(wait_queue_head_t *q, wait_queue_t *wait) |
119 | { | 120 | { |
120 | wait->flags |= WQ_FLAG_EXCLUSIVE; | 121 | wait->flags |= WQ_FLAG_EXCLUSIVE; |
121 | __add_wait_queue(q, wait); | 122 | __add_wait_queue(q, wait); |
@@ -127,23 +128,22 @@ static inline void __add_wait_queue_tail(wait_queue_head_t *head, | |||
127 | list_add_tail(&new->task_list, &head->task_list); | 128 | list_add_tail(&new->task_list, &head->task_list); |
128 | } | 129 | } |
129 | 130 | ||
130 | static inline void __add_wait_queue_tail_exclusive(wait_queue_head_t *q, | 131 | static inline void |
131 | wait_queue_t *wait) | 132 | __add_wait_queue_tail_exclusive(wait_queue_head_t *q, wait_queue_t *wait) |
132 | { | 133 | { |
133 | wait->flags |= WQ_FLAG_EXCLUSIVE; | 134 | wait->flags |= WQ_FLAG_EXCLUSIVE; |
134 | __add_wait_queue_tail(q, wait); | 135 | __add_wait_queue_tail(q, wait); |
135 | } | 136 | } |
136 | 137 | ||
137 | static inline void __remove_wait_queue(wait_queue_head_t *head, | 138 | static inline void |
138 | wait_queue_t *old) | 139 | __remove_wait_queue(wait_queue_head_t *head, wait_queue_t *old) |
139 | { | 140 | { |
140 | list_del(&old->task_list); | 141 | list_del(&old->task_list); |
141 | } | 142 | } |
142 | 143 | ||
143 | void __wake_up(wait_queue_head_t *q, unsigned int mode, int nr, void *key); | 144 | void __wake_up(wait_queue_head_t *q, unsigned int mode, int nr, void *key); |
144 | void __wake_up_locked_key(wait_queue_head_t *q, unsigned int mode, void *key); | 145 | void __wake_up_locked_key(wait_queue_head_t *q, unsigned int mode, void *key); |
145 | void __wake_up_sync_key(wait_queue_head_t *q, unsigned int mode, int nr, | 146 | void __wake_up_sync_key(wait_queue_head_t *q, unsigned int mode, int nr, void *key); |
146 | void *key); | ||
147 | void __wake_up_locked(wait_queue_head_t *q, unsigned int mode, int nr); | 147 | void __wake_up_locked(wait_queue_head_t *q, unsigned int mode, int nr); |
148 | void __wake_up_sync(wait_queue_head_t *q, unsigned int mode, int nr); | 148 | void __wake_up_sync(wait_queue_head_t *q, unsigned int mode, int nr); |
149 | void __wake_up_bit(wait_queue_head_t *, void *, int); | 149 | void __wake_up_bit(wait_queue_head_t *, void *, int); |
@@ -170,27 +170,64 @@ wait_queue_head_t *bit_waitqueue(void *, int); | |||
170 | /* | 170 | /* |
171 | * Wakeup macros to be used to report events to the targets. | 171 | * Wakeup macros to be used to report events to the targets. |
172 | */ | 172 | */ |
173 | #define wake_up_poll(x, m) \ | 173 | #define wake_up_poll(x, m) \ |
174 | __wake_up(x, TASK_NORMAL, 1, (void *) (m)) | 174 | __wake_up(x, TASK_NORMAL, 1, (void *) (m)) |
175 | #define wake_up_locked_poll(x, m) \ | 175 | #define wake_up_locked_poll(x, m) \ |
176 | __wake_up_locked_key((x), TASK_NORMAL, (void *) (m)) | 176 | __wake_up_locked_key((x), TASK_NORMAL, (void *) (m)) |
177 | #define wake_up_interruptible_poll(x, m) \ | 177 | #define wake_up_interruptible_poll(x, m) \ |
178 | __wake_up(x, TASK_INTERRUPTIBLE, 1, (void *) (m)) | 178 | __wake_up(x, TASK_INTERRUPTIBLE, 1, (void *) (m)) |
179 | #define wake_up_interruptible_sync_poll(x, m) \ | 179 | #define wake_up_interruptible_sync_poll(x, m) \ |
180 | __wake_up_sync_key((x), TASK_INTERRUPTIBLE, 1, (void *) (m)) | 180 | __wake_up_sync_key((x), TASK_INTERRUPTIBLE, 1, (void *) (m)) |
181 | 181 | ||
182 | #define __wait_event(wq, condition) \ | 182 | #define ___wait_cond_timeout(condition) \ |
183 | do { \ | 183 | ({ \ |
184 | DEFINE_WAIT(__wait); \ | 184 | bool __cond = (condition); \ |
185 | if (__cond && !__ret) \ | ||
186 | __ret = 1; \ | ||
187 | __cond || !__ret; \ | ||
188 | }) | ||
189 | |||
190 | #define ___wait_is_interruptible(state) \ | ||
191 | (!__builtin_constant_p(state) || \ | ||
192 | state == TASK_INTERRUPTIBLE || state == TASK_KILLABLE) \ | ||
193 | |||
194 | #define ___wait_event(wq, condition, state, exclusive, ret, cmd) \ | ||
195 | ({ \ | ||
196 | __label__ __out; \ | ||
197 | wait_queue_t __wait; \ | ||
198 | long __ret = ret; \ | ||
199 | \ | ||
200 | INIT_LIST_HEAD(&__wait.task_list); \ | ||
201 | if (exclusive) \ | ||
202 | __wait.flags = WQ_FLAG_EXCLUSIVE; \ | ||
203 | else \ | ||
204 | __wait.flags = 0; \ | ||
185 | \ | 205 | \ |
186 | for (;;) { \ | 206 | for (;;) { \ |
187 | prepare_to_wait(&wq, &__wait, TASK_UNINTERRUPTIBLE); \ | 207 | long __int = prepare_to_wait_event(&wq, &__wait, state);\ |
208 | \ | ||
188 | if (condition) \ | 209 | if (condition) \ |
189 | break; \ | 210 | break; \ |
190 | schedule(); \ | 211 | \ |
212 | if (___wait_is_interruptible(state) && __int) { \ | ||
213 | __ret = __int; \ | ||
214 | if (exclusive) { \ | ||
215 | abort_exclusive_wait(&wq, &__wait, \ | ||
216 | state, NULL); \ | ||
217 | goto __out; \ | ||
218 | } \ | ||
219 | break; \ | ||
220 | } \ | ||
221 | \ | ||
222 | cmd; \ | ||
191 | } \ | 223 | } \ |
192 | finish_wait(&wq, &__wait); \ | 224 | finish_wait(&wq, &__wait); \ |
193 | } while (0) | 225 | __out: __ret; \ |
226 | }) | ||
227 | |||
228 | #define __wait_event(wq, condition) \ | ||
229 | (void)___wait_event(wq, condition, TASK_UNINTERRUPTIBLE, 0, 0, \ | ||
230 | schedule()) | ||
194 | 231 | ||
195 | /** | 232 | /** |
196 | * wait_event - sleep until a condition gets true | 233 | * wait_event - sleep until a condition gets true |
@@ -204,29 +241,17 @@ do { \ | |||
204 | * wake_up() has to be called after changing any variable that could | 241 | * wake_up() has to be called after changing any variable that could |
205 | * change the result of the wait condition. | 242 | * change the result of the wait condition. |
206 | */ | 243 | */ |
207 | #define wait_event(wq, condition) \ | 244 | #define wait_event(wq, condition) \ |
208 | do { \ | 245 | do { \ |
209 | if (condition) \ | 246 | if (condition) \ |
210 | break; \ | 247 | break; \ |
211 | __wait_event(wq, condition); \ | 248 | __wait_event(wq, condition); \ |
212 | } while (0) | 249 | } while (0) |
213 | 250 | ||
214 | #define __wait_event_timeout(wq, condition, ret) \ | 251 | #define __wait_event_timeout(wq, condition, timeout) \ |
215 | do { \ | 252 | ___wait_event(wq, ___wait_cond_timeout(condition), \ |
216 | DEFINE_WAIT(__wait); \ | 253 | TASK_UNINTERRUPTIBLE, 0, timeout, \ |
217 | \ | 254 | __ret = schedule_timeout(__ret)) |
218 | for (;;) { \ | ||
219 | prepare_to_wait(&wq, &__wait, TASK_UNINTERRUPTIBLE); \ | ||
220 | if (condition) \ | ||
221 | break; \ | ||
222 | ret = schedule_timeout(ret); \ | ||
223 | if (!ret) \ | ||
224 | break; \ | ||
225 | } \ | ||
226 | if (!ret && (condition)) \ | ||
227 | ret = 1; \ | ||
228 | finish_wait(&wq, &__wait); \ | ||
229 | } while (0) | ||
230 | 255 | ||
231 | /** | 256 | /** |
232 | * wait_event_timeout - sleep until a condition gets true or a timeout elapses | 257 | * wait_event_timeout - sleep until a condition gets true or a timeout elapses |
@@ -248,29 +273,40 @@ do { \ | |||
248 | #define wait_event_timeout(wq, condition, timeout) \ | 273 | #define wait_event_timeout(wq, condition, timeout) \ |
249 | ({ \ | 274 | ({ \ |
250 | long __ret = timeout; \ | 275 | long __ret = timeout; \ |
251 | if (!(condition)) \ | 276 | if (!___wait_cond_timeout(condition)) \ |
252 | __wait_event_timeout(wq, condition, __ret); \ | 277 | __ret = __wait_event_timeout(wq, condition, timeout); \ |
253 | __ret; \ | 278 | __ret; \ |
254 | }) | 279 | }) |
255 | 280 | ||
256 | #define __wait_event_interruptible(wq, condition, ret) \ | 281 | #define __wait_event_cmd(wq, condition, cmd1, cmd2) \ |
282 | (void)___wait_event(wq, condition, TASK_UNINTERRUPTIBLE, 0, 0, \ | ||
283 | cmd1; schedule(); cmd2) | ||
284 | |||
285 | /** | ||
286 | * wait_event_cmd - sleep until a condition gets true | ||
287 | * @wq: the waitqueue to wait on | ||
288 | * @condition: a C expression for the event to wait for | ||
289 | * cmd1: the command will be executed before sleep | ||
290 | * cmd2: the command will be executed after sleep | ||
291 | * | ||
292 | * The process is put to sleep (TASK_UNINTERRUPTIBLE) until the | ||
293 | * @condition evaluates to true. The @condition is checked each time | ||
294 | * the waitqueue @wq is woken up. | ||
295 | * | ||
296 | * wake_up() has to be called after changing any variable that could | ||
297 | * change the result of the wait condition. | ||
298 | */ | ||
299 | #define wait_event_cmd(wq, condition, cmd1, cmd2) \ | ||
257 | do { \ | 300 | do { \ |
258 | DEFINE_WAIT(__wait); \ | 301 | if (condition) \ |
259 | \ | ||
260 | for (;;) { \ | ||
261 | prepare_to_wait(&wq, &__wait, TASK_INTERRUPTIBLE); \ | ||
262 | if (condition) \ | ||
263 | break; \ | ||
264 | if (!signal_pending(current)) { \ | ||
265 | schedule(); \ | ||
266 | continue; \ | ||
267 | } \ | ||
268 | ret = -ERESTARTSYS; \ | ||
269 | break; \ | 302 | break; \ |
270 | } \ | 303 | __wait_event_cmd(wq, condition, cmd1, cmd2); \ |
271 | finish_wait(&wq, &__wait); \ | ||
272 | } while (0) | 304 | } while (0) |
273 | 305 | ||
306 | #define __wait_event_interruptible(wq, condition) \ | ||
307 | ___wait_event(wq, condition, TASK_INTERRUPTIBLE, 0, 0, \ | ||
308 | schedule()) | ||
309 | |||
274 | /** | 310 | /** |
275 | * wait_event_interruptible - sleep until a condition gets true | 311 | * wait_event_interruptible - sleep until a condition gets true |
276 | * @wq: the waitqueue to wait on | 312 | * @wq: the waitqueue to wait on |
@@ -290,31 +326,14 @@ do { \ | |||
290 | ({ \ | 326 | ({ \ |
291 | int __ret = 0; \ | 327 | int __ret = 0; \ |
292 | if (!(condition)) \ | 328 | if (!(condition)) \ |
293 | __wait_event_interruptible(wq, condition, __ret); \ | 329 | __ret = __wait_event_interruptible(wq, condition); \ |
294 | __ret; \ | 330 | __ret; \ |
295 | }) | 331 | }) |
296 | 332 | ||
297 | #define __wait_event_interruptible_timeout(wq, condition, ret) \ | 333 | #define __wait_event_interruptible_timeout(wq, condition, timeout) \ |
298 | do { \ | 334 | ___wait_event(wq, ___wait_cond_timeout(condition), \ |
299 | DEFINE_WAIT(__wait); \ | 335 | TASK_INTERRUPTIBLE, 0, timeout, \ |
300 | \ | 336 | __ret = schedule_timeout(__ret)) |
301 | for (;;) { \ | ||
302 | prepare_to_wait(&wq, &__wait, TASK_INTERRUPTIBLE); \ | ||
303 | if (condition) \ | ||
304 | break; \ | ||
305 | if (!signal_pending(current)) { \ | ||
306 | ret = schedule_timeout(ret); \ | ||
307 | if (!ret) \ | ||
308 | break; \ | ||
309 | continue; \ | ||
310 | } \ | ||
311 | ret = -ERESTARTSYS; \ | ||
312 | break; \ | ||
313 | } \ | ||
314 | if (!ret && (condition)) \ | ||
315 | ret = 1; \ | ||
316 | finish_wait(&wq, &__wait); \ | ||
317 | } while (0) | ||
318 | 337 | ||
319 | /** | 338 | /** |
320 | * wait_event_interruptible_timeout - sleep until a condition gets true or a timeout elapses | 339 | * wait_event_interruptible_timeout - sleep until a condition gets true or a timeout elapses |
@@ -337,15 +356,15 @@ do { \ | |||
337 | #define wait_event_interruptible_timeout(wq, condition, timeout) \ | 356 | #define wait_event_interruptible_timeout(wq, condition, timeout) \ |
338 | ({ \ | 357 | ({ \ |
339 | long __ret = timeout; \ | 358 | long __ret = timeout; \ |
340 | if (!(condition)) \ | 359 | if (!___wait_cond_timeout(condition)) \ |
341 | __wait_event_interruptible_timeout(wq, condition, __ret); \ | 360 | __ret = __wait_event_interruptible_timeout(wq, \ |
361 | condition, timeout); \ | ||
342 | __ret; \ | 362 | __ret; \ |
343 | }) | 363 | }) |
344 | 364 | ||
345 | #define __wait_event_hrtimeout(wq, condition, timeout, state) \ | 365 | #define __wait_event_hrtimeout(wq, condition, timeout, state) \ |
346 | ({ \ | 366 | ({ \ |
347 | int __ret = 0; \ | 367 | int __ret = 0; \ |
348 | DEFINE_WAIT(__wait); \ | ||
349 | struct hrtimer_sleeper __t; \ | 368 | struct hrtimer_sleeper __t; \ |
350 | \ | 369 | \ |
351 | hrtimer_init_on_stack(&__t.timer, CLOCK_MONOTONIC, \ | 370 | hrtimer_init_on_stack(&__t.timer, CLOCK_MONOTONIC, \ |
@@ -356,25 +375,15 @@ do { \ | |||
356 | current->timer_slack_ns, \ | 375 | current->timer_slack_ns, \ |
357 | HRTIMER_MODE_REL); \ | 376 | HRTIMER_MODE_REL); \ |
358 | \ | 377 | \ |
359 | for (;;) { \ | 378 | __ret = ___wait_event(wq, condition, state, 0, 0, \ |
360 | prepare_to_wait(&wq, &__wait, state); \ | ||
361 | if (condition) \ | ||
362 | break; \ | ||
363 | if (state == TASK_INTERRUPTIBLE && \ | ||
364 | signal_pending(current)) { \ | ||
365 | __ret = -ERESTARTSYS; \ | ||
366 | break; \ | ||
367 | } \ | ||
368 | if (!__t.task) { \ | 379 | if (!__t.task) { \ |
369 | __ret = -ETIME; \ | 380 | __ret = -ETIME; \ |
370 | break; \ | 381 | break; \ |
371 | } \ | 382 | } \ |
372 | schedule(); \ | 383 | schedule()); \ |
373 | } \ | ||
374 | \ | 384 | \ |
375 | hrtimer_cancel(&__t.timer); \ | 385 | hrtimer_cancel(&__t.timer); \ |
376 | destroy_hrtimer_on_stack(&__t.timer); \ | 386 | destroy_hrtimer_on_stack(&__t.timer); \ |
377 | finish_wait(&wq, &__wait); \ | ||
378 | __ret; \ | 387 | __ret; \ |
379 | }) | 388 | }) |
380 | 389 | ||
@@ -428,33 +437,15 @@ do { \ | |||
428 | __ret; \ | 437 | __ret; \ |
429 | }) | 438 | }) |
430 | 439 | ||
431 | #define __wait_event_interruptible_exclusive(wq, condition, ret) \ | 440 | #define __wait_event_interruptible_exclusive(wq, condition) \ |
432 | do { \ | 441 | ___wait_event(wq, condition, TASK_INTERRUPTIBLE, 1, 0, \ |
433 | DEFINE_WAIT(__wait); \ | 442 | schedule()) |
434 | \ | ||
435 | for (;;) { \ | ||
436 | prepare_to_wait_exclusive(&wq, &__wait, \ | ||
437 | TASK_INTERRUPTIBLE); \ | ||
438 | if (condition) { \ | ||
439 | finish_wait(&wq, &__wait); \ | ||
440 | break; \ | ||
441 | } \ | ||
442 | if (!signal_pending(current)) { \ | ||
443 | schedule(); \ | ||
444 | continue; \ | ||
445 | } \ | ||
446 | ret = -ERESTARTSYS; \ | ||
447 | abort_exclusive_wait(&wq, &__wait, \ | ||
448 | TASK_INTERRUPTIBLE, NULL); \ | ||
449 | break; \ | ||
450 | } \ | ||
451 | } while (0) | ||
452 | 443 | ||
453 | #define wait_event_interruptible_exclusive(wq, condition) \ | 444 | #define wait_event_interruptible_exclusive(wq, condition) \ |
454 | ({ \ | 445 | ({ \ |
455 | int __ret = 0; \ | 446 | int __ret = 0; \ |
456 | if (!(condition)) \ | 447 | if (!(condition)) \ |
457 | __wait_event_interruptible_exclusive(wq, condition, __ret);\ | 448 | __ret = __wait_event_interruptible_exclusive(wq, condition);\ |
458 | __ret; \ | 449 | __ret; \ |
459 | }) | 450 | }) |
460 | 451 | ||
@@ -606,24 +597,8 @@ do { \ | |||
606 | ? 0 : __wait_event_interruptible_locked(wq, condition, 1, 1)) | 597 | ? 0 : __wait_event_interruptible_locked(wq, condition, 1, 1)) |
607 | 598 | ||
608 | 599 | ||
609 | 600 | #define __wait_event_killable(wq, condition) \ | |
610 | #define __wait_event_killable(wq, condition, ret) \ | 601 | ___wait_event(wq, condition, TASK_KILLABLE, 0, 0, schedule()) |
611 | do { \ | ||
612 | DEFINE_WAIT(__wait); \ | ||
613 | \ | ||
614 | for (;;) { \ | ||
615 | prepare_to_wait(&wq, &__wait, TASK_KILLABLE); \ | ||
616 | if (condition) \ | ||
617 | break; \ | ||
618 | if (!fatal_signal_pending(current)) { \ | ||
619 | schedule(); \ | ||
620 | continue; \ | ||
621 | } \ | ||
622 | ret = -ERESTARTSYS; \ | ||
623 | break; \ | ||
624 | } \ | ||
625 | finish_wait(&wq, &__wait); \ | ||
626 | } while (0) | ||
627 | 602 | ||
628 | /** | 603 | /** |
629 | * wait_event_killable - sleep until a condition gets true | 604 | * wait_event_killable - sleep until a condition gets true |
@@ -644,26 +619,17 @@ do { \ | |||
644 | ({ \ | 619 | ({ \ |
645 | int __ret = 0; \ | 620 | int __ret = 0; \ |
646 | if (!(condition)) \ | 621 | if (!(condition)) \ |
647 | __wait_event_killable(wq, condition, __ret); \ | 622 | __ret = __wait_event_killable(wq, condition); \ |
648 | __ret; \ | 623 | __ret; \ |
649 | }) | 624 | }) |
650 | 625 | ||
651 | 626 | ||
652 | #define __wait_event_lock_irq(wq, condition, lock, cmd) \ | 627 | #define __wait_event_lock_irq(wq, condition, lock, cmd) \ |
653 | do { \ | 628 | (void)___wait_event(wq, condition, TASK_UNINTERRUPTIBLE, 0, 0, \ |
654 | DEFINE_WAIT(__wait); \ | 629 | spin_unlock_irq(&lock); \ |
655 | \ | 630 | cmd; \ |
656 | for (;;) { \ | 631 | schedule(); \ |
657 | prepare_to_wait(&wq, &__wait, TASK_UNINTERRUPTIBLE); \ | 632 | spin_lock_irq(&lock)) |
658 | if (condition) \ | ||
659 | break; \ | ||
660 | spin_unlock_irq(&lock); \ | ||
661 | cmd; \ | ||
662 | schedule(); \ | ||
663 | spin_lock_irq(&lock); \ | ||
664 | } \ | ||
665 | finish_wait(&wq, &__wait); \ | ||
666 | } while (0) | ||
667 | 633 | ||
668 | /** | 634 | /** |
669 | * wait_event_lock_irq_cmd - sleep until a condition gets true. The | 635 | * wait_event_lock_irq_cmd - sleep until a condition gets true. The |
@@ -723,26 +689,12 @@ do { \ | |||
723 | } while (0) | 689 | } while (0) |
724 | 690 | ||
725 | 691 | ||
726 | #define __wait_event_interruptible_lock_irq(wq, condition, \ | 692 | #define __wait_event_interruptible_lock_irq(wq, condition, lock, cmd) \ |
727 | lock, ret, cmd) \ | 693 | ___wait_event(wq, condition, TASK_INTERRUPTIBLE, 0, 0, \ |
728 | do { \ | 694 | spin_unlock_irq(&lock); \ |
729 | DEFINE_WAIT(__wait); \ | 695 | cmd; \ |
730 | \ | 696 | schedule(); \ |
731 | for (;;) { \ | 697 | spin_lock_irq(&lock)) |
732 | prepare_to_wait(&wq, &__wait, TASK_INTERRUPTIBLE); \ | ||
733 | if (condition) \ | ||
734 | break; \ | ||
735 | if (signal_pending(current)) { \ | ||
736 | ret = -ERESTARTSYS; \ | ||
737 | break; \ | ||
738 | } \ | ||
739 | spin_unlock_irq(&lock); \ | ||
740 | cmd; \ | ||
741 | schedule(); \ | ||
742 | spin_lock_irq(&lock); \ | ||
743 | } \ | ||
744 | finish_wait(&wq, &__wait); \ | ||
745 | } while (0) | ||
746 | 698 | ||
747 | /** | 699 | /** |
748 | * wait_event_interruptible_lock_irq_cmd - sleep until a condition gets true. | 700 | * wait_event_interruptible_lock_irq_cmd - sleep until a condition gets true. |
@@ -772,10 +724,9 @@ do { \ | |||
772 | #define wait_event_interruptible_lock_irq_cmd(wq, condition, lock, cmd) \ | 724 | #define wait_event_interruptible_lock_irq_cmd(wq, condition, lock, cmd) \ |
773 | ({ \ | 725 | ({ \ |
774 | int __ret = 0; \ | 726 | int __ret = 0; \ |
775 | \ | ||
776 | if (!(condition)) \ | 727 | if (!(condition)) \ |
777 | __wait_event_interruptible_lock_irq(wq, condition, \ | 728 | __ret = __wait_event_interruptible_lock_irq(wq, \ |
778 | lock, __ret, cmd); \ | 729 | condition, lock, cmd); \ |
779 | __ret; \ | 730 | __ret; \ |
780 | }) | 731 | }) |
781 | 732 | ||
@@ -804,39 +755,24 @@ do { \ | |||
804 | #define wait_event_interruptible_lock_irq(wq, condition, lock) \ | 755 | #define wait_event_interruptible_lock_irq(wq, condition, lock) \ |
805 | ({ \ | 756 | ({ \ |
806 | int __ret = 0; \ | 757 | int __ret = 0; \ |
807 | \ | ||
808 | if (!(condition)) \ | 758 | if (!(condition)) \ |
809 | __wait_event_interruptible_lock_irq(wq, condition, \ | 759 | __ret = __wait_event_interruptible_lock_irq(wq, \ |
810 | lock, __ret, ); \ | 760 | condition, lock,); \ |
811 | __ret; \ | 761 | __ret; \ |
812 | }) | 762 | }) |
813 | 763 | ||
814 | #define __wait_event_interruptible_lock_irq_timeout(wq, condition, \ | 764 | #define __wait_event_interruptible_lock_irq_timeout(wq, condition, \ |
815 | lock, ret) \ | 765 | lock, timeout) \ |
816 | do { \ | 766 | ___wait_event(wq, ___wait_cond_timeout(condition), \ |
817 | DEFINE_WAIT(__wait); \ | 767 | TASK_INTERRUPTIBLE, 0, timeout, \ |
818 | \ | 768 | spin_unlock_irq(&lock); \ |
819 | for (;;) { \ | 769 | __ret = schedule_timeout(__ret); \ |
820 | prepare_to_wait(&wq, &__wait, TASK_INTERRUPTIBLE); \ | 770 | spin_lock_irq(&lock)); |
821 | if (condition) \ | ||
822 | break; \ | ||
823 | if (signal_pending(current)) { \ | ||
824 | ret = -ERESTARTSYS; \ | ||
825 | break; \ | ||
826 | } \ | ||
827 | spin_unlock_irq(&lock); \ | ||
828 | ret = schedule_timeout(ret); \ | ||
829 | spin_lock_irq(&lock); \ | ||
830 | if (!ret) \ | ||
831 | break; \ | ||
832 | } \ | ||
833 | finish_wait(&wq, &__wait); \ | ||
834 | } while (0) | ||
835 | 771 | ||
836 | /** | 772 | /** |
837 | * wait_event_interruptible_lock_irq_timeout - sleep until a condition gets true or a timeout elapses. | 773 | * wait_event_interruptible_lock_irq_timeout - sleep until a condition gets |
838 | * The condition is checked under the lock. This is expected | 774 | * true or a timeout elapses. The condition is checked under |
839 | * to be called with the lock taken. | 775 | * the lock. This is expected to be called with the lock taken. |
840 | * @wq: the waitqueue to wait on | 776 | * @wq: the waitqueue to wait on |
841 | * @condition: a C expression for the event to wait for | 777 | * @condition: a C expression for the event to wait for |
842 | * @lock: a locked spinlock_t, which will be released before schedule() | 778 | * @lock: a locked spinlock_t, which will be released before schedule() |
@@ -860,11 +796,10 @@ do { \ | |||
860 | #define wait_event_interruptible_lock_irq_timeout(wq, condition, lock, \ | 796 | #define wait_event_interruptible_lock_irq_timeout(wq, condition, lock, \ |
861 | timeout) \ | 797 | timeout) \ |
862 | ({ \ | 798 | ({ \ |
863 | int __ret = timeout; \ | 799 | long __ret = timeout; \ |
864 | \ | 800 | if (!___wait_cond_timeout(condition)) \ |
865 | if (!(condition)) \ | 801 | __ret = __wait_event_interruptible_lock_irq_timeout( \ |
866 | __wait_event_interruptible_lock_irq_timeout( \ | 802 | wq, condition, lock, timeout); \ |
867 | wq, condition, lock, __ret); \ | ||
868 | __ret; \ | 803 | __ret; \ |
869 | }) | 804 | }) |
870 | 805 | ||
@@ -875,20 +810,18 @@ do { \ | |||
875 | * We plan to remove these interfaces. | 810 | * We plan to remove these interfaces. |
876 | */ | 811 | */ |
877 | extern void sleep_on(wait_queue_head_t *q); | 812 | extern void sleep_on(wait_queue_head_t *q); |
878 | extern long sleep_on_timeout(wait_queue_head_t *q, | 813 | extern long sleep_on_timeout(wait_queue_head_t *q, signed long timeout); |
879 | signed long timeout); | ||
880 | extern void interruptible_sleep_on(wait_queue_head_t *q); | 814 | extern void interruptible_sleep_on(wait_queue_head_t *q); |
881 | extern long interruptible_sleep_on_timeout(wait_queue_head_t *q, | 815 | extern long interruptible_sleep_on_timeout(wait_queue_head_t *q, signed long timeout); |
882 | signed long timeout); | ||
883 | 816 | ||
884 | /* | 817 | /* |
885 | * Waitqueues which are removed from the waitqueue_head at wakeup time | 818 | * Waitqueues which are removed from the waitqueue_head at wakeup time |
886 | */ | 819 | */ |
887 | void prepare_to_wait(wait_queue_head_t *q, wait_queue_t *wait, int state); | 820 | void prepare_to_wait(wait_queue_head_t *q, wait_queue_t *wait, int state); |
888 | void prepare_to_wait_exclusive(wait_queue_head_t *q, wait_queue_t *wait, int state); | 821 | void prepare_to_wait_exclusive(wait_queue_head_t *q, wait_queue_t *wait, int state); |
822 | long prepare_to_wait_event(wait_queue_head_t *q, wait_queue_t *wait, int state); | ||
889 | void finish_wait(wait_queue_head_t *q, wait_queue_t *wait); | 823 | void finish_wait(wait_queue_head_t *q, wait_queue_t *wait); |
890 | void abort_exclusive_wait(wait_queue_head_t *q, wait_queue_t *wait, | 824 | void abort_exclusive_wait(wait_queue_head_t *q, wait_queue_t *wait, unsigned int mode, void *key); |
891 | unsigned int mode, void *key); | ||
892 | int autoremove_wake_function(wait_queue_t *wait, unsigned mode, int sync, void *key); | 825 | int autoremove_wake_function(wait_queue_t *wait, unsigned mode, int sync, void *key); |
893 | int wake_bit_function(wait_queue_t *wait, unsigned mode, int sync, void *key); | 826 | int wake_bit_function(wait_queue_t *wait, unsigned mode, int sync, void *key); |
894 | 827 | ||
@@ -934,8 +867,8 @@ int wake_bit_function(wait_queue_t *wait, unsigned mode, int sync, void *key); | |||
934 | * One uses wait_on_bit() where one is waiting for the bit to clear, | 867 | * One uses wait_on_bit() where one is waiting for the bit to clear, |
935 | * but has no intention of setting it. | 868 | * but has no intention of setting it. |
936 | */ | 869 | */ |
937 | static inline int wait_on_bit(void *word, int bit, | 870 | static inline int |
938 | int (*action)(void *), unsigned mode) | 871 | wait_on_bit(void *word, int bit, int (*action)(void *), unsigned mode) |
939 | { | 872 | { |
940 | if (!test_bit(bit, word)) | 873 | if (!test_bit(bit, word)) |
941 | return 0; | 874 | return 0; |
@@ -958,8 +891,8 @@ static inline int wait_on_bit(void *word, int bit, | |||
958 | * One uses wait_on_bit_lock() where one is waiting for the bit to | 891 | * One uses wait_on_bit_lock() where one is waiting for the bit to |
959 | * clear with the intention of setting it, and when done, clearing it. | 892 | * clear with the intention of setting it, and when done, clearing it. |
960 | */ | 893 | */ |
961 | static inline int wait_on_bit_lock(void *word, int bit, | 894 | static inline int |
962 | int (*action)(void *), unsigned mode) | 895 | wait_on_bit_lock(void *word, int bit, int (*action)(void *), unsigned mode) |
963 | { | 896 | { |
964 | if (!test_and_set_bit(bit, word)) | 897 | if (!test_and_set_bit(bit, word)) |
965 | return 0; | 898 | return 0; |
@@ -983,5 +916,5 @@ int wait_on_atomic_t(atomic_t *val, int (*action)(atomic_t *), unsigned mode) | |||
983 | return 0; | 916 | return 0; |
984 | return out_of_line_wait_on_atomic_t(val, action, mode); | 917 | return out_of_line_wait_on_atomic_t(val, action, mode); |
985 | } | 918 | } |
986 | 919 | ||
987 | #endif | 920 | #endif /* _LINUX_WAIT_H */ |