diff options
Diffstat (limited to 'lib/rwsem.c')
-rw-r--r-- | lib/rwsem.c | 61 |
1 files changed, 34 insertions, 27 deletions
diff --git a/lib/rwsem.c b/lib/rwsem.c index ceba8e28807a..917fd946b495 100644 --- a/lib/rwsem.c +++ b/lib/rwsem.c | |||
@@ -41,7 +41,7 @@ struct rwsem_waiter { | |||
41 | * - if we come here from up_xxxx(), then: | 41 | * - if we come here from up_xxxx(), then: |
42 | * - the 'active part' of count (&0x0000ffff) reached 0 (but may have changed) | 42 | * - the 'active part' of count (&0x0000ffff) reached 0 (but may have changed) |
43 | * - the 'waiting part' of count (&0xffff0000) is -ve (and will still be so) | 43 | * - the 'waiting part' of count (&0xffff0000) is -ve (and will still be so) |
44 | * - there must be someone on the queue | 44 | * - there must be someone on the queue |
45 | * - the spinlock must be held by the caller | 45 | * - the spinlock must be held by the caller |
46 | * - woken process blocks are discarded from the list after having task zeroed | 46 | * - woken process blocks are discarded from the list after having task zeroed |
47 | * - writers are only woken if downgrading is false | 47 | * - writers are only woken if downgrading is false |
@@ -54,26 +54,23 @@ __rwsem_do_wake(struct rw_semaphore *sem, int downgrading) | |||
54 | struct list_head *next; | 54 | struct list_head *next; |
55 | signed long oldcount, woken, loop; | 55 | signed long oldcount, woken, loop; |
56 | 56 | ||
57 | waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list); | ||
58 | if (!(waiter->flags & RWSEM_WAITING_FOR_WRITE)) | ||
59 | goto readers_only; | ||
60 | |||
57 | if (downgrading) | 61 | if (downgrading) |
58 | goto dont_wake_writers; | 62 | goto out; |
59 | 63 | ||
60 | /* if we came through an up_xxxx() call, we only only wake someone up | 64 | /* There's a writer at the front of the queue - try to grant it the |
61 | * if we can transition the active part of the count from 0 -> 1 | 65 | * write lock. However, we only wake this writer if we can transition |
66 | * the active part of the count from 0 -> 1 | ||
62 | */ | 67 | */ |
63 | try_again: | 68 | try_again_write: |
64 | oldcount = rwsem_atomic_update(RWSEM_ACTIVE_BIAS, sem) | 69 | oldcount = rwsem_atomic_update(RWSEM_ACTIVE_BIAS, sem) |
65 | - RWSEM_ACTIVE_BIAS; | 70 | - RWSEM_ACTIVE_BIAS; |
66 | if (oldcount & RWSEM_ACTIVE_MASK) | 71 | if (oldcount & RWSEM_ACTIVE_MASK) |
67 | goto undo; | 72 | /* Someone grabbed the sem already */ |
68 | 73 | goto undo_write; | |
69 | waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list); | ||
70 | |||
71 | /* try to grant a single write lock if there's a writer at the front | ||
72 | * of the queue - note we leave the 'active part' of the count | ||
73 | * incremented by 1 and the waiting part incremented by 0x00010000 | ||
74 | */ | ||
75 | if (!(waiter->flags & RWSEM_WAITING_FOR_WRITE)) | ||
76 | goto readers_only; | ||
77 | 74 | ||
78 | /* We must be careful not to touch 'waiter' after we set ->task = NULL. | 75 | /* We must be careful not to touch 'waiter' after we set ->task = NULL. |
79 | * It is an allocated on the waiter's stack and may become invalid at | 76 | * It is an allocated on the waiter's stack and may become invalid at |
@@ -87,18 +84,24 @@ __rwsem_do_wake(struct rw_semaphore *sem, int downgrading) | |||
87 | put_task_struct(tsk); | 84 | put_task_struct(tsk); |
88 | goto out; | 85 | goto out; |
89 | 86 | ||
90 | /* don't want to wake any writers */ | 87 | readers_only: |
91 | dont_wake_writers: | 88 | if (downgrading) |
92 | waiter = list_entry(sem->wait_list.next, struct rwsem_waiter, list); | 89 | goto wake_readers; |
93 | if (waiter->flags & RWSEM_WAITING_FOR_WRITE) | 90 | |
94 | goto out; | 91 | /* if we came through an up_xxxx() call, we only only wake someone up |
92 | * if we can transition the active part of the count from 0 -> 1 */ | ||
93 | try_again_read: | ||
94 | oldcount = rwsem_atomic_update(RWSEM_ACTIVE_BIAS, sem) | ||
95 | - RWSEM_ACTIVE_BIAS; | ||
96 | if (oldcount & RWSEM_ACTIVE_MASK) | ||
97 | /* Someone grabbed the sem already */ | ||
98 | goto undo_read; | ||
95 | 99 | ||
96 | /* grant an infinite number of read locks to the readers at the front | 100 | wake_readers: |
97 | * of the queue | 101 | /* Grant an infinite number of read locks to the readers at the front |
98 | * - note we increment the 'active part' of the count by the number of | 102 | * of the queue. Note we increment the 'active part' of the count by |
99 | * readers before waking any processes up | 103 | * the number of readers before waking any processes up. |
100 | */ | 104 | */ |
101 | readers_only: | ||
102 | woken = 0; | 105 | woken = 0; |
103 | do { | 106 | do { |
104 | woken++; | 107 | woken++; |
@@ -138,10 +141,14 @@ __rwsem_do_wake(struct rw_semaphore *sem, int downgrading) | |||
138 | 141 | ||
139 | /* undo the change to the active count, but check for a transition | 142 | /* undo the change to the active count, but check for a transition |
140 | * 1->0 */ | 143 | * 1->0 */ |
141 | undo: | 144 | undo_write: |
145 | if (rwsem_atomic_update(-RWSEM_ACTIVE_BIAS, sem) & RWSEM_ACTIVE_MASK) | ||
146 | goto out; | ||
147 | goto try_again_write; | ||
148 | undo_read: | ||
142 | if (rwsem_atomic_update(-RWSEM_ACTIVE_BIAS, sem) & RWSEM_ACTIVE_MASK) | 149 | if (rwsem_atomic_update(-RWSEM_ACTIVE_BIAS, sem) & RWSEM_ACTIVE_MASK) |
143 | goto out; | 150 | goto out; |
144 | goto try_again; | 151 | goto try_again_read; |
145 | } | 152 | } |
146 | 153 | ||
147 | /* | 154 | /* |