diff options
Diffstat (limited to 'litmus/prioq_lock.c')
-rw-r--r-- | litmus/prioq_lock.c | 1162 |
1 files changed, 1162 insertions, 0 deletions
diff --git a/litmus/prioq_lock.c b/litmus/prioq_lock.c new file mode 100644 index 000000000000..0091e4c1901e --- /dev/null +++ b/litmus/prioq_lock.c | |||
@@ -0,0 +1,1162 @@ | |||
1 | #include <linux/slab.h> | ||
2 | #include <linux/uaccess.h> | ||
3 | |||
4 | #include <litmus/trace.h> | ||
5 | #include <litmus/sched_plugin.h> | ||
6 | #include <litmus/prioq_lock.h> | ||
7 | |||
8 | #include <litmus/litmus_proc.h> | ||
9 | |||
10 | |||
11 | #if defined(CONFIG_LITMUS_AFFINITY_LOCKING) && defined(CONFIG_LITMUS_NVIDIA) | ||
12 | #include <litmus/gpu_affinity.h> | ||
13 | #endif | ||
14 | |||
15 | static void __attribute__((unused)) | ||
16 | __dump_lock_info(struct prioq_mutex *mutex) | ||
17 | { | ||
18 | #ifdef CONFIG_SCHED_DEBUG_TRACE | ||
19 | TRACE_CUR("%s (mutex: %p):\n", mutex->litmus_lock.name, mutex); | ||
20 | TRACE_CUR("owner: %s/%d (inh: %s/%d)\n", | ||
21 | (mutex->owner) ? | ||
22 | mutex->owner->comm : "null", | ||
23 | (mutex->owner) ? | ||
24 | mutex->owner->pid : 0, | ||
25 | (mutex->owner && tsk_rt(mutex->owner)->inh_task) ? | ||
26 | tsk_rt(mutex->owner)->inh_task->comm : "null", | ||
27 | (mutex->owner && tsk_rt(mutex->owner)->inh_task) ? | ||
28 | tsk_rt(mutex->owner)->inh_task->pid : 0); | ||
29 | TRACE_CUR("hp waiter: %s/%d (inh: %s/%d)\n", | ||
30 | (mutex->hp_waiter) ? | ||
31 | mutex->hp_waiter->comm : "null", | ||
32 | (mutex->hp_waiter) ? | ||
33 | mutex->hp_waiter->pid : 0, | ||
34 | (mutex->hp_waiter && tsk_rt(mutex->hp_waiter)->inh_task) ? | ||
35 | tsk_rt(mutex->hp_waiter)->inh_task->comm : "null", | ||
36 | (mutex->hp_waiter && tsk_rt(mutex->hp_waiter)->inh_task) ? | ||
37 | tsk_rt(mutex->hp_waiter)->inh_task->pid : 0); | ||
38 | TRACE_CUR("blocked tasks, front to back:\n"); | ||
39 | if (waitqueue_active(&mutex->wait)) { | ||
40 | wait_queue_t *q; | ||
41 | struct list_head *pos; | ||
42 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
43 | dgl_wait_state_t *dgl_wait = NULL; | ||
44 | #endif | ||
45 | list_for_each(pos, &mutex->wait.task_list) { | ||
46 | struct task_struct *blocked_task; | ||
47 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
48 | int enabled = 1; | ||
49 | #endif | ||
50 | q = list_entry(pos, wait_queue_t, task_list); | ||
51 | |||
52 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
53 | if(q->func == dgl_wake_up) { | ||
54 | dgl_wait = (dgl_wait_state_t*) q->private; | ||
55 | blocked_task = dgl_wait->task; | ||
56 | |||
57 | if(tsk_rt(blocked_task)->blocked_lock != &mutex->litmus_lock) | ||
58 | enabled = 0; | ||
59 | } | ||
60 | else { | ||
61 | blocked_task = (struct task_struct*) q->private; | ||
62 | } | ||
63 | #else | ||
64 | blocked_task = (struct task_struct*) q->private; | ||
65 | #endif | ||
66 | TRACE_CUR("\t%s/%d (inh: %s/%d)" | ||
67 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
68 | " DGL enabled: %d" | ||
69 | #endif | ||
70 | "\n", | ||
71 | blocked_task->comm, blocked_task->pid, | ||
72 | (tsk_rt(blocked_task)->inh_task) ? | ||
73 | tsk_rt(blocked_task)->inh_task->comm : "null", | ||
74 | (tsk_rt(blocked_task)->inh_task) ? | ||
75 | tsk_rt(blocked_task)->inh_task->pid : 0 | ||
76 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
77 | , enabled | ||
78 | #endif | ||
79 | ); | ||
80 | } | ||
81 | } | ||
82 | else { | ||
83 | TRACE_CUR("\t<NONE>\n"); | ||
84 | } | ||
85 | #endif | ||
86 | } | ||
87 | |||
88 | static void __add_wait_queue_sorted(wait_queue_head_t *q, wait_queue_t *add_node) | ||
89 | { | ||
90 | struct list_head *pq = &(q->task_list); | ||
91 | wait_queue_t *q_node; | ||
92 | struct task_struct *queued_task; | ||
93 | struct task_struct *add_task; | ||
94 | struct list_head *pos; | ||
95 | |||
96 | if (list_empty(pq)) { | ||
97 | list_add_tail(&add_node->task_list, pq); | ||
98 | return; | ||
99 | } | ||
100 | |||
101 | add_task = get_queued_task(add_node); | ||
102 | |||
103 | /* less priority than tail? if so, go to tail */ | ||
104 | q_node = list_entry(pq->prev, wait_queue_t, task_list); | ||
105 | queued_task = get_queued_task(q_node); | ||
106 | if (litmus->compare(queued_task, add_task)) { | ||
107 | list_add_tail(&add_node->task_list, pq); | ||
108 | return; | ||
109 | } | ||
110 | |||
111 | /* belongs at head or between nodes */ | ||
112 | list_for_each(pos, pq) { | ||
113 | q_node = list_entry(pos, wait_queue_t, task_list); | ||
114 | queued_task = get_queued_task(q_node); | ||
115 | if(litmus->compare(add_task, queued_task)) { | ||
116 | list_add(&add_node->task_list, pos->prev); | ||
117 | return; | ||
118 | } | ||
119 | } | ||
120 | |||
121 | BUG(); | ||
122 | } | ||
123 | |||
124 | static inline void __add_wait_queue_sorted_exclusive(wait_queue_head_t *q, wait_queue_t *wait) | ||
125 | { | ||
126 | wait->flags |= WQ_FLAG_EXCLUSIVE; | ||
127 | __add_wait_queue_sorted(q, wait); | ||
128 | } | ||
129 | |||
130 | static void __prioq_increase_pos(struct prioq_mutex *mutex, struct task_struct *t) | ||
131 | { | ||
132 | wait_queue_t *q; | ||
133 | struct list_head *pos; | ||
134 | struct task_struct *queued; | ||
135 | |||
136 | /* TODO: Make this efficient instead of remove/add */ | ||
137 | list_for_each(pos, &mutex->wait.task_list) { | ||
138 | q = list_entry(pos, wait_queue_t, task_list); | ||
139 | queued = get_queued_task(q); | ||
140 | if (queued == t) { | ||
141 | __remove_wait_queue(&mutex->wait, q); | ||
142 | __add_wait_queue_sorted(&mutex->wait, q); | ||
143 | return; | ||
144 | } | ||
145 | } | ||
146 | |||
147 | BUG(); | ||
148 | } | ||
149 | |||
150 | static void __prioq_decrease_pos(struct prioq_mutex *mutex, struct task_struct *t) | ||
151 | { | ||
152 | wait_queue_t *q; | ||
153 | struct list_head *pos; | ||
154 | struct task_struct *queued; | ||
155 | |||
156 | /* TODO: Make this efficient instead of remove/add */ | ||
157 | list_for_each(pos, &mutex->wait.task_list) { | ||
158 | q = list_entry(pos, wait_queue_t, task_list); | ||
159 | queued = get_queued_task(q); | ||
160 | if (queued == t) { | ||
161 | __remove_wait_queue(&mutex->wait, q); | ||
162 | __add_wait_queue_sorted(&mutex->wait, q); | ||
163 | return; | ||
164 | } | ||
165 | } | ||
166 | |||
167 | BUG(); | ||
168 | } | ||
169 | |||
170 | |||
171 | /* caller is responsible for locking */ | ||
172 | static struct task_struct* __prioq_mutex_find_hp_waiter(struct prioq_mutex *mutex, | ||
173 | struct task_struct* skip) | ||
174 | { | ||
175 | wait_queue_t *q; | ||
176 | struct list_head *pos; | ||
177 | struct task_struct *queued = NULL, *found = NULL; | ||
178 | |||
179 | /* list in sorted order. higher-prio tasks likely at the front. */ | ||
180 | list_for_each(pos, &mutex->wait.task_list) { | ||
181 | q = list_entry(pos, wait_queue_t, task_list); | ||
182 | queued = get_queued_task(q); | ||
183 | |||
184 | /* Compare task prios, find high prio task. */ | ||
185 | if (queued && queued != skip && litmus->compare(queued, found)) { | ||
186 | found = queued; | ||
187 | } | ||
188 | } | ||
189 | return found; | ||
190 | } | ||
191 | |||
192 | |||
193 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
194 | |||
195 | int prioq_mutex_is_owner(struct litmus_lock *l, struct task_struct *t) | ||
196 | { | ||
197 | struct prioq_mutex *mutex = prioq_mutex_from_lock(l); | ||
198 | return(mutex->owner == t); | ||
199 | } | ||
200 | |||
201 | // return 1 if resource was immediatly acquired. | ||
202 | // Assumes mutex->lock is held. | ||
203 | // Must set task state to TASK_UNINTERRUPTIBLE if task blocks. | ||
204 | int prioq_mutex_dgl_lock(struct litmus_lock *l, dgl_wait_state_t* dgl_wait, | ||
205 | wait_queue_t* wq_node) | ||
206 | { | ||
207 | struct prioq_mutex *mutex = prioq_mutex_from_lock(l); | ||
208 | struct task_struct *t = dgl_wait->task; | ||
209 | |||
210 | int acquired_immediatly = 0; | ||
211 | |||
212 | BUG_ON(t != current); | ||
213 | |||
214 | if (mutex->owner) { | ||
215 | TRACE_TASK(t, "Enqueuing on lock %d (held by %s/%d).\n", | ||
216 | l->ident, mutex->owner->comm, mutex->owner->pid); | ||
217 | |||
218 | init_dgl_waitqueue_entry(wq_node, dgl_wait); | ||
219 | |||
220 | set_task_state(t, TASK_UNINTERRUPTIBLE); | ||
221 | __add_wait_queue_sorted_exclusive(&mutex->wait, wq_node); | ||
222 | } else { | ||
223 | TRACE_TASK(t, "Acquired lock %d with no blocking.\n", l->ident); | ||
224 | |||
225 | /* it's ours now */ | ||
226 | mutex->owner = t; | ||
227 | |||
228 | raw_spin_lock(&tsk_rt(t)->hp_blocked_tasks_lock); | ||
229 | binheap_add(&l->nest.hp_binheap_node, &tsk_rt(t)->hp_blocked_tasks, | ||
230 | struct nested_info, hp_binheap_node); | ||
231 | raw_spin_unlock(&tsk_rt(t)->hp_blocked_tasks_lock); | ||
232 | |||
233 | acquired_immediatly = 1; | ||
234 | } | ||
235 | |||
236 | return acquired_immediatly; | ||
237 | } | ||
238 | |||
239 | void prioq_mutex_enable_priority(struct litmus_lock *l, | ||
240 | dgl_wait_state_t* dgl_wait) | ||
241 | { | ||
242 | struct prioq_mutex *mutex = prioq_mutex_from_lock(l); | ||
243 | struct task_struct *t = dgl_wait->task; | ||
244 | struct task_struct *owner = mutex->owner; | ||
245 | unsigned long flags = 0; // these are unused under DGL coarse-grain locking | ||
246 | |||
247 | BUG_ON(owner == t); | ||
248 | |||
249 | tsk_rt(t)->blocked_lock = l; | ||
250 | mb(); | ||
251 | |||
252 | if (litmus->compare(t, mutex->hp_waiter)) { | ||
253 | struct task_struct *old_max_eff_prio; | ||
254 | struct task_struct *new_max_eff_prio; | ||
255 | struct task_struct *new_prio = NULL; | ||
256 | |||
257 | if(mutex->hp_waiter) | ||
258 | TRACE_TASK(t, "has higher prio than hp_waiter (%s/%d).\n", | ||
259 | mutex->hp_waiter->comm, mutex->hp_waiter->pid); | ||
260 | else | ||
261 | TRACE_TASK(t, "has higher prio than hp_waiter (NIL).\n"); | ||
262 | |||
263 | raw_spin_lock(&tsk_rt(owner)->hp_blocked_tasks_lock); | ||
264 | |||
265 | old_max_eff_prio = top_priority(&tsk_rt(owner)->hp_blocked_tasks); | ||
266 | mutex->hp_waiter = t; | ||
267 | l->nest.hp_waiter_eff_prio = effective_priority(mutex->hp_waiter); | ||
268 | binheap_decrease(&l->nest.hp_binheap_node, | ||
269 | &tsk_rt(owner)->hp_blocked_tasks); | ||
270 | new_max_eff_prio = top_priority(&tsk_rt(owner)->hp_blocked_tasks); | ||
271 | |||
272 | if(new_max_eff_prio != old_max_eff_prio) { | ||
273 | TRACE_TASK(t, "is new hp_waiter.\n"); | ||
274 | |||
275 | if ((effective_priority(owner) == old_max_eff_prio) || | ||
276 | (litmus->__compare(new_max_eff_prio, BASE, owner, EFFECTIVE))){ | ||
277 | new_prio = new_max_eff_prio; | ||
278 | } | ||
279 | } | ||
280 | else { | ||
281 | TRACE_TASK(t, "no change in max_eff_prio of heap.\n"); | ||
282 | } | ||
283 | |||
284 | if(new_prio) { | ||
285 | litmus->nested_increase_prio(owner, new_prio, | ||
286 | &mutex->lock, flags); // unlocks lock. | ||
287 | } | ||
288 | else { | ||
289 | raw_spin_unlock(&tsk_rt(owner)->hp_blocked_tasks_lock); | ||
290 | unlock_fine_irqrestore(&mutex->lock, flags); | ||
291 | } | ||
292 | } | ||
293 | else { | ||
294 | TRACE_TASK(t, "no change in hp_waiter.\n"); | ||
295 | unlock_fine_irqrestore(&mutex->lock, flags); | ||
296 | } | ||
297 | } | ||
298 | |||
299 | static void select_next_lock_if_primary(struct litmus_lock *l, | ||
300 | dgl_wait_state_t *dgl_wait) | ||
301 | { | ||
302 | if(tsk_rt(dgl_wait->task)->blocked_lock == l) { | ||
303 | TRACE_CUR("Lock %d in DGL was primary for %s/%d.\n", | ||
304 | l->ident, dgl_wait->task->comm, dgl_wait->task->pid); | ||
305 | tsk_rt(dgl_wait->task)->blocked_lock = NULL; | ||
306 | mb(); | ||
307 | select_next_lock(dgl_wait /*, l*/); // pick the next lock to be blocked on | ||
308 | } | ||
309 | else { | ||
310 | TRACE_CUR("Got lock early! Lock %d in DGL was NOT primary for %s/%d.\n", | ||
311 | l->ident, dgl_wait->task->comm, dgl_wait->task->pid); | ||
312 | } | ||
313 | } | ||
314 | #endif | ||
315 | |||
316 | |||
317 | |||
318 | |||
319 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
320 | |||
321 | int prioq_mutex_dgl_can_quick_lock(struct litmus_lock *l, struct task_struct *t) | ||
322 | { | ||
323 | struct prioq_mutex *mutex = prioq_mutex_from_lock(l); | ||
324 | |||
325 | if(!mutex->owner && mutex->hp_waiter == t) { | ||
326 | wait_queue_t *front = list_entry(mutex->wait.task_list.next, wait_queue_t, task_list); | ||
327 | struct task_struct *at_front = get_queued_task(front); | ||
328 | if(t == at_front) { | ||
329 | return 1; | ||
330 | } | ||
331 | } | ||
332 | return 0; | ||
333 | } | ||
334 | |||
335 | void prioq_mutex_dgl_quick_lock(struct litmus_lock *l, struct litmus_lock *cur_lock, | ||
336 | struct task_struct* t, wait_queue_t *q) | ||
337 | { | ||
338 | struct prioq_mutex *mutex = prioq_mutex_from_lock(l); | ||
339 | |||
340 | BUG_ON(mutex->owner); | ||
341 | BUG_ON(mutex->hp_waiter != t); | ||
342 | BUG_ON(t != get_queued_task(list_entry(mutex->wait.task_list.next, wait_queue_t, task_list))); | ||
343 | |||
344 | |||
345 | mutex->owner = t; | ||
346 | |||
347 | if (l != cur_lock) { | ||
348 | /* we have to update the state of the other lock for it */ | ||
349 | __remove_wait_queue(&mutex->wait, q); | ||
350 | |||
351 | mutex->hp_waiter = __prioq_mutex_find_hp_waiter(mutex, t); | ||
352 | l->nest.hp_waiter_eff_prio = (mutex->hp_waiter) ? | ||
353 | effective_priority(mutex->hp_waiter) : | ||
354 | NULL; | ||
355 | |||
356 | if (mutex->hp_waiter) | ||
357 | TRACE_TASK(mutex->hp_waiter, "is new highest-prio waiter\n"); | ||
358 | else | ||
359 | TRACE("no further waiters\n"); | ||
360 | |||
361 | raw_spin_lock(&tsk_rt(t)->hp_blocked_tasks_lock); | ||
362 | |||
363 | binheap_add(&l->nest.hp_binheap_node, | ||
364 | &tsk_rt(t)->hp_blocked_tasks, | ||
365 | struct nested_info, hp_binheap_node); | ||
366 | |||
367 | raw_spin_unlock(&tsk_rt(t)->hp_blocked_tasks_lock); | ||
368 | } | ||
369 | } | ||
370 | #endif | ||
371 | |||
372 | |||
373 | int prioq_mutex_lock(struct litmus_lock* l) | ||
374 | { | ||
375 | struct task_struct *t = current; | ||
376 | struct task_struct *owner; | ||
377 | struct prioq_mutex *mutex = prioq_mutex_from_lock(l); | ||
378 | wait_queue_t wait; | ||
379 | unsigned long flags; | ||
380 | |||
381 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
382 | raw_spinlock_t *dgl_lock; | ||
383 | #endif | ||
384 | |||
385 | if (!is_realtime(t)) | ||
386 | return -EPERM; | ||
387 | |||
388 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
389 | dgl_lock = litmus->get_dgl_spinlock(t); | ||
390 | #endif | ||
391 | |||
392 | lock_global_irqsave(dgl_lock, flags); | ||
393 | lock_fine_irqsave(&mutex->lock, flags); | ||
394 | |||
395 | /* block if there is an owner, or if hp_waiter is blocked for DGL and | ||
396 | * prio(t) < prio(hp_waiter) */ | ||
397 | if (mutex->owner) { | ||
398 | TRACE_TASK(t, "Blocking on lock %d (held by %s/%d).\n", | ||
399 | l->ident, mutex->owner->comm, mutex->owner->pid); | ||
400 | |||
401 | #if defined(CONFIG_LITMUS_AFFINITY_LOCKING) && defined(CONFIG_LITMUS_NVIDIA) | ||
402 | // KLUDGE: don't count this suspension as time in the critical gpu | ||
403 | // critical section | ||
404 | if(tsk_rt(t)->held_gpus) { | ||
405 | tsk_rt(t)->suspend_gpu_tracker_on_block = 1; | ||
406 | } | ||
407 | #endif | ||
408 | |||
409 | /* resource is not free => must suspend and wait */ | ||
410 | |||
411 | owner = mutex->owner; | ||
412 | |||
413 | init_waitqueue_entry(&wait, t); | ||
414 | |||
415 | tsk_rt(t)->blocked_lock = l; /* record where we are blocked */ | ||
416 | mb(); // needed? | ||
417 | |||
418 | /* FIXME: interruptible would be nice some day */ | ||
419 | set_task_state(t, TASK_UNINTERRUPTIBLE); | ||
420 | |||
421 | __add_wait_queue_sorted_exclusive(&mutex->wait, &wait); | ||
422 | |||
423 | /* check if we need to activate priority inheritance */ | ||
424 | if (litmus->compare(t, mutex->hp_waiter)) { | ||
425 | |||
426 | struct task_struct *old_max_eff_prio; | ||
427 | struct task_struct *new_max_eff_prio; | ||
428 | struct task_struct *new_prio = NULL; | ||
429 | |||
430 | if(mutex->hp_waiter) | ||
431 | TRACE_TASK(t, "has higher prio than hp_waiter (%s/%d).\n", | ||
432 | mutex->hp_waiter->comm, mutex->hp_waiter->pid); | ||
433 | else | ||
434 | TRACE_TASK(t, "has higher prio than hp_waiter (NIL).\n"); | ||
435 | |||
436 | raw_spin_lock(&tsk_rt(owner)->hp_blocked_tasks_lock); | ||
437 | |||
438 | old_max_eff_prio = top_priority(&tsk_rt(owner)->hp_blocked_tasks); | ||
439 | mutex->hp_waiter = t; | ||
440 | |||
441 | TRACE_TASK(t, "prioq_mutex %d state after enqeue in priority queue\n", l->ident); | ||
442 | __dump_lock_info(mutex); | ||
443 | |||
444 | l->nest.hp_waiter_eff_prio = effective_priority(mutex->hp_waiter); | ||
445 | binheap_decrease(&l->nest.hp_binheap_node, | ||
446 | &tsk_rt(owner)->hp_blocked_tasks); | ||
447 | new_max_eff_prio = top_priority(&tsk_rt(owner)->hp_blocked_tasks); | ||
448 | |||
449 | if(new_max_eff_prio != old_max_eff_prio) { | ||
450 | TRACE_TASK(t, "is new hp_waiter.\n"); | ||
451 | |||
452 | if ((effective_priority(owner) == old_max_eff_prio) || | ||
453 | (litmus->__compare(new_max_eff_prio, BASE, owner, EFFECTIVE))){ | ||
454 | new_prio = new_max_eff_prio; | ||
455 | } | ||
456 | } | ||
457 | else { | ||
458 | TRACE_TASK(t, "no change in max_eff_prio of heap.\n"); | ||
459 | } | ||
460 | |||
461 | if(new_prio) { | ||
462 | litmus->nested_increase_prio(owner, new_prio, &mutex->lock, | ||
463 | flags); // unlocks lock. | ||
464 | } | ||
465 | else { | ||
466 | raw_spin_unlock(&tsk_rt(owner)->hp_blocked_tasks_lock); | ||
467 | unlock_fine_irqrestore(&mutex->lock, flags); | ||
468 | } | ||
469 | } | ||
470 | else { | ||
471 | TRACE_TASK(t, "no change in hp_waiter.\n"); | ||
472 | |||
473 | TRACE_TASK(t, "prioq_mutex %d state after enqeue in priority queue\n", l->ident); | ||
474 | __dump_lock_info(mutex); | ||
475 | |||
476 | unlock_fine_irqrestore(&mutex->lock, flags); | ||
477 | } | ||
478 | |||
479 | unlock_global_irqrestore(dgl_lock, flags); | ||
480 | |||
481 | TS_LOCK_SUSPEND; | ||
482 | |||
483 | /* We depend on the FIFO order. Thus, we don't need to recheck | ||
484 | * when we wake up; we are guaranteed to have the lock since | ||
485 | * there is only one wake up per release. | ||
486 | */ | ||
487 | |||
488 | suspend_for_lock(); | ||
489 | |||
490 | TS_LOCK_RESUME; | ||
491 | |||
492 | /* Since we hold the lock, no other task will change | ||
493 | * ->owner. We can thus check it without acquiring the spin | ||
494 | * lock. */ | ||
495 | BUG_ON(mutex->owner != t); | ||
496 | |||
497 | TRACE_TASK(t, "Acquired lock %d.\n", l->ident); | ||
498 | |||
499 | } else { | ||
500 | TRACE_TASK(t, "Acquired lock %d with no blocking.\n", l->ident); | ||
501 | |||
502 | /* it's ours now */ | ||
503 | mutex->owner = t; | ||
504 | |||
505 | raw_spin_lock(&tsk_rt(mutex->owner)->hp_blocked_tasks_lock); | ||
506 | binheap_add(&l->nest.hp_binheap_node, &tsk_rt(t)->hp_blocked_tasks, | ||
507 | struct nested_info, hp_binheap_node); | ||
508 | raw_spin_unlock(&tsk_rt(mutex->owner)->hp_blocked_tasks_lock); | ||
509 | |||
510 | |||
511 | unlock_fine_irqrestore(&mutex->lock, flags); | ||
512 | unlock_global_irqrestore(dgl_lock, flags); | ||
513 | } | ||
514 | |||
515 | return 0; | ||
516 | } | ||
517 | |||
518 | |||
519 | |||
520 | int prioq_mutex_unlock(struct litmus_lock* l) | ||
521 | { | ||
522 | struct task_struct *t = current, *next = NULL; | ||
523 | struct prioq_mutex *mutex = prioq_mutex_from_lock(l); | ||
524 | unsigned long flags; | ||
525 | |||
526 | struct task_struct *old_max_eff_prio; | ||
527 | |||
528 | int wake_up_task = 1; | ||
529 | |||
530 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
531 | dgl_wait_state_t *dgl_wait = NULL; | ||
532 | raw_spinlock_t *dgl_lock = litmus->get_dgl_spinlock(t); | ||
533 | #endif | ||
534 | |||
535 | int err = 0; | ||
536 | |||
537 | if (mutex->owner != t) { | ||
538 | err = -EINVAL; | ||
539 | return err; | ||
540 | } | ||
541 | |||
542 | lock_global_irqsave(dgl_lock, flags); | ||
543 | lock_fine_irqsave(&mutex->lock, flags); | ||
544 | |||
545 | raw_spin_lock(&tsk_rt(t)->hp_blocked_tasks_lock); | ||
546 | |||
547 | TRACE_TASK(t, "Freeing lock %d\n", l->ident); | ||
548 | |||
549 | old_max_eff_prio = top_priority(&tsk_rt(t)->hp_blocked_tasks); | ||
550 | binheap_delete(&l->nest.hp_binheap_node, &tsk_rt(t)->hp_blocked_tasks); | ||
551 | |||
552 | if(tsk_rt(t)->inh_task){ | ||
553 | struct task_struct *new_max_eff_prio = | ||
554 | top_priority(&tsk_rt(t)->hp_blocked_tasks); | ||
555 | |||
556 | if((new_max_eff_prio == NULL) || | ||
557 | /* there was a change in eff prio */ | ||
558 | ( (new_max_eff_prio != old_max_eff_prio) && | ||
559 | /* and owner had the old eff prio */ | ||
560 | (effective_priority(t) == old_max_eff_prio)) ) | ||
561 | { | ||
562 | // old_max_eff_prio > new_max_eff_prio | ||
563 | |||
564 | if(litmus->__compare(new_max_eff_prio, BASE, t, EFFECTIVE)) { | ||
565 | TRACE_TASK(t, "new_max_eff_prio > task's eff_prio-- new_max_eff_prio: %s/%d task: %s/%d [%s/%d]\n", | ||
566 | new_max_eff_prio->comm, new_max_eff_prio->pid, | ||
567 | t->comm, t->pid, tsk_rt(t)->inh_task->comm, | ||
568 | tsk_rt(t)->inh_task->pid); | ||
569 | WARN_ON(1); | ||
570 | } | ||
571 | |||
572 | litmus->decrease_prio(t, new_max_eff_prio); | ||
573 | } | ||
574 | } | ||
575 | |||
576 | if(binheap_empty(&tsk_rt(t)->hp_blocked_tasks) && | ||
577 | tsk_rt(t)->inh_task != NULL) | ||
578 | { | ||
579 | WARN_ON(tsk_rt(t)->inh_task != NULL); | ||
580 | TRACE_TASK(t, "No more locks are held, but eff_prio = %s/%d\n", | ||
581 | tsk_rt(t)->inh_task->comm, tsk_rt(t)->inh_task->pid); | ||
582 | } | ||
583 | |||
584 | raw_spin_unlock(&tsk_rt(t)->hp_blocked_tasks_lock); | ||
585 | |||
586 | |||
587 | /* check if there are jobs waiting for this resource */ | ||
588 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
589 | __waitqueue_dgl_remove_first(&mutex->wait, &dgl_wait, &next); | ||
590 | if(dgl_wait) { | ||
591 | next = dgl_wait->task; | ||
592 | } | ||
593 | #else | ||
594 | |||
595 | next = __waitqueue_remove_first(&mutex->wait); | ||
596 | #endif | ||
597 | if (next) { | ||
598 | /* next becomes the resouce holder */ | ||
599 | mutex->owner = next; | ||
600 | TRACE_CUR("lock %d ownership passed to %s/%d\n", l->ident, next->comm, next->pid); | ||
601 | |||
602 | /* determine new hp_waiter if necessary */ | ||
603 | if (next == mutex->hp_waiter) { | ||
604 | |||
605 | TRACE_TASK(next, "was highest-prio waiter\n"); | ||
606 | /* next has the highest priority --- it doesn't need to | ||
607 | * inherit. However, we need to make sure that the | ||
608 | * next-highest priority in the queue is reflected in | ||
609 | * hp_waiter. */ | ||
610 | mutex->hp_waiter = __prioq_mutex_find_hp_waiter(mutex, next); | ||
611 | l->nest.hp_waiter_eff_prio = (mutex->hp_waiter) ? | ||
612 | effective_priority(mutex->hp_waiter) : | ||
613 | NULL; | ||
614 | |||
615 | if (mutex->hp_waiter) | ||
616 | TRACE_TASK(mutex->hp_waiter, "is new highest-prio waiter\n"); | ||
617 | else | ||
618 | TRACE("no further waiters\n"); | ||
619 | |||
620 | raw_spin_lock(&tsk_rt(next)->hp_blocked_tasks_lock); | ||
621 | |||
622 | binheap_add(&l->nest.hp_binheap_node, | ||
623 | &tsk_rt(next)->hp_blocked_tasks, | ||
624 | struct nested_info, hp_binheap_node); | ||
625 | |||
626 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
627 | if(dgl_wait) { | ||
628 | select_next_lock_if_primary(l, dgl_wait); | ||
629 | --(dgl_wait->nr_remaining); | ||
630 | wake_up_task = (dgl_wait->nr_remaining == 0); | ||
631 | } | ||
632 | #endif | ||
633 | |||
634 | raw_spin_unlock(&tsk_rt(next)->hp_blocked_tasks_lock); | ||
635 | } | ||
636 | else { | ||
637 | /* Well, if 'next' is not the highest-priority waiter, | ||
638 | * then it (probably) ought to inherit the highest-priority | ||
639 | * waiter's priority. */ | ||
640 | TRACE_TASK(next, "is not hp_waiter of lock %d.\n", l->ident); | ||
641 | |||
642 | raw_spin_lock(&tsk_rt(next)->hp_blocked_tasks_lock); | ||
643 | |||
644 | binheap_add(&l->nest.hp_binheap_node, | ||
645 | &tsk_rt(next)->hp_blocked_tasks, | ||
646 | struct nested_info, hp_binheap_node); | ||
647 | |||
648 | /* It is possible that 'next' *should* be the hp_waiter, but isn't | ||
649 | * because that update hasn't yet executed (update operation is | ||
650 | * probably blocked on mutex->lock). So only inherit if the top of | ||
651 | * 'next's top heap node is indeed the effective prio. of hp_waiter. | ||
652 | * (We use l->hp_waiter_eff_prio instead of effective_priority(hp_waiter) | ||
653 | * since the effective priority of hp_waiter can change (and the | ||
654 | * update has not made it to this lock).) | ||
655 | */ | ||
656 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
657 | if((l->nest.hp_waiter_eff_prio != NULL) && | ||
658 | (top_priority(&tsk_rt(next)->hp_blocked_tasks) == l->nest.hp_waiter_eff_prio)) | ||
659 | { | ||
660 | if(dgl_wait && tsk_rt(next)->blocked_lock) { | ||
661 | if(litmus->__compare(l->nest.hp_waiter_eff_prio, BASE, next, EFFECTIVE)) { | ||
662 | litmus->nested_increase_prio(next, l->nest.hp_waiter_eff_prio, &mutex->lock, flags); // unlocks lock && hp_blocked_tasks_lock. | ||
663 | goto out; // all spinlocks are released. bail out now. | ||
664 | } | ||
665 | } | ||
666 | else { | ||
667 | litmus->increase_prio(next, l->nest.hp_waiter_eff_prio); | ||
668 | } | ||
669 | } | ||
670 | |||
671 | raw_spin_unlock(&tsk_rt(next)->hp_blocked_tasks_lock); | ||
672 | #else | ||
673 | if(likely(top_priority(&tsk_rt(next)->hp_blocked_tasks) == l->nest.hp_waiter_eff_prio)) | ||
674 | { | ||
675 | litmus->increase_prio(next, l->nest.hp_waiter_eff_prio); | ||
676 | } | ||
677 | raw_spin_unlock(&tsk_rt(next)->hp_blocked_tasks_lock); | ||
678 | #endif | ||
679 | } | ||
680 | |||
681 | if(wake_up_task) { | ||
682 | TRACE_TASK(next, "waking up since it is no longer blocked.\n"); | ||
683 | |||
684 | tsk_rt(next)->blocked_lock = NULL; | ||
685 | mb(); | ||
686 | |||
687 | #if defined(CONFIG_LITMUS_AFFINITY_LOCKING) && defined(CONFIG_LITMUS_NVIDIA) | ||
688 | // re-enable tracking | ||
689 | if(tsk_rt(next)->held_gpus) { | ||
690 | tsk_rt(next)->suspend_gpu_tracker_on_block = 0; | ||
691 | } | ||
692 | #endif | ||
693 | |||
694 | wake_up_process(next); | ||
695 | } | ||
696 | else { | ||
697 | TRACE_TASK(next, "is still blocked.\n"); | ||
698 | } | ||
699 | } | ||
700 | else { | ||
701 | /* becomes available */ | ||
702 | mutex->owner = NULL; | ||
703 | } | ||
704 | |||
705 | unlock_fine_irqrestore(&mutex->lock, flags); | ||
706 | |||
707 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
708 | out: | ||
709 | #endif | ||
710 | unlock_global_irqrestore(dgl_lock, flags); | ||
711 | |||
712 | TRACE_TASK(t, "-- Freed lock %d --\n", l->ident); | ||
713 | |||
714 | return err; | ||
715 | } | ||
716 | |||
717 | |||
718 | void prioq_mutex_propagate_increase_inheritance(struct litmus_lock* l, | ||
719 | struct task_struct* t, | ||
720 | raw_spinlock_t* to_unlock, | ||
721 | unsigned long irqflags) | ||
722 | { | ||
723 | struct prioq_mutex *mutex = prioq_mutex_from_lock(l); | ||
724 | |||
725 | // relay-style locking | ||
726 | lock_fine(&mutex->lock); | ||
727 | unlock_fine(to_unlock); | ||
728 | |||
729 | __prioq_increase_pos(mutex, t); | ||
730 | |||
731 | if(tsk_rt(t)->blocked_lock == l) { // prevent race on tsk_rt(t)->blocked | ||
732 | struct task_struct *owner = mutex->owner; | ||
733 | |||
734 | struct task_struct *old_max_eff_prio; | ||
735 | struct task_struct *new_max_eff_prio; | ||
736 | |||
737 | raw_spin_lock(&tsk_rt(owner)->hp_blocked_tasks_lock); | ||
738 | |||
739 | old_max_eff_prio = top_priority(&tsk_rt(owner)->hp_blocked_tasks); | ||
740 | |||
741 | if((t != mutex->hp_waiter) && litmus->compare(t, mutex->hp_waiter)) { | ||
742 | TRACE_TASK(t, "is new highest-prio waiter by propagation.\n"); | ||
743 | mutex->hp_waiter = t; | ||
744 | |||
745 | TRACE_TASK(t, "prioq_mutex %d state after prio increase in priority queue\n", l->ident); | ||
746 | __dump_lock_info(mutex); | ||
747 | } | ||
748 | else { | ||
749 | TRACE_TASK(t, "prioq_mutex %d state after prio increase in priority queue\n", l->ident); | ||
750 | __dump_lock_info(mutex); | ||
751 | } | ||
752 | |||
753 | if(t == mutex->hp_waiter) { | ||
754 | // reflect the decreased priority in the heap node. | ||
755 | l->nest.hp_waiter_eff_prio = effective_priority(mutex->hp_waiter); | ||
756 | |||
757 | BUG_ON(!binheap_is_in_heap(&l->nest.hp_binheap_node)); | ||
758 | BUG_ON(!binheap_is_in_this_heap(&l->nest.hp_binheap_node, | ||
759 | &tsk_rt(owner)->hp_blocked_tasks)); | ||
760 | |||
761 | binheap_decrease(&l->nest.hp_binheap_node, | ||
762 | &tsk_rt(owner)->hp_blocked_tasks); | ||
763 | } | ||
764 | |||
765 | new_max_eff_prio = top_priority(&tsk_rt(owner)->hp_blocked_tasks); | ||
766 | |||
767 | |||
768 | if(new_max_eff_prio != old_max_eff_prio) { | ||
769 | // new_max_eff_prio > old_max_eff_prio holds. | ||
770 | if ((effective_priority(owner) == old_max_eff_prio) || | ||
771 | (litmus->__compare(new_max_eff_prio, BASE, owner, EFFECTIVE))) { | ||
772 | TRACE_CUR("Propagating inheritance to holder of lock %d.\n", | ||
773 | l->ident); | ||
774 | |||
775 | // beware: recursion | ||
776 | litmus->nested_increase_prio(owner, new_max_eff_prio, | ||
777 | &mutex->lock, irqflags); // unlocks mutex->lock | ||
778 | } | ||
779 | else { | ||
780 | TRACE_CUR("Lower priority than holder %s/%d. No propagation.\n", | ||
781 | owner->comm, owner->pid); | ||
782 | raw_spin_unlock(&tsk_rt(owner)->hp_blocked_tasks_lock); | ||
783 | unlock_fine_irqrestore(&mutex->lock, irqflags); | ||
784 | } | ||
785 | } | ||
786 | else { | ||
787 | TRACE_TASK(mutex->owner, "No change in maxiumum effective priority.\n"); | ||
788 | raw_spin_unlock(&tsk_rt(owner)->hp_blocked_tasks_lock); | ||
789 | unlock_fine_irqrestore(&mutex->lock, irqflags); | ||
790 | } | ||
791 | } | ||
792 | else { | ||
793 | struct litmus_lock *still_blocked; | ||
794 | |||
795 | TRACE_TASK(t, "prioq_mutex %d state after prio increase in priority queue\n", l->ident); | ||
796 | __dump_lock_info(mutex); | ||
797 | |||
798 | still_blocked = tsk_rt(t)->blocked_lock; | ||
799 | |||
800 | TRACE_TASK(t, "is not blocked on lock %d.\n", l->ident); | ||
801 | if(still_blocked) { | ||
802 | TRACE_TASK(t, "is still blocked on a lock though (lock %d).\n", | ||
803 | still_blocked->ident); | ||
804 | if(still_blocked->ops->propagate_increase_inheritance) { | ||
805 | /* due to relay-style nesting of spinlocks (acq. A, acq. B, free A, free B) | ||
806 | we know that task 't' has not released any locks behind us in this | ||
807 | chain. Propagation just needs to catch up with task 't'. */ | ||
808 | still_blocked->ops->propagate_increase_inheritance(still_blocked, | ||
809 | t, | ||
810 | &mutex->lock, | ||
811 | irqflags); | ||
812 | } | ||
813 | else { | ||
814 | TRACE_TASK(t, | ||
815 | "Inheritor is blocked on lock (%p) that does not " | ||
816 | "support nesting!\n", | ||
817 | still_blocked); | ||
818 | unlock_fine_irqrestore(&mutex->lock, irqflags); | ||
819 | } | ||
820 | } | ||
821 | else { | ||
822 | unlock_fine_irqrestore(&mutex->lock, irqflags); | ||
823 | } | ||
824 | } | ||
825 | } | ||
826 | |||
827 | |||
828 | void prioq_mutex_propagate_decrease_inheritance(struct litmus_lock* l, | ||
829 | struct task_struct* t, | ||
830 | raw_spinlock_t* to_unlock, | ||
831 | unsigned long irqflags) | ||
832 | { | ||
833 | struct prioq_mutex *mutex = prioq_mutex_from_lock(l); | ||
834 | |||
835 | // relay-style locking | ||
836 | lock_fine(&mutex->lock); | ||
837 | unlock_fine(to_unlock); | ||
838 | |||
839 | __prioq_decrease_pos(mutex, t); | ||
840 | |||
841 | if(tsk_rt(t)->blocked_lock == l) { // prevent race on tsk_rt(t)->blocked | ||
842 | if(t == mutex->hp_waiter) { | ||
843 | struct task_struct *owner = mutex->owner; | ||
844 | |||
845 | struct task_struct *old_max_eff_prio; | ||
846 | struct task_struct *new_max_eff_prio; | ||
847 | |||
848 | raw_spin_lock(&tsk_rt(owner)->hp_blocked_tasks_lock); | ||
849 | |||
850 | old_max_eff_prio = top_priority(&tsk_rt(owner)->hp_blocked_tasks); | ||
851 | |||
852 | binheap_delete(&l->nest.hp_binheap_node, &tsk_rt(owner)->hp_blocked_tasks); | ||
853 | mutex->hp_waiter = __prioq_mutex_find_hp_waiter(mutex, NULL); | ||
854 | |||
855 | TRACE_TASK(t, "prioq_mutex %d state after prio decrease in priority queue\n", l->ident); | ||
856 | __dump_lock_info(mutex); | ||
857 | |||
858 | l->nest.hp_waiter_eff_prio = (mutex->hp_waiter) ? | ||
859 | effective_priority(mutex->hp_waiter) : NULL; | ||
860 | binheap_add(&l->nest.hp_binheap_node, | ||
861 | &tsk_rt(owner)->hp_blocked_tasks, | ||
862 | struct nested_info, hp_binheap_node); | ||
863 | |||
864 | new_max_eff_prio = top_priority(&tsk_rt(owner)->hp_blocked_tasks); | ||
865 | |||
866 | if((old_max_eff_prio != new_max_eff_prio) && | ||
867 | (effective_priority(owner) == old_max_eff_prio)) | ||
868 | { | ||
869 | // Need to set new effective_priority for owner | ||
870 | |||
871 | struct task_struct *decreased_prio; | ||
872 | |||
873 | TRACE_CUR("Propagating decreased inheritance to holder of lock %d.\n", | ||
874 | l->ident); | ||
875 | |||
876 | if(litmus->__compare(new_max_eff_prio, BASE, owner, BASE)) { | ||
877 | TRACE_CUR("%s/%d has greater base priority than base priority of owner (%s/%d) of lock %d.\n", | ||
878 | (new_max_eff_prio) ? new_max_eff_prio->comm : "null", | ||
879 | (new_max_eff_prio) ? new_max_eff_prio->pid : 0, | ||
880 | owner->comm, | ||
881 | owner->pid, | ||
882 | l->ident); | ||
883 | |||
884 | decreased_prio = new_max_eff_prio; | ||
885 | } | ||
886 | else { | ||
887 | TRACE_CUR("%s/%d has lesser base priority than base priority of owner (%s/%d) of lock %d.\n", | ||
888 | (new_max_eff_prio) ? new_max_eff_prio->comm : "null", | ||
889 | (new_max_eff_prio) ? new_max_eff_prio->pid : 0, | ||
890 | owner->comm, | ||
891 | owner->pid, | ||
892 | l->ident); | ||
893 | |||
894 | decreased_prio = NULL; | ||
895 | } | ||
896 | |||
897 | // beware: recursion | ||
898 | litmus->nested_decrease_prio(owner, decreased_prio, &mutex->lock, irqflags); // will unlock mutex->lock | ||
899 | } | ||
900 | else { | ||
901 | raw_spin_unlock(&tsk_rt(owner)->hp_blocked_tasks_lock); | ||
902 | unlock_fine_irqrestore(&mutex->lock, irqflags); | ||
903 | } | ||
904 | } | ||
905 | else { | ||
906 | TRACE_TASK(t, "prioq_mutex %d state after prio decrease in priority queue\n", l->ident); | ||
907 | __dump_lock_info(mutex); | ||
908 | |||
909 | TRACE_TASK(t, "is not hp_waiter. No propagation.\n"); | ||
910 | unlock_fine_irqrestore(&mutex->lock, irqflags); | ||
911 | } | ||
912 | } | ||
913 | else { | ||
914 | struct litmus_lock *still_blocked; | ||
915 | |||
916 | TRACE_TASK(t, "prioq_mutex %d state after prio decrease in priority queue\n", l->ident); | ||
917 | __dump_lock_info(mutex); | ||
918 | |||
919 | still_blocked = tsk_rt(t)->blocked_lock; | ||
920 | |||
921 | TRACE_TASK(t, "is not blocked on lock %d.\n", l->ident); | ||
922 | if(still_blocked) { | ||
923 | TRACE_TASK(t, "is still blocked on a lock though (lock %d).\n", | ||
924 | still_blocked->ident); | ||
925 | if(still_blocked->ops->propagate_decrease_inheritance) { | ||
926 | /* due to linked nesting of spinlocks (acq. A, acq. B, free A, free B) | ||
927 | we know that task 't' has not released any locks behind us in this | ||
928 | chain. propagation just needs to catch up with task 't' */ | ||
929 | still_blocked->ops->propagate_decrease_inheritance(still_blocked, | ||
930 | t, | ||
931 | &mutex->lock, | ||
932 | irqflags); | ||
933 | } | ||
934 | else { | ||
935 | TRACE_TASK(t, "Inheritor is blocked on lock (%p) that does not support nesting!\n", | ||
936 | still_blocked); | ||
937 | unlock_fine_irqrestore(&mutex->lock, irqflags); | ||
938 | } | ||
939 | } | ||
940 | else { | ||
941 | unlock_fine_irqrestore(&mutex->lock, irqflags); | ||
942 | } | ||
943 | } | ||
944 | } | ||
945 | |||
946 | |||
947 | int prioq_mutex_close(struct litmus_lock* l) | ||
948 | { | ||
949 | struct task_struct *t = current; | ||
950 | struct prioq_mutex *mutex = prioq_mutex_from_lock(l); | ||
951 | unsigned long flags; | ||
952 | |||
953 | int owner; | ||
954 | |||
955 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
956 | raw_spinlock_t *dgl_lock = litmus->get_dgl_spinlock(t); | ||
957 | #endif | ||
958 | |||
959 | lock_global_irqsave(dgl_lock, flags); | ||
960 | lock_fine_irqsave(&mutex->lock, flags); | ||
961 | |||
962 | owner = (mutex->owner == t); | ||
963 | |||
964 | unlock_fine_irqrestore(&mutex->lock, flags); | ||
965 | unlock_global_irqrestore(dgl_lock, flags); | ||
966 | |||
967 | /* | ||
968 | TODO: Currently panic. FIX THIS! | ||
969 | if (owner) | ||
970 | prioq_mutex_unlock(l); | ||
971 | */ | ||
972 | |||
973 | return 0; | ||
974 | } | ||
975 | |||
976 | void prioq_mutex_free(struct litmus_lock* lock) | ||
977 | { | ||
978 | kfree(prioq_mutex_from_lock(lock)); | ||
979 | } | ||
980 | |||
981 | |||
982 | /* The following may race if DGLs are enabled. Only examine /proc if things | ||
983 | appear to be locked up. TODO: FIX THIS! Must find an elegant way to transmit | ||
984 | DGL lock to function. */ | ||
985 | static int prioq_proc_print(char *page, char **start, off_t off, int count, int *eof, void *data) | ||
986 | { | ||
987 | struct prioq_mutex *mutex = prioq_mutex_from_lock((struct litmus_lock*)data); | ||
988 | |||
989 | int attempts = 0; | ||
990 | const int max_attempts = 10; | ||
991 | int locked = 0; | ||
992 | unsigned long flags; | ||
993 | |||
994 | int size = count; | ||
995 | char *next = page; | ||
996 | int w; | ||
997 | |||
998 | while(attempts < max_attempts) | ||
999 | { | ||
1000 | locked = raw_spin_trylock_irqsave(&mutex->lock, flags); | ||
1001 | |||
1002 | if (unlikely(!locked)) { | ||
1003 | ++attempts; | ||
1004 | cpu_relax(); | ||
1005 | } | ||
1006 | else { | ||
1007 | break; | ||
1008 | } | ||
1009 | } | ||
1010 | |||
1011 | if (locked) { | ||
1012 | w = scnprintf(next, size, "%s (mutex: %p, data: %p):\n", mutex->litmus_lock.name, mutex, data); | ||
1013 | size -= w; | ||
1014 | next += w; | ||
1015 | |||
1016 | w = scnprintf(next, size, | ||
1017 | "owner: %s/%d (inh: %s/%d)\n", | ||
1018 | (mutex->owner) ? | ||
1019 | mutex->owner->comm : "null", | ||
1020 | (mutex->owner) ? | ||
1021 | mutex->owner->pid : 0, | ||
1022 | (mutex->owner && tsk_rt(mutex->owner)->inh_task) ? | ||
1023 | tsk_rt(mutex->owner)->inh_task->comm : "null", | ||
1024 | (mutex->owner && tsk_rt(mutex->owner)->inh_task) ? | ||
1025 | tsk_rt(mutex->owner)->inh_task->pid : 0); | ||
1026 | size -= w; | ||
1027 | next += w; | ||
1028 | |||
1029 | w = scnprintf(next, size, | ||
1030 | "hp waiter: %s/%d (inh: %s/%d)\n", | ||
1031 | (mutex->hp_waiter) ? | ||
1032 | mutex->hp_waiter->comm : "null", | ||
1033 | (mutex->hp_waiter) ? | ||
1034 | mutex->hp_waiter->pid : 0, | ||
1035 | (mutex->hp_waiter && tsk_rt(mutex->hp_waiter)->inh_task) ? | ||
1036 | tsk_rt(mutex->hp_waiter)->inh_task->comm : "null", | ||
1037 | (mutex->hp_waiter && tsk_rt(mutex->hp_waiter)->inh_task) ? | ||
1038 | tsk_rt(mutex->hp_waiter)->inh_task->pid : 0); | ||
1039 | size -= w; | ||
1040 | next += w; | ||
1041 | |||
1042 | w = scnprintf(next, size, "\nblocked tasks, front to back:\n"); | ||
1043 | size -= w; | ||
1044 | next += w; | ||
1045 | |||
1046 | if (waitqueue_active(&mutex->wait)) { | ||
1047 | wait_queue_t *q; | ||
1048 | struct list_head *pos; | ||
1049 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
1050 | dgl_wait_state_t *dgl_wait = NULL; | ||
1051 | #endif | ||
1052 | list_for_each(pos, &mutex->wait.task_list) { | ||
1053 | struct task_struct *blocked_task; | ||
1054 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
1055 | int enabled = 1; | ||
1056 | #endif | ||
1057 | q = list_entry(pos, wait_queue_t, task_list); | ||
1058 | |||
1059 | blocked_task = get_queued_task(q); | ||
1060 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
1061 | if(q->func == dgl_wake_up) { | ||
1062 | dgl_wait = (dgl_wait_state_t*) q->private; | ||
1063 | blocked_task = dgl_wait->task; | ||
1064 | |||
1065 | if(tsk_rt(blocked_task)->blocked_lock != &mutex->litmus_lock) | ||
1066 | enabled = 0; | ||
1067 | } | ||
1068 | else { | ||
1069 | blocked_task = (struct task_struct*) q->private; | ||
1070 | } | ||
1071 | #else | ||
1072 | blocked_task = (struct task_struct*) q->private; | ||
1073 | #endif | ||
1074 | |||
1075 | w = scnprintf(next, size, | ||
1076 | "\t%s/%d (inh: %s/%d)" | ||
1077 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
1078 | " DGL enabled: %d" | ||
1079 | #endif | ||
1080 | "\n", | ||
1081 | blocked_task->comm, blocked_task->pid, | ||
1082 | (tsk_rt(blocked_task)->inh_task) ? | ||
1083 | tsk_rt(blocked_task)->inh_task->comm : "null", | ||
1084 | (tsk_rt(blocked_task)->inh_task) ? | ||
1085 | tsk_rt(blocked_task)->inh_task->pid : 0 | ||
1086 | #ifdef CONFIG_LITMUS_DGL_SUPPORT | ||
1087 | , enabled | ||
1088 | #endif | ||
1089 | ); | ||
1090 | size -= w; | ||
1091 | next += w; | ||
1092 | } | ||
1093 | } | ||
1094 | else { | ||
1095 | w = scnprintf(next, size, "\t<NONE>\n"); | ||
1096 | size -= w; | ||
1097 | next += w; | ||
1098 | } | ||
1099 | |||
1100 | raw_spin_unlock_irqrestore(&mutex->lock, flags); | ||
1101 | } | ||
1102 | else { | ||
1103 | w = scnprintf(next, size, "%s is busy.\n", mutex->litmus_lock.name); | ||
1104 | size -= w; | ||
1105 | next += w; | ||
1106 | } | ||
1107 | |||
1108 | return count - size; | ||
1109 | } | ||
1110 | |||
1111 | static void prioq_proc_add(struct litmus_lock* l) | ||
1112 | { | ||
1113 | snprintf(l->name, LOCK_NAME_LEN, "prioq-%d", l->ident); | ||
1114 | |||
1115 | l->proc_entry = litmus_add_proc_lock(l, prioq_proc_print); | ||
1116 | } | ||
1117 | |||
1118 | static void prioq_proc_remove(struct litmus_lock* l) | ||
1119 | { | ||
1120 | litmus_remove_proc_lock(l); | ||
1121 | } | ||
1122 | |||
1123 | static struct litmus_lock_proc_ops prioq_proc_ops = | ||
1124 | { | ||
1125 | .add = prioq_proc_add, | ||
1126 | .remove = prioq_proc_remove | ||
1127 | }; | ||
1128 | |||
1129 | |||
1130 | struct litmus_lock* prioq_mutex_new(struct litmus_lock_ops* ops) | ||
1131 | { | ||
1132 | struct prioq_mutex* mutex; | ||
1133 | |||
1134 | mutex = kmalloc(sizeof(*mutex), GFP_KERNEL); | ||
1135 | if (!mutex) | ||
1136 | return NULL; | ||
1137 | memset(mutex, 0, sizeof(*mutex)); | ||
1138 | |||
1139 | mutex->litmus_lock.ops = ops; | ||
1140 | mutex->owner = NULL; | ||
1141 | mutex->hp_waiter = NULL; | ||
1142 | init_waitqueue_head(&mutex->wait); | ||
1143 | |||
1144 | |||
1145 | #ifdef CONFIG_DEBUG_SPINLOCK | ||
1146 | { | ||
1147 | __raw_spin_lock_init(&mutex->lock, | ||
1148 | ((struct litmus_lock*)mutex)->cheat_lockdep, | ||
1149 | &((struct litmus_lock*)mutex)->key); | ||
1150 | } | ||
1151 | #else | ||
1152 | raw_spin_lock_init(&mutex->lock); | ||
1153 | #endif | ||
1154 | |||
1155 | ((struct litmus_lock*)mutex)->nest.hp_waiter_ptr = &mutex->hp_waiter; | ||
1156 | ((struct litmus_lock*)mutex)->nest.owner_ptr = &mutex->owner; | ||
1157 | |||
1158 | ((struct litmus_lock*)mutex)->proc = &prioq_proc_ops; | ||
1159 | |||
1160 | return &mutex->litmus_lock; | ||
1161 | } | ||
1162 | |||