diff options
author | Steven Whitehouse <swhiteho@redhat.com> | 2008-05-21 12:03:22 -0400 |
---|---|---|
committer | Steven Whitehouse <swhiteho@redhat.com> | 2008-06-27 04:39:22 -0400 |
commit | 6802e3400ff4549525930ee744030c36fce9cc73 (patch) | |
tree | db889bf5337c1d3bb12ebbf571c3c1cad1040496 /fs/gfs2/glock.c | |
parent | 543cf4cb3fe6f6cae3651ba918b9c56200b257d0 (diff) |
[GFS2] Clean up the glock core
This patch implements a number of cleanups to the core of the
GFS2 glock code. As a result a lot of code is removed. It looks
like a really big change, but actually a large part of this patch
is either removing or moving existing code.
There are some new bits too though, such as the new run_queue()
function which is considerably streamlined. Highlights of this
patch include:
o Fixes a cluster coherency bug during SH -> EX lock conversions
o Removes the "glmutex" code in favour of a single bit lock
o Removes the ->go_xmote_bh() for inodes since it was duplicating
->go_lock()
o We now only use the ->lm_lock() function for both locks and
unlocks (i.e. unlock is a lock with target mode LM_ST_UNLOCKED)
o The fast path is considerably shortly, giving performance gains
especially with lock_nolock
o The glock_workqueue is now used for all the callbacks from the DLM
which allows us to simplify the lock_dlm module (see following patch)
o The way is now open to make further changes such as eliminating the two
threads (gfs2_glockd and gfs2_scand) in favour of a more efficient
scheme.
This patch has undergone extensive testing with various test suites
so it should be pretty stable by now.
Signed-off-by: Steven Whitehouse <swhiteho@redhat.com>
Cc: Bob Peterson <rpeterso@redhat.com>
Diffstat (limited to 'fs/gfs2/glock.c')
-rw-r--r-- | fs/gfs2/glock.c | 1611 |
1 files changed, 644 insertions, 967 deletions
diff --git a/fs/gfs2/glock.c b/fs/gfs2/glock.c index d636b3e80f5d..519a54cc0b7b 100644 --- a/fs/gfs2/glock.c +++ b/fs/gfs2/glock.c | |||
@@ -45,21 +45,19 @@ struct gfs2_gl_hash_bucket { | |||
45 | struct hlist_head hb_list; | 45 | struct hlist_head hb_list; |
46 | }; | 46 | }; |
47 | 47 | ||
48 | struct glock_iter { | 48 | struct gfs2_glock_iter { |
49 | int hash; /* hash bucket index */ | 49 | int hash; /* hash bucket index */ |
50 | struct gfs2_sbd *sdp; /* incore superblock */ | 50 | struct gfs2_sbd *sdp; /* incore superblock */ |
51 | struct gfs2_glock *gl; /* current glock struct */ | 51 | struct gfs2_glock *gl; /* current glock struct */ |
52 | struct seq_file *seq; /* sequence file for debugfs */ | 52 | char string[512]; /* scratch space */ |
53 | char string[512]; /* scratch space */ | ||
54 | }; | 53 | }; |
55 | 54 | ||
56 | typedef void (*glock_examiner) (struct gfs2_glock * gl); | 55 | typedef void (*glock_examiner) (struct gfs2_glock * gl); |
57 | 56 | ||
58 | static int gfs2_dump_lockstate(struct gfs2_sbd *sdp); | 57 | static int gfs2_dump_lockstate(struct gfs2_sbd *sdp); |
59 | static int dump_glock(struct glock_iter *gi, struct gfs2_glock *gl); | 58 | static int __dump_glock(struct seq_file *seq, const struct gfs2_glock *gl); |
60 | static void gfs2_glock_xmote_th(struct gfs2_glock *gl, struct gfs2_holder *gh); | 59 | #define GLOCK_BUG_ON(gl,x) do { if (unlikely(x)) { __dump_glock(NULL, gl); BUG(); } } while(0) |
61 | static void gfs2_glock_drop_th(struct gfs2_glock *gl); | 60 | static void do_xmote(struct gfs2_glock *gl, struct gfs2_holder *gh, unsigned int target); |
62 | static void run_queue(struct gfs2_glock *gl); | ||
63 | 61 | ||
64 | static DECLARE_RWSEM(gfs2_umount_flush_sem); | 62 | static DECLARE_RWSEM(gfs2_umount_flush_sem); |
65 | static struct dentry *gfs2_root; | 63 | static struct dentry *gfs2_root; |
@@ -123,33 +121,6 @@ static inline rwlock_t *gl_lock_addr(unsigned int x) | |||
123 | #endif | 121 | #endif |
124 | 122 | ||
125 | /** | 123 | /** |
126 | * relaxed_state_ok - is a requested lock compatible with the current lock mode? | ||
127 | * @actual: the current state of the lock | ||
128 | * @requested: the lock state that was requested by the caller | ||
129 | * @flags: the modifier flags passed in by the caller | ||
130 | * | ||
131 | * Returns: 1 if the locks are compatible, 0 otherwise | ||
132 | */ | ||
133 | |||
134 | static inline int relaxed_state_ok(unsigned int actual, unsigned requested, | ||
135 | int flags) | ||
136 | { | ||
137 | if (actual == requested) | ||
138 | return 1; | ||
139 | |||
140 | if (flags & GL_EXACT) | ||
141 | return 0; | ||
142 | |||
143 | if (actual == LM_ST_EXCLUSIVE && requested == LM_ST_SHARED) | ||
144 | return 1; | ||
145 | |||
146 | if (actual != LM_ST_UNLOCKED && (flags & LM_FLAG_ANY)) | ||
147 | return 1; | ||
148 | |||
149 | return 0; | ||
150 | } | ||
151 | |||
152 | /** | ||
153 | * gl_hash() - Turn glock number into hash bucket number | 124 | * gl_hash() - Turn glock number into hash bucket number |
154 | * @lock: The glock number | 125 | * @lock: The glock number |
155 | * | 126 | * |
@@ -211,17 +182,14 @@ static void gfs2_glock_hold(struct gfs2_glock *gl) | |||
211 | int gfs2_glock_put(struct gfs2_glock *gl) | 182 | int gfs2_glock_put(struct gfs2_glock *gl) |
212 | { | 183 | { |
213 | int rv = 0; | 184 | int rv = 0; |
214 | struct gfs2_sbd *sdp = gl->gl_sbd; | ||
215 | 185 | ||
216 | write_lock(gl_lock_addr(gl->gl_hash)); | 186 | write_lock(gl_lock_addr(gl->gl_hash)); |
217 | if (atomic_dec_and_test(&gl->gl_ref)) { | 187 | if (atomic_dec_and_test(&gl->gl_ref)) { |
218 | hlist_del(&gl->gl_list); | 188 | hlist_del(&gl->gl_list); |
219 | write_unlock(gl_lock_addr(gl->gl_hash)); | 189 | write_unlock(gl_lock_addr(gl->gl_hash)); |
220 | gfs2_assert(sdp, gl->gl_state == LM_ST_UNLOCKED); | 190 | GLOCK_BUG_ON(gl, gl->gl_state != LM_ST_UNLOCKED); |
221 | gfs2_assert(sdp, list_empty(&gl->gl_reclaim)); | 191 | GLOCK_BUG_ON(gl, !list_empty(&gl->gl_reclaim)); |
222 | gfs2_assert(sdp, list_empty(&gl->gl_holders)); | 192 | GLOCK_BUG_ON(gl, !list_empty(&gl->gl_holders)); |
223 | gfs2_assert(sdp, list_empty(&gl->gl_waiters1)); | ||
224 | gfs2_assert(sdp, list_empty(&gl->gl_waiters3)); | ||
225 | glock_free(gl); | 193 | glock_free(gl); |
226 | rv = 1; | 194 | rv = 1; |
227 | goto out; | 195 | goto out; |
@@ -281,16 +249,382 @@ static struct gfs2_glock *gfs2_glock_find(const struct gfs2_sbd *sdp, | |||
281 | return gl; | 249 | return gl; |
282 | } | 250 | } |
283 | 251 | ||
252 | /** | ||
253 | * may_grant - check if its ok to grant a new lock | ||
254 | * @gl: The glock | ||
255 | * @gh: The lock request which we wish to grant | ||
256 | * | ||
257 | * Returns: true if its ok to grant the lock | ||
258 | */ | ||
259 | |||
260 | static inline int may_grant(const struct gfs2_glock *gl, const struct gfs2_holder *gh) | ||
261 | { | ||
262 | const struct gfs2_holder *gh_head = list_entry(gl->gl_holders.next, const struct gfs2_holder, gh_list); | ||
263 | if ((gh->gh_state == LM_ST_EXCLUSIVE || | ||
264 | gh_head->gh_state == LM_ST_EXCLUSIVE) && gh != gh_head) | ||
265 | return 0; | ||
266 | if (gl->gl_state == gh->gh_state) | ||
267 | return 1; | ||
268 | if (gh->gh_flags & GL_EXACT) | ||
269 | return 0; | ||
270 | if (gh->gh_state == LM_ST_SHARED && gl->gl_state == LM_ST_EXCLUSIVE) | ||
271 | return 1; | ||
272 | if (gl->gl_state != LM_ST_UNLOCKED && (gh->gh_flags & LM_FLAG_ANY)) | ||
273 | return 1; | ||
274 | return 0; | ||
275 | } | ||
276 | |||
277 | static void gfs2_holder_wake(struct gfs2_holder *gh) | ||
278 | { | ||
279 | clear_bit(HIF_WAIT, &gh->gh_iflags); | ||
280 | smp_mb__after_clear_bit(); | ||
281 | wake_up_bit(&gh->gh_iflags, HIF_WAIT); | ||
282 | } | ||
283 | |||
284 | /** | ||
285 | * do_promote - promote as many requests as possible on the current queue | ||
286 | * @gl: The glock | ||
287 | * | ||
288 | * Returns: true if there is a blocked holder at the head of the list | ||
289 | */ | ||
290 | |||
291 | static int do_promote(struct gfs2_glock *gl) | ||
292 | { | ||
293 | const struct gfs2_glock_operations *glops = gl->gl_ops; | ||
294 | struct gfs2_holder *gh, *tmp; | ||
295 | int ret; | ||
296 | |||
297 | restart: | ||
298 | list_for_each_entry_safe(gh, tmp, &gl->gl_holders, gh_list) { | ||
299 | if (test_bit(HIF_HOLDER, &gh->gh_iflags)) | ||
300 | continue; | ||
301 | if (may_grant(gl, gh)) { | ||
302 | if (gh->gh_list.prev == &gl->gl_holders && | ||
303 | glops->go_lock) { | ||
304 | spin_unlock(&gl->gl_spin); | ||
305 | /* FIXME: eliminate this eventually */ | ||
306 | ret = glops->go_lock(gh); | ||
307 | spin_lock(&gl->gl_spin); | ||
308 | if (ret) { | ||
309 | gh->gh_error = ret; | ||
310 | list_del_init(&gh->gh_list); | ||
311 | gfs2_holder_wake(gh); | ||
312 | goto restart; | ||
313 | } | ||
314 | set_bit(HIF_HOLDER, &gh->gh_iflags); | ||
315 | gfs2_holder_wake(gh); | ||
316 | goto restart; | ||
317 | } | ||
318 | set_bit(HIF_HOLDER, &gh->gh_iflags); | ||
319 | gfs2_holder_wake(gh); | ||
320 | continue; | ||
321 | } | ||
322 | if (gh->gh_list.prev == &gl->gl_holders) | ||
323 | return 1; | ||
324 | break; | ||
325 | } | ||
326 | return 0; | ||
327 | } | ||
328 | |||
329 | /** | ||
330 | * do_error - Something unexpected has happened during a lock request | ||
331 | * | ||
332 | */ | ||
333 | |||
334 | static inline void do_error(struct gfs2_glock *gl, const int ret) | ||
335 | { | ||
336 | struct gfs2_holder *gh, *tmp; | ||
337 | |||
338 | list_for_each_entry_safe(gh, tmp, &gl->gl_holders, gh_list) { | ||
339 | if (test_bit(HIF_HOLDER, &gh->gh_iflags)) | ||
340 | continue; | ||
341 | if (ret & LM_OUT_ERROR) | ||
342 | gh->gh_error = -EIO; | ||
343 | else if (gh->gh_flags & (LM_FLAG_TRY | LM_FLAG_TRY_1CB)) | ||
344 | gh->gh_error = GLR_TRYFAILED; | ||
345 | else | ||
346 | continue; | ||
347 | list_del_init(&gh->gh_list); | ||
348 | gfs2_holder_wake(gh); | ||
349 | } | ||
350 | } | ||
351 | |||
352 | /** | ||
353 | * find_first_waiter - find the first gh that's waiting for the glock | ||
354 | * @gl: the glock | ||
355 | */ | ||
356 | |||
357 | static inline struct gfs2_holder *find_first_waiter(const struct gfs2_glock *gl) | ||
358 | { | ||
359 | struct gfs2_holder *gh; | ||
360 | |||
361 | list_for_each_entry(gh, &gl->gl_holders, gh_list) { | ||
362 | if (!test_bit(HIF_HOLDER, &gh->gh_iflags)) | ||
363 | return gh; | ||
364 | } | ||
365 | return NULL; | ||
366 | } | ||
367 | |||
368 | /** | ||
369 | * state_change - record that the glock is now in a different state | ||
370 | * @gl: the glock | ||
371 | * @new_state the new state | ||
372 | * | ||
373 | */ | ||
374 | |||
375 | static void state_change(struct gfs2_glock *gl, unsigned int new_state) | ||
376 | { | ||
377 | int held1, held2; | ||
378 | |||
379 | held1 = (gl->gl_state != LM_ST_UNLOCKED); | ||
380 | held2 = (new_state != LM_ST_UNLOCKED); | ||
381 | |||
382 | if (held1 != held2) { | ||
383 | if (held2) | ||
384 | gfs2_glock_hold(gl); | ||
385 | else | ||
386 | gfs2_glock_put(gl); | ||
387 | } | ||
388 | |||
389 | gl->gl_state = new_state; | ||
390 | gl->gl_tchange = jiffies; | ||
391 | } | ||
392 | |||
393 | static void gfs2_demote_wake(struct gfs2_glock *gl) | ||
394 | { | ||
395 | gl->gl_demote_state = LM_ST_EXCLUSIVE; | ||
396 | clear_bit(GLF_DEMOTE, &gl->gl_flags); | ||
397 | smp_mb__after_clear_bit(); | ||
398 | wake_up_bit(&gl->gl_flags, GLF_DEMOTE); | ||
399 | } | ||
400 | |||
401 | /** | ||
402 | * finish_xmote - The DLM has replied to one of our lock requests | ||
403 | * @gl: The glock | ||
404 | * @ret: The status from the DLM | ||
405 | * | ||
406 | */ | ||
407 | |||
408 | static void finish_xmote(struct gfs2_glock *gl, unsigned int ret) | ||
409 | { | ||
410 | const struct gfs2_glock_operations *glops = gl->gl_ops; | ||
411 | struct gfs2_holder *gh; | ||
412 | unsigned state = ret & LM_OUT_ST_MASK; | ||
413 | |||
414 | spin_lock(&gl->gl_spin); | ||
415 | state_change(gl, state); | ||
416 | gh = find_first_waiter(gl); | ||
417 | |||
418 | /* Demote to UN request arrived during demote to SH or DF */ | ||
419 | if (test_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags) && | ||
420 | state != LM_ST_UNLOCKED && gl->gl_demote_state == LM_ST_UNLOCKED) | ||
421 | gl->gl_target = LM_ST_UNLOCKED; | ||
422 | |||
423 | /* Check for state != intended state */ | ||
424 | if (unlikely(state != gl->gl_target)) { | ||
425 | if (gh && !test_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags)) { | ||
426 | /* move to back of queue and try next entry */ | ||
427 | if (ret & LM_OUT_CANCELED) { | ||
428 | if ((gh->gh_flags & LM_FLAG_PRIORITY) == 0) | ||
429 | list_move_tail(&gh->gh_list, &gl->gl_holders); | ||
430 | gh = find_first_waiter(gl); | ||
431 | gl->gl_target = gh->gh_state; | ||
432 | goto retry; | ||
433 | } | ||
434 | /* Some error or failed "try lock" - report it */ | ||
435 | if ((ret & LM_OUT_ERROR) || | ||
436 | (gh->gh_flags & (LM_FLAG_TRY | LM_FLAG_TRY_1CB))) { | ||
437 | gl->gl_target = gl->gl_state; | ||
438 | do_error(gl, ret); | ||
439 | goto out; | ||
440 | } | ||
441 | } | ||
442 | switch(state) { | ||
443 | /* Unlocked due to conversion deadlock, try again */ | ||
444 | case LM_ST_UNLOCKED: | ||
445 | retry: | ||
446 | do_xmote(gl, gh, gl->gl_target); | ||
447 | break; | ||
448 | /* Conversion fails, unlock and try again */ | ||
449 | case LM_ST_SHARED: | ||
450 | case LM_ST_DEFERRED: | ||
451 | do_xmote(gl, gh, LM_ST_UNLOCKED); | ||
452 | break; | ||
453 | default: /* Everything else */ | ||
454 | printk(KERN_ERR "GFS2: wanted %u got %u\n", gl->gl_target, state); | ||
455 | GLOCK_BUG_ON(gl, 1); | ||
456 | } | ||
457 | spin_unlock(&gl->gl_spin); | ||
458 | gfs2_glock_put(gl); | ||
459 | return; | ||
460 | } | ||
461 | |||
462 | /* Fast path - we got what we asked for */ | ||
463 | if (test_and_clear_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags)) | ||
464 | gfs2_demote_wake(gl); | ||
465 | if (state != LM_ST_UNLOCKED) { | ||
466 | if (glops->go_xmote_bh) { | ||
467 | int rv; | ||
468 | spin_unlock(&gl->gl_spin); | ||
469 | rv = glops->go_xmote_bh(gl, gh); | ||
470 | if (rv == -EAGAIN) | ||
471 | return; | ||
472 | spin_lock(&gl->gl_spin); | ||
473 | if (rv) { | ||
474 | do_error(gl, rv); | ||
475 | goto out; | ||
476 | } | ||
477 | } | ||
478 | do_promote(gl); | ||
479 | } | ||
480 | out: | ||
481 | clear_bit(GLF_LOCK, &gl->gl_flags); | ||
482 | spin_unlock(&gl->gl_spin); | ||
483 | gfs2_glock_put(gl); | ||
484 | } | ||
485 | |||
486 | static unsigned int gfs2_lm_lock(struct gfs2_sbd *sdp, void *lock, | ||
487 | unsigned int cur_state, unsigned int req_state, | ||
488 | unsigned int flags) | ||
489 | { | ||
490 | int ret = LM_OUT_ERROR; | ||
491 | if (likely(!test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) | ||
492 | ret = sdp->sd_lockstruct.ls_ops->lm_lock(lock, cur_state, | ||
493 | req_state, flags); | ||
494 | return ret; | ||
495 | } | ||
496 | |||
497 | /** | ||
498 | * do_xmote - Calls the DLM to change the state of a lock | ||
499 | * @gl: The lock state | ||
500 | * @gh: The holder (only for promotes) | ||
501 | * @target: The target lock state | ||
502 | * | ||
503 | */ | ||
504 | |||
505 | static void do_xmote(struct gfs2_glock *gl, struct gfs2_holder *gh, unsigned int target) | ||
506 | { | ||
507 | const struct gfs2_glock_operations *glops = gl->gl_ops; | ||
508 | struct gfs2_sbd *sdp = gl->gl_sbd; | ||
509 | unsigned int lck_flags = gh ? gh->gh_flags : 0; | ||
510 | int ret; | ||
511 | |||
512 | lck_flags &= (LM_FLAG_TRY | LM_FLAG_TRY_1CB | LM_FLAG_NOEXP | | ||
513 | LM_FLAG_PRIORITY); | ||
514 | BUG_ON(gl->gl_state == target); | ||
515 | BUG_ON(gl->gl_state == gl->gl_target); | ||
516 | if ((target == LM_ST_UNLOCKED || target == LM_ST_DEFERRED) && | ||
517 | glops->go_inval) { | ||
518 | set_bit(GLF_INVALIDATE_IN_PROGRESS, &gl->gl_flags); | ||
519 | do_error(gl, 0); /* Fail queued try locks */ | ||
520 | } | ||
521 | spin_unlock(&gl->gl_spin); | ||
522 | if (glops->go_xmote_th) | ||
523 | glops->go_xmote_th(gl); | ||
524 | if (test_bit(GLF_INVALIDATE_IN_PROGRESS, &gl->gl_flags)) | ||
525 | glops->go_inval(gl, target == LM_ST_DEFERRED ? 0 : DIO_METADATA); | ||
526 | clear_bit(GLF_INVALIDATE_IN_PROGRESS, &gl->gl_flags); | ||
527 | |||
528 | gfs2_glock_hold(gl); | ||
529 | if (target != LM_ST_UNLOCKED && (gl->gl_state == LM_ST_SHARED || | ||
530 | gl->gl_state == LM_ST_DEFERRED) && | ||
531 | !(lck_flags & (LM_FLAG_TRY | LM_FLAG_TRY_1CB))) | ||
532 | lck_flags |= LM_FLAG_TRY_1CB; | ||
533 | ret = gfs2_lm_lock(sdp, gl->gl_lock, gl->gl_state, target, lck_flags); | ||
534 | |||
535 | if (!(ret & LM_OUT_ASYNC)) { | ||
536 | finish_xmote(gl, ret); | ||
537 | gfs2_glock_hold(gl); | ||
538 | if (queue_delayed_work(glock_workqueue, &gl->gl_work, 0) == 0) | ||
539 | gfs2_glock_put(gl); | ||
540 | } else { | ||
541 | GLOCK_BUG_ON(gl, ret != LM_OUT_ASYNC); | ||
542 | } | ||
543 | spin_lock(&gl->gl_spin); | ||
544 | } | ||
545 | |||
546 | /** | ||
547 | * find_first_holder - find the first "holder" gh | ||
548 | * @gl: the glock | ||
549 | */ | ||
550 | |||
551 | static inline struct gfs2_holder *find_first_holder(const struct gfs2_glock *gl) | ||
552 | { | ||
553 | struct gfs2_holder *gh; | ||
554 | |||
555 | if (!list_empty(&gl->gl_holders)) { | ||
556 | gh = list_entry(gl->gl_holders.next, struct gfs2_holder, gh_list); | ||
557 | if (test_bit(HIF_HOLDER, &gh->gh_iflags)) | ||
558 | return gh; | ||
559 | } | ||
560 | return NULL; | ||
561 | } | ||
562 | |||
563 | /** | ||
564 | * run_queue - do all outstanding tasks related to a glock | ||
565 | * @gl: The glock in question | ||
566 | * @nonblock: True if we must not block in run_queue | ||
567 | * | ||
568 | */ | ||
569 | |||
570 | static void run_queue(struct gfs2_glock *gl, const int nonblock) | ||
571 | { | ||
572 | struct gfs2_holder *gh = NULL; | ||
573 | |||
574 | if (test_and_set_bit(GLF_LOCK, &gl->gl_flags)) | ||
575 | return; | ||
576 | |||
577 | GLOCK_BUG_ON(gl, test_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags)); | ||
578 | |||
579 | if (test_bit(GLF_DEMOTE, &gl->gl_flags) && | ||
580 | gl->gl_demote_state != gl->gl_state) { | ||
581 | if (find_first_holder(gl)) | ||
582 | goto out; | ||
583 | if (nonblock) | ||
584 | goto out_sched; | ||
585 | set_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags); | ||
586 | gl->gl_target = gl->gl_demote_state; | ||
587 | } else { | ||
588 | if (test_bit(GLF_DEMOTE, &gl->gl_flags)) | ||
589 | gfs2_demote_wake(gl); | ||
590 | if (do_promote(gl) == 0) | ||
591 | goto out; | ||
592 | gh = find_first_waiter(gl); | ||
593 | gl->gl_target = gh->gh_state; | ||
594 | if (!(gh->gh_flags & (LM_FLAG_TRY | LM_FLAG_TRY_1CB))) | ||
595 | do_error(gl, 0); /* Fail queued try locks */ | ||
596 | } | ||
597 | do_xmote(gl, gh, gl->gl_target); | ||
598 | return; | ||
599 | |||
600 | out_sched: | ||
601 | gfs2_glock_hold(gl); | ||
602 | if (queue_delayed_work(glock_workqueue, &gl->gl_work, 0) == 0) | ||
603 | gfs2_glock_put(gl); | ||
604 | out: | ||
605 | clear_bit(GLF_LOCK, &gl->gl_flags); | ||
606 | } | ||
607 | |||
284 | static void glock_work_func(struct work_struct *work) | 608 | static void glock_work_func(struct work_struct *work) |
285 | { | 609 | { |
610 | unsigned long delay = 0; | ||
286 | struct gfs2_glock *gl = container_of(work, struct gfs2_glock, gl_work.work); | 611 | struct gfs2_glock *gl = container_of(work, struct gfs2_glock, gl_work.work); |
287 | 612 | ||
613 | if (test_and_clear_bit(GLF_REPLY_PENDING, &gl->gl_flags)) | ||
614 | finish_xmote(gl, gl->gl_reply); | ||
288 | spin_lock(&gl->gl_spin); | 615 | spin_lock(&gl->gl_spin); |
289 | if (test_and_clear_bit(GLF_PENDING_DEMOTE, &gl->gl_flags)) | 616 | if (test_and_clear_bit(GLF_PENDING_DEMOTE, &gl->gl_flags)) { |
290 | set_bit(GLF_DEMOTE, &gl->gl_flags); | 617 | unsigned long holdtime, now = jiffies; |
291 | run_queue(gl); | 618 | holdtime = gl->gl_tchange + gl->gl_ops->go_min_hold_time; |
619 | if (time_before(now, holdtime)) | ||
620 | delay = holdtime - now; | ||
621 | set_bit(delay ? GLF_PENDING_DEMOTE : GLF_DEMOTE, &gl->gl_flags); | ||
622 | } | ||
623 | run_queue(gl, 0); | ||
292 | spin_unlock(&gl->gl_spin); | 624 | spin_unlock(&gl->gl_spin); |
293 | gfs2_glock_put(gl); | 625 | if (!delay || |
626 | queue_delayed_work(glock_workqueue, &gl->gl_work, delay) == 0) | ||
627 | gfs2_glock_put(gl); | ||
294 | } | 628 | } |
295 | 629 | ||
296 | static int gfs2_lm_get_lock(struct gfs2_sbd *sdp, struct lm_lockname *name, | 630 | static int gfs2_lm_get_lock(struct gfs2_sbd *sdp, struct lm_lockname *name, |
@@ -342,12 +676,10 @@ int gfs2_glock_get(struct gfs2_sbd *sdp, u64 number, | |||
342 | gl->gl_name = name; | 676 | gl->gl_name = name; |
343 | atomic_set(&gl->gl_ref, 1); | 677 | atomic_set(&gl->gl_ref, 1); |
344 | gl->gl_state = LM_ST_UNLOCKED; | 678 | gl->gl_state = LM_ST_UNLOCKED; |
679 | gl->gl_target = LM_ST_UNLOCKED; | ||
345 | gl->gl_demote_state = LM_ST_EXCLUSIVE; | 680 | gl->gl_demote_state = LM_ST_EXCLUSIVE; |
346 | gl->gl_hash = hash; | 681 | gl->gl_hash = hash; |
347 | gl->gl_owner_pid = NULL; | ||
348 | gl->gl_ip = 0; | ||
349 | gl->gl_ops = glops; | 682 | gl->gl_ops = glops; |
350 | gl->gl_req_gh = NULL; | ||
351 | gl->gl_stamp = jiffies; | 683 | gl->gl_stamp = jiffies; |
352 | gl->gl_tchange = jiffies; | 684 | gl->gl_tchange = jiffies; |
353 | gl->gl_object = NULL; | 685 | gl->gl_object = NULL; |
@@ -447,13 +779,6 @@ void gfs2_holder_uninit(struct gfs2_holder *gh) | |||
447 | gh->gh_ip = 0; | 779 | gh->gh_ip = 0; |
448 | } | 780 | } |
449 | 781 | ||
450 | static void gfs2_holder_wake(struct gfs2_holder *gh) | ||
451 | { | ||
452 | clear_bit(HIF_WAIT, &gh->gh_iflags); | ||
453 | smp_mb__after_clear_bit(); | ||
454 | wake_up_bit(&gh->gh_iflags, HIF_WAIT); | ||
455 | } | ||
456 | |||
457 | static int just_schedule(void *word) | 782 | static int just_schedule(void *word) |
458 | { | 783 | { |
459 | schedule(); | 784 | schedule(); |
@@ -466,14 +791,6 @@ static void wait_on_holder(struct gfs2_holder *gh) | |||
466 | wait_on_bit(&gh->gh_iflags, HIF_WAIT, just_schedule, TASK_UNINTERRUPTIBLE); | 791 | wait_on_bit(&gh->gh_iflags, HIF_WAIT, just_schedule, TASK_UNINTERRUPTIBLE); |
467 | } | 792 | } |
468 | 793 | ||
469 | static void gfs2_demote_wake(struct gfs2_glock *gl) | ||
470 | { | ||
471 | gl->gl_demote_state = LM_ST_EXCLUSIVE; | ||
472 | clear_bit(GLF_DEMOTE, &gl->gl_flags); | ||
473 | smp_mb__after_clear_bit(); | ||
474 | wake_up_bit(&gl->gl_flags, GLF_DEMOTE); | ||
475 | } | ||
476 | |||
477 | static void wait_on_demote(struct gfs2_glock *gl) | 794 | static void wait_on_demote(struct gfs2_glock *gl) |
478 | { | 795 | { |
479 | might_sleep(); | 796 | might_sleep(); |
@@ -481,217 +798,6 @@ static void wait_on_demote(struct gfs2_glock *gl) | |||
481 | } | 798 | } |
482 | 799 | ||
483 | /** | 800 | /** |
484 | * rq_mutex - process a mutex request in the queue | ||
485 | * @gh: the glock holder | ||
486 | * | ||
487 | * Returns: 1 if the queue is blocked | ||
488 | */ | ||
489 | |||
490 | static int rq_mutex(struct gfs2_holder *gh) | ||
491 | { | ||
492 | struct gfs2_glock *gl = gh->gh_gl; | ||
493 | |||
494 | list_del_init(&gh->gh_list); | ||
495 | /* gh->gh_error never examined. */ | ||
496 | set_bit(GLF_LOCK, &gl->gl_flags); | ||
497 | clear_bit(HIF_WAIT, &gh->gh_iflags); | ||
498 | smp_mb(); | ||
499 | wake_up_bit(&gh->gh_iflags, HIF_WAIT); | ||
500 | |||
501 | return 1; | ||
502 | } | ||
503 | |||
504 | /** | ||
505 | * rq_promote - process a promote request in the queue | ||
506 | * @gh: the glock holder | ||
507 | * | ||
508 | * Acquire a new inter-node lock, or change a lock state to more restrictive. | ||
509 | * | ||
510 | * Returns: 1 if the queue is blocked | ||
511 | */ | ||
512 | |||
513 | static int rq_promote(struct gfs2_holder *gh) | ||
514 | { | ||
515 | struct gfs2_glock *gl = gh->gh_gl; | ||
516 | |||
517 | if (!relaxed_state_ok(gl->gl_state, gh->gh_state, gh->gh_flags)) { | ||
518 | if (list_empty(&gl->gl_holders)) { | ||
519 | gl->gl_req_gh = gh; | ||
520 | set_bit(GLF_LOCK, &gl->gl_flags); | ||
521 | spin_unlock(&gl->gl_spin); | ||
522 | gfs2_glock_xmote_th(gh->gh_gl, gh); | ||
523 | spin_lock(&gl->gl_spin); | ||
524 | } | ||
525 | return 1; | ||
526 | } | ||
527 | |||
528 | if (list_empty(&gl->gl_holders)) { | ||
529 | set_bit(HIF_FIRST, &gh->gh_iflags); | ||
530 | set_bit(GLF_LOCK, &gl->gl_flags); | ||
531 | } else { | ||
532 | struct gfs2_holder *next_gh; | ||
533 | if (gh->gh_state == LM_ST_EXCLUSIVE) | ||
534 | return 1; | ||
535 | next_gh = list_entry(gl->gl_holders.next, struct gfs2_holder, | ||
536 | gh_list); | ||
537 | if (next_gh->gh_state == LM_ST_EXCLUSIVE) | ||
538 | return 1; | ||
539 | } | ||
540 | |||
541 | list_move_tail(&gh->gh_list, &gl->gl_holders); | ||
542 | gh->gh_error = 0; | ||
543 | set_bit(HIF_HOLDER, &gh->gh_iflags); | ||
544 | |||
545 | gfs2_holder_wake(gh); | ||
546 | |||
547 | return 0; | ||
548 | } | ||
549 | |||
550 | /** | ||
551 | * rq_demote - process a demote request in the queue | ||
552 | * @gh: the glock holder | ||
553 | * | ||
554 | * Returns: 1 if the queue is blocked | ||
555 | */ | ||
556 | |||
557 | static int rq_demote(struct gfs2_glock *gl) | ||
558 | { | ||
559 | if (!list_empty(&gl->gl_holders)) | ||
560 | return 1; | ||
561 | |||
562 | if (gl->gl_state == gl->gl_demote_state || | ||
563 | gl->gl_state == LM_ST_UNLOCKED) { | ||
564 | gfs2_demote_wake(gl); | ||
565 | return 0; | ||
566 | } | ||
567 | |||
568 | set_bit(GLF_LOCK, &gl->gl_flags); | ||
569 | set_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags); | ||
570 | |||
571 | if (gl->gl_demote_state == LM_ST_UNLOCKED || | ||
572 | gl->gl_state != LM_ST_EXCLUSIVE) { | ||
573 | spin_unlock(&gl->gl_spin); | ||
574 | gfs2_glock_drop_th(gl); | ||
575 | } else { | ||
576 | spin_unlock(&gl->gl_spin); | ||
577 | gfs2_glock_xmote_th(gl, NULL); | ||
578 | } | ||
579 | |||
580 | spin_lock(&gl->gl_spin); | ||
581 | clear_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags); | ||
582 | |||
583 | return 0; | ||
584 | } | ||
585 | |||
586 | /** | ||
587 | * run_queue - process holder structures on a glock | ||
588 | * @gl: the glock | ||
589 | * | ||
590 | */ | ||
591 | static void run_queue(struct gfs2_glock *gl) | ||
592 | { | ||
593 | struct gfs2_holder *gh; | ||
594 | int blocked = 1; | ||
595 | |||
596 | for (;;) { | ||
597 | if (test_bit(GLF_LOCK, &gl->gl_flags)) | ||
598 | break; | ||
599 | |||
600 | if (!list_empty(&gl->gl_waiters1)) { | ||
601 | gh = list_entry(gl->gl_waiters1.next, | ||
602 | struct gfs2_holder, gh_list); | ||
603 | blocked = rq_mutex(gh); | ||
604 | } else if (test_bit(GLF_DEMOTE, &gl->gl_flags)) { | ||
605 | blocked = rq_demote(gl); | ||
606 | if (test_bit(GLF_WAITERS2, &gl->gl_flags) && | ||
607 | !blocked) { | ||
608 | set_bit(GLF_DEMOTE, &gl->gl_flags); | ||
609 | gl->gl_demote_state = LM_ST_UNLOCKED; | ||
610 | } | ||
611 | clear_bit(GLF_WAITERS2, &gl->gl_flags); | ||
612 | } else if (!list_empty(&gl->gl_waiters3)) { | ||
613 | gh = list_entry(gl->gl_waiters3.next, | ||
614 | struct gfs2_holder, gh_list); | ||
615 | blocked = rq_promote(gh); | ||
616 | } else | ||
617 | break; | ||
618 | |||
619 | if (blocked) | ||
620 | break; | ||
621 | } | ||
622 | } | ||
623 | |||
624 | /** | ||
625 | * gfs2_glmutex_lock - acquire a local lock on a glock | ||
626 | * @gl: the glock | ||
627 | * | ||
628 | * Gives caller exclusive access to manipulate a glock structure. | ||
629 | */ | ||
630 | |||
631 | static void gfs2_glmutex_lock(struct gfs2_glock *gl) | ||
632 | { | ||
633 | spin_lock(&gl->gl_spin); | ||
634 | if (test_and_set_bit(GLF_LOCK, &gl->gl_flags)) { | ||
635 | struct gfs2_holder gh; | ||
636 | |||
637 | gfs2_holder_init(gl, 0, 0, &gh); | ||
638 | set_bit(HIF_WAIT, &gh.gh_iflags); | ||
639 | list_add_tail(&gh.gh_list, &gl->gl_waiters1); | ||
640 | spin_unlock(&gl->gl_spin); | ||
641 | wait_on_holder(&gh); | ||
642 | gfs2_holder_uninit(&gh); | ||
643 | } else { | ||
644 | gl->gl_owner_pid = get_pid(task_pid(current)); | ||
645 | gl->gl_ip = (unsigned long)__builtin_return_address(0); | ||
646 | spin_unlock(&gl->gl_spin); | ||
647 | } | ||
648 | } | ||
649 | |||
650 | /** | ||
651 | * gfs2_glmutex_trylock - try to acquire a local lock on a glock | ||
652 | * @gl: the glock | ||
653 | * | ||
654 | * Returns: 1 if the glock is acquired | ||
655 | */ | ||
656 | |||
657 | static int gfs2_glmutex_trylock(struct gfs2_glock *gl) | ||
658 | { | ||
659 | int acquired = 1; | ||
660 | |||
661 | spin_lock(&gl->gl_spin); | ||
662 | if (test_and_set_bit(GLF_LOCK, &gl->gl_flags)) { | ||
663 | acquired = 0; | ||
664 | } else { | ||
665 | gl->gl_owner_pid = get_pid(task_pid(current)); | ||
666 | gl->gl_ip = (unsigned long)__builtin_return_address(0); | ||
667 | } | ||
668 | spin_unlock(&gl->gl_spin); | ||
669 | |||
670 | return acquired; | ||
671 | } | ||
672 | |||
673 | /** | ||
674 | * gfs2_glmutex_unlock - release a local lock on a glock | ||
675 | * @gl: the glock | ||
676 | * | ||
677 | */ | ||
678 | |||
679 | static void gfs2_glmutex_unlock(struct gfs2_glock *gl) | ||
680 | { | ||
681 | struct pid *pid; | ||
682 | |||
683 | spin_lock(&gl->gl_spin); | ||
684 | clear_bit(GLF_LOCK, &gl->gl_flags); | ||
685 | pid = gl->gl_owner_pid; | ||
686 | gl->gl_owner_pid = NULL; | ||
687 | gl->gl_ip = 0; | ||
688 | run_queue(gl); | ||
689 | spin_unlock(&gl->gl_spin); | ||
690 | |||
691 | put_pid(pid); | ||
692 | } | ||
693 | |||
694 | /** | ||
695 | * handle_callback - process a demote request | 801 | * handle_callback - process a demote request |
696 | * @gl: the glock | 802 | * @gl: the glock |
697 | * @state: the state the caller wants us to change to | 803 | * @state: the state the caller wants us to change to |
@@ -705,398 +811,45 @@ static void handle_callback(struct gfs2_glock *gl, unsigned int state, | |||
705 | { | 811 | { |
706 | int bit = delay ? GLF_PENDING_DEMOTE : GLF_DEMOTE; | 812 | int bit = delay ? GLF_PENDING_DEMOTE : GLF_DEMOTE; |
707 | 813 | ||
708 | spin_lock(&gl->gl_spin); | ||
709 | set_bit(bit, &gl->gl_flags); | 814 | set_bit(bit, &gl->gl_flags); |
710 | if (gl->gl_demote_state == LM_ST_EXCLUSIVE) { | 815 | if (gl->gl_demote_state == LM_ST_EXCLUSIVE) { |
711 | gl->gl_demote_state = state; | 816 | gl->gl_demote_state = state; |
712 | gl->gl_demote_time = jiffies; | 817 | gl->gl_demote_time = jiffies; |
713 | if (remote && gl->gl_ops->go_type == LM_TYPE_IOPEN && | 818 | if (remote && gl->gl_ops->go_type == LM_TYPE_IOPEN && |
714 | gl->gl_object) { | 819 | gl->gl_object) |
715 | gfs2_glock_schedule_for_reclaim(gl); | 820 | gfs2_glock_schedule_for_reclaim(gl); |
716 | spin_unlock(&gl->gl_spin); | ||
717 | return; | ||
718 | } | ||
719 | } else if (gl->gl_demote_state != LM_ST_UNLOCKED && | 821 | } else if (gl->gl_demote_state != LM_ST_UNLOCKED && |
720 | gl->gl_demote_state != state) { | 822 | gl->gl_demote_state != state) { |
721 | if (test_bit(GLF_DEMOTE_IN_PROGRESS, &gl->gl_flags)) | 823 | gl->gl_demote_state = LM_ST_UNLOCKED; |
722 | set_bit(GLF_WAITERS2, &gl->gl_flags); | ||
723 | else | ||
724 | gl->gl_demote_state = LM_ST_UNLOCKED; | ||
725 | } | 824 | } |
726 | spin_unlock(&gl->gl_spin); | ||
727 | } | 825 | } |
728 | 826 | ||
729 | /** | 827 | /** |
730 | * state_change - record that the glock is now in a different state | 828 | * gfs2_glock_wait - wait on a glock acquisition |
731 | * @gl: the glock | ||
732 | * @new_state the new state | ||
733 | * | ||
734 | */ | ||
735 | |||
736 | static void state_change(struct gfs2_glock *gl, unsigned int new_state) | ||
737 | { | ||
738 | int held1, held2; | ||
739 | |||
740 | held1 = (gl->gl_state != LM_ST_UNLOCKED); | ||
741 | held2 = (new_state != LM_ST_UNLOCKED); | ||
742 | |||
743 | if (held1 != held2) { | ||
744 | if (held2) | ||
745 | gfs2_glock_hold(gl); | ||
746 | else | ||
747 | gfs2_glock_put(gl); | ||
748 | } | ||
749 | |||
750 | gl->gl_state = new_state; | ||
751 | gl->gl_tchange = jiffies; | ||
752 | } | ||
753 | |||
754 | /** | ||
755 | * drop_bh - Called after a lock module unlock completes | ||
756 | * @gl: the glock | ||
757 | * @ret: the return status | ||
758 | * | ||
759 | * Doesn't wake up the process waiting on the struct gfs2_holder (if any) | ||
760 | * Doesn't drop the reference on the glock the top half took out | ||
761 | * | ||
762 | */ | ||
763 | |||
764 | static void drop_bh(struct gfs2_glock *gl, unsigned int ret) | ||
765 | { | ||
766 | struct gfs2_sbd *sdp = gl->gl_sbd; | ||
767 | struct gfs2_holder *gh = gl->gl_req_gh; | ||
768 | |||
769 | gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); | ||
770 | gfs2_assert_warn(sdp, list_empty(&gl->gl_holders)); | ||
771 | gfs2_assert_warn(sdp, !ret); | ||
772 | |||
773 | state_change(gl, LM_ST_UNLOCKED); | ||
774 | |||
775 | if (test_and_clear_bit(GLF_CONV_DEADLK, &gl->gl_flags)) { | ||
776 | spin_lock(&gl->gl_spin); | ||
777 | gh->gh_error = 0; | ||
778 | spin_unlock(&gl->gl_spin); | ||
779 | gfs2_glock_xmote_th(gl, gl->gl_req_gh); | ||
780 | gfs2_glock_put(gl); | ||
781 | return; | ||
782 | } | ||
783 | |||
784 | spin_lock(&gl->gl_spin); | ||
785 | gfs2_demote_wake(gl); | ||
786 | clear_bit(GLF_LOCK, &gl->gl_flags); | ||
787 | spin_unlock(&gl->gl_spin); | ||
788 | gfs2_glock_put(gl); | ||
789 | } | ||
790 | |||
791 | /** | ||
792 | * xmote_bh - Called after the lock module is done acquiring a lock | ||
793 | * @gl: The glock in question | ||
794 | * @ret: the int returned from the lock module | ||
795 | * | ||
796 | */ | ||
797 | |||
798 | static void xmote_bh(struct gfs2_glock *gl, unsigned int ret) | ||
799 | { | ||
800 | struct gfs2_sbd *sdp = gl->gl_sbd; | ||
801 | const struct gfs2_glock_operations *glops = gl->gl_ops; | ||
802 | struct gfs2_holder *gh = gl->gl_req_gh; | ||
803 | int op_done = 1; | ||
804 | |||
805 | if (!gh && (ret & LM_OUT_ST_MASK) == LM_ST_UNLOCKED) { | ||
806 | drop_bh(gl, ret); | ||
807 | return; | ||
808 | } | ||
809 | |||
810 | gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); | ||
811 | gfs2_assert_warn(sdp, list_empty(&gl->gl_holders)); | ||
812 | gfs2_assert_warn(sdp, !(ret & LM_OUT_ASYNC)); | ||
813 | |||
814 | state_change(gl, ret & LM_OUT_ST_MASK); | ||
815 | |||
816 | /* Deal with each possible exit condition */ | ||
817 | |||
818 | if (!gh) { | ||
819 | gl->gl_stamp = jiffies; | ||
820 | if (ret & LM_OUT_CANCELED) { | ||
821 | op_done = 0; | ||
822 | } else { | ||
823 | spin_lock(&gl->gl_spin); | ||
824 | if (gl->gl_state != gl->gl_demote_state) { | ||
825 | spin_unlock(&gl->gl_spin); | ||
826 | gfs2_glock_drop_th(gl); | ||
827 | gfs2_glock_put(gl); | ||
828 | return; | ||
829 | } | ||
830 | gfs2_demote_wake(gl); | ||
831 | spin_unlock(&gl->gl_spin); | ||
832 | } | ||
833 | } else { | ||
834 | spin_lock(&gl->gl_spin); | ||
835 | if (ret & LM_OUT_CONV_DEADLK) { | ||
836 | gh->gh_error = 0; | ||
837 | set_bit(GLF_CONV_DEADLK, &gl->gl_flags); | ||
838 | spin_unlock(&gl->gl_spin); | ||
839 | gfs2_glock_drop_th(gl); | ||
840 | gfs2_glock_put(gl); | ||
841 | return; | ||
842 | } | ||
843 | list_del_init(&gh->gh_list); | ||
844 | gh->gh_error = -EIO; | ||
845 | if (unlikely(test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) | ||
846 | goto out; | ||
847 | gh->gh_error = GLR_CANCELED; | ||
848 | if (ret & LM_OUT_CANCELED) | ||
849 | goto out; | ||
850 | if (relaxed_state_ok(gl->gl_state, gh->gh_state, gh->gh_flags)) { | ||
851 | list_add_tail(&gh->gh_list, &gl->gl_holders); | ||
852 | gh->gh_error = 0; | ||
853 | set_bit(HIF_HOLDER, &gh->gh_iflags); | ||
854 | set_bit(HIF_FIRST, &gh->gh_iflags); | ||
855 | op_done = 0; | ||
856 | goto out; | ||
857 | } | ||
858 | gh->gh_error = GLR_TRYFAILED; | ||
859 | if (gh->gh_flags & (LM_FLAG_TRY | LM_FLAG_TRY_1CB)) | ||
860 | goto out; | ||
861 | gh->gh_error = -EINVAL; | ||
862 | if (gfs2_assert_withdraw(sdp, 0) == -1) | ||
863 | fs_err(sdp, "ret = 0x%.8X\n", ret); | ||
864 | out: | ||
865 | spin_unlock(&gl->gl_spin); | ||
866 | } | ||
867 | |||
868 | if (glops->go_xmote_bh) | ||
869 | glops->go_xmote_bh(gl); | ||
870 | |||
871 | if (op_done) { | ||
872 | spin_lock(&gl->gl_spin); | ||
873 | gl->gl_req_gh = NULL; | ||
874 | clear_bit(GLF_LOCK, &gl->gl_flags); | ||
875 | spin_unlock(&gl->gl_spin); | ||
876 | } | ||
877 | |||
878 | gfs2_glock_put(gl); | ||
879 | |||
880 | if (gh) | ||
881 | gfs2_holder_wake(gh); | ||
882 | } | ||
883 | |||
884 | static unsigned int gfs2_lm_lock(struct gfs2_sbd *sdp, void *lock, | ||
885 | unsigned int cur_state, unsigned int req_state, | ||
886 | unsigned int flags) | ||
887 | { | ||
888 | int ret = 0; | ||
889 | if (likely(!test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) | ||
890 | ret = sdp->sd_lockstruct.ls_ops->lm_lock(lock, cur_state, | ||
891 | req_state, flags); | ||
892 | return ret; | ||
893 | } | ||
894 | |||
895 | /** | ||
896 | * gfs2_glock_xmote_th - Call into the lock module to acquire or change a glock | ||
897 | * @gl: The glock in question | ||
898 | * @state: the requested state | ||
899 | * @flags: modifier flags to the lock call | ||
900 | * | ||
901 | */ | ||
902 | |||
903 | static void gfs2_glock_xmote_th(struct gfs2_glock *gl, struct gfs2_holder *gh) | ||
904 | { | ||
905 | struct gfs2_sbd *sdp = gl->gl_sbd; | ||
906 | int flags = gh ? gh->gh_flags : 0; | ||
907 | unsigned state = gh ? gh->gh_state : gl->gl_demote_state; | ||
908 | const struct gfs2_glock_operations *glops = gl->gl_ops; | ||
909 | int lck_flags = flags & (LM_FLAG_TRY | LM_FLAG_TRY_1CB | | ||
910 | LM_FLAG_NOEXP | LM_FLAG_ANY | | ||
911 | LM_FLAG_PRIORITY); | ||
912 | unsigned int lck_ret; | ||
913 | |||
914 | if (glops->go_xmote_th) | ||
915 | glops->go_xmote_th(gl); | ||
916 | if (state == LM_ST_DEFERRED && glops->go_inval) | ||
917 | glops->go_inval(gl, DIO_METADATA); | ||
918 | |||
919 | gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); | ||
920 | gfs2_assert_warn(sdp, list_empty(&gl->gl_holders)); | ||
921 | gfs2_assert_warn(sdp, state != LM_ST_UNLOCKED); | ||
922 | gfs2_assert_warn(sdp, state != gl->gl_state); | ||
923 | |||
924 | gfs2_glock_hold(gl); | ||
925 | |||
926 | lck_ret = gfs2_lm_lock(sdp, gl->gl_lock, gl->gl_state, state, lck_flags); | ||
927 | |||
928 | if (gfs2_assert_withdraw(sdp, !(lck_ret & LM_OUT_ERROR))) | ||
929 | return; | ||
930 | |||
931 | if (lck_ret & LM_OUT_ASYNC) | ||
932 | gfs2_assert_warn(sdp, lck_ret == LM_OUT_ASYNC); | ||
933 | else | ||
934 | xmote_bh(gl, lck_ret); | ||
935 | } | ||
936 | |||
937 | static unsigned int gfs2_lm_unlock(struct gfs2_sbd *sdp, void *lock, | ||
938 | unsigned int cur_state) | ||
939 | { | ||
940 | int ret = 0; | ||
941 | if (likely(!test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) | ||
942 | ret = sdp->sd_lockstruct.ls_ops->lm_unlock(lock, cur_state); | ||
943 | return ret; | ||
944 | } | ||
945 | |||
946 | /** | ||
947 | * gfs2_glock_drop_th - call into the lock module to unlock a lock | ||
948 | * @gl: the glock | ||
949 | * | ||
950 | */ | ||
951 | |||
952 | static void gfs2_glock_drop_th(struct gfs2_glock *gl) | ||
953 | { | ||
954 | struct gfs2_sbd *sdp = gl->gl_sbd; | ||
955 | const struct gfs2_glock_operations *glops = gl->gl_ops; | ||
956 | unsigned int ret; | ||
957 | |||
958 | if (glops->go_xmote_th) | ||
959 | glops->go_xmote_th(gl); | ||
960 | if (glops->go_inval) | ||
961 | glops->go_inval(gl, DIO_METADATA); | ||
962 | |||
963 | gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); | ||
964 | gfs2_assert_warn(sdp, list_empty(&gl->gl_holders)); | ||
965 | gfs2_assert_warn(sdp, gl->gl_state != LM_ST_UNLOCKED); | ||
966 | |||
967 | gfs2_glock_hold(gl); | ||
968 | |||
969 | ret = gfs2_lm_unlock(sdp, gl->gl_lock, gl->gl_state); | ||
970 | |||
971 | if (gfs2_assert_withdraw(sdp, !(ret & LM_OUT_ERROR))) | ||
972 | return; | ||
973 | |||
974 | if (!ret) | ||
975 | drop_bh(gl, ret); | ||
976 | else | ||
977 | gfs2_assert_warn(sdp, ret == LM_OUT_ASYNC); | ||
978 | } | ||
979 | |||
980 | /** | ||
981 | * do_cancels - cancel requests for locks stuck waiting on an expire flag | ||
982 | * @gh: the LM_FLAG_PRIORITY holder waiting to acquire the lock | ||
983 | * | ||
984 | * Don't cancel GL_NOCANCEL requests. | ||
985 | */ | ||
986 | |||
987 | static void do_cancels(struct gfs2_holder *gh) | ||
988 | { | ||
989 | struct gfs2_glock *gl = gh->gh_gl; | ||
990 | struct gfs2_sbd *sdp = gl->gl_sbd; | ||
991 | |||
992 | spin_lock(&gl->gl_spin); | ||
993 | |||
994 | while (gl->gl_req_gh != gh && | ||
995 | !test_bit(HIF_HOLDER, &gh->gh_iflags) && | ||
996 | !list_empty(&gh->gh_list)) { | ||
997 | if (!(gl->gl_req_gh && (gl->gl_req_gh->gh_flags & GL_NOCANCEL))) { | ||
998 | spin_unlock(&gl->gl_spin); | ||
999 | if (likely(!test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) | ||
1000 | sdp->sd_lockstruct.ls_ops->lm_cancel(gl->gl_lock); | ||
1001 | msleep(100); | ||
1002 | spin_lock(&gl->gl_spin); | ||
1003 | } else { | ||
1004 | spin_unlock(&gl->gl_spin); | ||
1005 | msleep(100); | ||
1006 | spin_lock(&gl->gl_spin); | ||
1007 | } | ||
1008 | } | ||
1009 | |||
1010 | spin_unlock(&gl->gl_spin); | ||
1011 | } | ||
1012 | |||
1013 | /** | ||
1014 | * glock_wait_internal - wait on a glock acquisition | ||
1015 | * @gh: the glock holder | 829 | * @gh: the glock holder |
1016 | * | 830 | * |
1017 | * Returns: 0 on success | 831 | * Returns: 0 on success |
1018 | */ | 832 | */ |
1019 | 833 | ||
1020 | static int glock_wait_internal(struct gfs2_holder *gh) | 834 | int gfs2_glock_wait(struct gfs2_holder *gh) |
1021 | { | 835 | { |
1022 | struct gfs2_glock *gl = gh->gh_gl; | ||
1023 | struct gfs2_sbd *sdp = gl->gl_sbd; | ||
1024 | const struct gfs2_glock_operations *glops = gl->gl_ops; | ||
1025 | |||
1026 | if (test_bit(HIF_ABORTED, &gh->gh_iflags)) | ||
1027 | return -EIO; | ||
1028 | |||
1029 | if (gh->gh_flags & (LM_FLAG_TRY | LM_FLAG_TRY_1CB)) { | ||
1030 | spin_lock(&gl->gl_spin); | ||
1031 | if (gl->gl_req_gh != gh && | ||
1032 | !test_bit(HIF_HOLDER, &gh->gh_iflags) && | ||
1033 | !list_empty(&gh->gh_list)) { | ||
1034 | list_del_init(&gh->gh_list); | ||
1035 | gh->gh_error = GLR_TRYFAILED; | ||
1036 | run_queue(gl); | ||
1037 | spin_unlock(&gl->gl_spin); | ||
1038 | return gh->gh_error; | ||
1039 | } | ||
1040 | spin_unlock(&gl->gl_spin); | ||
1041 | } | ||
1042 | |||
1043 | if (gh->gh_flags & LM_FLAG_PRIORITY) | ||
1044 | do_cancels(gh); | ||
1045 | |||
1046 | wait_on_holder(gh); | 836 | wait_on_holder(gh); |
1047 | if (gh->gh_error) | ||
1048 | return gh->gh_error; | ||
1049 | |||
1050 | gfs2_assert_withdraw(sdp, test_bit(HIF_HOLDER, &gh->gh_iflags)); | ||
1051 | gfs2_assert_withdraw(sdp, relaxed_state_ok(gl->gl_state, gh->gh_state, | ||
1052 | gh->gh_flags)); | ||
1053 | |||
1054 | if (test_bit(HIF_FIRST, &gh->gh_iflags)) { | ||
1055 | gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); | ||
1056 | |||
1057 | if (glops->go_lock) { | ||
1058 | gh->gh_error = glops->go_lock(gh); | ||
1059 | if (gh->gh_error) { | ||
1060 | spin_lock(&gl->gl_spin); | ||
1061 | list_del_init(&gh->gh_list); | ||
1062 | spin_unlock(&gl->gl_spin); | ||
1063 | } | ||
1064 | } | ||
1065 | |||
1066 | spin_lock(&gl->gl_spin); | ||
1067 | gl->gl_req_gh = NULL; | ||
1068 | clear_bit(GLF_LOCK, &gl->gl_flags); | ||
1069 | run_queue(gl); | ||
1070 | spin_unlock(&gl->gl_spin); | ||
1071 | } | ||
1072 | |||
1073 | return gh->gh_error; | 837 | return gh->gh_error; |
1074 | } | 838 | } |
1075 | 839 | ||
1076 | static inline struct gfs2_holder * | 840 | void gfs2_print_dbg(struct seq_file *seq, const char *fmt, ...) |
1077 | find_holder_by_owner(struct list_head *head, struct pid *pid) | ||
1078 | { | ||
1079 | struct gfs2_holder *gh; | ||
1080 | |||
1081 | list_for_each_entry(gh, head, gh_list) { | ||
1082 | if (gh->gh_owner_pid == pid) | ||
1083 | return gh; | ||
1084 | } | ||
1085 | |||
1086 | return NULL; | ||
1087 | } | ||
1088 | |||
1089 | static void print_dbg(struct glock_iter *gi, const char *fmt, ...) | ||
1090 | { | 841 | { |
1091 | va_list args; | 842 | va_list args; |
1092 | 843 | ||
1093 | va_start(args, fmt); | 844 | va_start(args, fmt); |
1094 | if (gi) { | 845 | if (seq) { |
846 | struct gfs2_glock_iter *gi = seq->private; | ||
1095 | vsprintf(gi->string, fmt, args); | 847 | vsprintf(gi->string, fmt, args); |
1096 | seq_printf(gi->seq, gi->string); | 848 | seq_printf(seq, gi->string); |
1097 | } | 849 | } else { |
1098 | else | 850 | printk(KERN_ERR " "); |
1099 | vprintk(fmt, args); | 851 | vprintk(fmt, args); |
852 | } | ||
1100 | va_end(args); | 853 | va_end(args); |
1101 | } | 854 | } |
1102 | 855 | ||
@@ -1104,50 +857,75 @@ static void print_dbg(struct glock_iter *gi, const char *fmt, ...) | |||
1104 | * add_to_queue - Add a holder to the wait queue (but look for recursion) | 857 | * add_to_queue - Add a holder to the wait queue (but look for recursion) |
1105 | * @gh: the holder structure to add | 858 | * @gh: the holder structure to add |
1106 | * | 859 | * |
860 | * Eventually we should move the recursive locking trap to a | ||
861 | * debugging option or something like that. This is the fast | ||
862 | * path and needs to have the minimum number of distractions. | ||
863 | * | ||
1107 | */ | 864 | */ |
1108 | 865 | ||
1109 | static void add_to_queue(struct gfs2_holder *gh) | 866 | static inline void add_to_queue(struct gfs2_holder *gh) |
1110 | { | 867 | { |
1111 | struct gfs2_glock *gl = gh->gh_gl; | 868 | struct gfs2_glock *gl = gh->gh_gl; |
1112 | struct gfs2_holder *existing; | 869 | struct gfs2_sbd *sdp = gl->gl_sbd; |
870 | struct list_head *insert_pt = NULL; | ||
871 | struct gfs2_holder *gh2; | ||
872 | int try_lock = 0; | ||
1113 | 873 | ||
1114 | BUG_ON(gh->gh_owner_pid == NULL); | 874 | BUG_ON(gh->gh_owner_pid == NULL); |
1115 | if (test_and_set_bit(HIF_WAIT, &gh->gh_iflags)) | 875 | if (test_and_set_bit(HIF_WAIT, &gh->gh_iflags)) |
1116 | BUG(); | 876 | BUG(); |
1117 | 877 | ||
1118 | if (!(gh->gh_flags & GL_FLOCK)) { | 878 | if (gh->gh_flags & (LM_FLAG_TRY | LM_FLAG_TRY_1CB)) { |
1119 | existing = find_holder_by_owner(&gl->gl_holders, | 879 | if (test_bit(GLF_LOCK, &gl->gl_flags)) |
1120 | gh->gh_owner_pid); | 880 | try_lock = 1; |
1121 | if (existing) { | 881 | if (test_bit(GLF_INVALIDATE_IN_PROGRESS, &gl->gl_flags)) |
1122 | print_symbol(KERN_WARNING "original: %s\n", | 882 | goto fail; |
1123 | existing->gh_ip); | 883 | } |
1124 | printk(KERN_INFO "pid : %d\n", | 884 | |
1125 | pid_nr(existing->gh_owner_pid)); | 885 | list_for_each_entry(gh2, &gl->gl_holders, gh_list) { |
1126 | printk(KERN_INFO "lock type : %d lock state : %d\n", | 886 | if (unlikely(gh2->gh_owner_pid == gh->gh_owner_pid && |
1127 | existing->gh_gl->gl_name.ln_type, | 887 | (gh->gh_gl->gl_ops->go_type != LM_TYPE_FLOCK))) |
1128 | existing->gh_gl->gl_state); | 888 | goto trap_recursive; |
1129 | print_symbol(KERN_WARNING "new: %s\n", gh->gh_ip); | 889 | if (try_lock && |
1130 | printk(KERN_INFO "pid : %d\n", | 890 | !(gh2->gh_flags & (LM_FLAG_TRY | LM_FLAG_TRY_1CB)) && |
1131 | pid_nr(gh->gh_owner_pid)); | 891 | !may_grant(gl, gh)) { |
1132 | printk(KERN_INFO "lock type : %d lock state : %d\n", | 892 | fail: |
1133 | gl->gl_name.ln_type, gl->gl_state); | 893 | gh->gh_error = GLR_TRYFAILED; |
1134 | BUG(); | 894 | gfs2_holder_wake(gh); |
1135 | } | 895 | return; |
1136 | |||
1137 | existing = find_holder_by_owner(&gl->gl_waiters3, | ||
1138 | gh->gh_owner_pid); | ||
1139 | if (existing) { | ||
1140 | print_symbol(KERN_WARNING "original: %s\n", | ||
1141 | existing->gh_ip); | ||
1142 | print_symbol(KERN_WARNING "new: %s\n", gh->gh_ip); | ||
1143 | BUG(); | ||
1144 | } | 896 | } |
897 | if (test_bit(HIF_HOLDER, &gh2->gh_iflags)) | ||
898 | continue; | ||
899 | if (unlikely((gh->gh_flags & LM_FLAG_PRIORITY) && !insert_pt)) | ||
900 | insert_pt = &gh2->gh_list; | ||
901 | } | ||
902 | if (likely(insert_pt == NULL)) { | ||
903 | list_add_tail(&gh->gh_list, &gl->gl_holders); | ||
904 | if (unlikely(gh->gh_flags & LM_FLAG_PRIORITY)) | ||
905 | goto do_cancel; | ||
906 | return; | ||
1145 | } | 907 | } |
908 | list_add_tail(&gh->gh_list, insert_pt); | ||
909 | do_cancel: | ||
910 | gh = list_entry(gl->gl_holders.next, struct gfs2_holder, gh_list); | ||
911 | if (!(gh->gh_flags & LM_FLAG_PRIORITY)) { | ||
912 | spin_unlock(&gl->gl_spin); | ||
913 | sdp->sd_lockstruct.ls_ops->lm_cancel(gl->gl_lock); | ||
914 | spin_lock(&gl->gl_spin); | ||
915 | } | ||
916 | return; | ||
1146 | 917 | ||
1147 | if (gh->gh_flags & LM_FLAG_PRIORITY) | 918 | trap_recursive: |
1148 | list_add(&gh->gh_list, &gl->gl_waiters3); | 919 | print_symbol(KERN_ERR "original: %s\n", gh2->gh_ip); |
1149 | else | 920 | printk(KERN_ERR "pid: %d\n", pid_nr(gh2->gh_owner_pid)); |
1150 | list_add_tail(&gh->gh_list, &gl->gl_waiters3); | 921 | printk(KERN_ERR "lock type: %d req lock state : %d\n", |
922 | gh2->gh_gl->gl_name.ln_type, gh2->gh_state); | ||
923 | print_symbol(KERN_ERR "new: %s\n", gh->gh_ip); | ||
924 | printk(KERN_ERR "pid: %d\n", pid_nr(gh->gh_owner_pid)); | ||
925 | printk(KERN_ERR "lock type: %d req lock state : %d\n", | ||
926 | gh->gh_gl->gl_name.ln_type, gh->gh_state); | ||
927 | __dump_glock(NULL, gl); | ||
928 | BUG(); | ||
1151 | } | 929 | } |
1152 | 930 | ||
1153 | /** | 931 | /** |
@@ -1165,24 +943,16 @@ int gfs2_glock_nq(struct gfs2_holder *gh) | |||
1165 | struct gfs2_sbd *sdp = gl->gl_sbd; | 943 | struct gfs2_sbd *sdp = gl->gl_sbd; |
1166 | int error = 0; | 944 | int error = 0; |
1167 | 945 | ||
1168 | restart: | 946 | if (unlikely(test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) |
1169 | if (unlikely(test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) { | ||
1170 | set_bit(HIF_ABORTED, &gh->gh_iflags); | ||
1171 | return -EIO; | 947 | return -EIO; |
1172 | } | ||
1173 | 948 | ||
1174 | spin_lock(&gl->gl_spin); | 949 | spin_lock(&gl->gl_spin); |
1175 | add_to_queue(gh); | 950 | add_to_queue(gh); |
1176 | run_queue(gl); | 951 | run_queue(gl, 1); |
1177 | spin_unlock(&gl->gl_spin); | 952 | spin_unlock(&gl->gl_spin); |
1178 | 953 | ||
1179 | if (!(gh->gh_flags & GL_ASYNC)) { | 954 | if (!(gh->gh_flags & GL_ASYNC)) |
1180 | error = glock_wait_internal(gh); | 955 | error = gfs2_glock_wait(gh); |
1181 | if (error == GLR_CANCELED) { | ||
1182 | msleep(100); | ||
1183 | goto restart; | ||
1184 | } | ||
1185 | } | ||
1186 | 956 | ||
1187 | return error; | 957 | return error; |
1188 | } | 958 | } |
@@ -1196,48 +966,7 @@ restart: | |||
1196 | 966 | ||
1197 | int gfs2_glock_poll(struct gfs2_holder *gh) | 967 | int gfs2_glock_poll(struct gfs2_holder *gh) |
1198 | { | 968 | { |
1199 | struct gfs2_glock *gl = gh->gh_gl; | 969 | return test_bit(HIF_WAIT, &gh->gh_iflags) ? 0 : 1; |
1200 | int ready = 0; | ||
1201 | |||
1202 | spin_lock(&gl->gl_spin); | ||
1203 | |||
1204 | if (test_bit(HIF_HOLDER, &gh->gh_iflags)) | ||
1205 | ready = 1; | ||
1206 | else if (list_empty(&gh->gh_list)) { | ||
1207 | if (gh->gh_error == GLR_CANCELED) { | ||
1208 | spin_unlock(&gl->gl_spin); | ||
1209 | msleep(100); | ||
1210 | if (gfs2_glock_nq(gh)) | ||
1211 | return 1; | ||
1212 | return 0; | ||
1213 | } else | ||
1214 | ready = 1; | ||
1215 | } | ||
1216 | |||
1217 | spin_unlock(&gl->gl_spin); | ||
1218 | |||
1219 | return ready; | ||
1220 | } | ||
1221 | |||
1222 | /** | ||
1223 | * gfs2_glock_wait - wait for a lock acquisition that ended in a GLR_ASYNC | ||
1224 | * @gh: the holder structure | ||
1225 | * | ||
1226 | * Returns: 0, GLR_TRYFAILED, or errno on failure | ||
1227 | */ | ||
1228 | |||
1229 | int gfs2_glock_wait(struct gfs2_holder *gh) | ||
1230 | { | ||
1231 | int error; | ||
1232 | |||
1233 | error = glock_wait_internal(gh); | ||
1234 | if (error == GLR_CANCELED) { | ||
1235 | msleep(100); | ||
1236 | gh->gh_flags &= ~GL_ASYNC; | ||
1237 | error = gfs2_glock_nq(gh); | ||
1238 | } | ||
1239 | |||
1240 | return error; | ||
1241 | } | 970 | } |
1242 | 971 | ||
1243 | /** | 972 | /** |
@@ -1251,26 +980,30 @@ void gfs2_glock_dq(struct gfs2_holder *gh) | |||
1251 | struct gfs2_glock *gl = gh->gh_gl; | 980 | struct gfs2_glock *gl = gh->gh_gl; |
1252 | const struct gfs2_glock_operations *glops = gl->gl_ops; | 981 | const struct gfs2_glock_operations *glops = gl->gl_ops; |
1253 | unsigned delay = 0; | 982 | unsigned delay = 0; |
983 | int fast_path = 0; | ||
1254 | 984 | ||
985 | spin_lock(&gl->gl_spin); | ||
1255 | if (gh->gh_flags & GL_NOCACHE) | 986 | if (gh->gh_flags & GL_NOCACHE) |
1256 | handle_callback(gl, LM_ST_UNLOCKED, 0, 0); | 987 | handle_callback(gl, LM_ST_UNLOCKED, 0, 0); |
1257 | 988 | ||
1258 | gfs2_glmutex_lock(gl); | ||
1259 | |||
1260 | spin_lock(&gl->gl_spin); | ||
1261 | list_del_init(&gh->gh_list); | 989 | list_del_init(&gh->gh_list); |
1262 | 990 | if (find_first_holder(gl) == NULL) { | |
1263 | if (list_empty(&gl->gl_holders)) { | ||
1264 | if (glops->go_unlock) { | 991 | if (glops->go_unlock) { |
992 | GLOCK_BUG_ON(gl, test_and_set_bit(GLF_LOCK, &gl->gl_flags)); | ||
1265 | spin_unlock(&gl->gl_spin); | 993 | spin_unlock(&gl->gl_spin); |
1266 | glops->go_unlock(gh); | 994 | glops->go_unlock(gh); |
1267 | spin_lock(&gl->gl_spin); | 995 | spin_lock(&gl->gl_spin); |
996 | clear_bit(GLF_LOCK, &gl->gl_flags); | ||
1268 | } | 997 | } |
1269 | gl->gl_stamp = jiffies; | 998 | gl->gl_stamp = jiffies; |
999 | if (list_empty(&gl->gl_holders) && | ||
1000 | !test_bit(GLF_PENDING_DEMOTE, &gl->gl_flags) && | ||
1001 | !test_bit(GLF_DEMOTE, &gl->gl_flags)) | ||
1002 | fast_path = 1; | ||
1270 | } | 1003 | } |
1271 | |||
1272 | clear_bit(GLF_LOCK, &gl->gl_flags); | ||
1273 | spin_unlock(&gl->gl_spin); | 1004 | spin_unlock(&gl->gl_spin); |
1005 | if (likely(fast_path)) | ||
1006 | return; | ||
1274 | 1007 | ||
1275 | gfs2_glock_hold(gl); | 1008 | gfs2_glock_hold(gl); |
1276 | if (test_bit(GLF_PENDING_DEMOTE, &gl->gl_flags) && | 1009 | if (test_bit(GLF_PENDING_DEMOTE, &gl->gl_flags) && |
@@ -1469,20 +1202,14 @@ int gfs2_lvb_hold(struct gfs2_glock *gl) | |||
1469 | { | 1202 | { |
1470 | int error; | 1203 | int error; |
1471 | 1204 | ||
1472 | gfs2_glmutex_lock(gl); | ||
1473 | |||
1474 | if (!atomic_read(&gl->gl_lvb_count)) { | 1205 | if (!atomic_read(&gl->gl_lvb_count)) { |
1475 | error = gfs2_lm_hold_lvb(gl->gl_sbd, gl->gl_lock, &gl->gl_lvb); | 1206 | error = gfs2_lm_hold_lvb(gl->gl_sbd, gl->gl_lock, &gl->gl_lvb); |
1476 | if (error) { | 1207 | if (error) |
1477 | gfs2_glmutex_unlock(gl); | ||
1478 | return error; | 1208 | return error; |
1479 | } | ||
1480 | gfs2_glock_hold(gl); | 1209 | gfs2_glock_hold(gl); |
1481 | } | 1210 | } |
1482 | atomic_inc(&gl->gl_lvb_count); | 1211 | atomic_inc(&gl->gl_lvb_count); |
1483 | 1212 | ||
1484 | gfs2_glmutex_unlock(gl); | ||
1485 | |||
1486 | return 0; | 1213 | return 0; |
1487 | } | 1214 | } |
1488 | 1215 | ||
@@ -1497,8 +1224,6 @@ void gfs2_lvb_unhold(struct gfs2_glock *gl) | |||
1497 | struct gfs2_sbd *sdp = gl->gl_sbd; | 1224 | struct gfs2_sbd *sdp = gl->gl_sbd; |
1498 | 1225 | ||
1499 | gfs2_glock_hold(gl); | 1226 | gfs2_glock_hold(gl); |
1500 | gfs2_glmutex_lock(gl); | ||
1501 | |||
1502 | gfs2_assert(gl->gl_sbd, atomic_read(&gl->gl_lvb_count) > 0); | 1227 | gfs2_assert(gl->gl_sbd, atomic_read(&gl->gl_lvb_count) > 0); |
1503 | if (atomic_dec_and_test(&gl->gl_lvb_count)) { | 1228 | if (atomic_dec_and_test(&gl->gl_lvb_count)) { |
1504 | if (likely(!test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) | 1229 | if (likely(!test_bit(SDF_SHUTDOWN, &sdp->sd_flags))) |
@@ -1506,8 +1231,6 @@ void gfs2_lvb_unhold(struct gfs2_glock *gl) | |||
1506 | gl->gl_lvb = NULL; | 1231 | gl->gl_lvb = NULL; |
1507 | gfs2_glock_put(gl); | 1232 | gfs2_glock_put(gl); |
1508 | } | 1233 | } |
1509 | |||
1510 | gfs2_glmutex_unlock(gl); | ||
1511 | gfs2_glock_put(gl); | 1234 | gfs2_glock_put(gl); |
1512 | } | 1235 | } |
1513 | 1236 | ||
@@ -1527,7 +1250,9 @@ static void blocking_cb(struct gfs2_sbd *sdp, struct lm_lockname *name, | |||
1527 | if (time_before(now, holdtime)) | 1250 | if (time_before(now, holdtime)) |
1528 | delay = holdtime - now; | 1251 | delay = holdtime - now; |
1529 | 1252 | ||
1253 | spin_lock(&gl->gl_spin); | ||
1530 | handle_callback(gl, state, 1, delay); | 1254 | handle_callback(gl, state, 1, delay); |
1255 | spin_unlock(&gl->gl_spin); | ||
1531 | if (queue_delayed_work(glock_workqueue, &gl->gl_work, delay) == 0) | 1256 | if (queue_delayed_work(glock_workqueue, &gl->gl_work, delay) == 0) |
1532 | gfs2_glock_put(gl); | 1257 | gfs2_glock_put(gl); |
1533 | } | 1258 | } |
@@ -1568,7 +1293,8 @@ void gfs2_glock_cb(void *cb_data, unsigned int type, void *data) | |||
1568 | gl = gfs2_glock_find(sdp, &async->lc_name); | 1293 | gl = gfs2_glock_find(sdp, &async->lc_name); |
1569 | if (gfs2_assert_warn(sdp, gl)) | 1294 | if (gfs2_assert_warn(sdp, gl)) |
1570 | return; | 1295 | return; |
1571 | xmote_bh(gl, async->lc_ret); | 1296 | gl->gl_reply = async->lc_ret; |
1297 | set_bit(GLF_REPLY_PENDING, &gl->gl_flags); | ||
1572 | if (queue_delayed_work(glock_workqueue, &gl->gl_work, 0) == 0) | 1298 | if (queue_delayed_work(glock_workqueue, &gl->gl_work, 0) == 0) |
1573 | gfs2_glock_put(gl); | 1299 | gfs2_glock_put(gl); |
1574 | up_read(&gfs2_umount_flush_sem); | 1300 | up_read(&gfs2_umount_flush_sem); |
@@ -1646,6 +1372,7 @@ void gfs2_glock_schedule_for_reclaim(struct gfs2_glock *gl) | |||
1646 | void gfs2_reclaim_glock(struct gfs2_sbd *sdp) | 1372 | void gfs2_reclaim_glock(struct gfs2_sbd *sdp) |
1647 | { | 1373 | { |
1648 | struct gfs2_glock *gl; | 1374 | struct gfs2_glock *gl; |
1375 | int done_callback = 0; | ||
1649 | 1376 | ||
1650 | spin_lock(&sdp->sd_reclaim_lock); | 1377 | spin_lock(&sdp->sd_reclaim_lock); |
1651 | if (list_empty(&sdp->sd_reclaim_list)) { | 1378 | if (list_empty(&sdp->sd_reclaim_list)) { |
@@ -1660,14 +1387,16 @@ void gfs2_reclaim_glock(struct gfs2_sbd *sdp) | |||
1660 | atomic_dec(&sdp->sd_reclaim_count); | 1387 | atomic_dec(&sdp->sd_reclaim_count); |
1661 | atomic_inc(&sdp->sd_reclaimed); | 1388 | atomic_inc(&sdp->sd_reclaimed); |
1662 | 1389 | ||
1663 | if (gfs2_glmutex_trylock(gl)) { | 1390 | spin_lock(&gl->gl_spin); |
1664 | if (list_empty(&gl->gl_holders) && | 1391 | if (find_first_holder(gl) == NULL && |
1665 | gl->gl_state != LM_ST_UNLOCKED && demote_ok(gl)) | 1392 | gl->gl_state != LM_ST_UNLOCKED && demote_ok(gl)) { |
1666 | handle_callback(gl, LM_ST_UNLOCKED, 0, 0); | 1393 | handle_callback(gl, LM_ST_UNLOCKED, 0, 0); |
1667 | gfs2_glmutex_unlock(gl); | 1394 | done_callback = 1; |
1668 | } | 1395 | } |
1669 | 1396 | spin_unlock(&gl->gl_spin); | |
1670 | gfs2_glock_put(gl); | 1397 | if (!done_callback || |
1398 | queue_delayed_work(glock_workqueue, &gl->gl_work, 0) == 0) | ||
1399 | gfs2_glock_put(gl); | ||
1671 | } | 1400 | } |
1672 | 1401 | ||
1673 | /** | 1402 | /** |
@@ -1724,18 +1453,14 @@ static void scan_glock(struct gfs2_glock *gl) | |||
1724 | { | 1453 | { |
1725 | if (gl->gl_ops == &gfs2_inode_glops && gl->gl_object) | 1454 | if (gl->gl_ops == &gfs2_inode_glops && gl->gl_object) |
1726 | return; | 1455 | return; |
1456 | if (test_bit(GLF_LOCK, &gl->gl_flags)) | ||
1457 | return; | ||
1727 | 1458 | ||
1728 | if (gfs2_glmutex_trylock(gl)) { | 1459 | spin_lock(&gl->gl_spin); |
1729 | if (list_empty(&gl->gl_holders) && | 1460 | if (find_first_holder(gl) == NULL && |
1730 | gl->gl_state != LM_ST_UNLOCKED && demote_ok(gl)) | 1461 | gl->gl_state != LM_ST_UNLOCKED && demote_ok(gl)) |
1731 | goto out_schedule; | 1462 | gfs2_glock_schedule_for_reclaim(gl); |
1732 | gfs2_glmutex_unlock(gl); | 1463 | spin_unlock(&gl->gl_spin); |
1733 | } | ||
1734 | return; | ||
1735 | |||
1736 | out_schedule: | ||
1737 | gfs2_glmutex_unlock(gl); | ||
1738 | gfs2_glock_schedule_for_reclaim(gl); | ||
1739 | } | 1464 | } |
1740 | 1465 | ||
1741 | /** | 1466 | /** |
@@ -1760,12 +1485,13 @@ static void clear_glock(struct gfs2_glock *gl) | |||
1760 | spin_unlock(&sdp->sd_reclaim_lock); | 1485 | spin_unlock(&sdp->sd_reclaim_lock); |
1761 | } | 1486 | } |
1762 | 1487 | ||
1763 | if (gfs2_glmutex_trylock(gl)) { | 1488 | spin_lock(&gl->gl_spin); |
1764 | if (list_empty(&gl->gl_holders) && | 1489 | if (find_first_holder(gl) == NULL && gl->gl_state != LM_ST_UNLOCKED) |
1765 | gl->gl_state != LM_ST_UNLOCKED) | 1490 | handle_callback(gl, LM_ST_UNLOCKED, 0, 0); |
1766 | handle_callback(gl, LM_ST_UNLOCKED, 0, 0); | 1491 | spin_unlock(&gl->gl_spin); |
1767 | gfs2_glmutex_unlock(gl); | 1492 | gfs2_glock_hold(gl); |
1768 | } | 1493 | if (queue_delayed_work(glock_workqueue, &gl->gl_work, 0) == 0) |
1494 | gfs2_glock_put(gl); | ||
1769 | } | 1495 | } |
1770 | 1496 | ||
1771 | /** | 1497 | /** |
@@ -1810,180 +1536,164 @@ void gfs2_gl_hash_clear(struct gfs2_sbd *sdp, int wait) | |||
1810 | } | 1536 | } |
1811 | } | 1537 | } |
1812 | 1538 | ||
1813 | /* | 1539 | static const char *state2str(unsigned state) |
1814 | * Diagnostic routines to help debug distributed deadlock | ||
1815 | */ | ||
1816 | |||
1817 | static void gfs2_print_symbol(struct glock_iter *gi, const char *fmt, | ||
1818 | unsigned long address) | ||
1819 | { | 1540 | { |
1820 | char buffer[KSYM_SYMBOL_LEN]; | 1541 | switch(state) { |
1821 | 1542 | case LM_ST_UNLOCKED: | |
1822 | sprint_symbol(buffer, address); | 1543 | return "UN"; |
1823 | print_dbg(gi, fmt, buffer); | 1544 | case LM_ST_SHARED: |
1545 | return "SH"; | ||
1546 | case LM_ST_DEFERRED: | ||
1547 | return "DF"; | ||
1548 | case LM_ST_EXCLUSIVE: | ||
1549 | return "EX"; | ||
1550 | } | ||
1551 | return "??"; | ||
1552 | } | ||
1553 | |||
1554 | static const char *hflags2str(char *buf, unsigned flags, unsigned long iflags) | ||
1555 | { | ||
1556 | char *p = buf; | ||
1557 | if (flags & LM_FLAG_TRY) | ||
1558 | *p++ = 't'; | ||
1559 | if (flags & LM_FLAG_TRY_1CB) | ||
1560 | *p++ = 'T'; | ||
1561 | if (flags & LM_FLAG_NOEXP) | ||
1562 | *p++ = 'e'; | ||
1563 | if (flags & LM_FLAG_ANY) | ||
1564 | *p++ = 'a'; | ||
1565 | if (flags & LM_FLAG_PRIORITY) | ||
1566 | *p++ = 'p'; | ||
1567 | if (flags & GL_ASYNC) | ||
1568 | *p++ = 'a'; | ||
1569 | if (flags & GL_EXACT) | ||
1570 | *p++ = 'E'; | ||
1571 | if (flags & GL_ATIME) | ||
1572 | *p++ = 'a'; | ||
1573 | if (flags & GL_NOCACHE) | ||
1574 | *p++ = 'c'; | ||
1575 | if (test_bit(HIF_HOLDER, &iflags)) | ||
1576 | *p++ = 'H'; | ||
1577 | if (test_bit(HIF_WAIT, &iflags)) | ||
1578 | *p++ = 'W'; | ||
1579 | if (test_bit(HIF_FIRST, &iflags)) | ||
1580 | *p++ = 'F'; | ||
1581 | *p = 0; | ||
1582 | return buf; | ||
1824 | } | 1583 | } |
1825 | 1584 | ||
1826 | /** | 1585 | /** |
1827 | * dump_holder - print information about a glock holder | 1586 | * dump_holder - print information about a glock holder |
1828 | * @str: a string naming the type of holder | 1587 | * @seq: the seq_file struct |
1829 | * @gh: the glock holder | 1588 | * @gh: the glock holder |
1830 | * | 1589 | * |
1831 | * Returns: 0 on success, -ENOBUFS when we run out of space | 1590 | * Returns: 0 on success, -ENOBUFS when we run out of space |
1832 | */ | 1591 | */ |
1833 | 1592 | ||
1834 | static int dump_holder(struct glock_iter *gi, char *str, | 1593 | static int dump_holder(struct seq_file *seq, const struct gfs2_holder *gh) |
1835 | struct gfs2_holder *gh) | ||
1836 | { | 1594 | { |
1837 | unsigned int x; | 1595 | struct task_struct *gh_owner = NULL; |
1838 | struct task_struct *gh_owner; | 1596 | char buffer[KSYM_SYMBOL_LEN]; |
1597 | char flags_buf[32]; | ||
1839 | 1598 | ||
1840 | print_dbg(gi, " %s\n", str); | 1599 | sprint_symbol(buffer, gh->gh_ip); |
1841 | if (gh->gh_owner_pid) { | 1600 | if (gh->gh_owner_pid) |
1842 | print_dbg(gi, " owner = %ld ", | ||
1843 | (long)pid_nr(gh->gh_owner_pid)); | ||
1844 | gh_owner = pid_task(gh->gh_owner_pid, PIDTYPE_PID); | 1601 | gh_owner = pid_task(gh->gh_owner_pid, PIDTYPE_PID); |
1845 | if (gh_owner) | 1602 | gfs2_print_dbg(seq, " H: s:%s f:%s e:%d p:%ld [%s] %s\n", |
1846 | print_dbg(gi, "(%s)\n", gh_owner->comm); | 1603 | state2str(gh->gh_state), |
1847 | else | 1604 | hflags2str(flags_buf, gh->gh_flags, gh->gh_iflags), |
1848 | print_dbg(gi, "(ended)\n"); | 1605 | gh->gh_error, |
1849 | } else | 1606 | gh->gh_owner_pid ? (long)pid_nr(gh->gh_owner_pid) : -1, |
1850 | print_dbg(gi, " owner = -1\n"); | 1607 | gh_owner ? gh_owner->comm : "(ended)", buffer); |
1851 | print_dbg(gi, " gh_state = %u\n", gh->gh_state); | ||
1852 | print_dbg(gi, " gh_flags ="); | ||
1853 | for (x = 0; x < 32; x++) | ||
1854 | if (gh->gh_flags & (1 << x)) | ||
1855 | print_dbg(gi, " %u", x); | ||
1856 | print_dbg(gi, " \n"); | ||
1857 | print_dbg(gi, " error = %d\n", gh->gh_error); | ||
1858 | print_dbg(gi, " gh_iflags ="); | ||
1859 | for (x = 0; x < 32; x++) | ||
1860 | if (test_bit(x, &gh->gh_iflags)) | ||
1861 | print_dbg(gi, " %u", x); | ||
1862 | print_dbg(gi, " \n"); | ||
1863 | gfs2_print_symbol(gi, " initialized at: %s\n", gh->gh_ip); | ||
1864 | |||
1865 | return 0; | 1608 | return 0; |
1866 | } | 1609 | } |
1867 | 1610 | ||
1868 | /** | 1611 | static const char *gflags2str(char *buf, const unsigned long *gflags) |
1869 | * dump_inode - print information about an inode | 1612 | { |
1870 | * @ip: the inode | 1613 | char *p = buf; |
1871 | * | 1614 | if (test_bit(GLF_LOCK, gflags)) |
1872 | * Returns: 0 on success, -ENOBUFS when we run out of space | 1615 | *p++ = 'l'; |
1873 | */ | 1616 | if (test_bit(GLF_STICKY, gflags)) |
1874 | 1617 | *p++ = 's'; | |
1875 | static int dump_inode(struct glock_iter *gi, struct gfs2_inode *ip) | 1618 | if (test_bit(GLF_DEMOTE, gflags)) |
1876 | { | 1619 | *p++ = 'D'; |
1877 | unsigned int x; | 1620 | if (test_bit(GLF_PENDING_DEMOTE, gflags)) |
1878 | 1621 | *p++ = 'd'; | |
1879 | print_dbg(gi, " Inode:\n"); | 1622 | if (test_bit(GLF_DEMOTE_IN_PROGRESS, gflags)) |
1880 | print_dbg(gi, " num = %llu/%llu\n", | 1623 | *p++ = 'p'; |
1881 | (unsigned long long)ip->i_no_formal_ino, | 1624 | if (test_bit(GLF_DIRTY, gflags)) |
1882 | (unsigned long long)ip->i_no_addr); | 1625 | *p++ = 'y'; |
1883 | print_dbg(gi, " type = %u\n", IF2DT(ip->i_inode.i_mode)); | 1626 | if (test_bit(GLF_LFLUSH, gflags)) |
1884 | print_dbg(gi, " i_flags ="); | 1627 | *p++ = 'f'; |
1885 | for (x = 0; x < 32; x++) | 1628 | if (test_bit(GLF_INVALIDATE_IN_PROGRESS, gflags)) |
1886 | if (test_bit(x, &ip->i_flags)) | 1629 | *p++ = 'i'; |
1887 | print_dbg(gi, " %u", x); | 1630 | if (test_bit(GLF_REPLY_PENDING, gflags)) |
1888 | print_dbg(gi, " \n"); | 1631 | *p++ = 'r'; |
1889 | return 0; | 1632 | *p = 0; |
1633 | return buf; | ||
1890 | } | 1634 | } |
1891 | 1635 | ||
1892 | /** | 1636 | /** |
1893 | * dump_glock - print information about a glock | 1637 | * __dump_glock - print information about a glock |
1638 | * @seq: The seq_file struct | ||
1894 | * @gl: the glock | 1639 | * @gl: the glock |
1895 | * @count: where we are in the buffer | 1640 | * |
1641 | * The file format is as follows: | ||
1642 | * One line per object, capital letters are used to indicate objects | ||
1643 | * G = glock, I = Inode, R = rgrp, H = holder. Glocks are not indented, | ||
1644 | * other objects are indented by a single space and follow the glock to | ||
1645 | * which they are related. Fields are indicated by lower case letters | ||
1646 | * followed by a colon and the field value, except for strings which are in | ||
1647 | * [] so that its possible to see if they are composed of spaces for | ||
1648 | * example. The field's are n = number (id of the object), f = flags, | ||
1649 | * t = type, s = state, r = refcount, e = error, p = pid. | ||
1896 | * | 1650 | * |
1897 | * Returns: 0 on success, -ENOBUFS when we run out of space | 1651 | * Returns: 0 on success, -ENOBUFS when we run out of space |
1898 | */ | 1652 | */ |
1899 | 1653 | ||
1900 | static int dump_glock(struct glock_iter *gi, struct gfs2_glock *gl) | 1654 | static int __dump_glock(struct seq_file *seq, const struct gfs2_glock *gl) |
1901 | { | 1655 | { |
1902 | struct gfs2_holder *gh; | 1656 | const struct gfs2_glock_operations *glops = gl->gl_ops; |
1903 | unsigned int x; | 1657 | unsigned long long dtime; |
1904 | int error = -ENOBUFS; | 1658 | const struct gfs2_holder *gh; |
1905 | struct task_struct *gl_owner; | 1659 | char gflags_buf[32]; |
1660 | int error = 0; | ||
1906 | 1661 | ||
1907 | spin_lock(&gl->gl_spin); | 1662 | dtime = jiffies - gl->gl_demote_time; |
1663 | dtime *= 1000000/HZ; /* demote time in uSec */ | ||
1664 | if (!test_bit(GLF_DEMOTE, &gl->gl_flags)) | ||
1665 | dtime = 0; | ||
1666 | gfs2_print_dbg(seq, "G: s:%s n:%u/%llu f:%s t:%s d:%s/%llu l:%d a:%d r:%d\n", | ||
1667 | state2str(gl->gl_state), | ||
1668 | gl->gl_name.ln_type, | ||
1669 | (unsigned long long)gl->gl_name.ln_number, | ||
1670 | gflags2str(gflags_buf, &gl->gl_flags), | ||
1671 | state2str(gl->gl_target), | ||
1672 | state2str(gl->gl_demote_state), dtime, | ||
1673 | atomic_read(&gl->gl_lvb_count), | ||
1674 | atomic_read(&gl->gl_ail_count), | ||
1675 | atomic_read(&gl->gl_ref)); | ||
1908 | 1676 | ||
1909 | print_dbg(gi, "Glock 0x%p (%u, 0x%llx)\n", gl, gl->gl_name.ln_type, | ||
1910 | (unsigned long long)gl->gl_name.ln_number); | ||
1911 | print_dbg(gi, " gl_flags ="); | ||
1912 | for (x = 0; x < 32; x++) { | ||
1913 | if (test_bit(x, &gl->gl_flags)) | ||
1914 | print_dbg(gi, " %u", x); | ||
1915 | } | ||
1916 | if (!test_bit(GLF_LOCK, &gl->gl_flags)) | ||
1917 | print_dbg(gi, " (unlocked)"); | ||
1918 | print_dbg(gi, " \n"); | ||
1919 | print_dbg(gi, " gl_ref = %d\n", atomic_read(&gl->gl_ref)); | ||
1920 | print_dbg(gi, " gl_state = %u\n", gl->gl_state); | ||
1921 | if (gl->gl_owner_pid) { | ||
1922 | gl_owner = pid_task(gl->gl_owner_pid, PIDTYPE_PID); | ||
1923 | if (gl_owner) | ||
1924 | print_dbg(gi, " gl_owner = pid %d (%s)\n", | ||
1925 | pid_nr(gl->gl_owner_pid), gl_owner->comm); | ||
1926 | else | ||
1927 | print_dbg(gi, " gl_owner = %d (ended)\n", | ||
1928 | pid_nr(gl->gl_owner_pid)); | ||
1929 | } else | ||
1930 | print_dbg(gi, " gl_owner = -1\n"); | ||
1931 | print_dbg(gi, " gl_ip = %lu\n", gl->gl_ip); | ||
1932 | print_dbg(gi, " req_gh = %s\n", (gl->gl_req_gh) ? "yes" : "no"); | ||
1933 | print_dbg(gi, " lvb_count = %d\n", atomic_read(&gl->gl_lvb_count)); | ||
1934 | print_dbg(gi, " object = %s\n", (gl->gl_object) ? "yes" : "no"); | ||
1935 | print_dbg(gi, " reclaim = %s\n", | ||
1936 | (list_empty(&gl->gl_reclaim)) ? "no" : "yes"); | ||
1937 | if (gl->gl_aspace) | ||
1938 | print_dbg(gi, " aspace = 0x%p nrpages = %lu\n", gl->gl_aspace, | ||
1939 | gl->gl_aspace->i_mapping->nrpages); | ||
1940 | else | ||
1941 | print_dbg(gi, " aspace = no\n"); | ||
1942 | print_dbg(gi, " ail = %d\n", atomic_read(&gl->gl_ail_count)); | ||
1943 | if (gl->gl_req_gh) { | ||
1944 | error = dump_holder(gi, "Request", gl->gl_req_gh); | ||
1945 | if (error) | ||
1946 | goto out; | ||
1947 | } | ||
1948 | list_for_each_entry(gh, &gl->gl_holders, gh_list) { | 1677 | list_for_each_entry(gh, &gl->gl_holders, gh_list) { |
1949 | error = dump_holder(gi, "Holder", gh); | 1678 | error = dump_holder(seq, gh); |
1950 | if (error) | ||
1951 | goto out; | ||
1952 | } | ||
1953 | list_for_each_entry(gh, &gl->gl_waiters1, gh_list) { | ||
1954 | error = dump_holder(gi, "Waiter1", gh); | ||
1955 | if (error) | ||
1956 | goto out; | ||
1957 | } | ||
1958 | list_for_each_entry(gh, &gl->gl_waiters3, gh_list) { | ||
1959 | error = dump_holder(gi, "Waiter3", gh); | ||
1960 | if (error) | 1679 | if (error) |
1961 | goto out; | 1680 | goto out; |
1962 | } | 1681 | } |
1963 | if (test_bit(GLF_DEMOTE, &gl->gl_flags)) { | 1682 | if (gl->gl_state != LM_ST_UNLOCKED && glops->go_dump) |
1964 | print_dbg(gi, " Demotion req to state %u (%llu uS ago)\n", | 1683 | error = glops->go_dump(seq, gl); |
1965 | gl->gl_demote_state, (unsigned long long) | ||
1966 | (jiffies - gl->gl_demote_time)*(1000000/HZ)); | ||
1967 | } | ||
1968 | if (gl->gl_ops == &gfs2_inode_glops && gl->gl_object) { | ||
1969 | if (!test_bit(GLF_LOCK, &gl->gl_flags) && | ||
1970 | list_empty(&gl->gl_holders)) { | ||
1971 | error = dump_inode(gi, gl->gl_object); | ||
1972 | if (error) | ||
1973 | goto out; | ||
1974 | } else { | ||
1975 | error = -ENOBUFS; | ||
1976 | print_dbg(gi, " Inode: busy\n"); | ||
1977 | } | ||
1978 | } | ||
1979 | |||
1980 | error = 0; | ||
1981 | |||
1982 | out: | 1684 | out: |
1983 | spin_unlock(&gl->gl_spin); | ||
1984 | return error; | 1685 | return error; |
1985 | } | 1686 | } |
1986 | 1687 | ||
1688 | static int dump_glock(struct seq_file *seq, struct gfs2_glock *gl) | ||
1689 | { | ||
1690 | int ret; | ||
1691 | spin_lock(&gl->gl_spin); | ||
1692 | ret = __dump_glock(seq, gl); | ||
1693 | spin_unlock(&gl->gl_spin); | ||
1694 | return ret; | ||
1695 | } | ||
1696 | |||
1987 | /** | 1697 | /** |
1988 | * gfs2_dump_lockstate - print out the current lockstate | 1698 | * gfs2_dump_lockstate - print out the current lockstate |
1989 | * @sdp: the filesystem | 1699 | * @sdp: the filesystem |
@@ -2086,7 +1796,7 @@ void gfs2_glock_exit(void) | |||
2086 | module_param(scand_secs, uint, S_IRUGO|S_IWUSR); | 1796 | module_param(scand_secs, uint, S_IRUGO|S_IWUSR); |
2087 | MODULE_PARM_DESC(scand_secs, "The number of seconds between scand runs"); | 1797 | MODULE_PARM_DESC(scand_secs, "The number of seconds between scand runs"); |
2088 | 1798 | ||
2089 | static int gfs2_glock_iter_next(struct glock_iter *gi) | 1799 | static int gfs2_glock_iter_next(struct gfs2_glock_iter *gi) |
2090 | { | 1800 | { |
2091 | struct gfs2_glock *gl; | 1801 | struct gfs2_glock *gl; |
2092 | 1802 | ||
@@ -2104,7 +1814,7 @@ restart: | |||
2104 | gfs2_glock_put(gl); | 1814 | gfs2_glock_put(gl); |
2105 | if (gl && gi->gl == NULL) | 1815 | if (gl && gi->gl == NULL) |
2106 | gi->hash++; | 1816 | gi->hash++; |
2107 | while(gi->gl == NULL) { | 1817 | while (gi->gl == NULL) { |
2108 | if (gi->hash >= GFS2_GL_HASH_SIZE) | 1818 | if (gi->hash >= GFS2_GL_HASH_SIZE) |
2109 | return 1; | 1819 | return 1; |
2110 | read_lock(gl_lock_addr(gi->hash)); | 1820 | read_lock(gl_lock_addr(gi->hash)); |
@@ -2122,58 +1832,34 @@ restart: | |||
2122 | return 0; | 1832 | return 0; |
2123 | } | 1833 | } |
2124 | 1834 | ||
2125 | static void gfs2_glock_iter_free(struct glock_iter *gi) | 1835 | static void gfs2_glock_iter_free(struct gfs2_glock_iter *gi) |
2126 | { | 1836 | { |
2127 | if (gi->gl) | 1837 | if (gi->gl) |
2128 | gfs2_glock_put(gi->gl); | 1838 | gfs2_glock_put(gi->gl); |
2129 | kfree(gi); | ||
2130 | } | ||
2131 | |||
2132 | static struct glock_iter *gfs2_glock_iter_init(struct gfs2_sbd *sdp) | ||
2133 | { | ||
2134 | struct glock_iter *gi; | ||
2135 | |||
2136 | gi = kmalloc(sizeof (*gi), GFP_KERNEL); | ||
2137 | if (!gi) | ||
2138 | return NULL; | ||
2139 | |||
2140 | gi->sdp = sdp; | ||
2141 | gi->hash = 0; | ||
2142 | gi->seq = NULL; | ||
2143 | gi->gl = NULL; | 1839 | gi->gl = NULL; |
2144 | memset(gi->string, 0, sizeof(gi->string)); | ||
2145 | |||
2146 | if (gfs2_glock_iter_next(gi)) { | ||
2147 | gfs2_glock_iter_free(gi); | ||
2148 | return NULL; | ||
2149 | } | ||
2150 | |||
2151 | return gi; | ||
2152 | } | 1840 | } |
2153 | 1841 | ||
2154 | static void *gfs2_glock_seq_start(struct seq_file *file, loff_t *pos) | 1842 | static void *gfs2_glock_seq_start(struct seq_file *seq, loff_t *pos) |
2155 | { | 1843 | { |
2156 | struct glock_iter *gi; | 1844 | struct gfs2_glock_iter *gi = seq->private; |
2157 | loff_t n = *pos; | 1845 | loff_t n = *pos; |
2158 | 1846 | ||
2159 | gi = gfs2_glock_iter_init(file->private); | 1847 | gi->hash = 0; |
2160 | if (!gi) | ||
2161 | return NULL; | ||
2162 | 1848 | ||
2163 | while(n--) { | 1849 | do { |
2164 | if (gfs2_glock_iter_next(gi)) { | 1850 | if (gfs2_glock_iter_next(gi)) { |
2165 | gfs2_glock_iter_free(gi); | 1851 | gfs2_glock_iter_free(gi); |
2166 | return NULL; | 1852 | return NULL; |
2167 | } | 1853 | } |
2168 | } | 1854 | } while (n--); |
2169 | 1855 | ||
2170 | return gi; | 1856 | return gi->gl; |
2171 | } | 1857 | } |
2172 | 1858 | ||
2173 | static void *gfs2_glock_seq_next(struct seq_file *file, void *iter_ptr, | 1859 | static void *gfs2_glock_seq_next(struct seq_file *seq, void *iter_ptr, |
2174 | loff_t *pos) | 1860 | loff_t *pos) |
2175 | { | 1861 | { |
2176 | struct glock_iter *gi = iter_ptr; | 1862 | struct gfs2_glock_iter *gi = seq->private; |
2177 | 1863 | ||
2178 | (*pos)++; | 1864 | (*pos)++; |
2179 | 1865 | ||
@@ -2182,24 +1868,18 @@ static void *gfs2_glock_seq_next(struct seq_file *file, void *iter_ptr, | |||
2182 | return NULL; | 1868 | return NULL; |
2183 | } | 1869 | } |
2184 | 1870 | ||
2185 | return gi; | 1871 | return gi->gl; |
2186 | } | 1872 | } |
2187 | 1873 | ||
2188 | static void gfs2_glock_seq_stop(struct seq_file *file, void *iter_ptr) | 1874 | static void gfs2_glock_seq_stop(struct seq_file *seq, void *iter_ptr) |
2189 | { | 1875 | { |
2190 | struct glock_iter *gi = iter_ptr; | 1876 | struct gfs2_glock_iter *gi = seq->private; |
2191 | if (gi) | 1877 | gfs2_glock_iter_free(gi); |
2192 | gfs2_glock_iter_free(gi); | ||
2193 | } | 1878 | } |
2194 | 1879 | ||
2195 | static int gfs2_glock_seq_show(struct seq_file *file, void *iter_ptr) | 1880 | static int gfs2_glock_seq_show(struct seq_file *seq, void *iter_ptr) |
2196 | { | 1881 | { |
2197 | struct glock_iter *gi = iter_ptr; | 1882 | return dump_glock(seq, iter_ptr); |
2198 | |||
2199 | gi->seq = file; | ||
2200 | dump_glock(gi, gi->gl); | ||
2201 | |||
2202 | return 0; | ||
2203 | } | 1883 | } |
2204 | 1884 | ||
2205 | static const struct seq_operations gfs2_glock_seq_ops = { | 1885 | static const struct seq_operations gfs2_glock_seq_ops = { |
@@ -2211,17 +1891,14 @@ static const struct seq_operations gfs2_glock_seq_ops = { | |||
2211 | 1891 | ||
2212 | static int gfs2_debugfs_open(struct inode *inode, struct file *file) | 1892 | static int gfs2_debugfs_open(struct inode *inode, struct file *file) |
2213 | { | 1893 | { |
2214 | struct seq_file *seq; | 1894 | int ret = seq_open_private(file, &gfs2_glock_seq_ops, |
2215 | int ret; | 1895 | sizeof(struct gfs2_glock_iter)); |
2216 | 1896 | if (ret == 0) { | |
2217 | ret = seq_open(file, &gfs2_glock_seq_ops); | 1897 | struct seq_file *seq = file->private_data; |
2218 | if (ret) | 1898 | struct gfs2_glock_iter *gi = seq->private; |
2219 | return ret; | 1899 | gi->sdp = inode->i_private; |
2220 | 1900 | } | |
2221 | seq = file->private_data; | 1901 | return ret; |
2222 | seq->private = inode->i_private; | ||
2223 | |||
2224 | return 0; | ||
2225 | } | 1902 | } |
2226 | 1903 | ||
2227 | static const struct file_operations gfs2_debug_fops = { | 1904 | static const struct file_operations gfs2_debug_fops = { |
@@ -2229,7 +1906,7 @@ static const struct file_operations gfs2_debug_fops = { | |||
2229 | .open = gfs2_debugfs_open, | 1906 | .open = gfs2_debugfs_open, |
2230 | .read = seq_read, | 1907 | .read = seq_read, |
2231 | .llseek = seq_lseek, | 1908 | .llseek = seq_lseek, |
2232 | .release = seq_release | 1909 | .release = seq_release_private, |
2233 | }; | 1910 | }; |
2234 | 1911 | ||
2235 | int gfs2_create_debugfs_file(struct gfs2_sbd *sdp) | 1912 | int gfs2_create_debugfs_file(struct gfs2_sbd *sdp) |