aboutsummaryrefslogtreecommitdiffstats
path: root/fs/locks.c
diff options
context:
space:
mode:
authorJeff Layton <jlayton@redhat.com>2013-06-21 08:58:16 -0400
committerAl Viro <viro@zeniv.linux.org.uk>2013-06-29 04:57:43 -0400
commit4e8c765d384e549f9b542ea0bd42e2aa227e1404 (patch)
tree23600131f3c0e8fa0d84707ccb19344f14627804 /fs/locks.c
parent1c8c601a8c0dc59fe64907dcd9d512a3d181ddc7 (diff)
locks: avoid taking global lock if possible when waking up blocked waiters
Since we always hold the i_lock when inserting a new waiter onto the fl_block list, we can avoid taking the global lock at all if we find that it's empty when we go to wake up blocked waiters. Signed-off-by: Jeff Layton <jlayton@redhat.com> Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
Diffstat (limited to 'fs/locks.c')
-rw-r--r--fs/locks.c15
1 files changed, 14 insertions, 1 deletions
diff --git a/fs/locks.c b/fs/locks.c
index ce302d43822b..84e269fc4c69 100644
--- a/fs/locks.c
+++ b/fs/locks.c
@@ -548,7 +548,10 @@ static void locks_delete_block(struct file_lock *waiter)
548 * the order they blocked. The documentation doesn't require this but 548 * the order they blocked. The documentation doesn't require this but
549 * it seems like the reasonable thing to do. 549 * it seems like the reasonable thing to do.
550 * 550 *
551 * Must be called with file_lock_lock held! 551 * Must be called with both the i_lock and file_lock_lock held. The fl_block
552 * list itself is protected by the file_lock_list, but by ensuring that the
553 * i_lock is also held on insertions we can avoid taking the file_lock_lock
554 * in some cases when we see that the fl_block list is empty.
552 */ 555 */
553static void __locks_insert_block(struct file_lock *blocker, 556static void __locks_insert_block(struct file_lock *blocker,
554 struct file_lock *waiter) 557 struct file_lock *waiter)
@@ -576,6 +579,16 @@ static void locks_insert_block(struct file_lock *blocker,
576 */ 579 */
577static void locks_wake_up_blocks(struct file_lock *blocker) 580static void locks_wake_up_blocks(struct file_lock *blocker)
578{ 581{
582 /*
583 * Avoid taking global lock if list is empty. This is safe since new
584 * blocked requests are only added to the list under the i_lock, and
585 * the i_lock is always held here. Note that removal from the fl_block
586 * list does not require the i_lock, so we must recheck list_empty()
587 * after acquiring the file_lock_lock.
588 */
589 if (list_empty(&blocker->fl_block))
590 return;
591
579 spin_lock(&file_lock_lock); 592 spin_lock(&file_lock_lock);
580 while (!list_empty(&blocker->fl_block)) { 593 while (!list_empty(&blocker->fl_block)) {
581 struct file_lock *waiter; 594 struct file_lock *waiter;