diff options
author | Jason Low <jason.low2@hp.com> | 2014-06-11 14:37:22 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@kernel.org> | 2014-07-05 05:25:42 -0400 |
commit | 0d968dd8c6aced585b86fa7ba8ce4573bf19e848 (patch) | |
tree | 08d135faf746fc75290f3f16c71eb651b6fbe948 /kernel/locking | |
parent | 1e820c9608eace237e2c519d8fd9074aec479d81 (diff) |
locking/mutexes: Try to acquire mutex only if it is unlocked
Upon entering the slowpath in __mutex_lock_common(), we try once more to
acquire the mutex. We only try to acquire if (lock->count >= 0). However,
what we actually want here is to try to acquire if the mutex is unlocked
(lock->count == 1).
This patch changes it so that we only try-acquire the mutex upon entering
the slowpath if it is unlocked, rather than if the lock count is non-negative.
This helps further reduce unnecessary atomic xchg() operations.
Furthermore, this patch uses !mutex_is_locked(lock) to do the initial
checks for if the lock is free rather than directly calling atomic_read()
on the lock->count, in order to improve readability.
Signed-off-by: Jason Low <jason.low2@hp.com>
Acked-by: Waiman Long <Waiman.Long@hp.com>
Signed-off-by: Peter Zijlstra <peterz@infradead.org>
Cc: akpm@linux-foundation.org
Cc: tim.c.chen@linux.intel.com
Cc: paulmck@linux.vnet.ibm.com
Cc: rostedt@goodmis.org
Cc: davidlohr@hp.com
Cc: scott.norton@hp.com
Cc: aswin@hp.com
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Link: http://lkml.kernel.org/r/1402511843-4721-4-git-send-email-jason.low2@hp.com
Signed-off-by: Ingo Molnar <mingo@kernel.org>
Diffstat (limited to 'kernel/locking')
-rw-r--r-- | kernel/locking/mutex.c | 7 |
1 files changed, 4 insertions, 3 deletions
diff --git a/kernel/locking/mutex.c b/kernel/locking/mutex.c index 4bd95465af55..e4d997bb7d70 100644 --- a/kernel/locking/mutex.c +++ b/kernel/locking/mutex.c | |||
@@ -432,7 +432,8 @@ __mutex_lock_common(struct mutex *lock, long state, unsigned int subclass, | |||
432 | if (owner && !mutex_spin_on_owner(lock, owner)) | 432 | if (owner && !mutex_spin_on_owner(lock, owner)) |
433 | break; | 433 | break; |
434 | 434 | ||
435 | if ((atomic_read(&lock->count) == 1) && | 435 | /* Try to acquire the mutex if it is unlocked. */ |
436 | if (!mutex_is_locked(lock) && | ||
436 | (atomic_cmpxchg(&lock->count, 1, 0) == 1)) { | 437 | (atomic_cmpxchg(&lock->count, 1, 0) == 1)) { |
437 | lock_acquired(&lock->dep_map, ip); | 438 | lock_acquired(&lock->dep_map, ip); |
438 | if (use_ww_ctx) { | 439 | if (use_ww_ctx) { |
@@ -479,9 +480,9 @@ slowpath: | |||
479 | 480 | ||
480 | /* | 481 | /* |
481 | * Once more, try to acquire the lock. Only try-lock the mutex if | 482 | * Once more, try to acquire the lock. Only try-lock the mutex if |
482 | * lock->count >= 0 to reduce unnecessary xchg operations. | 483 | * it is unlocked to reduce unnecessary xchg() operations. |
483 | */ | 484 | */ |
484 | if (atomic_read(&lock->count) >= 0 && (atomic_xchg(&lock->count, 0) == 1)) | 485 | if (!mutex_is_locked(lock) && (atomic_xchg(&lock->count, 0) == 1)) |
485 | goto skip_wait; | 486 | goto skip_wait; |
486 | 487 | ||
487 | debug_mutex_lock_common(lock, &waiter); | 488 | debug_mutex_lock_common(lock, &waiter); |