aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/locking
diff options
context:
space:
mode:
authorJason Low <jason.low2@hp.com>2014-06-11 14:37:22 -0400
committerIngo Molnar <mingo@kernel.org>2014-07-05 05:25:42 -0400
commit0d968dd8c6aced585b86fa7ba8ce4573bf19e848 (patch)
tree08d135faf746fc75290f3f16c71eb651b6fbe948 /kernel/locking
parent1e820c9608eace237e2c519d8fd9074aec479d81 (diff)
locking/mutexes: Try to acquire mutex only if it is unlocked
Upon entering the slowpath in __mutex_lock_common(), we try once more to acquire the mutex. We only try to acquire if (lock->count >= 0). However, what we actually want here is to try to acquire if the mutex is unlocked (lock->count == 1). This patch changes it so that we only try-acquire the mutex upon entering the slowpath if it is unlocked, rather than if the lock count is non-negative. This helps further reduce unnecessary atomic xchg() operations. Furthermore, this patch uses !mutex_is_locked(lock) to do the initial checks for if the lock is free rather than directly calling atomic_read() on the lock->count, in order to improve readability. Signed-off-by: Jason Low <jason.low2@hp.com> Acked-by: Waiman Long <Waiman.Long@hp.com> Signed-off-by: Peter Zijlstra <peterz@infradead.org> Cc: akpm@linux-foundation.org Cc: tim.c.chen@linux.intel.com Cc: paulmck@linux.vnet.ibm.com Cc: rostedt@goodmis.org Cc: davidlohr@hp.com Cc: scott.norton@hp.com Cc: aswin@hp.com Cc: Linus Torvalds <torvalds@linux-foundation.org> Link: http://lkml.kernel.org/r/1402511843-4721-4-git-send-email-jason.low2@hp.com Signed-off-by: Ingo Molnar <mingo@kernel.org>
Diffstat (limited to 'kernel/locking')
-rw-r--r--kernel/locking/mutex.c7
1 files changed, 4 insertions, 3 deletions
diff --git a/kernel/locking/mutex.c b/kernel/locking/mutex.c
index 4bd95465af55..e4d997bb7d70 100644
--- a/kernel/locking/mutex.c
+++ b/kernel/locking/mutex.c
@@ -432,7 +432,8 @@ __mutex_lock_common(struct mutex *lock, long state, unsigned int subclass,
432 if (owner && !mutex_spin_on_owner(lock, owner)) 432 if (owner && !mutex_spin_on_owner(lock, owner))
433 break; 433 break;
434 434
435 if ((atomic_read(&lock->count) == 1) && 435 /* Try to acquire the mutex if it is unlocked. */
436 if (!mutex_is_locked(lock) &&
436 (atomic_cmpxchg(&lock->count, 1, 0) == 1)) { 437 (atomic_cmpxchg(&lock->count, 1, 0) == 1)) {
437 lock_acquired(&lock->dep_map, ip); 438 lock_acquired(&lock->dep_map, ip);
438 if (use_ww_ctx) { 439 if (use_ww_ctx) {
@@ -479,9 +480,9 @@ slowpath:
479 480
480 /* 481 /*
481 * Once more, try to acquire the lock. Only try-lock the mutex if 482 * Once more, try to acquire the lock. Only try-lock the mutex if
482 * lock->count >= 0 to reduce unnecessary xchg operations. 483 * it is unlocked to reduce unnecessary xchg() operations.
483 */ 484 */
484 if (atomic_read(&lock->count) >= 0 && (atomic_xchg(&lock->count, 0) == 1)) 485 if (!mutex_is_locked(lock) && (atomic_xchg(&lock->count, 0) == 1))
485 goto skip_wait; 486 goto skip_wait;
486 487
487 debug_mutex_lock_common(lock, &waiter); 488 debug_mutex_lock_common(lock, &waiter);