aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorTim Chen <tim.c.chen@linux.intel.com>2014-01-21 18:36:00 -0500
committerIngo Molnar <mingo@kernel.org>2014-01-28 07:13:27 -0500
commite72246748ff006ab928bc774e276e6ef5542f9c5 (patch)
treeb3021f1615d2088ce20fc02bd61e9f2baab72dd1
parentaff7385b5a16bca6b8d9243f01a9ea5a5b411e1d (diff)
locking/mutexes/mcs: Restructure the MCS lock defines and locking code into its own file
We will need the MCS lock code for doing optimistic spinning for rwsem and queued rwlock. Extracting the MCS code from mutex.c and put into its own file allow us to reuse this code easily. We also inline mcs_spin_lock and mcs_spin_unlock functions for better efficiency. Note that using the smp_load_acquire/smp_store_release pair used in mcs_lock and mcs_unlock is not sufficient to form a full memory barrier across cpus for many architectures (except x86). For applications that absolutely need a full barrier across multiple cpus with mcs_unlock and mcs_lock pair, smp_mb__after_unlock_lock() should be used after mcs_lock. Reviewed-by: Paul E. McKenney <paulmck@linux.vnet.ibm.com> Signed-off-by: Tim Chen <tim.c.chen@linux.intel.com> Signed-off-by: Davidlohr Bueso <davidlohr@hp.com> Signed-off-by: Peter Zijlstra <peterz@infradead.org> Cc: Linus Torvalds <torvalds@linux-foundation.org> Cc: Andrew Morton <akpm@linux-foundation.org> Cc: Thomas Gleixner <tglx@linutronix.de> Link: http://lkml.kernel.org/r/1390347360.3138.63.camel@schen9-DESK Signed-off-by: Ingo Molnar <mingo@kernel.org>
-rw-r--r--include/linux/mcs_spinlock.h77
-rw-r--r--include/linux/mutex.h5
-rw-r--r--kernel/locking/mutex.c68
3 files changed, 87 insertions, 63 deletions
diff --git a/include/linux/mcs_spinlock.h b/include/linux/mcs_spinlock.h
new file mode 100644
index 000000000000..9578ef81940b
--- /dev/null
+++ b/include/linux/mcs_spinlock.h
@@ -0,0 +1,77 @@
1/*
2 * MCS lock defines
3 *
4 * This file contains the main data structure and API definitions of MCS lock.
5 *
6 * The MCS lock (proposed by Mellor-Crummey and Scott) is a simple spin-lock
7 * with the desirable properties of being fair, and with each cpu trying
8 * to acquire the lock spinning on a local variable.
9 * It avoids expensive cache bouncings that common test-and-set spin-lock
10 * implementations incur.
11 */
12#ifndef __LINUX_MCS_SPINLOCK_H
13#define __LINUX_MCS_SPINLOCK_H
14
15struct mcs_spinlock {
16 struct mcs_spinlock *next;
17 int locked; /* 1 if lock acquired */
18};
19
20/*
21 * Note: the smp_load_acquire/smp_store_release pair is not
22 * sufficient to form a full memory barrier across
23 * cpus for many architectures (except x86) for mcs_unlock and mcs_lock.
24 * For applications that need a full barrier across multiple cpus
25 * with mcs_unlock and mcs_lock pair, smp_mb__after_unlock_lock() should be
26 * used after mcs_lock.
27 */
28static inline
29void mcs_spin_lock(struct mcs_spinlock **lock, struct mcs_spinlock *node)
30{
31 struct mcs_spinlock *prev;
32
33 /* Init node */
34 node->locked = 0;
35 node->next = NULL;
36
37 prev = xchg(lock, node);
38 if (likely(prev == NULL)) {
39 /* Lock acquired */
40 node->locked = 1;
41 return;
42 }
43 ACCESS_ONCE(prev->next) = node;
44 /*
45 * Wait until the lock holder passes the lock down.
46 * Using smp_load_acquire() provides a memory barrier that
47 * ensures subsequent operations happen after the lock is acquired.
48 */
49 while (!(smp_load_acquire(&node->locked)))
50 arch_mutex_cpu_relax();
51}
52
53static inline
54void mcs_spin_unlock(struct mcs_spinlock **lock, struct mcs_spinlock *node)
55{
56 struct mcs_spinlock *next = ACCESS_ONCE(node->next);
57
58 if (likely(!next)) {
59 /*
60 * Release the lock by setting it to NULL
61 */
62 if (cmpxchg(lock, node, NULL) == node)
63 return;
64 /* Wait until the next pointer is set */
65 while (!(next = ACCESS_ONCE(node->next)))
66 arch_mutex_cpu_relax();
67 }
68 /*
69 * Pass lock to next waiter.
70 * smp_store_release() provides a memory barrier to ensure
71 * all operations in the critical section has been completed
72 * before unlocking.
73 */
74 smp_store_release(&next->locked, 1);
75}
76
77#endif /* __LINUX_MCS_SPINLOCK_H */
diff --git a/include/linux/mutex.h b/include/linux/mutex.h
index d3181936c138..c482e1d2cc49 100644
--- a/include/linux/mutex.h
+++ b/include/linux/mutex.h
@@ -46,6 +46,7 @@
46 * - detects multi-task circular deadlocks and prints out all affected 46 * - detects multi-task circular deadlocks and prints out all affected
47 * locks and tasks (and only those tasks) 47 * locks and tasks (and only those tasks)
48 */ 48 */
49struct mcs_spinlock;
49struct mutex { 50struct mutex {
50 /* 1: unlocked, 0: locked, negative: locked, possible waiters */ 51 /* 1: unlocked, 0: locked, negative: locked, possible waiters */
51 atomic_t count; 52 atomic_t count;
@@ -55,7 +56,7 @@ struct mutex {
55 struct task_struct *owner; 56 struct task_struct *owner;
56#endif 57#endif
57#ifdef CONFIG_MUTEX_SPIN_ON_OWNER 58#ifdef CONFIG_MUTEX_SPIN_ON_OWNER
58 void *spin_mlock; /* Spinner MCS lock */ 59 struct mcs_spinlock *mcs_lock; /* Spinner MCS lock */
59#endif 60#endif
60#ifdef CONFIG_DEBUG_MUTEXES 61#ifdef CONFIG_DEBUG_MUTEXES
61 const char *name; 62 const char *name;
@@ -179,4 +180,4 @@ extern int atomic_dec_and_mutex_lock(atomic_t *cnt, struct mutex *lock);
179# define arch_mutex_cpu_relax() cpu_relax() 180# define arch_mutex_cpu_relax() cpu_relax()
180#endif 181#endif
181 182
182#endif 183#endif /* __LINUX_MUTEX_H */
diff --git a/kernel/locking/mutex.c b/kernel/locking/mutex.c
index fbbd2eda867e..45fe1b5293d6 100644
--- a/kernel/locking/mutex.c
+++ b/kernel/locking/mutex.c
@@ -25,6 +25,7 @@
25#include <linux/spinlock.h> 25#include <linux/spinlock.h>
26#include <linux/interrupt.h> 26#include <linux/interrupt.h>
27#include <linux/debug_locks.h> 27#include <linux/debug_locks.h>
28#include <linux/mcs_spinlock.h>
28 29
29/* 30/*
30 * In the DEBUG case we are using the "NULL fastpath" for mutexes, 31 * In the DEBUG case we are using the "NULL fastpath" for mutexes,
@@ -52,7 +53,7 @@ __mutex_init(struct mutex *lock, const char *name, struct lock_class_key *key)
52 INIT_LIST_HEAD(&lock->wait_list); 53 INIT_LIST_HEAD(&lock->wait_list);
53 mutex_clear_owner(lock); 54 mutex_clear_owner(lock);
54#ifdef CONFIG_MUTEX_SPIN_ON_OWNER 55#ifdef CONFIG_MUTEX_SPIN_ON_OWNER
55 lock->spin_mlock = NULL; 56 lock->mcs_lock = NULL;
56#endif 57#endif
57 58
58 debug_mutex_init(lock, name, key); 59 debug_mutex_init(lock, name, key);
@@ -111,62 +112,7 @@ EXPORT_SYMBOL(mutex_lock);
111 * more or less simultaneously, the spinners need to acquire a MCS lock 112 * more or less simultaneously, the spinners need to acquire a MCS lock
112 * first before spinning on the owner field. 113 * first before spinning on the owner field.
113 * 114 *
114 * We don't inline mspin_lock() so that perf can correctly account for the
115 * time spent in this lock function.
116 */ 115 */
117struct mspin_node {
118 struct mspin_node *next ;
119 int locked; /* 1 if lock acquired */
120};
121#define MLOCK(mutex) ((struct mspin_node **)&((mutex)->spin_mlock))
122
123static noinline
124void mspin_lock(struct mspin_node **lock, struct mspin_node *node)
125{
126 struct mspin_node *prev;
127
128 /* Init node */
129 node->locked = 0;
130 node->next = NULL;
131
132 prev = xchg(lock, node);
133 if (likely(prev == NULL)) {
134 /* Lock acquired */
135 node->locked = 1;
136 return;
137 }
138 ACCESS_ONCE(prev->next) = node;
139 /*
140 * Wait until the lock holder passes the lock down.
141 * Using smp_load_acquire() provides a memory barrier that
142 * ensures subsequent operations happen after the lock is acquired.
143 */
144 while (!(smp_load_acquire(&node->locked)))
145 arch_mutex_cpu_relax();
146}
147
148static void mspin_unlock(struct mspin_node **lock, struct mspin_node *node)
149{
150 struct mspin_node *next = ACCESS_ONCE(node->next);
151
152 if (likely(!next)) {
153 /*
154 * Release the lock by setting it to NULL
155 */
156 if (cmpxchg(lock, node, NULL) == node)
157 return;
158 /* Wait until the next pointer is set */
159 while (!(next = ACCESS_ONCE(node->next)))
160 arch_mutex_cpu_relax();
161 }
162 /*
163 * Pass lock to next waiter.
164 * smp_store_release() provides a memory barrier to ensure
165 * all operations in the critical section has been completed
166 * before unlocking.
167 */
168 smp_store_release(&next->locked, 1);
169}
170 116
171/* 117/*
172 * Mutex spinning code migrated from kernel/sched/core.c 118 * Mutex spinning code migrated from kernel/sched/core.c
@@ -456,7 +402,7 @@ __mutex_lock_common(struct mutex *lock, long state, unsigned int subclass,
456 402
457 for (;;) { 403 for (;;) {
458 struct task_struct *owner; 404 struct task_struct *owner;
459 struct mspin_node node; 405 struct mcs_spinlock node;
460 406
461 if (use_ww_ctx && ww_ctx->acquired > 0) { 407 if (use_ww_ctx && ww_ctx->acquired > 0) {
462 struct ww_mutex *ww; 408 struct ww_mutex *ww;
@@ -478,10 +424,10 @@ __mutex_lock_common(struct mutex *lock, long state, unsigned int subclass,
478 * If there's an owner, wait for it to either 424 * If there's an owner, wait for it to either
479 * release the lock or go to sleep. 425 * release the lock or go to sleep.
480 */ 426 */
481 mspin_lock(MLOCK(lock), &node); 427 mcs_spin_lock(&lock->mcs_lock, &node);
482 owner = ACCESS_ONCE(lock->owner); 428 owner = ACCESS_ONCE(lock->owner);
483 if (owner && !mutex_spin_on_owner(lock, owner)) { 429 if (owner && !mutex_spin_on_owner(lock, owner)) {
484 mspin_unlock(MLOCK(lock), &node); 430 mcs_spin_unlock(&lock->mcs_lock, &node);
485 goto slowpath; 431 goto slowpath;
486 } 432 }
487 433
@@ -496,11 +442,11 @@ __mutex_lock_common(struct mutex *lock, long state, unsigned int subclass,
496 } 442 }
497 443
498 mutex_set_owner(lock); 444 mutex_set_owner(lock);
499 mspin_unlock(MLOCK(lock), &node); 445 mcs_spin_unlock(&lock->mcs_lock, &node);
500 preempt_enable(); 446 preempt_enable();
501 return 0; 447 return 0;
502 } 448 }
503 mspin_unlock(MLOCK(lock), &node); 449 mcs_spin_unlock(&lock->mcs_lock, &node);
504 450
505 /* 451 /*
506 * When there's no owner, we might have preempted between the 452 * When there's no owner, we might have preempted between the