aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorSteven Whitehouse <swhiteho@redhat.com>2007-01-22 13:09:04 -0500
committerSteven Whitehouse <swhiteho@redhat.com>2007-02-05 13:37:32 -0500
commit12132933c4fdeb458195a9388287d550c8476edf (patch)
tree037a43592c395a7d60561628fdf24e80b5901489
parentbd44e2b007bc9024bce3357c185b38c73f87c3dd (diff)
[GFS2] Remove queue_empty() function
This function is not longer required since we do not do recursive locking in the glock layer. As a result all its callers can be replaceed with list_empty() calls. Signed-off-by: Steven Whitehouse <swhiteho@redhat.com>
-rw-r--r--fs/gfs2/glock.c38
1 files changed, 7 insertions, 31 deletions
diff --git a/fs/gfs2/glock.c b/fs/gfs2/glock.c
index 5b772bb0210f..1509481b8ca6 100644
--- a/fs/gfs2/glock.c
+++ b/fs/gfs2/glock.c
@@ -211,30 +211,6 @@ out:
211} 211}
212 212
213/** 213/**
214 * queue_empty - check to see if a glock's queue is empty
215 * @gl: the glock
216 * @head: the head of the queue to check
217 *
218 * This function protects the list in the event that a process already
219 * has a holder on the list and is adding a second holder for itself.
220 * The glmutex lock is what generally prevents processes from working
221 * on the same glock at once, but the special case of adding a second
222 * holder for yourself ("recursive" locking) doesn't involve locking
223 * glmutex, making the spin lock necessary.
224 *
225 * Returns: 1 if the queue is empty
226 */
227
228static inline int queue_empty(struct gfs2_glock *gl, struct list_head *head)
229{
230 int empty;
231 spin_lock(&gl->gl_spin);
232 empty = list_empty(head);
233 spin_unlock(&gl->gl_spin);
234 return empty;
235}
236
237/**
238 * search_bucket() - Find struct gfs2_glock by lock number 214 * search_bucket() - Find struct gfs2_glock by lock number
239 * @bucket: the bucket to search 215 * @bucket: the bucket to search
240 * @name: The lock name 216 * @name: The lock name
@@ -814,7 +790,7 @@ static void xmote_bh(struct gfs2_glock *gl, unsigned int ret)
814 int op_done = 1; 790 int op_done = 1;
815 791
816 gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); 792 gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags));
817 gfs2_assert_warn(sdp, queue_empty(gl, &gl->gl_holders)); 793 gfs2_assert_warn(sdp, list_empty(&gl->gl_holders));
818 gfs2_assert_warn(sdp, !(ret & LM_OUT_ASYNC)); 794 gfs2_assert_warn(sdp, !(ret & LM_OUT_ASYNC));
819 795
820 state_change(gl, ret & LM_OUT_ST_MASK); 796 state_change(gl, ret & LM_OUT_ST_MASK);
@@ -925,7 +901,7 @@ void gfs2_glock_xmote_th(struct gfs2_holder *gh)
925 glops->go_xmote_th(gl); 901 glops->go_xmote_th(gl);
926 902
927 gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); 903 gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags));
928 gfs2_assert_warn(sdp, queue_empty(gl, &gl->gl_holders)); 904 gfs2_assert_warn(sdp, list_empty(&gl->gl_holders));
929 gfs2_assert_warn(sdp, state != LM_ST_UNLOCKED); 905 gfs2_assert_warn(sdp, state != LM_ST_UNLOCKED);
930 gfs2_assert_warn(sdp, state != gl->gl_state); 906 gfs2_assert_warn(sdp, state != gl->gl_state);
931 907
@@ -960,7 +936,7 @@ static void drop_bh(struct gfs2_glock *gl, unsigned int ret)
960 struct gfs2_holder *gh = gl->gl_req_gh; 936 struct gfs2_holder *gh = gl->gl_req_gh;
961 937
962 gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); 938 gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags));
963 gfs2_assert_warn(sdp, queue_empty(gl, &gl->gl_holders)); 939 gfs2_assert_warn(sdp, list_empty(&gl->gl_holders));
964 gfs2_assert_warn(sdp, !ret); 940 gfs2_assert_warn(sdp, !ret);
965 941
966 state_change(gl, LM_ST_UNLOCKED); 942 state_change(gl, LM_ST_UNLOCKED);
@@ -1007,7 +983,7 @@ static void gfs2_glock_drop_th(struct gfs2_glock *gl)
1007 glops->go_drop_th(gl); 983 glops->go_drop_th(gl);
1008 984
1009 gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags)); 985 gfs2_assert_warn(sdp, test_bit(GLF_LOCK, &gl->gl_flags));
1010 gfs2_assert_warn(sdp, queue_empty(gl, &gl->gl_holders)); 986 gfs2_assert_warn(sdp, list_empty(&gl->gl_holders));
1011 gfs2_assert_warn(sdp, gl->gl_state != LM_ST_UNLOCKED); 987 gfs2_assert_warn(sdp, gl->gl_state != LM_ST_UNLOCKED);
1012 988
1013 gfs2_glock_hold(gl); 989 gfs2_glock_hold(gl);
@@ -1697,7 +1673,7 @@ void gfs2_reclaim_glock(struct gfs2_sbd *sdp)
1697 atomic_inc(&sdp->sd_reclaimed); 1673 atomic_inc(&sdp->sd_reclaimed);
1698 1674
1699 if (gfs2_glmutex_trylock(gl)) { 1675 if (gfs2_glmutex_trylock(gl)) {
1700 if (queue_empty(gl, &gl->gl_holders) && 1676 if (list_empty(&gl->gl_holders) &&
1701 gl->gl_state != LM_ST_UNLOCKED && demote_ok(gl)) 1677 gl->gl_state != LM_ST_UNLOCKED && demote_ok(gl))
1702 handle_callback(gl, LM_ST_UNLOCKED); 1678 handle_callback(gl, LM_ST_UNLOCKED);
1703 gfs2_glmutex_unlock(gl); 1679 gfs2_glmutex_unlock(gl);
@@ -1761,7 +1737,7 @@ static void scan_glock(struct gfs2_glock *gl)
1761 return; 1737 return;
1762 1738
1763 if (gfs2_glmutex_trylock(gl)) { 1739 if (gfs2_glmutex_trylock(gl)) {
1764 if (queue_empty(gl, &gl->gl_holders) && 1740 if (list_empty(&gl->gl_holders) &&
1765 gl->gl_state != LM_ST_UNLOCKED && demote_ok(gl)) 1741 gl->gl_state != LM_ST_UNLOCKED && demote_ok(gl))
1766 goto out_schedule; 1742 goto out_schedule;
1767 gfs2_glmutex_unlock(gl); 1743 gfs2_glmutex_unlock(gl);
@@ -1810,7 +1786,7 @@ static void clear_glock(struct gfs2_glock *gl)
1810 } 1786 }
1811 1787
1812 if (gfs2_glmutex_trylock(gl)) { 1788 if (gfs2_glmutex_trylock(gl)) {
1813 if (queue_empty(gl, &gl->gl_holders) && 1789 if (list_empty(gl, &gl->gl_holders) &&
1814 gl->gl_state != LM_ST_UNLOCKED) 1790 gl->gl_state != LM_ST_UNLOCKED)
1815 handle_callback(gl, LM_ST_UNLOCKED); 1791 handle_callback(gl, LM_ST_UNLOCKED);
1816 gfs2_glmutex_unlock(gl); 1792 gfs2_glmutex_unlock(gl);