diff options
author | Tom Tucker <tom@opengridcomputing.com> | 2007-12-30 22:08:10 -0500 |
---|---|---|
committer | J. Bruce Fields <bfields@citi.umich.edu> | 2008-02-01 16:42:12 -0500 |
commit | 8c7b0172a1db8120d25ecb4eff69664c52ee7639 (patch) | |
tree | d048ab4c5c378d2a90e5dd52a09dee4e24712cc9 /net | |
parent | def13d7401e9b95bbd34c20057ebeb2972708b1b (diff) |
svc: Make deferral processing xprt independent
This patch moves the transport independent sk_deferred list to the svc_xprt
structure and updates the svc_deferred_req structure to keep pointers to
svc_xprt's directly. The deferral processing code is also moved out of the
transport dependent recvfrom functions and into the generic svc_recv path.
Signed-off-by: Tom Tucker <tom@opengridcomputing.com>
Acked-by: Neil Brown <neilb@suse.de>
Reviewed-by: Chuck Lever <chuck.lever@oracle.com>
Reviewed-by: Greg Banks <gnb@sgi.com>
Signed-off-by: J. Bruce Fields <bfields@citi.umich.edu>
Diffstat (limited to 'net')
-rw-r--r-- | net/sunrpc/svc_xprt.c | 1 | ||||
-rw-r--r-- | net/sunrpc/svcsock.c | 57 |
2 files changed, 26 insertions, 32 deletions
diff --git a/net/sunrpc/svc_xprt.c b/net/sunrpc/svc_xprt.c index d2ac130b9040..023aeb0ecfa9 100644 --- a/net/sunrpc/svc_xprt.c +++ b/net/sunrpc/svc_xprt.c | |||
@@ -102,6 +102,7 @@ void svc_xprt_init(struct svc_xprt_class *xcl, struct svc_xprt *xprt, | |||
102 | xprt->xpt_server = serv; | 102 | xprt->xpt_server = serv; |
103 | INIT_LIST_HEAD(&xprt->xpt_list); | 103 | INIT_LIST_HEAD(&xprt->xpt_list); |
104 | INIT_LIST_HEAD(&xprt->xpt_ready); | 104 | INIT_LIST_HEAD(&xprt->xpt_ready); |
105 | INIT_LIST_HEAD(&xprt->xpt_deferred); | ||
105 | mutex_init(&xprt->xpt_mutex); | 106 | mutex_init(&xprt->xpt_mutex); |
106 | spin_lock_init(&xprt->xpt_lock); | 107 | spin_lock_init(&xprt->xpt_lock); |
107 | } | 108 | } |
diff --git a/net/sunrpc/svcsock.c b/net/sunrpc/svcsock.c index 5c9422c9a980..9d0a9e6c0e10 100644 --- a/net/sunrpc/svcsock.c +++ b/net/sunrpc/svcsock.c | |||
@@ -89,7 +89,7 @@ static void svc_close_xprt(struct svc_xprt *xprt); | |||
89 | static void svc_sock_detach(struct svc_xprt *); | 89 | static void svc_sock_detach(struct svc_xprt *); |
90 | static void svc_sock_free(struct svc_xprt *); | 90 | static void svc_sock_free(struct svc_xprt *); |
91 | 91 | ||
92 | static struct svc_deferred_req *svc_deferred_dequeue(struct svc_sock *svsk); | 92 | static struct svc_deferred_req *svc_deferred_dequeue(struct svc_xprt *xprt); |
93 | static int svc_deferred_recv(struct svc_rqst *rqstp); | 93 | static int svc_deferred_recv(struct svc_rqst *rqstp); |
94 | static struct cache_deferred_req *svc_defer(struct cache_req *req); | 94 | static struct cache_deferred_req *svc_defer(struct cache_req *req); |
95 | static struct svc_xprt *svc_create_socket(struct svc_serv *, int, | 95 | static struct svc_xprt *svc_create_socket(struct svc_serv *, int, |
@@ -771,11 +771,6 @@ svc_udp_recvfrom(struct svc_rqst *rqstp) | |||
771 | (serv->sv_nrthreads+3) * serv->sv_max_mesg, | 771 | (serv->sv_nrthreads+3) * serv->sv_max_mesg, |
772 | (serv->sv_nrthreads+3) * serv->sv_max_mesg); | 772 | (serv->sv_nrthreads+3) * serv->sv_max_mesg); |
773 | 773 | ||
774 | if ((rqstp->rq_deferred = svc_deferred_dequeue(svsk))) { | ||
775 | svc_xprt_received(&svsk->sk_xprt); | ||
776 | return svc_deferred_recv(rqstp); | ||
777 | } | ||
778 | |||
779 | clear_bit(XPT_DATA, &svsk->sk_xprt.xpt_flags); | 774 | clear_bit(XPT_DATA, &svsk->sk_xprt.xpt_flags); |
780 | skb = NULL; | 775 | skb = NULL; |
781 | err = kernel_recvmsg(svsk->sk_sock, &msg, NULL, | 776 | err = kernel_recvmsg(svsk->sk_sock, &msg, NULL, |
@@ -1138,11 +1133,6 @@ svc_tcp_recvfrom(struct svc_rqst *rqstp) | |||
1138 | test_bit(XPT_CONN, &svsk->sk_xprt.xpt_flags), | 1133 | test_bit(XPT_CONN, &svsk->sk_xprt.xpt_flags), |
1139 | test_bit(XPT_CLOSE, &svsk->sk_xprt.xpt_flags)); | 1134 | test_bit(XPT_CLOSE, &svsk->sk_xprt.xpt_flags)); |
1140 | 1135 | ||
1141 | if ((rqstp->rq_deferred = svc_deferred_dequeue(svsk))) { | ||
1142 | svc_xprt_received(&svsk->sk_xprt); | ||
1143 | return svc_deferred_recv(rqstp); | ||
1144 | } | ||
1145 | |||
1146 | if (test_and_clear_bit(XPT_CHNGBUF, &svsk->sk_xprt.xpt_flags)) | 1136 | if (test_and_clear_bit(XPT_CHNGBUF, &svsk->sk_xprt.xpt_flags)) |
1147 | /* sndbuf needs to have room for one request | 1137 | /* sndbuf needs to have room for one request |
1148 | * per thread, otherwise we can stall even when the | 1138 | * per thread, otherwise we can stall even when the |
@@ -1601,7 +1591,12 @@ svc_recv(struct svc_rqst *rqstp, long timeout) | |||
1601 | dprintk("svc: server %p, pool %u, socket %p, inuse=%d\n", | 1591 | dprintk("svc: server %p, pool %u, socket %p, inuse=%d\n", |
1602 | rqstp, pool->sp_id, svsk, | 1592 | rqstp, pool->sp_id, svsk, |
1603 | atomic_read(&svsk->sk_xprt.xpt_ref.refcount)); | 1593 | atomic_read(&svsk->sk_xprt.xpt_ref.refcount)); |
1604 | len = svsk->sk_xprt.xpt_ops->xpo_recvfrom(rqstp); | 1594 | rqstp->rq_deferred = svc_deferred_dequeue(&svsk->sk_xprt); |
1595 | if (rqstp->rq_deferred) { | ||
1596 | svc_xprt_received(&svsk->sk_xprt); | ||
1597 | len = svc_deferred_recv(rqstp); | ||
1598 | } else | ||
1599 | len = svsk->sk_xprt.xpt_ops->xpo_recvfrom(rqstp); | ||
1605 | dprintk("svc: got len=%d\n", len); | 1600 | dprintk("svc: got len=%d\n", len); |
1606 | } | 1601 | } |
1607 | 1602 | ||
@@ -1758,7 +1753,6 @@ static struct svc_sock *svc_setup_socket(struct svc_serv *serv, | |||
1758 | svsk->sk_ostate = inet->sk_state_change; | 1753 | svsk->sk_ostate = inet->sk_state_change; |
1759 | svsk->sk_odata = inet->sk_data_ready; | 1754 | svsk->sk_odata = inet->sk_data_ready; |
1760 | svsk->sk_owspace = inet->sk_write_space; | 1755 | svsk->sk_owspace = inet->sk_write_space; |
1761 | INIT_LIST_HEAD(&svsk->sk_deferred); | ||
1762 | 1756 | ||
1763 | /* Initialize the socket */ | 1757 | /* Initialize the socket */ |
1764 | if (sock->type == SOCK_DGRAM) | 1758 | if (sock->type == SOCK_DGRAM) |
@@ -1976,22 +1970,21 @@ void svc_close_all(struct list_head *xprt_list) | |||
1976 | static void svc_revisit(struct cache_deferred_req *dreq, int too_many) | 1970 | static void svc_revisit(struct cache_deferred_req *dreq, int too_many) |
1977 | { | 1971 | { |
1978 | struct svc_deferred_req *dr = container_of(dreq, struct svc_deferred_req, handle); | 1972 | struct svc_deferred_req *dr = container_of(dreq, struct svc_deferred_req, handle); |
1979 | struct svc_sock *svsk; | 1973 | struct svc_xprt *xprt = dr->xprt; |
1980 | 1974 | ||
1981 | if (too_many) { | 1975 | if (too_many) { |
1982 | svc_xprt_put(&dr->svsk->sk_xprt); | 1976 | svc_xprt_put(xprt); |
1983 | kfree(dr); | 1977 | kfree(dr); |
1984 | return; | 1978 | return; |
1985 | } | 1979 | } |
1986 | dprintk("revisit queued\n"); | 1980 | dprintk("revisit queued\n"); |
1987 | svsk = dr->svsk; | 1981 | dr->xprt = NULL; |
1988 | dr->svsk = NULL; | 1982 | spin_lock(&xprt->xpt_lock); |
1989 | spin_lock(&svsk->sk_xprt.xpt_lock); | 1983 | list_add(&dr->handle.recent, &xprt->xpt_deferred); |
1990 | list_add(&dr->handle.recent, &svsk->sk_deferred); | 1984 | spin_unlock(&xprt->xpt_lock); |
1991 | spin_unlock(&svsk->sk_xprt.xpt_lock); | 1985 | set_bit(XPT_DEFERRED, &xprt->xpt_flags); |
1992 | set_bit(XPT_DEFERRED, &svsk->sk_xprt.xpt_flags); | 1986 | svc_xprt_enqueue(xprt); |
1993 | svc_xprt_enqueue(&svsk->sk_xprt); | 1987 | svc_xprt_put(xprt); |
1994 | svc_xprt_put(&svsk->sk_xprt); | ||
1995 | } | 1988 | } |
1996 | 1989 | ||
1997 | static struct cache_deferred_req * | 1990 | static struct cache_deferred_req * |
@@ -2022,7 +2015,7 @@ svc_defer(struct cache_req *req) | |||
2022 | memcpy(dr->args, rqstp->rq_arg.head[0].iov_base-skip, dr->argslen<<2); | 2015 | memcpy(dr->args, rqstp->rq_arg.head[0].iov_base-skip, dr->argslen<<2); |
2023 | } | 2016 | } |
2024 | svc_xprt_get(rqstp->rq_xprt); | 2017 | svc_xprt_get(rqstp->rq_xprt); |
2025 | dr->svsk = rqstp->rq_sock; | 2018 | dr->xprt = rqstp->rq_xprt; |
2026 | 2019 | ||
2027 | dr->handle.revisit = svc_revisit; | 2020 | dr->handle.revisit = svc_revisit; |
2028 | return &dr->handle; | 2021 | return &dr->handle; |
@@ -2048,21 +2041,21 @@ static int svc_deferred_recv(struct svc_rqst *rqstp) | |||
2048 | } | 2041 | } |
2049 | 2042 | ||
2050 | 2043 | ||
2051 | static struct svc_deferred_req *svc_deferred_dequeue(struct svc_sock *svsk) | 2044 | static struct svc_deferred_req *svc_deferred_dequeue(struct svc_xprt *xprt) |
2052 | { | 2045 | { |
2053 | struct svc_deferred_req *dr = NULL; | 2046 | struct svc_deferred_req *dr = NULL; |
2054 | 2047 | ||
2055 | if (!test_bit(XPT_DEFERRED, &svsk->sk_xprt.xpt_flags)) | 2048 | if (!test_bit(XPT_DEFERRED, &xprt->xpt_flags)) |
2056 | return NULL; | 2049 | return NULL; |
2057 | spin_lock(&svsk->sk_xprt.xpt_lock); | 2050 | spin_lock(&xprt->xpt_lock); |
2058 | clear_bit(XPT_DEFERRED, &svsk->sk_xprt.xpt_flags); | 2051 | clear_bit(XPT_DEFERRED, &xprt->xpt_flags); |
2059 | if (!list_empty(&svsk->sk_deferred)) { | 2052 | if (!list_empty(&xprt->xpt_deferred)) { |
2060 | dr = list_entry(svsk->sk_deferred.next, | 2053 | dr = list_entry(xprt->xpt_deferred.next, |
2061 | struct svc_deferred_req, | 2054 | struct svc_deferred_req, |
2062 | handle.recent); | 2055 | handle.recent); |
2063 | list_del_init(&dr->handle.recent); | 2056 | list_del_init(&dr->handle.recent); |
2064 | set_bit(XPT_DEFERRED, &svsk->sk_xprt.xpt_flags); | 2057 | set_bit(XPT_DEFERRED, &xprt->xpt_flags); |
2065 | } | 2058 | } |
2066 | spin_unlock(&svsk->sk_xprt.xpt_lock); | 2059 | spin_unlock(&xprt->xpt_lock); |
2067 | return dr; | 2060 | return dr; |
2068 | } | 2061 | } |