aboutsummaryrefslogtreecommitdiffstats
path: root/net
diff options
context:
space:
mode:
authorTom Tucker <tom@opengridcomputing.com>2007-12-30 22:08:16 -0500
committerJ. Bruce Fields <bfields@citi.umich.edu>2008-02-01 16:42:12 -0500
commitc36adb2a7f9132b37d4b669b2e2c04e46d5188b2 (patch)
tree0047a71515d96cdd4085adbd1691232db42c2b9f /net
parenteab996d4aca7a9d8621d2b98c00ce420df85eaed (diff)
svc: Make svc_recv transport neutral
All of the transport field and functions used by svc_recv are now transport independent. Change the svc_recv function to use the svc_xprt structure directly instead of the transport specific svc_sock structure. Signed-off-by: Tom Tucker <tom@opengridcomputing.com> Acked-by: Neil Brown <neilb@suse.de> Reviewed-by: Chuck Lever <chuck.lever@oracle.com> Reviewed-by: Greg Banks <gnb@sgi.com> Signed-off-by: J. Bruce Fields <bfields@citi.umich.edu>
Diffstat (limited to 'net')
-rw-r--r--net/sunrpc/svcsock.c63
1 files changed, 32 insertions, 31 deletions
diff --git a/net/sunrpc/svcsock.c b/net/sunrpc/svcsock.c
index 355ab8da54fe..fa57e9bc68e4 100644
--- a/net/sunrpc/svcsock.c
+++ b/net/sunrpc/svcsock.c
@@ -310,22 +310,21 @@ EXPORT_SYMBOL_GPL(svc_xprt_enqueue);
310/* 310/*
311 * Dequeue the first socket. Must be called with the pool->sp_lock held. 311 * Dequeue the first socket. Must be called with the pool->sp_lock held.
312 */ 312 */
313static inline struct svc_sock * 313static struct svc_xprt *svc_xprt_dequeue(struct svc_pool *pool)
314svc_sock_dequeue(struct svc_pool *pool)
315{ 314{
316 struct svc_sock *svsk; 315 struct svc_xprt *xprt;
317 316
318 if (list_empty(&pool->sp_sockets)) 317 if (list_empty(&pool->sp_sockets))
319 return NULL; 318 return NULL;
320 319
321 svsk = list_entry(pool->sp_sockets.next, 320 xprt = list_entry(pool->sp_sockets.next,
322 struct svc_sock, sk_xprt.xpt_ready); 321 struct svc_xprt, xpt_ready);
323 list_del_init(&svsk->sk_xprt.xpt_ready); 322 list_del_init(&xprt->xpt_ready);
324 323
325 dprintk("svc: socket %p dequeued, inuse=%d\n", 324 dprintk("svc: transport %p dequeued, inuse=%d\n",
326 svsk->sk_sk, atomic_read(&svsk->sk_xprt.xpt_ref.refcount)); 325 xprt, atomic_read(&xprt->xpt_ref.refcount));
327 326
328 return svsk; 327 return xprt;
329} 328}
330 329
331/* 330/*
@@ -1475,20 +1474,20 @@ static void svc_check_conn_limits(struct svc_serv *serv)
1475int 1474int
1476svc_recv(struct svc_rqst *rqstp, long timeout) 1475svc_recv(struct svc_rqst *rqstp, long timeout)
1477{ 1476{
1478 struct svc_sock *svsk = NULL; 1477 struct svc_xprt *xprt = NULL;
1479 struct svc_serv *serv = rqstp->rq_server; 1478 struct svc_serv *serv = rqstp->rq_server;
1480 struct svc_pool *pool = rqstp->rq_pool; 1479 struct svc_pool *pool = rqstp->rq_pool;
1481 int len, i; 1480 int len, i;
1482 int pages; 1481 int pages;
1483 struct xdr_buf *arg; 1482 struct xdr_buf *arg;
1484 DECLARE_WAITQUEUE(wait, current); 1483 DECLARE_WAITQUEUE(wait, current);
1485 1484
1486 dprintk("svc: server %p waiting for data (to = %ld)\n", 1485 dprintk("svc: server %p waiting for data (to = %ld)\n",
1487 rqstp, timeout); 1486 rqstp, timeout);
1488 1487
1489 if (rqstp->rq_sock) 1488 if (rqstp->rq_xprt)
1490 printk(KERN_ERR 1489 printk(KERN_ERR
1491 "svc_recv: service %p, socket not NULL!\n", 1490 "svc_recv: service %p, transport not NULL!\n",
1492 rqstp); 1491 rqstp);
1493 if (waitqueue_active(&rqstp->rq_wait)) 1492 if (waitqueue_active(&rqstp->rq_wait))
1494 printk(KERN_ERR 1493 printk(KERN_ERR
@@ -1525,11 +1524,12 @@ svc_recv(struct svc_rqst *rqstp, long timeout)
1525 return -EINTR; 1524 return -EINTR;
1526 1525
1527 spin_lock_bh(&pool->sp_lock); 1526 spin_lock_bh(&pool->sp_lock);
1528 if ((svsk = svc_sock_dequeue(pool)) != NULL) { 1527 xprt = svc_xprt_dequeue(pool);
1529 rqstp->rq_sock = svsk; 1528 if (xprt) {
1530 svc_xprt_get(&svsk->sk_xprt); 1529 rqstp->rq_xprt = xprt;
1530 svc_xprt_get(xprt);
1531 rqstp->rq_reserved = serv->sv_max_mesg; 1531 rqstp->rq_reserved = serv->sv_max_mesg;
1532 atomic_add(rqstp->rq_reserved, &svsk->sk_xprt.xpt_reserved); 1532 atomic_add(rqstp->rq_reserved, &xprt->xpt_reserved);
1533 } else { 1533 } else {
1534 /* No data pending. Go to sleep */ 1534 /* No data pending. Go to sleep */
1535 svc_thread_enqueue(pool, rqstp); 1535 svc_thread_enqueue(pool, rqstp);
@@ -1549,7 +1549,8 @@ svc_recv(struct svc_rqst *rqstp, long timeout)
1549 spin_lock_bh(&pool->sp_lock); 1549 spin_lock_bh(&pool->sp_lock);
1550 remove_wait_queue(&rqstp->rq_wait, &wait); 1550 remove_wait_queue(&rqstp->rq_wait, &wait);
1551 1551
1552 if (!(svsk = rqstp->rq_sock)) { 1552 xprt = rqstp->rq_xprt;
1553 if (!xprt) {
1553 svc_thread_dequeue(pool, rqstp); 1554 svc_thread_dequeue(pool, rqstp);
1554 spin_unlock_bh(&pool->sp_lock); 1555 spin_unlock_bh(&pool->sp_lock);
1555 dprintk("svc: server %p, no data yet\n", rqstp); 1556 dprintk("svc: server %p, no data yet\n", rqstp);
@@ -1559,32 +1560,32 @@ svc_recv(struct svc_rqst *rqstp, long timeout)
1559 spin_unlock_bh(&pool->sp_lock); 1560 spin_unlock_bh(&pool->sp_lock);
1560 1561
1561 len = 0; 1562 len = 0;
1562 if (test_bit(XPT_CLOSE, &svsk->sk_xprt.xpt_flags)) { 1563 if (test_bit(XPT_CLOSE, &xprt->xpt_flags)) {
1563 dprintk("svc_recv: found XPT_CLOSE\n"); 1564 dprintk("svc_recv: found XPT_CLOSE\n");
1564 svc_delete_xprt(&svsk->sk_xprt); 1565 svc_delete_xprt(xprt);
1565 } else if (test_bit(XPT_LISTENER, &svsk->sk_xprt.xpt_flags)) { 1566 } else if (test_bit(XPT_LISTENER, &xprt->xpt_flags)) {
1566 struct svc_xprt *newxpt; 1567 struct svc_xprt *newxpt;
1567 newxpt = svsk->sk_xprt.xpt_ops->xpo_accept(&svsk->sk_xprt); 1568 newxpt = xprt->xpt_ops->xpo_accept(xprt);
1568 if (newxpt) { 1569 if (newxpt) {
1569 /* 1570 /*
1570 * We know this module_get will succeed because the 1571 * We know this module_get will succeed because the
1571 * listener holds a reference too 1572 * listener holds a reference too
1572 */ 1573 */
1573 __module_get(newxpt->xpt_class->xcl_owner); 1574 __module_get(newxpt->xpt_class->xcl_owner);
1574 svc_check_conn_limits(svsk->sk_xprt.xpt_server); 1575 svc_check_conn_limits(xprt->xpt_server);
1575 svc_xprt_received(newxpt); 1576 svc_xprt_received(newxpt);
1576 } 1577 }
1577 svc_xprt_received(&svsk->sk_xprt); 1578 svc_xprt_received(xprt);
1578 } else { 1579 } else {
1579 dprintk("svc: server %p, pool %u, socket %p, inuse=%d\n", 1580 dprintk("svc: server %p, pool %u, transport %p, inuse=%d\n",
1580 rqstp, pool->sp_id, svsk, 1581 rqstp, pool->sp_id, xprt,
1581 atomic_read(&svsk->sk_xprt.xpt_ref.refcount)); 1582 atomic_read(&xprt->xpt_ref.refcount));
1582 rqstp->rq_deferred = svc_deferred_dequeue(&svsk->sk_xprt); 1583 rqstp->rq_deferred = svc_deferred_dequeue(xprt);
1583 if (rqstp->rq_deferred) { 1584 if (rqstp->rq_deferred) {
1584 svc_xprt_received(&svsk->sk_xprt); 1585 svc_xprt_received(xprt);
1585 len = svc_deferred_recv(rqstp); 1586 len = svc_deferred_recv(rqstp);
1586 } else 1587 } else
1587 len = svsk->sk_xprt.xpt_ops->xpo_recvfrom(rqstp); 1588 len = xprt->xpt_ops->xpo_recvfrom(rqstp);
1588 dprintk("svc: got len=%d\n", len); 1589 dprintk("svc: got len=%d\n", len);
1589 } 1590 }
1590 1591
@@ -1594,7 +1595,7 @@ svc_recv(struct svc_rqst *rqstp, long timeout)
1594 svc_xprt_release(rqstp); 1595 svc_xprt_release(rqstp);
1595 return -EAGAIN; 1596 return -EAGAIN;
1596 } 1597 }
1597 clear_bit(XPT_OLD, &svsk->sk_xprt.xpt_flags); 1598 clear_bit(XPT_OLD, &xprt->xpt_flags);
1598 1599
1599 rqstp->rq_secure = svc_port_is_privileged(svc_addr(rqstp)); 1600 rqstp->rq_secure = svc_port_is_privileged(svc_addr(rqstp));
1600 rqstp->rq_chandle.defer = svc_defer; 1601 rqstp->rq_chandle.defer = svc_defer;