aboutsummaryrefslogtreecommitdiffstats
path: root/net
diff options
context:
space:
mode:
authorChuck Lever <chuck.lever@oracle.com>2014-05-28 10:35:14 -0400
committerAnna Schumaker <Anna.Schumaker@Netapp.com>2014-06-04 08:56:53 -0400
commitc93c62231cf55df4a26bd08937efeea97e6fc5e8 (patch)
treefa9b5c277f6fad09e1157cba579816c760c985ae /net
parentc977dea22708688eae31774f70126c97aa4dfe83 (diff)
xprtrdma: Disconnect on registration failure
If rpcrdma_register_external() fails during request marshaling, the current RPC request is killed. Instead, this RPC should be retried after reconnecting the transport instance. The most likely reason for registration failure with FRMR is a failed post_send, which would be due to a remote transport disconnect or memory exhaustion. These issues can be recovered by a retry. Problems encountered in the marshaling logic itself will not be corrected by trying again, so these should still kill a request. Now that we've added a clean exit for marshaling errors, take the opportunity to defang some BUG_ON's. Signed-off-by: Chuck Lever <chuck.lever@oracle.com> Signed-off-by: Anna Schumaker <Anna.Schumaker@Netapp.com>
Diffstat (limited to 'net')
-rw-r--r--net/sunrpc/xprtrdma/rpc_rdma.c48
-rw-r--r--net/sunrpc/xprtrdma/transport.c17
2 files changed, 42 insertions, 23 deletions
diff --git a/net/sunrpc/xprtrdma/rpc_rdma.c b/net/sunrpc/xprtrdma/rpc_rdma.c
index 77b84cfa5c77..693966d3f33b 100644
--- a/net/sunrpc/xprtrdma/rpc_rdma.c
+++ b/net/sunrpc/xprtrdma/rpc_rdma.c
@@ -77,6 +77,8 @@ static const char transfertypes[][12] = {
77 * Prepare the passed-in xdr_buf into representation as RPC/RDMA chunk 77 * Prepare the passed-in xdr_buf into representation as RPC/RDMA chunk
78 * elements. Segments are then coalesced when registered, if possible 78 * elements. Segments are then coalesced when registered, if possible
79 * within the selected memreg mode. 79 * within the selected memreg mode.
80 *
81 * Returns positive number of segments converted, or a negative errno.
80 */ 82 */
81 83
82static int 84static int
@@ -103,12 +105,13 @@ rpcrdma_convert_iovs(struct xdr_buf *xdrbuf, unsigned int pos,
103 /* alloc the pagelist for receiving buffer */ 105 /* alloc the pagelist for receiving buffer */
104 ppages[p] = alloc_page(GFP_ATOMIC); 106 ppages[p] = alloc_page(GFP_ATOMIC);
105 if (!ppages[p]) 107 if (!ppages[p])
106 return 0; 108 return -ENOMEM;
107 } 109 }
108 seg[n].mr_page = ppages[p]; 110 seg[n].mr_page = ppages[p];
109 seg[n].mr_offset = (void *)(unsigned long) page_base; 111 seg[n].mr_offset = (void *)(unsigned long) page_base;
110 seg[n].mr_len = min_t(u32, PAGE_SIZE - page_base, len); 112 seg[n].mr_len = min_t(u32, PAGE_SIZE - page_base, len);
111 BUG_ON(seg[n].mr_len > PAGE_SIZE); 113 if (seg[n].mr_len > PAGE_SIZE)
114 return -EIO;
112 len -= seg[n].mr_len; 115 len -= seg[n].mr_len;
113 ++n; 116 ++n;
114 ++p; 117 ++p;
@@ -117,7 +120,7 @@ rpcrdma_convert_iovs(struct xdr_buf *xdrbuf, unsigned int pos,
117 120
118 /* Message overflows the seg array */ 121 /* Message overflows the seg array */
119 if (len && n == nsegs) 122 if (len && n == nsegs)
120 return 0; 123 return -EIO;
121 124
122 if (xdrbuf->tail[0].iov_len) { 125 if (xdrbuf->tail[0].iov_len) {
123 /* the rpcrdma protocol allows us to omit any trailing 126 /* the rpcrdma protocol allows us to omit any trailing
@@ -126,7 +129,7 @@ rpcrdma_convert_iovs(struct xdr_buf *xdrbuf, unsigned int pos,
126 return n; 129 return n;
127 if (n == nsegs) 130 if (n == nsegs)
128 /* Tail remains, but we're out of segments */ 131 /* Tail remains, but we're out of segments */
129 return 0; 132 return -EIO;
130 seg[n].mr_page = NULL; 133 seg[n].mr_page = NULL;
131 seg[n].mr_offset = xdrbuf->tail[0].iov_base; 134 seg[n].mr_offset = xdrbuf->tail[0].iov_base;
132 seg[n].mr_len = xdrbuf->tail[0].iov_len; 135 seg[n].mr_len = xdrbuf->tail[0].iov_len;
@@ -167,15 +170,17 @@ rpcrdma_convert_iovs(struct xdr_buf *xdrbuf, unsigned int pos,
167 * Reply chunk (a counted array): 170 * Reply chunk (a counted array):
168 * N elements: 171 * N elements:
169 * 1 - N - HLOO - HLOO - ... - HLOO 172 * 1 - N - HLOO - HLOO - ... - HLOO
173 *
174 * Returns positive RPC/RDMA header size, or negative errno.
170 */ 175 */
171 176
172static unsigned int 177static ssize_t
173rpcrdma_create_chunks(struct rpc_rqst *rqst, struct xdr_buf *target, 178rpcrdma_create_chunks(struct rpc_rqst *rqst, struct xdr_buf *target,
174 struct rpcrdma_msg *headerp, enum rpcrdma_chunktype type) 179 struct rpcrdma_msg *headerp, enum rpcrdma_chunktype type)
175{ 180{
176 struct rpcrdma_req *req = rpcr_to_rdmar(rqst); 181 struct rpcrdma_req *req = rpcr_to_rdmar(rqst);
177 struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(rqst->rq_xprt); 182 struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(rqst->rq_xprt);
178 int nsegs, nchunks = 0; 183 int n, nsegs, nchunks = 0;
179 unsigned int pos; 184 unsigned int pos;
180 struct rpcrdma_mr_seg *seg = req->rl_segments; 185 struct rpcrdma_mr_seg *seg = req->rl_segments;
181 struct rpcrdma_read_chunk *cur_rchunk = NULL; 186 struct rpcrdma_read_chunk *cur_rchunk = NULL;
@@ -201,11 +206,11 @@ rpcrdma_create_chunks(struct rpc_rqst *rqst, struct xdr_buf *target,
201 pos = target->head[0].iov_len; 206 pos = target->head[0].iov_len;
202 207
203 nsegs = rpcrdma_convert_iovs(target, pos, type, seg, RPCRDMA_MAX_SEGS); 208 nsegs = rpcrdma_convert_iovs(target, pos, type, seg, RPCRDMA_MAX_SEGS);
204 if (nsegs == 0) 209 if (nsegs < 0)
205 return 0; 210 return nsegs;
206 211
207 do { 212 do {
208 int n = rpcrdma_register_external(seg, nsegs, 213 n = rpcrdma_register_external(seg, nsegs,
209 cur_wchunk != NULL, r_xprt); 214 cur_wchunk != NULL, r_xprt);
210 if (n <= 0) 215 if (n <= 0)
211 goto out; 216 goto out;
@@ -277,7 +282,7 @@ out:
277 for (pos = 0; nchunks--;) 282 for (pos = 0; nchunks--;)
278 pos += rpcrdma_deregister_external( 283 pos += rpcrdma_deregister_external(
279 &req->rl_segments[pos], r_xprt); 284 &req->rl_segments[pos], r_xprt);
280 return 0; 285 return n;
281} 286}
282 287
283/* 288/*
@@ -359,6 +364,8 @@ rpcrdma_inline_pullup(struct rpc_rqst *rqst, int pad)
359 * [1] -- the RPC header/data, marshaled by RPC and the NFS protocol. 364 * [1] -- the RPC header/data, marshaled by RPC and the NFS protocol.
360 * [2] -- optional padding. 365 * [2] -- optional padding.
361 * [3] -- if padded, header only in [1] and data here. 366 * [3] -- if padded, header only in [1] and data here.
367 *
368 * Returns zero on success, otherwise a negative errno.
362 */ 369 */
363 370
364int 371int
@@ -368,7 +375,8 @@ rpcrdma_marshal_req(struct rpc_rqst *rqst)
368 struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(xprt); 375 struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(xprt);
369 struct rpcrdma_req *req = rpcr_to_rdmar(rqst); 376 struct rpcrdma_req *req = rpcr_to_rdmar(rqst);
370 char *base; 377 char *base;
371 size_t hdrlen, rpclen, padlen; 378 size_t rpclen, padlen;
379 ssize_t hdrlen;
372 enum rpcrdma_chunktype rtype, wtype; 380 enum rpcrdma_chunktype rtype, wtype;
373 struct rpcrdma_msg *headerp; 381 struct rpcrdma_msg *headerp;
374 382
@@ -439,7 +447,11 @@ rpcrdma_marshal_req(struct rpc_rqst *rqst)
439 /* The following simplification is not true forever */ 447 /* The following simplification is not true forever */
440 if (rtype != rpcrdma_noch && wtype == rpcrdma_replych) 448 if (rtype != rpcrdma_noch && wtype == rpcrdma_replych)
441 wtype = rpcrdma_noch; 449 wtype = rpcrdma_noch;
442 BUG_ON(rtype != rpcrdma_noch && wtype != rpcrdma_noch); 450 if (rtype != rpcrdma_noch && wtype != rpcrdma_noch) {
451 dprintk("RPC: %s: cannot marshal multiple chunk lists\n",
452 __func__);
453 return -EIO;
454 }
443 455
444 hdrlen = 28; /*sizeof *headerp;*/ 456 hdrlen = 28; /*sizeof *headerp;*/
445 padlen = 0; 457 padlen = 0;
@@ -464,8 +476,11 @@ rpcrdma_marshal_req(struct rpc_rqst *rqst)
464 headerp->rm_body.rm_padded.rm_pempty[1] = xdr_zero; 476 headerp->rm_body.rm_padded.rm_pempty[1] = xdr_zero;
465 headerp->rm_body.rm_padded.rm_pempty[2] = xdr_zero; 477 headerp->rm_body.rm_padded.rm_pempty[2] = xdr_zero;
466 hdrlen += 2 * sizeof(u32); /* extra words in padhdr */ 478 hdrlen += 2 * sizeof(u32); /* extra words in padhdr */
467 BUG_ON(wtype != rpcrdma_noch); 479 if (wtype != rpcrdma_noch) {
468 480 dprintk("RPC: %s: invalid chunk list\n",
481 __func__);
482 return -EIO;
483 }
469 } else { 484 } else {
470 headerp->rm_body.rm_nochunks.rm_empty[0] = xdr_zero; 485 headerp->rm_body.rm_nochunks.rm_empty[0] = xdr_zero;
471 headerp->rm_body.rm_nochunks.rm_empty[1] = xdr_zero; 486 headerp->rm_body.rm_nochunks.rm_empty[1] = xdr_zero;
@@ -500,9 +515,8 @@ rpcrdma_marshal_req(struct rpc_rqst *rqst)
500 hdrlen = rpcrdma_create_chunks(rqst, 515 hdrlen = rpcrdma_create_chunks(rqst,
501 &rqst->rq_rcv_buf, headerp, wtype); 516 &rqst->rq_rcv_buf, headerp, wtype);
502 } 517 }
503 518 if (hdrlen < 0)
504 if (hdrlen == 0) 519 return hdrlen;
505 return -1;
506 520
507 dprintk("RPC: %s: %s: hdrlen %zd rpclen %zd padlen %zd" 521 dprintk("RPC: %s: %s: hdrlen %zd rpclen %zd padlen %zd"
508 " headerp 0x%p base 0x%p lkey 0x%x\n", 522 " headerp 0x%p base 0x%p lkey 0x%x\n",
diff --git a/net/sunrpc/xprtrdma/transport.c b/net/sunrpc/xprtrdma/transport.c
index 93fe7753ff94..66f91f0d071a 100644
--- a/net/sunrpc/xprtrdma/transport.c
+++ b/net/sunrpc/xprtrdma/transport.c
@@ -595,13 +595,12 @@ xprt_rdma_send_request(struct rpc_task *task)
595 struct rpc_xprt *xprt = rqst->rq_xprt; 595 struct rpc_xprt *xprt = rqst->rq_xprt;
596 struct rpcrdma_req *req = rpcr_to_rdmar(rqst); 596 struct rpcrdma_req *req = rpcr_to_rdmar(rqst);
597 struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(xprt); 597 struct rpcrdma_xprt *r_xprt = rpcx_to_rdmax(xprt);
598 int rc;
598 599
599 /* marshal the send itself */ 600 if (req->rl_niovs == 0) {
600 if (req->rl_niovs == 0 && rpcrdma_marshal_req(rqst) != 0) { 601 rc = rpcrdma_marshal_req(rqst);
601 r_xprt->rx_stats.failed_marshal_count++; 602 if (rc < 0)
602 dprintk("RPC: %s: rpcrdma_marshal_req failed\n", 603 goto failed_marshal;
603 __func__);
604 return -EIO;
605 } 604 }
606 605
607 if (req->rl_reply == NULL) /* e.g. reconnection */ 606 if (req->rl_reply == NULL) /* e.g. reconnection */
@@ -625,6 +624,12 @@ xprt_rdma_send_request(struct rpc_task *task)
625 rqst->rq_bytes_sent = 0; 624 rqst->rq_bytes_sent = 0;
626 return 0; 625 return 0;
627 626
627failed_marshal:
628 r_xprt->rx_stats.failed_marshal_count++;
629 dprintk("RPC: %s: rpcrdma_marshal_req failed, status %i\n",
630 __func__, rc);
631 if (rc == -EIO)
632 return -EIO;
628drop_connection: 633drop_connection:
629 xprt_disconnect_done(xprt); 634 xprt_disconnect_done(xprt);
630 return -ENOTCONN; /* implies disconnect */ 635 return -ENOTCONN; /* implies disconnect */