diff options
author | Ralph Campbell <ralph.campbell@qlogic.com> | 2010-05-24 00:44:54 -0400 |
---|---|---|
committer | Roland Dreier <rolandd@cisco.com> | 2010-05-24 00:44:54 -0400 |
commit | f931551bafe1f10ded7f5282e2aa162c267a2e5d (patch) | |
tree | e81b4656a8116abf5fd0bc0bbc46560aff536159 /drivers/infiniband/hw/qib/qib_uc.c | |
parent | 9a6edb60ec10d86b1025a0cdad68fd89f1ddaf02 (diff) |
IB/qib: Add new qib driver for QLogic PCIe InfiniBand adapters
Add a low-level IB driver for QLogic PCIe adapters.
Signed-off-by: Ralph Campbell <ralph.campbell@qlogic.com>
Signed-off-by: Roland Dreier <rolandd@cisco.com>
Diffstat (limited to 'drivers/infiniband/hw/qib/qib_uc.c')
-rw-r--r-- | drivers/infiniband/hw/qib/qib_uc.c | 555 |
1 files changed, 555 insertions, 0 deletions
diff --git a/drivers/infiniband/hw/qib/qib_uc.c b/drivers/infiniband/hw/qib/qib_uc.c new file mode 100644 index 000000000000..6c7fe78cca64 --- /dev/null +++ b/drivers/infiniband/hw/qib/qib_uc.c | |||
@@ -0,0 +1,555 @@ | |||
1 | /* | ||
2 | * Copyright (c) 2006, 2007, 2008, 2009, 2010 QLogic Corporation. | ||
3 | * All rights reserved. | ||
4 | * Copyright (c) 2005, 2006 PathScale, Inc. All rights reserved. | ||
5 | * | ||
6 | * This software is available to you under a choice of one of two | ||
7 | * licenses. You may choose to be licensed under the terms of the GNU | ||
8 | * General Public License (GPL) Version 2, available from the file | ||
9 | * COPYING in the main directory of this source tree, or the | ||
10 | * OpenIB.org BSD license below: | ||
11 | * | ||
12 | * Redistribution and use in source and binary forms, with or | ||
13 | * without modification, are permitted provided that the following | ||
14 | * conditions are met: | ||
15 | * | ||
16 | * - Redistributions of source code must retain the above | ||
17 | * copyright notice, this list of conditions and the following | ||
18 | * disclaimer. | ||
19 | * | ||
20 | * - Redistributions in binary form must reproduce the above | ||
21 | * copyright notice, this list of conditions and the following | ||
22 | * disclaimer in the documentation and/or other materials | ||
23 | * provided with the distribution. | ||
24 | * | ||
25 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, | ||
26 | * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF | ||
27 | * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND | ||
28 | * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS | ||
29 | * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN | ||
30 | * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN | ||
31 | * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE | ||
32 | * SOFTWARE. | ||
33 | */ | ||
34 | |||
35 | #include "qib.h" | ||
36 | |||
37 | /* cut down ridiculously long IB macro names */ | ||
38 | #define OP(x) IB_OPCODE_UC_##x | ||
39 | |||
40 | /** | ||
41 | * qib_make_uc_req - construct a request packet (SEND, RDMA write) | ||
42 | * @qp: a pointer to the QP | ||
43 | * | ||
44 | * Return 1 if constructed; otherwise, return 0. | ||
45 | */ | ||
46 | int qib_make_uc_req(struct qib_qp *qp) | ||
47 | { | ||
48 | struct qib_other_headers *ohdr; | ||
49 | struct qib_swqe *wqe; | ||
50 | unsigned long flags; | ||
51 | u32 hwords; | ||
52 | u32 bth0; | ||
53 | u32 len; | ||
54 | u32 pmtu = ib_mtu_enum_to_int(qp->path_mtu); | ||
55 | int ret = 0; | ||
56 | |||
57 | spin_lock_irqsave(&qp->s_lock, flags); | ||
58 | |||
59 | if (!(ib_qib_state_ops[qp->state] & QIB_PROCESS_SEND_OK)) { | ||
60 | if (!(ib_qib_state_ops[qp->state] & QIB_FLUSH_SEND)) | ||
61 | goto bail; | ||
62 | /* We are in the error state, flush the work request. */ | ||
63 | if (qp->s_last == qp->s_head) | ||
64 | goto bail; | ||
65 | /* If DMAs are in progress, we can't flush immediately. */ | ||
66 | if (atomic_read(&qp->s_dma_busy)) { | ||
67 | qp->s_flags |= QIB_S_WAIT_DMA; | ||
68 | goto bail; | ||
69 | } | ||
70 | wqe = get_swqe_ptr(qp, qp->s_last); | ||
71 | qib_send_complete(qp, wqe, IB_WC_WR_FLUSH_ERR); | ||
72 | goto done; | ||
73 | } | ||
74 | |||
75 | ohdr = &qp->s_hdr.u.oth; | ||
76 | if (qp->remote_ah_attr.ah_flags & IB_AH_GRH) | ||
77 | ohdr = &qp->s_hdr.u.l.oth; | ||
78 | |||
79 | /* header size in 32-bit words LRH+BTH = (8+12)/4. */ | ||
80 | hwords = 5; | ||
81 | bth0 = 0; | ||
82 | |||
83 | /* Get the next send request. */ | ||
84 | wqe = get_swqe_ptr(qp, qp->s_cur); | ||
85 | qp->s_wqe = NULL; | ||
86 | switch (qp->s_state) { | ||
87 | default: | ||
88 | if (!(ib_qib_state_ops[qp->state] & | ||
89 | QIB_PROCESS_NEXT_SEND_OK)) | ||
90 | goto bail; | ||
91 | /* Check if send work queue is empty. */ | ||
92 | if (qp->s_cur == qp->s_head) | ||
93 | goto bail; | ||
94 | /* | ||
95 | * Start a new request. | ||
96 | */ | ||
97 | wqe->psn = qp->s_next_psn; | ||
98 | qp->s_psn = qp->s_next_psn; | ||
99 | qp->s_sge.sge = wqe->sg_list[0]; | ||
100 | qp->s_sge.sg_list = wqe->sg_list + 1; | ||
101 | qp->s_sge.num_sge = wqe->wr.num_sge; | ||
102 | qp->s_sge.total_len = wqe->length; | ||
103 | len = wqe->length; | ||
104 | qp->s_len = len; | ||
105 | switch (wqe->wr.opcode) { | ||
106 | case IB_WR_SEND: | ||
107 | case IB_WR_SEND_WITH_IMM: | ||
108 | if (len > pmtu) { | ||
109 | qp->s_state = OP(SEND_FIRST); | ||
110 | len = pmtu; | ||
111 | break; | ||
112 | } | ||
113 | if (wqe->wr.opcode == IB_WR_SEND) | ||
114 | qp->s_state = OP(SEND_ONLY); | ||
115 | else { | ||
116 | qp->s_state = | ||
117 | OP(SEND_ONLY_WITH_IMMEDIATE); | ||
118 | /* Immediate data comes after the BTH */ | ||
119 | ohdr->u.imm_data = wqe->wr.ex.imm_data; | ||
120 | hwords += 1; | ||
121 | } | ||
122 | if (wqe->wr.send_flags & IB_SEND_SOLICITED) | ||
123 | bth0 |= IB_BTH_SOLICITED; | ||
124 | qp->s_wqe = wqe; | ||
125 | if (++qp->s_cur >= qp->s_size) | ||
126 | qp->s_cur = 0; | ||
127 | break; | ||
128 | |||
129 | case IB_WR_RDMA_WRITE: | ||
130 | case IB_WR_RDMA_WRITE_WITH_IMM: | ||
131 | ohdr->u.rc.reth.vaddr = | ||
132 | cpu_to_be64(wqe->wr.wr.rdma.remote_addr); | ||
133 | ohdr->u.rc.reth.rkey = | ||
134 | cpu_to_be32(wqe->wr.wr.rdma.rkey); | ||
135 | ohdr->u.rc.reth.length = cpu_to_be32(len); | ||
136 | hwords += sizeof(struct ib_reth) / 4; | ||
137 | if (len > pmtu) { | ||
138 | qp->s_state = OP(RDMA_WRITE_FIRST); | ||
139 | len = pmtu; | ||
140 | break; | ||
141 | } | ||
142 | if (wqe->wr.opcode == IB_WR_RDMA_WRITE) | ||
143 | qp->s_state = OP(RDMA_WRITE_ONLY); | ||
144 | else { | ||
145 | qp->s_state = | ||
146 | OP(RDMA_WRITE_ONLY_WITH_IMMEDIATE); | ||
147 | /* Immediate data comes after the RETH */ | ||
148 | ohdr->u.rc.imm_data = wqe->wr.ex.imm_data; | ||
149 | hwords += 1; | ||
150 | if (wqe->wr.send_flags & IB_SEND_SOLICITED) | ||
151 | bth0 |= IB_BTH_SOLICITED; | ||
152 | } | ||
153 | qp->s_wqe = wqe; | ||
154 | if (++qp->s_cur >= qp->s_size) | ||
155 | qp->s_cur = 0; | ||
156 | break; | ||
157 | |||
158 | default: | ||
159 | goto bail; | ||
160 | } | ||
161 | break; | ||
162 | |||
163 | case OP(SEND_FIRST): | ||
164 | qp->s_state = OP(SEND_MIDDLE); | ||
165 | /* FALLTHROUGH */ | ||
166 | case OP(SEND_MIDDLE): | ||
167 | len = qp->s_len; | ||
168 | if (len > pmtu) { | ||
169 | len = pmtu; | ||
170 | break; | ||
171 | } | ||
172 | if (wqe->wr.opcode == IB_WR_SEND) | ||
173 | qp->s_state = OP(SEND_LAST); | ||
174 | else { | ||
175 | qp->s_state = OP(SEND_LAST_WITH_IMMEDIATE); | ||
176 | /* Immediate data comes after the BTH */ | ||
177 | ohdr->u.imm_data = wqe->wr.ex.imm_data; | ||
178 | hwords += 1; | ||
179 | } | ||
180 | if (wqe->wr.send_flags & IB_SEND_SOLICITED) | ||
181 | bth0 |= IB_BTH_SOLICITED; | ||
182 | qp->s_wqe = wqe; | ||
183 | if (++qp->s_cur >= qp->s_size) | ||
184 | qp->s_cur = 0; | ||
185 | break; | ||
186 | |||
187 | case OP(RDMA_WRITE_FIRST): | ||
188 | qp->s_state = OP(RDMA_WRITE_MIDDLE); | ||
189 | /* FALLTHROUGH */ | ||
190 | case OP(RDMA_WRITE_MIDDLE): | ||
191 | len = qp->s_len; | ||
192 | if (len > pmtu) { | ||
193 | len = pmtu; | ||
194 | break; | ||
195 | } | ||
196 | if (wqe->wr.opcode == IB_WR_RDMA_WRITE) | ||
197 | qp->s_state = OP(RDMA_WRITE_LAST); | ||
198 | else { | ||
199 | qp->s_state = | ||
200 | OP(RDMA_WRITE_LAST_WITH_IMMEDIATE); | ||
201 | /* Immediate data comes after the BTH */ | ||
202 | ohdr->u.imm_data = wqe->wr.ex.imm_data; | ||
203 | hwords += 1; | ||
204 | if (wqe->wr.send_flags & IB_SEND_SOLICITED) | ||
205 | bth0 |= IB_BTH_SOLICITED; | ||
206 | } | ||
207 | qp->s_wqe = wqe; | ||
208 | if (++qp->s_cur >= qp->s_size) | ||
209 | qp->s_cur = 0; | ||
210 | break; | ||
211 | } | ||
212 | qp->s_len -= len; | ||
213 | qp->s_hdrwords = hwords; | ||
214 | qp->s_cur_sge = &qp->s_sge; | ||
215 | qp->s_cur_size = len; | ||
216 | qib_make_ruc_header(qp, ohdr, bth0 | (qp->s_state << 24), | ||
217 | qp->s_next_psn++ & QIB_PSN_MASK); | ||
218 | done: | ||
219 | ret = 1; | ||
220 | goto unlock; | ||
221 | |||
222 | bail: | ||
223 | qp->s_flags &= ~QIB_S_BUSY; | ||
224 | unlock: | ||
225 | spin_unlock_irqrestore(&qp->s_lock, flags); | ||
226 | return ret; | ||
227 | } | ||
228 | |||
229 | /** | ||
230 | * qib_uc_rcv - handle an incoming UC packet | ||
231 | * @ibp: the port the packet came in on | ||
232 | * @hdr: the header of the packet | ||
233 | * @has_grh: true if the packet has a GRH | ||
234 | * @data: the packet data | ||
235 | * @tlen: the length of the packet | ||
236 | * @qp: the QP for this packet. | ||
237 | * | ||
238 | * This is called from qib_qp_rcv() to process an incoming UC packet | ||
239 | * for the given QP. | ||
240 | * Called at interrupt level. | ||
241 | */ | ||
242 | void qib_uc_rcv(struct qib_ibport *ibp, struct qib_ib_header *hdr, | ||
243 | int has_grh, void *data, u32 tlen, struct qib_qp *qp) | ||
244 | { | ||
245 | struct qib_other_headers *ohdr; | ||
246 | unsigned long flags; | ||
247 | u32 opcode; | ||
248 | u32 hdrsize; | ||
249 | u32 psn; | ||
250 | u32 pad; | ||
251 | struct ib_wc wc; | ||
252 | u32 pmtu = ib_mtu_enum_to_int(qp->path_mtu); | ||
253 | struct ib_reth *reth; | ||
254 | int ret; | ||
255 | |||
256 | /* Check for GRH */ | ||
257 | if (!has_grh) { | ||
258 | ohdr = &hdr->u.oth; | ||
259 | hdrsize = 8 + 12; /* LRH + BTH */ | ||
260 | } else { | ||
261 | ohdr = &hdr->u.l.oth; | ||
262 | hdrsize = 8 + 40 + 12; /* LRH + GRH + BTH */ | ||
263 | } | ||
264 | |||
265 | opcode = be32_to_cpu(ohdr->bth[0]); | ||
266 | spin_lock_irqsave(&qp->s_lock, flags); | ||
267 | if (qib_ruc_check_hdr(ibp, hdr, has_grh, qp, opcode)) | ||
268 | goto sunlock; | ||
269 | spin_unlock_irqrestore(&qp->s_lock, flags); | ||
270 | |||
271 | psn = be32_to_cpu(ohdr->bth[2]); | ||
272 | opcode >>= 24; | ||
273 | memset(&wc, 0, sizeof wc); | ||
274 | |||
275 | /* Prevent simultaneous processing after APM on different CPUs */ | ||
276 | spin_lock(&qp->r_lock); | ||
277 | |||
278 | /* Compare the PSN verses the expected PSN. */ | ||
279 | if (unlikely(qib_cmp24(psn, qp->r_psn) != 0)) { | ||
280 | /* | ||
281 | * Handle a sequence error. | ||
282 | * Silently drop any current message. | ||
283 | */ | ||
284 | qp->r_psn = psn; | ||
285 | inv: | ||
286 | if (qp->r_state == OP(SEND_FIRST) || | ||
287 | qp->r_state == OP(SEND_MIDDLE)) { | ||
288 | set_bit(QIB_R_REWIND_SGE, &qp->r_aflags); | ||
289 | qp->r_sge.num_sge = 0; | ||
290 | } else | ||
291 | while (qp->r_sge.num_sge) { | ||
292 | atomic_dec(&qp->r_sge.sge.mr->refcount); | ||
293 | if (--qp->r_sge.num_sge) | ||
294 | qp->r_sge.sge = *qp->r_sge.sg_list++; | ||
295 | } | ||
296 | qp->r_state = OP(SEND_LAST); | ||
297 | switch (opcode) { | ||
298 | case OP(SEND_FIRST): | ||
299 | case OP(SEND_ONLY): | ||
300 | case OP(SEND_ONLY_WITH_IMMEDIATE): | ||
301 | goto send_first; | ||
302 | |||
303 | case OP(RDMA_WRITE_FIRST): | ||
304 | case OP(RDMA_WRITE_ONLY): | ||
305 | case OP(RDMA_WRITE_ONLY_WITH_IMMEDIATE): | ||
306 | goto rdma_first; | ||
307 | |||
308 | default: | ||
309 | goto drop; | ||
310 | } | ||
311 | } | ||
312 | |||
313 | /* Check for opcode sequence errors. */ | ||
314 | switch (qp->r_state) { | ||
315 | case OP(SEND_FIRST): | ||
316 | case OP(SEND_MIDDLE): | ||
317 | if (opcode == OP(SEND_MIDDLE) || | ||
318 | opcode == OP(SEND_LAST) || | ||
319 | opcode == OP(SEND_LAST_WITH_IMMEDIATE)) | ||
320 | break; | ||
321 | goto inv; | ||
322 | |||
323 | case OP(RDMA_WRITE_FIRST): | ||
324 | case OP(RDMA_WRITE_MIDDLE): | ||
325 | if (opcode == OP(RDMA_WRITE_MIDDLE) || | ||
326 | opcode == OP(RDMA_WRITE_LAST) || | ||
327 | opcode == OP(RDMA_WRITE_LAST_WITH_IMMEDIATE)) | ||
328 | break; | ||
329 | goto inv; | ||
330 | |||
331 | default: | ||
332 | if (opcode == OP(SEND_FIRST) || | ||
333 | opcode == OP(SEND_ONLY) || | ||
334 | opcode == OP(SEND_ONLY_WITH_IMMEDIATE) || | ||
335 | opcode == OP(RDMA_WRITE_FIRST) || | ||
336 | opcode == OP(RDMA_WRITE_ONLY) || | ||
337 | opcode == OP(RDMA_WRITE_ONLY_WITH_IMMEDIATE)) | ||
338 | break; | ||
339 | goto inv; | ||
340 | } | ||
341 | |||
342 | if (qp->state == IB_QPS_RTR && !(qp->r_flags & QIB_R_COMM_EST)) { | ||
343 | qp->r_flags |= QIB_R_COMM_EST; | ||
344 | if (qp->ibqp.event_handler) { | ||
345 | struct ib_event ev; | ||
346 | |||
347 | ev.device = qp->ibqp.device; | ||
348 | ev.element.qp = &qp->ibqp; | ||
349 | ev.event = IB_EVENT_COMM_EST; | ||
350 | qp->ibqp.event_handler(&ev, qp->ibqp.qp_context); | ||
351 | } | ||
352 | } | ||
353 | |||
354 | /* OK, process the packet. */ | ||
355 | switch (opcode) { | ||
356 | case OP(SEND_FIRST): | ||
357 | case OP(SEND_ONLY): | ||
358 | case OP(SEND_ONLY_WITH_IMMEDIATE): | ||
359 | send_first: | ||
360 | if (test_and_clear_bit(QIB_R_REWIND_SGE, &qp->r_aflags)) | ||
361 | qp->r_sge = qp->s_rdma_read_sge; | ||
362 | else { | ||
363 | ret = qib_get_rwqe(qp, 0); | ||
364 | if (ret < 0) | ||
365 | goto op_err; | ||
366 | if (!ret) | ||
367 | goto drop; | ||
368 | /* | ||
369 | * qp->s_rdma_read_sge will be the owner | ||
370 | * of the mr references. | ||
371 | */ | ||
372 | qp->s_rdma_read_sge = qp->r_sge; | ||
373 | } | ||
374 | qp->r_rcv_len = 0; | ||
375 | if (opcode == OP(SEND_ONLY)) | ||
376 | goto send_last; | ||
377 | else if (opcode == OP(SEND_ONLY_WITH_IMMEDIATE)) | ||
378 | goto send_last_imm; | ||
379 | /* FALLTHROUGH */ | ||
380 | case OP(SEND_MIDDLE): | ||
381 | /* Check for invalid length PMTU or posted rwqe len. */ | ||
382 | if (unlikely(tlen != (hdrsize + pmtu + 4))) | ||
383 | goto rewind; | ||
384 | qp->r_rcv_len += pmtu; | ||
385 | if (unlikely(qp->r_rcv_len > qp->r_len)) | ||
386 | goto rewind; | ||
387 | qib_copy_sge(&qp->r_sge, data, pmtu, 0); | ||
388 | break; | ||
389 | |||
390 | case OP(SEND_LAST_WITH_IMMEDIATE): | ||
391 | send_last_imm: | ||
392 | wc.ex.imm_data = ohdr->u.imm_data; | ||
393 | hdrsize += 4; | ||
394 | wc.wc_flags = IB_WC_WITH_IMM; | ||
395 | /* FALLTHROUGH */ | ||
396 | case OP(SEND_LAST): | ||
397 | send_last: | ||
398 | /* Get the number of bytes the message was padded by. */ | ||
399 | pad = (be32_to_cpu(ohdr->bth[0]) >> 20) & 3; | ||
400 | /* Check for invalid length. */ | ||
401 | /* XXX LAST len should be >= 1 */ | ||
402 | if (unlikely(tlen < (hdrsize + pad + 4))) | ||
403 | goto rewind; | ||
404 | /* Don't count the CRC. */ | ||
405 | tlen -= (hdrsize + pad + 4); | ||
406 | wc.byte_len = tlen + qp->r_rcv_len; | ||
407 | if (unlikely(wc.byte_len > qp->r_len)) | ||
408 | goto rewind; | ||
409 | wc.opcode = IB_WC_RECV; | ||
410 | last_imm: | ||
411 | qib_copy_sge(&qp->r_sge, data, tlen, 0); | ||
412 | while (qp->s_rdma_read_sge.num_sge) { | ||
413 | atomic_dec(&qp->s_rdma_read_sge.sge.mr->refcount); | ||
414 | if (--qp->s_rdma_read_sge.num_sge) | ||
415 | qp->s_rdma_read_sge.sge = | ||
416 | *qp->s_rdma_read_sge.sg_list++; | ||
417 | } | ||
418 | wc.wr_id = qp->r_wr_id; | ||
419 | wc.status = IB_WC_SUCCESS; | ||
420 | wc.qp = &qp->ibqp; | ||
421 | wc.src_qp = qp->remote_qpn; | ||
422 | wc.slid = qp->remote_ah_attr.dlid; | ||
423 | wc.sl = qp->remote_ah_attr.sl; | ||
424 | /* Signal completion event if the solicited bit is set. */ | ||
425 | qib_cq_enter(to_icq(qp->ibqp.recv_cq), &wc, | ||
426 | (ohdr->bth[0] & | ||
427 | cpu_to_be32(IB_BTH_SOLICITED)) != 0); | ||
428 | break; | ||
429 | |||
430 | case OP(RDMA_WRITE_FIRST): | ||
431 | case OP(RDMA_WRITE_ONLY): | ||
432 | case OP(RDMA_WRITE_ONLY_WITH_IMMEDIATE): /* consume RWQE */ | ||
433 | rdma_first: | ||
434 | if (unlikely(!(qp->qp_access_flags & | ||
435 | IB_ACCESS_REMOTE_WRITE))) { | ||
436 | goto drop; | ||
437 | } | ||
438 | reth = &ohdr->u.rc.reth; | ||
439 | hdrsize += sizeof(*reth); | ||
440 | qp->r_len = be32_to_cpu(reth->length); | ||
441 | qp->r_rcv_len = 0; | ||
442 | qp->r_sge.sg_list = NULL; | ||
443 | if (qp->r_len != 0) { | ||
444 | u32 rkey = be32_to_cpu(reth->rkey); | ||
445 | u64 vaddr = be64_to_cpu(reth->vaddr); | ||
446 | int ok; | ||
447 | |||
448 | /* Check rkey */ | ||
449 | ok = qib_rkey_ok(qp, &qp->r_sge.sge, qp->r_len, | ||
450 | vaddr, rkey, IB_ACCESS_REMOTE_WRITE); | ||
451 | if (unlikely(!ok)) | ||
452 | goto drop; | ||
453 | qp->r_sge.num_sge = 1; | ||
454 | } else { | ||
455 | qp->r_sge.num_sge = 0; | ||
456 | qp->r_sge.sge.mr = NULL; | ||
457 | qp->r_sge.sge.vaddr = NULL; | ||
458 | qp->r_sge.sge.length = 0; | ||
459 | qp->r_sge.sge.sge_length = 0; | ||
460 | } | ||
461 | if (opcode == OP(RDMA_WRITE_ONLY)) | ||
462 | goto rdma_last; | ||
463 | else if (opcode == OP(RDMA_WRITE_ONLY_WITH_IMMEDIATE)) | ||
464 | goto rdma_last_imm; | ||
465 | /* FALLTHROUGH */ | ||
466 | case OP(RDMA_WRITE_MIDDLE): | ||
467 | /* Check for invalid length PMTU or posted rwqe len. */ | ||
468 | if (unlikely(tlen != (hdrsize + pmtu + 4))) | ||
469 | goto drop; | ||
470 | qp->r_rcv_len += pmtu; | ||
471 | if (unlikely(qp->r_rcv_len > qp->r_len)) | ||
472 | goto drop; | ||
473 | qib_copy_sge(&qp->r_sge, data, pmtu, 1); | ||
474 | break; | ||
475 | |||
476 | case OP(RDMA_WRITE_LAST_WITH_IMMEDIATE): | ||
477 | rdma_last_imm: | ||
478 | wc.ex.imm_data = ohdr->u.imm_data; | ||
479 | hdrsize += 4; | ||
480 | wc.wc_flags = IB_WC_WITH_IMM; | ||
481 | |||
482 | /* Get the number of bytes the message was padded by. */ | ||
483 | pad = (be32_to_cpu(ohdr->bth[0]) >> 20) & 3; | ||
484 | /* Check for invalid length. */ | ||
485 | /* XXX LAST len should be >= 1 */ | ||
486 | if (unlikely(tlen < (hdrsize + pad + 4))) | ||
487 | goto drop; | ||
488 | /* Don't count the CRC. */ | ||
489 | tlen -= (hdrsize + pad + 4); | ||
490 | if (unlikely(tlen + qp->r_rcv_len != qp->r_len)) | ||
491 | goto drop; | ||
492 | if (test_and_clear_bit(QIB_R_REWIND_SGE, &qp->r_aflags)) | ||
493 | while (qp->s_rdma_read_sge.num_sge) { | ||
494 | atomic_dec(&qp->s_rdma_read_sge.sge.mr-> | ||
495 | refcount); | ||
496 | if (--qp->s_rdma_read_sge.num_sge) | ||
497 | qp->s_rdma_read_sge.sge = | ||
498 | *qp->s_rdma_read_sge.sg_list++; | ||
499 | } | ||
500 | else { | ||
501 | ret = qib_get_rwqe(qp, 1); | ||
502 | if (ret < 0) | ||
503 | goto op_err; | ||
504 | if (!ret) | ||
505 | goto drop; | ||
506 | } | ||
507 | wc.byte_len = qp->r_len; | ||
508 | wc.opcode = IB_WC_RECV_RDMA_WITH_IMM; | ||
509 | goto last_imm; | ||
510 | |||
511 | case OP(RDMA_WRITE_LAST): | ||
512 | rdma_last: | ||
513 | /* Get the number of bytes the message was padded by. */ | ||
514 | pad = (be32_to_cpu(ohdr->bth[0]) >> 20) & 3; | ||
515 | /* Check for invalid length. */ | ||
516 | /* XXX LAST len should be >= 1 */ | ||
517 | if (unlikely(tlen < (hdrsize + pad + 4))) | ||
518 | goto drop; | ||
519 | /* Don't count the CRC. */ | ||
520 | tlen -= (hdrsize + pad + 4); | ||
521 | if (unlikely(tlen + qp->r_rcv_len != qp->r_len)) | ||
522 | goto drop; | ||
523 | qib_copy_sge(&qp->r_sge, data, tlen, 1); | ||
524 | while (qp->r_sge.num_sge) { | ||
525 | atomic_dec(&qp->r_sge.sge.mr->refcount); | ||
526 | if (--qp->r_sge.num_sge) | ||
527 | qp->r_sge.sge = *qp->r_sge.sg_list++; | ||
528 | } | ||
529 | break; | ||
530 | |||
531 | default: | ||
532 | /* Drop packet for unknown opcodes. */ | ||
533 | goto drop; | ||
534 | } | ||
535 | qp->r_psn++; | ||
536 | qp->r_state = opcode; | ||
537 | spin_unlock(&qp->r_lock); | ||
538 | return; | ||
539 | |||
540 | rewind: | ||
541 | set_bit(QIB_R_REWIND_SGE, &qp->r_aflags); | ||
542 | qp->r_sge.num_sge = 0; | ||
543 | drop: | ||
544 | ibp->n_pkt_drops++; | ||
545 | spin_unlock(&qp->r_lock); | ||
546 | return; | ||
547 | |||
548 | op_err: | ||
549 | qib_rc_error(qp, IB_WC_LOC_QP_OP_ERR); | ||
550 | spin_unlock(&qp->r_lock); | ||
551 | return; | ||
552 | |||
553 | sunlock: | ||
554 | spin_unlock_irqrestore(&qp->s_lock, flags); | ||
555 | } | ||