aboutsummaryrefslogtreecommitdiffstats
path: root/drivers/infiniband
diff options
context:
space:
mode:
Diffstat (limited to 'drivers/infiniband')
-rw-r--r--drivers/infiniband/ulp/iser/iscsi_iser.c114
-rw-r--r--drivers/infiniband/ulp/iser/iscsi_iser.h24
-rw-r--r--drivers/infiniband/ulp/iser/iser_initiator.c190
-rw-r--r--drivers/infiniband/ulp/iser/iser_memory.c77
4 files changed, 203 insertions, 202 deletions
diff --git a/drivers/infiniband/ulp/iser/iscsi_iser.c b/drivers/infiniband/ulp/iser/iscsi_iser.c
index baecca1ed42a..86d9c42f0d33 100644
--- a/drivers/infiniband/ulp/iser/iscsi_iser.c
+++ b/drivers/infiniband/ulp/iser/iscsi_iser.c
@@ -124,33 +124,33 @@ error:
124 124
125 125
126/** 126/**
127 * iscsi_iser_task_init - Initialize ctask 127 * iscsi_iser_task_init - Initialize task
128 * @ctask: iscsi ctask 128 * @task: iscsi task
129 * 129 *
130 * Initialize the ctask for the scsi command or mgmt command. 130 * Initialize the task for the scsi command or mgmt command.
131 */ 131 */
132static int 132static int
133iscsi_iser_task_init(struct iscsi_cmd_task *ctask) 133iscsi_iser_task_init(struct iscsi_task *task)
134{ 134{
135 struct iscsi_iser_conn *iser_conn = ctask->conn->dd_data; 135 struct iscsi_iser_conn *iser_conn = task->conn->dd_data;
136 struct iscsi_iser_cmd_task *iser_ctask = ctask->dd_data; 136 struct iscsi_iser_task *iser_task = task->dd_data;
137 137
138 /* mgmt ctask */ 138 /* mgmt task */
139 if (!ctask->sc) { 139 if (!task->sc) {
140 iser_ctask->desc.data = ctask->data; 140 iser_task->desc.data = task->data;
141 return 0; 141 return 0;
142 } 142 }
143 143
144 iser_ctask->command_sent = 0; 144 iser_task->command_sent = 0;
145 iser_ctask->iser_conn = iser_conn; 145 iser_task->iser_conn = iser_conn;
146 iser_ctask_rdma_init(iser_ctask); 146 iser_task_rdma_init(iser_task);
147 return 0; 147 return 0;
148} 148}
149 149
150/** 150/**
151 * iscsi_iser_mtask_xmit - xmit management(immediate) ctask 151 * iscsi_iser_mtask_xmit - xmit management(immediate) task
152 * @conn: iscsi connection 152 * @conn: iscsi connection
153 * @ctask: ctask management ctask 153 * @task: task management task
154 * 154 *
155 * Notes: 155 * Notes:
156 * The function can return -EAGAIN in which case caller must 156 * The function can return -EAGAIN in which case caller must
@@ -159,19 +159,19 @@ iscsi_iser_task_init(struct iscsi_cmd_task *ctask)
159 * 159 *
160 **/ 160 **/
161static int 161static int
162iscsi_iser_mtask_xmit(struct iscsi_conn *conn, struct iscsi_cmd_task *ctask) 162iscsi_iser_mtask_xmit(struct iscsi_conn *conn, struct iscsi_task *task)
163{ 163{
164 int error = 0; 164 int error = 0;
165 165
166 debug_scsi("ctask deq [cid %d itt 0x%x]\n", conn->id, ctask->itt); 166 debug_scsi("task deq [cid %d itt 0x%x]\n", conn->id, task->itt);
167 167
168 error = iser_send_control(conn, ctask); 168 error = iser_send_control(conn, task);
169 169
170 /* since iser xmits control with zero copy, ctasks can not be recycled 170 /* since iser xmits control with zero copy, tasks can not be recycled
171 * right after sending them. 171 * right after sending them.
172 * The recycling scheme is based on whether a response is expected 172 * The recycling scheme is based on whether a response is expected
173 * - if yes, the ctask is recycled at iscsi_complete_pdu 173 * - if yes, the task is recycled at iscsi_complete_pdu
174 * - if no, the ctask is recycled at iser_snd_completion 174 * - if no, the task is recycled at iser_snd_completion
175 */ 175 */
176 if (error && error != -ENOBUFS) 176 if (error && error != -ENOBUFS)
177 iscsi_conn_failure(conn, ISCSI_ERR_CONN_FAILED); 177 iscsi_conn_failure(conn, ISCSI_ERR_CONN_FAILED);
@@ -181,27 +181,27 @@ iscsi_iser_mtask_xmit(struct iscsi_conn *conn, struct iscsi_cmd_task *ctask)
181 181
182static int 182static int
183iscsi_iser_task_xmit_unsol_data(struct iscsi_conn *conn, 183iscsi_iser_task_xmit_unsol_data(struct iscsi_conn *conn,
184 struct iscsi_cmd_task *ctask) 184 struct iscsi_task *task)
185{ 185{
186 struct iscsi_data hdr; 186 struct iscsi_data hdr;
187 int error = 0; 187 int error = 0;
188 188
189 /* Send data-out PDUs while there's still unsolicited data to send */ 189 /* Send data-out PDUs while there's still unsolicited data to send */
190 while (ctask->unsol_count > 0) { 190 while (task->unsol_count > 0) {
191 iscsi_prep_unsolicit_data_pdu(ctask, &hdr); 191 iscsi_prep_unsolicit_data_pdu(task, &hdr);
192 debug_scsi("Sending data-out: itt 0x%x, data count %d\n", 192 debug_scsi("Sending data-out: itt 0x%x, data count %d\n",
193 hdr.itt, ctask->data_count); 193 hdr.itt, task->data_count);
194 194
195 /* the buffer description has been passed with the command */ 195 /* the buffer description has been passed with the command */
196 /* Send the command */ 196 /* Send the command */
197 error = iser_send_data_out(conn, ctask, &hdr); 197 error = iser_send_data_out(conn, task, &hdr);
198 if (error) { 198 if (error) {
199 ctask->unsol_datasn--; 199 task->unsol_datasn--;
200 goto iscsi_iser_task_xmit_unsol_data_exit; 200 goto iscsi_iser_task_xmit_unsol_data_exit;
201 } 201 }
202 ctask->unsol_count -= ctask->data_count; 202 task->unsol_count -= task->data_count;
203 debug_scsi("Need to send %d more as data-out PDUs\n", 203 debug_scsi("Need to send %d more as data-out PDUs\n",
204 ctask->unsol_count); 204 task->unsol_count);
205 } 205 }
206 206
207iscsi_iser_task_xmit_unsol_data_exit: 207iscsi_iser_task_xmit_unsol_data_exit:
@@ -209,37 +209,37 @@ iscsi_iser_task_xmit_unsol_data_exit:
209} 209}
210 210
211static int 211static int
212iscsi_iser_task_xmit(struct iscsi_cmd_task *ctask) 212iscsi_iser_task_xmit(struct iscsi_task *task)
213{ 213{
214 struct iscsi_conn *conn = ctask->conn; 214 struct iscsi_conn *conn = task->conn;
215 struct iscsi_iser_cmd_task *iser_ctask = ctask->dd_data; 215 struct iscsi_iser_task *iser_task = task->dd_data;
216 int error = 0; 216 int error = 0;
217 217
218 if (!ctask->sc) 218 if (!task->sc)
219 return iscsi_iser_mtask_xmit(conn, ctask); 219 return iscsi_iser_mtask_xmit(conn, task);
220 220
221 if (ctask->sc->sc_data_direction == DMA_TO_DEVICE) { 221 if (task->sc->sc_data_direction == DMA_TO_DEVICE) {
222 BUG_ON(scsi_bufflen(ctask->sc) == 0); 222 BUG_ON(scsi_bufflen(task->sc) == 0);
223 223
224 debug_scsi("cmd [itt %x total %d imm %d unsol_data %d\n", 224 debug_scsi("cmd [itt %x total %d imm %d unsol_data %d\n",
225 ctask->itt, scsi_bufflen(ctask->sc), 225 task->itt, scsi_bufflen(task->sc),
226 ctask->imm_count, ctask->unsol_count); 226 task->imm_count, task->unsol_count);
227 } 227 }
228 228
229 debug_scsi("ctask deq [cid %d itt 0x%x]\n", 229 debug_scsi("task deq [cid %d itt 0x%x]\n",
230 conn->id, ctask->itt); 230 conn->id, task->itt);
231 231
232 /* Send the cmd PDU */ 232 /* Send the cmd PDU */
233 if (!iser_ctask->command_sent) { 233 if (!iser_task->command_sent) {
234 error = iser_send_command(conn, ctask); 234 error = iser_send_command(conn, task);
235 if (error) 235 if (error)
236 goto iscsi_iser_task_xmit_exit; 236 goto iscsi_iser_task_xmit_exit;
237 iser_ctask->command_sent = 1; 237 iser_task->command_sent = 1;
238 } 238 }
239 239
240 /* Send unsolicited data-out PDU(s) if necessary */ 240 /* Send unsolicited data-out PDU(s) if necessary */
241 if (ctask->unsol_count) 241 if (task->unsol_count)
242 error = iscsi_iser_task_xmit_unsol_data(conn, ctask); 242 error = iscsi_iser_task_xmit_unsol_data(conn, task);
243 243
244 iscsi_iser_task_xmit_exit: 244 iscsi_iser_task_xmit_exit:
245 if (error && error != -ENOBUFS) 245 if (error && error != -ENOBUFS)
@@ -248,17 +248,17 @@ iscsi_iser_task_xmit(struct iscsi_cmd_task *ctask)
248} 248}
249 249
250static void 250static void
251iscsi_iser_cleanup_task(struct iscsi_conn *conn, struct iscsi_cmd_task *ctask) 251iscsi_iser_cleanup_task(struct iscsi_conn *conn, struct iscsi_task *task)
252{ 252{
253 struct iscsi_iser_cmd_task *iser_ctask = ctask->dd_data; 253 struct iscsi_iser_task *iser_task = task->dd_data;
254 254
255 /* mgmt tasks do not need special cleanup */ 255 /* mgmt tasks do not need special cleanup */
256 if (!ctask->sc) 256 if (!task->sc)
257 return; 257 return;
258 258
259 if (iser_ctask->status == ISER_TASK_STATUS_STARTED) { 259 if (iser_task->status == ISER_TASK_STATUS_STARTED) {
260 iser_ctask->status = ISER_TASK_STATUS_COMPLETED; 260 iser_task->status = ISER_TASK_STATUS_COMPLETED;
261 iser_ctask_rdma_finalize(iser_ctask); 261 iser_task_rdma_finalize(iser_task);
262 } 262 }
263} 263}
264 264
@@ -408,8 +408,8 @@ iscsi_iser_session_create(struct Scsi_Host *shost,
408 struct iscsi_cls_session *cls_session; 408 struct iscsi_cls_session *cls_session;
409 struct iscsi_session *session; 409 struct iscsi_session *session;
410 int i; 410 int i;
411 struct iscsi_cmd_task *ctask; 411 struct iscsi_task *task;
412 struct iscsi_iser_cmd_task *iser_ctask; 412 struct iscsi_iser_task *iser_task;
413 413
414 if (shost) { 414 if (shost) {
415 printk(KERN_ERR "iscsi_tcp: invalid shost %d.\n", 415 printk(KERN_ERR "iscsi_tcp: invalid shost %d.\n",
@@ -436,7 +436,7 @@ iscsi_iser_session_create(struct Scsi_Host *shost,
436 */ 436 */
437 cls_session = iscsi_session_setup(&iscsi_iser_transport, shost, 437 cls_session = iscsi_session_setup(&iscsi_iser_transport, shost,
438 ISCSI_DEF_XMIT_CMDS_MAX, 438 ISCSI_DEF_XMIT_CMDS_MAX,
439 sizeof(struct iscsi_iser_cmd_task), 439 sizeof(struct iscsi_iser_task),
440 initial_cmdsn); 440 initial_cmdsn);
441 if (!cls_session) 441 if (!cls_session)
442 goto remove_host; 442 goto remove_host;
@@ -445,10 +445,10 @@ iscsi_iser_session_create(struct Scsi_Host *shost,
445 shost->can_queue = session->scsi_cmds_max; 445 shost->can_queue = session->scsi_cmds_max;
446 /* libiscsi setup itts, data and pool so just set desc fields */ 446 /* libiscsi setup itts, data and pool so just set desc fields */
447 for (i = 0; i < session->cmds_max; i++) { 447 for (i = 0; i < session->cmds_max; i++) {
448 ctask = session->cmds[i]; 448 task = session->cmds[i];
449 iser_ctask = ctask->dd_data; 449 iser_task = task->dd_data;
450 ctask->hdr = (struct iscsi_cmd *)&iser_ctask->desc.iscsi_header; 450 task->hdr = (struct iscsi_cmd *)&iser_task->desc.iscsi_header;
451 ctask->hdr_max = sizeof(iser_ctask->desc.iscsi_header); 451 task->hdr_max = sizeof(iser_task->desc.iscsi_header);
452 } 452 }
453 return cls_session; 453 return cls_session;
454 454
diff --git a/drivers/infiniband/ulp/iser/iscsi_iser.h b/drivers/infiniband/ulp/iser/iscsi_iser.h
index 96a600f127c8..05431f270fe8 100644
--- a/drivers/infiniband/ulp/iser/iscsi_iser.h
+++ b/drivers/infiniband/ulp/iser/iscsi_iser.h
@@ -173,7 +173,7 @@ struct iser_data_buf {
173/* fwd declarations */ 173/* fwd declarations */
174struct iser_device; 174struct iser_device;
175struct iscsi_iser_conn; 175struct iscsi_iser_conn;
176struct iscsi_iser_cmd_task; 176struct iscsi_iser_task;
177 177
178struct iser_mem_reg { 178struct iser_mem_reg {
179 u32 lkey; 179 u32 lkey;
@@ -197,7 +197,7 @@ struct iser_regd_buf {
197#define MAX_REGD_BUF_VECTOR_LEN 2 197#define MAX_REGD_BUF_VECTOR_LEN 2
198 198
199struct iser_dto { 199struct iser_dto {
200 struct iscsi_iser_cmd_task *ctask; 200 struct iscsi_iser_task *task;
201 struct iser_conn *ib_conn; 201 struct iser_conn *ib_conn;
202 int notify_enable; 202 int notify_enable;
203 203
@@ -265,7 +265,7 @@ struct iscsi_iser_conn {
265 rwlock_t lock; 265 rwlock_t lock;
266}; 266};
267 267
268struct iscsi_iser_cmd_task { 268struct iscsi_iser_task {
269 struct iser_desc desc; 269 struct iser_desc desc;
270 struct iscsi_iser_conn *iser_conn; 270 struct iscsi_iser_conn *iser_conn;
271 enum iser_task_status status; 271 enum iser_task_status status;
@@ -299,13 +299,13 @@ extern int iser_debug_level;
299int iser_conn_set_full_featured_mode(struct iscsi_conn *conn); 299int iser_conn_set_full_featured_mode(struct iscsi_conn *conn);
300 300
301int iser_send_control(struct iscsi_conn *conn, 301int iser_send_control(struct iscsi_conn *conn,
302 struct iscsi_cmd_task *ctask); 302 struct iscsi_task *task);
303 303
304int iser_send_command(struct iscsi_conn *conn, 304int iser_send_command(struct iscsi_conn *conn,
305 struct iscsi_cmd_task *ctask); 305 struct iscsi_task *task);
306 306
307int iser_send_data_out(struct iscsi_conn *conn, 307int iser_send_data_out(struct iscsi_conn *conn,
308 struct iscsi_cmd_task *ctask, 308 struct iscsi_task *task,
309 struct iscsi_data *hdr); 309 struct iscsi_data *hdr);
310 310
311void iscsi_iser_recv(struct iscsi_conn *conn, 311void iscsi_iser_recv(struct iscsi_conn *conn,
@@ -326,9 +326,9 @@ void iser_rcv_completion(struct iser_desc *desc,
326 326
327void iser_snd_completion(struct iser_desc *desc); 327void iser_snd_completion(struct iser_desc *desc);
328 328
329void iser_ctask_rdma_init(struct iscsi_iser_cmd_task *ctask); 329void iser_task_rdma_init(struct iscsi_iser_task *task);
330 330
331void iser_ctask_rdma_finalize(struct iscsi_iser_cmd_task *ctask); 331void iser_task_rdma_finalize(struct iscsi_iser_task *task);
332 332
333void iser_dto_buffs_release(struct iser_dto *dto); 333void iser_dto_buffs_release(struct iser_dto *dto);
334 334
@@ -338,10 +338,10 @@ void iser_reg_single(struct iser_device *device,
338 struct iser_regd_buf *regd_buf, 338 struct iser_regd_buf *regd_buf,
339 enum dma_data_direction direction); 339 enum dma_data_direction direction);
340 340
341void iser_finalize_rdma_unaligned_sg(struct iscsi_iser_cmd_task *ctask, 341void iser_finalize_rdma_unaligned_sg(struct iscsi_iser_task *task,
342 enum iser_data_dir cmd_dir); 342 enum iser_data_dir cmd_dir);
343 343
344int iser_reg_rdma_mem(struct iscsi_iser_cmd_task *ctask, 344int iser_reg_rdma_mem(struct iscsi_iser_task *task,
345 enum iser_data_dir cmd_dir); 345 enum iser_data_dir cmd_dir);
346 346
347int iser_connect(struct iser_conn *ib_conn, 347int iser_connect(struct iser_conn *ib_conn,
@@ -361,10 +361,10 @@ int iser_post_send(struct iser_desc *tx_desc);
361int iser_conn_state_comp(struct iser_conn *ib_conn, 361int iser_conn_state_comp(struct iser_conn *ib_conn,
362 enum iser_ib_conn_state comp); 362 enum iser_ib_conn_state comp);
363 363
364int iser_dma_map_task_data(struct iscsi_iser_cmd_task *iser_ctask, 364int iser_dma_map_task_data(struct iscsi_iser_task *iser_task,
365 struct iser_data_buf *data, 365 struct iser_data_buf *data,
366 enum iser_data_dir iser_dir, 366 enum iser_data_dir iser_dir,
367 enum dma_data_direction dma_dir); 367 enum dma_data_direction dma_dir);
368 368
369void iser_dma_unmap_task_data(struct iscsi_iser_cmd_task *iser_ctask); 369void iser_dma_unmap_task_data(struct iscsi_iser_task *iser_task);
370#endif 370#endif
diff --git a/drivers/infiniband/ulp/iser/iser_initiator.c b/drivers/infiniband/ulp/iser/iser_initiator.c
index 4ea78fbeee95..35af60a23c61 100644
--- a/drivers/infiniband/ulp/iser/iser_initiator.c
+++ b/drivers/infiniband/ulp/iser/iser_initiator.c
@@ -66,46 +66,46 @@ static void iser_dto_add_regd_buff(struct iser_dto *dto,
66 66
67/* Register user buffer memory and initialize passive rdma 67/* Register user buffer memory and initialize passive rdma
68 * dto descriptor. Total data size is stored in 68 * dto descriptor. Total data size is stored in
69 * iser_ctask->data[ISER_DIR_IN].data_len 69 * iser_task->data[ISER_DIR_IN].data_len
70 */ 70 */
71static int iser_prepare_read_cmd(struct iscsi_cmd_task *ctask, 71static int iser_prepare_read_cmd(struct iscsi_task *task,
72 unsigned int edtl) 72 unsigned int edtl)
73 73
74{ 74{
75 struct iscsi_iser_cmd_task *iser_ctask = ctask->dd_data; 75 struct iscsi_iser_task *iser_task = task->dd_data;
76 struct iser_regd_buf *regd_buf; 76 struct iser_regd_buf *regd_buf;
77 int err; 77 int err;
78 struct iser_hdr *hdr = &iser_ctask->desc.iser_header; 78 struct iser_hdr *hdr = &iser_task->desc.iser_header;
79 struct iser_data_buf *buf_in = &iser_ctask->data[ISER_DIR_IN]; 79 struct iser_data_buf *buf_in = &iser_task->data[ISER_DIR_IN];
80 80
81 err = iser_dma_map_task_data(iser_ctask, 81 err = iser_dma_map_task_data(iser_task,
82 buf_in, 82 buf_in,
83 ISER_DIR_IN, 83 ISER_DIR_IN,
84 DMA_FROM_DEVICE); 84 DMA_FROM_DEVICE);
85 if (err) 85 if (err)
86 return err; 86 return err;
87 87
88 if (edtl > iser_ctask->data[ISER_DIR_IN].data_len) { 88 if (edtl > iser_task->data[ISER_DIR_IN].data_len) {
89 iser_err("Total data length: %ld, less than EDTL: " 89 iser_err("Total data length: %ld, less than EDTL: "
90 "%d, in READ cmd BHS itt: %d, conn: 0x%p\n", 90 "%d, in READ cmd BHS itt: %d, conn: 0x%p\n",
91 iser_ctask->data[ISER_DIR_IN].data_len, edtl, 91 iser_task->data[ISER_DIR_IN].data_len, edtl,
92 ctask->itt, iser_ctask->iser_conn); 92 task->itt, iser_task->iser_conn);
93 return -EINVAL; 93 return -EINVAL;
94 } 94 }
95 95
96 err = iser_reg_rdma_mem(iser_ctask,ISER_DIR_IN); 96 err = iser_reg_rdma_mem(iser_task,ISER_DIR_IN);
97 if (err) { 97 if (err) {
98 iser_err("Failed to set up Data-IN RDMA\n"); 98 iser_err("Failed to set up Data-IN RDMA\n");
99 return err; 99 return err;
100 } 100 }
101 regd_buf = &iser_ctask->rdma_regd[ISER_DIR_IN]; 101 regd_buf = &iser_task->rdma_regd[ISER_DIR_IN];
102 102
103 hdr->flags |= ISER_RSV; 103 hdr->flags |= ISER_RSV;
104 hdr->read_stag = cpu_to_be32(regd_buf->reg.rkey); 104 hdr->read_stag = cpu_to_be32(regd_buf->reg.rkey);
105 hdr->read_va = cpu_to_be64(regd_buf->reg.va); 105 hdr->read_va = cpu_to_be64(regd_buf->reg.va);
106 106
107 iser_dbg("Cmd itt:%d READ tags RKEY:%#.4X VA:%#llX\n", 107 iser_dbg("Cmd itt:%d READ tags RKEY:%#.4X VA:%#llX\n",
108 ctask->itt, regd_buf->reg.rkey, 108 task->itt, regd_buf->reg.rkey,
109 (unsigned long long)regd_buf->reg.va); 109 (unsigned long long)regd_buf->reg.va);
110 110
111 return 0; 111 return 0;
@@ -113,43 +113,43 @@ static int iser_prepare_read_cmd(struct iscsi_cmd_task *ctask,
113 113
114/* Register user buffer memory and initialize passive rdma 114/* Register user buffer memory and initialize passive rdma
115 * dto descriptor. Total data size is stored in 115 * dto descriptor. Total data size is stored in
116 * ctask->data[ISER_DIR_OUT].data_len 116 * task->data[ISER_DIR_OUT].data_len
117 */ 117 */
118static int 118static int
119iser_prepare_write_cmd(struct iscsi_cmd_task *ctask, 119iser_prepare_write_cmd(struct iscsi_task *task,
120 unsigned int imm_sz, 120 unsigned int imm_sz,
121 unsigned int unsol_sz, 121 unsigned int unsol_sz,
122 unsigned int edtl) 122 unsigned int edtl)
123{ 123{
124 struct iscsi_iser_cmd_task *iser_ctask = ctask->dd_data; 124 struct iscsi_iser_task *iser_task = task->dd_data;
125 struct iser_regd_buf *regd_buf; 125 struct iser_regd_buf *regd_buf;
126 int err; 126 int err;
127 struct iser_dto *send_dto = &iser_ctask->desc.dto; 127 struct iser_dto *send_dto = &iser_task->desc.dto;
128 struct iser_hdr *hdr = &iser_ctask->desc.iser_header; 128 struct iser_hdr *hdr = &iser_task->desc.iser_header;
129 struct iser_data_buf *buf_out = &iser_ctask->data[ISER_DIR_OUT]; 129 struct iser_data_buf *buf_out = &iser_task->data[ISER_DIR_OUT];
130 130
131 err = iser_dma_map_task_data(iser_ctask, 131 err = iser_dma_map_task_data(iser_task,
132 buf_out, 132 buf_out,
133 ISER_DIR_OUT, 133 ISER_DIR_OUT,
134 DMA_TO_DEVICE); 134 DMA_TO_DEVICE);
135 if (err) 135 if (err)
136 return err; 136 return err;
137 137
138 if (edtl > iser_ctask->data[ISER_DIR_OUT].data_len) { 138 if (edtl > iser_task->data[ISER_DIR_OUT].data_len) {
139 iser_err("Total data length: %ld, less than EDTL: %d, " 139 iser_err("Total data length: %ld, less than EDTL: %d, "
140 "in WRITE cmd BHS itt: %d, conn: 0x%p\n", 140 "in WRITE cmd BHS itt: %d, conn: 0x%p\n",
141 iser_ctask->data[ISER_DIR_OUT].data_len, 141 iser_task->data[ISER_DIR_OUT].data_len,
142 edtl, ctask->itt, ctask->conn); 142 edtl, task->itt, task->conn);
143 return -EINVAL; 143 return -EINVAL;
144 } 144 }
145 145
146 err = iser_reg_rdma_mem(iser_ctask,ISER_DIR_OUT); 146 err = iser_reg_rdma_mem(iser_task,ISER_DIR_OUT);
147 if (err != 0) { 147 if (err != 0) {
148 iser_err("Failed to register write cmd RDMA mem\n"); 148 iser_err("Failed to register write cmd RDMA mem\n");
149 return err; 149 return err;
150 } 150 }
151 151
152 regd_buf = &iser_ctask->rdma_regd[ISER_DIR_OUT]; 152 regd_buf = &iser_task->rdma_regd[ISER_DIR_OUT];
153 153
154 if (unsol_sz < edtl) { 154 if (unsol_sz < edtl) {
155 hdr->flags |= ISER_WSV; 155 hdr->flags |= ISER_WSV;
@@ -158,13 +158,13 @@ iser_prepare_write_cmd(struct iscsi_cmd_task *ctask,
158 158
159 iser_dbg("Cmd itt:%d, WRITE tags, RKEY:%#.4X " 159 iser_dbg("Cmd itt:%d, WRITE tags, RKEY:%#.4X "
160 "VA:%#llX + unsol:%d\n", 160 "VA:%#llX + unsol:%d\n",
161 ctask->itt, regd_buf->reg.rkey, 161 task->itt, regd_buf->reg.rkey,
162 (unsigned long long)regd_buf->reg.va, unsol_sz); 162 (unsigned long long)regd_buf->reg.va, unsol_sz);
163 } 163 }
164 164
165 if (imm_sz > 0) { 165 if (imm_sz > 0) {
166 iser_dbg("Cmd itt:%d, WRITE, adding imm.data sz: %d\n", 166 iser_dbg("Cmd itt:%d, WRITE, adding imm.data sz: %d\n",
167 ctask->itt, imm_sz); 167 task->itt, imm_sz);
168 iser_dto_add_regd_buff(send_dto, 168 iser_dto_add_regd_buff(send_dto,
169 regd_buf, 169 regd_buf,
170 0, 170 0,
@@ -300,13 +300,13 @@ int iser_conn_set_full_featured_mode(struct iscsi_conn *conn)
300} 300}
301 301
302static int 302static int
303iser_check_xmit(struct iscsi_conn *conn, void *ctask) 303iser_check_xmit(struct iscsi_conn *conn, void *task)
304{ 304{
305 struct iscsi_iser_conn *iser_conn = conn->dd_data; 305 struct iscsi_iser_conn *iser_conn = conn->dd_data;
306 306
307 if (atomic_read(&iser_conn->ib_conn->post_send_buf_count) == 307 if (atomic_read(&iser_conn->ib_conn->post_send_buf_count) ==
308 ISER_QP_MAX_REQ_DTOS) { 308 ISER_QP_MAX_REQ_DTOS) {
309 iser_dbg("%ld can't xmit ctask %p\n",jiffies,ctask); 309 iser_dbg("%ld can't xmit task %p\n",jiffies,task);
310 return -ENOBUFS; 310 return -ENOBUFS;
311 } 311 }
312 return 0; 312 return 0;
@@ -317,37 +317,37 @@ iser_check_xmit(struct iscsi_conn *conn, void *ctask)
317 * iser_send_command - send command PDU 317 * iser_send_command - send command PDU
318 */ 318 */
319int iser_send_command(struct iscsi_conn *conn, 319int iser_send_command(struct iscsi_conn *conn,
320 struct iscsi_cmd_task *ctask) 320 struct iscsi_task *task)
321{ 321{
322 struct iscsi_iser_conn *iser_conn = conn->dd_data; 322 struct iscsi_iser_conn *iser_conn = conn->dd_data;
323 struct iscsi_iser_cmd_task *iser_ctask = ctask->dd_data; 323 struct iscsi_iser_task *iser_task = task->dd_data;
324 struct iser_dto *send_dto = NULL; 324 struct iser_dto *send_dto = NULL;
325 unsigned long edtl; 325 unsigned long edtl;
326 int err = 0; 326 int err = 0;
327 struct iser_data_buf *data_buf; 327 struct iser_data_buf *data_buf;
328 328
329 struct iscsi_cmd *hdr = ctask->hdr; 329 struct iscsi_cmd *hdr = task->hdr;
330 struct scsi_cmnd *sc = ctask->sc; 330 struct scsi_cmnd *sc = task->sc;
331 331
332 if (!iser_conn_state_comp(iser_conn->ib_conn, ISER_CONN_UP)) { 332 if (!iser_conn_state_comp(iser_conn->ib_conn, ISER_CONN_UP)) {
333 iser_err("Failed to send, conn: 0x%p is not up\n", iser_conn->ib_conn); 333 iser_err("Failed to send, conn: 0x%p is not up\n", iser_conn->ib_conn);
334 return -EPERM; 334 return -EPERM;
335 } 335 }
336 if (iser_check_xmit(conn, ctask)) 336 if (iser_check_xmit(conn, task))
337 return -ENOBUFS; 337 return -ENOBUFS;
338 338
339 edtl = ntohl(hdr->data_length); 339 edtl = ntohl(hdr->data_length);
340 340
341 /* build the tx desc regd header and add it to the tx desc dto */ 341 /* build the tx desc regd header and add it to the tx desc dto */
342 iser_ctask->desc.type = ISCSI_TX_SCSI_COMMAND; 342 iser_task->desc.type = ISCSI_TX_SCSI_COMMAND;
343 send_dto = &iser_ctask->desc.dto; 343 send_dto = &iser_task->desc.dto;
344 send_dto->ctask = iser_ctask; 344 send_dto->task = iser_task;
345 iser_create_send_desc(iser_conn, &iser_ctask->desc); 345 iser_create_send_desc(iser_conn, &iser_task->desc);
346 346
347 if (hdr->flags & ISCSI_FLAG_CMD_READ) 347 if (hdr->flags & ISCSI_FLAG_CMD_READ)
348 data_buf = &iser_ctask->data[ISER_DIR_IN]; 348 data_buf = &iser_task->data[ISER_DIR_IN];
349 else 349 else
350 data_buf = &iser_ctask->data[ISER_DIR_OUT]; 350 data_buf = &iser_task->data[ISER_DIR_OUT];
351 351
352 if (scsi_sg_count(sc)) { /* using a scatter list */ 352 if (scsi_sg_count(sc)) { /* using a scatter list */
353 data_buf->buf = scsi_sglist(sc); 353 data_buf->buf = scsi_sglist(sc);
@@ -357,15 +357,15 @@ int iser_send_command(struct iscsi_conn *conn,
357 data_buf->data_len = scsi_bufflen(sc); 357 data_buf->data_len = scsi_bufflen(sc);
358 358
359 if (hdr->flags & ISCSI_FLAG_CMD_READ) { 359 if (hdr->flags & ISCSI_FLAG_CMD_READ) {
360 err = iser_prepare_read_cmd(ctask, edtl); 360 err = iser_prepare_read_cmd(task, edtl);
361 if (err) 361 if (err)
362 goto send_command_error; 362 goto send_command_error;
363 } 363 }
364 if (hdr->flags & ISCSI_FLAG_CMD_WRITE) { 364 if (hdr->flags & ISCSI_FLAG_CMD_WRITE) {
365 err = iser_prepare_write_cmd(ctask, 365 err = iser_prepare_write_cmd(task,
366 ctask->imm_count, 366 task->imm_count,
367 ctask->imm_count + 367 task->imm_count +
368 ctask->unsol_count, 368 task->unsol_count,
369 edtl); 369 edtl);
370 if (err) 370 if (err)
371 goto send_command_error; 371 goto send_command_error;
@@ -380,15 +380,15 @@ int iser_send_command(struct iscsi_conn *conn,
380 goto send_command_error; 380 goto send_command_error;
381 } 381 }
382 382
383 iser_ctask->status = ISER_TASK_STATUS_STARTED; 383 iser_task->status = ISER_TASK_STATUS_STARTED;
384 384
385 err = iser_post_send(&iser_ctask->desc); 385 err = iser_post_send(&iser_task->desc);
386 if (!err) 386 if (!err)
387 return 0; 387 return 0;
388 388
389send_command_error: 389send_command_error:
390 iser_dto_buffs_release(send_dto); 390 iser_dto_buffs_release(send_dto);
391 iser_err("conn %p failed ctask->itt %d err %d\n",conn, ctask->itt, err); 391 iser_err("conn %p failed task->itt %d err %d\n",conn, task->itt, err);
392 return err; 392 return err;
393} 393}
394 394
@@ -396,11 +396,11 @@ send_command_error:
396 * iser_send_data_out - send data out PDU 396 * iser_send_data_out - send data out PDU
397 */ 397 */
398int iser_send_data_out(struct iscsi_conn *conn, 398int iser_send_data_out(struct iscsi_conn *conn,
399 struct iscsi_cmd_task *ctask, 399 struct iscsi_task *task,
400 struct iscsi_data *hdr) 400 struct iscsi_data *hdr)
401{ 401{
402 struct iscsi_iser_conn *iser_conn = conn->dd_data; 402 struct iscsi_iser_conn *iser_conn = conn->dd_data;
403 struct iscsi_iser_cmd_task *iser_ctask = ctask->dd_data; 403 struct iscsi_iser_task *iser_task = task->dd_data;
404 struct iser_desc *tx_desc = NULL; 404 struct iser_desc *tx_desc = NULL;
405 struct iser_dto *send_dto = NULL; 405 struct iser_dto *send_dto = NULL;
406 unsigned long buf_offset; 406 unsigned long buf_offset;
@@ -413,7 +413,7 @@ int iser_send_data_out(struct iscsi_conn *conn,
413 return -EPERM; 413 return -EPERM;
414 } 414 }
415 415
416 if (iser_check_xmit(conn, ctask)) 416 if (iser_check_xmit(conn, task))
417 return -ENOBUFS; 417 return -ENOBUFS;
418 418
419 itt = (__force uint32_t)hdr->itt; 419 itt = (__force uint32_t)hdr->itt;
@@ -434,7 +434,7 @@ int iser_send_data_out(struct iscsi_conn *conn,
434 434
435 /* build the tx desc regd header and add it to the tx desc dto */ 435 /* build the tx desc regd header and add it to the tx desc dto */
436 send_dto = &tx_desc->dto; 436 send_dto = &tx_desc->dto;
437 send_dto->ctask = iser_ctask; 437 send_dto->task = iser_task;
438 iser_create_send_desc(iser_conn, tx_desc); 438 iser_create_send_desc(iser_conn, tx_desc);
439 439
440 iser_reg_single(iser_conn->ib_conn->device, 440 iser_reg_single(iser_conn->ib_conn->device,
@@ -442,15 +442,15 @@ int iser_send_data_out(struct iscsi_conn *conn,
442 442
443 /* all data was registered for RDMA, we can use the lkey */ 443 /* all data was registered for RDMA, we can use the lkey */
444 iser_dto_add_regd_buff(send_dto, 444 iser_dto_add_regd_buff(send_dto,
445 &iser_ctask->rdma_regd[ISER_DIR_OUT], 445 &iser_task->rdma_regd[ISER_DIR_OUT],
446 buf_offset, 446 buf_offset,
447 data_seg_len); 447 data_seg_len);
448 448
449 if (buf_offset + data_seg_len > iser_ctask->data[ISER_DIR_OUT].data_len) { 449 if (buf_offset + data_seg_len > iser_task->data[ISER_DIR_OUT].data_len) {
450 iser_err("Offset:%ld & DSL:%ld in Data-Out " 450 iser_err("Offset:%ld & DSL:%ld in Data-Out "
451 "inconsistent with total len:%ld, itt:%d\n", 451 "inconsistent with total len:%ld, itt:%d\n",
452 buf_offset, data_seg_len, 452 buf_offset, data_seg_len,
453 iser_ctask->data[ISER_DIR_OUT].data_len, itt); 453 iser_task->data[ISER_DIR_OUT].data_len, itt);
454 err = -EINVAL; 454 err = -EINVAL;
455 goto send_data_out_error; 455 goto send_data_out_error;
456 } 456 }
@@ -470,11 +470,11 @@ send_data_out_error:
470} 470}
471 471
472int iser_send_control(struct iscsi_conn *conn, 472int iser_send_control(struct iscsi_conn *conn,
473 struct iscsi_cmd_task *ctask) 473 struct iscsi_task *task)
474{ 474{
475 struct iscsi_iser_conn *iser_conn = conn->dd_data; 475 struct iscsi_iser_conn *iser_conn = conn->dd_data;
476 struct iscsi_iser_cmd_task *iser_ctask = ctask->dd_data; 476 struct iscsi_iser_task *iser_task = task->dd_data;
477 struct iser_desc *mdesc = &iser_ctask->desc; 477 struct iser_desc *mdesc = &iser_task->desc;
478 struct iser_dto *send_dto = NULL; 478 struct iser_dto *send_dto = NULL;
479 unsigned long data_seg_len; 479 unsigned long data_seg_len;
480 int err = 0; 480 int err = 0;
@@ -486,27 +486,27 @@ int iser_send_control(struct iscsi_conn *conn,
486 return -EPERM; 486 return -EPERM;
487 } 487 }
488 488
489 if (iser_check_xmit(conn, ctask)) 489 if (iser_check_xmit(conn, task))
490 return -ENOBUFS; 490 return -ENOBUFS;
491 491
492 /* build the tx desc regd header and add it to the tx desc dto */ 492 /* build the tx desc regd header and add it to the tx desc dto */
493 mdesc->type = ISCSI_TX_CONTROL; 493 mdesc->type = ISCSI_TX_CONTROL;
494 send_dto = &mdesc->dto; 494 send_dto = &mdesc->dto;
495 send_dto->ctask = NULL; 495 send_dto->task = NULL;
496 iser_create_send_desc(iser_conn, mdesc); 496 iser_create_send_desc(iser_conn, mdesc);
497 497
498 device = iser_conn->ib_conn->device; 498 device = iser_conn->ib_conn->device;
499 499
500 iser_reg_single(device, send_dto->regd[0], DMA_TO_DEVICE); 500 iser_reg_single(device, send_dto->regd[0], DMA_TO_DEVICE);
501 501
502 data_seg_len = ntoh24(ctask->hdr->dlength); 502 data_seg_len = ntoh24(task->hdr->dlength);
503 503
504 if (data_seg_len > 0) { 504 if (data_seg_len > 0) {
505 regd_buf = &mdesc->data_regd_buf; 505 regd_buf = &mdesc->data_regd_buf;
506 memset(regd_buf, 0, sizeof(struct iser_regd_buf)); 506 memset(regd_buf, 0, sizeof(struct iser_regd_buf));
507 regd_buf->device = device; 507 regd_buf->device = device;
508 regd_buf->virt_addr = ctask->data; 508 regd_buf->virt_addr = task->data;
509 regd_buf->data_size = ctask->data_count; 509 regd_buf->data_size = task->data_count;
510 iser_reg_single(device, regd_buf, 510 iser_reg_single(device, regd_buf,
511 DMA_TO_DEVICE); 511 DMA_TO_DEVICE);
512 iser_dto_add_regd_buff(send_dto, regd_buf, 512 iser_dto_add_regd_buff(send_dto, regd_buf,
@@ -538,8 +538,8 @@ void iser_rcv_completion(struct iser_desc *rx_desc,
538{ 538{
539 struct iser_dto *dto = &rx_desc->dto; 539 struct iser_dto *dto = &rx_desc->dto;
540 struct iscsi_iser_conn *conn = dto->ib_conn->iser_conn; 540 struct iscsi_iser_conn *conn = dto->ib_conn->iser_conn;
541 struct iscsi_cmd_task *ctask; 541 struct iscsi_task *task;
542 struct iscsi_iser_cmd_task *iser_ctask; 542 struct iscsi_iser_task *iser_task;
543 struct iscsi_hdr *hdr; 543 struct iscsi_hdr *hdr;
544 char *rx_data = NULL; 544 char *rx_data = NULL;
545 int rx_data_len = 0; 545 int rx_data_len = 0;
@@ -558,16 +558,16 @@ void iser_rcv_completion(struct iser_desc *rx_desc,
558 opcode = hdr->opcode & ISCSI_OPCODE_MASK; 558 opcode = hdr->opcode & ISCSI_OPCODE_MASK;
559 559
560 if (opcode == ISCSI_OP_SCSI_CMD_RSP) { 560 if (opcode == ISCSI_OP_SCSI_CMD_RSP) {
561 ctask = iscsi_itt_to_ctask(conn->iscsi_conn, hdr->itt); 561 task = iscsi_itt_to_ctask(conn->iscsi_conn, hdr->itt);
562 if (!ctask) 562 if (!task)
563 iser_err("itt can't be matched to ctask!!! " 563 iser_err("itt can't be matched to task!!! "
564 "conn %p opcode %d itt %d\n", 564 "conn %p opcode %d itt %d\n",
565 conn->iscsi_conn, opcode, hdr->itt); 565 conn->iscsi_conn, opcode, hdr->itt);
566 else { 566 else {
567 iser_ctask = ctask->dd_data; 567 iser_task = task->dd_data;
568 iser_dbg("itt %d ctask %p\n",hdr->itt, ctask); 568 iser_dbg("itt %d task %p\n",hdr->itt, task);
569 iser_ctask->status = ISER_TASK_STATUS_COMPLETED; 569 iser_task->status = ISER_TASK_STATUS_COMPLETED;
570 iser_ctask_rdma_finalize(iser_ctask); 570 iser_task_rdma_finalize(iser_task);
571 } 571 }
572 } 572 }
573 iser_dto_buffs_release(dto); 573 iser_dto_buffs_release(dto);
@@ -578,7 +578,7 @@ void iser_rcv_completion(struct iser_desc *rx_desc,
578 kmem_cache_free(ig.desc_cache, rx_desc); 578 kmem_cache_free(ig.desc_cache, rx_desc);
579 579
580 /* decrementing conn->post_recv_buf_count only --after-- freeing the * 580 /* decrementing conn->post_recv_buf_count only --after-- freeing the *
581 * ctask eliminates the need to worry on ctasks which are completed in * 581 * task eliminates the need to worry on tasks which are completed in *
582 * parallel to the execution of iser_conn_term. So the code that waits * 582 * parallel to the execution of iser_conn_term. So the code that waits *
583 * for the posted rx bufs refcount to become zero handles everything */ 583 * for the posted rx bufs refcount to become zero handles everything */
584 atomic_dec(&conn->ib_conn->post_recv_buf_count); 584 atomic_dec(&conn->ib_conn->post_recv_buf_count);
@@ -590,7 +590,7 @@ void iser_snd_completion(struct iser_desc *tx_desc)
590 struct iser_conn *ib_conn = dto->ib_conn; 590 struct iser_conn *ib_conn = dto->ib_conn;
591 struct iscsi_iser_conn *iser_conn = ib_conn->iser_conn; 591 struct iscsi_iser_conn *iser_conn = ib_conn->iser_conn;
592 struct iscsi_conn *conn = iser_conn->iscsi_conn; 592 struct iscsi_conn *conn = iser_conn->iscsi_conn;
593 struct iscsi_cmd_task *ctask; 593 struct iscsi_task *task;
594 int resume_tx = 0; 594 int resume_tx = 0;
595 595
596 iser_dbg("Initiator, Data sent dto=0x%p\n", dto); 596 iser_dbg("Initiator, Data sent dto=0x%p\n", dto);
@@ -613,31 +613,31 @@ void iser_snd_completion(struct iser_desc *tx_desc)
613 613
614 if (tx_desc->type == ISCSI_TX_CONTROL) { 614 if (tx_desc->type == ISCSI_TX_CONTROL) {
615 /* this arithmetic is legal by libiscsi dd_data allocation */ 615 /* this arithmetic is legal by libiscsi dd_data allocation */
616 ctask = (void *) ((long)(void *)tx_desc - 616 task = (void *) ((long)(void *)tx_desc -
617 sizeof(struct iscsi_cmd_task)); 617 sizeof(struct iscsi_task));
618 if (ctask->hdr->itt == RESERVED_ITT) 618 if (task->hdr->itt == RESERVED_ITT)
619 iscsi_put_ctask(ctask); 619 iscsi_put_task(task);
620 } 620 }
621} 621}
622 622
623void iser_ctask_rdma_init(struct iscsi_iser_cmd_task *iser_ctask) 623void iser_task_rdma_init(struct iscsi_iser_task *iser_task)
624 624
625{ 625{
626 iser_ctask->status = ISER_TASK_STATUS_INIT; 626 iser_task->status = ISER_TASK_STATUS_INIT;
627 627
628 iser_ctask->dir[ISER_DIR_IN] = 0; 628 iser_task->dir[ISER_DIR_IN] = 0;
629 iser_ctask->dir[ISER_DIR_OUT] = 0; 629 iser_task->dir[ISER_DIR_OUT] = 0;
630 630
631 iser_ctask->data[ISER_DIR_IN].data_len = 0; 631 iser_task->data[ISER_DIR_IN].data_len = 0;
632 iser_ctask->data[ISER_DIR_OUT].data_len = 0; 632 iser_task->data[ISER_DIR_OUT].data_len = 0;
633 633
634 memset(&iser_ctask->rdma_regd[ISER_DIR_IN], 0, 634 memset(&iser_task->rdma_regd[ISER_DIR_IN], 0,
635 sizeof(struct iser_regd_buf)); 635 sizeof(struct iser_regd_buf));
636 memset(&iser_ctask->rdma_regd[ISER_DIR_OUT], 0, 636 memset(&iser_task->rdma_regd[ISER_DIR_OUT], 0,
637 sizeof(struct iser_regd_buf)); 637 sizeof(struct iser_regd_buf));
638} 638}
639 639
640void iser_ctask_rdma_finalize(struct iscsi_iser_cmd_task *iser_ctask) 640void iser_task_rdma_finalize(struct iscsi_iser_task *iser_task)
641{ 641{
642 int deferred; 642 int deferred;
643 int is_rdma_aligned = 1; 643 int is_rdma_aligned = 1;
@@ -646,17 +646,17 @@ void iser_ctask_rdma_finalize(struct iscsi_iser_cmd_task *iser_ctask)
646 /* if we were reading, copy back to unaligned sglist, 646 /* if we were reading, copy back to unaligned sglist,
647 * anyway dma_unmap and free the copy 647 * anyway dma_unmap and free the copy
648 */ 648 */
649 if (iser_ctask->data_copy[ISER_DIR_IN].copy_buf != NULL) { 649 if (iser_task->data_copy[ISER_DIR_IN].copy_buf != NULL) {
650 is_rdma_aligned = 0; 650 is_rdma_aligned = 0;
651 iser_finalize_rdma_unaligned_sg(iser_ctask, ISER_DIR_IN); 651 iser_finalize_rdma_unaligned_sg(iser_task, ISER_DIR_IN);
652 } 652 }
653 if (iser_ctask->data_copy[ISER_DIR_OUT].copy_buf != NULL) { 653 if (iser_task->data_copy[ISER_DIR_OUT].copy_buf != NULL) {
654 is_rdma_aligned = 0; 654 is_rdma_aligned = 0;
655 iser_finalize_rdma_unaligned_sg(iser_ctask, ISER_DIR_OUT); 655 iser_finalize_rdma_unaligned_sg(iser_task, ISER_DIR_OUT);
656 } 656 }
657 657
658 if (iser_ctask->dir[ISER_DIR_IN]) { 658 if (iser_task->dir[ISER_DIR_IN]) {
659 regd = &iser_ctask->rdma_regd[ISER_DIR_IN]; 659 regd = &iser_task->rdma_regd[ISER_DIR_IN];
660 deferred = iser_regd_buff_release(regd); 660 deferred = iser_regd_buff_release(regd);
661 if (deferred) { 661 if (deferred) {
662 iser_err("%d references remain for BUF-IN rdma reg\n", 662 iser_err("%d references remain for BUF-IN rdma reg\n",
@@ -664,8 +664,8 @@ void iser_ctask_rdma_finalize(struct iscsi_iser_cmd_task *iser_ctask)
664 } 664 }
665 } 665 }
666 666
667 if (iser_ctask->dir[ISER_DIR_OUT]) { 667 if (iser_task->dir[ISER_DIR_OUT]) {
668 regd = &iser_ctask->rdma_regd[ISER_DIR_OUT]; 668 regd = &iser_task->rdma_regd[ISER_DIR_OUT];
669 deferred = iser_regd_buff_release(regd); 669 deferred = iser_regd_buff_release(regd);
670 if (deferred) { 670 if (deferred) {
671 iser_err("%d references remain for BUF-OUT rdma reg\n", 671 iser_err("%d references remain for BUF-OUT rdma reg\n",
@@ -675,7 +675,7 @@ void iser_ctask_rdma_finalize(struct iscsi_iser_cmd_task *iser_ctask)
675 675
676 /* if the data was unaligned, it was already unmapped and then copied */ 676 /* if the data was unaligned, it was already unmapped and then copied */
677 if (is_rdma_aligned) 677 if (is_rdma_aligned)
678 iser_dma_unmap_task_data(iser_ctask); 678 iser_dma_unmap_task_data(iser_task);
679} 679}
680 680
681void iser_dto_buffs_release(struct iser_dto *dto) 681void iser_dto_buffs_release(struct iser_dto *dto)
diff --git a/drivers/infiniband/ulp/iser/iser_memory.c b/drivers/infiniband/ulp/iser/iser_memory.c
index cac50c4dc159..48f2a601fc27 100644
--- a/drivers/infiniband/ulp/iser/iser_memory.c
+++ b/drivers/infiniband/ulp/iser/iser_memory.c
@@ -101,13 +101,13 @@ void iser_reg_single(struct iser_device *device,
101/** 101/**
102 * iser_start_rdma_unaligned_sg 102 * iser_start_rdma_unaligned_sg
103 */ 103 */
104static int iser_start_rdma_unaligned_sg(struct iscsi_iser_cmd_task *iser_ctask, 104static int iser_start_rdma_unaligned_sg(struct iscsi_iser_task *iser_task,
105 enum iser_data_dir cmd_dir) 105 enum iser_data_dir cmd_dir)
106{ 106{
107 int dma_nents; 107 int dma_nents;
108 struct ib_device *dev; 108 struct ib_device *dev;
109 char *mem = NULL; 109 char *mem = NULL;
110 struct iser_data_buf *data = &iser_ctask->data[cmd_dir]; 110 struct iser_data_buf *data = &iser_task->data[cmd_dir];
111 unsigned long cmd_data_len = data->data_len; 111 unsigned long cmd_data_len = data->data_len;
112 112
113 if (cmd_data_len > ISER_KMALLOC_THRESHOLD) 113 if (cmd_data_len > ISER_KMALLOC_THRESHOLD)
@@ -140,37 +140,37 @@ static int iser_start_rdma_unaligned_sg(struct iscsi_iser_cmd_task *iser_ctask,
140 } 140 }
141 } 141 }
142 142
143 sg_init_one(&iser_ctask->data_copy[cmd_dir].sg_single, mem, cmd_data_len); 143 sg_init_one(&iser_task->data_copy[cmd_dir].sg_single, mem, cmd_data_len);
144 iser_ctask->data_copy[cmd_dir].buf = 144 iser_task->data_copy[cmd_dir].buf =
145 &iser_ctask->data_copy[cmd_dir].sg_single; 145 &iser_task->data_copy[cmd_dir].sg_single;
146 iser_ctask->data_copy[cmd_dir].size = 1; 146 iser_task->data_copy[cmd_dir].size = 1;
147 147
148 iser_ctask->data_copy[cmd_dir].copy_buf = mem; 148 iser_task->data_copy[cmd_dir].copy_buf = mem;
149 149
150 dev = iser_ctask->iser_conn->ib_conn->device->ib_device; 150 dev = iser_task->iser_conn->ib_conn->device->ib_device;
151 dma_nents = ib_dma_map_sg(dev, 151 dma_nents = ib_dma_map_sg(dev,
152 &iser_ctask->data_copy[cmd_dir].sg_single, 152 &iser_task->data_copy[cmd_dir].sg_single,
153 1, 153 1,
154 (cmd_dir == ISER_DIR_OUT) ? 154 (cmd_dir == ISER_DIR_OUT) ?
155 DMA_TO_DEVICE : DMA_FROM_DEVICE); 155 DMA_TO_DEVICE : DMA_FROM_DEVICE);
156 BUG_ON(dma_nents == 0); 156 BUG_ON(dma_nents == 0);
157 157
158 iser_ctask->data_copy[cmd_dir].dma_nents = dma_nents; 158 iser_task->data_copy[cmd_dir].dma_nents = dma_nents;
159 return 0; 159 return 0;
160} 160}
161 161
162/** 162/**
163 * iser_finalize_rdma_unaligned_sg 163 * iser_finalize_rdma_unaligned_sg
164 */ 164 */
165void iser_finalize_rdma_unaligned_sg(struct iscsi_iser_cmd_task *iser_ctask, 165void iser_finalize_rdma_unaligned_sg(struct iscsi_iser_task *iser_task,
166 enum iser_data_dir cmd_dir) 166 enum iser_data_dir cmd_dir)
167{ 167{
168 struct ib_device *dev; 168 struct ib_device *dev;
169 struct iser_data_buf *mem_copy; 169 struct iser_data_buf *mem_copy;
170 unsigned long cmd_data_len; 170 unsigned long cmd_data_len;
171 171
172 dev = iser_ctask->iser_conn->ib_conn->device->ib_device; 172 dev = iser_task->iser_conn->ib_conn->device->ib_device;
173 mem_copy = &iser_ctask->data_copy[cmd_dir]; 173 mem_copy = &iser_task->data_copy[cmd_dir];
174 174
175 ib_dma_unmap_sg(dev, &mem_copy->sg_single, 1, 175 ib_dma_unmap_sg(dev, &mem_copy->sg_single, 1,
176 (cmd_dir == ISER_DIR_OUT) ? 176 (cmd_dir == ISER_DIR_OUT) ?
@@ -186,8 +186,8 @@ void iser_finalize_rdma_unaligned_sg(struct iscsi_iser_cmd_task *iser_ctask,
186 /* copy back read RDMA to unaligned sg */ 186 /* copy back read RDMA to unaligned sg */
187 mem = mem_copy->copy_buf; 187 mem = mem_copy->copy_buf;
188 188
189 sgl = (struct scatterlist *)iser_ctask->data[ISER_DIR_IN].buf; 189 sgl = (struct scatterlist *)iser_task->data[ISER_DIR_IN].buf;
190 sg_size = iser_ctask->data[ISER_DIR_IN].size; 190 sg_size = iser_task->data[ISER_DIR_IN].size;
191 191
192 p = mem; 192 p = mem;
193 for_each_sg(sgl, sg, sg_size, i) { 193 for_each_sg(sgl, sg, sg_size, i) {
@@ -200,7 +200,7 @@ void iser_finalize_rdma_unaligned_sg(struct iscsi_iser_cmd_task *iser_ctask,
200 } 200 }
201 } 201 }
202 202
203 cmd_data_len = iser_ctask->data[cmd_dir].data_len; 203 cmd_data_len = iser_task->data[cmd_dir].data_len;
204 204
205 if (cmd_data_len > ISER_KMALLOC_THRESHOLD) 205 if (cmd_data_len > ISER_KMALLOC_THRESHOLD)
206 free_pages((unsigned long)mem_copy->copy_buf, 206 free_pages((unsigned long)mem_copy->copy_buf,
@@ -378,15 +378,15 @@ static void iser_page_vec_build(struct iser_data_buf *data,
378 } 378 }
379} 379}
380 380
381int iser_dma_map_task_data(struct iscsi_iser_cmd_task *iser_ctask, 381int iser_dma_map_task_data(struct iscsi_iser_task *iser_task,
382 struct iser_data_buf *data, 382 struct iser_data_buf *data,
383 enum iser_data_dir iser_dir, 383 enum iser_data_dir iser_dir,
384 enum dma_data_direction dma_dir) 384 enum dma_data_direction dma_dir)
385{ 385{
386 struct ib_device *dev; 386 struct ib_device *dev;
387 387
388 iser_ctask->dir[iser_dir] = 1; 388 iser_task->dir[iser_dir] = 1;
389 dev = iser_ctask->iser_conn->ib_conn->device->ib_device; 389 dev = iser_task->iser_conn->ib_conn->device->ib_device;
390 390
391 data->dma_nents = ib_dma_map_sg(dev, data->buf, data->size, dma_dir); 391 data->dma_nents = ib_dma_map_sg(dev, data->buf, data->size, dma_dir);
392 if (data->dma_nents == 0) { 392 if (data->dma_nents == 0) {
@@ -396,20 +396,20 @@ int iser_dma_map_task_data(struct iscsi_iser_cmd_task *iser_ctask,
396 return 0; 396 return 0;
397} 397}
398 398
399void iser_dma_unmap_task_data(struct iscsi_iser_cmd_task *iser_ctask) 399void iser_dma_unmap_task_data(struct iscsi_iser_task *iser_task)
400{ 400{
401 struct ib_device *dev; 401 struct ib_device *dev;
402 struct iser_data_buf *data; 402 struct iser_data_buf *data;
403 403
404 dev = iser_ctask->iser_conn->ib_conn->device->ib_device; 404 dev = iser_task->iser_conn->ib_conn->device->ib_device;
405 405
406 if (iser_ctask->dir[ISER_DIR_IN]) { 406 if (iser_task->dir[ISER_DIR_IN]) {
407 data = &iser_ctask->data[ISER_DIR_IN]; 407 data = &iser_task->data[ISER_DIR_IN];
408 ib_dma_unmap_sg(dev, data->buf, data->size, DMA_FROM_DEVICE); 408 ib_dma_unmap_sg(dev, data->buf, data->size, DMA_FROM_DEVICE);
409 } 409 }
410 410
411 if (iser_ctask->dir[ISER_DIR_OUT]) { 411 if (iser_task->dir[ISER_DIR_OUT]) {
412 data = &iser_ctask->data[ISER_DIR_OUT]; 412 data = &iser_task->data[ISER_DIR_OUT];
413 ib_dma_unmap_sg(dev, data->buf, data->size, DMA_TO_DEVICE); 413 ib_dma_unmap_sg(dev, data->buf, data->size, DMA_TO_DEVICE);
414 } 414 }
415} 415}
@@ -420,21 +420,21 @@ void iser_dma_unmap_task_data(struct iscsi_iser_cmd_task *iser_ctask)
420 * 420 *
421 * returns 0 on success, errno code on failure 421 * returns 0 on success, errno code on failure
422 */ 422 */
423int iser_reg_rdma_mem(struct iscsi_iser_cmd_task *iser_ctask, 423int iser_reg_rdma_mem(struct iscsi_iser_task *iser_task,
424 enum iser_data_dir cmd_dir) 424 enum iser_data_dir cmd_dir)
425{ 425{
426 struct iscsi_conn *iscsi_conn = iser_ctask->iser_conn->iscsi_conn; 426 struct iscsi_conn *iscsi_conn = iser_task->iser_conn->iscsi_conn;
427 struct iser_conn *ib_conn = iser_ctask->iser_conn->ib_conn; 427 struct iser_conn *ib_conn = iser_task->iser_conn->ib_conn;
428 struct iser_device *device = ib_conn->device; 428 struct iser_device *device = ib_conn->device;
429 struct ib_device *ibdev = device->ib_device; 429 struct ib_device *ibdev = device->ib_device;
430 struct iser_data_buf *mem = &iser_ctask->data[cmd_dir]; 430 struct iser_data_buf *mem = &iser_task->data[cmd_dir];
431 struct iser_regd_buf *regd_buf; 431 struct iser_regd_buf *regd_buf;
432 int aligned_len; 432 int aligned_len;
433 int err; 433 int err;
434 int i; 434 int i;
435 struct scatterlist *sg; 435 struct scatterlist *sg;
436 436
437 regd_buf = &iser_ctask->rdma_regd[cmd_dir]; 437 regd_buf = &iser_task->rdma_regd[cmd_dir];
438 438
439 aligned_len = iser_data_buf_aligned_len(mem, ibdev); 439 aligned_len = iser_data_buf_aligned_len(mem, ibdev);
440 if (aligned_len != mem->dma_nents) { 440 if (aligned_len != mem->dma_nents) {
@@ -444,13 +444,13 @@ int iser_reg_rdma_mem(struct iscsi_iser_cmd_task *iser_ctask,
444 iser_data_buf_dump(mem, ibdev); 444 iser_data_buf_dump(mem, ibdev);
445 445
446 /* unmap the command data before accessing it */ 446 /* unmap the command data before accessing it */
447 iser_dma_unmap_task_data(iser_ctask); 447 iser_dma_unmap_task_data(iser_task);
448 448
449 /* allocate copy buf, if we are writing, copy the */ 449 /* allocate copy buf, if we are writing, copy the */
450 /* unaligned scatterlist, dma map the copy */ 450 /* unaligned scatterlist, dma map the copy */
451 if (iser_start_rdma_unaligned_sg(iser_ctask, cmd_dir) != 0) 451 if (iser_start_rdma_unaligned_sg(iser_task, cmd_dir) != 0)
452 return -ENOMEM; 452 return -ENOMEM;
453 mem = &iser_ctask->data_copy[cmd_dir]; 453 mem = &iser_task->data_copy[cmd_dir];
454 } 454 }
455 455
456 /* if there a single dma entry, FMR is not needed */ 456 /* if there a single dma entry, FMR is not needed */
@@ -474,8 +474,9 @@ int iser_reg_rdma_mem(struct iscsi_iser_cmd_task *iser_ctask,
474 err = iser_reg_page_vec(ib_conn, ib_conn->page_vec, &regd_buf->reg); 474 err = iser_reg_page_vec(ib_conn, ib_conn->page_vec, &regd_buf->reg);
475 if (err) { 475 if (err) {
476 iser_data_buf_dump(mem, ibdev); 476 iser_data_buf_dump(mem, ibdev);
477 iser_err("mem->dma_nents = %d (dlength = 0x%x)\n", mem->dma_nents, 477 iser_err("mem->dma_nents = %d (dlength = 0x%x)\n",
478 ntoh24(iser_ctask->desc.iscsi_header.dlength)); 478 mem->dma_nents,
479 ntoh24(iser_task->desc.iscsi_header.dlength));
479 iser_err("page_vec: data_size = 0x%x, length = %d, offset = 0x%x\n", 480 iser_err("page_vec: data_size = 0x%x, length = %d, offset = 0x%x\n",
480 ib_conn->page_vec->data_size, ib_conn->page_vec->length, 481 ib_conn->page_vec->data_size, ib_conn->page_vec->length,
481 ib_conn->page_vec->offset); 482 ib_conn->page_vec->offset);