diff options
| -rw-r--r-- | drivers/infiniband/ulp/iser/iser_memory.c | 53 |
1 files changed, 29 insertions, 24 deletions
diff --git a/drivers/infiniband/ulp/iser/iser_memory.c b/drivers/infiniband/ulp/iser/iser_memory.c index a9335080ae69..2c3f4b144a1a 100644 --- a/drivers/infiniband/ulp/iser/iser_memory.c +++ b/drivers/infiniband/ulp/iser/iser_memory.c | |||
| @@ -45,13 +45,19 @@ | |||
| 45 | * iser_start_rdma_unaligned_sg | 45 | * iser_start_rdma_unaligned_sg |
| 46 | */ | 46 | */ |
| 47 | static int iser_start_rdma_unaligned_sg(struct iscsi_iser_task *iser_task, | 47 | static int iser_start_rdma_unaligned_sg(struct iscsi_iser_task *iser_task, |
| 48 | struct iser_data_buf *data, | ||
| 49 | struct iser_data_buf *data_copy, | ||
| 48 | enum iser_data_dir cmd_dir) | 50 | enum iser_data_dir cmd_dir) |
| 49 | { | 51 | { |
| 50 | int dma_nents; | 52 | struct ib_device *dev = iser_task->iser_conn->ib_conn->device->ib_device; |
| 51 | struct ib_device *dev; | 53 | struct scatterlist *sgl = (struct scatterlist *)data->buf; |
| 54 | struct scatterlist *sg; | ||
| 52 | char *mem = NULL; | 55 | char *mem = NULL; |
| 53 | struct iser_data_buf *data = &iser_task->data[cmd_dir]; | 56 | unsigned long cmd_data_len = 0; |
| 54 | unsigned long cmd_data_len = data->data_len; | 57 | int dma_nents, i; |
| 58 | |||
| 59 | for_each_sg(sgl, sg, data->size, i) | ||
| 60 | cmd_data_len += ib_sg_dma_len(dev, sg); | ||
| 55 | 61 | ||
| 56 | if (cmd_data_len > ISER_KMALLOC_THRESHOLD) | 62 | if (cmd_data_len > ISER_KMALLOC_THRESHOLD) |
| 57 | mem = (void *)__get_free_pages(GFP_ATOMIC, | 63 | mem = (void *)__get_free_pages(GFP_ATOMIC, |
| @@ -61,17 +67,16 @@ static int iser_start_rdma_unaligned_sg(struct iscsi_iser_task *iser_task, | |||
| 61 | 67 | ||
| 62 | if (mem == NULL) { | 68 | if (mem == NULL) { |
| 63 | iser_err("Failed to allocate mem size %d %d for copying sglist\n", | 69 | iser_err("Failed to allocate mem size %d %d for copying sglist\n", |
| 64 | data->size,(int)cmd_data_len); | 70 | data->size, (int)cmd_data_len); |
| 65 | return -ENOMEM; | 71 | return -ENOMEM; |
| 66 | } | 72 | } |
| 67 | 73 | ||
| 68 | if (cmd_dir == ISER_DIR_OUT) { | 74 | if (cmd_dir == ISER_DIR_OUT) { |
| 69 | /* copy the unaligned sg the buffer which is used for RDMA */ | 75 | /* copy the unaligned sg the buffer which is used for RDMA */ |
| 70 | struct scatterlist *sgl = (struct scatterlist *)data->buf; | ||
| 71 | struct scatterlist *sg; | ||
| 72 | int i; | 76 | int i; |
| 73 | char *p, *from; | 77 | char *p, *from; |
| 74 | 78 | ||
| 79 | sgl = (struct scatterlist *)data->buf; | ||
| 75 | p = mem; | 80 | p = mem; |
| 76 | for_each_sg(sgl, sg, data->size, i) { | 81 | for_each_sg(sgl, sg, data->size, i) { |
| 77 | from = kmap_atomic(sg_page(sg)); | 82 | from = kmap_atomic(sg_page(sg)); |
| @@ -83,22 +88,19 @@ static int iser_start_rdma_unaligned_sg(struct iscsi_iser_task *iser_task, | |||
| 83 | } | 88 | } |
| 84 | } | 89 | } |
| 85 | 90 | ||
| 86 | sg_init_one(&iser_task->data_copy[cmd_dir].sg_single, mem, cmd_data_len); | 91 | sg_init_one(&data_copy->sg_single, mem, cmd_data_len); |
| 87 | iser_task->data_copy[cmd_dir].buf = | 92 | data_copy->buf = &data_copy->sg_single; |
| 88 | &iser_task->data_copy[cmd_dir].sg_single; | 93 | data_copy->size = 1; |
| 89 | iser_task->data_copy[cmd_dir].size = 1; | 94 | data_copy->copy_buf = mem; |
| 90 | 95 | ||
| 91 | iser_task->data_copy[cmd_dir].copy_buf = mem; | 96 | dma_nents = ib_dma_map_sg(dev, &data_copy->sg_single, 1, |
| 92 | |||
| 93 | dev = iser_task->iser_conn->ib_conn->device->ib_device; | ||
| 94 | dma_nents = ib_dma_map_sg(dev, | ||
| 95 | &iser_task->data_copy[cmd_dir].sg_single, | ||
| 96 | 1, | ||
| 97 | (cmd_dir == ISER_DIR_OUT) ? | 97 | (cmd_dir == ISER_DIR_OUT) ? |
| 98 | DMA_TO_DEVICE : DMA_FROM_DEVICE); | 98 | DMA_TO_DEVICE : DMA_FROM_DEVICE); |
| 99 | BUG_ON(dma_nents == 0); | 99 | BUG_ON(dma_nents == 0); |
| 100 | 100 | ||
| 101 | iser_task->data_copy[cmd_dir].dma_nents = dma_nents; | 101 | data_copy->dma_nents = dma_nents; |
| 102 | data_copy->data_len = cmd_data_len; | ||
| 103 | |||
| 102 | return 0; | 104 | return 0; |
| 103 | } | 105 | } |
| 104 | 106 | ||
| @@ -341,11 +343,12 @@ void iser_dma_unmap_task_data(struct iscsi_iser_task *iser_task, | |||
| 341 | 343 | ||
| 342 | static int fall_to_bounce_buf(struct iscsi_iser_task *iser_task, | 344 | static int fall_to_bounce_buf(struct iscsi_iser_task *iser_task, |
| 343 | struct ib_device *ibdev, | 345 | struct ib_device *ibdev, |
| 346 | struct iser_data_buf *mem, | ||
| 347 | struct iser_data_buf *mem_copy, | ||
| 344 | enum iser_data_dir cmd_dir, | 348 | enum iser_data_dir cmd_dir, |
| 345 | int aligned_len) | 349 | int aligned_len) |
| 346 | { | 350 | { |
| 347 | struct iscsi_conn *iscsi_conn = iser_task->iser_conn->iscsi_conn; | 351 | struct iscsi_conn *iscsi_conn = iser_task->iser_conn->iscsi_conn; |
| 348 | struct iser_data_buf *mem = &iser_task->data[cmd_dir]; | ||
| 349 | 352 | ||
| 350 | iscsi_conn->fmr_unalign_cnt++; | 353 | iscsi_conn->fmr_unalign_cnt++; |
| 351 | iser_warn("rdma alignment violation (%d/%d aligned) or FMR not supported\n", | 354 | iser_warn("rdma alignment violation (%d/%d aligned) or FMR not supported\n", |
| @@ -355,12 +358,12 @@ static int fall_to_bounce_buf(struct iscsi_iser_task *iser_task, | |||
| 355 | iser_data_buf_dump(mem, ibdev); | 358 | iser_data_buf_dump(mem, ibdev); |
| 356 | 359 | ||
| 357 | /* unmap the command data before accessing it */ | 360 | /* unmap the command data before accessing it */ |
| 358 | iser_dma_unmap_task_data(iser_task, &iser_task->data[cmd_dir]); | 361 | iser_dma_unmap_task_data(iser_task, mem); |
| 359 | 362 | ||
| 360 | /* allocate copy buf, if we are writing, copy the */ | 363 | /* allocate copy buf, if we are writing, copy the */ |
| 361 | /* unaligned scatterlist, dma map the copy */ | 364 | /* unaligned scatterlist, dma map the copy */ |
| 362 | if (iser_start_rdma_unaligned_sg(iser_task, cmd_dir) != 0) | 365 | if (iser_start_rdma_unaligned_sg(iser_task, mem, mem_copy, cmd_dir) != 0) |
| 363 | return -ENOMEM; | 366 | return -ENOMEM; |
| 364 | 367 | ||
| 365 | return 0; | 368 | return 0; |
| 366 | } | 369 | } |
| @@ -388,7 +391,8 @@ int iser_reg_rdma_mem_fmr(struct iscsi_iser_task *iser_task, | |||
| 388 | 391 | ||
| 389 | aligned_len = iser_data_buf_aligned_len(mem, ibdev); | 392 | aligned_len = iser_data_buf_aligned_len(mem, ibdev); |
| 390 | if (aligned_len != mem->dma_nents) { | 393 | if (aligned_len != mem->dma_nents) { |
| 391 | err = fall_to_bounce_buf(iser_task, ibdev, | 394 | err = fall_to_bounce_buf(iser_task, ibdev, mem, |
| 395 | &iser_task->data_copy[cmd_dir], | ||
| 392 | cmd_dir, aligned_len); | 396 | cmd_dir, aligned_len); |
| 393 | if (err) { | 397 | if (err) { |
| 394 | iser_err("failed to allocate bounce buffer\n"); | 398 | iser_err("failed to allocate bounce buffer\n"); |
| @@ -536,7 +540,8 @@ int iser_reg_rdma_mem_fastreg(struct iscsi_iser_task *iser_task, | |||
| 536 | 540 | ||
| 537 | aligned_len = iser_data_buf_aligned_len(mem, ibdev); | 541 | aligned_len = iser_data_buf_aligned_len(mem, ibdev); |
| 538 | if (aligned_len != mem->dma_nents) { | 542 | if (aligned_len != mem->dma_nents) { |
| 539 | err = fall_to_bounce_buf(iser_task, ibdev, | 543 | err = fall_to_bounce_buf(iser_task, ibdev, mem, |
| 544 | &iser_task->data_copy[cmd_dir], | ||
| 540 | cmd_dir, aligned_len); | 545 | cmd_dir, aligned_len); |
| 541 | if (err) { | 546 | if (err) { |
| 542 | iser_err("failed to allocate bounce buffer\n"); | 547 | iser_err("failed to allocate bounce buffer\n"); |
