diff options
author | Linus Torvalds <torvalds@woody.linux-foundation.org> | 2007-02-06 17:55:33 -0500 |
---|---|---|
committer | Linus Torvalds <torvalds@woody.linux-foundation.org> | 2007-02-06 17:55:33 -0500 |
commit | 0c7d3757116c59b3eedd9aa6dfd7ae0a1341f5c2 (patch) | |
tree | 61dbcee09d925a839b34df3b8149e26972dc5402 /drivers | |
parent | 2442d3109943bafbdfc4f0495e3d10eeedc8390c (diff) | |
parent | b45bfcc1ae084aa98c0350b8c33c8b57540b0acc (diff) |
Merge branch 'for-linus' of master.kernel.org:/pub/scm/linux/kernel/git/roland/infiniband
* 'for-linus' of master.kernel.org:/pub/scm/linux/kernel/git/roland/infiniband:
IB/ehca: Remove obsolete prototypes
IB/ehca: Remove use of do_mmap()
RDMA/addr: Handle ethernet neighbour updates during route resolution
IB: Make sure struct ib_user_mad.data is aligned
IB/srp: Don't wait for response when QP is in error state.
IB: Return qp pointer as part of ib_wc
IB: Include <linux/kref.h> explicitly in <rdma/ib_verbs.h>
Diffstat (limited to 'drivers')
-rw-r--r-- | drivers/infiniband/core/addr.c | 3 | ||||
-rw-r--r-- | drivers/infiniband/core/mad.c | 11 | ||||
-rw-r--r-- | drivers/infiniband/core/uverbs_cmd.c | 2 | ||||
-rw-r--r-- | drivers/infiniband/hw/amso1100/c2_cq.c | 2 | ||||
-rw-r--r-- | drivers/infiniband/hw/ehca/ehca_classes.h | 29 | ||||
-rw-r--r-- | drivers/infiniband/hw/ehca/ehca_cq.c | 65 | ||||
-rw-r--r-- | drivers/infiniband/hw/ehca/ehca_iverbs.h | 8 | ||||
-rw-r--r-- | drivers/infiniband/hw/ehca/ehca_main.c | 6 | ||||
-rw-r--r-- | drivers/infiniband/hw/ehca/ehca_qp.c | 78 | ||||
-rw-r--r-- | drivers/infiniband/hw/ehca/ehca_reqs.c | 2 | ||||
-rw-r--r-- | drivers/infiniband/hw/ehca/ehca_uverbs.c | 395 | ||||
-rw-r--r-- | drivers/infiniband/hw/ipath/ipath_qp.c | 2 | ||||
-rw-r--r-- | drivers/infiniband/hw/ipath/ipath_rc.c | 8 | ||||
-rw-r--r-- | drivers/infiniband/hw/ipath/ipath_ruc.c | 8 | ||||
-rw-r--r-- | drivers/infiniband/hw/ipath/ipath_uc.c | 4 | ||||
-rw-r--r-- | drivers/infiniband/hw/ipath/ipath_ud.c | 8 | ||||
-rw-r--r-- | drivers/infiniband/hw/mthca/mthca_cmd.c | 2 | ||||
-rw-r--r-- | drivers/infiniband/hw/mthca/mthca_cq.c | 2 | ||||
-rw-r--r-- | drivers/infiniband/ulp/srp/ib_srp.c | 7 | ||||
-rw-r--r-- | drivers/infiniband/ulp/srp/ib_srp.h | 1 |
20 files changed, 240 insertions, 403 deletions
diff --git a/drivers/infiniband/core/addr.c b/drivers/infiniband/core/addr.c index af939796750d..d2bb5a9a303f 100644 --- a/drivers/infiniband/core/addr.c +++ b/drivers/infiniband/core/addr.c | |||
@@ -360,8 +360,7 @@ static int netevent_callback(struct notifier_block *self, unsigned long event, | |||
360 | if (event == NETEVENT_NEIGH_UPDATE) { | 360 | if (event == NETEVENT_NEIGH_UPDATE) { |
361 | struct neighbour *neigh = ctx; | 361 | struct neighbour *neigh = ctx; |
362 | 362 | ||
363 | if (neigh->dev->type == ARPHRD_INFINIBAND && | 363 | if (neigh->nud_state & NUD_VALID) { |
364 | (neigh->nud_state & NUD_VALID)) { | ||
365 | set_timeout(jiffies); | 364 | set_timeout(jiffies); |
366 | } | 365 | } |
367 | } | 366 | } |
diff --git a/drivers/infiniband/core/mad.c b/drivers/infiniband/core/mad.c index 5ed141ebd1c8..13efd4170349 100644 --- a/drivers/infiniband/core/mad.c +++ b/drivers/infiniband/core/mad.c | |||
@@ -642,7 +642,8 @@ static void snoop_recv(struct ib_mad_qp_info *qp_info, | |||
642 | spin_unlock_irqrestore(&qp_info->snoop_lock, flags); | 642 | spin_unlock_irqrestore(&qp_info->snoop_lock, flags); |
643 | } | 643 | } |
644 | 644 | ||
645 | static void build_smp_wc(u64 wr_id, u16 slid, u16 pkey_index, u8 port_num, | 645 | static void build_smp_wc(struct ib_qp *qp, |
646 | u64 wr_id, u16 slid, u16 pkey_index, u8 port_num, | ||
646 | struct ib_wc *wc) | 647 | struct ib_wc *wc) |
647 | { | 648 | { |
648 | memset(wc, 0, sizeof *wc); | 649 | memset(wc, 0, sizeof *wc); |
@@ -652,7 +653,7 @@ static void build_smp_wc(u64 wr_id, u16 slid, u16 pkey_index, u8 port_num, | |||
652 | wc->pkey_index = pkey_index; | 653 | wc->pkey_index = pkey_index; |
653 | wc->byte_len = sizeof(struct ib_mad) + sizeof(struct ib_grh); | 654 | wc->byte_len = sizeof(struct ib_mad) + sizeof(struct ib_grh); |
654 | wc->src_qp = IB_QP0; | 655 | wc->src_qp = IB_QP0; |
655 | wc->qp_num = IB_QP0; | 656 | wc->qp = qp; |
656 | wc->slid = slid; | 657 | wc->slid = slid; |
657 | wc->sl = 0; | 658 | wc->sl = 0; |
658 | wc->dlid_path_bits = 0; | 659 | wc->dlid_path_bits = 0; |
@@ -713,7 +714,8 @@ static int handle_outgoing_dr_smp(struct ib_mad_agent_private *mad_agent_priv, | |||
713 | goto out; | 714 | goto out; |
714 | } | 715 | } |
715 | 716 | ||
716 | build_smp_wc(send_wr->wr_id, be16_to_cpu(smp->dr_slid), | 717 | build_smp_wc(mad_agent_priv->agent.qp, |
718 | send_wr->wr_id, be16_to_cpu(smp->dr_slid), | ||
717 | send_wr->wr.ud.pkey_index, | 719 | send_wr->wr.ud.pkey_index, |
718 | send_wr->wr.ud.port_num, &mad_wc); | 720 | send_wr->wr.ud.port_num, &mad_wc); |
719 | 721 | ||
@@ -2355,7 +2357,8 @@ static void local_completions(struct work_struct *work) | |||
2355 | * Defined behavior is to complete response | 2357 | * Defined behavior is to complete response |
2356 | * before request | 2358 | * before request |
2357 | */ | 2359 | */ |
2358 | build_smp_wc((unsigned long) local->mad_send_wr, | 2360 | build_smp_wc(recv_mad_agent->agent.qp, |
2361 | (unsigned long) local->mad_send_wr, | ||
2359 | be16_to_cpu(IB_LID_PERMISSIVE), | 2362 | be16_to_cpu(IB_LID_PERMISSIVE), |
2360 | 0, recv_mad_agent->agent.port_num, &wc); | 2363 | 0, recv_mad_agent->agent.port_num, &wc); |
2361 | 2364 | ||
diff --git a/drivers/infiniband/core/uverbs_cmd.c b/drivers/infiniband/core/uverbs_cmd.c index 743247ec065e..df1efbc10882 100644 --- a/drivers/infiniband/core/uverbs_cmd.c +++ b/drivers/infiniband/core/uverbs_cmd.c | |||
@@ -933,7 +933,7 @@ ssize_t ib_uverbs_poll_cq(struct ib_uverbs_file *file, | |||
933 | resp->wc[i].vendor_err = wc[i].vendor_err; | 933 | resp->wc[i].vendor_err = wc[i].vendor_err; |
934 | resp->wc[i].byte_len = wc[i].byte_len; | 934 | resp->wc[i].byte_len = wc[i].byte_len; |
935 | resp->wc[i].imm_data = (__u32 __force) wc[i].imm_data; | 935 | resp->wc[i].imm_data = (__u32 __force) wc[i].imm_data; |
936 | resp->wc[i].qp_num = wc[i].qp_num; | 936 | resp->wc[i].qp_num = wc[i].qp->qp_num; |
937 | resp->wc[i].src_qp = wc[i].src_qp; | 937 | resp->wc[i].src_qp = wc[i].src_qp; |
938 | resp->wc[i].wc_flags = wc[i].wc_flags; | 938 | resp->wc[i].wc_flags = wc[i].wc_flags; |
939 | resp->wc[i].pkey_index = wc[i].pkey_index; | 939 | resp->wc[i].pkey_index = wc[i].pkey_index; |
diff --git a/drivers/infiniband/hw/amso1100/c2_cq.c b/drivers/infiniband/hw/amso1100/c2_cq.c index 05c9154d46f4..5175c99ee586 100644 --- a/drivers/infiniband/hw/amso1100/c2_cq.c +++ b/drivers/infiniband/hw/amso1100/c2_cq.c | |||
@@ -153,7 +153,7 @@ static inline int c2_poll_one(struct c2_dev *c2dev, | |||
153 | 153 | ||
154 | entry->status = c2_cqe_status_to_openib(c2_wr_get_result(ce)); | 154 | entry->status = c2_cqe_status_to_openib(c2_wr_get_result(ce)); |
155 | entry->wr_id = ce->hdr.context; | 155 | entry->wr_id = ce->hdr.context; |
156 | entry->qp_num = ce->handle; | 156 | entry->qp = &qp->ibqp; |
157 | entry->wc_flags = 0; | 157 | entry->wc_flags = 0; |
158 | entry->slid = 0; | 158 | entry->slid = 0; |
159 | entry->sl = 0; | 159 | entry->sl = 0; |
diff --git a/drivers/infiniband/hw/ehca/ehca_classes.h b/drivers/infiniband/hw/ehca/ehca_classes.h index 1c722032319c..cf95ee474b0f 100644 --- a/drivers/infiniband/hw/ehca/ehca_classes.h +++ b/drivers/infiniband/hw/ehca/ehca_classes.h | |||
@@ -119,13 +119,14 @@ struct ehca_qp { | |||
119 | struct ipz_qp_handle ipz_qp_handle; | 119 | struct ipz_qp_handle ipz_qp_handle; |
120 | struct ehca_pfqp pf; | 120 | struct ehca_pfqp pf; |
121 | struct ib_qp_init_attr init_attr; | 121 | struct ib_qp_init_attr init_attr; |
122 | u64 uspace_squeue; | ||
123 | u64 uspace_rqueue; | ||
124 | u64 uspace_fwh; | ||
125 | struct ehca_cq *send_cq; | 122 | struct ehca_cq *send_cq; |
126 | struct ehca_cq *recv_cq; | 123 | struct ehca_cq *recv_cq; |
127 | unsigned int sqerr_purgeflag; | 124 | unsigned int sqerr_purgeflag; |
128 | struct hlist_node list_entries; | 125 | struct hlist_node list_entries; |
126 | /* mmap counter for resources mapped into user space */ | ||
127 | u32 mm_count_squeue; | ||
128 | u32 mm_count_rqueue; | ||
129 | u32 mm_count_galpa; | ||
129 | }; | 130 | }; |
130 | 131 | ||
131 | /* must be power of 2 */ | 132 | /* must be power of 2 */ |
@@ -142,13 +143,14 @@ struct ehca_cq { | |||
142 | struct ipz_cq_handle ipz_cq_handle; | 143 | struct ipz_cq_handle ipz_cq_handle; |
143 | struct ehca_pfcq pf; | 144 | struct ehca_pfcq pf; |
144 | spinlock_t cb_lock; | 145 | spinlock_t cb_lock; |
145 | u64 uspace_queue; | ||
146 | u64 uspace_fwh; | ||
147 | struct hlist_head qp_hashtab[QP_HASHTAB_LEN]; | 146 | struct hlist_head qp_hashtab[QP_HASHTAB_LEN]; |
148 | struct list_head entry; | 147 | struct list_head entry; |
149 | u32 nr_callbacks; | 148 | u32 nr_callbacks; |
150 | spinlock_t task_lock; | 149 | spinlock_t task_lock; |
151 | u32 ownpid; | 150 | u32 ownpid; |
151 | /* mmap counter for resources mapped into user space */ | ||
152 | u32 mm_count_queue; | ||
153 | u32 mm_count_galpa; | ||
152 | }; | 154 | }; |
153 | 155 | ||
154 | enum ehca_mr_flag { | 156 | enum ehca_mr_flag { |
@@ -248,20 +250,6 @@ struct ehca_ucontext { | |||
248 | struct ib_ucontext ib_ucontext; | 250 | struct ib_ucontext ib_ucontext; |
249 | }; | 251 | }; |
250 | 252 | ||
251 | struct ehca_module *ehca_module_new(void); | ||
252 | |||
253 | int ehca_module_delete(struct ehca_module *me); | ||
254 | |||
255 | int ehca_eq_ctor(struct ehca_eq *eq); | ||
256 | |||
257 | int ehca_eq_dtor(struct ehca_eq *eq); | ||
258 | |||
259 | struct ehca_shca *ehca_shca_new(void); | ||
260 | |||
261 | int ehca_shca_delete(struct ehca_shca *me); | ||
262 | |||
263 | struct ehca_sport *ehca_sport_new(struct ehca_shca *anchor); | ||
264 | |||
265 | int ehca_init_pd_cache(void); | 253 | int ehca_init_pd_cache(void); |
266 | void ehca_cleanup_pd_cache(void); | 254 | void ehca_cleanup_pd_cache(void); |
267 | int ehca_init_cq_cache(void); | 255 | int ehca_init_cq_cache(void); |
@@ -283,7 +271,6 @@ extern int ehca_port_act_time; | |||
283 | extern int ehca_use_hp_mr; | 271 | extern int ehca_use_hp_mr; |
284 | 272 | ||
285 | struct ipzu_queue_resp { | 273 | struct ipzu_queue_resp { |
286 | u64 queue; /* points to first queue entry */ | ||
287 | u32 qe_size; /* queue entry size */ | 274 | u32 qe_size; /* queue entry size */ |
288 | u32 act_nr_of_sg; | 275 | u32 act_nr_of_sg; |
289 | u32 queue_length; /* queue length allocated in bytes */ | 276 | u32 queue_length; /* queue length allocated in bytes */ |
@@ -296,7 +283,6 @@ struct ehca_create_cq_resp { | |||
296 | u32 cq_number; | 283 | u32 cq_number; |
297 | u32 token; | 284 | u32 token; |
298 | struct ipzu_queue_resp ipz_queue; | 285 | struct ipzu_queue_resp ipz_queue; |
299 | struct h_galpas galpas; | ||
300 | }; | 286 | }; |
301 | 287 | ||
302 | struct ehca_create_qp_resp { | 288 | struct ehca_create_qp_resp { |
@@ -309,7 +295,6 @@ struct ehca_create_qp_resp { | |||
309 | u32 dummy; /* padding for 8 byte alignment */ | 295 | u32 dummy; /* padding for 8 byte alignment */ |
310 | struct ipzu_queue_resp ipz_squeue; | 296 | struct ipzu_queue_resp ipz_squeue; |
311 | struct ipzu_queue_resp ipz_rqueue; | 297 | struct ipzu_queue_resp ipz_rqueue; |
312 | struct h_galpas galpas; | ||
313 | }; | 298 | }; |
314 | 299 | ||
315 | struct ehca_alloc_cq_parms { | 300 | struct ehca_alloc_cq_parms { |
diff --git a/drivers/infiniband/hw/ehca/ehca_cq.c b/drivers/infiniband/hw/ehca/ehca_cq.c index 6074c897f51c..9291a86ca053 100644 --- a/drivers/infiniband/hw/ehca/ehca_cq.c +++ b/drivers/infiniband/hw/ehca/ehca_cq.c | |||
@@ -267,7 +267,6 @@ struct ib_cq *ehca_create_cq(struct ib_device *device, int cqe, | |||
267 | if (context) { | 267 | if (context) { |
268 | struct ipz_queue *ipz_queue = &my_cq->ipz_queue; | 268 | struct ipz_queue *ipz_queue = &my_cq->ipz_queue; |
269 | struct ehca_create_cq_resp resp; | 269 | struct ehca_create_cq_resp resp; |
270 | struct vm_area_struct *vma; | ||
271 | memset(&resp, 0, sizeof(resp)); | 270 | memset(&resp, 0, sizeof(resp)); |
272 | resp.cq_number = my_cq->cq_number; | 271 | resp.cq_number = my_cq->cq_number; |
273 | resp.token = my_cq->token; | 272 | resp.token = my_cq->token; |
@@ -276,40 +275,14 @@ struct ib_cq *ehca_create_cq(struct ib_device *device, int cqe, | |||
276 | resp.ipz_queue.queue_length = ipz_queue->queue_length; | 275 | resp.ipz_queue.queue_length = ipz_queue->queue_length; |
277 | resp.ipz_queue.pagesize = ipz_queue->pagesize; | 276 | resp.ipz_queue.pagesize = ipz_queue->pagesize; |
278 | resp.ipz_queue.toggle_state = ipz_queue->toggle_state; | 277 | resp.ipz_queue.toggle_state = ipz_queue->toggle_state; |
279 | ret = ehca_mmap_nopage(((u64)(my_cq->token) << 32) | 0x12000000, | ||
280 | ipz_queue->queue_length, | ||
281 | (void**)&resp.ipz_queue.queue, | ||
282 | &vma); | ||
283 | if (ret) { | ||
284 | ehca_err(device, "Could not mmap queue pages"); | ||
285 | cq = ERR_PTR(ret); | ||
286 | goto create_cq_exit4; | ||
287 | } | ||
288 | my_cq->uspace_queue = resp.ipz_queue.queue; | ||
289 | resp.galpas = my_cq->galpas; | ||
290 | ret = ehca_mmap_register(my_cq->galpas.user.fw_handle, | ||
291 | (void**)&resp.galpas.kernel.fw_handle, | ||
292 | &vma); | ||
293 | if (ret) { | ||
294 | ehca_err(device, "Could not mmap fw_handle"); | ||
295 | cq = ERR_PTR(ret); | ||
296 | goto create_cq_exit5; | ||
297 | } | ||
298 | my_cq->uspace_fwh = (u64)resp.galpas.kernel.fw_handle; | ||
299 | if (ib_copy_to_udata(udata, &resp, sizeof(resp))) { | 278 | if (ib_copy_to_udata(udata, &resp, sizeof(resp))) { |
300 | ehca_err(device, "Copy to udata failed."); | 279 | ehca_err(device, "Copy to udata failed."); |
301 | goto create_cq_exit6; | 280 | goto create_cq_exit4; |
302 | } | 281 | } |
303 | } | 282 | } |
304 | 283 | ||
305 | return cq; | 284 | return cq; |
306 | 285 | ||
307 | create_cq_exit6: | ||
308 | ehca_munmap(my_cq->uspace_fwh, EHCA_PAGESIZE); | ||
309 | |||
310 | create_cq_exit5: | ||
311 | ehca_munmap(my_cq->uspace_queue, my_cq->ipz_queue.queue_length); | ||
312 | |||
313 | create_cq_exit4: | 286 | create_cq_exit4: |
314 | ipz_queue_dtor(&my_cq->ipz_queue); | 287 | ipz_queue_dtor(&my_cq->ipz_queue); |
315 | 288 | ||
@@ -333,7 +306,6 @@ create_cq_exit1: | |||
333 | int ehca_destroy_cq(struct ib_cq *cq) | 306 | int ehca_destroy_cq(struct ib_cq *cq) |
334 | { | 307 | { |
335 | u64 h_ret; | 308 | u64 h_ret; |
336 | int ret; | ||
337 | struct ehca_cq *my_cq = container_of(cq, struct ehca_cq, ib_cq); | 309 | struct ehca_cq *my_cq = container_of(cq, struct ehca_cq, ib_cq); |
338 | int cq_num = my_cq->cq_number; | 310 | int cq_num = my_cq->cq_number; |
339 | struct ib_device *device = cq->device; | 311 | struct ib_device *device = cq->device; |
@@ -343,6 +315,20 @@ int ehca_destroy_cq(struct ib_cq *cq) | |||
343 | u32 cur_pid = current->tgid; | 315 | u32 cur_pid = current->tgid; |
344 | unsigned long flags; | 316 | unsigned long flags; |
345 | 317 | ||
318 | if (cq->uobject) { | ||
319 | if (my_cq->mm_count_galpa || my_cq->mm_count_queue) { | ||
320 | ehca_err(device, "Resources still referenced in " | ||
321 | "user space cq_num=%x", my_cq->cq_number); | ||
322 | return -EINVAL; | ||
323 | } | ||
324 | if (my_cq->ownpid != cur_pid) { | ||
325 | ehca_err(device, "Invalid caller pid=%x ownpid=%x " | ||
326 | "cq_num=%x", | ||
327 | cur_pid, my_cq->ownpid, my_cq->cq_number); | ||
328 | return -EINVAL; | ||
329 | } | ||
330 | } | ||
331 | |||
346 | spin_lock_irqsave(&ehca_cq_idr_lock, flags); | 332 | spin_lock_irqsave(&ehca_cq_idr_lock, flags); |
347 | while (my_cq->nr_callbacks) { | 333 | while (my_cq->nr_callbacks) { |
348 | spin_unlock_irqrestore(&ehca_cq_idr_lock, flags); | 334 | spin_unlock_irqrestore(&ehca_cq_idr_lock, flags); |
@@ -353,25 +339,6 @@ int ehca_destroy_cq(struct ib_cq *cq) | |||
353 | idr_remove(&ehca_cq_idr, my_cq->token); | 339 | idr_remove(&ehca_cq_idr, my_cq->token); |
354 | spin_unlock_irqrestore(&ehca_cq_idr_lock, flags); | 340 | spin_unlock_irqrestore(&ehca_cq_idr_lock, flags); |
355 | 341 | ||
356 | if (my_cq->uspace_queue && my_cq->ownpid != cur_pid) { | ||
357 | ehca_err(device, "Invalid caller pid=%x ownpid=%x", | ||
358 | cur_pid, my_cq->ownpid); | ||
359 | return -EINVAL; | ||
360 | } | ||
361 | |||
362 | /* un-mmap if vma alloc */ | ||
363 | if (my_cq->uspace_queue ) { | ||
364 | ret = ehca_munmap(my_cq->uspace_queue, | ||
365 | my_cq->ipz_queue.queue_length); | ||
366 | if (ret) | ||
367 | ehca_err(device, "Could not munmap queue ehca_cq=%p " | ||
368 | "cq_num=%x", my_cq, cq_num); | ||
369 | ret = ehca_munmap(my_cq->uspace_fwh, EHCA_PAGESIZE); | ||
370 | if (ret) | ||
371 | ehca_err(device, "Could not munmap fwh ehca_cq=%p " | ||
372 | "cq_num=%x", my_cq, cq_num); | ||
373 | } | ||
374 | |||
375 | h_ret = hipz_h_destroy_cq(adapter_handle, my_cq, 0); | 342 | h_ret = hipz_h_destroy_cq(adapter_handle, my_cq, 0); |
376 | if (h_ret == H_R_STATE) { | 343 | if (h_ret == H_R_STATE) { |
377 | /* cq in err: read err data and destroy it forcibly */ | 344 | /* cq in err: read err data and destroy it forcibly */ |
@@ -400,7 +367,7 @@ int ehca_resize_cq(struct ib_cq *cq, int cqe, struct ib_udata *udata) | |||
400 | struct ehca_cq *my_cq = container_of(cq, struct ehca_cq, ib_cq); | 367 | struct ehca_cq *my_cq = container_of(cq, struct ehca_cq, ib_cq); |
401 | u32 cur_pid = current->tgid; | 368 | u32 cur_pid = current->tgid; |
402 | 369 | ||
403 | if (my_cq->uspace_queue && my_cq->ownpid != cur_pid) { | 370 | if (cq->uobject && my_cq->ownpid != cur_pid) { |
404 | ehca_err(cq->device, "Invalid caller pid=%x ownpid=%x", | 371 | ehca_err(cq->device, "Invalid caller pid=%x ownpid=%x", |
405 | cur_pid, my_cq->ownpid); | 372 | cur_pid, my_cq->ownpid); |
406 | return -EINVAL; | 373 | return -EINVAL; |
diff --git a/drivers/infiniband/hw/ehca/ehca_iverbs.h b/drivers/infiniband/hw/ehca/ehca_iverbs.h index cd7789f0d08e..95fd59fb4528 100644 --- a/drivers/infiniband/hw/ehca/ehca_iverbs.h +++ b/drivers/infiniband/hw/ehca/ehca_iverbs.h | |||
@@ -171,14 +171,6 @@ int ehca_mmap(struct ib_ucontext *context, struct vm_area_struct *vma); | |||
171 | 171 | ||
172 | void ehca_poll_eqs(unsigned long data); | 172 | void ehca_poll_eqs(unsigned long data); |
173 | 173 | ||
174 | int ehca_mmap_nopage(u64 foffset,u64 length,void **mapped, | ||
175 | struct vm_area_struct **vma); | ||
176 | |||
177 | int ehca_mmap_register(u64 physical,void **mapped, | ||
178 | struct vm_area_struct **vma); | ||
179 | |||
180 | int ehca_munmap(unsigned long addr, size_t len); | ||
181 | |||
182 | #ifdef CONFIG_PPC_64K_PAGES | 174 | #ifdef CONFIG_PPC_64K_PAGES |
183 | void *ehca_alloc_fw_ctrlblock(gfp_t flags); | 175 | void *ehca_alloc_fw_ctrlblock(gfp_t flags); |
184 | void ehca_free_fw_ctrlblock(void *ptr); | 176 | void ehca_free_fw_ctrlblock(void *ptr); |
diff --git a/drivers/infiniband/hw/ehca/ehca_main.c b/drivers/infiniband/hw/ehca/ehca_main.c index 6574fbbaead5..1155bcf48212 100644 --- a/drivers/infiniband/hw/ehca/ehca_main.c +++ b/drivers/infiniband/hw/ehca/ehca_main.c | |||
@@ -52,7 +52,7 @@ | |||
52 | MODULE_LICENSE("Dual BSD/GPL"); | 52 | MODULE_LICENSE("Dual BSD/GPL"); |
53 | MODULE_AUTHOR("Christoph Raisch <raisch@de.ibm.com>"); | 53 | MODULE_AUTHOR("Christoph Raisch <raisch@de.ibm.com>"); |
54 | MODULE_DESCRIPTION("IBM eServer HCA InfiniBand Device Driver"); | 54 | MODULE_DESCRIPTION("IBM eServer HCA InfiniBand Device Driver"); |
55 | MODULE_VERSION("SVNEHCA_0019"); | 55 | MODULE_VERSION("SVNEHCA_0020"); |
56 | 56 | ||
57 | int ehca_open_aqp1 = 0; | 57 | int ehca_open_aqp1 = 0; |
58 | int ehca_debug_level = 0; | 58 | int ehca_debug_level = 0; |
@@ -288,7 +288,7 @@ int ehca_init_device(struct ehca_shca *shca) | |||
288 | strlcpy(shca->ib_device.name, "ehca%d", IB_DEVICE_NAME_MAX); | 288 | strlcpy(shca->ib_device.name, "ehca%d", IB_DEVICE_NAME_MAX); |
289 | shca->ib_device.owner = THIS_MODULE; | 289 | shca->ib_device.owner = THIS_MODULE; |
290 | 290 | ||
291 | shca->ib_device.uverbs_abi_ver = 5; | 291 | shca->ib_device.uverbs_abi_ver = 6; |
292 | shca->ib_device.uverbs_cmd_mask = | 292 | shca->ib_device.uverbs_cmd_mask = |
293 | (1ull << IB_USER_VERBS_CMD_GET_CONTEXT) | | 293 | (1ull << IB_USER_VERBS_CMD_GET_CONTEXT) | |
294 | (1ull << IB_USER_VERBS_CMD_QUERY_DEVICE) | | 294 | (1ull << IB_USER_VERBS_CMD_QUERY_DEVICE) | |
@@ -790,7 +790,7 @@ int __init ehca_module_init(void) | |||
790 | int ret; | 790 | int ret; |
791 | 791 | ||
792 | printk(KERN_INFO "eHCA Infiniband Device Driver " | 792 | printk(KERN_INFO "eHCA Infiniband Device Driver " |
793 | "(Rel.: SVNEHCA_0019)\n"); | 793 | "(Rel.: SVNEHCA_0020)\n"); |
794 | idr_init(&ehca_qp_idr); | 794 | idr_init(&ehca_qp_idr); |
795 | idr_init(&ehca_cq_idr); | 795 | idr_init(&ehca_cq_idr); |
796 | spin_lock_init(&ehca_qp_idr_lock); | 796 | spin_lock_init(&ehca_qp_idr_lock); |
diff --git a/drivers/infiniband/hw/ehca/ehca_qp.c b/drivers/infiniband/hw/ehca/ehca_qp.c index 34b85556d01e..95efef921f1d 100644 --- a/drivers/infiniband/hw/ehca/ehca_qp.c +++ b/drivers/infiniband/hw/ehca/ehca_qp.c | |||
@@ -637,7 +637,6 @@ struct ib_qp *ehca_create_qp(struct ib_pd *pd, | |||
637 | struct ipz_queue *ipz_rqueue = &my_qp->ipz_rqueue; | 637 | struct ipz_queue *ipz_rqueue = &my_qp->ipz_rqueue; |
638 | struct ipz_queue *ipz_squeue = &my_qp->ipz_squeue; | 638 | struct ipz_queue *ipz_squeue = &my_qp->ipz_squeue; |
639 | struct ehca_create_qp_resp resp; | 639 | struct ehca_create_qp_resp resp; |
640 | struct vm_area_struct * vma; | ||
641 | memset(&resp, 0, sizeof(resp)); | 640 | memset(&resp, 0, sizeof(resp)); |
642 | 641 | ||
643 | resp.qp_num = my_qp->real_qp_num; | 642 | resp.qp_num = my_qp->real_qp_num; |
@@ -651,59 +650,21 @@ struct ib_qp *ehca_create_qp(struct ib_pd *pd, | |||
651 | resp.ipz_rqueue.queue_length = ipz_rqueue->queue_length; | 650 | resp.ipz_rqueue.queue_length = ipz_rqueue->queue_length; |
652 | resp.ipz_rqueue.pagesize = ipz_rqueue->pagesize; | 651 | resp.ipz_rqueue.pagesize = ipz_rqueue->pagesize; |
653 | resp.ipz_rqueue.toggle_state = ipz_rqueue->toggle_state; | 652 | resp.ipz_rqueue.toggle_state = ipz_rqueue->toggle_state; |
654 | ret = ehca_mmap_nopage(((u64)(my_qp->token) << 32) | 0x22000000, | ||
655 | ipz_rqueue->queue_length, | ||
656 | (void**)&resp.ipz_rqueue.queue, | ||
657 | &vma); | ||
658 | if (ret) { | ||
659 | ehca_err(pd->device, "Could not mmap rqueue pages"); | ||
660 | goto create_qp_exit3; | ||
661 | } | ||
662 | my_qp->uspace_rqueue = resp.ipz_rqueue.queue; | ||
663 | /* squeue properties */ | 653 | /* squeue properties */ |
664 | resp.ipz_squeue.qe_size = ipz_squeue->qe_size; | 654 | resp.ipz_squeue.qe_size = ipz_squeue->qe_size; |
665 | resp.ipz_squeue.act_nr_of_sg = ipz_squeue->act_nr_of_sg; | 655 | resp.ipz_squeue.act_nr_of_sg = ipz_squeue->act_nr_of_sg; |
666 | resp.ipz_squeue.queue_length = ipz_squeue->queue_length; | 656 | resp.ipz_squeue.queue_length = ipz_squeue->queue_length; |
667 | resp.ipz_squeue.pagesize = ipz_squeue->pagesize; | 657 | resp.ipz_squeue.pagesize = ipz_squeue->pagesize; |
668 | resp.ipz_squeue.toggle_state = ipz_squeue->toggle_state; | 658 | resp.ipz_squeue.toggle_state = ipz_squeue->toggle_state; |
669 | ret = ehca_mmap_nopage(((u64)(my_qp->token) << 32) | 0x23000000, | ||
670 | ipz_squeue->queue_length, | ||
671 | (void**)&resp.ipz_squeue.queue, | ||
672 | &vma); | ||
673 | if (ret) { | ||
674 | ehca_err(pd->device, "Could not mmap squeue pages"); | ||
675 | goto create_qp_exit4; | ||
676 | } | ||
677 | my_qp->uspace_squeue = resp.ipz_squeue.queue; | ||
678 | /* fw_handle */ | ||
679 | resp.galpas = my_qp->galpas; | ||
680 | ret = ehca_mmap_register(my_qp->galpas.user.fw_handle, | ||
681 | (void**)&resp.galpas.kernel.fw_handle, | ||
682 | &vma); | ||
683 | if (ret) { | ||
684 | ehca_err(pd->device, "Could not mmap fw_handle"); | ||
685 | goto create_qp_exit5; | ||
686 | } | ||
687 | my_qp->uspace_fwh = (u64)resp.galpas.kernel.fw_handle; | ||
688 | |||
689 | if (ib_copy_to_udata(udata, &resp, sizeof resp)) { | 659 | if (ib_copy_to_udata(udata, &resp, sizeof resp)) { |
690 | ehca_err(pd->device, "Copy to udata failed"); | 660 | ehca_err(pd->device, "Copy to udata failed"); |
691 | ret = -EINVAL; | 661 | ret = -EINVAL; |
692 | goto create_qp_exit6; | 662 | goto create_qp_exit3; |
693 | } | 663 | } |
694 | } | 664 | } |
695 | 665 | ||
696 | return &my_qp->ib_qp; | 666 | return &my_qp->ib_qp; |
697 | 667 | ||
698 | create_qp_exit6: | ||
699 | ehca_munmap(my_qp->uspace_fwh, EHCA_PAGESIZE); | ||
700 | |||
701 | create_qp_exit5: | ||
702 | ehca_munmap(my_qp->uspace_squeue, my_qp->ipz_squeue.queue_length); | ||
703 | |||
704 | create_qp_exit4: | ||
705 | ehca_munmap(my_qp->uspace_rqueue, my_qp->ipz_rqueue.queue_length); | ||
706 | |||
707 | create_qp_exit3: | 668 | create_qp_exit3: |
708 | ipz_queue_dtor(&my_qp->ipz_rqueue); | 669 | ipz_queue_dtor(&my_qp->ipz_rqueue); |
709 | ipz_queue_dtor(&my_qp->ipz_squeue); | 670 | ipz_queue_dtor(&my_qp->ipz_squeue); |
@@ -931,7 +892,7 @@ static int internal_modify_qp(struct ib_qp *ibqp, | |||
931 | my_qp->qp_type == IB_QPT_SMI) && | 892 | my_qp->qp_type == IB_QPT_SMI) && |
932 | statetrans == IB_QPST_SQE2RTS) { | 893 | statetrans == IB_QPST_SQE2RTS) { |
933 | /* mark next free wqe if kernel */ | 894 | /* mark next free wqe if kernel */ |
934 | if (my_qp->uspace_squeue == 0) { | 895 | if (!ibqp->uobject) { |
935 | struct ehca_wqe *wqe; | 896 | struct ehca_wqe *wqe; |
936 | /* lock send queue */ | 897 | /* lock send queue */ |
937 | spin_lock_irqsave(&my_qp->spinlock_s, spl_flags); | 898 | spin_lock_irqsave(&my_qp->spinlock_s, spl_flags); |
@@ -1417,11 +1378,18 @@ int ehca_destroy_qp(struct ib_qp *ibqp) | |||
1417 | enum ib_qp_type qp_type; | 1378 | enum ib_qp_type qp_type; |
1418 | unsigned long flags; | 1379 | unsigned long flags; |
1419 | 1380 | ||
1420 | if (my_pd->ib_pd.uobject && my_pd->ib_pd.uobject->context && | 1381 | if (ibqp->uobject) { |
1421 | my_pd->ownpid != cur_pid) { | 1382 | if (my_qp->mm_count_galpa || |
1422 | ehca_err(ibqp->device, "Invalid caller pid=%x ownpid=%x", | 1383 | my_qp->mm_count_rqueue || my_qp->mm_count_squeue) { |
1423 | cur_pid, my_pd->ownpid); | 1384 | ehca_err(ibqp->device, "Resources still referenced in " |
1424 | return -EINVAL; | 1385 | "user space qp_num=%x", ibqp->qp_num); |
1386 | return -EINVAL; | ||
1387 | } | ||
1388 | if (my_pd->ownpid != cur_pid) { | ||
1389 | ehca_err(ibqp->device, "Invalid caller pid=%x ownpid=%x", | ||
1390 | cur_pid, my_pd->ownpid); | ||
1391 | return -EINVAL; | ||
1392 | } | ||
1425 | } | 1393 | } |
1426 | 1394 | ||
1427 | if (my_qp->send_cq) { | 1395 | if (my_qp->send_cq) { |
@@ -1439,24 +1407,6 @@ int ehca_destroy_qp(struct ib_qp *ibqp) | |||
1439 | idr_remove(&ehca_qp_idr, my_qp->token); | 1407 | idr_remove(&ehca_qp_idr, my_qp->token); |
1440 | spin_unlock_irqrestore(&ehca_qp_idr_lock, flags); | 1408 | spin_unlock_irqrestore(&ehca_qp_idr_lock, flags); |
1441 | 1409 | ||
1442 | /* un-mmap if vma alloc */ | ||
1443 | if (my_qp->uspace_rqueue) { | ||
1444 | ret = ehca_munmap(my_qp->uspace_rqueue, | ||
1445 | my_qp->ipz_rqueue.queue_length); | ||
1446 | if (ret) | ||
1447 | ehca_err(ibqp->device, "Could not munmap rqueue " | ||
1448 | "qp_num=%x", qp_num); | ||
1449 | ret = ehca_munmap(my_qp->uspace_squeue, | ||
1450 | my_qp->ipz_squeue.queue_length); | ||
1451 | if (ret) | ||
1452 | ehca_err(ibqp->device, "Could not munmap squeue " | ||
1453 | "qp_num=%x", qp_num); | ||
1454 | ret = ehca_munmap(my_qp->uspace_fwh, EHCA_PAGESIZE); | ||
1455 | if (ret) | ||
1456 | ehca_err(ibqp->device, "Could not munmap fwh qp_num=%x", | ||
1457 | qp_num); | ||
1458 | } | ||
1459 | |||
1460 | h_ret = hipz_h_destroy_qp(shca->ipz_hca_handle, my_qp); | 1410 | h_ret = hipz_h_destroy_qp(shca->ipz_hca_handle, my_qp); |
1461 | if (h_ret != H_SUCCESS) { | 1411 | if (h_ret != H_SUCCESS) { |
1462 | ehca_err(ibqp->device, "hipz_h_destroy_qp() failed rc=%lx " | 1412 | ehca_err(ibqp->device, "hipz_h_destroy_qp() failed rc=%lx " |
diff --git a/drivers/infiniband/hw/ehca/ehca_reqs.c b/drivers/infiniband/hw/ehca/ehca_reqs.c index b46bda1bf85d..08d3f892d9f3 100644 --- a/drivers/infiniband/hw/ehca/ehca_reqs.c +++ b/drivers/infiniband/hw/ehca/ehca_reqs.c | |||
@@ -579,7 +579,7 @@ poll_cq_one_read_cqe: | |||
579 | } else | 579 | } else |
580 | wc->status = IB_WC_SUCCESS; | 580 | wc->status = IB_WC_SUCCESS; |
581 | 581 | ||
582 | wc->qp_num = cqe->local_qp_number; | 582 | wc->qp = NULL; |
583 | wc->byte_len = cqe->nr_bytes_transferred; | 583 | wc->byte_len = cqe->nr_bytes_transferred; |
584 | wc->pkey_index = cqe->pkey_index; | 584 | wc->pkey_index = cqe->pkey_index; |
585 | wc->slid = cqe->rlid; | 585 | wc->slid = cqe->rlid; |
diff --git a/drivers/infiniband/hw/ehca/ehca_uverbs.c b/drivers/infiniband/hw/ehca/ehca_uverbs.c index e08764e4aef2..73db920b6945 100644 --- a/drivers/infiniband/hw/ehca/ehca_uverbs.c +++ b/drivers/infiniband/hw/ehca/ehca_uverbs.c | |||
@@ -68,105 +68,183 @@ int ehca_dealloc_ucontext(struct ib_ucontext *context) | |||
68 | return 0; | 68 | return 0; |
69 | } | 69 | } |
70 | 70 | ||
71 | struct page *ehca_nopage(struct vm_area_struct *vma, | 71 | static void ehca_mm_open(struct vm_area_struct *vma) |
72 | unsigned long address, int *type) | ||
73 | { | 72 | { |
74 | struct page *mypage = NULL; | 73 | u32 *count = (u32*)vma->vm_private_data; |
75 | u64 fileoffset = vma->vm_pgoff << PAGE_SHIFT; | 74 | if (!count) { |
76 | u32 idr_handle = fileoffset >> 32; | 75 | ehca_gen_err("Invalid vma struct vm_start=%lx vm_end=%lx", |
77 | u32 q_type = (fileoffset >> 28) & 0xF; /* CQ, QP,... */ | 76 | vma->vm_start, vma->vm_end); |
78 | u32 rsrc_type = (fileoffset >> 24) & 0xF; /* sq,rq,cmnd_window */ | 77 | return; |
79 | u32 cur_pid = current->tgid; | 78 | } |
80 | unsigned long flags; | 79 | (*count)++; |
81 | struct ehca_cq *cq; | 80 | if (!(*count)) |
82 | struct ehca_qp *qp; | 81 | ehca_gen_err("Use count overflow vm_start=%lx vm_end=%lx", |
83 | struct ehca_pd *pd; | 82 | vma->vm_start, vma->vm_end); |
84 | u64 offset; | 83 | ehca_gen_dbg("vm_start=%lx vm_end=%lx count=%x", |
85 | void *vaddr; | 84 | vma->vm_start, vma->vm_end, *count); |
85 | } | ||
86 | 86 | ||
87 | switch (q_type) { | 87 | static void ehca_mm_close(struct vm_area_struct *vma) |
88 | case 1: /* CQ */ | 88 | { |
89 | spin_lock_irqsave(&ehca_cq_idr_lock, flags); | 89 | u32 *count = (u32*)vma->vm_private_data; |
90 | cq = idr_find(&ehca_cq_idr, idr_handle); | 90 | if (!count) { |
91 | spin_unlock_irqrestore(&ehca_cq_idr_lock, flags); | 91 | ehca_gen_err("Invalid vma struct vm_start=%lx vm_end=%lx", |
92 | vma->vm_start, vma->vm_end); | ||
93 | return; | ||
94 | } | ||
95 | (*count)--; | ||
96 | ehca_gen_dbg("vm_start=%lx vm_end=%lx count=%x", | ||
97 | vma->vm_start, vma->vm_end, *count); | ||
98 | } | ||
92 | 99 | ||
93 | /* make sure this mmap really belongs to the authorized user */ | 100 | static struct vm_operations_struct vm_ops = { |
94 | if (!cq) { | 101 | .open = ehca_mm_open, |
95 | ehca_gen_err("cq is NULL ret=NOPAGE_SIGBUS"); | 102 | .close = ehca_mm_close, |
96 | return NOPAGE_SIGBUS; | 103 | }; |
104 | |||
105 | static int ehca_mmap_fw(struct vm_area_struct *vma, struct h_galpas *galpas, | ||
106 | u32 *mm_count) | ||
107 | { | ||
108 | int ret; | ||
109 | u64 vsize, physical; | ||
110 | |||
111 | vsize = vma->vm_end - vma->vm_start; | ||
112 | if (vsize != EHCA_PAGESIZE) { | ||
113 | ehca_gen_err("invalid vsize=%lx", vma->vm_end - vma->vm_start); | ||
114 | return -EINVAL; | ||
115 | } | ||
116 | |||
117 | physical = galpas->user.fw_handle; | ||
118 | vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); | ||
119 | ehca_gen_dbg("vsize=%lx physical=%lx", vsize, physical); | ||
120 | /* VM_IO | VM_RESERVED are set by remap_pfn_range() */ | ||
121 | ret = remap_pfn_range(vma, vma->vm_start, physical >> PAGE_SHIFT, | ||
122 | vsize, vma->vm_page_prot); | ||
123 | if (unlikely(ret)) { | ||
124 | ehca_gen_err("remap_pfn_range() failed ret=%x", ret); | ||
125 | return -ENOMEM; | ||
126 | } | ||
127 | |||
128 | vma->vm_private_data = mm_count; | ||
129 | (*mm_count)++; | ||
130 | vma->vm_ops = &vm_ops; | ||
131 | |||
132 | return 0; | ||
133 | } | ||
134 | |||
135 | static int ehca_mmap_queue(struct vm_area_struct *vma, struct ipz_queue *queue, | ||
136 | u32 *mm_count) | ||
137 | { | ||
138 | int ret; | ||
139 | u64 start, ofs; | ||
140 | struct page *page; | ||
141 | |||
142 | vma->vm_flags |= VM_RESERVED; | ||
143 | start = vma->vm_start; | ||
144 | for (ofs = 0; ofs < queue->queue_length; ofs += PAGE_SIZE) { | ||
145 | u64 virt_addr = (u64)ipz_qeit_calc(queue, ofs); | ||
146 | page = virt_to_page(virt_addr); | ||
147 | ret = vm_insert_page(vma, start, page); | ||
148 | if (unlikely(ret)) { | ||
149 | ehca_gen_err("vm_insert_page() failed rc=%x", ret); | ||
150 | return ret; | ||
97 | } | 151 | } |
152 | start += PAGE_SIZE; | ||
153 | } | ||
154 | vma->vm_private_data = mm_count; | ||
155 | (*mm_count)++; | ||
156 | vma->vm_ops = &vm_ops; | ||
98 | 157 | ||
99 | if (cq->ownpid != cur_pid) { | 158 | return 0; |
159 | } | ||
160 | |||
161 | static int ehca_mmap_cq(struct vm_area_struct *vma, struct ehca_cq *cq, | ||
162 | u32 rsrc_type) | ||
163 | { | ||
164 | int ret; | ||
165 | |||
166 | switch (rsrc_type) { | ||
167 | case 1: /* galpa fw handle */ | ||
168 | ehca_dbg(cq->ib_cq.device, "cq_num=%x fw", cq->cq_number); | ||
169 | ret = ehca_mmap_fw(vma, &cq->galpas, &cq->mm_count_galpa); | ||
170 | if (unlikely(ret)) { | ||
100 | ehca_err(cq->ib_cq.device, | 171 | ehca_err(cq->ib_cq.device, |
101 | "Invalid caller pid=%x ownpid=%x", | 172 | "ehca_mmap_fw() failed rc=%x cq_num=%x", |
102 | cur_pid, cq->ownpid); | 173 | ret, cq->cq_number); |
103 | return NOPAGE_SIGBUS; | 174 | return ret; |
104 | } | 175 | } |
176 | break; | ||
105 | 177 | ||
106 | if (rsrc_type == 2) { | 178 | case 2: /* cq queue_addr */ |
107 | ehca_dbg(cq->ib_cq.device, "cq=%p cq queuearea", cq); | 179 | ehca_dbg(cq->ib_cq.device, "cq_num=%x queue", cq->cq_number); |
108 | offset = address - vma->vm_start; | 180 | ret = ehca_mmap_queue(vma, &cq->ipz_queue, &cq->mm_count_queue); |
109 | vaddr = ipz_qeit_calc(&cq->ipz_queue, offset); | 181 | if (unlikely(ret)) { |
110 | ehca_dbg(cq->ib_cq.device, "offset=%lx vaddr=%p", | 182 | ehca_err(cq->ib_cq.device, |
111 | offset, vaddr); | 183 | "ehca_mmap_queue() failed rc=%x cq_num=%x", |
112 | mypage = virt_to_page(vaddr); | 184 | ret, cq->cq_number); |
185 | return ret; | ||
113 | } | 186 | } |
114 | break; | 187 | break; |
115 | 188 | ||
116 | case 2: /* QP */ | 189 | default: |
117 | spin_lock_irqsave(&ehca_qp_idr_lock, flags); | 190 | ehca_err(cq->ib_cq.device, "bad resource type=%x cq_num=%x", |
118 | qp = idr_find(&ehca_qp_idr, idr_handle); | 191 | rsrc_type, cq->cq_number); |
119 | spin_unlock_irqrestore(&ehca_qp_idr_lock, flags); | 192 | return -EINVAL; |
193 | } | ||
120 | 194 | ||
121 | /* make sure this mmap really belongs to the authorized user */ | 195 | return 0; |
122 | if (!qp) { | 196 | } |
123 | ehca_gen_err("qp is NULL ret=NOPAGE_SIGBUS"); | 197 | |
124 | return NOPAGE_SIGBUS; | 198 | static int ehca_mmap_qp(struct vm_area_struct *vma, struct ehca_qp *qp, |
199 | u32 rsrc_type) | ||
200 | { | ||
201 | int ret; | ||
202 | |||
203 | switch (rsrc_type) { | ||
204 | case 1: /* galpa fw handle */ | ||
205 | ehca_dbg(qp->ib_qp.device, "qp_num=%x fw", qp->ib_qp.qp_num); | ||
206 | ret = ehca_mmap_fw(vma, &qp->galpas, &qp->mm_count_galpa); | ||
207 | if (unlikely(ret)) { | ||
208 | ehca_err(qp->ib_qp.device, | ||
209 | "remap_pfn_range() failed ret=%x qp_num=%x", | ||
210 | ret, qp->ib_qp.qp_num); | ||
211 | return -ENOMEM; | ||
125 | } | 212 | } |
213 | break; | ||
126 | 214 | ||
127 | pd = container_of(qp->ib_qp.pd, struct ehca_pd, ib_pd); | 215 | case 2: /* qp rqueue_addr */ |
128 | if (pd->ownpid != cur_pid) { | 216 | ehca_dbg(qp->ib_qp.device, "qp_num=%x rqueue", |
217 | qp->ib_qp.qp_num); | ||
218 | ret = ehca_mmap_queue(vma, &qp->ipz_rqueue, &qp->mm_count_rqueue); | ||
219 | if (unlikely(ret)) { | ||
129 | ehca_err(qp->ib_qp.device, | 220 | ehca_err(qp->ib_qp.device, |
130 | "Invalid caller pid=%x ownpid=%x", | 221 | "ehca_mmap_queue(rq) failed rc=%x qp_num=%x", |
131 | cur_pid, pd->ownpid); | 222 | ret, qp->ib_qp.qp_num); |
132 | return NOPAGE_SIGBUS; | 223 | return ret; |
133 | } | 224 | } |
225 | break; | ||
134 | 226 | ||
135 | if (rsrc_type == 2) { /* rqueue */ | 227 | case 3: /* qp squeue_addr */ |
136 | ehca_dbg(qp->ib_qp.device, "qp=%p qp rqueuearea", qp); | 228 | ehca_dbg(qp->ib_qp.device, "qp_num=%x squeue", |
137 | offset = address - vma->vm_start; | 229 | qp->ib_qp.qp_num); |
138 | vaddr = ipz_qeit_calc(&qp->ipz_rqueue, offset); | 230 | ret = ehca_mmap_queue(vma, &qp->ipz_squeue, &qp->mm_count_squeue); |
139 | ehca_dbg(qp->ib_qp.device, "offset=%lx vaddr=%p", | 231 | if (unlikely(ret)) { |
140 | offset, vaddr); | 232 | ehca_err(qp->ib_qp.device, |
141 | mypage = virt_to_page(vaddr); | 233 | "ehca_mmap_queue(sq) failed rc=%x qp_num=%x", |
142 | } else if (rsrc_type == 3) { /* squeue */ | 234 | ret, qp->ib_qp.qp_num); |
143 | ehca_dbg(qp->ib_qp.device, "qp=%p qp squeuearea", qp); | 235 | return ret; |
144 | offset = address - vma->vm_start; | ||
145 | vaddr = ipz_qeit_calc(&qp->ipz_squeue, offset); | ||
146 | ehca_dbg(qp->ib_qp.device, "offset=%lx vaddr=%p", | ||
147 | offset, vaddr); | ||
148 | mypage = virt_to_page(vaddr); | ||
149 | } | 236 | } |
150 | break; | 237 | break; |
151 | 238 | ||
152 | default: | 239 | default: |
153 | ehca_gen_err("bad queue type %x", q_type); | 240 | ehca_err(qp->ib_qp.device, "bad resource type=%x qp=num=%x", |
154 | return NOPAGE_SIGBUS; | 241 | rsrc_type, qp->ib_qp.qp_num); |
155 | } | 242 | return -EINVAL; |
156 | |||
157 | if (!mypage) { | ||
158 | ehca_gen_err("Invalid page adr==NULL ret=NOPAGE_SIGBUS"); | ||
159 | return NOPAGE_SIGBUS; | ||
160 | } | 243 | } |
161 | get_page(mypage); | ||
162 | 244 | ||
163 | return mypage; | 245 | return 0; |
164 | } | 246 | } |
165 | 247 | ||
166 | static struct vm_operations_struct ehcau_vm_ops = { | ||
167 | .nopage = ehca_nopage, | ||
168 | }; | ||
169 | |||
170 | int ehca_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) | 248 | int ehca_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) |
171 | { | 249 | { |
172 | u64 fileoffset = vma->vm_pgoff << PAGE_SHIFT; | 250 | u64 fileoffset = vma->vm_pgoff << PAGE_SHIFT; |
@@ -175,7 +253,6 @@ int ehca_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) | |||
175 | u32 rsrc_type = (fileoffset >> 24) & 0xF; /* sq,rq,cmnd_window */ | 253 | u32 rsrc_type = (fileoffset >> 24) & 0xF; /* sq,rq,cmnd_window */ |
176 | u32 cur_pid = current->tgid; | 254 | u32 cur_pid = current->tgid; |
177 | u32 ret; | 255 | u32 ret; |
178 | u64 vsize, physical; | ||
179 | unsigned long flags; | 256 | unsigned long flags; |
180 | struct ehca_cq *cq; | 257 | struct ehca_cq *cq; |
181 | struct ehca_qp *qp; | 258 | struct ehca_qp *qp; |
@@ -201,44 +278,12 @@ int ehca_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) | |||
201 | if (!cq->ib_cq.uobject || cq->ib_cq.uobject->context != context) | 278 | if (!cq->ib_cq.uobject || cq->ib_cq.uobject->context != context) |
202 | return -EINVAL; | 279 | return -EINVAL; |
203 | 280 | ||
204 | switch (rsrc_type) { | 281 | ret = ehca_mmap_cq(vma, cq, rsrc_type); |
205 | case 1: /* galpa fw handle */ | 282 | if (unlikely(ret)) { |
206 | ehca_dbg(cq->ib_cq.device, "cq=%p cq triggerarea", cq); | 283 | ehca_err(cq->ib_cq.device, |
207 | vma->vm_flags |= VM_RESERVED; | 284 | "ehca_mmap_cq() failed rc=%x cq_num=%x", |
208 | vsize = vma->vm_end - vma->vm_start; | 285 | ret, cq->cq_number); |
209 | if (vsize != EHCA_PAGESIZE) { | 286 | return ret; |
210 | ehca_err(cq->ib_cq.device, "invalid vsize=%lx", | ||
211 | vma->vm_end - vma->vm_start); | ||
212 | return -EINVAL; | ||
213 | } | ||
214 | |||
215 | physical = cq->galpas.user.fw_handle; | ||
216 | vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); | ||
217 | vma->vm_flags |= VM_IO | VM_RESERVED; | ||
218 | |||
219 | ehca_dbg(cq->ib_cq.device, | ||
220 | "vsize=%lx physical=%lx", vsize, physical); | ||
221 | ret = remap_pfn_range(vma, vma->vm_start, | ||
222 | physical >> PAGE_SHIFT, vsize, | ||
223 | vma->vm_page_prot); | ||
224 | if (ret) { | ||
225 | ehca_err(cq->ib_cq.device, | ||
226 | "remap_pfn_range() failed ret=%x", | ||
227 | ret); | ||
228 | return -ENOMEM; | ||
229 | } | ||
230 | break; | ||
231 | |||
232 | case 2: /* cq queue_addr */ | ||
233 | ehca_dbg(cq->ib_cq.device, "cq=%p cq q_addr", cq); | ||
234 | vma->vm_flags |= VM_RESERVED; | ||
235 | vma->vm_ops = &ehcau_vm_ops; | ||
236 | break; | ||
237 | |||
238 | default: | ||
239 | ehca_err(cq->ib_cq.device, "bad resource type %x", | ||
240 | rsrc_type); | ||
241 | return -EINVAL; | ||
242 | } | 287 | } |
243 | break; | 288 | break; |
244 | 289 | ||
@@ -262,50 +307,12 @@ int ehca_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) | |||
262 | if (!qp->ib_qp.uobject || qp->ib_qp.uobject->context != context) | 307 | if (!qp->ib_qp.uobject || qp->ib_qp.uobject->context != context) |
263 | return -EINVAL; | 308 | return -EINVAL; |
264 | 309 | ||
265 | switch (rsrc_type) { | 310 | ret = ehca_mmap_qp(vma, qp, rsrc_type); |
266 | case 1: /* galpa fw handle */ | 311 | if (unlikely(ret)) { |
267 | ehca_dbg(qp->ib_qp.device, "qp=%p qp triggerarea", qp); | 312 | ehca_err(qp->ib_qp.device, |
268 | vma->vm_flags |= VM_RESERVED; | 313 | "ehca_mmap_qp() failed rc=%x qp_num=%x", |
269 | vsize = vma->vm_end - vma->vm_start; | 314 | ret, qp->ib_qp.qp_num); |
270 | if (vsize != EHCA_PAGESIZE) { | 315 | return ret; |
271 | ehca_err(qp->ib_qp.device, "invalid vsize=%lx", | ||
272 | vma->vm_end - vma->vm_start); | ||
273 | return -EINVAL; | ||
274 | } | ||
275 | |||
276 | physical = qp->galpas.user.fw_handle; | ||
277 | vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); | ||
278 | vma->vm_flags |= VM_IO | VM_RESERVED; | ||
279 | |||
280 | ehca_dbg(qp->ib_qp.device, "vsize=%lx physical=%lx", | ||
281 | vsize, physical); | ||
282 | ret = remap_pfn_range(vma, vma->vm_start, | ||
283 | physical >> PAGE_SHIFT, vsize, | ||
284 | vma->vm_page_prot); | ||
285 | if (ret) { | ||
286 | ehca_err(qp->ib_qp.device, | ||
287 | "remap_pfn_range() failed ret=%x", | ||
288 | ret); | ||
289 | return -ENOMEM; | ||
290 | } | ||
291 | break; | ||
292 | |||
293 | case 2: /* qp rqueue_addr */ | ||
294 | ehca_dbg(qp->ib_qp.device, "qp=%p qp rqueue_addr", qp); | ||
295 | vma->vm_flags |= VM_RESERVED; | ||
296 | vma->vm_ops = &ehcau_vm_ops; | ||
297 | break; | ||
298 | |||
299 | case 3: /* qp squeue_addr */ | ||
300 | ehca_dbg(qp->ib_qp.device, "qp=%p qp squeue_addr", qp); | ||
301 | vma->vm_flags |= VM_RESERVED; | ||
302 | vma->vm_ops = &ehcau_vm_ops; | ||
303 | break; | ||
304 | |||
305 | default: | ||
306 | ehca_err(qp->ib_qp.device, "bad resource type %x", | ||
307 | rsrc_type); | ||
308 | return -EINVAL; | ||
309 | } | 316 | } |
310 | break; | 317 | break; |
311 | 318 | ||
@@ -316,77 +323,3 @@ int ehca_mmap(struct ib_ucontext *context, struct vm_area_struct *vma) | |||
316 | 323 | ||
317 | return 0; | 324 | return 0; |
318 | } | 325 | } |
319 | |||
320 | int ehca_mmap_nopage(u64 foffset, u64 length, void **mapped, | ||
321 | struct vm_area_struct **vma) | ||
322 | { | ||
323 | down_write(¤t->mm->mmap_sem); | ||
324 | *mapped = (void*)do_mmap(NULL,0, length, PROT_WRITE, | ||
325 | MAP_SHARED | MAP_ANONYMOUS, | ||
326 | foffset); | ||
327 | up_write(¤t->mm->mmap_sem); | ||
328 | if (!(*mapped)) { | ||
329 | ehca_gen_err("couldn't mmap foffset=%lx length=%lx", | ||
330 | foffset, length); | ||
331 | return -EINVAL; | ||
332 | } | ||
333 | |||
334 | *vma = find_vma(current->mm, (u64)*mapped); | ||
335 | if (!(*vma)) { | ||
336 | down_write(¤t->mm->mmap_sem); | ||
337 | do_munmap(current->mm, 0, length); | ||
338 | up_write(¤t->mm->mmap_sem); | ||
339 | ehca_gen_err("couldn't find vma queue=%p", *mapped); | ||
340 | return -EINVAL; | ||
341 | } | ||
342 | (*vma)->vm_flags |= VM_RESERVED; | ||
343 | (*vma)->vm_ops = &ehcau_vm_ops; | ||
344 | |||
345 | return 0; | ||
346 | } | ||
347 | |||
348 | int ehca_mmap_register(u64 physical, void **mapped, | ||
349 | struct vm_area_struct **vma) | ||
350 | { | ||
351 | int ret; | ||
352 | unsigned long vsize; | ||
353 | /* ehca hw supports only 4k page */ | ||
354 | ret = ehca_mmap_nopage(0, EHCA_PAGESIZE, mapped, vma); | ||
355 | if (ret) { | ||
356 | ehca_gen_err("could'nt mmap physical=%lx", physical); | ||
357 | return ret; | ||
358 | } | ||
359 | |||
360 | (*vma)->vm_flags |= VM_RESERVED; | ||
361 | vsize = (*vma)->vm_end - (*vma)->vm_start; | ||
362 | if (vsize != EHCA_PAGESIZE) { | ||
363 | ehca_gen_err("invalid vsize=%lx", | ||
364 | (*vma)->vm_end - (*vma)->vm_start); | ||
365 | return -EINVAL; | ||
366 | } | ||
367 | |||
368 | (*vma)->vm_page_prot = pgprot_noncached((*vma)->vm_page_prot); | ||
369 | (*vma)->vm_flags |= VM_IO | VM_RESERVED; | ||
370 | |||
371 | ret = remap_pfn_range((*vma), (*vma)->vm_start, | ||
372 | physical >> PAGE_SHIFT, vsize, | ||
373 | (*vma)->vm_page_prot); | ||
374 | if (ret) { | ||
375 | ehca_gen_err("remap_pfn_range() failed ret=%x", ret); | ||
376 | return -ENOMEM; | ||
377 | } | ||
378 | |||
379 | return 0; | ||
380 | |||
381 | } | ||
382 | |||
383 | int ehca_munmap(unsigned long addr, size_t len) { | ||
384 | int ret = 0; | ||
385 | struct mm_struct *mm = current->mm; | ||
386 | if (mm) { | ||
387 | down_write(&mm->mmap_sem); | ||
388 | ret = do_munmap(mm, addr, len); | ||
389 | up_write(&mm->mmap_sem); | ||
390 | } | ||
391 | return ret; | ||
392 | } | ||
diff --git a/drivers/infiniband/hw/ipath/ipath_qp.c b/drivers/infiniband/hw/ipath/ipath_qp.c index 46c1c89bf6ae..64f07b19349f 100644 --- a/drivers/infiniband/hw/ipath/ipath_qp.c +++ b/drivers/infiniband/hw/ipath/ipath_qp.c | |||
@@ -379,7 +379,7 @@ void ipath_error_qp(struct ipath_qp *qp, enum ib_wc_status err) | |||
379 | wc.vendor_err = 0; | 379 | wc.vendor_err = 0; |
380 | wc.byte_len = 0; | 380 | wc.byte_len = 0; |
381 | wc.imm_data = 0; | 381 | wc.imm_data = 0; |
382 | wc.qp_num = qp->ibqp.qp_num; | 382 | wc.qp = &qp->ibqp; |
383 | wc.src_qp = 0; | 383 | wc.src_qp = 0; |
384 | wc.wc_flags = 0; | 384 | wc.wc_flags = 0; |
385 | wc.pkey_index = 0; | 385 | wc.pkey_index = 0; |
diff --git a/drivers/infiniband/hw/ipath/ipath_rc.c b/drivers/infiniband/hw/ipath/ipath_rc.c index ce6038743c5c..5ff20cb04494 100644 --- a/drivers/infiniband/hw/ipath/ipath_rc.c +++ b/drivers/infiniband/hw/ipath/ipath_rc.c | |||
@@ -702,7 +702,7 @@ void ipath_restart_rc(struct ipath_qp *qp, u32 psn, struct ib_wc *wc) | |||
702 | wc->opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; | 702 | wc->opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; |
703 | wc->vendor_err = 0; | 703 | wc->vendor_err = 0; |
704 | wc->byte_len = 0; | 704 | wc->byte_len = 0; |
705 | wc->qp_num = qp->ibqp.qp_num; | 705 | wc->qp = &qp->ibqp; |
706 | wc->src_qp = qp->remote_qpn; | 706 | wc->src_qp = qp->remote_qpn; |
707 | wc->pkey_index = 0; | 707 | wc->pkey_index = 0; |
708 | wc->slid = qp->remote_ah_attr.dlid; | 708 | wc->slid = qp->remote_ah_attr.dlid; |
@@ -836,7 +836,7 @@ static int do_rc_ack(struct ipath_qp *qp, u32 aeth, u32 psn, int opcode) | |||
836 | wc.opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; | 836 | wc.opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; |
837 | wc.vendor_err = 0; | 837 | wc.vendor_err = 0; |
838 | wc.byte_len = wqe->length; | 838 | wc.byte_len = wqe->length; |
839 | wc.qp_num = qp->ibqp.qp_num; | 839 | wc.qp = &qp->ibqp; |
840 | wc.src_qp = qp->remote_qpn; | 840 | wc.src_qp = qp->remote_qpn; |
841 | wc.pkey_index = 0; | 841 | wc.pkey_index = 0; |
842 | wc.slid = qp->remote_ah_attr.dlid; | 842 | wc.slid = qp->remote_ah_attr.dlid; |
@@ -951,7 +951,7 @@ static int do_rc_ack(struct ipath_qp *qp, u32 aeth, u32 psn, int opcode) | |||
951 | wc.opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; | 951 | wc.opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; |
952 | wc.vendor_err = 0; | 952 | wc.vendor_err = 0; |
953 | wc.byte_len = 0; | 953 | wc.byte_len = 0; |
954 | wc.qp_num = qp->ibqp.qp_num; | 954 | wc.qp = &qp->ibqp; |
955 | wc.src_qp = qp->remote_qpn; | 955 | wc.src_qp = qp->remote_qpn; |
956 | wc.pkey_index = 0; | 956 | wc.pkey_index = 0; |
957 | wc.slid = qp->remote_ah_attr.dlid; | 957 | wc.slid = qp->remote_ah_attr.dlid; |
@@ -1511,7 +1511,7 @@ void ipath_rc_rcv(struct ipath_ibdev *dev, struct ipath_ib_header *hdr, | |||
1511 | wc.status = IB_WC_SUCCESS; | 1511 | wc.status = IB_WC_SUCCESS; |
1512 | wc.opcode = IB_WC_RECV; | 1512 | wc.opcode = IB_WC_RECV; |
1513 | wc.vendor_err = 0; | 1513 | wc.vendor_err = 0; |
1514 | wc.qp_num = qp->ibqp.qp_num; | 1514 | wc.qp = &qp->ibqp; |
1515 | wc.src_qp = qp->remote_qpn; | 1515 | wc.src_qp = qp->remote_qpn; |
1516 | wc.pkey_index = 0; | 1516 | wc.pkey_index = 0; |
1517 | wc.slid = qp->remote_ah_attr.dlid; | 1517 | wc.slid = qp->remote_ah_attr.dlid; |
diff --git a/drivers/infiniband/hw/ipath/ipath_ruc.c b/drivers/infiniband/hw/ipath/ipath_ruc.c index f7530512045d..e86cb171872e 100644 --- a/drivers/infiniband/hw/ipath/ipath_ruc.c +++ b/drivers/infiniband/hw/ipath/ipath_ruc.c | |||
@@ -137,7 +137,7 @@ bad_lkey: | |||
137 | wc.vendor_err = 0; | 137 | wc.vendor_err = 0; |
138 | wc.byte_len = 0; | 138 | wc.byte_len = 0; |
139 | wc.imm_data = 0; | 139 | wc.imm_data = 0; |
140 | wc.qp_num = qp->ibqp.qp_num; | 140 | wc.qp = &qp->ibqp; |
141 | wc.src_qp = 0; | 141 | wc.src_qp = 0; |
142 | wc.wc_flags = 0; | 142 | wc.wc_flags = 0; |
143 | wc.pkey_index = 0; | 143 | wc.pkey_index = 0; |
@@ -336,7 +336,7 @@ again: | |||
336 | wc.opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; | 336 | wc.opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; |
337 | wc.vendor_err = 0; | 337 | wc.vendor_err = 0; |
338 | wc.byte_len = 0; | 338 | wc.byte_len = 0; |
339 | wc.qp_num = sqp->ibqp.qp_num; | 339 | wc.qp = &sqp->ibqp; |
340 | wc.src_qp = sqp->remote_qpn; | 340 | wc.src_qp = sqp->remote_qpn; |
341 | wc.pkey_index = 0; | 341 | wc.pkey_index = 0; |
342 | wc.slid = sqp->remote_ah_attr.dlid; | 342 | wc.slid = sqp->remote_ah_attr.dlid; |
@@ -426,7 +426,7 @@ again: | |||
426 | wc.status = IB_WC_SUCCESS; | 426 | wc.status = IB_WC_SUCCESS; |
427 | wc.vendor_err = 0; | 427 | wc.vendor_err = 0; |
428 | wc.byte_len = wqe->length; | 428 | wc.byte_len = wqe->length; |
429 | wc.qp_num = qp->ibqp.qp_num; | 429 | wc.qp = &qp->ibqp; |
430 | wc.src_qp = qp->remote_qpn; | 430 | wc.src_qp = qp->remote_qpn; |
431 | /* XXX do we know which pkey matched? Only needed for GSI. */ | 431 | /* XXX do we know which pkey matched? Only needed for GSI. */ |
432 | wc.pkey_index = 0; | 432 | wc.pkey_index = 0; |
@@ -447,7 +447,7 @@ send_comp: | |||
447 | wc.opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; | 447 | wc.opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; |
448 | wc.vendor_err = 0; | 448 | wc.vendor_err = 0; |
449 | wc.byte_len = wqe->length; | 449 | wc.byte_len = wqe->length; |
450 | wc.qp_num = sqp->ibqp.qp_num; | 450 | wc.qp = &sqp->ibqp; |
451 | wc.src_qp = 0; | 451 | wc.src_qp = 0; |
452 | wc.pkey_index = 0; | 452 | wc.pkey_index = 0; |
453 | wc.slid = 0; | 453 | wc.slid = 0; |
diff --git a/drivers/infiniband/hw/ipath/ipath_uc.c b/drivers/infiniband/hw/ipath/ipath_uc.c index e636cfd67a82..325d6634ff53 100644 --- a/drivers/infiniband/hw/ipath/ipath_uc.c +++ b/drivers/infiniband/hw/ipath/ipath_uc.c | |||
@@ -49,7 +49,7 @@ static void complete_last_send(struct ipath_qp *qp, struct ipath_swqe *wqe, | |||
49 | wc->opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; | 49 | wc->opcode = ib_ipath_wc_opcode[wqe->wr.opcode]; |
50 | wc->vendor_err = 0; | 50 | wc->vendor_err = 0; |
51 | wc->byte_len = wqe->length; | 51 | wc->byte_len = wqe->length; |
52 | wc->qp_num = qp->ibqp.qp_num; | 52 | wc->qp = &qp->ibqp; |
53 | wc->src_qp = qp->remote_qpn; | 53 | wc->src_qp = qp->remote_qpn; |
54 | wc->pkey_index = 0; | 54 | wc->pkey_index = 0; |
55 | wc->slid = qp->remote_ah_attr.dlid; | 55 | wc->slid = qp->remote_ah_attr.dlid; |
@@ -411,7 +411,7 @@ void ipath_uc_rcv(struct ipath_ibdev *dev, struct ipath_ib_header *hdr, | |||
411 | wc.status = IB_WC_SUCCESS; | 411 | wc.status = IB_WC_SUCCESS; |
412 | wc.opcode = IB_WC_RECV; | 412 | wc.opcode = IB_WC_RECV; |
413 | wc.vendor_err = 0; | 413 | wc.vendor_err = 0; |
414 | wc.qp_num = qp->ibqp.qp_num; | 414 | wc.qp = &qp->ibqp; |
415 | wc.src_qp = qp->remote_qpn; | 415 | wc.src_qp = qp->remote_qpn; |
416 | wc.pkey_index = 0; | 416 | wc.pkey_index = 0; |
417 | wc.slid = qp->remote_ah_attr.dlid; | 417 | wc.slid = qp->remote_ah_attr.dlid; |
diff --git a/drivers/infiniband/hw/ipath/ipath_ud.c b/drivers/infiniband/hw/ipath/ipath_ud.c index 49f1102af8b3..9a3e54664ee4 100644 --- a/drivers/infiniband/hw/ipath/ipath_ud.c +++ b/drivers/infiniband/hw/ipath/ipath_ud.c | |||
@@ -66,7 +66,7 @@ bad_lkey: | |||
66 | wc.vendor_err = 0; | 66 | wc.vendor_err = 0; |
67 | wc.byte_len = 0; | 67 | wc.byte_len = 0; |
68 | wc.imm_data = 0; | 68 | wc.imm_data = 0; |
69 | wc.qp_num = qp->ibqp.qp_num; | 69 | wc.qp = &qp->ibqp; |
70 | wc.src_qp = 0; | 70 | wc.src_qp = 0; |
71 | wc.wc_flags = 0; | 71 | wc.wc_flags = 0; |
72 | wc.pkey_index = 0; | 72 | wc.pkey_index = 0; |
@@ -255,7 +255,7 @@ static void ipath_ud_loopback(struct ipath_qp *sqp, | |||
255 | wc->status = IB_WC_SUCCESS; | 255 | wc->status = IB_WC_SUCCESS; |
256 | wc->opcode = IB_WC_RECV; | 256 | wc->opcode = IB_WC_RECV; |
257 | wc->vendor_err = 0; | 257 | wc->vendor_err = 0; |
258 | wc->qp_num = qp->ibqp.qp_num; | 258 | wc->qp = &qp->ibqp; |
259 | wc->src_qp = sqp->ibqp.qp_num; | 259 | wc->src_qp = sqp->ibqp.qp_num; |
260 | /* XXX do we know which pkey matched? Only needed for GSI. */ | 260 | /* XXX do we know which pkey matched? Only needed for GSI. */ |
261 | wc->pkey_index = 0; | 261 | wc->pkey_index = 0; |
@@ -474,7 +474,7 @@ done: | |||
474 | wc.vendor_err = 0; | 474 | wc.vendor_err = 0; |
475 | wc.opcode = IB_WC_SEND; | 475 | wc.opcode = IB_WC_SEND; |
476 | wc.byte_len = len; | 476 | wc.byte_len = len; |
477 | wc.qp_num = qp->ibqp.qp_num; | 477 | wc.qp = &qp->ibqp; |
478 | wc.src_qp = 0; | 478 | wc.src_qp = 0; |
479 | wc.wc_flags = 0; | 479 | wc.wc_flags = 0; |
480 | /* XXX initialize other fields? */ | 480 | /* XXX initialize other fields? */ |
@@ -651,7 +651,7 @@ void ipath_ud_rcv(struct ipath_ibdev *dev, struct ipath_ib_header *hdr, | |||
651 | wc.status = IB_WC_SUCCESS; | 651 | wc.status = IB_WC_SUCCESS; |
652 | wc.opcode = IB_WC_RECV; | 652 | wc.opcode = IB_WC_RECV; |
653 | wc.vendor_err = 0; | 653 | wc.vendor_err = 0; |
654 | wc.qp_num = qp->ibqp.qp_num; | 654 | wc.qp = &qp->ibqp; |
655 | wc.src_qp = src_qp; | 655 | wc.src_qp = src_qp; |
656 | /* XXX do we know which pkey matched? Only needed for GSI. */ | 656 | /* XXX do we know which pkey matched? Only needed for GSI. */ |
657 | wc.pkey_index = 0; | 657 | wc.pkey_index = 0; |
diff --git a/drivers/infiniband/hw/mthca/mthca_cmd.c b/drivers/infiniband/hw/mthca/mthca_cmd.c index 768df7265b81..968d1519761c 100644 --- a/drivers/infiniband/hw/mthca/mthca_cmd.c +++ b/drivers/infiniband/hw/mthca/mthca_cmd.c | |||
@@ -1854,7 +1854,7 @@ int mthca_MAD_IFC(struct mthca_dev *dev, int ignore_mkey, int ignore_bkey, | |||
1854 | 1854 | ||
1855 | memset(inbox + 256, 0, 256); | 1855 | memset(inbox + 256, 0, 256); |
1856 | 1856 | ||
1857 | MTHCA_PUT(inbox, in_wc->qp_num, MAD_IFC_MY_QPN_OFFSET); | 1857 | MTHCA_PUT(inbox, in_wc->qp->qp_num, MAD_IFC_MY_QPN_OFFSET); |
1858 | MTHCA_PUT(inbox, in_wc->src_qp, MAD_IFC_RQPN_OFFSET); | 1858 | MTHCA_PUT(inbox, in_wc->src_qp, MAD_IFC_RQPN_OFFSET); |
1859 | 1859 | ||
1860 | val = in_wc->sl << 4; | 1860 | val = in_wc->sl << 4; |
diff --git a/drivers/infiniband/hw/mthca/mthca_cq.c b/drivers/infiniband/hw/mthca/mthca_cq.c index 1159c8a0f2c5..efd79ef109a6 100644 --- a/drivers/infiniband/hw/mthca/mthca_cq.c +++ b/drivers/infiniband/hw/mthca/mthca_cq.c | |||
@@ -534,7 +534,7 @@ static inline int mthca_poll_one(struct mthca_dev *dev, | |||
534 | } | 534 | } |
535 | } | 535 | } |
536 | 536 | ||
537 | entry->qp_num = (*cur_qp)->qpn; | 537 | entry->qp = &(*cur_qp)->ibqp; |
538 | 538 | ||
539 | if (is_send) { | 539 | if (is_send) { |
540 | wq = &(*cur_qp)->sq; | 540 | wq = &(*cur_qp)->sq; |
diff --git a/drivers/infiniband/ulp/srp/ib_srp.c b/drivers/infiniband/ulp/srp/ib_srp.c index 72611fd15103..5e8ac577f0ad 100644 --- a/drivers/infiniband/ulp/srp/ib_srp.c +++ b/drivers/infiniband/ulp/srp/ib_srp.c | |||
@@ -548,6 +548,7 @@ static int srp_reconnect_target(struct srp_target_port *target) | |||
548 | target->tx_head = 0; | 548 | target->tx_head = 0; |
549 | target->tx_tail = 0; | 549 | target->tx_tail = 0; |
550 | 550 | ||
551 | target->qp_in_error = 0; | ||
551 | ret = srp_connect_target(target); | 552 | ret = srp_connect_target(target); |
552 | if (ret) | 553 | if (ret) |
553 | goto err; | 554 | goto err; |
@@ -878,6 +879,7 @@ static void srp_completion(struct ib_cq *cq, void *target_ptr) | |||
878 | printk(KERN_ERR PFX "failed %s status %d\n", | 879 | printk(KERN_ERR PFX "failed %s status %d\n", |
879 | wc.wr_id & SRP_OP_RECV ? "receive" : "send", | 880 | wc.wr_id & SRP_OP_RECV ? "receive" : "send", |
880 | wc.status); | 881 | wc.status); |
882 | target->qp_in_error = 1; | ||
881 | break; | 883 | break; |
882 | } | 884 | } |
883 | 885 | ||
@@ -1337,6 +1339,8 @@ static int srp_abort(struct scsi_cmnd *scmnd) | |||
1337 | 1339 | ||
1338 | printk(KERN_ERR "SRP abort called\n"); | 1340 | printk(KERN_ERR "SRP abort called\n"); |
1339 | 1341 | ||
1342 | if (target->qp_in_error) | ||
1343 | return FAILED; | ||
1340 | if (srp_find_req(target, scmnd, &req)) | 1344 | if (srp_find_req(target, scmnd, &req)) |
1341 | return FAILED; | 1345 | return FAILED; |
1342 | if (srp_send_tsk_mgmt(target, req, SRP_TSK_ABORT_TASK)) | 1346 | if (srp_send_tsk_mgmt(target, req, SRP_TSK_ABORT_TASK)) |
@@ -1365,6 +1369,8 @@ static int srp_reset_device(struct scsi_cmnd *scmnd) | |||
1365 | 1369 | ||
1366 | printk(KERN_ERR "SRP reset_device called\n"); | 1370 | printk(KERN_ERR "SRP reset_device called\n"); |
1367 | 1371 | ||
1372 | if (target->qp_in_error) | ||
1373 | return FAILED; | ||
1368 | if (srp_find_req(target, scmnd, &req)) | 1374 | if (srp_find_req(target, scmnd, &req)) |
1369 | return FAILED; | 1375 | return FAILED; |
1370 | if (srp_send_tsk_mgmt(target, req, SRP_TSK_LUN_RESET)) | 1376 | if (srp_send_tsk_mgmt(target, req, SRP_TSK_LUN_RESET)) |
@@ -1801,6 +1807,7 @@ static ssize_t srp_create_target(struct class_device *class_dev, | |||
1801 | goto err_free; | 1807 | goto err_free; |
1802 | } | 1808 | } |
1803 | 1809 | ||
1810 | target->qp_in_error = 0; | ||
1804 | ret = srp_connect_target(target); | 1811 | ret = srp_connect_target(target); |
1805 | if (ret) { | 1812 | if (ret) { |
1806 | printk(KERN_ERR PFX "Connection failed\n"); | 1813 | printk(KERN_ERR PFX "Connection failed\n"); |
diff --git a/drivers/infiniband/ulp/srp/ib_srp.h b/drivers/infiniband/ulp/srp/ib_srp.h index c21772317b86..2f3319c719a5 100644 --- a/drivers/infiniband/ulp/srp/ib_srp.h +++ b/drivers/infiniband/ulp/srp/ib_srp.h | |||
@@ -158,6 +158,7 @@ struct srp_target_port { | |||
158 | struct completion done; | 158 | struct completion done; |
159 | int status; | 159 | int status; |
160 | enum srp_target_state state; | 160 | enum srp_target_state state; |
161 | int qp_in_error; | ||
161 | }; | 162 | }; |
162 | 163 | ||
163 | struct srp_iu { | 164 | struct srp_iu { |