diff options
author | Linus Torvalds <torvalds@woody.linux-foundation.org> | 2007-10-23 12:56:11 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@woody.linux-foundation.org> | 2007-10-23 12:56:11 -0400 |
commit | 0b776eb5426752d4e53354ac89e3710d857e09a7 (patch) | |
tree | 1eebeeaabab90de5834b32e72d2e259dc8a4a635 /drivers/infiniband/hw/ehca | |
parent | 0d6810091cdbd05efeb31654c6a41a6cbdfdd2c8 (diff) | |
parent | 77109cc2823f025ccd66ebd9b88fbab90437b2d8 (diff) |
Merge branch 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/roland/infiniband
* 'for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/roland/infiniband:
mlx4_core: Increase command timeout for INIT_HCA to 10 seconds
IPoIB/cm: Use common CQ for CM send completions
IB/uverbs: Fix checking of userspace object ownership
IB/mlx4: Sanity check userspace send queue sizes
IPoIB: Rewrite "if (!likely(...))" as "if (unlikely(!(...)))"
IB/ehca: Enable large page MRs by default
IB/ehca: Change meaning of hca_cap_mr_pgsize
IB/ehca: Fix ehca_encode_hwpage_size() and alloc_fmr()
IB/ehca: Fix masking error in {,re}reg_phys_mr()
IB/ehca: Supply QP token for SRQ base QPs
IPoIB: Use round_jiffies() for ah_reap_task
RDMA/cma: Fix deadlock destroying listen requests
RDMA/cma: Add locking around QP accesses
IB/mthca: Avoid alignment traps when writing doorbells
mlx4_core: Kill mlx4_write64_raw()
Diffstat (limited to 'drivers/infiniband/hw/ehca')
-rw-r--r-- | drivers/infiniband/hw/ehca/ehca_classes.h | 1 | ||||
-rw-r--r-- | drivers/infiniband/hw/ehca/ehca_hca.c | 1 | ||||
-rw-r--r-- | drivers/infiniband/hw/ehca/ehca_main.c | 20 | ||||
-rw-r--r-- | drivers/infiniband/hw/ehca/ehca_mrmw.c | 57 | ||||
-rw-r--r-- | drivers/infiniband/hw/ehca/ehca_qp.c | 4 |
5 files changed, 46 insertions, 37 deletions
diff --git a/drivers/infiniband/hw/ehca/ehca_classes.h b/drivers/infiniband/hw/ehca/ehca_classes.h index 3f2d68cff764..2d660ae189e5 100644 --- a/drivers/infiniband/hw/ehca/ehca_classes.h +++ b/drivers/infiniband/hw/ehca/ehca_classes.h | |||
@@ -323,7 +323,6 @@ extern int ehca_static_rate; | |||
323 | extern int ehca_port_act_time; | 323 | extern int ehca_port_act_time; |
324 | extern int ehca_use_hp_mr; | 324 | extern int ehca_use_hp_mr; |
325 | extern int ehca_scaling_code; | 325 | extern int ehca_scaling_code; |
326 | extern int ehca_mr_largepage; | ||
327 | 326 | ||
328 | struct ipzu_queue_resp { | 327 | struct ipzu_queue_resp { |
329 | u32 qe_size; /* queue entry size */ | 328 | u32 qe_size; /* queue entry size */ |
diff --git a/drivers/infiniband/hw/ehca/ehca_hca.c b/drivers/infiniband/hw/ehca/ehca_hca.c index 4aa3ffa6a19f..15806d140461 100644 --- a/drivers/infiniband/hw/ehca/ehca_hca.c +++ b/drivers/infiniband/hw/ehca/ehca_hca.c | |||
@@ -77,6 +77,7 @@ int ehca_query_device(struct ib_device *ibdev, struct ib_device_attr *props) | |||
77 | } | 77 | } |
78 | 78 | ||
79 | memset(props, 0, sizeof(struct ib_device_attr)); | 79 | memset(props, 0, sizeof(struct ib_device_attr)); |
80 | props->page_size_cap = shca->hca_cap_mr_pgsize; | ||
80 | props->fw_ver = rblock->hw_ver; | 81 | props->fw_ver = rblock->hw_ver; |
81 | props->max_mr_size = rblock->max_mr_size; | 82 | props->max_mr_size = rblock->max_mr_size; |
82 | props->vendor_id = rblock->vendor_id >> 8; | 83 | props->vendor_id = rblock->vendor_id >> 8; |
diff --git a/drivers/infiniband/hw/ehca/ehca_main.c b/drivers/infiniband/hw/ehca/ehca_main.c index 7a7dab890f6d..c6cd38c5321f 100644 --- a/drivers/infiniband/hw/ehca/ehca_main.c +++ b/drivers/infiniband/hw/ehca/ehca_main.c | |||
@@ -65,7 +65,7 @@ int ehca_port_act_time = 30; | |||
65 | int ehca_poll_all_eqs = 1; | 65 | int ehca_poll_all_eqs = 1; |
66 | int ehca_static_rate = -1; | 66 | int ehca_static_rate = -1; |
67 | int ehca_scaling_code = 0; | 67 | int ehca_scaling_code = 0; |
68 | int ehca_mr_largepage = 0; | 68 | int ehca_mr_largepage = 1; |
69 | 69 | ||
70 | module_param_named(open_aqp1, ehca_open_aqp1, int, S_IRUGO); | 70 | module_param_named(open_aqp1, ehca_open_aqp1, int, S_IRUGO); |
71 | module_param_named(debug_level, ehca_debug_level, int, S_IRUGO); | 71 | module_param_named(debug_level, ehca_debug_level, int, S_IRUGO); |
@@ -260,13 +260,20 @@ static struct cap_descr { | |||
260 | { HCA_CAP_MINI_QP, "HCA_CAP_MINI_QP" }, | 260 | { HCA_CAP_MINI_QP, "HCA_CAP_MINI_QP" }, |
261 | }; | 261 | }; |
262 | 262 | ||
263 | int ehca_sense_attributes(struct ehca_shca *shca) | 263 | static int ehca_sense_attributes(struct ehca_shca *shca) |
264 | { | 264 | { |
265 | int i, ret = 0; | 265 | int i, ret = 0; |
266 | u64 h_ret; | 266 | u64 h_ret; |
267 | struct hipz_query_hca *rblock; | 267 | struct hipz_query_hca *rblock; |
268 | struct hipz_query_port *port; | 268 | struct hipz_query_port *port; |
269 | 269 | ||
270 | static const u32 pgsize_map[] = { | ||
271 | HCA_CAP_MR_PGSIZE_4K, 0x1000, | ||
272 | HCA_CAP_MR_PGSIZE_64K, 0x10000, | ||
273 | HCA_CAP_MR_PGSIZE_1M, 0x100000, | ||
274 | HCA_CAP_MR_PGSIZE_16M, 0x1000000, | ||
275 | }; | ||
276 | |||
270 | rblock = ehca_alloc_fw_ctrlblock(GFP_KERNEL); | 277 | rblock = ehca_alloc_fw_ctrlblock(GFP_KERNEL); |
271 | if (!rblock) { | 278 | if (!rblock) { |
272 | ehca_gen_err("Cannot allocate rblock memory."); | 279 | ehca_gen_err("Cannot allocate rblock memory."); |
@@ -329,8 +336,15 @@ int ehca_sense_attributes(struct ehca_shca *shca) | |||
329 | if (EHCA_BMASK_GET(hca_cap_descr[i].mask, shca->hca_cap)) | 336 | if (EHCA_BMASK_GET(hca_cap_descr[i].mask, shca->hca_cap)) |
330 | ehca_gen_dbg(" %s", hca_cap_descr[i].descr); | 337 | ehca_gen_dbg(" %s", hca_cap_descr[i].descr); |
331 | 338 | ||
332 | shca->hca_cap_mr_pgsize = rblock->memory_page_size_supported; | 339 | /* translate supported MR page sizes; always support 4K */ |
340 | shca->hca_cap_mr_pgsize = EHCA_PAGESIZE; | ||
341 | if (ehca_mr_largepage) { /* support extra sizes only if enabled */ | ||
342 | for (i = 0; i < ARRAY_SIZE(pgsize_map); i += 2) | ||
343 | if (rblock->memory_page_size_supported & pgsize_map[i]) | ||
344 | shca->hca_cap_mr_pgsize |= pgsize_map[i + 1]; | ||
345 | } | ||
333 | 346 | ||
347 | /* query max MTU from first port -- it's the same for all ports */ | ||
334 | port = (struct hipz_query_port *)rblock; | 348 | port = (struct hipz_query_port *)rblock; |
335 | h_ret = hipz_h_query_port(shca->ipz_hca_handle, 1, port); | 349 | h_ret = hipz_h_query_port(shca->ipz_hca_handle, 1, port); |
336 | if (h_ret != H_SUCCESS) { | 350 | if (h_ret != H_SUCCESS) { |
diff --git a/drivers/infiniband/hw/ehca/ehca_mrmw.c b/drivers/infiniband/hw/ehca/ehca_mrmw.c index ead7230d7738..e239bbf54da1 100644 --- a/drivers/infiniband/hw/ehca/ehca_mrmw.c +++ b/drivers/infiniband/hw/ehca/ehca_mrmw.c | |||
@@ -72,24 +72,14 @@ enum ehca_mr_pgsize { | |||
72 | 72 | ||
73 | static u32 ehca_encode_hwpage_size(u32 pgsize) | 73 | static u32 ehca_encode_hwpage_size(u32 pgsize) |
74 | { | 74 | { |
75 | u32 idx = 0; | 75 | int log = ilog2(pgsize); |
76 | pgsize >>= 12; | 76 | WARN_ON(log < 12 || log > 24 || log & 3); |
77 | /* | 77 | return (log - 12) / 4; |
78 | * map mr page size into hw code: | ||
79 | * 0, 1, 2, 3 for 4K, 64K, 1M, 64M | ||
80 | */ | ||
81 | while (!(pgsize & 1)) { | ||
82 | idx++; | ||
83 | pgsize >>= 4; | ||
84 | } | ||
85 | return idx; | ||
86 | } | 78 | } |
87 | 79 | ||
88 | static u64 ehca_get_max_hwpage_size(struct ehca_shca *shca) | 80 | static u64 ehca_get_max_hwpage_size(struct ehca_shca *shca) |
89 | { | 81 | { |
90 | if (shca->hca_cap_mr_pgsize & HCA_CAP_MR_PGSIZE_16M) | 82 | return 1UL << ilog2(shca->hca_cap_mr_pgsize); |
91 | return EHCA_MR_PGSIZE16M; | ||
92 | return EHCA_MR_PGSIZE4K; | ||
93 | } | 83 | } |
94 | 84 | ||
95 | static struct ehca_mr *ehca_mr_new(void) | 85 | static struct ehca_mr *ehca_mr_new(void) |
@@ -259,7 +249,7 @@ struct ib_mr *ehca_reg_phys_mr(struct ib_pd *pd, | |||
259 | pginfo.u.phy.num_phys_buf = num_phys_buf; | 249 | pginfo.u.phy.num_phys_buf = num_phys_buf; |
260 | pginfo.u.phy.phys_buf_array = phys_buf_array; | 250 | pginfo.u.phy.phys_buf_array = phys_buf_array; |
261 | pginfo.next_hwpage = | 251 | pginfo.next_hwpage = |
262 | ((u64)iova_start & ~(hw_pgsize - 1)) / hw_pgsize; | 252 | ((u64)iova_start & ~PAGE_MASK) / hw_pgsize; |
263 | 253 | ||
264 | ret = ehca_reg_mr(shca, e_mr, iova_start, size, mr_access_flags, | 254 | ret = ehca_reg_mr(shca, e_mr, iova_start, size, mr_access_flags, |
265 | e_pd, &pginfo, &e_mr->ib.ib_mr.lkey, | 255 | e_pd, &pginfo, &e_mr->ib.ib_mr.lkey, |
@@ -296,7 +286,7 @@ struct ib_mr *ehca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, | |||
296 | container_of(pd->device, struct ehca_shca, ib_device); | 286 | container_of(pd->device, struct ehca_shca, ib_device); |
297 | struct ehca_pd *e_pd = container_of(pd, struct ehca_pd, ib_pd); | 287 | struct ehca_pd *e_pd = container_of(pd, struct ehca_pd, ib_pd); |
298 | struct ehca_mr_pginfo pginfo; | 288 | struct ehca_mr_pginfo pginfo; |
299 | int ret; | 289 | int ret, page_shift; |
300 | u32 num_kpages; | 290 | u32 num_kpages; |
301 | u32 num_hwpages; | 291 | u32 num_hwpages; |
302 | u64 hwpage_size; | 292 | u64 hwpage_size; |
@@ -351,19 +341,20 @@ struct ib_mr *ehca_reg_user_mr(struct ib_pd *pd, u64 start, u64 length, | |||
351 | /* determine number of MR pages */ | 341 | /* determine number of MR pages */ |
352 | num_kpages = NUM_CHUNKS((virt % PAGE_SIZE) + length, PAGE_SIZE); | 342 | num_kpages = NUM_CHUNKS((virt % PAGE_SIZE) + length, PAGE_SIZE); |
353 | /* select proper hw_pgsize */ | 343 | /* select proper hw_pgsize */ |
354 | if (ehca_mr_largepage && | 344 | page_shift = PAGE_SHIFT; |
355 | (shca->hca_cap_mr_pgsize & HCA_CAP_MR_PGSIZE_16M)) { | 345 | if (e_mr->umem->hugetlb) { |
356 | int page_shift = PAGE_SHIFT; | 346 | /* determine page_shift, clamp between 4K and 16M */ |
357 | if (e_mr->umem->hugetlb) { | 347 | page_shift = (fls64(length - 1) + 3) & ~3; |
358 | /* determine page_shift, clamp between 4K and 16M */ | 348 | page_shift = min(max(page_shift, EHCA_MR_PGSHIFT4K), |
359 | page_shift = (fls64(length - 1) + 3) & ~3; | 349 | EHCA_MR_PGSHIFT16M); |
360 | page_shift = min(max(page_shift, EHCA_MR_PGSHIFT4K), | 350 | } |
361 | EHCA_MR_PGSHIFT16M); | 351 | hwpage_size = 1UL << page_shift; |
362 | } | 352 | |
363 | hwpage_size = 1UL << page_shift; | 353 | /* now that we have the desired page size, shift until it's |
364 | } else | 354 | * supported, too. 4K is always supported, so this terminates. |
365 | hwpage_size = EHCA_MR_PGSIZE4K; /* ehca1 only supports 4k */ | 355 | */ |
366 | ehca_dbg(pd->device, "hwpage_size=%lx", hwpage_size); | 356 | while (!(hwpage_size & shca->hca_cap_mr_pgsize)) |
357 | hwpage_size >>= 4; | ||
367 | 358 | ||
368 | reg_user_mr_fallback: | 359 | reg_user_mr_fallback: |
369 | num_hwpages = NUM_CHUNKS((virt % hwpage_size) + length, hwpage_size); | 360 | num_hwpages = NUM_CHUNKS((virt % hwpage_size) + length, hwpage_size); |
@@ -547,7 +538,7 @@ int ehca_rereg_phys_mr(struct ib_mr *mr, | |||
547 | pginfo.u.phy.num_phys_buf = num_phys_buf; | 538 | pginfo.u.phy.num_phys_buf = num_phys_buf; |
548 | pginfo.u.phy.phys_buf_array = phys_buf_array; | 539 | pginfo.u.phy.phys_buf_array = phys_buf_array; |
549 | pginfo.next_hwpage = | 540 | pginfo.next_hwpage = |
550 | ((u64)iova_start & ~(hw_pgsize - 1)) / hw_pgsize; | 541 | ((u64)iova_start & ~PAGE_MASK) / hw_pgsize; |
551 | } | 542 | } |
552 | if (mr_rereg_mask & IB_MR_REREG_ACCESS) | 543 | if (mr_rereg_mask & IB_MR_REREG_ACCESS) |
553 | new_acl = mr_access_flags; | 544 | new_acl = mr_access_flags; |
@@ -809,8 +800,9 @@ struct ib_fmr *ehca_alloc_fmr(struct ib_pd *pd, | |||
809 | ib_fmr = ERR_PTR(-EINVAL); | 800 | ib_fmr = ERR_PTR(-EINVAL); |
810 | goto alloc_fmr_exit0; | 801 | goto alloc_fmr_exit0; |
811 | } | 802 | } |
812 | hw_pgsize = ehca_get_max_hwpage_size(shca); | 803 | |
813 | if ((1 << fmr_attr->page_shift) != hw_pgsize) { | 804 | hw_pgsize = 1 << fmr_attr->page_shift; |
805 | if (!(hw_pgsize & shca->hca_cap_mr_pgsize)) { | ||
814 | ehca_err(pd->device, "unsupported fmr_attr->page_shift=%x", | 806 | ehca_err(pd->device, "unsupported fmr_attr->page_shift=%x", |
815 | fmr_attr->page_shift); | 807 | fmr_attr->page_shift); |
816 | ib_fmr = ERR_PTR(-EINVAL); | 808 | ib_fmr = ERR_PTR(-EINVAL); |
@@ -826,6 +818,7 @@ struct ib_fmr *ehca_alloc_fmr(struct ib_pd *pd, | |||
826 | 818 | ||
827 | /* register MR on HCA */ | 819 | /* register MR on HCA */ |
828 | memset(&pginfo, 0, sizeof(pginfo)); | 820 | memset(&pginfo, 0, sizeof(pginfo)); |
821 | pginfo.hwpage_size = hw_pgsize; | ||
829 | /* | 822 | /* |
830 | * pginfo.num_hwpages==0, ie register_rpages() will not be called | 823 | * pginfo.num_hwpages==0, ie register_rpages() will not be called |
831 | * but deferred to map_phys_fmr() | 824 | * but deferred to map_phys_fmr() |
diff --git a/drivers/infiniband/hw/ehca/ehca_qp.c b/drivers/infiniband/hw/ehca/ehca_qp.c index e2bd62be11e7..de182648b282 100644 --- a/drivers/infiniband/hw/ehca/ehca_qp.c +++ b/drivers/infiniband/hw/ehca/ehca_qp.c | |||
@@ -451,7 +451,6 @@ static struct ehca_qp *internal_create_qp( | |||
451 | has_srq = 1; | 451 | has_srq = 1; |
452 | parms.ext_type = EQPT_SRQBASE; | 452 | parms.ext_type = EQPT_SRQBASE; |
453 | parms.srq_qpn = my_srq->real_qp_num; | 453 | parms.srq_qpn = my_srq->real_qp_num; |
454 | parms.srq_token = my_srq->token; | ||
455 | } | 454 | } |
456 | 455 | ||
457 | if (is_llqp && has_srq) { | 456 | if (is_llqp && has_srq) { |
@@ -583,6 +582,9 @@ static struct ehca_qp *internal_create_qp( | |||
583 | goto create_qp_exit1; | 582 | goto create_qp_exit1; |
584 | } | 583 | } |
585 | 584 | ||
585 | if (has_srq) | ||
586 | parms.srq_token = my_qp->token; | ||
587 | |||
586 | parms.servicetype = ibqptype2servicetype(qp_type); | 588 | parms.servicetype = ibqptype2servicetype(qp_type); |
587 | if (parms.servicetype < 0) { | 589 | if (parms.servicetype < 0) { |
588 | ret = -EINVAL; | 590 | ret = -EINVAL; |