diff options
Diffstat (limited to 'drivers')
-rw-r--r-- | drivers/infiniband/core/cma.c | 190 | ||||
-rw-r--r-- | drivers/infiniband/core/ucma.c | 7 |
2 files changed, 130 insertions, 67 deletions
diff --git a/drivers/infiniband/core/cma.c b/drivers/infiniband/core/cma.c index eff5e46f005c..99dde874fbbd 100644 --- a/drivers/infiniband/core/cma.c +++ b/drivers/infiniband/core/cma.c | |||
@@ -148,6 +148,7 @@ struct rdma_id_private { | |||
148 | u32 qp_num; | 148 | u32 qp_num; |
149 | u8 srq; | 149 | u8 srq; |
150 | u8 tos; | 150 | u8 tos; |
151 | u8 reuseaddr; | ||
151 | }; | 152 | }; |
152 | 153 | ||
153 | struct cma_multicast { | 154 | struct cma_multicast { |
@@ -1579,50 +1580,6 @@ static void cma_listen_on_all(struct rdma_id_private *id_priv) | |||
1579 | mutex_unlock(&lock); | 1580 | mutex_unlock(&lock); |
1580 | } | 1581 | } |
1581 | 1582 | ||
1582 | int rdma_listen(struct rdma_cm_id *id, int backlog) | ||
1583 | { | ||
1584 | struct rdma_id_private *id_priv; | ||
1585 | int ret; | ||
1586 | |||
1587 | id_priv = container_of(id, struct rdma_id_private, id); | ||
1588 | if (id_priv->state == CMA_IDLE) { | ||
1589 | ((struct sockaddr *) &id->route.addr.src_addr)->sa_family = AF_INET; | ||
1590 | ret = rdma_bind_addr(id, (struct sockaddr *) &id->route.addr.src_addr); | ||
1591 | if (ret) | ||
1592 | return ret; | ||
1593 | } | ||
1594 | |||
1595 | if (!cma_comp_exch(id_priv, CMA_ADDR_BOUND, CMA_LISTEN)) | ||
1596 | return -EINVAL; | ||
1597 | |||
1598 | id_priv->backlog = backlog; | ||
1599 | if (id->device) { | ||
1600 | switch (rdma_node_get_transport(id->device->node_type)) { | ||
1601 | case RDMA_TRANSPORT_IB: | ||
1602 | ret = cma_ib_listen(id_priv); | ||
1603 | if (ret) | ||
1604 | goto err; | ||
1605 | break; | ||
1606 | case RDMA_TRANSPORT_IWARP: | ||
1607 | ret = cma_iw_listen(id_priv, backlog); | ||
1608 | if (ret) | ||
1609 | goto err; | ||
1610 | break; | ||
1611 | default: | ||
1612 | ret = -ENOSYS; | ||
1613 | goto err; | ||
1614 | } | ||
1615 | } else | ||
1616 | cma_listen_on_all(id_priv); | ||
1617 | |||
1618 | return 0; | ||
1619 | err: | ||
1620 | id_priv->backlog = 0; | ||
1621 | cma_comp_exch(id_priv, CMA_LISTEN, CMA_ADDR_BOUND); | ||
1622 | return ret; | ||
1623 | } | ||
1624 | EXPORT_SYMBOL(rdma_listen); | ||
1625 | |||
1626 | void rdma_set_service_type(struct rdma_cm_id *id, int tos) | 1583 | void rdma_set_service_type(struct rdma_cm_id *id, int tos) |
1627 | { | 1584 | { |
1628 | struct rdma_id_private *id_priv; | 1585 | struct rdma_id_private *id_priv; |
@@ -2105,6 +2062,25 @@ err: | |||
2105 | } | 2062 | } |
2106 | EXPORT_SYMBOL(rdma_resolve_addr); | 2063 | EXPORT_SYMBOL(rdma_resolve_addr); |
2107 | 2064 | ||
2065 | int rdma_set_reuseaddr(struct rdma_cm_id *id, int reuse) | ||
2066 | { | ||
2067 | struct rdma_id_private *id_priv; | ||
2068 | unsigned long flags; | ||
2069 | int ret; | ||
2070 | |||
2071 | id_priv = container_of(id, struct rdma_id_private, id); | ||
2072 | spin_lock_irqsave(&id_priv->lock, flags); | ||
2073 | if (id_priv->state == CMA_IDLE) { | ||
2074 | id_priv->reuseaddr = reuse; | ||
2075 | ret = 0; | ||
2076 | } else { | ||
2077 | ret = -EINVAL; | ||
2078 | } | ||
2079 | spin_unlock_irqrestore(&id_priv->lock, flags); | ||
2080 | return ret; | ||
2081 | } | ||
2082 | EXPORT_SYMBOL(rdma_set_reuseaddr); | ||
2083 | |||
2108 | static void cma_bind_port(struct rdma_bind_list *bind_list, | 2084 | static void cma_bind_port(struct rdma_bind_list *bind_list, |
2109 | struct rdma_id_private *id_priv) | 2085 | struct rdma_id_private *id_priv) |
2110 | { | 2086 | { |
@@ -2180,43 +2156,73 @@ retry: | |||
2180 | return -EADDRNOTAVAIL; | 2156 | return -EADDRNOTAVAIL; |
2181 | } | 2157 | } |
2182 | 2158 | ||
2183 | static int cma_use_port(struct idr *ps, struct rdma_id_private *id_priv) | 2159 | /* |
2160 | * Check that the requested port is available. This is called when trying to | ||
2161 | * bind to a specific port, or when trying to listen on a bound port. In | ||
2162 | * the latter case, the provided id_priv may already be on the bind_list, but | ||
2163 | * we still need to check that it's okay to start listening. | ||
2164 | */ | ||
2165 | static int cma_check_port(struct rdma_bind_list *bind_list, | ||
2166 | struct rdma_id_private *id_priv, uint8_t reuseaddr) | ||
2184 | { | 2167 | { |
2185 | struct rdma_id_private *cur_id; | 2168 | struct rdma_id_private *cur_id; |
2186 | struct sockaddr *addr, *cur_addr; | 2169 | struct sockaddr *addr, *cur_addr; |
2187 | struct rdma_bind_list *bind_list; | ||
2188 | struct hlist_node *node; | 2170 | struct hlist_node *node; |
2189 | unsigned short snum; | ||
2190 | 2171 | ||
2191 | addr = (struct sockaddr *) &id_priv->id.route.addr.src_addr; | 2172 | addr = (struct sockaddr *) &id_priv->id.route.addr.src_addr; |
2192 | snum = ntohs(cma_port(addr)); | 2173 | if (cma_any_addr(addr) && !reuseaddr) |
2193 | if (snum < PROT_SOCK && !capable(CAP_NET_BIND_SERVICE)) | ||
2194 | return -EACCES; | ||
2195 | |||
2196 | bind_list = idr_find(ps, snum); | ||
2197 | if (!bind_list) | ||
2198 | return cma_alloc_port(ps, id_priv, snum); | ||
2199 | |||
2200 | /* | ||
2201 | * We don't support binding to any address if anyone is bound to | ||
2202 | * a specific address on the same port. | ||
2203 | */ | ||
2204 | if (cma_any_addr(addr)) | ||
2205 | return -EADDRNOTAVAIL; | 2174 | return -EADDRNOTAVAIL; |
2206 | 2175 | ||
2207 | hlist_for_each_entry(cur_id, node, &bind_list->owners, node) { | 2176 | hlist_for_each_entry(cur_id, node, &bind_list->owners, node) { |
2208 | cur_addr = (struct sockaddr *) &cur_id->id.route.addr.src_addr; | 2177 | if (id_priv == cur_id) |
2209 | if (cma_any_addr(cur_addr)) | 2178 | continue; |
2210 | return -EADDRNOTAVAIL; | ||
2211 | 2179 | ||
2212 | if (!cma_addr_cmp(addr, cur_addr)) | 2180 | if ((cur_id->state == CMA_LISTEN) || |
2213 | return -EADDRINUSE; | 2181 | !reuseaddr || !cur_id->reuseaddr) { |
2214 | } | 2182 | cur_addr = (struct sockaddr *) &cur_id->id.route.addr.src_addr; |
2183 | if (cma_any_addr(cur_addr)) | ||
2184 | return -EADDRNOTAVAIL; | ||
2215 | 2185 | ||
2216 | cma_bind_port(bind_list, id_priv); | 2186 | if (!cma_addr_cmp(addr, cur_addr)) |
2187 | return -EADDRINUSE; | ||
2188 | } | ||
2189 | } | ||
2217 | return 0; | 2190 | return 0; |
2218 | } | 2191 | } |
2219 | 2192 | ||
2193 | static int cma_use_port(struct idr *ps, struct rdma_id_private *id_priv) | ||
2194 | { | ||
2195 | struct rdma_bind_list *bind_list; | ||
2196 | unsigned short snum; | ||
2197 | int ret; | ||
2198 | |||
2199 | snum = ntohs(cma_port((struct sockaddr *) &id_priv->id.route.addr.src_addr)); | ||
2200 | if (snum < PROT_SOCK && !capable(CAP_NET_BIND_SERVICE)) | ||
2201 | return -EACCES; | ||
2202 | |||
2203 | bind_list = idr_find(ps, snum); | ||
2204 | if (!bind_list) { | ||
2205 | ret = cma_alloc_port(ps, id_priv, snum); | ||
2206 | } else { | ||
2207 | ret = cma_check_port(bind_list, id_priv, id_priv->reuseaddr); | ||
2208 | if (!ret) | ||
2209 | cma_bind_port(bind_list, id_priv); | ||
2210 | } | ||
2211 | return ret; | ||
2212 | } | ||
2213 | |||
2214 | static int cma_bind_listen(struct rdma_id_private *id_priv) | ||
2215 | { | ||
2216 | struct rdma_bind_list *bind_list = id_priv->bind_list; | ||
2217 | int ret = 0; | ||
2218 | |||
2219 | mutex_lock(&lock); | ||
2220 | if (bind_list->owners.first->next) | ||
2221 | ret = cma_check_port(bind_list, id_priv, 0); | ||
2222 | mutex_unlock(&lock); | ||
2223 | return ret; | ||
2224 | } | ||
2225 | |||
2220 | static int cma_get_port(struct rdma_id_private *id_priv) | 2226 | static int cma_get_port(struct rdma_id_private *id_priv) |
2221 | { | 2227 | { |
2222 | struct idr *ps; | 2228 | struct idr *ps; |
@@ -2268,6 +2274,56 @@ static int cma_check_linklocal(struct rdma_dev_addr *dev_addr, | |||
2268 | return 0; | 2274 | return 0; |
2269 | } | 2275 | } |
2270 | 2276 | ||
2277 | int rdma_listen(struct rdma_cm_id *id, int backlog) | ||
2278 | { | ||
2279 | struct rdma_id_private *id_priv; | ||
2280 | int ret; | ||
2281 | |||
2282 | id_priv = container_of(id, struct rdma_id_private, id); | ||
2283 | if (id_priv->state == CMA_IDLE) { | ||
2284 | ((struct sockaddr *) &id->route.addr.src_addr)->sa_family = AF_INET; | ||
2285 | ret = rdma_bind_addr(id, (struct sockaddr *) &id->route.addr.src_addr); | ||
2286 | if (ret) | ||
2287 | return ret; | ||
2288 | } | ||
2289 | |||
2290 | if (!cma_comp_exch(id_priv, CMA_ADDR_BOUND, CMA_LISTEN)) | ||
2291 | return -EINVAL; | ||
2292 | |||
2293 | if (id_priv->reuseaddr) { | ||
2294 | ret = cma_bind_listen(id_priv); | ||
2295 | if (ret) | ||
2296 | goto err; | ||
2297 | } | ||
2298 | |||
2299 | id_priv->backlog = backlog; | ||
2300 | if (id->device) { | ||
2301 | switch (rdma_node_get_transport(id->device->node_type)) { | ||
2302 | case RDMA_TRANSPORT_IB: | ||
2303 | ret = cma_ib_listen(id_priv); | ||
2304 | if (ret) | ||
2305 | goto err; | ||
2306 | break; | ||
2307 | case RDMA_TRANSPORT_IWARP: | ||
2308 | ret = cma_iw_listen(id_priv, backlog); | ||
2309 | if (ret) | ||
2310 | goto err; | ||
2311 | break; | ||
2312 | default: | ||
2313 | ret = -ENOSYS; | ||
2314 | goto err; | ||
2315 | } | ||
2316 | } else | ||
2317 | cma_listen_on_all(id_priv); | ||
2318 | |||
2319 | return 0; | ||
2320 | err: | ||
2321 | id_priv->backlog = 0; | ||
2322 | cma_comp_exch(id_priv, CMA_LISTEN, CMA_ADDR_BOUND); | ||
2323 | return ret; | ||
2324 | } | ||
2325 | EXPORT_SYMBOL(rdma_listen); | ||
2326 | |||
2271 | int rdma_bind_addr(struct rdma_cm_id *id, struct sockaddr *addr) | 2327 | int rdma_bind_addr(struct rdma_cm_id *id, struct sockaddr *addr) |
2272 | { | 2328 | { |
2273 | struct rdma_id_private *id_priv; | 2329 | struct rdma_id_private *id_priv; |
diff --git a/drivers/infiniband/core/ucma.c b/drivers/infiniband/core/ucma.c index ec1e9da1488b..b3fa798525b2 100644 --- a/drivers/infiniband/core/ucma.c +++ b/drivers/infiniband/core/ucma.c | |||
@@ -883,6 +883,13 @@ static int ucma_set_option_id(struct ucma_context *ctx, int optname, | |||
883 | } | 883 | } |
884 | rdma_set_service_type(ctx->cm_id, *((u8 *) optval)); | 884 | rdma_set_service_type(ctx->cm_id, *((u8 *) optval)); |
885 | break; | 885 | break; |
886 | case RDMA_OPTION_ID_REUSEADDR: | ||
887 | if (optlen != sizeof(int)) { | ||
888 | ret = -EINVAL; | ||
889 | break; | ||
890 | } | ||
891 | ret = rdma_set_reuseaddr(ctx->cm_id, *((int *) optval) ? 1 : 0); | ||
892 | break; | ||
886 | default: | 893 | default: |
887 | ret = -ENOSYS; | 894 | ret = -ENOSYS; |
888 | } | 895 | } |