diff options
Diffstat (limited to 'net/decnet/dn_rules.c')
| -rw-r--r-- | net/decnet/dn_rules.c | 115 |
1 files changed, 62 insertions, 53 deletions
diff --git a/net/decnet/dn_rules.c b/net/decnet/dn_rules.c index 1060de70bc0c..446faafe2065 100644 --- a/net/decnet/dn_rules.c +++ b/net/decnet/dn_rules.c | |||
| @@ -27,6 +27,8 @@ | |||
| 27 | #include <linux/timer.h> | 27 | #include <linux/timer.h> |
| 28 | #include <linux/spinlock.h> | 28 | #include <linux/spinlock.h> |
| 29 | #include <linux/in_route.h> | 29 | #include <linux/in_route.h> |
| 30 | #include <linux/list.h> | ||
| 31 | #include <linux/rcupdate.h> | ||
| 30 | #include <asm/atomic.h> | 32 | #include <asm/atomic.h> |
| 31 | #include <asm/uaccess.h> | 33 | #include <asm/uaccess.h> |
| 32 | #include <net/neighbour.h> | 34 | #include <net/neighbour.h> |
| @@ -39,18 +41,18 @@ | |||
| 39 | 41 | ||
| 40 | struct dn_fib_rule | 42 | struct dn_fib_rule |
| 41 | { | 43 | { |
| 42 | struct dn_fib_rule *r_next; | 44 | struct hlist_node r_hlist; |
| 43 | atomic_t r_clntref; | 45 | atomic_t r_clntref; |
| 44 | u32 r_preference; | 46 | u32 r_preference; |
| 45 | unsigned char r_table; | 47 | unsigned char r_table; |
| 46 | unsigned char r_action; | 48 | unsigned char r_action; |
| 47 | unsigned char r_dst_len; | 49 | unsigned char r_dst_len; |
| 48 | unsigned char r_src_len; | 50 | unsigned char r_src_len; |
| 49 | dn_address r_src; | 51 | __le16 r_src; |
| 50 | dn_address r_srcmask; | 52 | __le16 r_srcmask; |
| 51 | dn_address r_dst; | 53 | __le16 r_dst; |
| 52 | dn_address r_dstmask; | 54 | __le16 r_dstmask; |
| 53 | dn_address r_srcmap; | 55 | __le16 r_srcmap; |
| 54 | u8 r_flags; | 56 | u8 r_flags; |
| 55 | #ifdef CONFIG_DECNET_ROUTE_FWMARK | 57 | #ifdef CONFIG_DECNET_ROUTE_FWMARK |
| 56 | u32 r_fwmark; | 58 | u32 r_fwmark; |
| @@ -58,6 +60,7 @@ struct dn_fib_rule | |||
| 58 | int r_ifindex; | 60 | int r_ifindex; |
| 59 | char r_ifname[IFNAMSIZ]; | 61 | char r_ifname[IFNAMSIZ]; |
| 60 | int r_dead; | 62 | int r_dead; |
| 63 | struct rcu_head rcu; | ||
| 61 | }; | 64 | }; |
| 62 | 65 | ||
| 63 | static struct dn_fib_rule default_rule = { | 66 | static struct dn_fib_rule default_rule = { |
| @@ -67,18 +70,17 @@ static struct dn_fib_rule default_rule = { | |||
| 67 | .r_action = RTN_UNICAST | 70 | .r_action = RTN_UNICAST |
| 68 | }; | 71 | }; |
| 69 | 72 | ||
| 70 | static struct dn_fib_rule *dn_fib_rules = &default_rule; | 73 | static struct hlist_head dn_fib_rules; |
| 71 | static DEFINE_RWLOCK(dn_fib_rules_lock); | ||
| 72 | |||
| 73 | 74 | ||
| 74 | int dn_fib_rtm_delrule(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg) | 75 | int dn_fib_rtm_delrule(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg) |
| 75 | { | 76 | { |
| 76 | struct rtattr **rta = arg; | 77 | struct rtattr **rta = arg; |
| 77 | struct rtmsg *rtm = NLMSG_DATA(nlh); | 78 | struct rtmsg *rtm = NLMSG_DATA(nlh); |
| 78 | struct dn_fib_rule *r, **rp; | 79 | struct dn_fib_rule *r; |
| 80 | struct hlist_node *node; | ||
| 79 | int err = -ESRCH; | 81 | int err = -ESRCH; |
| 80 | 82 | ||
| 81 | for(rp=&dn_fib_rules; (r=*rp) != NULL; rp = &r->r_next) { | 83 | hlist_for_each_entry(r, node, &dn_fib_rules, r_hlist) { |
| 82 | if ((!rta[RTA_SRC-1] || memcmp(RTA_DATA(rta[RTA_SRC-1]), &r->r_src, 2) == 0) && | 84 | if ((!rta[RTA_SRC-1] || memcmp(RTA_DATA(rta[RTA_SRC-1]), &r->r_src, 2) == 0) && |
| 83 | rtm->rtm_src_len == r->r_src_len && | 85 | rtm->rtm_src_len == r->r_src_len && |
| 84 | rtm->rtm_dst_len == r->r_dst_len && | 86 | rtm->rtm_dst_len == r->r_dst_len && |
| @@ -95,10 +97,8 @@ int dn_fib_rtm_delrule(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg) | |||
| 95 | if (r == &default_rule) | 97 | if (r == &default_rule) |
| 96 | break; | 98 | break; |
| 97 | 99 | ||
| 98 | write_lock_bh(&dn_fib_rules_lock); | 100 | hlist_del_rcu(&r->r_hlist); |
| 99 | *rp = r->r_next; | ||
| 100 | r->r_dead = 1; | 101 | r->r_dead = 1; |
| 101 | write_unlock_bh(&dn_fib_rules_lock); | ||
| 102 | dn_fib_rule_put(r); | 102 | dn_fib_rule_put(r); |
| 103 | err = 0; | 103 | err = 0; |
| 104 | break; | 104 | break; |
| @@ -108,11 +108,17 @@ int dn_fib_rtm_delrule(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg) | |||
| 108 | return err; | 108 | return err; |
| 109 | } | 109 | } |
| 110 | 110 | ||
| 111 | static inline void dn_fib_rule_put_rcu(struct rcu_head *head) | ||
| 112 | { | ||
| 113 | struct dn_fib_rule *r = container_of(head, struct dn_fib_rule, rcu); | ||
| 114 | kfree(r); | ||
| 115 | } | ||
| 116 | |||
| 111 | void dn_fib_rule_put(struct dn_fib_rule *r) | 117 | void dn_fib_rule_put(struct dn_fib_rule *r) |
| 112 | { | 118 | { |
| 113 | if (atomic_dec_and_test(&r->r_clntref)) { | 119 | if (atomic_dec_and_test(&r->r_clntref)) { |
| 114 | if (r->r_dead) | 120 | if (r->r_dead) |
| 115 | kfree(r); | 121 | call_rcu(&r->rcu, dn_fib_rule_put_rcu); |
| 116 | else | 122 | else |
| 117 | printk(KERN_DEBUG "Attempt to free alive dn_fib_rule\n"); | 123 | printk(KERN_DEBUG "Attempt to free alive dn_fib_rule\n"); |
| 118 | } | 124 | } |
| @@ -123,7 +129,8 @@ int dn_fib_rtm_newrule(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg) | |||
| 123 | { | 129 | { |
| 124 | struct rtattr **rta = arg; | 130 | struct rtattr **rta = arg; |
| 125 | struct rtmsg *rtm = NLMSG_DATA(nlh); | 131 | struct rtmsg *rtm = NLMSG_DATA(nlh); |
| 126 | struct dn_fib_rule *r, *new_r, **rp; | 132 | struct dn_fib_rule *r, *new_r, *last = NULL; |
| 133 | struct hlist_node *node = NULL; | ||
| 127 | unsigned char table_id; | 134 | unsigned char table_id; |
| 128 | 135 | ||
| 129 | if (rtm->rtm_src_len > 16 || rtm->rtm_dst_len > 16) | 136 | if (rtm->rtm_src_len > 16 || rtm->rtm_dst_len > 16) |
| @@ -149,6 +156,7 @@ int dn_fib_rtm_newrule(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg) | |||
| 149 | if (!new_r) | 156 | if (!new_r) |
| 150 | return -ENOMEM; | 157 | return -ENOMEM; |
| 151 | memset(new_r, 0, sizeof(*new_r)); | 158 | memset(new_r, 0, sizeof(*new_r)); |
| 159 | |||
| 152 | if (rta[RTA_SRC-1]) | 160 | if (rta[RTA_SRC-1]) |
| 153 | memcpy(&new_r->r_src, RTA_DATA(rta[RTA_SRC-1]), 2); | 161 | memcpy(&new_r->r_src, RTA_DATA(rta[RTA_SRC-1]), 2); |
| 154 | if (rta[RTA_DST-1]) | 162 | if (rta[RTA_DST-1]) |
| @@ -179,27 +187,26 @@ int dn_fib_rtm_newrule(struct sk_buff *skb, struct nlmsghdr *nlh, void *arg) | |||
| 179 | } | 187 | } |
| 180 | } | 188 | } |
| 181 | 189 | ||
| 182 | rp = &dn_fib_rules; | 190 | r = container_of(dn_fib_rules.first, struct dn_fib_rule, r_hlist); |
| 183 | if (!new_r->r_preference) { | 191 | if (!new_r->r_preference) { |
| 184 | r = dn_fib_rules; | 192 | if (r && r->r_hlist.next != NULL) { |
| 185 | if (r && (r = r->r_next) != NULL) { | 193 | r = container_of(r->r_hlist.next, struct dn_fib_rule, r_hlist); |
| 186 | rp = &dn_fib_rules->r_next; | ||
| 187 | if (r->r_preference) | 194 | if (r->r_preference) |
| 188 | new_r->r_preference = r->r_preference - 1; | 195 | new_r->r_preference = r->r_preference - 1; |
| 189 | } | 196 | } |
| 190 | } | 197 | } |
| 191 | 198 | ||
| 192 | while((r=*rp) != NULL) { | 199 | hlist_for_each_entry(r, node, &dn_fib_rules, r_hlist) { |
| 193 | if (r->r_preference > new_r->r_preference) | 200 | if (r->r_preference > new_r->r_preference) |
| 194 | break; | 201 | break; |
| 195 | rp = &r->r_next; | 202 | last = r; |
| 196 | } | 203 | } |
| 197 | |||
| 198 | new_r->r_next = r; | ||
| 199 | atomic_inc(&new_r->r_clntref); | 204 | atomic_inc(&new_r->r_clntref); |
| 200 | write_lock_bh(&dn_fib_rules_lock); | 205 | |
| 201 | *rp = new_r; | 206 | if (last) |
| 202 | write_unlock_bh(&dn_fib_rules_lock); | 207 | hlist_add_after_rcu(&last->r_hlist, &new_r->r_hlist); |
| 208 | else | ||
| 209 | hlist_add_before_rcu(&new_r->r_hlist, &r->r_hlist); | ||
| 203 | return 0; | 210 | return 0; |
| 204 | } | 211 | } |
| 205 | 212 | ||
| @@ -208,12 +215,14 @@ int dn_fib_lookup(const struct flowi *flp, struct dn_fib_res *res) | |||
| 208 | { | 215 | { |
| 209 | struct dn_fib_rule *r, *policy; | 216 | struct dn_fib_rule *r, *policy; |
| 210 | struct dn_fib_table *tb; | 217 | struct dn_fib_table *tb; |
| 211 | dn_address saddr = flp->fld_src; | 218 | __le16 saddr = flp->fld_src; |
| 212 | dn_address daddr = flp->fld_dst; | 219 | __le16 daddr = flp->fld_dst; |
| 220 | struct hlist_node *node; | ||
| 213 | int err; | 221 | int err; |
| 214 | 222 | ||
| 215 | read_lock(&dn_fib_rules_lock); | 223 | rcu_read_lock(); |
| 216 | for(r = dn_fib_rules; r; r = r->r_next) { | 224 | |
| 225 | hlist_for_each_entry_rcu(r, node, &dn_fib_rules, r_hlist) { | ||
| 217 | if (((saddr^r->r_src) & r->r_srcmask) || | 226 | if (((saddr^r->r_src) & r->r_srcmask) || |
| 218 | ((daddr^r->r_dst) & r->r_dstmask) || | 227 | ((daddr^r->r_dst) & r->r_dstmask) || |
| 219 | #ifdef CONFIG_DECNET_ROUTE_FWMARK | 228 | #ifdef CONFIG_DECNET_ROUTE_FWMARK |
| @@ -228,14 +237,14 @@ int dn_fib_lookup(const struct flowi *flp, struct dn_fib_res *res) | |||
| 228 | policy = r; | 237 | policy = r; |
| 229 | break; | 238 | break; |
| 230 | case RTN_UNREACHABLE: | 239 | case RTN_UNREACHABLE: |
| 231 | read_unlock(&dn_fib_rules_lock); | 240 | rcu_read_unlock(); |
| 232 | return -ENETUNREACH; | 241 | return -ENETUNREACH; |
| 233 | default: | 242 | default: |
| 234 | case RTN_BLACKHOLE: | 243 | case RTN_BLACKHOLE: |
| 235 | read_unlock(&dn_fib_rules_lock); | 244 | rcu_read_unlock(); |
| 236 | return -EINVAL; | 245 | return -EINVAL; |
| 237 | case RTN_PROHIBIT: | 246 | case RTN_PROHIBIT: |
| 238 | read_unlock(&dn_fib_rules_lock); | 247 | rcu_read_unlock(); |
| 239 | return -EACCES; | 248 | return -EACCES; |
| 240 | } | 249 | } |
| 241 | 250 | ||
| @@ -246,20 +255,20 @@ int dn_fib_lookup(const struct flowi *flp, struct dn_fib_res *res) | |||
| 246 | res->r = policy; | 255 | res->r = policy; |
| 247 | if (policy) | 256 | if (policy) |
| 248 | atomic_inc(&policy->r_clntref); | 257 | atomic_inc(&policy->r_clntref); |
| 249 | read_unlock(&dn_fib_rules_lock); | 258 | rcu_read_unlock(); |
| 250 | return 0; | 259 | return 0; |
| 251 | } | 260 | } |
| 252 | if (err < 0 && err != -EAGAIN) { | 261 | if (err < 0 && err != -EAGAIN) { |
| 253 | read_unlock(&dn_fib_rules_lock); | 262 | rcu_read_unlock(); |
| 254 | return err; | 263 | return err; |
| 255 | } | 264 | } |
| 256 | } | 265 | } |
| 257 | 266 | ||
| 258 | read_unlock(&dn_fib_rules_lock); | 267 | rcu_read_unlock(); |
| 259 | return -ESRCH; | 268 | return -ESRCH; |
| 260 | } | 269 | } |
| 261 | 270 | ||
| 262 | unsigned dnet_addr_type(__u16 addr) | 271 | unsigned dnet_addr_type(__le16 addr) |
| 263 | { | 272 | { |
| 264 | struct flowi fl = { .nl_u = { .dn_u = { .daddr = addr } } }; | 273 | struct flowi fl = { .nl_u = { .dn_u = { .daddr = addr } } }; |
| 265 | struct dn_fib_res res; | 274 | struct dn_fib_res res; |
| @@ -277,7 +286,7 @@ unsigned dnet_addr_type(__u16 addr) | |||
| 277 | return ret; | 286 | return ret; |
| 278 | } | 287 | } |
| 279 | 288 | ||
| 280 | __u16 dn_fib_rules_policy(__u16 saddr, struct dn_fib_res *res, unsigned *flags) | 289 | __le16 dn_fib_rules_policy(__le16 saddr, struct dn_fib_res *res, unsigned *flags) |
| 281 | { | 290 | { |
| 282 | struct dn_fib_rule *r = res->r; | 291 | struct dn_fib_rule *r = res->r; |
| 283 | 292 | ||
| @@ -297,27 +306,23 @@ __u16 dn_fib_rules_policy(__u16 saddr, struct dn_fib_res *res, unsigned *flags) | |||
| 297 | 306 | ||
| 298 | static void dn_fib_rules_detach(struct net_device *dev) | 307 | static void dn_fib_rules_detach(struct net_device *dev) |
| 299 | { | 308 | { |
| 309 | struct hlist_node *node; | ||
| 300 | struct dn_fib_rule *r; | 310 | struct dn_fib_rule *r; |
| 301 | 311 | ||
| 302 | for(r = dn_fib_rules; r; r = r->r_next) { | 312 | hlist_for_each_entry(r, node, &dn_fib_rules, r_hlist) { |
| 303 | if (r->r_ifindex == dev->ifindex) { | 313 | if (r->r_ifindex == dev->ifindex) |
| 304 | write_lock_bh(&dn_fib_rules_lock); | ||
| 305 | r->r_ifindex = -1; | 314 | r->r_ifindex = -1; |
| 306 | write_unlock_bh(&dn_fib_rules_lock); | ||
| 307 | } | ||
| 308 | } | 315 | } |
| 309 | } | 316 | } |
| 310 | 317 | ||
| 311 | static void dn_fib_rules_attach(struct net_device *dev) | 318 | static void dn_fib_rules_attach(struct net_device *dev) |
| 312 | { | 319 | { |
| 320 | struct hlist_node *node; | ||
| 313 | struct dn_fib_rule *r; | 321 | struct dn_fib_rule *r; |
| 314 | 322 | ||
| 315 | for(r = dn_fib_rules; r; r = r->r_next) { | 323 | hlist_for_each_entry(r, node, &dn_fib_rules, r_hlist) { |
| 316 | if (r->r_ifindex == -1 && strcmp(dev->name, r->r_ifname) == 0) { | 324 | if (r->r_ifindex == -1 && strcmp(dev->name, r->r_ifname) == 0) |
| 317 | write_lock_bh(&dn_fib_rules_lock); | ||
| 318 | r->r_ifindex = dev->ifindex; | 325 | r->r_ifindex = dev->ifindex; |
| 319 | write_unlock_bh(&dn_fib_rules_lock); | ||
| 320 | } | ||
| 321 | } | 326 | } |
| 322 | } | 327 | } |
| 323 | 328 | ||
| @@ -387,18 +392,20 @@ rtattr_failure: | |||
| 387 | 392 | ||
| 388 | int dn_fib_dump_rules(struct sk_buff *skb, struct netlink_callback *cb) | 393 | int dn_fib_dump_rules(struct sk_buff *skb, struct netlink_callback *cb) |
| 389 | { | 394 | { |
| 390 | int idx; | 395 | int idx = 0; |
| 391 | int s_idx = cb->args[0]; | 396 | int s_idx = cb->args[0]; |
| 392 | struct dn_fib_rule *r; | 397 | struct dn_fib_rule *r; |
| 398 | struct hlist_node *node; | ||
| 393 | 399 | ||
| 394 | read_lock(&dn_fib_rules_lock); | 400 | rcu_read_lock(); |
| 395 | for(r = dn_fib_rules, idx = 0; r; r = r->r_next, idx++) { | 401 | hlist_for_each_entry(r, node, &dn_fib_rules, r_hlist) { |
| 396 | if (idx < s_idx) | 402 | if (idx < s_idx) |
| 397 | continue; | 403 | continue; |
| 398 | if (dn_fib_fill_rule(skb, r, cb, NLM_F_MULTI) < 0) | 404 | if (dn_fib_fill_rule(skb, r, cb, NLM_F_MULTI) < 0) |
| 399 | break; | 405 | break; |
| 406 | idx++; | ||
| 400 | } | 407 | } |
| 401 | read_unlock(&dn_fib_rules_lock); | 408 | rcu_read_unlock(); |
| 402 | cb->args[0] = idx; | 409 | cb->args[0] = idx; |
| 403 | 410 | ||
| 404 | return skb->len; | 411 | return skb->len; |
| @@ -406,6 +413,8 @@ int dn_fib_dump_rules(struct sk_buff *skb, struct netlink_callback *cb) | |||
| 406 | 413 | ||
| 407 | void __init dn_fib_rules_init(void) | 414 | void __init dn_fib_rules_init(void) |
| 408 | { | 415 | { |
| 416 | INIT_HLIST_HEAD(&dn_fib_rules); | ||
| 417 | hlist_add_head(&default_rule.r_hlist, &dn_fib_rules); | ||
| 409 | register_netdevice_notifier(&dn_fib_rules_notifier); | 418 | register_netdevice_notifier(&dn_fib_rules_notifier); |
| 410 | } | 419 | } |
| 411 | 420 | ||
