diff options
author | James Morris <james.l.morris@oracle.com> | 2014-11-19 05:32:12 -0500 |
---|---|---|
committer | James Morris <james.l.morris@oracle.com> | 2014-11-19 05:32:12 -0500 |
commit | b10778a00d40b3d9fdaaf5891e802794781ff71c (patch) | |
tree | 6ba4cbac86eecedc3f30650e7f764ecf00c83898 /net/hsr/hsr_device.c | |
parent | 594081ee7145cc30a3977cb4e218f81213b63dc5 (diff) | |
parent | bfe01a5ba2490f299e1d2d5508cbbbadd897bbe9 (diff) |
Merge commit 'v3.17' into next
Diffstat (limited to 'net/hsr/hsr_device.c')
-rw-r--r-- | net/hsr/hsr_device.c | 580 |
1 files changed, 242 insertions, 338 deletions
diff --git a/net/hsr/hsr_device.c b/net/hsr/hsr_device.c index e5302b7f7ca9..a138d75751df 100644 --- a/net/hsr/hsr_device.c +++ b/net/hsr/hsr_device.c | |||
@@ -1,4 +1,4 @@ | |||
1 | /* Copyright 2011-2013 Autronica Fire and Security AS | 1 | /* Copyright 2011-2014 Autronica Fire and Security AS |
2 | * | 2 | * |
3 | * This program is free software; you can redistribute it and/or modify it | 3 | * This program is free software; you can redistribute it and/or modify it |
4 | * under the terms of the GNU General Public License as published by the Free | 4 | * under the terms of the GNU General Public License as published by the Free |
@@ -6,7 +6,7 @@ | |||
6 | * any later version. | 6 | * any later version. |
7 | * | 7 | * |
8 | * Author(s): | 8 | * Author(s): |
9 | * 2011-2013 Arvid Brodin, arvid.brodin@xdin.com | 9 | * 2011-2014 Arvid Brodin, arvid.brodin@alten.se |
10 | * | 10 | * |
11 | * This file contains device methods for creating, using and destroying | 11 | * This file contains device methods for creating, using and destroying |
12 | * virtual HSR devices. | 12 | * virtual HSR devices. |
@@ -15,12 +15,13 @@ | |||
15 | #include <linux/netdevice.h> | 15 | #include <linux/netdevice.h> |
16 | #include <linux/skbuff.h> | 16 | #include <linux/skbuff.h> |
17 | #include <linux/etherdevice.h> | 17 | #include <linux/etherdevice.h> |
18 | #include <linux/if_arp.h> | ||
19 | #include <linux/rtnetlink.h> | 18 | #include <linux/rtnetlink.h> |
20 | #include <linux/pkt_sched.h> | 19 | #include <linux/pkt_sched.h> |
21 | #include "hsr_device.h" | 20 | #include "hsr_device.h" |
21 | #include "hsr_slave.h" | ||
22 | #include "hsr_framereg.h" | 22 | #include "hsr_framereg.h" |
23 | #include "hsr_main.h" | 23 | #include "hsr_main.h" |
24 | #include "hsr_forward.h" | ||
24 | 25 | ||
25 | 26 | ||
26 | static bool is_admin_up(struct net_device *dev) | 27 | static bool is_admin_up(struct net_device *dev) |
@@ -45,75 +46,108 @@ static void __hsr_set_operstate(struct net_device *dev, int transition) | |||
45 | } | 46 | } |
46 | } | 47 | } |
47 | 48 | ||
48 | void hsr_set_operstate(struct net_device *hsr_dev, struct net_device *slave1, | 49 | static void hsr_set_operstate(struct hsr_port *master, bool has_carrier) |
49 | struct net_device *slave2) | ||
50 | { | 50 | { |
51 | if (!is_admin_up(hsr_dev)) { | 51 | if (!is_admin_up(master->dev)) { |
52 | __hsr_set_operstate(hsr_dev, IF_OPER_DOWN); | 52 | __hsr_set_operstate(master->dev, IF_OPER_DOWN); |
53 | return; | 53 | return; |
54 | } | 54 | } |
55 | 55 | ||
56 | if (is_slave_up(slave1) || is_slave_up(slave2)) | 56 | if (has_carrier) |
57 | __hsr_set_operstate(hsr_dev, IF_OPER_UP); | 57 | __hsr_set_operstate(master->dev, IF_OPER_UP); |
58 | else | 58 | else |
59 | __hsr_set_operstate(hsr_dev, IF_OPER_LOWERLAYERDOWN); | 59 | __hsr_set_operstate(master->dev, IF_OPER_LOWERLAYERDOWN); |
60 | } | 60 | } |
61 | 61 | ||
62 | void hsr_set_carrier(struct net_device *hsr_dev, struct net_device *slave1, | 62 | static bool hsr_check_carrier(struct hsr_port *master) |
63 | struct net_device *slave2) | ||
64 | { | 63 | { |
65 | if (is_slave_up(slave1) || is_slave_up(slave2)) | 64 | struct hsr_port *port; |
66 | netif_carrier_on(hsr_dev); | 65 | bool has_carrier; |
66 | |||
67 | has_carrier = false; | ||
68 | |||
69 | rcu_read_lock(); | ||
70 | hsr_for_each_port(master->hsr, port) | ||
71 | if ((port->type != HSR_PT_MASTER) && is_slave_up(port->dev)) { | ||
72 | has_carrier = true; | ||
73 | break; | ||
74 | } | ||
75 | rcu_read_unlock(); | ||
76 | |||
77 | if (has_carrier) | ||
78 | netif_carrier_on(master->dev); | ||
67 | else | 79 | else |
68 | netif_carrier_off(hsr_dev); | 80 | netif_carrier_off(master->dev); |
81 | |||
82 | return has_carrier; | ||
69 | } | 83 | } |
70 | 84 | ||
71 | 85 | ||
72 | void hsr_check_announce(struct net_device *hsr_dev, int old_operstate) | 86 | static void hsr_check_announce(struct net_device *hsr_dev, |
87 | unsigned char old_operstate) | ||
73 | { | 88 | { |
74 | struct hsr_priv *hsr_priv; | 89 | struct hsr_priv *hsr; |
75 | 90 | ||
76 | hsr_priv = netdev_priv(hsr_dev); | 91 | hsr = netdev_priv(hsr_dev); |
77 | 92 | ||
78 | if ((hsr_dev->operstate == IF_OPER_UP) && (old_operstate != IF_OPER_UP)) { | 93 | if ((hsr_dev->operstate == IF_OPER_UP) && (old_operstate != IF_OPER_UP)) { |
79 | /* Went up */ | 94 | /* Went up */ |
80 | hsr_priv->announce_count = 0; | 95 | hsr->announce_count = 0; |
81 | hsr_priv->announce_timer.expires = jiffies + | 96 | hsr->announce_timer.expires = jiffies + |
82 | msecs_to_jiffies(HSR_ANNOUNCE_INTERVAL); | 97 | msecs_to_jiffies(HSR_ANNOUNCE_INTERVAL); |
83 | add_timer(&hsr_priv->announce_timer); | 98 | add_timer(&hsr->announce_timer); |
84 | } | 99 | } |
85 | 100 | ||
86 | if ((hsr_dev->operstate != IF_OPER_UP) && (old_operstate == IF_OPER_UP)) | 101 | if ((hsr_dev->operstate != IF_OPER_UP) && (old_operstate == IF_OPER_UP)) |
87 | /* Went down */ | 102 | /* Went down */ |
88 | del_timer(&hsr_priv->announce_timer); | 103 | del_timer(&hsr->announce_timer); |
89 | } | 104 | } |
90 | 105 | ||
91 | 106 | void hsr_check_carrier_and_operstate(struct hsr_priv *hsr) | |
92 | int hsr_get_max_mtu(struct hsr_priv *hsr_priv) | ||
93 | { | 107 | { |
94 | int mtu_max; | 108 | struct hsr_port *master; |
95 | 109 | unsigned char old_operstate; | |
96 | if (hsr_priv->slave[0] && hsr_priv->slave[1]) | 110 | bool has_carrier; |
97 | mtu_max = min(hsr_priv->slave[0]->mtu, hsr_priv->slave[1]->mtu); | ||
98 | else if (hsr_priv->slave[0]) | ||
99 | mtu_max = hsr_priv->slave[0]->mtu; | ||
100 | else if (hsr_priv->slave[1]) | ||
101 | mtu_max = hsr_priv->slave[1]->mtu; | ||
102 | else | ||
103 | mtu_max = HSR_TAGLEN; | ||
104 | 111 | ||
105 | return mtu_max - HSR_TAGLEN; | 112 | master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); |
113 | /* netif_stacked_transfer_operstate() cannot be used here since | ||
114 | * it doesn't set IF_OPER_LOWERLAYERDOWN (?) | ||
115 | */ | ||
116 | old_operstate = master->dev->operstate; | ||
117 | has_carrier = hsr_check_carrier(master); | ||
118 | hsr_set_operstate(master, has_carrier); | ||
119 | hsr_check_announce(master->dev, old_operstate); | ||
106 | } | 120 | } |
107 | 121 | ||
122 | int hsr_get_max_mtu(struct hsr_priv *hsr) | ||
123 | { | ||
124 | unsigned int mtu_max; | ||
125 | struct hsr_port *port; | ||
126 | |||
127 | mtu_max = ETH_DATA_LEN; | ||
128 | rcu_read_lock(); | ||
129 | hsr_for_each_port(hsr, port) | ||
130 | if (port->type != HSR_PT_MASTER) | ||
131 | mtu_max = min(port->dev->mtu, mtu_max); | ||
132 | rcu_read_unlock(); | ||
133 | |||
134 | if (mtu_max < HSR_HLEN) | ||
135 | return 0; | ||
136 | return mtu_max - HSR_HLEN; | ||
137 | } | ||
138 | |||
139 | |||
108 | static int hsr_dev_change_mtu(struct net_device *dev, int new_mtu) | 140 | static int hsr_dev_change_mtu(struct net_device *dev, int new_mtu) |
109 | { | 141 | { |
110 | struct hsr_priv *hsr_priv; | 142 | struct hsr_priv *hsr; |
143 | struct hsr_port *master; | ||
111 | 144 | ||
112 | hsr_priv = netdev_priv(dev); | 145 | hsr = netdev_priv(dev); |
146 | master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); | ||
113 | 147 | ||
114 | if (new_mtu > hsr_get_max_mtu(hsr_priv)) { | 148 | if (new_mtu > hsr_get_max_mtu(hsr)) { |
115 | netdev_info(hsr_priv->dev, "A HSR master's MTU cannot be greater than the smallest MTU of its slaves minus the HSR Tag length (%d octets).\n", | 149 | netdev_info(master->dev, "A HSR master's MTU cannot be greater than the smallest MTU of its slaves minus the HSR Tag length (%d octets).\n", |
116 | HSR_TAGLEN); | 150 | HSR_HLEN); |
117 | return -EINVAL; | 151 | return -EINVAL; |
118 | } | 152 | } |
119 | 153 | ||
@@ -124,164 +158,95 @@ static int hsr_dev_change_mtu(struct net_device *dev, int new_mtu) | |||
124 | 158 | ||
125 | static int hsr_dev_open(struct net_device *dev) | 159 | static int hsr_dev_open(struct net_device *dev) |
126 | { | 160 | { |
127 | struct hsr_priv *hsr_priv; | 161 | struct hsr_priv *hsr; |
128 | int i; | 162 | struct hsr_port *port; |
129 | char *slave_name; | 163 | char designation; |
130 | 164 | ||
131 | hsr_priv = netdev_priv(dev); | 165 | hsr = netdev_priv(dev); |
166 | designation = '\0'; | ||
132 | 167 | ||
133 | for (i = 0; i < HSR_MAX_SLAVE; i++) { | 168 | rcu_read_lock(); |
134 | if (hsr_priv->slave[i]) | 169 | hsr_for_each_port(hsr, port) { |
135 | slave_name = hsr_priv->slave[i]->name; | 170 | if (port->type == HSR_PT_MASTER) |
136 | else | 171 | continue; |
137 | slave_name = "null"; | 172 | switch (port->type) { |
138 | 173 | case HSR_PT_SLAVE_A: | |
139 | if (!is_slave_up(hsr_priv->slave[i])) | 174 | designation = 'A'; |
140 | netdev_warn(dev, "Slave %c (%s) is not up; please bring it up to get a working HSR network\n", | 175 | break; |
141 | 'A' + i, slave_name); | 176 | case HSR_PT_SLAVE_B: |
177 | designation = 'B'; | ||
178 | break; | ||
179 | default: | ||
180 | designation = '?'; | ||
181 | } | ||
182 | if (!is_slave_up(port->dev)) | ||
183 | netdev_warn(dev, "Slave %c (%s) is not up; please bring it up to get a fully working HSR network\n", | ||
184 | designation, port->dev->name); | ||
142 | } | 185 | } |
186 | rcu_read_unlock(); | ||
187 | |||
188 | if (designation == '\0') | ||
189 | netdev_warn(dev, "No slave devices configured\n"); | ||
143 | 190 | ||
144 | return 0; | 191 | return 0; |
145 | } | 192 | } |
146 | 193 | ||
194 | |||
147 | static int hsr_dev_close(struct net_device *dev) | 195 | static int hsr_dev_close(struct net_device *dev) |
148 | { | 196 | { |
149 | /* Nothing to do here. We could try to restore the state of the slaves | 197 | /* Nothing to do here. */ |
150 | * to what they were before being changed by the hsr master dev's state, | ||
151 | * but they might have been changed manually in the mean time too, so | ||
152 | * taking them up or down here might be confusing and is probably not a | ||
153 | * good idea. | ||
154 | */ | ||
155 | return 0; | 198 | return 0; |
156 | } | 199 | } |
157 | 200 | ||
158 | 201 | ||
159 | static void hsr_fill_tag(struct hsr_ethhdr *hsr_ethhdr, struct hsr_priv *hsr_priv) | 202 | static netdev_features_t hsr_features_recompute(struct hsr_priv *hsr, |
203 | netdev_features_t features) | ||
160 | { | 204 | { |
161 | unsigned long irqflags; | 205 | netdev_features_t mask; |
206 | struct hsr_port *port; | ||
162 | 207 | ||
163 | /* IEC 62439-1:2010, p 48, says the 4-bit "path" field can take values | 208 | mask = features; |
164 | * between 0001-1001 ("ring identifier", for regular HSR frames), | ||
165 | * or 1111 ("HSR management", supervision frames). Unfortunately, the | ||
166 | * spec writers forgot to explain what a "ring identifier" is, or | ||
167 | * how it is used. So we just set this to 0001 for regular frames, | ||
168 | * and 1111 for supervision frames. | ||
169 | */ | ||
170 | set_hsr_tag_path(&hsr_ethhdr->hsr_tag, 0x1); | ||
171 | 209 | ||
172 | /* IEC 62439-1:2010, p 12: "The link service data unit in an Ethernet | 210 | /* Mask out all features that, if supported by one device, should be |
173 | * frame is the content of the frame located between the Length/Type | 211 | * enabled for all devices (see NETIF_F_ONE_FOR_ALL). |
174 | * field and the Frame Check Sequence." | ||
175 | * | 212 | * |
176 | * IEC 62439-3, p 48, specifies the "original LPDU" to include the | 213 | * Anything that's off in mask will not be enabled - so only things |
177 | * original "LT" field (what "LT" means is not explained anywhere as | 214 | * that were in features originally, and also is in NETIF_F_ONE_FOR_ALL, |
178 | * far as I can see - perhaps "Length/Type"?). So LSDU_size might | 215 | * may become enabled. |
179 | * equal original length + 2. | ||
180 | * Also, the fact that this field is not used anywhere (might be used | ||
181 | * by a RedBox connecting HSR and PRP nets?) means I cannot test its | ||
182 | * correctness. Instead of guessing, I set this to 0 here, to make any | ||
183 | * problems immediately apparent. Anyone using this driver with PRP/HSR | ||
184 | * RedBoxes might need to fix this... | ||
185 | */ | 216 | */ |
186 | set_hsr_tag_LSDU_size(&hsr_ethhdr->hsr_tag, 0); | 217 | features &= ~NETIF_F_ONE_FOR_ALL; |
187 | 218 | hsr_for_each_port(hsr, port) | |
188 | spin_lock_irqsave(&hsr_priv->seqnr_lock, irqflags); | 219 | features = netdev_increment_features(features, |
189 | hsr_ethhdr->hsr_tag.sequence_nr = htons(hsr_priv->sequence_nr); | 220 | port->dev->features, |
190 | hsr_priv->sequence_nr++; | 221 | mask); |
191 | spin_unlock_irqrestore(&hsr_priv->seqnr_lock, irqflags); | ||
192 | 222 | ||
193 | hsr_ethhdr->hsr_tag.encap_proto = hsr_ethhdr->ethhdr.h_proto; | 223 | return features; |
194 | |||
195 | hsr_ethhdr->ethhdr.h_proto = htons(ETH_P_PRP); | ||
196 | } | 224 | } |
197 | 225 | ||
198 | static int slave_xmit(struct sk_buff *skb, struct hsr_priv *hsr_priv, | 226 | static netdev_features_t hsr_fix_features(struct net_device *dev, |
199 | enum hsr_dev_idx dev_idx) | 227 | netdev_features_t features) |
200 | { | 228 | { |
201 | struct hsr_ethhdr *hsr_ethhdr; | 229 | struct hsr_priv *hsr = netdev_priv(dev); |
202 | |||
203 | hsr_ethhdr = (struct hsr_ethhdr *) skb->data; | ||
204 | 230 | ||
205 | skb->dev = hsr_priv->slave[dev_idx]; | 231 | return hsr_features_recompute(hsr, features); |
206 | |||
207 | hsr_addr_subst_dest(hsr_priv, &hsr_ethhdr->ethhdr, dev_idx); | ||
208 | |||
209 | /* Address substitution (IEC62439-3 pp 26, 50): replace mac | ||
210 | * address of outgoing frame with that of the outgoing slave's. | ||
211 | */ | ||
212 | ether_addr_copy(hsr_ethhdr->ethhdr.h_source, skb->dev->dev_addr); | ||
213 | |||
214 | return dev_queue_xmit(skb); | ||
215 | } | 232 | } |
216 | 233 | ||
217 | 234 | ||
218 | static int hsr_dev_xmit(struct sk_buff *skb, struct net_device *dev) | 235 | static int hsr_dev_xmit(struct sk_buff *skb, struct net_device *dev) |
219 | { | 236 | { |
220 | struct hsr_priv *hsr_priv; | 237 | struct hsr_priv *hsr = netdev_priv(dev); |
221 | struct hsr_ethhdr *hsr_ethhdr; | 238 | struct hsr_port *master; |
222 | struct sk_buff *skb2; | ||
223 | int res1, res2; | ||
224 | |||
225 | hsr_priv = netdev_priv(dev); | ||
226 | hsr_ethhdr = (struct hsr_ethhdr *) skb->data; | ||
227 | |||
228 | if ((skb->protocol != htons(ETH_P_PRP)) || | ||
229 | (hsr_ethhdr->ethhdr.h_proto != htons(ETH_P_PRP))) { | ||
230 | hsr_fill_tag(hsr_ethhdr, hsr_priv); | ||
231 | skb->protocol = htons(ETH_P_PRP); | ||
232 | } | ||
233 | |||
234 | skb2 = pskb_copy(skb, GFP_ATOMIC); | ||
235 | |||
236 | res1 = NET_XMIT_DROP; | ||
237 | if (likely(hsr_priv->slave[HSR_DEV_SLAVE_A])) | ||
238 | res1 = slave_xmit(skb, hsr_priv, HSR_DEV_SLAVE_A); | ||
239 | 239 | ||
240 | res2 = NET_XMIT_DROP; | 240 | master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); |
241 | if (likely(skb2 && hsr_priv->slave[HSR_DEV_SLAVE_B])) | 241 | skb->dev = master->dev; |
242 | res2 = slave_xmit(skb2, hsr_priv, HSR_DEV_SLAVE_B); | 242 | hsr_forward_skb(skb, master); |
243 | |||
244 | if (likely(res1 == NET_XMIT_SUCCESS || res1 == NET_XMIT_CN || | ||
245 | res2 == NET_XMIT_SUCCESS || res2 == NET_XMIT_CN)) { | ||
246 | hsr_priv->dev->stats.tx_packets++; | ||
247 | hsr_priv->dev->stats.tx_bytes += skb->len; | ||
248 | } else { | ||
249 | hsr_priv->dev->stats.tx_dropped++; | ||
250 | } | ||
251 | 243 | ||
252 | return NETDEV_TX_OK; | 244 | return NETDEV_TX_OK; |
253 | } | 245 | } |
254 | 246 | ||
255 | 247 | ||
256 | static int hsr_header_create(struct sk_buff *skb, struct net_device *dev, | ||
257 | unsigned short type, const void *daddr, | ||
258 | const void *saddr, unsigned int len) | ||
259 | { | ||
260 | int res; | ||
261 | |||
262 | /* Make room for the HSR tag now. We will fill it in later (in | ||
263 | * hsr_dev_xmit) | ||
264 | */ | ||
265 | if (skb_headroom(skb) < HSR_TAGLEN + ETH_HLEN) | ||
266 | return -ENOBUFS; | ||
267 | skb_push(skb, HSR_TAGLEN); | ||
268 | |||
269 | /* To allow VLAN/HSR combos we should probably use | ||
270 | * res = dev_hard_header(skb, dev, type, daddr, saddr, len + HSR_TAGLEN); | ||
271 | * here instead. It would require other changes too, though - e.g. | ||
272 | * separate headers for each slave etc... | ||
273 | */ | ||
274 | res = eth_header(skb, dev, type, daddr, saddr, len + HSR_TAGLEN); | ||
275 | if (res <= 0) | ||
276 | return res; | ||
277 | skb_reset_mac_header(skb); | ||
278 | |||
279 | return res + HSR_TAGLEN; | ||
280 | } | ||
281 | |||
282 | |||
283 | static const struct header_ops hsr_header_ops = { | 248 | static const struct header_ops hsr_header_ops = { |
284 | .create = hsr_header_create, | 249 | .create = eth_header, |
285 | .parse = eth_header_parse, | 250 | .parse = eth_header_parse, |
286 | }; | 251 | }; |
287 | 252 | ||
@@ -291,67 +256,63 @@ static const struct header_ops hsr_header_ops = { | |||
291 | */ | 256 | */ |
292 | static int hsr_pad(int size) | 257 | static int hsr_pad(int size) |
293 | { | 258 | { |
294 | const int min_size = ETH_ZLEN - HSR_TAGLEN - ETH_HLEN; | 259 | const int min_size = ETH_ZLEN - HSR_HLEN - ETH_HLEN; |
295 | 260 | ||
296 | if (size >= min_size) | 261 | if (size >= min_size) |
297 | return size; | 262 | return size; |
298 | return min_size; | 263 | return min_size; |
299 | } | 264 | } |
300 | 265 | ||
301 | static void send_hsr_supervision_frame(struct net_device *hsr_dev, u8 type) | 266 | static void send_hsr_supervision_frame(struct hsr_port *master, u8 type) |
302 | { | 267 | { |
303 | struct hsr_priv *hsr_priv; | ||
304 | struct sk_buff *skb; | 268 | struct sk_buff *skb; |
305 | int hlen, tlen; | 269 | int hlen, tlen; |
306 | struct hsr_sup_tag *hsr_stag; | 270 | struct hsr_sup_tag *hsr_stag; |
307 | struct hsr_sup_payload *hsr_sp; | 271 | struct hsr_sup_payload *hsr_sp; |
308 | unsigned long irqflags; | 272 | unsigned long irqflags; |
309 | 273 | ||
310 | hlen = LL_RESERVED_SPACE(hsr_dev); | 274 | hlen = LL_RESERVED_SPACE(master->dev); |
311 | tlen = hsr_dev->needed_tailroom; | 275 | tlen = master->dev->needed_tailroom; |
312 | skb = alloc_skb(hsr_pad(sizeof(struct hsr_sup_payload)) + hlen + tlen, | 276 | skb = alloc_skb(hsr_pad(sizeof(struct hsr_sup_payload)) + hlen + tlen, |
313 | GFP_ATOMIC); | 277 | GFP_ATOMIC); |
314 | 278 | ||
315 | if (skb == NULL) | 279 | if (skb == NULL) |
316 | return; | 280 | return; |
317 | 281 | ||
318 | hsr_priv = netdev_priv(hsr_dev); | ||
319 | |||
320 | skb_reserve(skb, hlen); | 282 | skb_reserve(skb, hlen); |
321 | 283 | ||
322 | skb->dev = hsr_dev; | 284 | skb->dev = master->dev; |
323 | skb->protocol = htons(ETH_P_PRP); | 285 | skb->protocol = htons(ETH_P_PRP); |
324 | skb->priority = TC_PRIO_CONTROL; | 286 | skb->priority = TC_PRIO_CONTROL; |
325 | 287 | ||
326 | if (dev_hard_header(skb, skb->dev, ETH_P_PRP, | 288 | if (dev_hard_header(skb, skb->dev, ETH_P_PRP, |
327 | hsr_priv->sup_multicast_addr, | 289 | master->hsr->sup_multicast_addr, |
328 | skb->dev->dev_addr, skb->len) < 0) | 290 | skb->dev->dev_addr, skb->len) <= 0) |
329 | goto out; | 291 | goto out; |
292 | skb_reset_mac_header(skb); | ||
330 | 293 | ||
331 | skb_pull(skb, sizeof(struct ethhdr)); | 294 | hsr_stag = (typeof(hsr_stag)) skb_put(skb, sizeof(*hsr_stag)); |
332 | hsr_stag = (typeof(hsr_stag)) skb->data; | ||
333 | 295 | ||
334 | set_hsr_stag_path(hsr_stag, 0xf); | 296 | set_hsr_stag_path(hsr_stag, 0xf); |
335 | set_hsr_stag_HSR_Ver(hsr_stag, 0); | 297 | set_hsr_stag_HSR_Ver(hsr_stag, 0); |
336 | 298 | ||
337 | spin_lock_irqsave(&hsr_priv->seqnr_lock, irqflags); | 299 | spin_lock_irqsave(&master->hsr->seqnr_lock, irqflags); |
338 | hsr_stag->sequence_nr = htons(hsr_priv->sequence_nr); | 300 | hsr_stag->sequence_nr = htons(master->hsr->sequence_nr); |
339 | hsr_priv->sequence_nr++; | 301 | master->hsr->sequence_nr++; |
340 | spin_unlock_irqrestore(&hsr_priv->seqnr_lock, irqflags); | 302 | spin_unlock_irqrestore(&master->hsr->seqnr_lock, irqflags); |
341 | 303 | ||
342 | hsr_stag->HSR_TLV_Type = type; | 304 | hsr_stag->HSR_TLV_Type = type; |
343 | hsr_stag->HSR_TLV_Length = 12; | 305 | hsr_stag->HSR_TLV_Length = 12; |
344 | 306 | ||
345 | skb_push(skb, sizeof(struct ethhdr)); | ||
346 | |||
347 | /* Payload: MacAddressA */ | 307 | /* Payload: MacAddressA */ |
348 | hsr_sp = (typeof(hsr_sp)) skb_put(skb, sizeof(*hsr_sp)); | 308 | hsr_sp = (typeof(hsr_sp)) skb_put(skb, sizeof(*hsr_sp)); |
349 | ether_addr_copy(hsr_sp->MacAddressA, hsr_dev->dev_addr); | 309 | ether_addr_copy(hsr_sp->MacAddressA, master->dev->dev_addr); |
350 | 310 | ||
351 | dev_queue_xmit(skb); | 311 | hsr_forward_skb(skb, master); |
352 | return; | 312 | return; |
353 | 313 | ||
354 | out: | 314 | out: |
315 | WARN_ON_ONCE("HSR: Could not send supervision frame\n"); | ||
355 | kfree_skb(skb); | 316 | kfree_skb(skb); |
356 | } | 317 | } |
357 | 318 | ||
@@ -360,59 +321,32 @@ out: | |||
360 | */ | 321 | */ |
361 | static void hsr_announce(unsigned long data) | 322 | static void hsr_announce(unsigned long data) |
362 | { | 323 | { |
363 | struct hsr_priv *hsr_priv; | 324 | struct hsr_priv *hsr; |
325 | struct hsr_port *master; | ||
364 | 326 | ||
365 | hsr_priv = (struct hsr_priv *) data; | 327 | hsr = (struct hsr_priv *) data; |
366 | 328 | ||
367 | if (hsr_priv->announce_count < 3) { | 329 | rcu_read_lock(); |
368 | send_hsr_supervision_frame(hsr_priv->dev, HSR_TLV_ANNOUNCE); | 330 | master = hsr_port_get_hsr(hsr, HSR_PT_MASTER); |
369 | hsr_priv->announce_count++; | 331 | |
332 | if (hsr->announce_count < 3) { | ||
333 | send_hsr_supervision_frame(master, HSR_TLV_ANNOUNCE); | ||
334 | hsr->announce_count++; | ||
370 | } else { | 335 | } else { |
371 | send_hsr_supervision_frame(hsr_priv->dev, HSR_TLV_LIFE_CHECK); | 336 | send_hsr_supervision_frame(master, HSR_TLV_LIFE_CHECK); |
372 | } | 337 | } |
373 | 338 | ||
374 | if (hsr_priv->announce_count < 3) | 339 | if (hsr->announce_count < 3) |
375 | hsr_priv->announce_timer.expires = jiffies + | 340 | hsr->announce_timer.expires = jiffies + |
376 | msecs_to_jiffies(HSR_ANNOUNCE_INTERVAL); | 341 | msecs_to_jiffies(HSR_ANNOUNCE_INTERVAL); |
377 | else | 342 | else |
378 | hsr_priv->announce_timer.expires = jiffies + | 343 | hsr->announce_timer.expires = jiffies + |
379 | msecs_to_jiffies(HSR_LIFE_CHECK_INTERVAL); | 344 | msecs_to_jiffies(HSR_LIFE_CHECK_INTERVAL); |
380 | 345 | ||
381 | if (is_admin_up(hsr_priv->dev)) | 346 | if (is_admin_up(master->dev)) |
382 | add_timer(&hsr_priv->announce_timer); | 347 | add_timer(&hsr->announce_timer); |
383 | } | ||
384 | |||
385 | |||
386 | static void restore_slaves(struct net_device *hsr_dev) | ||
387 | { | ||
388 | struct hsr_priv *hsr_priv; | ||
389 | int i; | ||
390 | int res; | ||
391 | |||
392 | hsr_priv = netdev_priv(hsr_dev); | ||
393 | |||
394 | rtnl_lock(); | ||
395 | |||
396 | /* Restore promiscuity */ | ||
397 | for (i = 0; i < HSR_MAX_SLAVE; i++) { | ||
398 | if (!hsr_priv->slave[i]) | ||
399 | continue; | ||
400 | res = dev_set_promiscuity(hsr_priv->slave[i], -1); | ||
401 | if (res) | ||
402 | netdev_info(hsr_dev, | ||
403 | "Cannot restore slave promiscuity (%s, %d)\n", | ||
404 | hsr_priv->slave[i]->name, res); | ||
405 | } | ||
406 | |||
407 | rtnl_unlock(); | ||
408 | } | ||
409 | |||
410 | static void reclaim_hsr_dev(struct rcu_head *rh) | ||
411 | { | ||
412 | struct hsr_priv *hsr_priv; | ||
413 | 348 | ||
414 | hsr_priv = container_of(rh, struct hsr_priv, rcu_head); | 349 | rcu_read_unlock(); |
415 | free_netdev(hsr_priv->dev); | ||
416 | } | 350 | } |
417 | 351 | ||
418 | 352 | ||
@@ -421,14 +355,18 @@ static void reclaim_hsr_dev(struct rcu_head *rh) | |||
421 | */ | 355 | */ |
422 | static void hsr_dev_destroy(struct net_device *hsr_dev) | 356 | static void hsr_dev_destroy(struct net_device *hsr_dev) |
423 | { | 357 | { |
424 | struct hsr_priv *hsr_priv; | 358 | struct hsr_priv *hsr; |
359 | struct hsr_port *port; | ||
425 | 360 | ||
426 | hsr_priv = netdev_priv(hsr_dev); | 361 | hsr = netdev_priv(hsr_dev); |
362 | hsr_for_each_port(hsr, port) | ||
363 | hsr_del_port(port); | ||
427 | 364 | ||
428 | del_timer(&hsr_priv->announce_timer); | 365 | del_timer_sync(&hsr->prune_timer); |
429 | unregister_hsr_master(hsr_priv); /* calls list_del_rcu on hsr_priv */ | 366 | del_timer_sync(&hsr->announce_timer); |
430 | restore_slaves(hsr_dev); | 367 | |
431 | call_rcu(&hsr_priv->rcu_head, reclaim_hsr_dev); /* reclaim hsr_priv */ | 368 | synchronize_rcu(); |
369 | free_netdev(hsr_dev); | ||
432 | } | 370 | } |
433 | 371 | ||
434 | static const struct net_device_ops hsr_device_ops = { | 372 | static const struct net_device_ops hsr_device_ops = { |
@@ -436,62 +374,51 @@ static const struct net_device_ops hsr_device_ops = { | |||
436 | .ndo_open = hsr_dev_open, | 374 | .ndo_open = hsr_dev_open, |
437 | .ndo_stop = hsr_dev_close, | 375 | .ndo_stop = hsr_dev_close, |
438 | .ndo_start_xmit = hsr_dev_xmit, | 376 | .ndo_start_xmit = hsr_dev_xmit, |
377 | .ndo_fix_features = hsr_fix_features, | ||
439 | }; | 378 | }; |
440 | 379 | ||
380 | static struct device_type hsr_type = { | ||
381 | .name = "hsr", | ||
382 | }; | ||
441 | 383 | ||
442 | void hsr_dev_setup(struct net_device *dev) | 384 | void hsr_dev_setup(struct net_device *dev) |
443 | { | 385 | { |
444 | random_ether_addr(dev->dev_addr); | 386 | random_ether_addr(dev->dev_addr); |
445 | 387 | ||
446 | ether_setup(dev); | 388 | ether_setup(dev); |
447 | dev->header_ops = &hsr_header_ops; | 389 | dev->header_ops = &hsr_header_ops; |
448 | dev->netdev_ops = &hsr_device_ops; | 390 | dev->netdev_ops = &hsr_device_ops; |
449 | dev->tx_queue_len = 0; | 391 | SET_NETDEV_DEVTYPE(dev, &hsr_type); |
392 | dev->tx_queue_len = 0; | ||
450 | 393 | ||
451 | dev->destructor = hsr_dev_destroy; | 394 | dev->destructor = hsr_dev_destroy; |
395 | |||
396 | dev->hw_features = NETIF_F_SG | NETIF_F_FRAGLIST | NETIF_F_HIGHDMA | | ||
397 | NETIF_F_GSO_MASK | NETIF_F_HW_CSUM | | ||
398 | NETIF_F_HW_VLAN_CTAG_TX; | ||
399 | |||
400 | dev->features = dev->hw_features; | ||
401 | |||
402 | /* Prevent recursive tx locking */ | ||
403 | dev->features |= NETIF_F_LLTX; | ||
404 | /* VLAN on top of HSR needs testing and probably some work on | ||
405 | * hsr_header_create() etc. | ||
406 | */ | ||
407 | dev->features |= NETIF_F_VLAN_CHALLENGED; | ||
408 | /* Not sure about this. Taken from bridge code. netdev_features.h says | ||
409 | * it means "Does not change network namespaces". | ||
410 | */ | ||
411 | dev->features |= NETIF_F_NETNS_LOCAL; | ||
452 | } | 412 | } |
453 | 413 | ||
454 | 414 | ||
455 | /* Return true if dev is a HSR master; return false otherwise. | 415 | /* Return true if dev is a HSR master; return false otherwise. |
456 | */ | 416 | */ |
457 | bool is_hsr_master(struct net_device *dev) | 417 | inline bool is_hsr_master(struct net_device *dev) |
458 | { | 418 | { |
459 | return (dev->netdev_ops->ndo_start_xmit == hsr_dev_xmit); | 419 | return (dev->netdev_ops->ndo_start_xmit == hsr_dev_xmit); |
460 | } | 420 | } |
461 | 421 | ||
462 | static int check_slave_ok(struct net_device *dev) | ||
463 | { | ||
464 | /* Don't allow HSR on non-ethernet like devices */ | ||
465 | if ((dev->flags & IFF_LOOPBACK) || (dev->type != ARPHRD_ETHER) || | ||
466 | (dev->addr_len != ETH_ALEN)) { | ||
467 | netdev_info(dev, "Cannot use loopback or non-ethernet device as HSR slave.\n"); | ||
468 | return -EINVAL; | ||
469 | } | ||
470 | |||
471 | /* Don't allow enslaving hsr devices */ | ||
472 | if (is_hsr_master(dev)) { | ||
473 | netdev_info(dev, "Cannot create trees of HSR devices.\n"); | ||
474 | return -EINVAL; | ||
475 | } | ||
476 | |||
477 | if (is_hsr_slave(dev)) { | ||
478 | netdev_info(dev, "This device is already a HSR slave.\n"); | ||
479 | return -EINVAL; | ||
480 | } | ||
481 | |||
482 | if (dev->priv_flags & IFF_802_1Q_VLAN) { | ||
483 | netdev_info(dev, "HSR on top of VLAN is not yet supported in this driver.\n"); | ||
484 | return -EINVAL; | ||
485 | } | ||
486 | |||
487 | /* HSR over bonded devices has not been tested, but I'm not sure it | ||
488 | * won't work... | ||
489 | */ | ||
490 | |||
491 | return 0; | ||
492 | } | ||
493 | |||
494 | |||
495 | /* Default multicast address for HSR Supervision frames */ | 422 | /* Default multicast address for HSR Supervision frames */ |
496 | static const unsigned char def_multicast_addr[ETH_ALEN] __aligned(2) = { | 423 | static const unsigned char def_multicast_addr[ETH_ALEN] __aligned(2) = { |
497 | 0x01, 0x15, 0x4e, 0x00, 0x01, 0x00 | 424 | 0x01, 0x15, 0x4e, 0x00, 0x01, 0x00 |
@@ -500,97 +427,74 @@ static const unsigned char def_multicast_addr[ETH_ALEN] __aligned(2) = { | |||
500 | int hsr_dev_finalize(struct net_device *hsr_dev, struct net_device *slave[2], | 427 | int hsr_dev_finalize(struct net_device *hsr_dev, struct net_device *slave[2], |
501 | unsigned char multicast_spec) | 428 | unsigned char multicast_spec) |
502 | { | 429 | { |
503 | struct hsr_priv *hsr_priv; | 430 | struct hsr_priv *hsr; |
504 | int i; | 431 | struct hsr_port *port; |
505 | int res; | 432 | int res; |
506 | 433 | ||
507 | hsr_priv = netdev_priv(hsr_dev); | 434 | hsr = netdev_priv(hsr_dev); |
508 | hsr_priv->dev = hsr_dev; | 435 | INIT_LIST_HEAD(&hsr->ports); |
509 | INIT_LIST_HEAD(&hsr_priv->node_db); | 436 | INIT_LIST_HEAD(&hsr->node_db); |
510 | INIT_LIST_HEAD(&hsr_priv->self_node_db); | 437 | INIT_LIST_HEAD(&hsr->self_node_db); |
511 | for (i = 0; i < HSR_MAX_SLAVE; i++) | ||
512 | hsr_priv->slave[i] = slave[i]; | ||
513 | |||
514 | spin_lock_init(&hsr_priv->seqnr_lock); | ||
515 | /* Overflow soon to find bugs easier: */ | ||
516 | hsr_priv->sequence_nr = USHRT_MAX - 1024; | ||
517 | |||
518 | init_timer(&hsr_priv->announce_timer); | ||
519 | hsr_priv->announce_timer.function = hsr_announce; | ||
520 | hsr_priv->announce_timer.data = (unsigned long) hsr_priv; | ||
521 | 438 | ||
522 | ether_addr_copy(hsr_priv->sup_multicast_addr, def_multicast_addr); | 439 | ether_addr_copy(hsr_dev->dev_addr, slave[0]->dev_addr); |
523 | hsr_priv->sup_multicast_addr[ETH_ALEN - 1] = multicast_spec; | ||
524 | 440 | ||
525 | /* FIXME: should I modify the value of these? | 441 | /* Make sure we recognize frames from ourselves in hsr_rcv() */ |
526 | * | 442 | res = hsr_create_self_node(&hsr->self_node_db, hsr_dev->dev_addr, |
527 | * - hsr_dev->flags - i.e. | 443 | slave[1]->dev_addr); |
528 | * IFF_MASTER/SLAVE? | 444 | if (res < 0) |
529 | * - hsr_dev->priv_flags - i.e. | 445 | return res; |
530 | * IFF_EBRIDGE? | ||
531 | * IFF_TX_SKB_SHARING? | ||
532 | * IFF_HSR_MASTER/SLAVE? | ||
533 | */ | ||
534 | 446 | ||
535 | for (i = 0; i < HSR_MAX_SLAVE; i++) { | 447 | spin_lock_init(&hsr->seqnr_lock); |
536 | res = check_slave_ok(slave[i]); | 448 | /* Overflow soon to find bugs easier: */ |
537 | if (res) | 449 | hsr->sequence_nr = HSR_SEQNR_START; |
538 | return res; | ||
539 | } | ||
540 | 450 | ||
541 | hsr_dev->features = slave[0]->features & slave[1]->features; | 451 | init_timer(&hsr->announce_timer); |
542 | /* Prevent recursive tx locking */ | 452 | hsr->announce_timer.function = hsr_announce; |
543 | hsr_dev->features |= NETIF_F_LLTX; | 453 | hsr->announce_timer.data = (unsigned long) hsr; |
544 | /* VLAN on top of HSR needs testing and probably some work on | ||
545 | * hsr_header_create() etc. | ||
546 | */ | ||
547 | hsr_dev->features |= NETIF_F_VLAN_CHALLENGED; | ||
548 | 454 | ||
549 | /* Set hsr_dev's MAC address to that of mac_slave1 */ | 455 | init_timer(&hsr->prune_timer); |
550 | ether_addr_copy(hsr_dev->dev_addr, hsr_priv->slave[0]->dev_addr); | 456 | hsr->prune_timer.function = hsr_prune_nodes; |
457 | hsr->prune_timer.data = (unsigned long) hsr; | ||
551 | 458 | ||
552 | /* Set required header length */ | 459 | ether_addr_copy(hsr->sup_multicast_addr, def_multicast_addr); |
553 | for (i = 0; i < HSR_MAX_SLAVE; i++) { | 460 | hsr->sup_multicast_addr[ETH_ALEN - 1] = multicast_spec; |
554 | if (slave[i]->hard_header_len + HSR_TAGLEN > | ||
555 | hsr_dev->hard_header_len) | ||
556 | hsr_dev->hard_header_len = | ||
557 | slave[i]->hard_header_len + HSR_TAGLEN; | ||
558 | } | ||
559 | 461 | ||
560 | /* MTU */ | 462 | /* FIXME: should I modify the value of these? |
561 | for (i = 0; i < HSR_MAX_SLAVE; i++) | 463 | * |
562 | if (slave[i]->mtu - HSR_TAGLEN < hsr_dev->mtu) | 464 | * - hsr_dev->flags - i.e. |
563 | hsr_dev->mtu = slave[i]->mtu - HSR_TAGLEN; | 465 | * IFF_MASTER/SLAVE? |
466 | * - hsr_dev->priv_flags - i.e. | ||
467 | * IFF_EBRIDGE? | ||
468 | * IFF_TX_SKB_SHARING? | ||
469 | * IFF_HSR_MASTER/SLAVE? | ||
470 | */ | ||
564 | 471 | ||
565 | /* Make sure the 1st call to netif_carrier_on() gets through */ | 472 | /* Make sure the 1st call to netif_carrier_on() gets through */ |
566 | netif_carrier_off(hsr_dev); | 473 | netif_carrier_off(hsr_dev); |
567 | 474 | ||
568 | /* Promiscuity */ | 475 | res = hsr_add_port(hsr, hsr_dev, HSR_PT_MASTER); |
569 | for (i = 0; i < HSR_MAX_SLAVE; i++) { | 476 | if (res) |
570 | res = dev_set_promiscuity(slave[i], 1); | 477 | return res; |
571 | if (res) { | ||
572 | netdev_info(hsr_dev, "Cannot set slave promiscuity (%s, %d)\n", | ||
573 | slave[i]->name, res); | ||
574 | goto fail; | ||
575 | } | ||
576 | } | ||
577 | 478 | ||
578 | /* Make sure we recognize frames from ourselves in hsr_rcv() */ | 479 | res = register_netdevice(hsr_dev); |
579 | res = hsr_create_self_node(&hsr_priv->self_node_db, | 480 | if (res) |
580 | hsr_dev->dev_addr, | ||
581 | hsr_priv->slave[1]->dev_addr); | ||
582 | if (res < 0) | ||
583 | goto fail; | 481 | goto fail; |
584 | 482 | ||
585 | res = register_netdevice(hsr_dev); | 483 | res = hsr_add_port(hsr, slave[0], HSR_PT_SLAVE_A); |
484 | if (res) | ||
485 | goto fail; | ||
486 | res = hsr_add_port(hsr, slave[1], HSR_PT_SLAVE_B); | ||
586 | if (res) | 487 | if (res) |
587 | goto fail; | 488 | goto fail; |
588 | 489 | ||
589 | register_hsr_master(hsr_priv); | 490 | hsr->prune_timer.expires = jiffies + msecs_to_jiffies(PRUNE_PERIOD); |
491 | add_timer(&hsr->prune_timer); | ||
590 | 492 | ||
591 | return 0; | 493 | return 0; |
592 | 494 | ||
593 | fail: | 495 | fail: |
594 | restore_slaves(hsr_dev); | 496 | hsr_for_each_port(hsr, port) |
497 | hsr_del_port(port); | ||
498 | |||
595 | return res; | 499 | return res; |
596 | } | 500 | } |