aboutsummaryrefslogtreecommitdiffstats
path: root/net/openvswitch/datapath.c
diff options
context:
space:
mode:
authorPravin B Shelar <pshelar@nicira.com>2012-02-22 22:58:59 -0500
committerJesse Gross <jesse@nicira.com>2012-08-22 17:48:55 -0400
commit46df7b814548849deee01f50bc75f8f5ae8cd767 (patch)
tree1663fa6ae46edcba3787c0a4ff839365bb52bcbe /net/openvswitch/datapath.c
parent0d7614f09c1ebdbaa1599a5aba7593f147bf96ee (diff)
openvswitch: Add support for network namespaces.
Following patch adds support for network namespace to openvswitch. Since it must release devices when namespaces are destroyed, a side effect of this patch is that the module no longer keeps a refcount but instead cleans up any state when it is unloaded. Signed-off-by: Pravin B Shelar <pshelar@nicira.com> Signed-off-by: Jesse Gross <jesse@nicira.com>
Diffstat (limited to 'net/openvswitch/datapath.c')
-rw-r--r--net/openvswitch/datapath.c269
1 files changed, 162 insertions, 107 deletions
diff --git a/net/openvswitch/datapath.c b/net/openvswitch/datapath.c
index d8277d29e710..cad39fca75a9 100644
--- a/net/openvswitch/datapath.c
+++ b/net/openvswitch/datapath.c
@@ -49,12 +49,29 @@
49#include <linux/dmi.h> 49#include <linux/dmi.h>
50#include <linux/workqueue.h> 50#include <linux/workqueue.h>
51#include <net/genetlink.h> 51#include <net/genetlink.h>
52#include <net/net_namespace.h>
53#include <net/netns/generic.h>
52 54
53#include "datapath.h" 55#include "datapath.h"
54#include "flow.h" 56#include "flow.h"
55#include "vport-internal_dev.h" 57#include "vport-internal_dev.h"
56 58
57/** 59/**
60 * struct ovs_net - Per net-namespace data for ovs.
61 * @dps: List of datapaths to enable dumping them all out.
62 * Protected by genl_mutex.
63 */
64struct ovs_net {
65 struct list_head dps;
66};
67
68static int ovs_net_id __read_mostly;
69
70#define REHASH_FLOW_INTERVAL (10 * 60 * HZ)
71static void rehash_flow_table(struct work_struct *work);
72static DECLARE_DELAYED_WORK(rehash_flow_wq, rehash_flow_table);
73
74/**
58 * DOC: Locking: 75 * DOC: Locking:
59 * 76 *
60 * Writes to device state (add/remove datapath, port, set operations on vports, 77 * Writes to device state (add/remove datapath, port, set operations on vports,
@@ -71,29 +88,21 @@
71 * each other. 88 * each other.
72 */ 89 */
73 90
74/* Global list of datapaths to enable dumping them all out.
75 * Protected by genl_mutex.
76 */
77static LIST_HEAD(dps);
78
79#define REHASH_FLOW_INTERVAL (10 * 60 * HZ)
80static void rehash_flow_table(struct work_struct *work);
81static DECLARE_DELAYED_WORK(rehash_flow_wq, rehash_flow_table);
82
83static struct vport *new_vport(const struct vport_parms *); 91static struct vport *new_vport(const struct vport_parms *);
84static int queue_gso_packets(int dp_ifindex, struct sk_buff *, 92static int queue_gso_packets(struct net *, int dp_ifindex, struct sk_buff *,
85 const struct dp_upcall_info *); 93 const struct dp_upcall_info *);
86static int queue_userspace_packet(int dp_ifindex, struct sk_buff *, 94static int queue_userspace_packet(struct net *, int dp_ifindex,
95 struct sk_buff *,
87 const struct dp_upcall_info *); 96 const struct dp_upcall_info *);
88 97
89/* Must be called with rcu_read_lock, genl_mutex, or RTNL lock. */ 98/* Must be called with rcu_read_lock, genl_mutex, or RTNL lock. */
90static struct datapath *get_dp(int dp_ifindex) 99static struct datapath *get_dp(struct net *net, int dp_ifindex)
91{ 100{
92 struct datapath *dp = NULL; 101 struct datapath *dp = NULL;
93 struct net_device *dev; 102 struct net_device *dev;
94 103
95 rcu_read_lock(); 104 rcu_read_lock();
96 dev = dev_get_by_index_rcu(&init_net, dp_ifindex); 105 dev = dev_get_by_index_rcu(net, dp_ifindex);
97 if (dev) { 106 if (dev) {
98 struct vport *vport = ovs_internal_dev_get_vport(dev); 107 struct vport *vport = ovs_internal_dev_get_vport(dev);
99 if (vport) 108 if (vport)
@@ -135,6 +144,7 @@ static void destroy_dp_rcu(struct rcu_head *rcu)
135 144
136 ovs_flow_tbl_destroy((__force struct flow_table *)dp->table); 145 ovs_flow_tbl_destroy((__force struct flow_table *)dp->table);
137 free_percpu(dp->stats_percpu); 146 free_percpu(dp->stats_percpu);
147 release_net(ovs_dp_get_net(dp));
138 kfree(dp); 148 kfree(dp);
139} 149}
140 150
@@ -220,11 +230,12 @@ static struct genl_family dp_packet_genl_family = {
220 .hdrsize = sizeof(struct ovs_header), 230 .hdrsize = sizeof(struct ovs_header),
221 .name = OVS_PACKET_FAMILY, 231 .name = OVS_PACKET_FAMILY,
222 .version = OVS_PACKET_VERSION, 232 .version = OVS_PACKET_VERSION,
223 .maxattr = OVS_PACKET_ATTR_MAX 233 .maxattr = OVS_PACKET_ATTR_MAX,
234 .netnsok = true
224}; 235};
225 236
226int ovs_dp_upcall(struct datapath *dp, struct sk_buff *skb, 237int ovs_dp_upcall(struct datapath *dp, struct sk_buff *skb,
227 const struct dp_upcall_info *upcall_info) 238 const struct dp_upcall_info *upcall_info)
228{ 239{
229 struct dp_stats_percpu *stats; 240 struct dp_stats_percpu *stats;
230 int dp_ifindex; 241 int dp_ifindex;
@@ -242,9 +253,9 @@ int ovs_dp_upcall(struct datapath *dp, struct sk_buff *skb,
242 } 253 }
243 254
244 if (!skb_is_gso(skb)) 255 if (!skb_is_gso(skb))
245 err = queue_userspace_packet(dp_ifindex, skb, upcall_info); 256 err = queue_userspace_packet(ovs_dp_get_net(dp), dp_ifindex, skb, upcall_info);
246 else 257 else
247 err = queue_gso_packets(dp_ifindex, skb, upcall_info); 258 err = queue_gso_packets(ovs_dp_get_net(dp), dp_ifindex, skb, upcall_info);
248 if (err) 259 if (err)
249 goto err; 260 goto err;
250 261
@@ -260,7 +271,8 @@ err:
260 return err; 271 return err;
261} 272}
262 273
263static int queue_gso_packets(int dp_ifindex, struct sk_buff *skb, 274static int queue_gso_packets(struct net *net, int dp_ifindex,
275 struct sk_buff *skb,
264 const struct dp_upcall_info *upcall_info) 276 const struct dp_upcall_info *upcall_info)
265{ 277{
266 unsigned short gso_type = skb_shinfo(skb)->gso_type; 278 unsigned short gso_type = skb_shinfo(skb)->gso_type;
@@ -276,7 +288,7 @@ static int queue_gso_packets(int dp_ifindex, struct sk_buff *skb,
276 /* Queue all of the segments. */ 288 /* Queue all of the segments. */
277 skb = segs; 289 skb = segs;
278 do { 290 do {
279 err = queue_userspace_packet(dp_ifindex, skb, upcall_info); 291 err = queue_userspace_packet(net, dp_ifindex, skb, upcall_info);
280 if (err) 292 if (err)
281 break; 293 break;
282 294
@@ -306,7 +318,8 @@ static int queue_gso_packets(int dp_ifindex, struct sk_buff *skb,
306 return err; 318 return err;
307} 319}
308 320
309static int queue_userspace_packet(int dp_ifindex, struct sk_buff *skb, 321static int queue_userspace_packet(struct net *net, int dp_ifindex,
322 struct sk_buff *skb,
310 const struct dp_upcall_info *upcall_info) 323 const struct dp_upcall_info *upcall_info)
311{ 324{
312 struct ovs_header *upcall; 325 struct ovs_header *upcall;
@@ -362,7 +375,7 @@ static int queue_userspace_packet(int dp_ifindex, struct sk_buff *skb,
362 375
363 skb_copy_and_csum_dev(skb, nla_data(nla)); 376 skb_copy_and_csum_dev(skb, nla_data(nla));
364 377
365 err = genlmsg_unicast(&init_net, user_skb, upcall_info->pid); 378 err = genlmsg_unicast(net, user_skb, upcall_info->pid);
366 379
367out: 380out:
368 kfree_skb(nskb); 381 kfree_skb(nskb);
@@ -370,15 +383,10 @@ out:
370} 383}
371 384
372/* Called with genl_mutex. */ 385/* Called with genl_mutex. */
373static int flush_flows(int dp_ifindex) 386static int flush_flows(struct datapath *dp)
374{ 387{
375 struct flow_table *old_table; 388 struct flow_table *old_table;
376 struct flow_table *new_table; 389 struct flow_table *new_table;
377 struct datapath *dp;
378
379 dp = get_dp(dp_ifindex);
380 if (!dp)
381 return -ENODEV;
382 390
383 old_table = genl_dereference(dp->table); 391 old_table = genl_dereference(dp->table);
384 new_table = ovs_flow_tbl_alloc(TBL_MIN_BUCKETS); 392 new_table = ovs_flow_tbl_alloc(TBL_MIN_BUCKETS);
@@ -668,7 +676,7 @@ static int ovs_packet_cmd_execute(struct sk_buff *skb, struct genl_info *info)
668 packet->priority = flow->key.phy.priority; 676 packet->priority = flow->key.phy.priority;
669 677
670 rcu_read_lock(); 678 rcu_read_lock();
671 dp = get_dp(ovs_header->dp_ifindex); 679 dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex);
672 err = -ENODEV; 680 err = -ENODEV;
673 if (!dp) 681 if (!dp)
674 goto err_unlock; 682 goto err_unlock;
@@ -742,7 +750,8 @@ static struct genl_family dp_flow_genl_family = {
742 .hdrsize = sizeof(struct ovs_header), 750 .hdrsize = sizeof(struct ovs_header),
743 .name = OVS_FLOW_FAMILY, 751 .name = OVS_FLOW_FAMILY,
744 .version = OVS_FLOW_VERSION, 752 .version = OVS_FLOW_VERSION,
745 .maxattr = OVS_FLOW_ATTR_MAX 753 .maxattr = OVS_FLOW_ATTR_MAX,
754 .netnsok = true
746}; 755};
747 756
748static struct genl_multicast_group ovs_dp_flow_multicast_group = { 757static struct genl_multicast_group ovs_dp_flow_multicast_group = {
@@ -894,7 +903,7 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info)
894 goto error; 903 goto error;
895 } 904 }
896 905
897 dp = get_dp(ovs_header->dp_ifindex); 906 dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex);
898 error = -ENODEV; 907 error = -ENODEV;
899 if (!dp) 908 if (!dp)
900 goto error; 909 goto error;
@@ -995,7 +1004,7 @@ static int ovs_flow_cmd_new_or_set(struct sk_buff *skb, struct genl_info *info)
995 ovs_dp_flow_multicast_group.id, info->nlhdr, 1004 ovs_dp_flow_multicast_group.id, info->nlhdr,
996 GFP_KERNEL); 1005 GFP_KERNEL);
997 else 1006 else
998 netlink_set_err(init_net.genl_sock, 0, 1007 netlink_set_err(sock_net(skb->sk)->genl_sock, 0,
999 ovs_dp_flow_multicast_group.id, PTR_ERR(reply)); 1008 ovs_dp_flow_multicast_group.id, PTR_ERR(reply));
1000 return 0; 1009 return 0;
1001 1010
@@ -1023,7 +1032,7 @@ static int ovs_flow_cmd_get(struct sk_buff *skb, struct genl_info *info)
1023 if (err) 1032 if (err)
1024 return err; 1033 return err;
1025 1034
1026 dp = get_dp(ovs_header->dp_ifindex); 1035 dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex);
1027 if (!dp) 1036 if (!dp)
1028 return -ENODEV; 1037 return -ENODEV;
1029 1038
@@ -1052,16 +1061,17 @@ static int ovs_flow_cmd_del(struct sk_buff *skb, struct genl_info *info)
1052 int err; 1061 int err;
1053 int key_len; 1062 int key_len;
1054 1063
1064 dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex);
1065 if (!dp)
1066 return -ENODEV;
1067
1055 if (!a[OVS_FLOW_ATTR_KEY]) 1068 if (!a[OVS_FLOW_ATTR_KEY])
1056 return flush_flows(ovs_header->dp_ifindex); 1069 return flush_flows(dp);
1070
1057 err = ovs_flow_from_nlattrs(&key, &key_len, a[OVS_FLOW_ATTR_KEY]); 1071 err = ovs_flow_from_nlattrs(&key, &key_len, a[OVS_FLOW_ATTR_KEY]);
1058 if (err) 1072 if (err)
1059 return err; 1073 return err;
1060 1074
1061 dp = get_dp(ovs_header->dp_ifindex);
1062 if (!dp)
1063 return -ENODEV;
1064
1065 table = genl_dereference(dp->table); 1075 table = genl_dereference(dp->table);
1066 flow = ovs_flow_tbl_lookup(table, &key, key_len); 1076 flow = ovs_flow_tbl_lookup(table, &key, key_len);
1067 if (!flow) 1077 if (!flow)
@@ -1090,7 +1100,7 @@ static int ovs_flow_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb)
1090 struct datapath *dp; 1100 struct datapath *dp;
1091 struct flow_table *table; 1101 struct flow_table *table;
1092 1102
1093 dp = get_dp(ovs_header->dp_ifindex); 1103 dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex);
1094 if (!dp) 1104 if (!dp)
1095 return -ENODEV; 1105 return -ENODEV;
1096 1106
@@ -1152,7 +1162,8 @@ static struct genl_family dp_datapath_genl_family = {
1152 .hdrsize = sizeof(struct ovs_header), 1162 .hdrsize = sizeof(struct ovs_header),
1153 .name = OVS_DATAPATH_FAMILY, 1163 .name = OVS_DATAPATH_FAMILY,
1154 .version = OVS_DATAPATH_VERSION, 1164 .version = OVS_DATAPATH_VERSION,
1155 .maxattr = OVS_DP_ATTR_MAX 1165 .maxattr = OVS_DP_ATTR_MAX,
1166 .netnsok = true
1156}; 1167};
1157 1168
1158static struct genl_multicast_group ovs_dp_datapath_multicast_group = { 1169static struct genl_multicast_group ovs_dp_datapath_multicast_group = {
@@ -1210,18 +1221,19 @@ static struct sk_buff *ovs_dp_cmd_build_info(struct datapath *dp, u32 pid,
1210} 1221}
1211 1222
1212/* Called with genl_mutex and optionally with RTNL lock also. */ 1223/* Called with genl_mutex and optionally with RTNL lock also. */
1213static struct datapath *lookup_datapath(struct ovs_header *ovs_header, 1224static struct datapath *lookup_datapath(struct net *net,
1225 struct ovs_header *ovs_header,
1214 struct nlattr *a[OVS_DP_ATTR_MAX + 1]) 1226 struct nlattr *a[OVS_DP_ATTR_MAX + 1])
1215{ 1227{
1216 struct datapath *dp; 1228 struct datapath *dp;
1217 1229
1218 if (!a[OVS_DP_ATTR_NAME]) 1230 if (!a[OVS_DP_ATTR_NAME])
1219 dp = get_dp(ovs_header->dp_ifindex); 1231 dp = get_dp(net, ovs_header->dp_ifindex);
1220 else { 1232 else {
1221 struct vport *vport; 1233 struct vport *vport;
1222 1234
1223 rcu_read_lock(); 1235 rcu_read_lock();
1224 vport = ovs_vport_locate(nla_data(a[OVS_DP_ATTR_NAME])); 1236 vport = ovs_vport_locate(net, nla_data(a[OVS_DP_ATTR_NAME]));
1225 dp = vport && vport->port_no == OVSP_LOCAL ? vport->dp : NULL; 1237 dp = vport && vport->port_no == OVSP_LOCAL ? vport->dp : NULL;
1226 rcu_read_unlock(); 1238 rcu_read_unlock();
1227 } 1239 }
@@ -1235,6 +1247,7 @@ static int ovs_dp_cmd_new(struct sk_buff *skb, struct genl_info *info)
1235 struct sk_buff *reply; 1247 struct sk_buff *reply;
1236 struct datapath *dp; 1248 struct datapath *dp;
1237 struct vport *vport; 1249 struct vport *vport;
1250 struct ovs_net *ovs_net;
1238 int err; 1251 int err;
1239 1252
1240 err = -EINVAL; 1253 err = -EINVAL;
@@ -1242,15 +1255,14 @@ static int ovs_dp_cmd_new(struct sk_buff *skb, struct genl_info *info)
1242 goto err; 1255 goto err;
1243 1256
1244 rtnl_lock(); 1257 rtnl_lock();
1245 err = -ENODEV;
1246 if (!try_module_get(THIS_MODULE))
1247 goto err_unlock_rtnl;
1248 1258
1249 err = -ENOMEM; 1259 err = -ENOMEM;
1250 dp = kzalloc(sizeof(*dp), GFP_KERNEL); 1260 dp = kzalloc(sizeof(*dp), GFP_KERNEL);
1251 if (dp == NULL) 1261 if (dp == NULL)
1252 goto err_put_module; 1262 goto err_unlock_rtnl;
1263
1253 INIT_LIST_HEAD(&dp->port_list); 1264 INIT_LIST_HEAD(&dp->port_list);
1265 ovs_dp_set_net(dp, hold_net(sock_net(skb->sk)));
1254 1266
1255 /* Allocate table. */ 1267 /* Allocate table. */
1256 err = -ENOMEM; 1268 err = -ENOMEM;
@@ -1287,7 +1299,8 @@ static int ovs_dp_cmd_new(struct sk_buff *skb, struct genl_info *info)
1287 if (IS_ERR(reply)) 1299 if (IS_ERR(reply))
1288 goto err_destroy_local_port; 1300 goto err_destroy_local_port;
1289 1301
1290 list_add_tail(&dp->list_node, &dps); 1302 ovs_net = net_generic(ovs_dp_get_net(dp), ovs_net_id);
1303 list_add_tail(&dp->list_node, &ovs_net->dps);
1291 rtnl_unlock(); 1304 rtnl_unlock();
1292 1305
1293 genl_notify(reply, genl_info_net(info), info->snd_pid, 1306 genl_notify(reply, genl_info_net(info), info->snd_pid,
@@ -1302,34 +1315,20 @@ err_destroy_percpu:
1302err_destroy_table: 1315err_destroy_table:
1303 ovs_flow_tbl_destroy(genl_dereference(dp->table)); 1316 ovs_flow_tbl_destroy(genl_dereference(dp->table));
1304err_free_dp: 1317err_free_dp:
1318 release_net(ovs_dp_get_net(dp));
1305 kfree(dp); 1319 kfree(dp);
1306err_put_module:
1307 module_put(THIS_MODULE);
1308err_unlock_rtnl: 1320err_unlock_rtnl:
1309 rtnl_unlock(); 1321 rtnl_unlock();
1310err: 1322err:
1311 return err; 1323 return err;
1312} 1324}
1313 1325
1314static int ovs_dp_cmd_del(struct sk_buff *skb, struct genl_info *info) 1326/* Called with genl_mutex. */
1327static void __dp_destroy(struct datapath *dp)
1315{ 1328{
1316 struct vport *vport, *next_vport; 1329 struct vport *vport, *next_vport;
1317 struct sk_buff *reply;
1318 struct datapath *dp;
1319 int err;
1320 1330
1321 rtnl_lock(); 1331 rtnl_lock();
1322 dp = lookup_datapath(info->userhdr, info->attrs);
1323 err = PTR_ERR(dp);
1324 if (IS_ERR(dp))
1325 goto exit_unlock;
1326
1327 reply = ovs_dp_cmd_build_info(dp, info->snd_pid,
1328 info->snd_seq, OVS_DP_CMD_DEL);
1329 err = PTR_ERR(reply);
1330 if (IS_ERR(reply))
1331 goto exit_unlock;
1332
1333 list_for_each_entry_safe(vport, next_vport, &dp->port_list, node) 1332 list_for_each_entry_safe(vport, next_vport, &dp->port_list, node)
1334 if (vport->port_no != OVSP_LOCAL) 1333 if (vport->port_no != OVSP_LOCAL)
1335 ovs_dp_detach_port(vport); 1334 ovs_dp_detach_port(vport);
@@ -1345,17 +1344,32 @@ static int ovs_dp_cmd_del(struct sk_buff *skb, struct genl_info *info)
1345 rtnl_unlock(); 1344 rtnl_unlock();
1346 1345
1347 call_rcu(&dp->rcu, destroy_dp_rcu); 1346 call_rcu(&dp->rcu, destroy_dp_rcu);
1348 module_put(THIS_MODULE); 1347}
1348
1349static int ovs_dp_cmd_del(struct sk_buff *skb, struct genl_info *info)
1350{
1351 struct sk_buff *reply;
1352 struct datapath *dp;
1353 int err;
1354
1355 dp = lookup_datapath(sock_net(skb->sk), info->userhdr, info->attrs);
1356 err = PTR_ERR(dp);
1357 if (IS_ERR(dp))
1358 return err;
1359
1360 reply = ovs_dp_cmd_build_info(dp, info->snd_pid,
1361 info->snd_seq, OVS_DP_CMD_DEL);
1362 err = PTR_ERR(reply);
1363 if (IS_ERR(reply))
1364 return err;
1365
1366 __dp_destroy(dp);
1349 1367
1350 genl_notify(reply, genl_info_net(info), info->snd_pid, 1368 genl_notify(reply, genl_info_net(info), info->snd_pid,
1351 ovs_dp_datapath_multicast_group.id, info->nlhdr, 1369 ovs_dp_datapath_multicast_group.id, info->nlhdr,
1352 GFP_KERNEL); 1370 GFP_KERNEL);
1353 1371
1354 return 0; 1372 return 0;
1355
1356exit_unlock:
1357 rtnl_unlock();
1358 return err;
1359} 1373}
1360 1374
1361static int ovs_dp_cmd_set(struct sk_buff *skb, struct genl_info *info) 1375static int ovs_dp_cmd_set(struct sk_buff *skb, struct genl_info *info)
@@ -1364,7 +1378,7 @@ static int ovs_dp_cmd_set(struct sk_buff *skb, struct genl_info *info)
1364 struct datapath *dp; 1378 struct datapath *dp;
1365 int err; 1379 int err;
1366 1380
1367 dp = lookup_datapath(info->userhdr, info->attrs); 1381 dp = lookup_datapath(sock_net(skb->sk), info->userhdr, info->attrs);
1368 if (IS_ERR(dp)) 1382 if (IS_ERR(dp))
1369 return PTR_ERR(dp); 1383 return PTR_ERR(dp);
1370 1384
@@ -1372,7 +1386,7 @@ static int ovs_dp_cmd_set(struct sk_buff *skb, struct genl_info *info)
1372 info->snd_seq, OVS_DP_CMD_NEW); 1386 info->snd_seq, OVS_DP_CMD_NEW);
1373 if (IS_ERR(reply)) { 1387 if (IS_ERR(reply)) {
1374 err = PTR_ERR(reply); 1388 err = PTR_ERR(reply);
1375 netlink_set_err(init_net.genl_sock, 0, 1389 netlink_set_err(sock_net(skb->sk)->genl_sock, 0,
1376 ovs_dp_datapath_multicast_group.id, err); 1390 ovs_dp_datapath_multicast_group.id, err);
1377 return 0; 1391 return 0;
1378 } 1392 }
@@ -1389,7 +1403,7 @@ static int ovs_dp_cmd_get(struct sk_buff *skb, struct genl_info *info)
1389 struct sk_buff *reply; 1403 struct sk_buff *reply;
1390 struct datapath *dp; 1404 struct datapath *dp;
1391 1405
1392 dp = lookup_datapath(info->userhdr, info->attrs); 1406 dp = lookup_datapath(sock_net(skb->sk), info->userhdr, info->attrs);
1393 if (IS_ERR(dp)) 1407 if (IS_ERR(dp))
1394 return PTR_ERR(dp); 1408 return PTR_ERR(dp);
1395 1409
@@ -1403,11 +1417,12 @@ static int ovs_dp_cmd_get(struct sk_buff *skb, struct genl_info *info)
1403 1417
1404static int ovs_dp_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb) 1418static int ovs_dp_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb)
1405{ 1419{
1420 struct ovs_net *ovs_net = net_generic(sock_net(skb->sk), ovs_net_id);
1406 struct datapath *dp; 1421 struct datapath *dp;
1407 int skip = cb->args[0]; 1422 int skip = cb->args[0];
1408 int i = 0; 1423 int i = 0;
1409 1424
1410 list_for_each_entry(dp, &dps, list_node) { 1425 list_for_each_entry(dp, &ovs_net->dps, list_node) {
1411 if (i >= skip && 1426 if (i >= skip &&
1412 ovs_dp_cmd_fill_info(dp, skb, NETLINK_CB(cb->skb).pid, 1427 ovs_dp_cmd_fill_info(dp, skb, NETLINK_CB(cb->skb).pid,
1413 cb->nlh->nlmsg_seq, NLM_F_MULTI, 1428 cb->nlh->nlmsg_seq, NLM_F_MULTI,
@@ -1459,7 +1474,8 @@ static struct genl_family dp_vport_genl_family = {
1459 .hdrsize = sizeof(struct ovs_header), 1474 .hdrsize = sizeof(struct ovs_header),
1460 .name = OVS_VPORT_FAMILY, 1475 .name = OVS_VPORT_FAMILY,
1461 .version = OVS_VPORT_VERSION, 1476 .version = OVS_VPORT_VERSION,
1462 .maxattr = OVS_VPORT_ATTR_MAX 1477 .maxattr = OVS_VPORT_ATTR_MAX,
1478 .netnsok = true
1463}; 1479};
1464 1480
1465struct genl_multicast_group ovs_dp_vport_multicast_group = { 1481struct genl_multicast_group ovs_dp_vport_multicast_group = {
@@ -1525,14 +1541,15 @@ struct sk_buff *ovs_vport_cmd_build_info(struct vport *vport, u32 pid,
1525} 1541}
1526 1542
1527/* Called with RTNL lock or RCU read lock. */ 1543/* Called with RTNL lock or RCU read lock. */
1528static struct vport *lookup_vport(struct ovs_header *ovs_header, 1544static struct vport *lookup_vport(struct net *net,
1545 struct ovs_header *ovs_header,
1529 struct nlattr *a[OVS_VPORT_ATTR_MAX + 1]) 1546 struct nlattr *a[OVS_VPORT_ATTR_MAX + 1])
1530{ 1547{
1531 struct datapath *dp; 1548 struct datapath *dp;
1532 struct vport *vport; 1549 struct vport *vport;
1533 1550
1534 if (a[OVS_VPORT_ATTR_NAME]) { 1551 if (a[OVS_VPORT_ATTR_NAME]) {
1535 vport = ovs_vport_locate(nla_data(a[OVS_VPORT_ATTR_NAME])); 1552 vport = ovs_vport_locate(net, nla_data(a[OVS_VPORT_ATTR_NAME]));
1536 if (!vport) 1553 if (!vport)
1537 return ERR_PTR(-ENODEV); 1554 return ERR_PTR(-ENODEV);
1538 if (ovs_header->dp_ifindex && 1555 if (ovs_header->dp_ifindex &&
@@ -1545,7 +1562,7 @@ static struct vport *lookup_vport(struct ovs_header *ovs_header,
1545 if (port_no >= DP_MAX_PORTS) 1562 if (port_no >= DP_MAX_PORTS)
1546 return ERR_PTR(-EFBIG); 1563 return ERR_PTR(-EFBIG);
1547 1564
1548 dp = get_dp(ovs_header->dp_ifindex); 1565 dp = get_dp(net, ovs_header->dp_ifindex);
1549 if (!dp) 1566 if (!dp)
1550 return ERR_PTR(-ENODEV); 1567 return ERR_PTR(-ENODEV);
1551 1568
@@ -1574,7 +1591,7 @@ static int ovs_vport_cmd_new(struct sk_buff *skb, struct genl_info *info)
1574 goto exit; 1591 goto exit;
1575 1592
1576 rtnl_lock(); 1593 rtnl_lock();
1577 dp = get_dp(ovs_header->dp_ifindex); 1594 dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex);
1578 err = -ENODEV; 1595 err = -ENODEV;
1579 if (!dp) 1596 if (!dp)
1580 goto exit_unlock; 1597 goto exit_unlock;
@@ -1638,7 +1655,7 @@ static int ovs_vport_cmd_set(struct sk_buff *skb, struct genl_info *info)
1638 int err; 1655 int err;
1639 1656
1640 rtnl_lock(); 1657 rtnl_lock();
1641 vport = lookup_vport(info->userhdr, a); 1658 vport = lookup_vport(sock_net(skb->sk), info->userhdr, a);
1642 err = PTR_ERR(vport); 1659 err = PTR_ERR(vport);
1643 if (IS_ERR(vport)) 1660 if (IS_ERR(vport))
1644 goto exit_unlock; 1661 goto exit_unlock;
@@ -1658,7 +1675,7 @@ static int ovs_vport_cmd_set(struct sk_buff *skb, struct genl_info *info)
1658 reply = ovs_vport_cmd_build_info(vport, info->snd_pid, info->snd_seq, 1675 reply = ovs_vport_cmd_build_info(vport, info->snd_pid, info->snd_seq,
1659 OVS_VPORT_CMD_NEW); 1676 OVS_VPORT_CMD_NEW);
1660 if (IS_ERR(reply)) { 1677 if (IS_ERR(reply)) {
1661 netlink_set_err(init_net.genl_sock, 0, 1678 netlink_set_err(sock_net(skb->sk)->genl_sock, 0,
1662 ovs_dp_vport_multicast_group.id, PTR_ERR(reply)); 1679 ovs_dp_vport_multicast_group.id, PTR_ERR(reply));
1663 goto exit_unlock; 1680 goto exit_unlock;
1664 } 1681 }
@@ -1679,7 +1696,7 @@ static int ovs_vport_cmd_del(struct sk_buff *skb, struct genl_info *info)
1679 int err; 1696 int err;
1680 1697
1681 rtnl_lock(); 1698 rtnl_lock();
1682 vport = lookup_vport(info->userhdr, a); 1699 vport = lookup_vport(sock_net(skb->sk), info->userhdr, a);
1683 err = PTR_ERR(vport); 1700 err = PTR_ERR(vport);
1684 if (IS_ERR(vport)) 1701 if (IS_ERR(vport))
1685 goto exit_unlock; 1702 goto exit_unlock;
@@ -1714,7 +1731,7 @@ static int ovs_vport_cmd_get(struct sk_buff *skb, struct genl_info *info)
1714 int err; 1731 int err;
1715 1732
1716 rcu_read_lock(); 1733 rcu_read_lock();
1717 vport = lookup_vport(ovs_header, a); 1734 vport = lookup_vport(sock_net(skb->sk), ovs_header, a);
1718 err = PTR_ERR(vport); 1735 err = PTR_ERR(vport);
1719 if (IS_ERR(vport)) 1736 if (IS_ERR(vport))
1720 goto exit_unlock; 1737 goto exit_unlock;
@@ -1741,7 +1758,7 @@ static int ovs_vport_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb)
1741 u32 port_no; 1758 u32 port_no;
1742 int retval; 1759 int retval;
1743 1760
1744 dp = get_dp(ovs_header->dp_ifindex); 1761 dp = get_dp(sock_net(skb->sk), ovs_header->dp_ifindex);
1745 if (!dp) 1762 if (!dp)
1746 return -ENODEV; 1763 return -ENODEV;
1747 1764
@@ -1766,28 +1783,6 @@ static int ovs_vport_cmd_dump(struct sk_buff *skb, struct netlink_callback *cb)
1766 return retval; 1783 return retval;
1767} 1784}
1768 1785
1769static void rehash_flow_table(struct work_struct *work)
1770{
1771 struct datapath *dp;
1772
1773 genl_lock();
1774
1775 list_for_each_entry(dp, &dps, list_node) {
1776 struct flow_table *old_table = genl_dereference(dp->table);
1777 struct flow_table *new_table;
1778
1779 new_table = ovs_flow_tbl_rehash(old_table);
1780 if (!IS_ERR(new_table)) {
1781 rcu_assign_pointer(dp->table, new_table);
1782 ovs_flow_tbl_deferred_destroy(old_table);
1783 }
1784 }
1785
1786 genl_unlock();
1787
1788 schedule_delayed_work(&rehash_flow_wq, REHASH_FLOW_INTERVAL);
1789}
1790
1791static struct genl_ops dp_vport_genl_ops[] = { 1786static struct genl_ops dp_vport_genl_ops[] = {
1792 { .cmd = OVS_VPORT_CMD_NEW, 1787 { .cmd = OVS_VPORT_CMD_NEW,
1793 .flags = GENL_ADMIN_PERM, /* Requires CAP_NET_ADMIN privilege. */ 1788 .flags = GENL_ADMIN_PERM, /* Requires CAP_NET_ADMIN privilege. */
@@ -1872,6 +1867,59 @@ error:
1872 return err; 1867 return err;
1873} 1868}
1874 1869
1870static void rehash_flow_table(struct work_struct *work)
1871{
1872 struct datapath *dp;
1873 struct net *net;
1874
1875 genl_lock();
1876 rtnl_lock();
1877 for_each_net(net) {
1878 struct ovs_net *ovs_net = net_generic(net, ovs_net_id);
1879
1880 list_for_each_entry(dp, &ovs_net->dps, list_node) {
1881 struct flow_table *old_table = genl_dereference(dp->table);
1882 struct flow_table *new_table;
1883
1884 new_table = ovs_flow_tbl_rehash(old_table);
1885 if (!IS_ERR(new_table)) {
1886 rcu_assign_pointer(dp->table, new_table);
1887 ovs_flow_tbl_deferred_destroy(old_table);
1888 }
1889 }
1890 }
1891 rtnl_unlock();
1892 genl_unlock();
1893
1894 schedule_delayed_work(&rehash_flow_wq, REHASH_FLOW_INTERVAL);
1895}
1896
1897static int __net_init ovs_init_net(struct net *net)
1898{
1899 struct ovs_net *ovs_net = net_generic(net, ovs_net_id);
1900
1901 INIT_LIST_HEAD(&ovs_net->dps);
1902 return 0;
1903}
1904
1905static void __net_exit ovs_exit_net(struct net *net)
1906{
1907 struct ovs_net *ovs_net = net_generic(net, ovs_net_id);
1908 struct datapath *dp, *dp_next;
1909
1910 genl_lock();
1911 list_for_each_entry_safe(dp, dp_next, &ovs_net->dps, list_node)
1912 __dp_destroy(dp);
1913 genl_unlock();
1914}
1915
1916static struct pernet_operations ovs_net_ops = {
1917 .init = ovs_init_net,
1918 .exit = ovs_exit_net,
1919 .id = &ovs_net_id,
1920 .size = sizeof(struct ovs_net),
1921};
1922
1875static int __init dp_init(void) 1923static int __init dp_init(void)
1876{ 1924{
1877 struct sk_buff *dummy_skb; 1925 struct sk_buff *dummy_skb;
@@ -1889,10 +1937,14 @@ static int __init dp_init(void)
1889 if (err) 1937 if (err)
1890 goto error_flow_exit; 1938 goto error_flow_exit;
1891 1939
1892 err = register_netdevice_notifier(&ovs_dp_device_notifier); 1940 err = register_pernet_device(&ovs_net_ops);
1893 if (err) 1941 if (err)
1894 goto error_vport_exit; 1942 goto error_vport_exit;
1895 1943
1944 err = register_netdevice_notifier(&ovs_dp_device_notifier);
1945 if (err)
1946 goto error_netns_exit;
1947
1896 err = dp_register_genl(); 1948 err = dp_register_genl();
1897 if (err < 0) 1949 if (err < 0)
1898 goto error_unreg_notifier; 1950 goto error_unreg_notifier;
@@ -1903,6 +1955,8 @@ static int __init dp_init(void)
1903 1955
1904error_unreg_notifier: 1956error_unreg_notifier:
1905 unregister_netdevice_notifier(&ovs_dp_device_notifier); 1957 unregister_netdevice_notifier(&ovs_dp_device_notifier);
1958error_netns_exit:
1959 unregister_pernet_device(&ovs_net_ops);
1906error_vport_exit: 1960error_vport_exit:
1907 ovs_vport_exit(); 1961 ovs_vport_exit();
1908error_flow_exit: 1962error_flow_exit:
@@ -1914,9 +1968,10 @@ error:
1914static void dp_cleanup(void) 1968static void dp_cleanup(void)
1915{ 1969{
1916 cancel_delayed_work_sync(&rehash_flow_wq); 1970 cancel_delayed_work_sync(&rehash_flow_wq);
1917 rcu_barrier();
1918 dp_unregister_genl(ARRAY_SIZE(dp_genl_families)); 1971 dp_unregister_genl(ARRAY_SIZE(dp_genl_families));
1919 unregister_netdevice_notifier(&ovs_dp_device_notifier); 1972 unregister_netdevice_notifier(&ovs_dp_device_notifier);
1973 unregister_pernet_device(&ovs_net_ops);
1974 rcu_barrier();
1920 ovs_vport_exit(); 1975 ovs_vport_exit();
1921 ovs_flow_exit(); 1976 ovs_flow_exit();
1922} 1977}