diff options
author | Eric Dumazet <eric.dumazet@gmail.com> | 2012-07-30 21:08:23 -0400 |
---|---|---|
committer | David S. Miller <davem@davemloft.net> | 2012-07-31 17:41:38 -0400 |
commit | 54764bb647b2e847c512acf8d443df965da35000 (patch) | |
tree | 3918f6f42679d0ebdcef230f28285f99d178be62 /net/core | |
parent | 5a0d513b622ee41e117fc37e26e27e8ef42e8dae (diff) |
ipv4: Restore old dst_free() behavior.
commit 404e0a8b6a55 (net: ipv4: fix RCU races on dst refcounts) tried
to solve a race but added a problem at device/fib dismantle time :
We really want to call dst_free() as soon as possible, even if sockets
still have dst in their cache.
dst_release() calls in free_fib_info_rcu() are not welcomed.
Root of the problem was that now we also cache output routes (in
nh_rth_output), we must use call_rcu() instead of call_rcu_bh() in
rt_free(), because output route lookups are done in process context.
Based on feedback and initial patch from David Miller (adding another
call_rcu_bh() call in fib, but it appears it was not the right fix)
I left the inet_sk_rx_dst_set() helper and added __rcu attributes
to nh_rth_output and nh_rth_input to better document what is going on in
this code.
Signed-off-by: Eric Dumazet <edumazet@google.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'net/core')
-rw-r--r-- | net/core/dst.c | 26 |
1 files changed, 5 insertions, 21 deletions
diff --git a/net/core/dst.c b/net/core/dst.c index d9e33ebe170f..069d51d29414 100644 --- a/net/core/dst.c +++ b/net/core/dst.c | |||
@@ -258,15 +258,6 @@ again: | |||
258 | } | 258 | } |
259 | EXPORT_SYMBOL(dst_destroy); | 259 | EXPORT_SYMBOL(dst_destroy); |
260 | 260 | ||
261 | static void dst_rcu_destroy(struct rcu_head *head) | ||
262 | { | ||
263 | struct dst_entry *dst = container_of(head, struct dst_entry, rcu_head); | ||
264 | |||
265 | dst = dst_destroy(dst); | ||
266 | if (dst) | ||
267 | __dst_free(dst); | ||
268 | } | ||
269 | |||
270 | void dst_release(struct dst_entry *dst) | 261 | void dst_release(struct dst_entry *dst) |
271 | { | 262 | { |
272 | if (dst) { | 263 | if (dst) { |
@@ -274,14 +265,10 @@ void dst_release(struct dst_entry *dst) | |||
274 | 265 | ||
275 | newrefcnt = atomic_dec_return(&dst->__refcnt); | 266 | newrefcnt = atomic_dec_return(&dst->__refcnt); |
276 | WARN_ON(newrefcnt < 0); | 267 | WARN_ON(newrefcnt < 0); |
277 | if (unlikely(dst->flags & (DST_NOCACHE | DST_RCU_FREE)) && !newrefcnt) { | 268 | if (unlikely(dst->flags & DST_NOCACHE) && !newrefcnt) { |
278 | if (dst->flags & DST_RCU_FREE) { | 269 | dst = dst_destroy(dst); |
279 | call_rcu_bh(&dst->rcu_head, dst_rcu_destroy); | 270 | if (dst) |
280 | } else { | 271 | __dst_free(dst); |
281 | dst = dst_destroy(dst); | ||
282 | if (dst) | ||
283 | __dst_free(dst); | ||
284 | } | ||
285 | } | 272 | } |
286 | } | 273 | } |
287 | } | 274 | } |
@@ -333,14 +320,11 @@ EXPORT_SYMBOL(__dst_destroy_metrics_generic); | |||
333 | */ | 320 | */ |
334 | void skb_dst_set_noref(struct sk_buff *skb, struct dst_entry *dst) | 321 | void skb_dst_set_noref(struct sk_buff *skb, struct dst_entry *dst) |
335 | { | 322 | { |
336 | bool hold; | ||
337 | |||
338 | WARN_ON(!rcu_read_lock_held() && !rcu_read_lock_bh_held()); | 323 | WARN_ON(!rcu_read_lock_held() && !rcu_read_lock_bh_held()); |
339 | /* If dst not in cache, we must take a reference, because | 324 | /* If dst not in cache, we must take a reference, because |
340 | * dst_release() will destroy dst as soon as its refcount becomes zero | 325 | * dst_release() will destroy dst as soon as its refcount becomes zero |
341 | */ | 326 | */ |
342 | hold = (dst->flags & (DST_NOCACHE | DST_RCU_FREE)) == DST_NOCACHE; | 327 | if (unlikely(dst->flags & DST_NOCACHE)) { |
343 | if (unlikely(hold)) { | ||
344 | dst_hold(dst); | 328 | dst_hold(dst); |
345 | skb_dst_set(skb, dst); | 329 | skb_dst_set(skb, dst); |
346 | } else { | 330 | } else { |