diff options
author | Jarek Poplawski <jarkao2@o2.pl> | 2007-06-29 01:11:47 -0400 |
---|---|---|
committer | David S. Miller <davem@davemloft.net> | 2007-06-29 01:11:47 -0400 |
commit | 17200811cf539b9107a99a39bf71ba3567966285 (patch) | |
tree | 11763c9163f8d521acc74c9b89faa0210860b2f1 /net/core/netpoll.c | |
parent | 25243633c29b72c4edd5fe9cfcbd76aa5eef8b36 (diff) |
[NETPOLL] netconsole: fix soft lockup when removing module
#1
Until kernel ver. 2.6.21 (including) cancel_rearming_delayed_work()
required a work function should always (unconditionally) rearm with
delay > 0 - otherwise it would endlessly loop. This patch replaces
this function with cancel_delayed_work(). Later kernel versions don't
require this, so here it's only for uniformity.
#2
After deleting a timer in cancel_[rearming_]delayed_work() there could
stay a last skb queued in npinfo->txq causing a memory leak after
kfree(npinfo).
Initial patch & testing by: Jason Wessel <jason.wessel@windriver.com>
Signed-off-by: Jarek Poplawski <jarkao2@o2.pl>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'net/core/netpoll.c')
-rw-r--r-- | net/core/netpoll.c | 11 |
1 files changed, 9 insertions, 2 deletions
diff --git a/net/core/netpoll.c b/net/core/netpoll.c index f8e74e511ce..cf40ff91ac0 100644 --- a/net/core/netpoll.c +++ b/net/core/netpoll.c | |||
@@ -72,7 +72,8 @@ static void queue_process(struct work_struct *work) | |||
72 | netif_tx_unlock(dev); | 72 | netif_tx_unlock(dev); |
73 | local_irq_restore(flags); | 73 | local_irq_restore(flags); |
74 | 74 | ||
75 | schedule_delayed_work(&npinfo->tx_work, HZ/10); | 75 | if (atomic_read(&npinfo->refcnt)) |
76 | schedule_delayed_work(&npinfo->tx_work, HZ/10); | ||
76 | return; | 77 | return; |
77 | } | 78 | } |
78 | netif_tx_unlock(dev); | 79 | netif_tx_unlock(dev); |
@@ -785,9 +786,15 @@ void netpoll_cleanup(struct netpoll *np) | |||
785 | if (atomic_dec_and_test(&npinfo->refcnt)) { | 786 | if (atomic_dec_and_test(&npinfo->refcnt)) { |
786 | skb_queue_purge(&npinfo->arp_tx); | 787 | skb_queue_purge(&npinfo->arp_tx); |
787 | skb_queue_purge(&npinfo->txq); | 788 | skb_queue_purge(&npinfo->txq); |
788 | cancel_rearming_delayed_work(&npinfo->tx_work); | 789 | cancel_delayed_work(&npinfo->tx_work); |
789 | flush_scheduled_work(); | 790 | flush_scheduled_work(); |
790 | 791 | ||
792 | /* clean after last, unfinished work */ | ||
793 | if (!skb_queue_empty(&npinfo->txq)) { | ||
794 | struct sk_buff *skb; | ||
795 | skb = __skb_dequeue(&npinfo->txq); | ||
796 | kfree_skb(skb); | ||
797 | } | ||
791 | kfree(npinfo); | 798 | kfree(npinfo); |
792 | } | 799 | } |
793 | } | 800 | } |