aboutsummaryrefslogtreecommitdiffstats
path: root/net
diff options
context:
space:
mode:
authorJarek Poplawski <jarkao2@o2.pl>2007-06-29 01:11:47 -0400
committerDavid S. Miller <davem@davemloft.net>2007-06-29 01:11:47 -0400
commit17200811cf539b9107a99a39bf71ba3567966285 (patch)
tree11763c9163f8d521acc74c9b89faa0210860b2f1 /net
parent25243633c29b72c4edd5fe9cfcbd76aa5eef8b36 (diff)
[NETPOLL] netconsole: fix soft lockup when removing module
#1 Until kernel ver. 2.6.21 (including) cancel_rearming_delayed_work() required a work function should always (unconditionally) rearm with delay > 0 - otherwise it would endlessly loop. This patch replaces this function with cancel_delayed_work(). Later kernel versions don't require this, so here it's only for uniformity. #2 After deleting a timer in cancel_[rearming_]delayed_work() there could stay a last skb queued in npinfo->txq causing a memory leak after kfree(npinfo). Initial patch & testing by: Jason Wessel <jason.wessel@windriver.com> Signed-off-by: Jarek Poplawski <jarkao2@o2.pl> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: David S. Miller <davem@davemloft.net>
Diffstat (limited to 'net')
-rw-r--r--net/core/netpoll.c11
1 files changed, 9 insertions, 2 deletions
diff --git a/net/core/netpoll.c b/net/core/netpoll.c
index f8e74e511ce6..cf40ff91ac01 100644
--- a/net/core/netpoll.c
+++ b/net/core/netpoll.c
@@ -72,7 +72,8 @@ static void queue_process(struct work_struct *work)
72 netif_tx_unlock(dev); 72 netif_tx_unlock(dev);
73 local_irq_restore(flags); 73 local_irq_restore(flags);
74 74
75 schedule_delayed_work(&npinfo->tx_work, HZ/10); 75 if (atomic_read(&npinfo->refcnt))
76 schedule_delayed_work(&npinfo->tx_work, HZ/10);
76 return; 77 return;
77 } 78 }
78 netif_tx_unlock(dev); 79 netif_tx_unlock(dev);
@@ -785,9 +786,15 @@ void netpoll_cleanup(struct netpoll *np)
785 if (atomic_dec_and_test(&npinfo->refcnt)) { 786 if (atomic_dec_and_test(&npinfo->refcnt)) {
786 skb_queue_purge(&npinfo->arp_tx); 787 skb_queue_purge(&npinfo->arp_tx);
787 skb_queue_purge(&npinfo->txq); 788 skb_queue_purge(&npinfo->txq);
788 cancel_rearming_delayed_work(&npinfo->tx_work); 789 cancel_delayed_work(&npinfo->tx_work);
789 flush_scheduled_work(); 790 flush_scheduled_work();
790 791
792 /* clean after last, unfinished work */
793 if (!skb_queue_empty(&npinfo->txq)) {
794 struct sk_buff *skb;
795 skb = __skb_dequeue(&npinfo->txq);
796 kfree_skb(skb);
797 }
791 kfree(npinfo); 798 kfree(npinfo);
792 } 799 }
793 } 800 }