diff options
author | Mandeep Singh Baines <msb@google.com> | 2009-01-15 14:08:40 -0500 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2009-01-16 08:06:04 -0500 |
commit | e162b39a368f0401e41b558f430c354d12a85b37 (patch) | |
tree | 3fb7e4d48f398d62e5074e7e3dd183cc54f59820 /kernel | |
parent | c903ff837909ccada51243307d4239f86af40179 (diff) |
softlockup: decouple hung tasks check from softlockup detection
Decoupling allows:
* hung tasks check to happen at very low priority
* hung tasks check and softlockup to be enabled/disabled independently
at compile and/or run-time
* individual panic settings to be enabled disabled independently
at compile and/or run-time
* softlockup threshold to be reduced without increasing hung tasks
poll frequency (hung task check is expensive relative to softlock watchdog)
* hung task check to be zero over-head when disabled at run-time
Signed-off-by: Mandeep Singh Baines <msb@google.com>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel')
-rw-r--r-- | kernel/Makefile | 1 | ||||
-rw-r--r-- | kernel/hung_task.c | 198 | ||||
-rw-r--r-- | kernel/softlockup.c | 100 | ||||
-rw-r--r-- | kernel/sysctl.c | 15 |
4 files changed, 213 insertions, 101 deletions
diff --git a/kernel/Makefile b/kernel/Makefile index 2aebc4cd7878..979745f1b4bc 100644 --- a/kernel/Makefile +++ b/kernel/Makefile | |||
@@ -74,6 +74,7 @@ obj-$(CONFIG_AUDIT_TREE) += audit_tree.o | |||
74 | obj-$(CONFIG_KPROBES) += kprobes.o | 74 | obj-$(CONFIG_KPROBES) += kprobes.o |
75 | obj-$(CONFIG_KGDB) += kgdb.o | 75 | obj-$(CONFIG_KGDB) += kgdb.o |
76 | obj-$(CONFIG_DETECT_SOFTLOCKUP) += softlockup.o | 76 | obj-$(CONFIG_DETECT_SOFTLOCKUP) += softlockup.o |
77 | obj-$(CONFIG_DETECT_HUNG_TASK) += hung_task.o | ||
77 | obj-$(CONFIG_GENERIC_HARDIRQS) += irq/ | 78 | obj-$(CONFIG_GENERIC_HARDIRQS) += irq/ |
78 | obj-$(CONFIG_SECCOMP) += seccomp.o | 79 | obj-$(CONFIG_SECCOMP) += seccomp.o |
79 | obj-$(CONFIG_RCU_TORTURE_TEST) += rcutorture.o | 80 | obj-$(CONFIG_RCU_TORTURE_TEST) += rcutorture.o |
diff --git a/kernel/hung_task.c b/kernel/hung_task.c new file mode 100644 index 000000000000..ba5a77cad3bb --- /dev/null +++ b/kernel/hung_task.c | |||
@@ -0,0 +1,198 @@ | |||
1 | /* | ||
2 | * Detect Hung Task | ||
3 | * | ||
4 | * kernel/hung_task.c - kernel thread for detecting tasks stuck in D state | ||
5 | * | ||
6 | */ | ||
7 | |||
8 | #include <linux/mm.h> | ||
9 | #include <linux/cpu.h> | ||
10 | #include <linux/nmi.h> | ||
11 | #include <linux/init.h> | ||
12 | #include <linux/delay.h> | ||
13 | #include <linux/freezer.h> | ||
14 | #include <linux/kthread.h> | ||
15 | #include <linux/lockdep.h> | ||
16 | #include <linux/module.h> | ||
17 | #include <linux/sysctl.h> | ||
18 | |||
19 | /* | ||
20 | * Have a reasonable limit on the number of tasks checked: | ||
21 | */ | ||
22 | unsigned long __read_mostly sysctl_hung_task_check_count = 1024; | ||
23 | |||
24 | /* | ||
25 | * Zero means infinite timeout - no checking done: | ||
26 | */ | ||
27 | unsigned long __read_mostly sysctl_hung_task_timeout_secs = 120; | ||
28 | static unsigned long __read_mostly hung_task_poll_jiffies; | ||
29 | |||
30 | unsigned long __read_mostly sysctl_hung_task_warnings = 10; | ||
31 | |||
32 | static int __read_mostly did_panic; | ||
33 | |||
34 | static struct task_struct *watchdog_task; | ||
35 | |||
36 | /* | ||
37 | * Should we panic (and reboot, if panic_timeout= is set) when a | ||
38 | * hung task is detected: | ||
39 | */ | ||
40 | unsigned int __read_mostly sysctl_hung_task_panic = | ||
41 | CONFIG_BOOTPARAM_HUNG_TASK_PANIC_VALUE; | ||
42 | |||
43 | static int __init hung_task_panic_setup(char *str) | ||
44 | { | ||
45 | sysctl_hung_task_panic = simple_strtoul(str, NULL, 0); | ||
46 | |||
47 | return 1; | ||
48 | } | ||
49 | __setup("hung_task_panic=", hung_task_panic_setup); | ||
50 | |||
51 | static int | ||
52 | hung_task_panic(struct notifier_block *this, unsigned long event, void *ptr) | ||
53 | { | ||
54 | did_panic = 1; | ||
55 | |||
56 | return NOTIFY_DONE; | ||
57 | } | ||
58 | |||
59 | static struct notifier_block panic_block = { | ||
60 | .notifier_call = hung_task_panic, | ||
61 | }; | ||
62 | |||
63 | /* | ||
64 | * Returns seconds, approximately. We don't need nanosecond | ||
65 | * resolution, and we don't need to waste time with a big divide when | ||
66 | * 2^30ns == 1.074s. | ||
67 | */ | ||
68 | static unsigned long get_timestamp(void) | ||
69 | { | ||
70 | int this_cpu = raw_smp_processor_id(); | ||
71 | |||
72 | return cpu_clock(this_cpu) >> 30LL; /* 2^30 ~= 10^9 */ | ||
73 | } | ||
74 | |||
75 | static void check_hung_task(struct task_struct *t, unsigned long now) | ||
76 | { | ||
77 | unsigned long switch_count = t->nvcsw + t->nivcsw; | ||
78 | |||
79 | if (t->flags & PF_FROZEN) | ||
80 | return; | ||
81 | |||
82 | if (switch_count != t->last_switch_count || !t->last_switch_timestamp) { | ||
83 | t->last_switch_count = switch_count; | ||
84 | t->last_switch_timestamp = now; | ||
85 | return; | ||
86 | } | ||
87 | if ((long)(now - t->last_switch_timestamp) < | ||
88 | sysctl_hung_task_timeout_secs) | ||
89 | return; | ||
90 | if (!sysctl_hung_task_warnings) | ||
91 | return; | ||
92 | sysctl_hung_task_warnings--; | ||
93 | |||
94 | /* | ||
95 | * Ok, the task did not get scheduled for more than 2 minutes, | ||
96 | * complain: | ||
97 | */ | ||
98 | printk(KERN_ERR "INFO: task %s:%d blocked for more than " | ||
99 | "%ld seconds.\n", t->comm, t->pid, | ||
100 | sysctl_hung_task_timeout_secs); | ||
101 | printk(KERN_ERR "\"echo 0 > /proc/sys/kernel/hung_task_timeout_secs\"" | ||
102 | " disables this message.\n"); | ||
103 | sched_show_task(t); | ||
104 | __debug_show_held_locks(t); | ||
105 | |||
106 | t->last_switch_timestamp = now; | ||
107 | touch_nmi_watchdog(); | ||
108 | |||
109 | if (sysctl_hung_task_panic) | ||
110 | panic("hung_task: blocked tasks"); | ||
111 | } | ||
112 | |||
113 | /* | ||
114 | * Check whether a TASK_UNINTERRUPTIBLE does not get woken up for | ||
115 | * a really long time (120 seconds). If that happens, print out | ||
116 | * a warning. | ||
117 | */ | ||
118 | static void check_hung_uninterruptible_tasks(void) | ||
119 | { | ||
120 | int max_count = sysctl_hung_task_check_count; | ||
121 | unsigned long now = get_timestamp(); | ||
122 | struct task_struct *g, *t; | ||
123 | |||
124 | /* | ||
125 | * If the system crashed already then all bets are off, | ||
126 | * do not report extra hung tasks: | ||
127 | */ | ||
128 | if (test_taint(TAINT_DIE) || did_panic) | ||
129 | return; | ||
130 | |||
131 | read_lock(&tasklist_lock); | ||
132 | do_each_thread(g, t) { | ||
133 | if (!--max_count) | ||
134 | goto unlock; | ||
135 | /* use "==" to skip the TASK_KILLABLE tasks waiting on NFS */ | ||
136 | if (t->state == TASK_UNINTERRUPTIBLE) | ||
137 | check_hung_task(t, now); | ||
138 | } while_each_thread(g, t); | ||
139 | unlock: | ||
140 | read_unlock(&tasklist_lock); | ||
141 | } | ||
142 | |||
143 | static void update_poll_jiffies(void) | ||
144 | { | ||
145 | /* timeout of 0 will disable the watchdog */ | ||
146 | if (sysctl_hung_task_timeout_secs == 0) | ||
147 | hung_task_poll_jiffies = MAX_SCHEDULE_TIMEOUT; | ||
148 | else | ||
149 | hung_task_poll_jiffies = sysctl_hung_task_timeout_secs * HZ / 2; | ||
150 | } | ||
151 | |||
152 | /* | ||
153 | * Process updating of timeout sysctl | ||
154 | */ | ||
155 | int proc_dohung_task_timeout_secs(struct ctl_table *table, int write, | ||
156 | struct file *filp, void __user *buffer, | ||
157 | size_t *lenp, loff_t *ppos) | ||
158 | { | ||
159 | int ret; | ||
160 | |||
161 | ret = proc_doulongvec_minmax(table, write, filp, buffer, lenp, ppos); | ||
162 | |||
163 | if (ret || !write) | ||
164 | goto out; | ||
165 | |||
166 | update_poll_jiffies(); | ||
167 | |||
168 | wake_up_process(watchdog_task); | ||
169 | |||
170 | out: | ||
171 | return ret; | ||
172 | } | ||
173 | |||
174 | /* | ||
175 | * kthread which checks for tasks stuck in D state | ||
176 | */ | ||
177 | static int watchdog(void *dummy) | ||
178 | { | ||
179 | set_user_nice(current, 0); | ||
180 | update_poll_jiffies(); | ||
181 | |||
182 | for ( ; ; ) { | ||
183 | while (schedule_timeout_interruptible(hung_task_poll_jiffies)); | ||
184 | check_hung_uninterruptible_tasks(); | ||
185 | } | ||
186 | |||
187 | return 0; | ||
188 | } | ||
189 | |||
190 | static int __init hung_task_init(void) | ||
191 | { | ||
192 | atomic_notifier_chain_register(&panic_notifier_list, &panic_block); | ||
193 | watchdog_task = kthread_run(watchdog, NULL, "khungtaskd"); | ||
194 | |||
195 | return 0; | ||
196 | } | ||
197 | |||
198 | module_init(hung_task_init); | ||
diff --git a/kernel/softlockup.c b/kernel/softlockup.c index 85d5a2455103..88796c330838 100644 --- a/kernel/softlockup.c +++ b/kernel/softlockup.c | |||
@@ -166,97 +166,11 @@ void softlockup_tick(void) | |||
166 | } | 166 | } |
167 | 167 | ||
168 | /* | 168 | /* |
169 | * Have a reasonable limit on the number of tasks checked: | ||
170 | */ | ||
171 | unsigned long __read_mostly sysctl_hung_task_check_count = 1024; | ||
172 | |||
173 | /* | ||
174 | * Zero means infinite timeout - no checking done: | ||
175 | */ | ||
176 | unsigned long __read_mostly sysctl_hung_task_timeout_secs = 480; | ||
177 | |||
178 | unsigned long __read_mostly sysctl_hung_task_warnings = 10; | ||
179 | |||
180 | /* | ||
181 | * Only do the hung-tasks check on one CPU: | ||
182 | */ | ||
183 | static int check_cpu __read_mostly = -1; | ||
184 | |||
185 | static void check_hung_task(struct task_struct *t, unsigned long now) | ||
186 | { | ||
187 | unsigned long switch_count = t->nvcsw + t->nivcsw; | ||
188 | |||
189 | if (t->flags & PF_FROZEN) | ||
190 | return; | ||
191 | |||
192 | if (switch_count != t->last_switch_count || !t->last_switch_timestamp) { | ||
193 | t->last_switch_count = switch_count; | ||
194 | t->last_switch_timestamp = now; | ||
195 | return; | ||
196 | } | ||
197 | if ((long)(now - t->last_switch_timestamp) < | ||
198 | sysctl_hung_task_timeout_secs) | ||
199 | return; | ||
200 | if (!sysctl_hung_task_warnings) | ||
201 | return; | ||
202 | sysctl_hung_task_warnings--; | ||
203 | |||
204 | /* | ||
205 | * Ok, the task did not get scheduled for more than 2 minutes, | ||
206 | * complain: | ||
207 | */ | ||
208 | printk(KERN_ERR "INFO: task %s:%d blocked for more than " | ||
209 | "%ld seconds.\n", t->comm, t->pid, | ||
210 | sysctl_hung_task_timeout_secs); | ||
211 | printk(KERN_ERR "\"echo 0 > /proc/sys/kernel/hung_task_timeout_secs\"" | ||
212 | " disables this message.\n"); | ||
213 | sched_show_task(t); | ||
214 | __debug_show_held_locks(t); | ||
215 | |||
216 | t->last_switch_timestamp = now; | ||
217 | touch_nmi_watchdog(); | ||
218 | |||
219 | if (softlockup_panic) | ||
220 | panic("softlockup: blocked tasks"); | ||
221 | } | ||
222 | |||
223 | /* | ||
224 | * Check whether a TASK_UNINTERRUPTIBLE does not get woken up for | ||
225 | * a really long time (120 seconds). If that happens, print out | ||
226 | * a warning. | ||
227 | */ | ||
228 | static void check_hung_uninterruptible_tasks(int this_cpu) | ||
229 | { | ||
230 | int max_count = sysctl_hung_task_check_count; | ||
231 | unsigned long now = get_timestamp(this_cpu); | ||
232 | struct task_struct *g, *t; | ||
233 | |||
234 | /* | ||
235 | * If the system crashed already then all bets are off, | ||
236 | * do not report extra hung tasks: | ||
237 | */ | ||
238 | if (test_taint(TAINT_DIE) || did_panic) | ||
239 | return; | ||
240 | |||
241 | read_lock(&tasklist_lock); | ||
242 | do_each_thread(g, t) { | ||
243 | if (!--max_count) | ||
244 | goto unlock; | ||
245 | /* use "==" to skip the TASK_KILLABLE tasks waiting on NFS */ | ||
246 | if (t->state == TASK_UNINTERRUPTIBLE) | ||
247 | check_hung_task(t, now); | ||
248 | } while_each_thread(g, t); | ||
249 | unlock: | ||
250 | read_unlock(&tasklist_lock); | ||
251 | } | ||
252 | |||
253 | /* | ||
254 | * The watchdog thread - runs every second and touches the timestamp. | 169 | * The watchdog thread - runs every second and touches the timestamp. |
255 | */ | 170 | */ |
256 | static int watchdog(void *__bind_cpu) | 171 | static int watchdog(void *__bind_cpu) |
257 | { | 172 | { |
258 | struct sched_param param = { .sched_priority = MAX_RT_PRIO-1 }; | 173 | struct sched_param param = { .sched_priority = MAX_RT_PRIO-1 }; |
259 | int this_cpu = (long)__bind_cpu; | ||
260 | 174 | ||
261 | sched_setscheduler(current, SCHED_FIFO, ¶m); | 175 | sched_setscheduler(current, SCHED_FIFO, ¶m); |
262 | 176 | ||
@@ -276,11 +190,6 @@ static int watchdog(void *__bind_cpu) | |||
276 | if (kthread_should_stop()) | 190 | if (kthread_should_stop()) |
277 | break; | 191 | break; |
278 | 192 | ||
279 | if (this_cpu == check_cpu) { | ||
280 | if (sysctl_hung_task_timeout_secs) | ||
281 | check_hung_uninterruptible_tasks(this_cpu); | ||
282 | } | ||
283 | |||
284 | set_current_state(TASK_INTERRUPTIBLE); | 193 | set_current_state(TASK_INTERRUPTIBLE); |
285 | } | 194 | } |
286 | __set_current_state(TASK_RUNNING); | 195 | __set_current_state(TASK_RUNNING); |
@@ -312,18 +221,9 @@ cpu_callback(struct notifier_block *nfb, unsigned long action, void *hcpu) | |||
312 | break; | 221 | break; |
313 | case CPU_ONLINE: | 222 | case CPU_ONLINE: |
314 | case CPU_ONLINE_FROZEN: | 223 | case CPU_ONLINE_FROZEN: |
315 | check_cpu = cpumask_any(cpu_online_mask); | ||
316 | wake_up_process(per_cpu(watchdog_task, hotcpu)); | 224 | wake_up_process(per_cpu(watchdog_task, hotcpu)); |
317 | break; | 225 | break; |
318 | #ifdef CONFIG_HOTPLUG_CPU | 226 | #ifdef CONFIG_HOTPLUG_CPU |
319 | case CPU_DOWN_PREPARE: | ||
320 | case CPU_DOWN_PREPARE_FROZEN: | ||
321 | if (hotcpu == check_cpu) { | ||
322 | /* Pick any other online cpu. */ | ||
323 | check_cpu = cpumask_any_but(cpu_online_mask, hotcpu); | ||
324 | } | ||
325 | break; | ||
326 | |||
327 | case CPU_UP_CANCELED: | 227 | case CPU_UP_CANCELED: |
328 | case CPU_UP_CANCELED_FROZEN: | 228 | case CPU_UP_CANCELED_FROZEN: |
329 | if (!per_cpu(watchdog_task, hotcpu)) | 229 | if (!per_cpu(watchdog_task, hotcpu)) |
diff --git a/kernel/sysctl.c b/kernel/sysctl.c index 596dc31a7116..2481ed30d2b5 100644 --- a/kernel/sysctl.c +++ b/kernel/sysctl.c | |||
@@ -805,6 +805,19 @@ static struct ctl_table kern_table[] = { | |||
805 | .extra1 = &neg_one, | 805 | .extra1 = &neg_one, |
806 | .extra2 = &sixty, | 806 | .extra2 = &sixty, |
807 | }, | 807 | }, |
808 | #endif | ||
809 | #ifdef CONFIG_DETECT_HUNG_TASK | ||
810 | { | ||
811 | .ctl_name = CTL_UNNUMBERED, | ||
812 | .procname = "hung_task_panic", | ||
813 | .data = &sysctl_hung_task_panic, | ||
814 | .maxlen = sizeof(int), | ||
815 | .mode = 0644, | ||
816 | .proc_handler = &proc_dointvec_minmax, | ||
817 | .strategy = &sysctl_intvec, | ||
818 | .extra1 = &zero, | ||
819 | .extra2 = &one, | ||
820 | }, | ||
808 | { | 821 | { |
809 | .ctl_name = CTL_UNNUMBERED, | 822 | .ctl_name = CTL_UNNUMBERED, |
810 | .procname = "hung_task_check_count", | 823 | .procname = "hung_task_check_count", |
@@ -820,7 +833,7 @@ static struct ctl_table kern_table[] = { | |||
820 | .data = &sysctl_hung_task_timeout_secs, | 833 | .data = &sysctl_hung_task_timeout_secs, |
821 | .maxlen = sizeof(unsigned long), | 834 | .maxlen = sizeof(unsigned long), |
822 | .mode = 0644, | 835 | .mode = 0644, |
823 | .proc_handler = &proc_doulongvec_minmax, | 836 | .proc_handler = &proc_dohung_task_timeout_secs, |
824 | .strategy = &sysctl_intvec, | 837 | .strategy = &sysctl_intvec, |
825 | }, | 838 | }, |
826 | { | 839 | { |