diff options
| -rw-r--r-- | include/linux/sched.h | 18 | ||||
| -rw-r--r-- | kernel/Makefile | 1 | ||||
| -rw-r--r-- | kernel/fork.c | 8 | ||||
| -rw-r--r-- | kernel/hung_task.c | 217 | ||||
| -rw-r--r-- | kernel/softlockup.c | 100 | ||||
| -rw-r--r-- | kernel/sysctl.c | 15 | ||||
| -rw-r--r-- | lib/Kconfig.debug | 38 |
7 files changed, 283 insertions, 114 deletions
diff --git a/include/linux/sched.h b/include/linux/sched.h index c96140210d1c..98e1fe51601d 100644 --- a/include/linux/sched.h +++ b/include/linux/sched.h | |||
| @@ -300,17 +300,11 @@ extern int proc_dosoftlockup_thresh(struct ctl_table *table, int write, | |||
| 300 | struct file *filp, void __user *buffer, | 300 | struct file *filp, void __user *buffer, |
| 301 | size_t *lenp, loff_t *ppos); | 301 | size_t *lenp, loff_t *ppos); |
| 302 | extern unsigned int softlockup_panic; | 302 | extern unsigned int softlockup_panic; |
| 303 | extern unsigned long sysctl_hung_task_check_count; | ||
| 304 | extern unsigned long sysctl_hung_task_timeout_secs; | ||
| 305 | extern unsigned long sysctl_hung_task_warnings; | ||
| 306 | extern int softlockup_thresh; | 303 | extern int softlockup_thresh; |
| 307 | #else | 304 | #else |
| 308 | static inline void softlockup_tick(void) | 305 | static inline void softlockup_tick(void) |
| 309 | { | 306 | { |
| 310 | } | 307 | } |
| 311 | static inline void spawn_softlockup_task(void) | ||
| 312 | { | ||
| 313 | } | ||
| 314 | static inline void touch_softlockup_watchdog(void) | 308 | static inline void touch_softlockup_watchdog(void) |
| 315 | { | 309 | { |
| 316 | } | 310 | } |
| @@ -319,6 +313,15 @@ static inline void touch_all_softlockup_watchdogs(void) | |||
| 319 | } | 313 | } |
| 320 | #endif | 314 | #endif |
| 321 | 315 | ||
| 316 | #ifdef CONFIG_DETECT_HUNG_TASK | ||
| 317 | extern unsigned int sysctl_hung_task_panic; | ||
| 318 | extern unsigned long sysctl_hung_task_check_count; | ||
| 319 | extern unsigned long sysctl_hung_task_timeout_secs; | ||
| 320 | extern unsigned long sysctl_hung_task_warnings; | ||
| 321 | extern int proc_dohung_task_timeout_secs(struct ctl_table *table, int write, | ||
| 322 | struct file *filp, void __user *buffer, | ||
| 323 | size_t *lenp, loff_t *ppos); | ||
| 324 | #endif | ||
| 322 | 325 | ||
| 323 | /* Attach to any functions which should be ignored in wchan output. */ | 326 | /* Attach to any functions which should be ignored in wchan output. */ |
| 324 | #define __sched __attribute__((__section__(".sched.text"))) | 327 | #define __sched __attribute__((__section__(".sched.text"))) |
| @@ -1255,9 +1258,8 @@ struct task_struct { | |||
| 1255 | /* ipc stuff */ | 1258 | /* ipc stuff */ |
| 1256 | struct sysv_sem sysvsem; | 1259 | struct sysv_sem sysvsem; |
| 1257 | #endif | 1260 | #endif |
| 1258 | #ifdef CONFIG_DETECT_SOFTLOCKUP | 1261 | #ifdef CONFIG_DETECT_HUNG_TASK |
| 1259 | /* hung task detection */ | 1262 | /* hung task detection */ |
| 1260 | unsigned long last_switch_timestamp; | ||
| 1261 | unsigned long last_switch_count; | 1263 | unsigned long last_switch_count; |
| 1262 | #endif | 1264 | #endif |
| 1263 | /* CPU-specific state of this task */ | 1265 | /* CPU-specific state of this task */ |
diff --git a/kernel/Makefile b/kernel/Makefile index bab1dffe37e9..42423665660a 100644 --- a/kernel/Makefile +++ b/kernel/Makefile | |||
| @@ -74,6 +74,7 @@ obj-$(CONFIG_AUDIT_TREE) += audit_tree.o | |||
| 74 | obj-$(CONFIG_KPROBES) += kprobes.o | 74 | obj-$(CONFIG_KPROBES) += kprobes.o |
| 75 | obj-$(CONFIG_KGDB) += kgdb.o | 75 | obj-$(CONFIG_KGDB) += kgdb.o |
| 76 | obj-$(CONFIG_DETECT_SOFTLOCKUP) += softlockup.o | 76 | obj-$(CONFIG_DETECT_SOFTLOCKUP) += softlockup.o |
| 77 | obj-$(CONFIG_DETECT_HUNG_TASK) += hung_task.o | ||
| 77 | obj-$(CONFIG_GENERIC_HARDIRQS) += irq/ | 78 | obj-$(CONFIG_GENERIC_HARDIRQS) += irq/ |
| 78 | obj-$(CONFIG_SECCOMP) += seccomp.o | 79 | obj-$(CONFIG_SECCOMP) += seccomp.o |
| 79 | obj-$(CONFIG_RCU_TORTURE_TEST) += rcutorture.o | 80 | obj-$(CONFIG_RCU_TORTURE_TEST) += rcutorture.o |
diff --git a/kernel/fork.c b/kernel/fork.c index 660c2b8765bc..989c7c202b3d 100644 --- a/kernel/fork.c +++ b/kernel/fork.c | |||
| @@ -645,6 +645,9 @@ static int copy_mm(unsigned long clone_flags, struct task_struct * tsk) | |||
| 645 | 645 | ||
| 646 | tsk->min_flt = tsk->maj_flt = 0; | 646 | tsk->min_flt = tsk->maj_flt = 0; |
| 647 | tsk->nvcsw = tsk->nivcsw = 0; | 647 | tsk->nvcsw = tsk->nivcsw = 0; |
| 648 | #ifdef CONFIG_DETECT_HUNG_TASK | ||
| 649 | tsk->last_switch_count = tsk->nvcsw + tsk->nivcsw; | ||
| 650 | #endif | ||
| 648 | 651 | ||
| 649 | tsk->mm = NULL; | 652 | tsk->mm = NULL; |
| 650 | tsk->active_mm = NULL; | 653 | tsk->active_mm = NULL; |
| @@ -1032,11 +1035,6 @@ static struct task_struct *copy_process(unsigned long clone_flags, | |||
| 1032 | 1035 | ||
| 1033 | p->default_timer_slack_ns = current->timer_slack_ns; | 1036 | p->default_timer_slack_ns = current->timer_slack_ns; |
| 1034 | 1037 | ||
| 1035 | #ifdef CONFIG_DETECT_SOFTLOCKUP | ||
| 1036 | p->last_switch_count = 0; | ||
| 1037 | p->last_switch_timestamp = 0; | ||
| 1038 | #endif | ||
| 1039 | |||
| 1040 | task_io_accounting_init(&p->ioac); | 1038 | task_io_accounting_init(&p->ioac); |
| 1041 | acct_clear_integrals(p); | 1039 | acct_clear_integrals(p); |
| 1042 | 1040 | ||
diff --git a/kernel/hung_task.c b/kernel/hung_task.c new file mode 100644 index 000000000000..022a4927b785 --- /dev/null +++ b/kernel/hung_task.c | |||
| @@ -0,0 +1,217 @@ | |||
| 1 | /* | ||
| 2 | * Detect Hung Task | ||
| 3 | * | ||
| 4 | * kernel/hung_task.c - kernel thread for detecting tasks stuck in D state | ||
| 5 | * | ||
| 6 | */ | ||
| 7 | |||
| 8 | #include <linux/mm.h> | ||
| 9 | #include <linux/cpu.h> | ||
| 10 | #include <linux/nmi.h> | ||
| 11 | #include <linux/init.h> | ||
| 12 | #include <linux/delay.h> | ||
| 13 | #include <linux/freezer.h> | ||
| 14 | #include <linux/kthread.h> | ||
| 15 | #include <linux/lockdep.h> | ||
| 16 | #include <linux/module.h> | ||
| 17 | #include <linux/sysctl.h> | ||
| 18 | |||
| 19 | /* | ||
| 20 | * The number of tasks checked: | ||
| 21 | */ | ||
| 22 | unsigned long __read_mostly sysctl_hung_task_check_count = PID_MAX_LIMIT; | ||
| 23 | |||
| 24 | /* | ||
| 25 | * Limit number of tasks checked in a batch. | ||
| 26 | * | ||
| 27 | * This value controls the preemptibility of khungtaskd since preemption | ||
| 28 | * is disabled during the critical section. It also controls the size of | ||
| 29 | * the RCU grace period. So it needs to be upper-bound. | ||
| 30 | */ | ||
| 31 | #define HUNG_TASK_BATCHING 1024 | ||
| 32 | |||
| 33 | /* | ||
| 34 | * Zero means infinite timeout - no checking done: | ||
| 35 | */ | ||
| 36 | unsigned long __read_mostly sysctl_hung_task_timeout_secs = 120; | ||
| 37 | |||
| 38 | unsigned long __read_mostly sysctl_hung_task_warnings = 10; | ||
| 39 | |||
| 40 | static int __read_mostly did_panic; | ||
| 41 | |||
| 42 | static struct task_struct *watchdog_task; | ||
| 43 | |||
| 44 | /* | ||
| 45 | * Should we panic (and reboot, if panic_timeout= is set) when a | ||
| 46 | * hung task is detected: | ||
| 47 | */ | ||
| 48 | unsigned int __read_mostly sysctl_hung_task_panic = | ||
| 49 | CONFIG_BOOTPARAM_HUNG_TASK_PANIC_VALUE; | ||
| 50 | |||
| 51 | static int __init hung_task_panic_setup(char *str) | ||
| 52 | { | ||
| 53 | sysctl_hung_task_panic = simple_strtoul(str, NULL, 0); | ||
| 54 | |||
| 55 | return 1; | ||
| 56 | } | ||
| 57 | __setup("hung_task_panic=", hung_task_panic_setup); | ||
| 58 | |||
| 59 | static int | ||
| 60 | hung_task_panic(struct notifier_block *this, unsigned long event, void *ptr) | ||
| 61 | { | ||
| 62 | did_panic = 1; | ||
| 63 | |||
| 64 | return NOTIFY_DONE; | ||
| 65 | } | ||
| 66 | |||
| 67 | static struct notifier_block panic_block = { | ||
| 68 | .notifier_call = hung_task_panic, | ||
| 69 | }; | ||
| 70 | |||
| 71 | static void check_hung_task(struct task_struct *t, unsigned long timeout) | ||
| 72 | { | ||
| 73 | unsigned long switch_count = t->nvcsw + t->nivcsw; | ||
| 74 | |||
| 75 | /* | ||
| 76 | * Ensure the task is not frozen. | ||
| 77 | * Also, when a freshly created task is scheduled once, changes | ||
| 78 | * its state to TASK_UNINTERRUPTIBLE without having ever been | ||
| 79 | * switched out once, it musn't be checked. | ||
| 80 | */ | ||
| 81 | if (unlikely(t->flags & PF_FROZEN || !switch_count)) | ||
| 82 | return; | ||
| 83 | |||
| 84 | if (switch_count != t->last_switch_count) { | ||
| 85 | t->last_switch_count = switch_count; | ||
| 86 | return; | ||
| 87 | } | ||
| 88 | if (!sysctl_hung_task_warnings) | ||
| 89 | return; | ||
| 90 | sysctl_hung_task_warnings--; | ||
| 91 | |||
| 92 | /* | ||
| 93 | * Ok, the task did not get scheduled for more than 2 minutes, | ||
| 94 | * complain: | ||
| 95 | */ | ||
| 96 | printk(KERN_ERR "INFO: task %s:%d blocked for more than " | ||
| 97 | "%ld seconds.\n", t->comm, t->pid, timeout); | ||
| 98 | printk(KERN_ERR "\"echo 0 > /proc/sys/kernel/hung_task_timeout_secs\"" | ||
| 99 | " disables this message.\n"); | ||
| 100 | sched_show_task(t); | ||
| 101 | __debug_show_held_locks(t); | ||
| 102 | |||
| 103 | touch_nmi_watchdog(); | ||
| 104 | |||
| 105 | if (sysctl_hung_task_panic) | ||
| 106 | panic("hung_task: blocked tasks"); | ||
| 107 | } | ||
| 108 | |||
| 109 | /* | ||
| 110 | * To avoid extending the RCU grace period for an unbounded amount of time, | ||
| 111 | * periodically exit the critical section and enter a new one. | ||
| 112 | * | ||
| 113 | * For preemptible RCU it is sufficient to call rcu_read_unlock in order | ||
| 114 | * exit the grace period. For classic RCU, a reschedule is required. | ||
| 115 | */ | ||
| 116 | static void rcu_lock_break(struct task_struct *g, struct task_struct *t) | ||
| 117 | { | ||
| 118 | get_task_struct(g); | ||
| 119 | get_task_struct(t); | ||
| 120 | rcu_read_unlock(); | ||
| 121 | cond_resched(); | ||
| 122 | rcu_read_lock(); | ||
| 123 | put_task_struct(t); | ||
| 124 | put_task_struct(g); | ||
| 125 | } | ||
| 126 | |||
| 127 | /* | ||
| 128 | * Check whether a TASK_UNINTERRUPTIBLE does not get woken up for | ||
| 129 | * a really long time (120 seconds). If that happens, print out | ||
| 130 | * a warning. | ||
| 131 | */ | ||
| 132 | static void check_hung_uninterruptible_tasks(unsigned long timeout) | ||
| 133 | { | ||
| 134 | int max_count = sysctl_hung_task_check_count; | ||
| 135 | int batch_count = HUNG_TASK_BATCHING; | ||
| 136 | struct task_struct *g, *t; | ||
| 137 | |||
| 138 | /* | ||
| 139 | * If the system crashed already then all bets are off, | ||
| 140 | * do not report extra hung tasks: | ||
| 141 | */ | ||
| 142 | if (test_taint(TAINT_DIE) || did_panic) | ||
| 143 | return; | ||
| 144 | |||
| 145 | rcu_read_lock(); | ||
| 146 | do_each_thread(g, t) { | ||
| 147 | if (!--max_count) | ||
| 148 | goto unlock; | ||
| 149 | if (!--batch_count) { | ||
| 150 | batch_count = HUNG_TASK_BATCHING; | ||
| 151 | rcu_lock_break(g, t); | ||
| 152 | /* Exit if t or g was unhashed during refresh. */ | ||
| 153 | if (t->state == TASK_DEAD || g->state == TASK_DEAD) | ||
| 154 | goto unlock; | ||
| 155 | } | ||
| 156 | /* use "==" to skip the TASK_KILLABLE tasks waiting on NFS */ | ||
| 157 | if (t->state == TASK_UNINTERRUPTIBLE) | ||
| 158 | check_hung_task(t, timeout); | ||
| 159 | } while_each_thread(g, t); | ||
| 160 | unlock: | ||
| 161 | rcu_read_unlock(); | ||
| 162 | } | ||
| 163 | |||
| 164 | static unsigned long timeout_jiffies(unsigned long timeout) | ||
| 165 | { | ||
| 166 | /* timeout of 0 will disable the watchdog */ | ||
| 167 | return timeout ? timeout * HZ : MAX_SCHEDULE_TIMEOUT; | ||
| 168 | } | ||
| 169 | |||
| 170 | /* | ||
| 171 | * Process updating of timeout sysctl | ||
| 172 | */ | ||
| 173 | int proc_dohung_task_timeout_secs(struct ctl_table *table, int write, | ||
| 174 | struct file *filp, void __user *buffer, | ||
| 175 | size_t *lenp, loff_t *ppos) | ||
| 176 | { | ||
| 177 | int ret; | ||
| 178 | |||
| 179 | ret = proc_doulongvec_minmax(table, write, filp, buffer, lenp, ppos); | ||
| 180 | |||
| 181 | if (ret || !write) | ||
| 182 | goto out; | ||
| 183 | |||
| 184 | wake_up_process(watchdog_task); | ||
| 185 | |||
| 186 | out: | ||
| 187 | return ret; | ||
| 188 | } | ||
| 189 | |||
| 190 | /* | ||
| 191 | * kthread which checks for tasks stuck in D state | ||
| 192 | */ | ||
| 193 | static int watchdog(void *dummy) | ||
| 194 | { | ||
| 195 | set_user_nice(current, 0); | ||
| 196 | |||
| 197 | for ( ; ; ) { | ||
| 198 | unsigned long timeout = sysctl_hung_task_timeout_secs; | ||
| 199 | |||
| 200 | while (schedule_timeout_interruptible(timeout_jiffies(timeout))) | ||
| 201 | timeout = sysctl_hung_task_timeout_secs; | ||
| 202 | |||
| 203 | check_hung_uninterruptible_tasks(timeout); | ||
| 204 | } | ||
| 205 | |||
| 206 | return 0; | ||
| 207 | } | ||
| 208 | |||
| 209 | static int __init hung_task_init(void) | ||
| 210 | { | ||
| 211 | atomic_notifier_chain_register(&panic_notifier_list, &panic_block); | ||
| 212 | watchdog_task = kthread_run(watchdog, NULL, "khungtaskd"); | ||
| 213 | |||
| 214 | return 0; | ||
| 215 | } | ||
| 216 | |||
| 217 | module_init(hung_task_init); | ||
diff --git a/kernel/softlockup.c b/kernel/softlockup.c index 85d5a2455103..88796c330838 100644 --- a/kernel/softlockup.c +++ b/kernel/softlockup.c | |||
| @@ -166,97 +166,11 @@ void softlockup_tick(void) | |||
| 166 | } | 166 | } |
| 167 | 167 | ||
| 168 | /* | 168 | /* |
| 169 | * Have a reasonable limit on the number of tasks checked: | ||
| 170 | */ | ||
| 171 | unsigned long __read_mostly sysctl_hung_task_check_count = 1024; | ||
| 172 | |||
| 173 | /* | ||
| 174 | * Zero means infinite timeout - no checking done: | ||
| 175 | */ | ||
| 176 | unsigned long __read_mostly sysctl_hung_task_timeout_secs = 480; | ||
| 177 | |||
| 178 | unsigned long __read_mostly sysctl_hung_task_warnings = 10; | ||
| 179 | |||
| 180 | /* | ||
| 181 | * Only do the hung-tasks check on one CPU: | ||
| 182 | */ | ||
| 183 | static int check_cpu __read_mostly = -1; | ||
| 184 | |||
| 185 | static void check_hung_task(struct task_struct *t, unsigned long now) | ||
| 186 | { | ||
| 187 | unsigned long switch_count = t->nvcsw + t->nivcsw; | ||
| 188 | |||
| 189 | if (t->flags & PF_FROZEN) | ||
| 190 | return; | ||
| 191 | |||
| 192 | if (switch_count != t->last_switch_count || !t->last_switch_timestamp) { | ||
| 193 | t->last_switch_count = switch_count; | ||
| 194 | t->last_switch_timestamp = now; | ||
| 195 | return; | ||
| 196 | } | ||
| 197 | if ((long)(now - t->last_switch_timestamp) < | ||
| 198 | sysctl_hung_task_timeout_secs) | ||
| 199 | return; | ||
| 200 | if (!sysctl_hung_task_warnings) | ||
| 201 | return; | ||
| 202 | sysctl_hung_task_warnings--; | ||
| 203 | |||
| 204 | /* | ||
| 205 | * Ok, the task did not get scheduled for more than 2 minutes, | ||
| 206 | * complain: | ||
| 207 | */ | ||
| 208 | printk(KERN_ERR "INFO: task %s:%d blocked for more than " | ||
| 209 | "%ld seconds.\n", t->comm, t->pid, | ||
| 210 | sysctl_hung_task_timeout_secs); | ||
| 211 | printk(KERN_ERR "\"echo 0 > /proc/sys/kernel/hung_task_timeout_secs\"" | ||
| 212 | " disables this message.\n"); | ||
| 213 | sched_show_task(t); | ||
| 214 | __debug_show_held_locks(t); | ||
| 215 | |||
| 216 | t->last_switch_timestamp = now; | ||
| 217 | touch_nmi_watchdog(); | ||
| 218 | |||
| 219 | if (softlockup_panic) | ||
| 220 | panic("softlockup: blocked tasks"); | ||
| 221 | } | ||
| 222 | |||
| 223 | /* | ||
| 224 | * Check whether a TASK_UNINTERRUPTIBLE does not get woken up for | ||
| 225 | * a really long time (120 seconds). If that happens, print out | ||
| 226 | * a warning. | ||
| 227 | */ | ||
| 228 | static void check_hung_uninterruptible_tasks(int this_cpu) | ||
| 229 | { | ||
| 230 | int max_count = sysctl_hung_task_check_count; | ||
| 231 | unsigned long now = get_timestamp(this_cpu); | ||
| 232 | struct task_struct *g, *t; | ||
| 233 | |||
| 234 | /* | ||
| 235 | * If the system crashed already then all bets are off, | ||
| 236 | * do not report extra hung tasks: | ||
| 237 | */ | ||
| 238 | if (test_taint(TAINT_DIE) || did_panic) | ||
| 239 | return; | ||
| 240 | |||
| 241 | read_lock(&tasklist_lock); | ||
| 242 | do_each_thread(g, t) { | ||
| 243 | if (!--max_count) | ||
| 244 | goto unlock; | ||
| 245 | /* use "==" to skip the TASK_KILLABLE tasks waiting on NFS */ | ||
| 246 | if (t->state == TASK_UNINTERRUPTIBLE) | ||
| 247 | check_hung_task(t, now); | ||
| 248 | } while_each_thread(g, t); | ||
| 249 | unlock: | ||
| 250 | read_unlock(&tasklist_lock); | ||
| 251 | } | ||
| 252 | |||
| 253 | /* | ||
| 254 | * The watchdog thread - runs every second and touches the timestamp. | 169 | * The watchdog thread - runs every second and touches the timestamp. |
| 255 | */ | 170 | */ |
| 256 | static int watchdog(void *__bind_cpu) | 171 | static int watchdog(void *__bind_cpu) |
| 257 | { | 172 | { |
| 258 | struct sched_param param = { .sched_priority = MAX_RT_PRIO-1 }; | 173 | struct sched_param param = { .sched_priority = MAX_RT_PRIO-1 }; |
| 259 | int this_cpu = (long)__bind_cpu; | ||
| 260 | 174 | ||
| 261 | sched_setscheduler(current, SCHED_FIFO, ¶m); | 175 | sched_setscheduler(current, SCHED_FIFO, ¶m); |
| 262 | 176 | ||
| @@ -276,11 +190,6 @@ static int watchdog(void *__bind_cpu) | |||
| 276 | if (kthread_should_stop()) | 190 | if (kthread_should_stop()) |
| 277 | break; | 191 | break; |
| 278 | 192 | ||
| 279 | if (this_cpu == check_cpu) { | ||
| 280 | if (sysctl_hung_task_timeout_secs) | ||
| 281 | check_hung_uninterruptible_tasks(this_cpu); | ||
| 282 | } | ||
| 283 | |||
| 284 | set_current_state(TASK_INTERRUPTIBLE); | 193 | set_current_state(TASK_INTERRUPTIBLE); |
| 285 | } | 194 | } |
| 286 | __set_current_state(TASK_RUNNING); | 195 | __set_current_state(TASK_RUNNING); |
| @@ -312,18 +221,9 @@ cpu_callback(struct notifier_block *nfb, unsigned long action, void *hcpu) | |||
| 312 | break; | 221 | break; |
| 313 | case CPU_ONLINE: | 222 | case CPU_ONLINE: |
| 314 | case CPU_ONLINE_FROZEN: | 223 | case CPU_ONLINE_FROZEN: |
| 315 | check_cpu = cpumask_any(cpu_online_mask); | ||
| 316 | wake_up_process(per_cpu(watchdog_task, hotcpu)); | 224 | wake_up_process(per_cpu(watchdog_task, hotcpu)); |
| 317 | break; | 225 | break; |
| 318 | #ifdef CONFIG_HOTPLUG_CPU | 226 | #ifdef CONFIG_HOTPLUG_CPU |
| 319 | case CPU_DOWN_PREPARE: | ||
| 320 | case CPU_DOWN_PREPARE_FROZEN: | ||
| 321 | if (hotcpu == check_cpu) { | ||
| 322 | /* Pick any other online cpu. */ | ||
| 323 | check_cpu = cpumask_any_but(cpu_online_mask, hotcpu); | ||
| 324 | } | ||
| 325 | break; | ||
| 326 | |||
| 327 | case CPU_UP_CANCELED: | 227 | case CPU_UP_CANCELED: |
| 328 | case CPU_UP_CANCELED_FROZEN: | 228 | case CPU_UP_CANCELED_FROZEN: |
| 329 | if (!per_cpu(watchdog_task, hotcpu)) | 229 | if (!per_cpu(watchdog_task, hotcpu)) |
diff --git a/kernel/sysctl.c b/kernel/sysctl.c index 72eb1a41dcab..4286b62b34a0 100644 --- a/kernel/sysctl.c +++ b/kernel/sysctl.c | |||
| @@ -814,6 +814,19 @@ static struct ctl_table kern_table[] = { | |||
| 814 | .extra1 = &neg_one, | 814 | .extra1 = &neg_one, |
| 815 | .extra2 = &sixty, | 815 | .extra2 = &sixty, |
| 816 | }, | 816 | }, |
| 817 | #endif | ||
| 818 | #ifdef CONFIG_DETECT_HUNG_TASK | ||
| 819 | { | ||
| 820 | .ctl_name = CTL_UNNUMBERED, | ||
| 821 | .procname = "hung_task_panic", | ||
| 822 | .data = &sysctl_hung_task_panic, | ||
| 823 | .maxlen = sizeof(int), | ||
| 824 | .mode = 0644, | ||
| 825 | .proc_handler = &proc_dointvec_minmax, | ||
| 826 | .strategy = &sysctl_intvec, | ||
| 827 | .extra1 = &zero, | ||
| 828 | .extra2 = &one, | ||
| 829 | }, | ||
| 817 | { | 830 | { |
| 818 | .ctl_name = CTL_UNNUMBERED, | 831 | .ctl_name = CTL_UNNUMBERED, |
| 819 | .procname = "hung_task_check_count", | 832 | .procname = "hung_task_check_count", |
| @@ -829,7 +842,7 @@ static struct ctl_table kern_table[] = { | |||
| 829 | .data = &sysctl_hung_task_timeout_secs, | 842 | .data = &sysctl_hung_task_timeout_secs, |
| 830 | .maxlen = sizeof(unsigned long), | 843 | .maxlen = sizeof(unsigned long), |
| 831 | .mode = 0644, | 844 | .mode = 0644, |
| 832 | .proc_handler = &proc_doulongvec_minmax, | 845 | .proc_handler = &proc_dohung_task_timeout_secs, |
| 833 | .strategy = &sysctl_intvec, | 846 | .strategy = &sysctl_intvec, |
| 834 | }, | 847 | }, |
| 835 | { | 848 | { |
diff --git a/lib/Kconfig.debug b/lib/Kconfig.debug index 9638d99644af..c6e854f215fa 100644 --- a/lib/Kconfig.debug +++ b/lib/Kconfig.debug | |||
| @@ -186,6 +186,44 @@ config BOOTPARAM_SOFTLOCKUP_PANIC_VALUE | |||
| 186 | default 0 if !BOOTPARAM_SOFTLOCKUP_PANIC | 186 | default 0 if !BOOTPARAM_SOFTLOCKUP_PANIC |
| 187 | default 1 if BOOTPARAM_SOFTLOCKUP_PANIC | 187 | default 1 if BOOTPARAM_SOFTLOCKUP_PANIC |
| 188 | 188 | ||
| 189 | config DETECT_HUNG_TASK | ||
| 190 | bool "Detect Hung Tasks" | ||
| 191 | depends on DEBUG_KERNEL | ||
| 192 | default DETECT_SOFTLOCKUP | ||
| 193 | help | ||
| 194 | Say Y here to enable the kernel to detect "hung tasks", | ||
| 195 | which are bugs that cause the task to be stuck in | ||
| 196 | uninterruptible "D" state indefinitiley. | ||
| 197 | |||
| 198 | When a hung task is detected, the kernel will print the | ||
| 199 | current stack trace (which you should report), but the | ||
| 200 | task will stay in uninterruptible state. If lockdep is | ||
| 201 | enabled then all held locks will also be reported. This | ||
| 202 | feature has negligible overhead. | ||
| 203 | |||
| 204 | config BOOTPARAM_HUNG_TASK_PANIC | ||
| 205 | bool "Panic (Reboot) On Hung Tasks" | ||
| 206 | depends on DETECT_HUNG_TASK | ||
| 207 | help | ||
| 208 | Say Y here to enable the kernel to panic on "hung tasks", | ||
| 209 | which are bugs that cause the kernel to leave a task stuck | ||
| 210 | in uninterruptible "D" state. | ||
| 211 | |||
| 212 | The panic can be used in combination with panic_timeout, | ||
| 213 | to cause the system to reboot automatically after a | ||
| 214 | hung task has been detected. This feature is useful for | ||
| 215 | high-availability systems that have uptime guarantees and | ||
| 216 | where a hung tasks must be resolved ASAP. | ||
| 217 | |||
| 218 | Say N if unsure. | ||
| 219 | |||
| 220 | config BOOTPARAM_HUNG_TASK_PANIC_VALUE | ||
| 221 | int | ||
| 222 | depends on DETECT_HUNG_TASK | ||
| 223 | range 0 1 | ||
| 224 | default 0 if !BOOTPARAM_HUNG_TASK_PANIC | ||
| 225 | default 1 if BOOTPARAM_HUNG_TASK_PANIC | ||
| 226 | |||
| 189 | config SCHED_DEBUG | 227 | config SCHED_DEBUG |
| 190 | bool "Collect scheduler debugging info" | 228 | bool "Collect scheduler debugging info" |
| 191 | depends on DEBUG_KERNEL && PROC_FS | 229 | depends on DEBUG_KERNEL && PROC_FS |
