diff options
Diffstat (limited to 'kernel/watchdog.c')
-rw-r--r-- | kernel/watchdog.c | 52 |
1 files changed, 30 insertions, 22 deletions
diff --git a/kernel/watchdog.c b/kernel/watchdog.c index 14733d4d156b..6e63097fa73a 100644 --- a/kernel/watchdog.c +++ b/kernel/watchdog.c | |||
@@ -28,7 +28,7 @@ | |||
28 | #include <linux/perf_event.h> | 28 | #include <linux/perf_event.h> |
29 | 29 | ||
30 | int watchdog_enabled = 1; | 30 | int watchdog_enabled = 1; |
31 | int __read_mostly softlockup_thresh = 60; | 31 | int __read_mostly watchdog_thresh = 10; |
32 | 32 | ||
33 | static DEFINE_PER_CPU(unsigned long, watchdog_touch_ts); | 33 | static DEFINE_PER_CPU(unsigned long, watchdog_touch_ts); |
34 | static DEFINE_PER_CPU(struct task_struct *, softlockup_watchdog); | 34 | static DEFINE_PER_CPU(struct task_struct *, softlockup_watchdog); |
@@ -91,6 +91,17 @@ static int __init nosoftlockup_setup(char *str) | |||
91 | __setup("nosoftlockup", nosoftlockup_setup); | 91 | __setup("nosoftlockup", nosoftlockup_setup); |
92 | /* */ | 92 | /* */ |
93 | 93 | ||
94 | /* | ||
95 | * Hard-lockup warnings should be triggered after just a few seconds. Soft- | ||
96 | * lockups can have false positives under extreme conditions. So we generally | ||
97 | * want a higher threshold for soft lockups than for hard lockups. So we couple | ||
98 | * the thresholds with a factor: we make the soft threshold twice the amount of | ||
99 | * time the hard threshold is. | ||
100 | */ | ||
101 | static int get_softlockup_thresh() | ||
102 | { | ||
103 | return watchdog_thresh * 2; | ||
104 | } | ||
94 | 105 | ||
95 | /* | 106 | /* |
96 | * Returns seconds, approximately. We don't need nanosecond | 107 | * Returns seconds, approximately. We don't need nanosecond |
@@ -105,12 +116,12 @@ static unsigned long get_timestamp(int this_cpu) | |||
105 | static unsigned long get_sample_period(void) | 116 | static unsigned long get_sample_period(void) |
106 | { | 117 | { |
107 | /* | 118 | /* |
108 | * convert softlockup_thresh from seconds to ns | 119 | * convert watchdog_thresh from seconds to ns |
109 | * the divide by 5 is to give hrtimer 5 chances to | 120 | * the divide by 5 is to give hrtimer 5 chances to |
110 | * increment before the hardlockup detector generates | 121 | * increment before the hardlockup detector generates |
111 | * a warning | 122 | * a warning |
112 | */ | 123 | */ |
113 | return softlockup_thresh / 5 * NSEC_PER_SEC; | 124 | return get_softlockup_thresh() * (NSEC_PER_SEC / 5); |
114 | } | 125 | } |
115 | 126 | ||
116 | /* Commands for resetting the watchdog */ | 127 | /* Commands for resetting the watchdog */ |
@@ -182,7 +193,7 @@ static int is_softlockup(unsigned long touch_ts) | |||
182 | unsigned long now = get_timestamp(smp_processor_id()); | 193 | unsigned long now = get_timestamp(smp_processor_id()); |
183 | 194 | ||
184 | /* Warn about unreasonable delays: */ | 195 | /* Warn about unreasonable delays: */ |
185 | if (time_after(now, touch_ts + softlockup_thresh)) | 196 | if (time_after(now, touch_ts + get_softlockup_thresh())) |
186 | return now - touch_ts; | 197 | return now - touch_ts; |
187 | 198 | ||
188 | return 0; | 199 | return 0; |
@@ -359,7 +370,7 @@ static int watchdog_nmi_enable(int cpu) | |||
359 | 370 | ||
360 | /* Try to register using hardware perf events */ | 371 | /* Try to register using hardware perf events */ |
361 | wd_attr = &wd_hw_attr; | 372 | wd_attr = &wd_hw_attr; |
362 | wd_attr->sample_period = hw_nmi_get_sample_period(); | 373 | wd_attr->sample_period = hw_nmi_get_sample_period(watchdog_thresh); |
363 | event = perf_event_create_kernel_counter(wd_attr, cpu, NULL, watchdog_overflow_callback); | 374 | event = perf_event_create_kernel_counter(wd_attr, cpu, NULL, watchdog_overflow_callback); |
364 | if (!IS_ERR(event)) { | 375 | if (!IS_ERR(event)) { |
365 | printk(KERN_INFO "NMI watchdog enabled, takes one hw-pmu counter.\n"); | 376 | printk(KERN_INFO "NMI watchdog enabled, takes one hw-pmu counter.\n"); |
@@ -501,28 +512,25 @@ static void watchdog_disable_all_cpus(void) | |||
501 | /* sysctl functions */ | 512 | /* sysctl functions */ |
502 | #ifdef CONFIG_SYSCTL | 513 | #ifdef CONFIG_SYSCTL |
503 | /* | 514 | /* |
504 | * proc handler for /proc/sys/kernel/nmi_watchdog | 515 | * proc handler for /proc/sys/kernel/nmi_watchdog,watchdog_thresh |
505 | */ | 516 | */ |
506 | 517 | ||
507 | int proc_dowatchdog_enabled(struct ctl_table *table, int write, | 518 | int proc_dowatchdog(struct ctl_table *table, int write, |
508 | void __user *buffer, size_t *length, loff_t *ppos) | 519 | void __user *buffer, size_t *lenp, loff_t *ppos) |
509 | { | 520 | { |
510 | proc_dointvec(table, write, buffer, length, ppos); | 521 | int ret; |
511 | 522 | ||
512 | if (write) { | 523 | ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos); |
513 | if (watchdog_enabled) | 524 | if (ret || !write) |
514 | watchdog_enable_all_cpus(); | 525 | goto out; |
515 | else | ||
516 | watchdog_disable_all_cpus(); | ||
517 | } | ||
518 | return 0; | ||
519 | } | ||
520 | 526 | ||
521 | int proc_dowatchdog_thresh(struct ctl_table *table, int write, | 527 | if (watchdog_enabled && watchdog_thresh) |
522 | void __user *buffer, | 528 | watchdog_enable_all_cpus(); |
523 | size_t *lenp, loff_t *ppos) | 529 | else |
524 | { | 530 | watchdog_disable_all_cpus(); |
525 | return proc_dointvec_minmax(table, write, buffer, lenp, ppos); | 531 | |
532 | out: | ||
533 | return ret; | ||
526 | } | 534 | } |
527 | #endif /* CONFIG_SYSCTL */ | 535 | #endif /* CONFIG_SYSCTL */ |
528 | 536 | ||