diff options
author | Linus Torvalds <torvalds@linux-foundation.org> | 2015-04-13 16:33:20 -0400 |
---|---|---|
committer | Linus Torvalds <torvalds@linux-foundation.org> | 2015-04-13 16:33:20 -0400 |
commit | 07f2d8c63fa439613405760841e41fce3041023f (patch) | |
tree | d5ef4a4f4d00b6a04ff2a59e66a6c7238008dff5 | |
parent | ee799f41eb2bc0484711b0fc942fddf54248289f (diff) | |
parent | cee8f5a6c8c917613dd021552909d071b1dab592 (diff) |
Merge branch 'x86-ras-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip
Pull x86 RAS changes from Ingo Molnar:
"The main changes in this cycle were:
- Simplify the CMCI storm logic on Intel CPUs after yet another
report about a race in the code (Borislav Petkov)
- Enable the MCE threshold irq on AMD CPUs by default (Aravind
Gopalakrishnan)
- Add AMD-specific MCE-severity grading function. Further error
recovery actions will be based on its output (Aravind Gopalakrishnan)
- Documentation updates (Borislav Petkov)
- ... assorted fixes and cleanups"
* 'x86-ras-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tip/tip:
x86/mce/severity: Fix warning about indented braces
x86/mce: Define mce_severity function pointer
x86/mce: Add an AMD severities-grading function
x86/mce: Reindent __mcheck_cpu_apply_quirks() properly
x86/mce: Use safe MSR accesses for AMD quirk
x86/MCE/AMD: Enable thresholding interrupts by default if supported
x86/MCE: Make mce_panic() fatal machine check msg in the same pattern
x86/MCE/intel: Cleanup CMCI storm logic
Documentation/acpi/einj: Correct and streamline text
x86/MCE/AMD: Drop bogus const modifier from AMD's bank4_names()
-rw-r--r-- | Documentation/acpi/apei/einj.txt | 196 | ||||
-rw-r--r-- | arch/x86/include/asm/mce.h | 16 | ||||
-rw-r--r-- | arch/x86/kernel/cpu/mcheck/mce-internal.h | 11 | ||||
-rw-r--r-- | arch/x86/kernel/cpu/mcheck/mce-severity.c | 66 | ||||
-rw-r--r-- | arch/x86/kernel/cpu/mcheck/mce.c | 154 | ||||
-rw-r--r-- | arch/x86/kernel/cpu/mcheck/mce_amd.c | 11 | ||||
-rw-r--r-- | arch/x86/kernel/cpu/mcheck/mce_intel.c | 63 |
7 files changed, 336 insertions, 181 deletions
diff --git a/Documentation/acpi/apei/einj.txt b/Documentation/acpi/apei/einj.txt index f51861bcb07b..e550c8b98139 100644 --- a/Documentation/acpi/apei/einj.txt +++ b/Documentation/acpi/apei/einj.txt | |||
@@ -1,129 +1,177 @@ | |||
1 | APEI Error INJection | 1 | APEI Error INJection |
2 | ~~~~~~~~~~~~~~~~~~~~ | 2 | ~~~~~~~~~~~~~~~~~~~~ |
3 | 3 | ||
4 | EINJ provides a hardware error injection mechanism | 4 | EINJ provides a hardware error injection mechanism. It is very useful |
5 | It is very useful for debugging and testing of other APEI and RAS features. | 5 | for debugging and testing APEI and RAS features in general. |
6 | 6 | ||
7 | To use EINJ, make sure the following are enabled in your kernel | 7 | You need to check whether your BIOS supports EINJ first. For that, look |
8 | for early boot messages similar to this one: | ||
9 | |||
10 | ACPI: EINJ 0x000000007370A000 000150 (v01 INTEL 00000001 INTL 00000001) | ||
11 | |||
12 | which shows that the BIOS is exposing an EINJ table - it is the | ||
13 | mechanism through which the injection is done. | ||
14 | |||
15 | Alternatively, look in /sys/firmware/acpi/tables for an "EINJ" file, | ||
16 | which is a different representation of the same thing. | ||
17 | |||
18 | It doesn't necessarily mean that EINJ is not supported if those above | ||
19 | don't exist: before you give up, go into BIOS setup to see if the BIOS | ||
20 | has an option to enable error injection. Look for something called WHEA | ||
21 | or similar. Often, you need to enable an ACPI5 support option prior, in | ||
22 | order to see the APEI,EINJ,... functionality supported and exposed by | ||
23 | the BIOS menu. | ||
24 | |||
25 | To use EINJ, make sure the following are options enabled in your kernel | ||
8 | configuration: | 26 | configuration: |
9 | 27 | ||
10 | CONFIG_DEBUG_FS | 28 | CONFIG_DEBUG_FS |
11 | CONFIG_ACPI_APEI | 29 | CONFIG_ACPI_APEI |
12 | CONFIG_ACPI_APEI_EINJ | 30 | CONFIG_ACPI_APEI_EINJ |
13 | 31 | ||
14 | The user interface of EINJ is debug file system, under the | 32 | The EINJ user interface is in <debugfs mount point>/apei/einj. |
15 | directory apei/einj. The following files are provided. | 33 | |
34 | The following files belong to it: | ||
16 | 35 | ||
17 | - available_error_type | 36 | - available_error_type |
18 | Reading this file returns the error injection capability of the | 37 | |
19 | platform, that is, which error types are supported. The error type | 38 | This file shows which error types are supported: |
20 | definition is as follow, the left field is the error type value, the | 39 | |
21 | right field is error description. | 40 | Error Type Value Error Description |
22 | 41 | ================ ================= | |
23 | 0x00000001 Processor Correctable | 42 | 0x00000001 Processor Correctable |
24 | 0x00000002 Processor Uncorrectable non-fatal | 43 | 0x00000002 Processor Uncorrectable non-fatal |
25 | 0x00000004 Processor Uncorrectable fatal | 44 | 0x00000004 Processor Uncorrectable fatal |
26 | 0x00000008 Memory Correctable | 45 | 0x00000008 Memory Correctable |
27 | 0x00000010 Memory Uncorrectable non-fatal | 46 | 0x00000010 Memory Uncorrectable non-fatal |
28 | 0x00000020 Memory Uncorrectable fatal | 47 | 0x00000020 Memory Uncorrectable fatal |
29 | 0x00000040 PCI Express Correctable | 48 | 0x00000040 PCI Express Correctable |
30 | 0x00000080 PCI Express Uncorrectable fatal | 49 | 0x00000080 PCI Express Uncorrectable fatal |
31 | 0x00000100 PCI Express Uncorrectable non-fatal | 50 | 0x00000100 PCI Express Uncorrectable non-fatal |
32 | 0x00000200 Platform Correctable | 51 | 0x00000200 Platform Correctable |
33 | 0x00000400 Platform Uncorrectable non-fatal | 52 | 0x00000400 Platform Uncorrectable non-fatal |
34 | 0x00000800 Platform Uncorrectable fatal | 53 | 0x00000800 Platform Uncorrectable fatal |
35 | 54 | ||
36 | The format of file contents are as above, except there are only the | 55 | The format of the file contents are as above, except present are only |
37 | available error type lines. | 56 | the available error types. |
38 | 57 | ||
39 | - error_type | 58 | - error_type |
40 | This file is used to set the error type value. The error type value | 59 | |
41 | is defined in "available_error_type" description. | 60 | Set the value of the error type being injected. Possible error types |
61 | are defined in the file available_error_type above. | ||
42 | 62 | ||
43 | - error_inject | 63 | - error_inject |
44 | Write any integer to this file to trigger the error | 64 | |
45 | injection. Before this, please specify all necessary error | 65 | Write any integer to this file to trigger the error injection. Make |
46 | parameters. | 66 | sure you have specified all necessary error parameters, i.e. this |
67 | write should be the last step when injecting errors. | ||
47 | 68 | ||
48 | - flags | 69 | - flags |
49 | Present for kernel version 3.13 and above. Used to specify which | 70 | |
50 | of param{1..4} are valid and should be used by BIOS during injection. | 71 | Present for kernel versions 3.13 and above. Used to specify which |
51 | Value is a bitmask as specified in ACPI5.0 spec for the | 72 | of param{1..4} are valid and should be used by the firmware during |
73 | injection. Value is a bitmask as specified in ACPI5.0 spec for the | ||
52 | SET_ERROR_TYPE_WITH_ADDRESS data structure: | 74 | SET_ERROR_TYPE_WITH_ADDRESS data structure: |
53 | Bit 0 - Processor APIC field valid (see param3 below) | 75 | |
54 | Bit 1 - Memory address and mask valid (param1 and param2) | 76 | Bit 0 - Processor APIC field valid (see param3 below). |
55 | Bit 2 - PCIe (seg,bus,dev,fn) valid (param4 below) | 77 | Bit 1 - Memory address and mask valid (param1 and param2). |
56 | If set to zero, legacy behaviour is used where the type of injection | 78 | Bit 2 - PCIe (seg,bus,dev,fn) valid (see param4 below). |
57 | specifies just one bit set, and param1 is multiplexed. | 79 | |
80 | If set to zero, legacy behavior is mimicked where the type of | ||
81 | injection specifies just one bit set, and param1 is multiplexed. | ||
58 | 82 | ||
59 | - param1 | 83 | - param1 |
60 | This file is used to set the first error parameter value. Effect of | 84 | |
61 | parameter depends on error_type specified. For example, if error | 85 | This file is used to set the first error parameter value. Its effect |
62 | type is memory related type, the param1 should be a valid physical | 86 | depends on the error type specified in error_type. For example, if |
63 | memory address. [Unless "flag" is set - see above] | 87 | error type is memory related type, the param1 should be a valid |
88 | physical memory address. [Unless "flag" is set - see above] | ||
64 | 89 | ||
65 | - param2 | 90 | - param2 |
66 | This file is used to set the second error parameter value. Effect of | 91 | |
67 | parameter depends on error_type specified. For example, if error | 92 | Same use as param1 above. For example, if error type is of memory |
68 | type is memory related type, the param2 should be a physical memory | 93 | related type, then param2 should be a physical memory address mask. |
69 | address mask. Linux requires page or narrower granularity, say, | 94 | Linux requires page or narrower granularity, say, 0xfffffffffffff000. |
70 | 0xfffffffffffff000. | ||
71 | 95 | ||
72 | - param3 | 96 | - param3 |
73 | Used when the 0x1 bit is set in "flag" to specify the APIC id | 97 | |
98 | Used when the 0x1 bit is set in "flags" to specify the APIC id | ||
74 | 99 | ||
75 | - param4 | 100 | - param4 |
76 | Used when the 0x4 bit is set in "flag" to specify target PCIe device | 101 | Used when the 0x4 bit is set in "flags" to specify target PCIe device |
77 | 102 | ||
78 | - notrigger | 103 | - notrigger |
79 | The EINJ mechanism is a two step process. First inject the error, then | 104 | |
80 | perform some actions to trigger it. Setting "notrigger" to 1 skips the | 105 | The error injection mechanism is a two-step process. First inject the |
81 | trigger phase, which *may* allow the user to cause the error in some other | 106 | error, then perform some actions to trigger it. Setting "notrigger" |
82 | context by a simple access to the cpu, memory location, or device that is | 107 | to 1 skips the trigger phase, which *may* allow the user to cause the |
83 | the target of the error injection. Whether this actually works depends | 108 | error in some other context by a simple access to the CPU, memory |
84 | on what operations the BIOS actually includes in the trigger phase. | 109 | location, or device that is the target of the error injection. Whether |
85 | 110 | this actually works depends on what operations the BIOS actually | |
86 | BIOS versions based in the ACPI 4.0 specification have limited options | 111 | includes in the trigger phase. |
87 | to control where the errors are injected. Your BIOS may support an | 112 | |
88 | extension (enabled with the param_extension=1 module parameter, or | 113 | BIOS versions based on the ACPI 4.0 specification have limited options |
89 | boot command line einj.param_extension=1). This allows the address | 114 | in controlling where the errors are injected. Your BIOS may support an |
90 | and mask for memory injections to be specified by the param1 and | 115 | extension (enabled with the param_extension=1 module parameter, or boot |
91 | param2 files in apei/einj. | 116 | command line einj.param_extension=1). This allows the address and mask |
92 | 117 | for memory injections to be specified by the param1 and param2 files in | |
93 | BIOS versions using the ACPI 5.0 specification have more control over | 118 | apei/einj. |
94 | the target of the injection. For processor related errors (type 0x1, | 119 | |
95 | 0x2 and 0x4) the APICID of the target should be provided using the | 120 | BIOS versions based on the ACPI 5.0 specification have more control over |
96 | param1 file in apei/einj. For memory errors (type 0x8, 0x10 and 0x20) | 121 | the target of the injection. For processor-related errors (type 0x1, 0x2 |
97 | the address is set using param1 with a mask in param2 (0x0 is equivalent | 122 | and 0x4), you can set flags to 0x3 (param3 for bit 0, and param1 and |
98 | to all ones). For PCI express errors (type 0x40, 0x80 and 0x100) the | 123 | param2 for bit 1) so that you have more information added to the error |
99 | segment, bus, device and function are specified using param1: | 124 | signature being injected. The actual data passed is this: |
125 | |||
126 | memory_address = param1; | ||
127 | memory_address_range = param2; | ||
128 | apicid = param3; | ||
129 | pcie_sbdf = param4; | ||
130 | |||
131 | For memory errors (type 0x8, 0x10 and 0x20) the address is set using | ||
132 | param1 with a mask in param2 (0x0 is equivalent to all ones). For PCI | ||
133 | express errors (type 0x40, 0x80 and 0x100) the segment, bus, device and | ||
134 | function are specified using param1: | ||
100 | 135 | ||
101 | 31 24 23 16 15 11 10 8 7 0 | 136 | 31 24 23 16 15 11 10 8 7 0 |
102 | +-------------------------------------------------+ | 137 | +-------------------------------------------------+ |
103 | | segment | bus | device | function | reserved | | 138 | | segment | bus | device | function | reserved | |
104 | +-------------------------------------------------+ | 139 | +-------------------------------------------------+ |
105 | 140 | ||
106 | An ACPI 5.0 BIOS may also allow vendor specific errors to be injected. | 141 | Anyway, you get the idea, if there's doubt just take a look at the code |
142 | in drivers/acpi/apei/einj.c. | ||
143 | |||
144 | An ACPI 5.0 BIOS may also allow vendor-specific errors to be injected. | ||
107 | In this case a file named vendor will contain identifying information | 145 | In this case a file named vendor will contain identifying information |
108 | from the BIOS that hopefully will allow an application wishing to use | 146 | from the BIOS that hopefully will allow an application wishing to use |
109 | the vendor specific extension to tell that they are running on a BIOS | 147 | the vendor-specific extension to tell that they are running on a BIOS |
110 | that supports it. All vendor extensions have the 0x80000000 bit set in | 148 | that supports it. All vendor extensions have the 0x80000000 bit set in |
111 | error_type. A file vendor_flags controls the interpretation of param1 | 149 | error_type. A file vendor_flags controls the interpretation of param1 |
112 | and param2 (1 = PROCESSOR, 2 = MEMORY, 4 = PCI). See your BIOS vendor | 150 | and param2 (1 = PROCESSOR, 2 = MEMORY, 4 = PCI). See your BIOS vendor |
113 | documentation for details (and expect changes to this API if vendors | 151 | documentation for details (and expect changes to this API if vendors |
114 | creativity in using this feature expands beyond our expectations). | 152 | creativity in using this feature expands beyond our expectations). |
115 | 153 | ||
116 | Example: | 154 | |
155 | An error injection example: | ||
156 | |||
117 | # cd /sys/kernel/debug/apei/einj | 157 | # cd /sys/kernel/debug/apei/einj |
118 | # cat available_error_type # See which errors can be injected | 158 | # cat available_error_type # See which errors can be injected |
119 | 0x00000002 Processor Uncorrectable non-fatal | 159 | 0x00000002 Processor Uncorrectable non-fatal |
120 | 0x00000008 Memory Correctable | 160 | 0x00000008 Memory Correctable |
121 | 0x00000010 Memory Uncorrectable non-fatal | 161 | 0x00000010 Memory Uncorrectable non-fatal |
122 | # echo 0x12345000 > param1 # Set memory address for injection | 162 | # echo 0x12345000 > param1 # Set memory address for injection |
123 | # echo 0xfffffffffffff000 > param2 # Mask - anywhere in this page | 163 | # echo $((-1 << 12)) > param2 # Mask 0xfffffffffffff000 - anywhere in this page |
124 | # echo 0x8 > error_type # Choose correctable memory error | 164 | # echo 0x8 > error_type # Choose correctable memory error |
125 | # echo 1 > error_inject # Inject now | 165 | # echo 1 > error_inject # Inject now |
126 | 166 | ||
167 | You should see something like this in dmesg: | ||
168 | |||
169 | [22715.830801] EDAC sbridge MC3: HANDLING MCE MEMORY ERROR | ||
170 | [22715.834759] EDAC sbridge MC3: CPU 0: Machine Check Event: 0 Bank 7: 8c00004000010090 | ||
171 | [22715.834759] EDAC sbridge MC3: TSC 0 | ||
172 | [22715.834759] EDAC sbridge MC3: ADDR 12345000 EDAC sbridge MC3: MISC 144780c86 | ||
173 | [22715.834759] EDAC sbridge MC3: PROCESSOR 0:306e7 TIME 1422553404 SOCKET 0 APIC 0 | ||
174 | [22716.616173] EDAC MC3: 1 CE memory read error on CPU_SrcID#0_Channel#0_DIMM#0 (channel:0 slot:0 page:0x12345 offset:0x0 grain:32 syndrome:0x0 - area:DRAM err_code:0001:0090 socket:0 channel_mask:1 rank:0) | ||
127 | 175 | ||
128 | For more information about EINJ, please refer to ACPI specification | 176 | For more information about EINJ, please refer to ACPI specification |
129 | version 4.0, section 17.5 and ACPI 5.0, section 18.6. | 177 | version 4.0, section 17.5 and ACPI 5.0, section 18.6. |
diff --git a/arch/x86/include/asm/mce.h b/arch/x86/include/asm/mce.h index 9b3de99dc004..1f5a86d518db 100644 --- a/arch/x86/include/asm/mce.h +++ b/arch/x86/include/asm/mce.h | |||
@@ -116,6 +116,12 @@ struct mca_config { | |||
116 | u32 rip_msr; | 116 | u32 rip_msr; |
117 | }; | 117 | }; |
118 | 118 | ||
119 | struct mce_vendor_flags { | ||
120 | __u64 overflow_recov : 1, /* cpuid_ebx(80000007) */ | ||
121 | __reserved_0 : 63; | ||
122 | }; | ||
123 | extern struct mce_vendor_flags mce_flags; | ||
124 | |||
119 | extern struct mca_config mca_cfg; | 125 | extern struct mca_config mca_cfg; |
120 | extern void mce_register_decode_chain(struct notifier_block *nb); | 126 | extern void mce_register_decode_chain(struct notifier_block *nb); |
121 | extern void mce_unregister_decode_chain(struct notifier_block *nb); | 127 | extern void mce_unregister_decode_chain(struct notifier_block *nb); |
@@ -128,9 +134,11 @@ extern int mce_p5_enabled; | |||
128 | #ifdef CONFIG_X86_MCE | 134 | #ifdef CONFIG_X86_MCE |
129 | int mcheck_init(void); | 135 | int mcheck_init(void); |
130 | void mcheck_cpu_init(struct cpuinfo_x86 *c); | 136 | void mcheck_cpu_init(struct cpuinfo_x86 *c); |
137 | void mcheck_vendor_init_severity(void); | ||
131 | #else | 138 | #else |
132 | static inline int mcheck_init(void) { return 0; } | 139 | static inline int mcheck_init(void) { return 0; } |
133 | static inline void mcheck_cpu_init(struct cpuinfo_x86 *c) {} | 140 | static inline void mcheck_cpu_init(struct cpuinfo_x86 *c) {} |
141 | static inline void mcheck_vendor_init_severity(void) {} | ||
134 | #endif | 142 | #endif |
135 | 143 | ||
136 | #ifdef CONFIG_X86_ANCIENT_MCE | 144 | #ifdef CONFIG_X86_ANCIENT_MCE |
@@ -183,11 +191,11 @@ typedef DECLARE_BITMAP(mce_banks_t, MAX_NR_BANKS); | |||
183 | DECLARE_PER_CPU(mce_banks_t, mce_poll_banks); | 191 | DECLARE_PER_CPU(mce_banks_t, mce_poll_banks); |
184 | 192 | ||
185 | enum mcp_flags { | 193 | enum mcp_flags { |
186 | MCP_TIMESTAMP = (1 << 0), /* log time stamp */ | 194 | MCP_TIMESTAMP = BIT(0), /* log time stamp */ |
187 | MCP_UC = (1 << 1), /* log uncorrected errors */ | 195 | MCP_UC = BIT(1), /* log uncorrected errors */ |
188 | MCP_DONTLOG = (1 << 2), /* only clear, don't log */ | 196 | MCP_DONTLOG = BIT(2), /* only clear, don't log */ |
189 | }; | 197 | }; |
190 | void machine_check_poll(enum mcp_flags flags, mce_banks_t *b); | 198 | bool machine_check_poll(enum mcp_flags flags, mce_banks_t *b); |
191 | 199 | ||
192 | int mce_notify_irq(void); | 200 | int mce_notify_irq(void); |
193 | 201 | ||
diff --git a/arch/x86/kernel/cpu/mcheck/mce-internal.h b/arch/x86/kernel/cpu/mcheck/mce-internal.h index 10b46906767f..fe32074b865b 100644 --- a/arch/x86/kernel/cpu/mcheck/mce-internal.h +++ b/arch/x86/kernel/cpu/mcheck/mce-internal.h | |||
@@ -14,6 +14,7 @@ enum severity_level { | |||
14 | }; | 14 | }; |
15 | 15 | ||
16 | #define ATTR_LEN 16 | 16 | #define ATTR_LEN 16 |
17 | #define INITIAL_CHECK_INTERVAL 5 * 60 /* 5 minutes */ | ||
17 | 18 | ||
18 | /* One object for each MCE bank, shared by all CPUs */ | 19 | /* One object for each MCE bank, shared by all CPUs */ |
19 | struct mce_bank { | 20 | struct mce_bank { |
@@ -23,20 +24,20 @@ struct mce_bank { | |||
23 | char attrname[ATTR_LEN]; /* attribute name */ | 24 | char attrname[ATTR_LEN]; /* attribute name */ |
24 | }; | 25 | }; |
25 | 26 | ||
26 | int mce_severity(struct mce *a, int tolerant, char **msg, bool is_excp); | 27 | extern int (*mce_severity)(struct mce *a, int tolerant, char **msg, bool is_excp); |
27 | struct dentry *mce_get_debugfs_dir(void); | 28 | struct dentry *mce_get_debugfs_dir(void); |
28 | 29 | ||
29 | extern struct mce_bank *mce_banks; | 30 | extern struct mce_bank *mce_banks; |
30 | extern mce_banks_t mce_banks_ce_disabled; | 31 | extern mce_banks_t mce_banks_ce_disabled; |
31 | 32 | ||
32 | #ifdef CONFIG_X86_MCE_INTEL | 33 | #ifdef CONFIG_X86_MCE_INTEL |
33 | unsigned long mce_intel_adjust_timer(unsigned long interval); | 34 | unsigned long cmci_intel_adjust_timer(unsigned long interval); |
34 | void mce_intel_cmci_poll(void); | 35 | bool mce_intel_cmci_poll(void); |
35 | void mce_intel_hcpu_update(unsigned long cpu); | 36 | void mce_intel_hcpu_update(unsigned long cpu); |
36 | void cmci_disable_bank(int bank); | 37 | void cmci_disable_bank(int bank); |
37 | #else | 38 | #else |
38 | # define mce_intel_adjust_timer mce_adjust_timer_default | 39 | # define cmci_intel_adjust_timer mce_adjust_timer_default |
39 | static inline void mce_intel_cmci_poll(void) { } | 40 | static inline bool mce_intel_cmci_poll(void) { return false; } |
40 | static inline void mce_intel_hcpu_update(unsigned long cpu) { } | 41 | static inline void mce_intel_hcpu_update(unsigned long cpu) { } |
41 | static inline void cmci_disable_bank(int bank) { } | 42 | static inline void cmci_disable_bank(int bank) { } |
42 | #endif | 43 | #endif |
diff --git a/arch/x86/kernel/cpu/mcheck/mce-severity.c b/arch/x86/kernel/cpu/mcheck/mce-severity.c index 8bb433043a7f..9c682c222071 100644 --- a/arch/x86/kernel/cpu/mcheck/mce-severity.c +++ b/arch/x86/kernel/cpu/mcheck/mce-severity.c | |||
@@ -186,7 +186,61 @@ static int error_context(struct mce *m) | |||
186 | return ((m->cs & 3) == 3) ? IN_USER : IN_KERNEL; | 186 | return ((m->cs & 3) == 3) ? IN_USER : IN_KERNEL; |
187 | } | 187 | } |
188 | 188 | ||
189 | int mce_severity(struct mce *m, int tolerant, char **msg, bool is_excp) | 189 | /* |
190 | * See AMD Error Scope Hierarchy table in a newer BKDG. For example | ||
191 | * 49125_15h_Models_30h-3Fh_BKDG.pdf, section "RAS Features" | ||
192 | */ | ||
193 | static int mce_severity_amd(struct mce *m, int tolerant, char **msg, bool is_excp) | ||
194 | { | ||
195 | enum context ctx = error_context(m); | ||
196 | |||
197 | /* Processor Context Corrupt, no need to fumble too much, die! */ | ||
198 | if (m->status & MCI_STATUS_PCC) | ||
199 | return MCE_PANIC_SEVERITY; | ||
200 | |||
201 | if (m->status & MCI_STATUS_UC) { | ||
202 | |||
203 | /* | ||
204 | * On older systems where overflow_recov flag is not present, we | ||
205 | * should simply panic if an error overflow occurs. If | ||
206 | * overflow_recov flag is present and set, then software can try | ||
207 | * to at least kill process to prolong system operation. | ||
208 | */ | ||
209 | if (mce_flags.overflow_recov) { | ||
210 | /* software can try to contain */ | ||
211 | if (!(m->mcgstatus & MCG_STATUS_RIPV) && (ctx == IN_KERNEL)) | ||
212 | return MCE_PANIC_SEVERITY; | ||
213 | |||
214 | /* kill current process */ | ||
215 | return MCE_AR_SEVERITY; | ||
216 | } else { | ||
217 | /* at least one error was not logged */ | ||
218 | if (m->status & MCI_STATUS_OVER) | ||
219 | return MCE_PANIC_SEVERITY; | ||
220 | } | ||
221 | |||
222 | /* | ||
223 | * For any other case, return MCE_UC_SEVERITY so that we log the | ||
224 | * error and exit #MC handler. | ||
225 | */ | ||
226 | return MCE_UC_SEVERITY; | ||
227 | } | ||
228 | |||
229 | /* | ||
230 | * deferred error: poll handler catches these and adds to mce_ring so | ||
231 | * memory-failure can take recovery actions. | ||
232 | */ | ||
233 | if (m->status & MCI_STATUS_DEFERRED) | ||
234 | return MCE_DEFERRED_SEVERITY; | ||
235 | |||
236 | /* | ||
237 | * corrected error: poll handler catches these and passes responsibility | ||
238 | * of decoding the error to EDAC | ||
239 | */ | ||
240 | return MCE_KEEP_SEVERITY; | ||
241 | } | ||
242 | |||
243 | static int mce_severity_intel(struct mce *m, int tolerant, char **msg, bool is_excp) | ||
190 | { | 244 | { |
191 | enum exception excp = (is_excp ? EXCP_CONTEXT : NO_EXCP); | 245 | enum exception excp = (is_excp ? EXCP_CONTEXT : NO_EXCP); |
192 | enum context ctx = error_context(m); | 246 | enum context ctx = error_context(m); |
@@ -216,6 +270,16 @@ int mce_severity(struct mce *m, int tolerant, char **msg, bool is_excp) | |||
216 | } | 270 | } |
217 | } | 271 | } |
218 | 272 | ||
273 | /* Default to mce_severity_intel */ | ||
274 | int (*mce_severity)(struct mce *m, int tolerant, char **msg, bool is_excp) = | ||
275 | mce_severity_intel; | ||
276 | |||
277 | void __init mcheck_vendor_init_severity(void) | ||
278 | { | ||
279 | if (boot_cpu_data.x86_vendor == X86_VENDOR_AMD) | ||
280 | mce_severity = mce_severity_amd; | ||
281 | } | ||
282 | |||
219 | #ifdef CONFIG_DEBUG_FS | 283 | #ifdef CONFIG_DEBUG_FS |
220 | static void *s_start(struct seq_file *f, loff_t *pos) | 284 | static void *s_start(struct seq_file *f, loff_t *pos) |
221 | { | 285 | { |
diff --git a/arch/x86/kernel/cpu/mcheck/mce.c b/arch/x86/kernel/cpu/mcheck/mce.c index 3c036cb4a370..e535533d5ab8 100644 --- a/arch/x86/kernel/cpu/mcheck/mce.c +++ b/arch/x86/kernel/cpu/mcheck/mce.c | |||
@@ -60,11 +60,12 @@ static DEFINE_MUTEX(mce_chrdev_read_mutex); | |||
60 | #define CREATE_TRACE_POINTS | 60 | #define CREATE_TRACE_POINTS |
61 | #include <trace/events/mce.h> | 61 | #include <trace/events/mce.h> |
62 | 62 | ||
63 | #define SPINUNIT 100 /* 100ns */ | 63 | #define SPINUNIT 100 /* 100ns */ |
64 | 64 | ||
65 | DEFINE_PER_CPU(unsigned, mce_exception_count); | 65 | DEFINE_PER_CPU(unsigned, mce_exception_count); |
66 | 66 | ||
67 | struct mce_bank *mce_banks __read_mostly; | 67 | struct mce_bank *mce_banks __read_mostly; |
68 | struct mce_vendor_flags mce_flags __read_mostly; | ||
68 | 69 | ||
69 | struct mca_config mca_cfg __read_mostly = { | 70 | struct mca_config mca_cfg __read_mostly = { |
70 | .bootlog = -1, | 71 | .bootlog = -1, |
@@ -89,9 +90,6 @@ static DECLARE_WAIT_QUEUE_HEAD(mce_chrdev_wait); | |||
89 | static DEFINE_PER_CPU(struct mce, mces_seen); | 90 | static DEFINE_PER_CPU(struct mce, mces_seen); |
90 | static int cpu_missing; | 91 | static int cpu_missing; |
91 | 92 | ||
92 | /* CMCI storm detection filter */ | ||
93 | static DEFINE_PER_CPU(unsigned long, mce_polled_error); | ||
94 | |||
95 | /* | 93 | /* |
96 | * MCA banks polled by the period polling timer for corrected events. | 94 | * MCA banks polled by the period polling timer for corrected events. |
97 | * With Intel CMCI, this only has MCA banks which do not support CMCI (if any). | 95 | * With Intel CMCI, this only has MCA banks which do not support CMCI (if any). |
@@ -622,8 +620,9 @@ DEFINE_PER_CPU(unsigned, mce_poll_count); | |||
622 | * is already totally * confused. In this case it's likely it will | 620 | * is already totally * confused. In this case it's likely it will |
623 | * not fully execute the machine check handler either. | 621 | * not fully execute the machine check handler either. |
624 | */ | 622 | */ |
625 | void machine_check_poll(enum mcp_flags flags, mce_banks_t *b) | 623 | bool machine_check_poll(enum mcp_flags flags, mce_banks_t *b) |
626 | { | 624 | { |
625 | bool error_logged = false; | ||
627 | struct mce m; | 626 | struct mce m; |
628 | int severity; | 627 | int severity; |
629 | int i; | 628 | int i; |
@@ -646,7 +645,7 @@ void machine_check_poll(enum mcp_flags flags, mce_banks_t *b) | |||
646 | if (!(m.status & MCI_STATUS_VAL)) | 645 | if (!(m.status & MCI_STATUS_VAL)) |
647 | continue; | 646 | continue; |
648 | 647 | ||
649 | this_cpu_write(mce_polled_error, 1); | 648 | |
650 | /* | 649 | /* |
651 | * Uncorrected or signalled events are handled by the exception | 650 | * Uncorrected or signalled events are handled by the exception |
652 | * handler when it is enabled, so don't process those here. | 651 | * handler when it is enabled, so don't process those here. |
@@ -679,8 +678,10 @@ void machine_check_poll(enum mcp_flags flags, mce_banks_t *b) | |||
679 | * Don't get the IP here because it's unlikely to | 678 | * Don't get the IP here because it's unlikely to |
680 | * have anything to do with the actual error location. | 679 | * have anything to do with the actual error location. |
681 | */ | 680 | */ |
682 | if (!(flags & MCP_DONTLOG) && !mca_cfg.dont_log_ce) | 681 | if (!(flags & MCP_DONTLOG) && !mca_cfg.dont_log_ce) { |
682 | error_logged = true; | ||
683 | mce_log(&m); | 683 | mce_log(&m); |
684 | } | ||
684 | 685 | ||
685 | /* | 686 | /* |
686 | * Clear state for this bank. | 687 | * Clear state for this bank. |
@@ -694,6 +695,8 @@ void machine_check_poll(enum mcp_flags flags, mce_banks_t *b) | |||
694 | */ | 695 | */ |
695 | 696 | ||
696 | sync_core(); | 697 | sync_core(); |
698 | |||
699 | return error_logged; | ||
697 | } | 700 | } |
698 | EXPORT_SYMBOL_GPL(machine_check_poll); | 701 | EXPORT_SYMBOL_GPL(machine_check_poll); |
699 | 702 | ||
@@ -813,7 +816,7 @@ static void mce_reign(void) | |||
813 | * other CPUs. | 816 | * other CPUs. |
814 | */ | 817 | */ |
815 | if (m && global_worst >= MCE_PANIC_SEVERITY && mca_cfg.tolerant < 3) | 818 | if (m && global_worst >= MCE_PANIC_SEVERITY && mca_cfg.tolerant < 3) |
816 | mce_panic("Fatal Machine check", m, msg); | 819 | mce_panic("Fatal machine check", m, msg); |
817 | 820 | ||
818 | /* | 821 | /* |
819 | * For UC somewhere we let the CPU who detects it handle it. | 822 | * For UC somewhere we let the CPU who detects it handle it. |
@@ -826,7 +829,7 @@ static void mce_reign(void) | |||
826 | * source or one CPU is hung. Panic. | 829 | * source or one CPU is hung. Panic. |
827 | */ | 830 | */ |
828 | if (global_worst <= MCE_KEEP_SEVERITY && mca_cfg.tolerant < 3) | 831 | if (global_worst <= MCE_KEEP_SEVERITY && mca_cfg.tolerant < 3) |
829 | mce_panic("Machine check from unknown source", NULL, NULL); | 832 | mce_panic("Fatal machine check from unknown source", NULL, NULL); |
830 | 833 | ||
831 | /* | 834 | /* |
832 | * Now clear all the mces_seen so that they don't reappear on | 835 | * Now clear all the mces_seen so that they don't reappear on |
@@ -1258,7 +1261,7 @@ void mce_log_therm_throt_event(__u64 status) | |||
1258 | * poller finds an MCE, poll 2x faster. When the poller finds no more | 1261 | * poller finds an MCE, poll 2x faster. When the poller finds no more |
1259 | * errors, poll 2x slower (up to check_interval seconds). | 1262 | * errors, poll 2x slower (up to check_interval seconds). |
1260 | */ | 1263 | */ |
1261 | static unsigned long check_interval = 5 * 60; /* 5 minutes */ | 1264 | static unsigned long check_interval = INITIAL_CHECK_INTERVAL; |
1262 | 1265 | ||
1263 | static DEFINE_PER_CPU(unsigned long, mce_next_interval); /* in jiffies */ | 1266 | static DEFINE_PER_CPU(unsigned long, mce_next_interval); /* in jiffies */ |
1264 | static DEFINE_PER_CPU(struct timer_list, mce_timer); | 1267 | static DEFINE_PER_CPU(struct timer_list, mce_timer); |
@@ -1268,49 +1271,57 @@ static unsigned long mce_adjust_timer_default(unsigned long interval) | |||
1268 | return interval; | 1271 | return interval; |
1269 | } | 1272 | } |
1270 | 1273 | ||
1271 | static unsigned long (*mce_adjust_timer)(unsigned long interval) = | 1274 | static unsigned long (*mce_adjust_timer)(unsigned long interval) = mce_adjust_timer_default; |
1272 | mce_adjust_timer_default; | ||
1273 | 1275 | ||
1274 | static int cmc_error_seen(void) | 1276 | static void __restart_timer(struct timer_list *t, unsigned long interval) |
1275 | { | 1277 | { |
1276 | unsigned long *v = this_cpu_ptr(&mce_polled_error); | 1278 | unsigned long when = jiffies + interval; |
1279 | unsigned long flags; | ||
1280 | |||
1281 | local_irq_save(flags); | ||
1277 | 1282 | ||
1278 | return test_and_clear_bit(0, v); | 1283 | if (timer_pending(t)) { |
1284 | if (time_before(when, t->expires)) | ||
1285 | mod_timer_pinned(t, when); | ||
1286 | } else { | ||
1287 | t->expires = round_jiffies(when); | ||
1288 | add_timer_on(t, smp_processor_id()); | ||
1289 | } | ||
1290 | |||
1291 | local_irq_restore(flags); | ||
1279 | } | 1292 | } |
1280 | 1293 | ||
1281 | static void mce_timer_fn(unsigned long data) | 1294 | static void mce_timer_fn(unsigned long data) |
1282 | { | 1295 | { |
1283 | struct timer_list *t = this_cpu_ptr(&mce_timer); | 1296 | struct timer_list *t = this_cpu_ptr(&mce_timer); |
1297 | int cpu = smp_processor_id(); | ||
1284 | unsigned long iv; | 1298 | unsigned long iv; |
1285 | int notify; | ||
1286 | 1299 | ||
1287 | WARN_ON(smp_processor_id() != data); | 1300 | WARN_ON(cpu != data); |
1301 | |||
1302 | iv = __this_cpu_read(mce_next_interval); | ||
1288 | 1303 | ||
1289 | if (mce_available(this_cpu_ptr(&cpu_info))) { | 1304 | if (mce_available(this_cpu_ptr(&cpu_info))) { |
1290 | machine_check_poll(MCP_TIMESTAMP, | 1305 | machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_poll_banks)); |
1291 | this_cpu_ptr(&mce_poll_banks)); | 1306 | |
1292 | mce_intel_cmci_poll(); | 1307 | if (mce_intel_cmci_poll()) { |
1308 | iv = mce_adjust_timer(iv); | ||
1309 | goto done; | ||
1310 | } | ||
1293 | } | 1311 | } |
1294 | 1312 | ||
1295 | /* | 1313 | /* |
1296 | * Alert userspace if needed. If we logged an MCE, reduce the | 1314 | * Alert userspace if needed. If we logged an MCE, reduce the polling |
1297 | * polling interval, otherwise increase the polling interval. | 1315 | * interval, otherwise increase the polling interval. |
1298 | */ | 1316 | */ |
1299 | iv = __this_cpu_read(mce_next_interval); | 1317 | if (mce_notify_irq()) |
1300 | notify = mce_notify_irq(); | ||
1301 | notify |= cmc_error_seen(); | ||
1302 | if (notify) { | ||
1303 | iv = max(iv / 2, (unsigned long) HZ/100); | 1318 | iv = max(iv / 2, (unsigned long) HZ/100); |
1304 | } else { | 1319 | else |
1305 | iv = min(iv * 2, round_jiffies_relative(check_interval * HZ)); | 1320 | iv = min(iv * 2, round_jiffies_relative(check_interval * HZ)); |
1306 | iv = mce_adjust_timer(iv); | 1321 | |
1307 | } | 1322 | done: |
1308 | __this_cpu_write(mce_next_interval, iv); | 1323 | __this_cpu_write(mce_next_interval, iv); |
1309 | /* Might have become 0 after CMCI storm subsided */ | 1324 | __restart_timer(t, iv); |
1310 | if (iv) { | ||
1311 | t->expires = jiffies + iv; | ||
1312 | add_timer_on(t, smp_processor_id()); | ||
1313 | } | ||
1314 | } | 1325 | } |
1315 | 1326 | ||
1316 | /* | 1327 | /* |
@@ -1319,16 +1330,10 @@ static void mce_timer_fn(unsigned long data) | |||
1319 | void mce_timer_kick(unsigned long interval) | 1330 | void mce_timer_kick(unsigned long interval) |
1320 | { | 1331 | { |
1321 | struct timer_list *t = this_cpu_ptr(&mce_timer); | 1332 | struct timer_list *t = this_cpu_ptr(&mce_timer); |
1322 | unsigned long when = jiffies + interval; | ||
1323 | unsigned long iv = __this_cpu_read(mce_next_interval); | 1333 | unsigned long iv = __this_cpu_read(mce_next_interval); |
1324 | 1334 | ||
1325 | if (timer_pending(t)) { | 1335 | __restart_timer(t, interval); |
1326 | if (time_before(when, t->expires)) | 1336 | |
1327 | mod_timer_pinned(t, when); | ||
1328 | } else { | ||
1329 | t->expires = round_jiffies(when); | ||
1330 | add_timer_on(t, smp_processor_id()); | ||
1331 | } | ||
1332 | if (interval < iv) | 1337 | if (interval < iv) |
1333 | __this_cpu_write(mce_next_interval, interval); | 1338 | __this_cpu_write(mce_next_interval, interval); |
1334 | } | 1339 | } |
@@ -1525,45 +1530,46 @@ static int __mcheck_cpu_apply_quirks(struct cpuinfo_x86 *c) | |||
1525 | * Various K7s with broken bank 0 around. Always disable | 1530 | * Various K7s with broken bank 0 around. Always disable |
1526 | * by default. | 1531 | * by default. |
1527 | */ | 1532 | */ |
1528 | if (c->x86 == 6 && cfg->banks > 0) | 1533 | if (c->x86 == 6 && cfg->banks > 0) |
1529 | mce_banks[0].ctl = 0; | 1534 | mce_banks[0].ctl = 0; |
1530 | 1535 | ||
1531 | /* | 1536 | /* |
1532 | * Turn off MC4_MISC thresholding banks on those models since | 1537 | * overflow_recov is supported for F15h Models 00h-0fh |
1533 | * they're not supported there. | 1538 | * even though we don't have a CPUID bit for it. |
1534 | */ | 1539 | */ |
1535 | if (c->x86 == 0x15 && | 1540 | if (c->x86 == 0x15 && c->x86_model <= 0xf) |
1536 | (c->x86_model >= 0x10 && c->x86_model <= 0x1f)) { | 1541 | mce_flags.overflow_recov = 1; |
1537 | int i; | 1542 | |
1538 | u64 val, hwcr; | 1543 | /* |
1539 | bool need_toggle; | 1544 | * Turn off MC4_MISC thresholding banks on those models since |
1540 | u32 msrs[] = { | 1545 | * they're not supported there. |
1546 | */ | ||
1547 | if (c->x86 == 0x15 && | ||
1548 | (c->x86_model >= 0x10 && c->x86_model <= 0x1f)) { | ||
1549 | int i; | ||
1550 | u64 hwcr; | ||
1551 | bool need_toggle; | ||
1552 | u32 msrs[] = { | ||
1541 | 0x00000413, /* MC4_MISC0 */ | 1553 | 0x00000413, /* MC4_MISC0 */ |
1542 | 0xc0000408, /* MC4_MISC1 */ | 1554 | 0xc0000408, /* MC4_MISC1 */ |
1543 | }; | 1555 | }; |
1544 | 1556 | ||
1545 | rdmsrl(MSR_K7_HWCR, hwcr); | 1557 | rdmsrl(MSR_K7_HWCR, hwcr); |
1546 | 1558 | ||
1547 | /* McStatusWrEn has to be set */ | 1559 | /* McStatusWrEn has to be set */ |
1548 | need_toggle = !(hwcr & BIT(18)); | 1560 | need_toggle = !(hwcr & BIT(18)); |
1549 | 1561 | ||
1550 | if (need_toggle) | 1562 | if (need_toggle) |
1551 | wrmsrl(MSR_K7_HWCR, hwcr | BIT(18)); | 1563 | wrmsrl(MSR_K7_HWCR, hwcr | BIT(18)); |
1552 | 1564 | ||
1553 | for (i = 0; i < ARRAY_SIZE(msrs); i++) { | 1565 | /* Clear CntP bit safely */ |
1554 | rdmsrl(msrs[i], val); | 1566 | for (i = 0; i < ARRAY_SIZE(msrs); i++) |
1567 | msr_clear_bit(msrs[i], 62); | ||
1555 | 1568 | ||
1556 | /* CntP bit set? */ | 1569 | /* restore old settings */ |
1557 | if (val & BIT_64(62)) { | 1570 | if (need_toggle) |
1558 | val &= ~BIT_64(62); | 1571 | wrmsrl(MSR_K7_HWCR, hwcr); |
1559 | wrmsrl(msrs[i], val); | 1572 | } |
1560 | } | ||
1561 | } | ||
1562 | |||
1563 | /* restore old settings */ | ||
1564 | if (need_toggle) | ||
1565 | wrmsrl(MSR_K7_HWCR, hwcr); | ||
1566 | } | ||
1567 | } | 1573 | } |
1568 | 1574 | ||
1569 | if (c->x86_vendor == X86_VENDOR_INTEL) { | 1575 | if (c->x86_vendor == X86_VENDOR_INTEL) { |
@@ -1629,10 +1635,11 @@ static void __mcheck_cpu_init_vendor(struct cpuinfo_x86 *c) | |||
1629 | switch (c->x86_vendor) { | 1635 | switch (c->x86_vendor) { |
1630 | case X86_VENDOR_INTEL: | 1636 | case X86_VENDOR_INTEL: |
1631 | mce_intel_feature_init(c); | 1637 | mce_intel_feature_init(c); |
1632 | mce_adjust_timer = mce_intel_adjust_timer; | 1638 | mce_adjust_timer = cmci_intel_adjust_timer; |
1633 | break; | 1639 | break; |
1634 | case X86_VENDOR_AMD: | 1640 | case X86_VENDOR_AMD: |
1635 | mce_amd_feature_init(c); | 1641 | mce_amd_feature_init(c); |
1642 | mce_flags.overflow_recov = cpuid_ebx(0x80000007) & 0x1; | ||
1636 | break; | 1643 | break; |
1637 | default: | 1644 | default: |
1638 | break; | 1645 | break; |
@@ -2017,6 +2024,7 @@ __setup("mce", mcheck_enable); | |||
2017 | int __init mcheck_init(void) | 2024 | int __init mcheck_init(void) |
2018 | { | 2025 | { |
2019 | mcheck_intel_therm_init(); | 2026 | mcheck_intel_therm_init(); |
2027 | mcheck_vendor_init_severity(); | ||
2020 | 2028 | ||
2021 | return 0; | 2029 | return 0; |
2022 | } | 2030 | } |
diff --git a/arch/x86/kernel/cpu/mcheck/mce_amd.c b/arch/x86/kernel/cpu/mcheck/mce_amd.c index f1c3769bbd64..55ad9b37cae8 100644 --- a/arch/x86/kernel/cpu/mcheck/mce_amd.c +++ b/arch/x86/kernel/cpu/mcheck/mce_amd.c | |||
@@ -79,7 +79,7 @@ static inline bool is_shared_bank(int bank) | |||
79 | return (bank == 4); | 79 | return (bank == 4); |
80 | } | 80 | } |
81 | 81 | ||
82 | static const char * const bank4_names(struct threshold_block *b) | 82 | static const char *bank4_names(const struct threshold_block *b) |
83 | { | 83 | { |
84 | switch (b->address) { | 84 | switch (b->address) { |
85 | /* MSR4_MISC0 */ | 85 | /* MSR4_MISC0 */ |
@@ -250,6 +250,7 @@ void mce_amd_feature_init(struct cpuinfo_x86 *c) | |||
250 | if (!b.interrupt_capable) | 250 | if (!b.interrupt_capable) |
251 | goto init; | 251 | goto init; |
252 | 252 | ||
253 | b.interrupt_enable = 1; | ||
253 | new = (high & MASK_LVTOFF_HI) >> 20; | 254 | new = (high & MASK_LVTOFF_HI) >> 20; |
254 | offset = setup_APIC_mce(offset, new); | 255 | offset = setup_APIC_mce(offset, new); |
255 | 256 | ||
@@ -322,6 +323,8 @@ static void amd_threshold_interrupt(void) | |||
322 | log: | 323 | log: |
323 | mce_setup(&m); | 324 | mce_setup(&m); |
324 | rdmsrl(MSR_IA32_MCx_STATUS(bank), m.status); | 325 | rdmsrl(MSR_IA32_MCx_STATUS(bank), m.status); |
326 | if (!(m.status & MCI_STATUS_VAL)) | ||
327 | return; | ||
325 | m.misc = ((u64)high << 32) | low; | 328 | m.misc = ((u64)high << 32) | low; |
326 | m.bank = bank; | 329 | m.bank = bank; |
327 | mce_log(&m); | 330 | mce_log(&m); |
@@ -497,10 +500,12 @@ static int allocate_threshold_blocks(unsigned int cpu, unsigned int bank, | |||
497 | b->interrupt_capable = lvt_interrupt_supported(bank, high); | 500 | b->interrupt_capable = lvt_interrupt_supported(bank, high); |
498 | b->threshold_limit = THRESHOLD_MAX; | 501 | b->threshold_limit = THRESHOLD_MAX; |
499 | 502 | ||
500 | if (b->interrupt_capable) | 503 | if (b->interrupt_capable) { |
501 | threshold_ktype.default_attrs[2] = &interrupt_enable.attr; | 504 | threshold_ktype.default_attrs[2] = &interrupt_enable.attr; |
502 | else | 505 | b->interrupt_enable = 1; |
506 | } else { | ||
503 | threshold_ktype.default_attrs[2] = NULL; | 507 | threshold_ktype.default_attrs[2] = NULL; |
508 | } | ||
504 | 509 | ||
505 | INIT_LIST_HEAD(&b->miscj); | 510 | INIT_LIST_HEAD(&b->miscj); |
506 | 511 | ||
diff --git a/arch/x86/kernel/cpu/mcheck/mce_intel.c b/arch/x86/kernel/cpu/mcheck/mce_intel.c index b3c97bafc123..b4a41cf030ed 100644 --- a/arch/x86/kernel/cpu/mcheck/mce_intel.c +++ b/arch/x86/kernel/cpu/mcheck/mce_intel.c | |||
@@ -39,6 +39,15 @@ | |||
39 | static DEFINE_PER_CPU(mce_banks_t, mce_banks_owned); | 39 | static DEFINE_PER_CPU(mce_banks_t, mce_banks_owned); |
40 | 40 | ||
41 | /* | 41 | /* |
42 | * CMCI storm detection backoff counter | ||
43 | * | ||
44 | * During storm, we reset this counter to INITIAL_CHECK_INTERVAL in case we've | ||
45 | * encountered an error. If not, we decrement it by one. We signal the end of | ||
46 | * the CMCI storm when it reaches 0. | ||
47 | */ | ||
48 | static DEFINE_PER_CPU(int, cmci_backoff_cnt); | ||
49 | |||
50 | /* | ||
42 | * cmci_discover_lock protects against parallel discovery attempts | 51 | * cmci_discover_lock protects against parallel discovery attempts |
43 | * which could race against each other. | 52 | * which could race against each other. |
44 | */ | 53 | */ |
@@ -46,7 +55,7 @@ static DEFINE_RAW_SPINLOCK(cmci_discover_lock); | |||
46 | 55 | ||
47 | #define CMCI_THRESHOLD 1 | 56 | #define CMCI_THRESHOLD 1 |
48 | #define CMCI_POLL_INTERVAL (30 * HZ) | 57 | #define CMCI_POLL_INTERVAL (30 * HZ) |
49 | #define CMCI_STORM_INTERVAL (1 * HZ) | 58 | #define CMCI_STORM_INTERVAL (HZ) |
50 | #define CMCI_STORM_THRESHOLD 15 | 59 | #define CMCI_STORM_THRESHOLD 15 |
51 | 60 | ||
52 | static DEFINE_PER_CPU(unsigned long, cmci_time_stamp); | 61 | static DEFINE_PER_CPU(unsigned long, cmci_time_stamp); |
@@ -82,11 +91,21 @@ static int cmci_supported(int *banks) | |||
82 | return !!(cap & MCG_CMCI_P); | 91 | return !!(cap & MCG_CMCI_P); |
83 | } | 92 | } |
84 | 93 | ||
85 | void mce_intel_cmci_poll(void) | 94 | bool mce_intel_cmci_poll(void) |
86 | { | 95 | { |
87 | if (__this_cpu_read(cmci_storm_state) == CMCI_STORM_NONE) | 96 | if (__this_cpu_read(cmci_storm_state) == CMCI_STORM_NONE) |
88 | return; | 97 | return false; |
89 | machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned)); | 98 | |
99 | /* | ||
100 | * Reset the counter if we've logged an error in the last poll | ||
101 | * during the storm. | ||
102 | */ | ||
103 | if (machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned))) | ||
104 | this_cpu_write(cmci_backoff_cnt, INITIAL_CHECK_INTERVAL); | ||
105 | else | ||
106 | this_cpu_dec(cmci_backoff_cnt); | ||
107 | |||
108 | return true; | ||
90 | } | 109 | } |
91 | 110 | ||
92 | void mce_intel_hcpu_update(unsigned long cpu) | 111 | void mce_intel_hcpu_update(unsigned long cpu) |
@@ -97,31 +116,32 @@ void mce_intel_hcpu_update(unsigned long cpu) | |||
97 | per_cpu(cmci_storm_state, cpu) = CMCI_STORM_NONE; | 116 | per_cpu(cmci_storm_state, cpu) = CMCI_STORM_NONE; |
98 | } | 117 | } |
99 | 118 | ||
100 | unsigned long mce_intel_adjust_timer(unsigned long interval) | 119 | unsigned long cmci_intel_adjust_timer(unsigned long interval) |
101 | { | 120 | { |
102 | int r; | 121 | if ((this_cpu_read(cmci_backoff_cnt) > 0) && |
103 | 122 | (__this_cpu_read(cmci_storm_state) == CMCI_STORM_ACTIVE)) { | |
104 | if (interval < CMCI_POLL_INTERVAL) | 123 | mce_notify_irq(); |
105 | return interval; | 124 | return CMCI_STORM_INTERVAL; |
125 | } | ||
106 | 126 | ||
107 | switch (__this_cpu_read(cmci_storm_state)) { | 127 | switch (__this_cpu_read(cmci_storm_state)) { |
108 | case CMCI_STORM_ACTIVE: | 128 | case CMCI_STORM_ACTIVE: |
129 | |||
109 | /* | 130 | /* |
110 | * We switch back to interrupt mode once the poll timer has | 131 | * We switch back to interrupt mode once the poll timer has |
111 | * silenced itself. That means no events recorded and the | 132 | * silenced itself. That means no events recorded and the timer |
112 | * timer interval is back to our poll interval. | 133 | * interval is back to our poll interval. |
113 | */ | 134 | */ |
114 | __this_cpu_write(cmci_storm_state, CMCI_STORM_SUBSIDED); | 135 | __this_cpu_write(cmci_storm_state, CMCI_STORM_SUBSIDED); |
115 | r = atomic_sub_return(1, &cmci_storm_on_cpus); | 136 | if (!atomic_sub_return(1, &cmci_storm_on_cpus)) |
116 | if (r == 0) | ||
117 | pr_notice("CMCI storm subsided: switching to interrupt mode\n"); | 137 | pr_notice("CMCI storm subsided: switching to interrupt mode\n"); |
138 | |||
118 | /* FALLTHROUGH */ | 139 | /* FALLTHROUGH */ |
119 | 140 | ||
120 | case CMCI_STORM_SUBSIDED: | 141 | case CMCI_STORM_SUBSIDED: |
121 | /* | 142 | /* |
122 | * We wait for all cpus to go back to SUBSIDED | 143 | * We wait for all CPUs to go back to SUBSIDED state. When that |
123 | * state. When that happens we switch back to | 144 | * happens we switch back to interrupt mode. |
124 | * interrupt mode. | ||
125 | */ | 145 | */ |
126 | if (!atomic_read(&cmci_storm_on_cpus)) { | 146 | if (!atomic_read(&cmci_storm_on_cpus)) { |
127 | __this_cpu_write(cmci_storm_state, CMCI_STORM_NONE); | 147 | __this_cpu_write(cmci_storm_state, CMCI_STORM_NONE); |
@@ -130,10 +150,8 @@ unsigned long mce_intel_adjust_timer(unsigned long interval) | |||
130 | } | 150 | } |
131 | return CMCI_POLL_INTERVAL; | 151 | return CMCI_POLL_INTERVAL; |
132 | default: | 152 | default: |
133 | /* | 153 | |
134 | * We have shiny weather. Let the poll do whatever it | 154 | /* We have shiny weather. Let the poll do whatever it thinks. */ |
135 | * thinks. | ||
136 | */ | ||
137 | return interval; | 155 | return interval; |
138 | } | 156 | } |
139 | } | 157 | } |
@@ -178,7 +196,8 @@ static bool cmci_storm_detect(void) | |||
178 | cmci_storm_disable_banks(); | 196 | cmci_storm_disable_banks(); |
179 | __this_cpu_write(cmci_storm_state, CMCI_STORM_ACTIVE); | 197 | __this_cpu_write(cmci_storm_state, CMCI_STORM_ACTIVE); |
180 | r = atomic_add_return(1, &cmci_storm_on_cpus); | 198 | r = atomic_add_return(1, &cmci_storm_on_cpus); |
181 | mce_timer_kick(CMCI_POLL_INTERVAL); | 199 | mce_timer_kick(CMCI_STORM_INTERVAL); |
200 | this_cpu_write(cmci_backoff_cnt, INITIAL_CHECK_INTERVAL); | ||
182 | 201 | ||
183 | if (r == 1) | 202 | if (r == 1) |
184 | pr_notice("CMCI storm detected: switching to poll mode\n"); | 203 | pr_notice("CMCI storm detected: switching to poll mode\n"); |
@@ -195,6 +214,7 @@ static void intel_threshold_interrupt(void) | |||
195 | { | 214 | { |
196 | if (cmci_storm_detect()) | 215 | if (cmci_storm_detect()) |
197 | return; | 216 | return; |
217 | |||
198 | machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned)); | 218 | machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned)); |
199 | mce_notify_irq(); | 219 | mce_notify_irq(); |
200 | } | 220 | } |
@@ -286,6 +306,7 @@ void cmci_recheck(void) | |||
286 | 306 | ||
287 | if (!mce_available(raw_cpu_ptr(&cpu_info)) || !cmci_supported(&banks)) | 307 | if (!mce_available(raw_cpu_ptr(&cpu_info)) || !cmci_supported(&banks)) |
288 | return; | 308 | return; |
309 | |||
289 | local_irq_save(flags); | 310 | local_irq_save(flags); |
290 | machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned)); | 311 | machine_check_poll(MCP_TIMESTAMP, this_cpu_ptr(&mce_banks_owned)); |
291 | local_irq_restore(flags); | 312 | local_irq_restore(flags); |