diff options
author | Ingo Molnar <mingo@elte.hu> | 2009-06-17 06:52:15 -0400 |
---|---|---|
committer | Ingo Molnar <mingo@elte.hu> | 2009-06-17 06:56:49 -0400 |
commit | eadb8a091b27a840de7450f84ecff5ef13476424 (patch) | |
tree | 58c3782d40def63baa8167f3d31e3048cb4c7660 /arch/x86/power/cpu.c | |
parent | 73874005cd8800440be4299bd095387fff4b90ac (diff) | |
parent | 65795efbd380a832ae508b04dba8f8e53f0b84d9 (diff) |
Merge branch 'linus' into tracing/hw-breakpoints
Conflicts:
arch/x86/Kconfig
arch/x86/kernel/traps.c
arch/x86/power/cpu.c
arch/x86/power/cpu_32.c
kernel/Makefile
Semantic conflict:
arch/x86/kernel/hw_breakpoint.c
Merge reason: Resolve the conflicts, move from put_cpu_no_sched() to
put_cpu() in arch/x86/kernel/hw_breakpoint.c.
Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'arch/x86/power/cpu.c')
-rw-r--r-- | arch/x86/power/cpu.c | 241 |
1 files changed, 241 insertions, 0 deletions
diff --git a/arch/x86/power/cpu.c b/arch/x86/power/cpu.c new file mode 100644 index 000000000000..394cbb88987c --- /dev/null +++ b/arch/x86/power/cpu.c | |||
@@ -0,0 +1,241 @@ | |||
1 | /* | ||
2 | * Suspend support specific for i386/x86-64. | ||
3 | * | ||
4 | * Distribute under GPLv2 | ||
5 | * | ||
6 | * Copyright (c) 2007 Rafael J. Wysocki <rjw@sisk.pl> | ||
7 | * Copyright (c) 2002 Pavel Machek <pavel@suse.cz> | ||
8 | * Copyright (c) 2001 Patrick Mochel <mochel@osdl.org> | ||
9 | */ | ||
10 | |||
11 | #include <linux/suspend.h> | ||
12 | #include <linux/smp.h> | ||
13 | |||
14 | #include <asm/pgtable.h> | ||
15 | #include <asm/proto.h> | ||
16 | #include <asm/mtrr.h> | ||
17 | #include <asm/page.h> | ||
18 | #include <asm/mce.h> | ||
19 | #include <asm/xcr.h> | ||
20 | #include <asm/suspend.h> | ||
21 | #include <asm/debugreg.h> | ||
22 | |||
23 | #ifdef CONFIG_X86_32 | ||
24 | static struct saved_context saved_context; | ||
25 | |||
26 | unsigned long saved_context_ebx; | ||
27 | unsigned long saved_context_esp, saved_context_ebp; | ||
28 | unsigned long saved_context_esi, saved_context_edi; | ||
29 | unsigned long saved_context_eflags; | ||
30 | #else | ||
31 | /* CONFIG_X86_64 */ | ||
32 | struct saved_context saved_context; | ||
33 | #endif | ||
34 | |||
35 | /** | ||
36 | * __save_processor_state - save CPU registers before creating a | ||
37 | * hibernation image and before restoring the memory state from it | ||
38 | * @ctxt - structure to store the registers contents in | ||
39 | * | ||
40 | * NOTE: If there is a CPU register the modification of which by the | ||
41 | * boot kernel (ie. the kernel used for loading the hibernation image) | ||
42 | * might affect the operations of the restored target kernel (ie. the one | ||
43 | * saved in the hibernation image), then its contents must be saved by this | ||
44 | * function. In other words, if kernel A is hibernated and different | ||
45 | * kernel B is used for loading the hibernation image into memory, the | ||
46 | * kernel A's __save_processor_state() function must save all registers | ||
47 | * needed by kernel A, so that it can operate correctly after the resume | ||
48 | * regardless of what kernel B does in the meantime. | ||
49 | */ | ||
50 | static void __save_processor_state(struct saved_context *ctxt) | ||
51 | { | ||
52 | #ifdef CONFIG_X86_32 | ||
53 | mtrr_save_fixed_ranges(NULL); | ||
54 | #endif | ||
55 | kernel_fpu_begin(); | ||
56 | |||
57 | /* | ||
58 | * descriptor tables | ||
59 | */ | ||
60 | #ifdef CONFIG_X86_32 | ||
61 | store_gdt(&ctxt->gdt); | ||
62 | store_idt(&ctxt->idt); | ||
63 | #else | ||
64 | /* CONFIG_X86_64 */ | ||
65 | store_gdt((struct desc_ptr *)&ctxt->gdt_limit); | ||
66 | store_idt((struct desc_ptr *)&ctxt->idt_limit); | ||
67 | #endif | ||
68 | store_tr(ctxt->tr); | ||
69 | |||
70 | /* XMM0..XMM15 should be handled by kernel_fpu_begin(). */ | ||
71 | /* | ||
72 | * segment registers | ||
73 | */ | ||
74 | #ifdef CONFIG_X86_32 | ||
75 | savesegment(es, ctxt->es); | ||
76 | savesegment(fs, ctxt->fs); | ||
77 | savesegment(gs, ctxt->gs); | ||
78 | savesegment(ss, ctxt->ss); | ||
79 | #else | ||
80 | /* CONFIG_X86_64 */ | ||
81 | asm volatile ("movw %%ds, %0" : "=m" (ctxt->ds)); | ||
82 | asm volatile ("movw %%es, %0" : "=m" (ctxt->es)); | ||
83 | asm volatile ("movw %%fs, %0" : "=m" (ctxt->fs)); | ||
84 | asm volatile ("movw %%gs, %0" : "=m" (ctxt->gs)); | ||
85 | asm volatile ("movw %%ss, %0" : "=m" (ctxt->ss)); | ||
86 | |||
87 | rdmsrl(MSR_FS_BASE, ctxt->fs_base); | ||
88 | rdmsrl(MSR_GS_BASE, ctxt->gs_base); | ||
89 | rdmsrl(MSR_KERNEL_GS_BASE, ctxt->gs_kernel_base); | ||
90 | mtrr_save_fixed_ranges(NULL); | ||
91 | |||
92 | rdmsrl(MSR_EFER, ctxt->efer); | ||
93 | #endif | ||
94 | |||
95 | /* | ||
96 | * control registers | ||
97 | */ | ||
98 | ctxt->cr0 = read_cr0(); | ||
99 | ctxt->cr2 = read_cr2(); | ||
100 | ctxt->cr3 = read_cr3(); | ||
101 | #ifdef CONFIG_X86_32 | ||
102 | ctxt->cr4 = read_cr4_safe(); | ||
103 | #else | ||
104 | /* CONFIG_X86_64 */ | ||
105 | ctxt->cr4 = read_cr4(); | ||
106 | ctxt->cr8 = read_cr8(); | ||
107 | #endif | ||
108 | hw_breakpoint_disable(); | ||
109 | } | ||
110 | |||
111 | /* Needed by apm.c */ | ||
112 | void save_processor_state(void) | ||
113 | { | ||
114 | __save_processor_state(&saved_context); | ||
115 | } | ||
116 | #ifdef CONFIG_X86_32 | ||
117 | EXPORT_SYMBOL(save_processor_state); | ||
118 | #endif | ||
119 | |||
120 | static void do_fpu_end(void) | ||
121 | { | ||
122 | /* | ||
123 | * Restore FPU regs if necessary. | ||
124 | */ | ||
125 | kernel_fpu_end(); | ||
126 | } | ||
127 | |||
128 | static void fix_processor_context(void) | ||
129 | { | ||
130 | int cpu = smp_processor_id(); | ||
131 | struct tss_struct *t = &per_cpu(init_tss, cpu); | ||
132 | |||
133 | set_tss_desc(cpu, t); /* | ||
134 | * This just modifies memory; should not be | ||
135 | * necessary. But... This is necessary, because | ||
136 | * 386 hardware has concept of busy TSS or some | ||
137 | * similar stupidity. | ||
138 | */ | ||
139 | |||
140 | #ifdef CONFIG_X86_64 | ||
141 | get_cpu_gdt_table(cpu)[GDT_ENTRY_TSS].type = 9; | ||
142 | |||
143 | syscall_init(); /* This sets MSR_*STAR and related */ | ||
144 | #endif | ||
145 | load_TR_desc(); /* This does ltr */ | ||
146 | load_LDT(¤t->active_mm->context); /* This does lldt */ | ||
147 | |||
148 | /* | ||
149 | * Now maybe reload the debug registers | ||
150 | */ | ||
151 | load_debug_registers(); | ||
152 | } | ||
153 | |||
154 | /** | ||
155 | * __restore_processor_state - restore the contents of CPU registers saved | ||
156 | * by __save_processor_state() | ||
157 | * @ctxt - structure to load the registers contents from | ||
158 | */ | ||
159 | static void __restore_processor_state(struct saved_context *ctxt) | ||
160 | { | ||
161 | /* | ||
162 | * control registers | ||
163 | */ | ||
164 | /* cr4 was introduced in the Pentium CPU */ | ||
165 | #ifdef CONFIG_X86_32 | ||
166 | if (ctxt->cr4) | ||
167 | write_cr4(ctxt->cr4); | ||
168 | #else | ||
169 | /* CONFIG X86_64 */ | ||
170 | wrmsrl(MSR_EFER, ctxt->efer); | ||
171 | write_cr8(ctxt->cr8); | ||
172 | write_cr4(ctxt->cr4); | ||
173 | #endif | ||
174 | write_cr3(ctxt->cr3); | ||
175 | write_cr2(ctxt->cr2); | ||
176 | write_cr0(ctxt->cr0); | ||
177 | |||
178 | /* | ||
179 | * now restore the descriptor tables to their proper values | ||
180 | * ltr is done i fix_processor_context(). | ||
181 | */ | ||
182 | #ifdef CONFIG_X86_32 | ||
183 | load_gdt(&ctxt->gdt); | ||
184 | load_idt(&ctxt->idt); | ||
185 | #else | ||
186 | /* CONFIG_X86_64 */ | ||
187 | load_gdt((const struct desc_ptr *)&ctxt->gdt_limit); | ||
188 | load_idt((const struct desc_ptr *)&ctxt->idt_limit); | ||
189 | #endif | ||
190 | |||
191 | /* | ||
192 | * segment registers | ||
193 | */ | ||
194 | #ifdef CONFIG_X86_32 | ||
195 | loadsegment(es, ctxt->es); | ||
196 | loadsegment(fs, ctxt->fs); | ||
197 | loadsegment(gs, ctxt->gs); | ||
198 | loadsegment(ss, ctxt->ss); | ||
199 | |||
200 | /* | ||
201 | * sysenter MSRs | ||
202 | */ | ||
203 | if (boot_cpu_has(X86_FEATURE_SEP)) | ||
204 | enable_sep_cpu(); | ||
205 | #else | ||
206 | /* CONFIG_X86_64 */ | ||
207 | asm volatile ("movw %0, %%ds" :: "r" (ctxt->ds)); | ||
208 | asm volatile ("movw %0, %%es" :: "r" (ctxt->es)); | ||
209 | asm volatile ("movw %0, %%fs" :: "r" (ctxt->fs)); | ||
210 | load_gs_index(ctxt->gs); | ||
211 | asm volatile ("movw %0, %%ss" :: "r" (ctxt->ss)); | ||
212 | |||
213 | wrmsrl(MSR_FS_BASE, ctxt->fs_base); | ||
214 | wrmsrl(MSR_GS_BASE, ctxt->gs_base); | ||
215 | wrmsrl(MSR_KERNEL_GS_BASE, ctxt->gs_kernel_base); | ||
216 | #endif | ||
217 | |||
218 | /* | ||
219 | * restore XCR0 for xsave capable cpu's. | ||
220 | */ | ||
221 | if (cpu_has_xsave) | ||
222 | xsetbv(XCR_XFEATURE_ENABLED_MASK, pcntxt_mask); | ||
223 | |||
224 | fix_processor_context(); | ||
225 | |||
226 | do_fpu_end(); | ||
227 | mtrr_ap_init(); | ||
228 | |||
229 | #ifdef CONFIG_X86_32 | ||
230 | mcheck_init(&boot_cpu_data); | ||
231 | #endif | ||
232 | } | ||
233 | |||
234 | /* Needed by apm.c */ | ||
235 | void restore_processor_state(void) | ||
236 | { | ||
237 | __restore_processor_state(&saved_context); | ||
238 | } | ||
239 | #ifdef CONFIG_X86_32 | ||
240 | EXPORT_SYMBOL(restore_processor_state); | ||
241 | #endif | ||