diff options
| author | Rafael J. Wysocki <rjw@sisk.pl> | 2007-10-18 06:04:54 -0400 |
|---|---|---|
| committer | Linus Torvalds <torvalds@woody.linux-foundation.org> | 2007-10-18 17:37:19 -0400 |
| commit | c30bb68c26ad7e9277f2d1dfed8347f329e1cf27 (patch) | |
| tree | 6609a9a69446f58d076dadd452b250db732be47b /arch | |
| parent | d158cbdf39ffaec9dd5299fdfdfdd2c7897a71dc (diff) | |
Hibernation: Pass CR3 in the image header on x86_64
Since we already pass the address of restore_registers() in the image header,
we can also pass the value of the CR3 register from before the hibernation in
the same way. This will allow us to avoid using init_level4_pgt page tables
during the restore.
Signed-off-by: Rafael J. Wysocki <rjw@sisk.pl>
Acked-by: Pavel Machek <pavel@ucw.cz>
Cc: Andi Kleen <ak@suse.de>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'arch')
| -rw-r--r-- | arch/x86/kernel/suspend_64.c | 16 | ||||
| -rw-r--r-- | arch/x86/kernel/suspend_asm_64.S | 8 |
2 files changed, 17 insertions, 7 deletions
diff --git a/arch/x86/kernel/suspend_64.c b/arch/x86/kernel/suspend_64.c index 01fbfb018ca9..da10eef4c3ef 100644 --- a/arch/x86/kernel/suspend_64.c +++ b/arch/x86/kernel/suspend_64.c | |||
| @@ -156,6 +156,12 @@ extern int restore_image(void); | |||
| 156 | */ | 156 | */ |
| 157 | unsigned long restore_jump_address; | 157 | unsigned long restore_jump_address; |
| 158 | 158 | ||
| 159 | /* | ||
| 160 | * Value of the cr3 register from before the hibernation (this value is passed | ||
| 161 | * in the image header). | ||
| 162 | */ | ||
| 163 | unsigned long restore_cr3; | ||
| 164 | |||
| 159 | pgd_t *temp_level4_pgt; | 165 | pgd_t *temp_level4_pgt; |
| 160 | 166 | ||
| 161 | void *relocated_restore_code; | 167 | void *relocated_restore_code; |
| @@ -254,7 +260,8 @@ int pfn_is_nosave(unsigned long pfn) | |||
| 254 | 260 | ||
| 255 | struct restore_data_record { | 261 | struct restore_data_record { |
| 256 | unsigned long jump_address; | 262 | unsigned long jump_address; |
| 257 | unsigned long control; | 263 | unsigned long cr3; |
| 264 | unsigned long magic; | ||
| 258 | }; | 265 | }; |
| 259 | 266 | ||
| 260 | #define RESTORE_MAGIC 0x0123456789ABCDEFUL | 267 | #define RESTORE_MAGIC 0x0123456789ABCDEFUL |
| @@ -271,7 +278,8 @@ int arch_hibernation_header_save(void *addr, unsigned int max_size) | |||
| 271 | if (max_size < sizeof(struct restore_data_record)) | 278 | if (max_size < sizeof(struct restore_data_record)) |
| 272 | return -EOVERFLOW; | 279 | return -EOVERFLOW; |
| 273 | rdr->jump_address = restore_jump_address; | 280 | rdr->jump_address = restore_jump_address; |
| 274 | rdr->control = (restore_jump_address ^ RESTORE_MAGIC); | 281 | rdr->cr3 = restore_cr3; |
| 282 | rdr->magic = RESTORE_MAGIC; | ||
| 275 | return 0; | 283 | return 0; |
| 276 | } | 284 | } |
| 277 | 285 | ||
| @@ -285,7 +293,7 @@ int arch_hibernation_header_restore(void *addr) | |||
| 285 | struct restore_data_record *rdr = addr; | 293 | struct restore_data_record *rdr = addr; |
| 286 | 294 | ||
| 287 | restore_jump_address = rdr->jump_address; | 295 | restore_jump_address = rdr->jump_address; |
| 288 | return (rdr->control == (restore_jump_address ^ RESTORE_MAGIC)) ? | 296 | restore_cr3 = rdr->cr3; |
| 289 | 0 : -EINVAL; | 297 | return (rdr->magic == RESTORE_MAGIC) ? 0 : -EINVAL; |
| 290 | } | 298 | } |
| 291 | #endif /* CONFIG_HIBERNATION */ | 299 | #endif /* CONFIG_HIBERNATION */ |
diff --git a/arch/x86/kernel/suspend_asm_64.S b/arch/x86/kernel/suspend_asm_64.S index 40a209e0525c..48344b666d2c 100644 --- a/arch/x86/kernel/suspend_asm_64.S +++ b/arch/x86/kernel/suspend_asm_64.S | |||
| @@ -39,6 +39,9 @@ ENTRY(swsusp_arch_suspend) | |||
| 39 | /* save the address of restore_registers */ | 39 | /* save the address of restore_registers */ |
| 40 | movq $restore_registers, %rax | 40 | movq $restore_registers, %rax |
| 41 | movq %rax, restore_jump_address(%rip) | 41 | movq %rax, restore_jump_address(%rip) |
| 42 | /* save cr3 */ | ||
| 43 | movq %cr3, %rax | ||
| 44 | movq %rax, restore_cr3(%rip) | ||
| 42 | 45 | ||
| 43 | call swsusp_save | 46 | call swsusp_save |
| 44 | ret | 47 | ret |
| @@ -60,6 +63,7 @@ ENTRY(restore_image) | |||
| 60 | 63 | ||
| 61 | /* prepare to jump to the image kernel */ | 64 | /* prepare to jump to the image kernel */ |
| 62 | movq restore_jump_address(%rip), %rax | 65 | movq restore_jump_address(%rip), %rax |
| 66 | movq restore_cr3(%rip), %rbx | ||
| 63 | 67 | ||
| 64 | /* prepare to copy image data to their original locations */ | 68 | /* prepare to copy image data to their original locations */ |
| 65 | movq restore_pblist(%rip), %rdx | 69 | movq restore_pblist(%rip), %rdx |
| @@ -98,9 +102,7 @@ done: | |||
| 98 | 102 | ||
| 99 | ENTRY(restore_registers) | 103 | ENTRY(restore_registers) |
| 100 | /* go back to the original page tables */ | 104 | /* go back to the original page tables */ |
| 101 | movq $(init_level4_pgt - __START_KERNEL_map), %rax | 105 | movq %rbx, %cr3 |
| 102 | addq phys_base(%rip), %rax | ||
| 103 | movq %rax, %cr3 | ||
| 104 | 106 | ||
| 105 | /* Flush TLB, including "global" things (vmalloc) */ | 107 | /* Flush TLB, including "global" things (vmalloc) */ |
| 106 | movq mmu_cr4_features(%rip), %rax | 108 | movq mmu_cr4_features(%rip), %rax |
