diff options
| author | Linus Torvalds <torvalds@woody.linux-foundation.org> | 2007-05-05 17:55:20 -0400 |
|---|---|---|
| committer | Linus Torvalds <torvalds@woody.linux-foundation.org> | 2007-05-05 17:55:20 -0400 |
| commit | ea62ccd00fd0b6720b033adfc9984f31130ce195 (patch) | |
| tree | 9837b797b2466fffcb0af96c388b06eae9c3df18 /arch/x86_64/boot/compressed/head.S | |
| parent | 886a0768affe9a32f18c45f8e1393bca9ece5392 (diff) | |
| parent | 35060b6a9a4e1c89bc6fbea61090e302dbc61847 (diff) | |
Merge branch 'for-linus' of git://one.firstfloor.org/home/andi/git/linux-2.6
* 'for-linus' of git://one.firstfloor.org/home/andi/git/linux-2.6: (231 commits)
[PATCH] i386: Don't delete cpu_devs data to identify different x86 types in late_initcall
[PATCH] i386: type may be unused
[PATCH] i386: Some additional chipset register values validation.
[PATCH] i386: Add missing !X86_PAE dependincy to the 2G/2G split.
[PATCH] x86-64: Don't exclude asm-offsets.c in Documentation/dontdiff
[PATCH] i386: avoid redundant preempt_disable in __unlazy_fpu
[PATCH] i386: white space fixes in i387.h
[PATCH] i386: Drop noisy e820 debugging printks
[PATCH] x86-64: Fix allnoconfig error in genapic_flat.c
[PATCH] x86-64: Shut up warnings for vfat compat ioctls on other file systems
[PATCH] x86-64: Share identical video.S between i386 and x86-64
[PATCH] x86-64: Remove CONFIG_REORDER
[PATCH] x86-64: Print type and size correctly for unknown compat ioctls
[PATCH] i386: Remove copy_*_user BUG_ONs for (size < 0)
[PATCH] i386: Little cleanups in smpboot.c
[PATCH] x86-64: Don't enable NUMA for a single node in K8 NUMA scanning
[PATCH] x86: Use RDTSCP for synchronous get_cycles if possible
[PATCH] i386: Add X86_FEATURE_RDTSCP
[PATCH] i386: Implement X86_FEATURE_SYNC_RDTSC on i386
[PATCH] i386: Implement alternative_io for i386
...
Fix up trivial conflict in include/linux/highmem.h manually.
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
Diffstat (limited to 'arch/x86_64/boot/compressed/head.S')
| -rw-r--r-- | arch/x86_64/boot/compressed/head.S | 339 |
1 files changed, 251 insertions, 88 deletions
diff --git a/arch/x86_64/boot/compressed/head.S b/arch/x86_64/boot/compressed/head.S index 6f55565e4d42..f9d5692a0106 100644 --- a/arch/x86_64/boot/compressed/head.S +++ b/arch/x86_64/boot/compressed/head.S | |||
| @@ -26,116 +26,279 @@ | |||
| 26 | 26 | ||
| 27 | #include <linux/linkage.h> | 27 | #include <linux/linkage.h> |
| 28 | #include <asm/segment.h> | 28 | #include <asm/segment.h> |
| 29 | #include <asm/pgtable.h> | ||
| 29 | #include <asm/page.h> | 30 | #include <asm/page.h> |
| 31 | #include <asm/msr.h> | ||
| 30 | 32 | ||
| 33 | .section ".text.head" | ||
| 31 | .code32 | 34 | .code32 |
| 32 | .globl startup_32 | 35 | .globl startup_32 |
| 33 | 36 | ||
| 34 | startup_32: | 37 | startup_32: |
| 35 | cld | 38 | cld |
| 36 | cli | 39 | cli |
| 37 | movl $(__KERNEL_DS),%eax | 40 | movl $(__KERNEL_DS), %eax |
| 38 | movl %eax,%ds | 41 | movl %eax, %ds |
| 39 | movl %eax,%es | 42 | movl %eax, %es |
| 40 | movl %eax,%fs | 43 | movl %eax, %ss |
| 41 | movl %eax,%gs | 44 | |
| 42 | 45 | /* Calculate the delta between where we were compiled to run | |
| 43 | lss stack_start,%esp | 46 | * at and where we were actually loaded at. This can only be done |
| 44 | xorl %eax,%eax | 47 | * with a short local call on x86. Nothing else will tell us what |
| 45 | 1: incl %eax # check that A20 really IS enabled | 48 | * address we are running at. The reserved chunk of the real-mode |
| 46 | movl %eax,0x000000 # loop forever if it isn't | 49 | * data at 0x34-0x3f are used as the stack for this calculation. |
| 47 | cmpl %eax,0x100000 | 50 | * Only 4 bytes are needed. |
| 48 | je 1b | 51 | */ |
| 52 | leal 0x40(%esi), %esp | ||
| 53 | call 1f | ||
| 54 | 1: popl %ebp | ||
| 55 | subl $1b, %ebp | ||
| 56 | |||
| 57 | /* setup a stack and make sure cpu supports long mode. */ | ||
| 58 | movl $user_stack_end, %eax | ||
| 59 | addl %ebp, %eax | ||
| 60 | movl %eax, %esp | ||
| 61 | |||
| 62 | call verify_cpu | ||
| 63 | testl %eax, %eax | ||
| 64 | jnz no_longmode | ||
| 65 | |||
| 66 | /* Compute the delta between where we were compiled to run at | ||
| 67 | * and where the code will actually run at. | ||
| 68 | */ | ||
| 69 | /* %ebp contains the address we are loaded at by the boot loader and %ebx | ||
| 70 | * contains the address where we should move the kernel image temporarily | ||
| 71 | * for safe in-place decompression. | ||
| 72 | */ | ||
| 73 | |||
| 74 | #ifdef CONFIG_RELOCATABLE | ||
| 75 | movl %ebp, %ebx | ||
| 76 | addl $(LARGE_PAGE_SIZE -1), %ebx | ||
| 77 | andl $LARGE_PAGE_MASK, %ebx | ||
| 78 | #else | ||
| 79 | movl $CONFIG_PHYSICAL_START, %ebx | ||
| 80 | #endif | ||
| 81 | |||
| 82 | /* Replace the compressed data size with the uncompressed size */ | ||
| 83 | subl input_len(%ebp), %ebx | ||
| 84 | movl output_len(%ebp), %eax | ||
| 85 | addl %eax, %ebx | ||
| 86 | /* Add 8 bytes for every 32K input block */ | ||
| 87 | shrl $12, %eax | ||
| 88 | addl %eax, %ebx | ||
| 89 | /* Add 32K + 18 bytes of extra slack and align on a 4K boundary */ | ||
| 90 | addl $(32768 + 18 + 4095), %ebx | ||
| 91 | andl $~4095, %ebx | ||
| 49 | 92 | ||
| 50 | /* | 93 | /* |
| 51 | * Initialize eflags. Some BIOS's leave bits like NT set. This would | 94 | * Prepare for entering 64 bit mode |
| 52 | * confuse the debugger if this code is traced. | ||
| 53 | * XXX - best to initialize before switching to protected mode. | ||
| 54 | */ | 95 | */ |
| 55 | pushl $0 | 96 | |
| 56 | popfl | 97 | /* Load new GDT with the 64bit segments using 32bit descriptor */ |
| 98 | leal gdt(%ebp), %eax | ||
| 99 | movl %eax, gdt+2(%ebp) | ||
| 100 | lgdt gdt(%ebp) | ||
| 101 | |||
| 102 | /* Enable PAE mode */ | ||
| 103 | xorl %eax, %eax | ||
| 104 | orl $(1 << 5), %eax | ||
| 105 | movl %eax, %cr4 | ||
| 106 | |||
| 107 | /* | ||
| 108 | * Build early 4G boot pagetable | ||
| 109 | */ | ||
| 110 | /* Initialize Page tables to 0*/ | ||
| 111 | leal pgtable(%ebx), %edi | ||
| 112 | xorl %eax, %eax | ||
| 113 | movl $((4096*6)/4), %ecx | ||
| 114 | rep stosl | ||
| 115 | |||
| 116 | /* Build Level 4 */ | ||
| 117 | leal pgtable + 0(%ebx), %edi | ||
| 118 | leal 0x1007 (%edi), %eax | ||
| 119 | movl %eax, 0(%edi) | ||
| 120 | |||
| 121 | /* Build Level 3 */ | ||
| 122 | leal pgtable + 0x1000(%ebx), %edi | ||
| 123 | leal 0x1007(%edi), %eax | ||
| 124 | movl $4, %ecx | ||
| 125 | 1: movl %eax, 0x00(%edi) | ||
| 126 | addl $0x00001000, %eax | ||
| 127 | addl $8, %edi | ||
| 128 | decl %ecx | ||
| 129 | jnz 1b | ||
| 130 | |||
| 131 | /* Build Level 2 */ | ||
| 132 | leal pgtable + 0x2000(%ebx), %edi | ||
| 133 | movl $0x00000183, %eax | ||
| 134 | movl $2048, %ecx | ||
| 135 | 1: movl %eax, 0(%edi) | ||
| 136 | addl $0x00200000, %eax | ||
| 137 | addl $8, %edi | ||
| 138 | decl %ecx | ||
| 139 | jnz 1b | ||
| 140 | |||
| 141 | /* Enable the boot page tables */ | ||
| 142 | leal pgtable(%ebx), %eax | ||
| 143 | movl %eax, %cr3 | ||
| 144 | |||
| 145 | /* Enable Long mode in EFER (Extended Feature Enable Register) */ | ||
| 146 | movl $MSR_EFER, %ecx | ||
| 147 | rdmsr | ||
| 148 | btsl $_EFER_LME, %eax | ||
| 149 | wrmsr | ||
| 150 | |||
| 151 | /* Setup for the jump to 64bit mode | ||
| 152 | * | ||
| 153 | * When the jump is performend we will be in long mode but | ||
| 154 | * in 32bit compatibility mode with EFER.LME = 1, CS.L = 0, CS.D = 1 | ||
| 155 | * (and in turn EFER.LMA = 1). To jump into 64bit mode we use | ||
| 156 | * the new gdt/idt that has __KERNEL_CS with CS.L = 1. | ||
| 157 | * We place all of the values on our mini stack so lret can | ||
| 158 | * used to perform that far jump. | ||
| 159 | */ | ||
| 160 | pushl $__KERNEL_CS | ||
| 161 | leal startup_64(%ebp), %eax | ||
| 162 | pushl %eax | ||
| 163 | |||
| 164 | /* Enter paged protected Mode, activating Long Mode */ | ||
| 165 | movl $0x80000001, %eax /* Enable Paging and Protected mode */ | ||
| 166 | movl %eax, %cr0 | ||
| 167 | |||
| 168 | /* Jump from 32bit compatibility mode into 64bit mode. */ | ||
| 169 | lret | ||
| 170 | |||
| 171 | no_longmode: | ||
| 172 | /* This isn't an x86-64 CPU so hang */ | ||
| 173 | 1: | ||
| 174 | hlt | ||
| 175 | jmp 1b | ||
| 176 | |||
| 177 | #include "../../kernel/verify_cpu.S" | ||
| 178 | |||
| 179 | /* Be careful here startup_64 needs to be at a predictable | ||
| 180 | * address so I can export it in an ELF header. Bootloaders | ||
| 181 | * should look at the ELF header to find this address, as | ||
| 182 | * it may change in the future. | ||
| 183 | */ | ||
| 184 | .code64 | ||
| 185 | .org 0x200 | ||
| 186 | ENTRY(startup_64) | ||
| 187 | /* We come here either from startup_32 or directly from a | ||
| 188 | * 64bit bootloader. If we come here from a bootloader we depend on | ||
| 189 | * an identity mapped page table being provied that maps our | ||
| 190 | * entire text+data+bss and hopefully all of memory. | ||
| 191 | */ | ||
| 192 | |||
| 193 | /* Setup data segments. */ | ||
| 194 | xorl %eax, %eax | ||
| 195 | movl %eax, %ds | ||
| 196 | movl %eax, %es | ||
| 197 | movl %eax, %ss | ||
| 198 | |||
| 199 | /* Compute the decompressed kernel start address. It is where | ||
| 200 | * we were loaded at aligned to a 2M boundary. %rbp contains the | ||
| 201 | * decompressed kernel start address. | ||
| 202 | * | ||
| 203 | * If it is a relocatable kernel then decompress and run the kernel | ||
| 204 | * from load address aligned to 2MB addr, otherwise decompress and | ||
| 205 | * run the kernel from CONFIG_PHYSICAL_START | ||
| 206 | */ | ||
| 207 | |||
| 208 | /* Start with the delta to where the kernel will run at. */ | ||
| 209 | #ifdef CONFIG_RELOCATABLE | ||
| 210 | leaq startup_32(%rip) /* - $startup_32 */, %rbp | ||
| 211 | addq $(LARGE_PAGE_SIZE - 1), %rbp | ||
| 212 | andq $LARGE_PAGE_MASK, %rbp | ||
| 213 | movq %rbp, %rbx | ||
| 214 | #else | ||
| 215 | movq $CONFIG_PHYSICAL_START, %rbp | ||
| 216 | movq %rbp, %rbx | ||
| 217 | #endif | ||
| 218 | |||
| 219 | /* Replace the compressed data size with the uncompressed size */ | ||
| 220 | movl input_len(%rip), %eax | ||
| 221 | subq %rax, %rbx | ||
| 222 | movl output_len(%rip), %eax | ||
| 223 | addq %rax, %rbx | ||
| 224 | /* Add 8 bytes for every 32K input block */ | ||
| 225 | shrq $12, %rax | ||
| 226 | addq %rax, %rbx | ||
| 227 | /* Add 32K + 18 bytes of extra slack and align on a 4K boundary */ | ||
| 228 | addq $(32768 + 18 + 4095), %rbx | ||
| 229 | andq $~4095, %rbx | ||
| 230 | |||
| 231 | /* Copy the compressed kernel to the end of our buffer | ||
| 232 | * where decompression in place becomes safe. | ||
| 233 | */ | ||
| 234 | leaq _end(%rip), %r8 | ||
| 235 | leaq _end(%rbx), %r9 | ||
| 236 | movq $_end /* - $startup_32 */, %rcx | ||
| 237 | 1: subq $8, %r8 | ||
| 238 | subq $8, %r9 | ||
| 239 | movq 0(%r8), %rax | ||
| 240 | movq %rax, 0(%r9) | ||
| 241 | subq $8, %rcx | ||
| 242 | jnz 1b | ||
| 243 | |||
| 244 | /* | ||
| 245 | * Jump to the relocated address. | ||
| 246 | */ | ||
| 247 | leaq relocated(%rbx), %rax | ||
| 248 | jmp *%rax | ||
| 249 | |||
| 250 | .section ".text" | ||
| 251 | relocated: | ||
| 252 | |||
| 57 | /* | 253 | /* |
| 58 | * Clear BSS | 254 | * Clear BSS |
| 59 | */ | 255 | */ |
| 60 | xorl %eax,%eax | 256 | xorq %rax, %rax |
| 61 | movl $_edata,%edi | 257 | leaq _edata(%rbx), %rdi |
| 62 | movl $_end,%ecx | 258 | leaq _end(%rbx), %rcx |
| 63 | subl %edi,%ecx | 259 | subq %rdi, %rcx |
| 64 | cld | 260 | cld |
| 65 | rep | 261 | rep |
| 66 | stosb | 262 | stosb |
| 263 | |||
| 264 | /* Setup the stack */ | ||
| 265 | leaq user_stack_end(%rip), %rsp | ||
| 266 | |||
| 267 | /* zero EFLAGS after setting rsp */ | ||
| 268 | pushq $0 | ||
| 269 | popfq | ||
| 270 | |||
| 67 | /* | 271 | /* |
| 68 | * Do the decompression, and jump to the new kernel.. | 272 | * Do the decompression, and jump to the new kernel.. |
| 69 | */ | 273 | */ |
| 70 | subl $16,%esp # place for structure on the stack | 274 | pushq %rsi # Save the real mode argument |
| 71 | movl %esp,%eax | 275 | movq %rsi, %rdi # real mode address |
| 72 | pushl %esi # real mode pointer as second arg | 276 | leaq _heap(%rip), %rsi # _heap |
| 73 | pushl %eax # address of structure as first arg | 277 | leaq input_data(%rip), %rdx # input_data |
| 74 | call decompress_kernel | 278 | movl input_len(%rip), %eax |
| 75 | orl %eax,%eax | 279 | movq %rax, %rcx # input_len |
| 76 | jnz 3f | 280 | movq %rbp, %r8 # output |
| 77 | addl $8,%esp | 281 | call decompress_kernel |
| 78 | xorl %ebx,%ebx | 282 | popq %rsi |
| 79 | ljmp $(__KERNEL_CS), $__PHYSICAL_START | ||
| 80 | 283 | ||
| 81 | /* | ||
| 82 | * We come here, if we were loaded high. | ||
| 83 | * We need to move the move-in-place routine down to 0x1000 | ||
| 84 | * and then start it with the buffer addresses in registers, | ||
| 85 | * which we got from the stack. | ||
| 86 | */ | ||
| 87 | 3: | ||
| 88 | movl %esi,%ebx | ||
| 89 | movl $move_routine_start,%esi | ||
| 90 | movl $0x1000,%edi | ||
| 91 | movl $move_routine_end,%ecx | ||
| 92 | subl %esi,%ecx | ||
| 93 | addl $3,%ecx | ||
| 94 | shrl $2,%ecx | ||
| 95 | cld | ||
| 96 | rep | ||
| 97 | movsl | ||
| 98 | |||
| 99 | popl %esi # discard the address | ||
| 100 | addl $4,%esp # real mode pointer | ||
| 101 | popl %esi # low_buffer_start | ||
| 102 | popl %ecx # lcount | ||
| 103 | popl %edx # high_buffer_start | ||
| 104 | popl %eax # hcount | ||
| 105 | movl $__PHYSICAL_START,%edi | ||
| 106 | cli # make sure we don't get interrupted | ||
| 107 | ljmp $(__KERNEL_CS), $0x1000 # and jump to the move routine | ||
| 108 | 284 | ||
| 109 | /* | 285 | /* |
| 110 | * Routine (template) for moving the decompressed kernel in place, | 286 | * Jump to the decompressed kernel. |
| 111 | * if we were high loaded. This _must_ PIC-code ! | ||
| 112 | */ | 287 | */ |
| 113 | move_routine_start: | 288 | jmp *%rbp |
| 114 | movl %ecx,%ebp | ||
| 115 | shrl $2,%ecx | ||
| 116 | rep | ||
| 117 | movsl | ||
| 118 | movl %ebp,%ecx | ||
| 119 | andl $3,%ecx | ||
| 120 | rep | ||
| 121 | movsb | ||
| 122 | movl %edx,%esi | ||
| 123 | movl %eax,%ecx # NOTE: rep movsb won't move if %ecx == 0 | ||
| 124 | addl $3,%ecx | ||
| 125 | shrl $2,%ecx | ||
| 126 | rep | ||
| 127 | movsl | ||
| 128 | movl %ebx,%esi # Restore setup pointer | ||
| 129 | xorl %ebx,%ebx | ||
| 130 | ljmp $(__KERNEL_CS), $__PHYSICAL_START | ||
| 131 | move_routine_end: | ||
| 132 | 289 | ||
| 133 | 290 | .data | |
| 134 | /* Stack for uncompression */ | 291 | gdt: |
| 135 | .align 32 | 292 | .word gdt_end - gdt |
| 136 | user_stack: | 293 | .long gdt |
| 294 | .word 0 | ||
| 295 | .quad 0x0000000000000000 /* NULL descriptor */ | ||
| 296 | .quad 0x00af9a000000ffff /* __KERNEL_CS */ | ||
| 297 | .quad 0x00cf92000000ffff /* __KERNEL_DS */ | ||
| 298 | gdt_end: | ||
| 299 | .bss | ||
| 300 | /* Stack for uncompression */ | ||
| 301 | .balign 4 | ||
| 302 | user_stack: | ||
| 137 | .fill 4096,4,0 | 303 | .fill 4096,4,0 |
| 138 | stack_start: | 304 | user_stack_end: |
| 139 | .long user_stack+4096 | ||
| 140 | .word __KERNEL_DS | ||
| 141 | |||
