diff options
90 files changed, 1304 insertions, 773 deletions
diff --git a/Documentation/00-INDEX b/Documentation/00-INDEX index 5b5aba404aac..73060819ed99 100644 --- a/Documentation/00-INDEX +++ b/Documentation/00-INDEX | |||
| @@ -251,8 +251,6 @@ mono.txt | |||
| 251 | - how to execute Mono-based .NET binaries with the help of BINFMT_MISC. | 251 | - how to execute Mono-based .NET binaries with the help of BINFMT_MISC. |
| 252 | moxa-smartio | 252 | moxa-smartio |
| 253 | - file with info on installing/using Moxa multiport serial driver. | 253 | - file with info on installing/using Moxa multiport serial driver. |
| 254 | mtrr.txt | ||
| 255 | - how to use PPro Memory Type Range Registers to increase performance. | ||
| 256 | mutex-design.txt | 254 | mutex-design.txt |
| 257 | - info on the generic mutex subsystem. | 255 | - info on the generic mutex subsystem. |
| 258 | namespaces/ | 256 | namespaces/ |
diff --git a/Documentation/kernel-parameters.txt b/Documentation/kernel-parameters.txt index 1150444a21ab..329dcabe4c5e 100644 --- a/Documentation/kernel-parameters.txt +++ b/Documentation/kernel-parameters.txt | |||
| @@ -463,12 +463,6 @@ and is between 256 and 4096 characters. It is defined in the file | |||
| 463 | Range: 0 - 8192 | 463 | Range: 0 - 8192 |
| 464 | Default: 64 | 464 | Default: 64 |
| 465 | 465 | ||
| 466 | disable_8254_timer | ||
| 467 | enable_8254_timer | ||
| 468 | [IA32/X86_64] Disable/Enable interrupt 0 timer routing | ||
| 469 | over the 8254 in addition to over the IO-APIC. The | ||
| 470 | kernel tries to set a sensible default. | ||
| 471 | |||
| 472 | hpet= [X86-32,HPET] option to control HPET usage | 466 | hpet= [X86-32,HPET] option to control HPET usage |
| 473 | Format: { enable (default) | disable | force } | 467 | Format: { enable (default) | disable | force } |
| 474 | disable: disable HPET and use PIT instead | 468 | disable: disable HPET and use PIT instead |
| @@ -1882,6 +1876,12 @@ and is between 256 and 4096 characters. It is defined in the file | |||
| 1882 | shapers= [NET] | 1876 | shapers= [NET] |
| 1883 | Maximal number of shapers. | 1877 | Maximal number of shapers. |
| 1884 | 1878 | ||
| 1879 | show_msr= [x86] show boot-time MSR settings | ||
| 1880 | Format: { <integer> } | ||
| 1881 | Show boot-time (BIOS-initialized) MSR settings. | ||
| 1882 | The parameter means the number of CPUs to show, | ||
| 1883 | for example 1 means boot CPU only. | ||
| 1884 | |||
| 1885 | sim710= [SCSI,HW] | 1885 | sim710= [SCSI,HW] |
| 1886 | See header of drivers/scsi/sim710.c. | 1886 | See header of drivers/scsi/sim710.c. |
| 1887 | 1887 | ||
diff --git a/Documentation/x86/00-INDEX b/Documentation/x86/00-INDEX new file mode 100644 index 000000000000..dbe3377754af --- /dev/null +++ b/Documentation/x86/00-INDEX | |||
| @@ -0,0 +1,4 @@ | |||
| 1 | 00-INDEX | ||
| 2 | - this file | ||
| 3 | mtrr.txt | ||
| 4 | - how to use x86 Memory Type Range Registers to increase performance | ||
diff --git a/Documentation/x86/i386/boot.txt b/Documentation/x86/boot.txt index 147bfe511cdd..83c0033ee9e0 100644 --- a/Documentation/x86/i386/boot.txt +++ b/Documentation/x86/boot.txt | |||
| @@ -308,7 +308,7 @@ Protocol: 2.00+ | |||
| 308 | 308 | ||
| 309 | Field name: start_sys | 309 | Field name: start_sys |
| 310 | Type: read | 310 | Type: read |
| 311 | Offset/size: 0x20c/4 | 311 | Offset/size: 0x20c/2 |
| 312 | Protocol: 2.00+ | 312 | Protocol: 2.00+ |
| 313 | 313 | ||
| 314 | The load low segment (0x1000). Obsolete. | 314 | The load low segment (0x1000). Obsolete. |
diff --git a/Documentation/mtrr.txt b/Documentation/x86/mtrr.txt index c39ac395970e..cc071dc333c2 100644 --- a/Documentation/mtrr.txt +++ b/Documentation/x86/mtrr.txt | |||
| @@ -18,7 +18,7 @@ Richard Gooch | |||
| 18 | The AMD K6-2 (stepping 8 and above) and K6-3 processors have two | 18 | The AMD K6-2 (stepping 8 and above) and K6-3 processors have two |
| 19 | MTRRs. These are supported. The AMD Athlon family provide 8 Intel | 19 | MTRRs. These are supported. The AMD Athlon family provide 8 Intel |
| 20 | style MTRRs. | 20 | style MTRRs. |
| 21 | 21 | ||
| 22 | The Centaur C6 (WinChip) has 8 MCRs, allowing write-combining. These | 22 | The Centaur C6 (WinChip) has 8 MCRs, allowing write-combining. These |
| 23 | are supported. | 23 | are supported. |
| 24 | 24 | ||
| @@ -87,7 +87,7 @@ reg00: base=0x00000000 ( 0MB), size= 64MB: write-back, count=1 | |||
| 87 | reg01: base=0xfb000000 (4016MB), size= 16MB: write-combining, count=1 | 87 | reg01: base=0xfb000000 (4016MB), size= 16MB: write-combining, count=1 |
| 88 | reg02: base=0xfb000000 (4016MB), size= 4kB: uncachable, count=1 | 88 | reg02: base=0xfb000000 (4016MB), size= 4kB: uncachable, count=1 |
| 89 | 89 | ||
| 90 | Some cards (especially Voodoo Graphics boards) need this 4 kB area | 90 | Some cards (especially Voodoo Graphics boards) need this 4 kB area |
| 91 | excluded from the beginning of the region because it is used for | 91 | excluded from the beginning of the region because it is used for |
| 92 | registers. | 92 | registers. |
| 93 | 93 | ||
diff --git a/Documentation/x86/pat.txt b/Documentation/x86/pat.txt index 17965f927c15..c93ff5f4c0dd 100644 --- a/Documentation/x86/pat.txt +++ b/Documentation/x86/pat.txt | |||
| @@ -14,6 +14,10 @@ PAT allows for different types of memory attributes. The most commonly used | |||
| 14 | ones that will be supported at this time are Write-back, Uncached, | 14 | ones that will be supported at this time are Write-back, Uncached, |
| 15 | Write-combined and Uncached Minus. | 15 | Write-combined and Uncached Minus. |
| 16 | 16 | ||
| 17 | |||
| 18 | PAT APIs | ||
| 19 | -------- | ||
| 20 | |||
| 17 | There are many different APIs in the kernel that allows setting of memory | 21 | There are many different APIs in the kernel that allows setting of memory |
| 18 | attributes at the page level. In order to avoid aliasing, these interfaces | 22 | attributes at the page level. In order to avoid aliasing, these interfaces |
| 19 | should be used thoughtfully. Below is a table of interfaces available, | 23 | should be used thoughtfully. Below is a table of interfaces available, |
| @@ -26,38 +30,38 @@ address range to avoid any aliasing. | |||
| 26 | API | RAM | ACPI,... | Reserved/Holes | | 30 | API | RAM | ACPI,... | Reserved/Holes | |
| 27 | -----------------------|----------|------------|------------------| | 31 | -----------------------|----------|------------|------------------| |
| 28 | | | | | | 32 | | | | | |
| 29 | ioremap | -- | UC | UC | | 33 | ioremap | -- | UC- | UC- | |
| 30 | | | | | | 34 | | | | | |
| 31 | ioremap_cache | -- | WB | WB | | 35 | ioremap_cache | -- | WB | WB | |
| 32 | | | | | | 36 | | | | | |
| 33 | ioremap_nocache | -- | UC | UC | | 37 | ioremap_nocache | -- | UC- | UC- | |
| 34 | | | | | | 38 | | | | | |
| 35 | ioremap_wc | -- | -- | WC | | 39 | ioremap_wc | -- | -- | WC | |
| 36 | | | | | | 40 | | | | | |
| 37 | set_memory_uc | UC | -- | -- | | 41 | set_memory_uc | UC- | -- | -- | |
| 38 | set_memory_wb | | | | | 42 | set_memory_wb | | | | |
| 39 | | | | | | 43 | | | | | |
| 40 | set_memory_wc | WC | -- | -- | | 44 | set_memory_wc | WC | -- | -- | |
| 41 | set_memory_wb | | | | | 45 | set_memory_wb | | | | |
| 42 | | | | | | 46 | | | | | |
| 43 | pci sysfs resource | -- | -- | UC | | 47 | pci sysfs resource | -- | -- | UC- | |
| 44 | | | | | | 48 | | | | | |
| 45 | pci sysfs resource_wc | -- | -- | WC | | 49 | pci sysfs resource_wc | -- | -- | WC | |
| 46 | is IORESOURCE_PREFETCH| | | | | 50 | is IORESOURCE_PREFETCH| | | | |
| 47 | | | | | | 51 | | | | | |
| 48 | pci proc | -- | -- | UC | | 52 | pci proc | -- | -- | UC- | |
| 49 | !PCIIOC_WRITE_COMBINE | | | | | 53 | !PCIIOC_WRITE_COMBINE | | | | |
| 50 | | | | | | 54 | | | | | |
| 51 | pci proc | -- | -- | WC | | 55 | pci proc | -- | -- | WC | |
| 52 | PCIIOC_WRITE_COMBINE | | | | | 56 | PCIIOC_WRITE_COMBINE | | | | |
| 53 | | | | | | 57 | | | | | |
| 54 | /dev/mem | -- | UC | UC | | 58 | /dev/mem | -- | WB/WC/UC- | WB/WC/UC- | |
| 55 | read-write | | | | | 59 | read-write | | | | |
| 56 | | | | | | 60 | | | | | |
| 57 | /dev/mem | -- | UC | UC | | 61 | /dev/mem | -- | UC- | UC- | |
| 58 | mmap SYNC flag | | | | | 62 | mmap SYNC flag | | | | |
| 59 | | | | | | 63 | | | | | |
| 60 | /dev/mem | -- | WB/WC/UC | WB/WC/UC | | 64 | /dev/mem | -- | WB/WC/UC- | WB/WC/UC- | |
| 61 | mmap !SYNC flag | |(from exist-| (from exist- | | 65 | mmap !SYNC flag | |(from exist-| (from exist- | |
| 62 | and | | ing alias)| ing alias) | | 66 | and | | ing alias)| ing alias) | |
| 63 | any alias to this area| | | | | 67 | any alias to this area| | | | |
| @@ -68,7 +72,7 @@ pci proc | -- | -- | WC | | |||
| 68 | and | | | | | 72 | and | | | | |
| 69 | MTRR says WB | | | | | 73 | MTRR says WB | | | | |
| 70 | | | | | | 74 | | | | | |
| 71 | /dev/mem | -- | -- | UC_MINUS | | 75 | /dev/mem | -- | -- | UC- | |
| 72 | mmap !SYNC flag | | | | | 76 | mmap !SYNC flag | | | | |
| 73 | no alias to this area | | | | | 77 | no alias to this area | | | | |
| 74 | and | | | | | 78 | and | | | | |
| @@ -98,3 +102,35 @@ types. | |||
| 98 | 102 | ||
| 99 | Drivers should use set_memory_[uc|wc] to set access type for RAM ranges. | 103 | Drivers should use set_memory_[uc|wc] to set access type for RAM ranges. |
| 100 | 104 | ||
| 105 | |||
| 106 | PAT debugging | ||
| 107 | ------------- | ||
| 108 | |||
| 109 | With CONFIG_DEBUG_FS enabled, PAT memtype list can be examined by | ||
| 110 | |||
| 111 | # mount -t debugfs debugfs /sys/kernel/debug | ||
| 112 | # cat /sys/kernel/debug/x86/pat_memtype_list | ||
| 113 | PAT memtype list: | ||
| 114 | uncached-minus @ 0x7fadf000-0x7fae0000 | ||
| 115 | uncached-minus @ 0x7fb19000-0x7fb1a000 | ||
| 116 | uncached-minus @ 0x7fb1a000-0x7fb1b000 | ||
| 117 | uncached-minus @ 0x7fb1b000-0x7fb1c000 | ||
| 118 | uncached-minus @ 0x7fb1c000-0x7fb1d000 | ||
| 119 | uncached-minus @ 0x7fb1d000-0x7fb1e000 | ||
| 120 | uncached-minus @ 0x7fb1e000-0x7fb25000 | ||
| 121 | uncached-minus @ 0x7fb25000-0x7fb26000 | ||
| 122 | uncached-minus @ 0x7fb26000-0x7fb27000 | ||
| 123 | uncached-minus @ 0x7fb27000-0x7fb28000 | ||
| 124 | uncached-minus @ 0x7fb28000-0x7fb2e000 | ||
| 125 | uncached-minus @ 0x7fb2e000-0x7fb2f000 | ||
| 126 | uncached-minus @ 0x7fb2f000-0x7fb30000 | ||
| 127 | uncached-minus @ 0x7fb31000-0x7fb32000 | ||
| 128 | uncached-minus @ 0x80000000-0x90000000 | ||
| 129 | |||
| 130 | This list shows physical address ranges and various PAT settings used to | ||
| 131 | access those physical address ranges. | ||
| 132 | |||
| 133 | Another, more verbose way of getting PAT related debug messages is with | ||
| 134 | "debugpat" boot parameter. With this parameter, various debug messages are | ||
| 135 | printed to dmesg log. | ||
| 136 | |||
diff --git a/Documentation/x86/i386/usb-legacy-support.txt b/Documentation/x86/usb-legacy-support.txt index 1894cdfc69d9..1894cdfc69d9 100644 --- a/Documentation/x86/i386/usb-legacy-support.txt +++ b/Documentation/x86/usb-legacy-support.txt | |||
diff --git a/Documentation/x86/x86_64/boot-options.txt b/Documentation/x86/x86_64/boot-options.txt index b0c7b6c4abda..72ffb5373ec7 100644 --- a/Documentation/x86/x86_64/boot-options.txt +++ b/Documentation/x86/x86_64/boot-options.txt | |||
| @@ -54,10 +54,6 @@ APICs | |||
| 54 | apicmaintimer. Useful when your PIT timer is totally | 54 | apicmaintimer. Useful when your PIT timer is totally |
| 55 | broken. | 55 | broken. |
| 56 | 56 | ||
| 57 | disable_8254_timer / enable_8254_timer | ||
| 58 | Enable interrupt 0 timer routing over the 8254 in addition to over | ||
| 59 | the IO-APIC. The kernel tries to set a sensible default. | ||
| 60 | |||
| 61 | Early Console | 57 | Early Console |
| 62 | 58 | ||
| 63 | syntax: earlyprintk=vga | 59 | syntax: earlyprintk=vga |
diff --git a/Documentation/x86/i386/zero-page.txt b/Documentation/x86/zero-page.txt index 169ad423a3d1..169ad423a3d1 100644 --- a/Documentation/x86/i386/zero-page.txt +++ b/Documentation/x86/zero-page.txt | |||
diff --git a/arch/x86/Kconfig b/arch/x86/Kconfig index ebfd7ff82ade..97f0d2b6dc0c 100644 --- a/arch/x86/Kconfig +++ b/arch/x86/Kconfig | |||
| @@ -1021,7 +1021,7 @@ config HAVE_ARCH_ALLOC_REMAP | |||
| 1021 | 1021 | ||
| 1022 | config ARCH_FLATMEM_ENABLE | 1022 | config ARCH_FLATMEM_ENABLE |
| 1023 | def_bool y | 1023 | def_bool y |
| 1024 | depends on X86_32 && ARCH_SELECT_MEMORY_MODEL && X86_PC && !NUMA | 1024 | depends on X86_32 && ARCH_SELECT_MEMORY_MODEL && !NUMA |
| 1025 | 1025 | ||
| 1026 | config ARCH_DISCONTIGMEM_ENABLE | 1026 | config ARCH_DISCONTIGMEM_ENABLE |
| 1027 | def_bool y | 1027 | def_bool y |
| @@ -1037,7 +1037,7 @@ config ARCH_SPARSEMEM_DEFAULT | |||
| 1037 | 1037 | ||
| 1038 | config ARCH_SPARSEMEM_ENABLE | 1038 | config ARCH_SPARSEMEM_ENABLE |
| 1039 | def_bool y | 1039 | def_bool y |
| 1040 | depends on X86_64 || NUMA || (EXPERIMENTAL && X86_PC) | 1040 | depends on X86_64 || NUMA || (EXPERIMENTAL && X86_PC) || X86_GENERICARCH |
| 1041 | select SPARSEMEM_STATIC if X86_32 | 1041 | select SPARSEMEM_STATIC if X86_32 |
| 1042 | select SPARSEMEM_VMEMMAP_ENABLE if X86_64 | 1042 | select SPARSEMEM_VMEMMAP_ENABLE if X86_64 |
| 1043 | 1043 | ||
| @@ -1118,10 +1118,10 @@ config MTRR | |||
| 1118 | You can safely say Y even if your machine doesn't have MTRRs, you'll | 1118 | You can safely say Y even if your machine doesn't have MTRRs, you'll |
| 1119 | just add about 9 KB to your kernel. | 1119 | just add about 9 KB to your kernel. |
| 1120 | 1120 | ||
| 1121 | See <file:Documentation/mtrr.txt> for more information. | 1121 | See <file:Documentation/x86/mtrr.txt> for more information. |
| 1122 | 1122 | ||
| 1123 | config MTRR_SANITIZER | 1123 | config MTRR_SANITIZER |
| 1124 | bool | 1124 | def_bool y |
| 1125 | prompt "MTRR cleanup support" | 1125 | prompt "MTRR cleanup support" |
| 1126 | depends on MTRR | 1126 | depends on MTRR |
| 1127 | help | 1127 | help |
| @@ -1132,7 +1132,7 @@ config MTRR_SANITIZER | |||
| 1132 | The largest mtrr entry size for a continous block can be set with | 1132 | The largest mtrr entry size for a continous block can be set with |
| 1133 | mtrr_chunk_size. | 1133 | mtrr_chunk_size. |
| 1134 | 1134 | ||
| 1135 | If unsure, say N. | 1135 | If unsure, say Y. |
| 1136 | 1136 | ||
| 1137 | config MTRR_SANITIZER_ENABLE_DEFAULT | 1137 | config MTRR_SANITIZER_ENABLE_DEFAULT |
| 1138 | int "MTRR cleanup enable value (0-1)" | 1138 | int "MTRR cleanup enable value (0-1)" |
| @@ -1192,7 +1192,6 @@ config IRQBALANCE | |||
| 1192 | config SECCOMP | 1192 | config SECCOMP |
| 1193 | def_bool y | 1193 | def_bool y |
| 1194 | prompt "Enable seccomp to safely compute untrusted bytecode" | 1194 | prompt "Enable seccomp to safely compute untrusted bytecode" |
| 1195 | depends on PROC_FS | ||
| 1196 | help | 1195 | help |
| 1197 | This kernel feature is useful for number crunching applications | 1196 | This kernel feature is useful for number crunching applications |
| 1198 | that may need to compute untrusted bytecode during their | 1197 | that may need to compute untrusted bytecode during their |
| @@ -1200,7 +1199,7 @@ config SECCOMP | |||
| 1200 | the process as file descriptors supporting the read/write | 1199 | the process as file descriptors supporting the read/write |
| 1201 | syscalls, it's possible to isolate those applications in | 1200 | syscalls, it's possible to isolate those applications in |
| 1202 | their own address space using seccomp. Once seccomp is | 1201 | their own address space using seccomp. Once seccomp is |
| 1203 | enabled via /proc/<pid>/seccomp, it cannot be disabled | 1202 | enabled via prctl(PR_SET_SECCOMP), it cannot be disabled |
| 1204 | and the task is only allowed to execute a few safe syscalls | 1203 | and the task is only allowed to execute a few safe syscalls |
| 1205 | defined by each seccomp mode. | 1204 | defined by each seccomp mode. |
| 1206 | 1205 | ||
| @@ -1357,14 +1356,14 @@ config PHYSICAL_ALIGN | |||
| 1357 | Don't change this unless you know what you are doing. | 1356 | Don't change this unless you know what you are doing. |
| 1358 | 1357 | ||
| 1359 | config HOTPLUG_CPU | 1358 | config HOTPLUG_CPU |
| 1360 | bool "Support for suspend on SMP and hot-pluggable CPUs (EXPERIMENTAL)" | 1359 | bool "Support for hot-pluggable CPUs" |
| 1361 | depends on SMP && HOTPLUG && EXPERIMENTAL && !X86_VOYAGER | 1360 | depends on SMP && HOTPLUG && !X86_VOYAGER |
| 1362 | ---help--- | 1361 | ---help--- |
| 1363 | Say Y here to experiment with turning CPUs off and on, and to | 1362 | Say Y here to allow turning CPUs off and on. CPUs can be |
| 1364 | enable suspend on SMP systems. CPUs can be controlled through | 1363 | controlled through /sys/devices/system/cpu. |
| 1365 | /sys/devices/system/cpu. | 1364 | ( Note: power management support will enable this option |
| 1366 | Say N if you want to disable CPU hotplug and don't need to | 1365 | automatically on SMP systems. ) |
| 1367 | suspend. | 1366 | Say N if you want to disable CPU hotplug. |
| 1368 | 1367 | ||
| 1369 | config COMPAT_VDSO | 1368 | config COMPAT_VDSO |
| 1370 | def_bool y | 1369 | def_bool y |
| @@ -1379,6 +1378,51 @@ config COMPAT_VDSO | |||
| 1379 | 1378 | ||
| 1380 | If unsure, say Y. | 1379 | If unsure, say Y. |
| 1381 | 1380 | ||
| 1381 | config CMDLINE_BOOL | ||
| 1382 | bool "Built-in kernel command line" | ||
| 1383 | default n | ||
| 1384 | help | ||
| 1385 | Allow for specifying boot arguments to the kernel at | ||
| 1386 | build time. On some systems (e.g. embedded ones), it is | ||
| 1387 | necessary or convenient to provide some or all of the | ||
| 1388 | kernel boot arguments with the kernel itself (that is, | ||
| 1389 | to not rely on the boot loader to provide them.) | ||
| 1390 | |||
| 1391 | To compile command line arguments into the kernel, | ||
| 1392 | set this option to 'Y', then fill in the | ||
| 1393 | the boot arguments in CONFIG_CMDLINE. | ||
| 1394 | |||
| 1395 | Systems with fully functional boot loaders (i.e. non-embedded) | ||
| 1396 | should leave this option set to 'N'. | ||
| 1397 | |||
| 1398 | config CMDLINE | ||
| 1399 | string "Built-in kernel command string" | ||
| 1400 | depends on CMDLINE_BOOL | ||
| 1401 | default "" | ||
| 1402 | help | ||
| 1403 | Enter arguments here that should be compiled into the kernel | ||
| 1404 | image and used at boot time. If the boot loader provides a | ||
| 1405 | command line at boot time, it is appended to this string to | ||
| 1406 | form the full kernel command line, when the system boots. | ||
| 1407 | |||
| 1408 | However, you can use the CONFIG_CMDLINE_OVERRIDE option to | ||
| 1409 | change this behavior. | ||
| 1410 | |||
| 1411 | In most cases, the command line (whether built-in or provided | ||
| 1412 | by the boot loader) should specify the device for the root | ||
| 1413 | file system. | ||
| 1414 | |||
| 1415 | config CMDLINE_OVERRIDE | ||
| 1416 | bool "Built-in command line overrides boot loader arguments" | ||
| 1417 | default n | ||
| 1418 | depends on CMDLINE_BOOL | ||
| 1419 | help | ||
| 1420 | Set this option to 'Y' to have the kernel ignore the boot loader | ||
| 1421 | command line, and use ONLY the built-in command line. | ||
| 1422 | |||
| 1423 | This is used to work around broken boot loaders. This should | ||
| 1424 | be set to 'N' under normal conditions. | ||
| 1425 | |||
| 1382 | endmenu | 1426 | endmenu |
| 1383 | 1427 | ||
| 1384 | config ARCH_ENABLE_MEMORY_HOTPLUG | 1428 | config ARCH_ENABLE_MEMORY_HOTPLUG |
| @@ -1774,7 +1818,7 @@ config COMPAT_FOR_U64_ALIGNMENT | |||
| 1774 | 1818 | ||
| 1775 | config SYSVIPC_COMPAT | 1819 | config SYSVIPC_COMPAT |
| 1776 | def_bool y | 1820 | def_bool y |
| 1777 | depends on X86_64 && COMPAT && SYSVIPC | 1821 | depends on COMPAT && SYSVIPC |
| 1778 | 1822 | ||
| 1779 | endmenu | 1823 | endmenu |
| 1780 | 1824 | ||
diff --git a/arch/x86/boot/compressed/head_32.S b/arch/x86/boot/compressed/head_32.S index ba7736cf2ec7..29c5fbf08392 100644 --- a/arch/x86/boot/compressed/head_32.S +++ b/arch/x86/boot/compressed/head_32.S | |||
| @@ -137,14 +137,15 @@ relocated: | |||
| 137 | */ | 137 | */ |
| 138 | movl output_len(%ebx), %eax | 138 | movl output_len(%ebx), %eax |
| 139 | pushl %eax | 139 | pushl %eax |
| 140 | # push arguments for decompress_kernel: | ||
| 140 | pushl %ebp # output address | 141 | pushl %ebp # output address |
| 141 | movl input_len(%ebx), %eax | 142 | movl input_len(%ebx), %eax |
| 142 | pushl %eax # input_len | 143 | pushl %eax # input_len |
| 143 | leal input_data(%ebx), %eax | 144 | leal input_data(%ebx), %eax |
| 144 | pushl %eax # input_data | 145 | pushl %eax # input_data |
| 145 | leal boot_heap(%ebx), %eax | 146 | leal boot_heap(%ebx), %eax |
| 146 | pushl %eax # heap area as third argument | 147 | pushl %eax # heap area |
| 147 | pushl %esi # real mode pointer as second arg | 148 | pushl %esi # real mode pointer |
| 148 | call decompress_kernel | 149 | call decompress_kernel |
| 149 | addl $20, %esp | 150 | addl $20, %esp |
| 150 | popl %ecx | 151 | popl %ecx |
diff --git a/arch/x86/boot/compressed/misc.c b/arch/x86/boot/compressed/misc.c index aaf5a2131efc..5780d361105b 100644 --- a/arch/x86/boot/compressed/misc.c +++ b/arch/x86/boot/compressed/misc.c | |||
| @@ -27,7 +27,7 @@ | |||
| 27 | #include <linux/linkage.h> | 27 | #include <linux/linkage.h> |
| 28 | #include <linux/screen_info.h> | 28 | #include <linux/screen_info.h> |
| 29 | #include <linux/elf.h> | 29 | #include <linux/elf.h> |
| 30 | #include <asm/io.h> | 30 | #include <linux/io.h> |
| 31 | #include <asm/page.h> | 31 | #include <asm/page.h> |
| 32 | #include <asm/boot.h> | 32 | #include <asm/boot.h> |
| 33 | #include <asm/bootparam.h> | 33 | #include <asm/bootparam.h> |
| @@ -251,7 +251,7 @@ static void __putstr(int error, const char *s) | |||
| 251 | y--; | 251 | y--; |
| 252 | } | 252 | } |
| 253 | } else { | 253 | } else { |
| 254 | vidmem [(x + cols * y) * 2] = c; | 254 | vidmem[(x + cols * y) * 2] = c; |
| 255 | if (++x >= cols) { | 255 | if (++x >= cols) { |
| 256 | x = 0; | 256 | x = 0; |
| 257 | if (++y >= lines) { | 257 | if (++y >= lines) { |
| @@ -277,7 +277,8 @@ static void *memset(void *s, int c, unsigned n) | |||
| 277 | int i; | 277 | int i; |
| 278 | char *ss = s; | 278 | char *ss = s; |
| 279 | 279 | ||
| 280 | for (i = 0; i < n; i++) ss[i] = c; | 280 | for (i = 0; i < n; i++) |
| 281 | ss[i] = c; | ||
| 281 | return s; | 282 | return s; |
| 282 | } | 283 | } |
| 283 | 284 | ||
| @@ -287,7 +288,8 @@ static void *memcpy(void *dest, const void *src, unsigned n) | |||
| 287 | const char *s = src; | 288 | const char *s = src; |
| 288 | char *d = dest; | 289 | char *d = dest; |
| 289 | 290 | ||
| 290 | for (i = 0; i < n; i++) d[i] = s[i]; | 291 | for (i = 0; i < n; i++) |
| 292 | d[i] = s[i]; | ||
| 291 | return dest; | 293 | return dest; |
| 292 | } | 294 | } |
| 293 | 295 | ||
diff --git a/arch/x86/boot/header.S b/arch/x86/boot/header.S index af86e431acfa..b993062e9a5f 100644 --- a/arch/x86/boot/header.S +++ b/arch/x86/boot/header.S | |||
| @@ -30,7 +30,6 @@ SYSSEG = DEF_SYSSEG /* system loaded at 0x10000 (65536) */ | |||
| 30 | SYSSIZE = DEF_SYSSIZE /* system size: # of 16-byte clicks */ | 30 | SYSSIZE = DEF_SYSSIZE /* system size: # of 16-byte clicks */ |
| 31 | /* to be loaded */ | 31 | /* to be loaded */ |
| 32 | ROOT_DEV = 0 /* ROOT_DEV is now written by "build" */ | 32 | ROOT_DEV = 0 /* ROOT_DEV is now written by "build" */ |
| 33 | SWAP_DEV = 0 /* SWAP_DEV is now written by "build" */ | ||
| 34 | 33 | ||
| 35 | #ifndef SVGA_MODE | 34 | #ifndef SVGA_MODE |
| 36 | #define SVGA_MODE ASK_VGA | 35 | #define SVGA_MODE ASK_VGA |
diff --git a/arch/x86/configs/i386_defconfig b/arch/x86/configs/i386_defconfig index 104275e191a8..ef9a52005ec9 100644 --- a/arch/x86/configs/i386_defconfig +++ b/arch/x86/configs/i386_defconfig | |||
| @@ -1,7 +1,7 @@ | |||
| 1 | # | 1 | # |
| 2 | # Automatically generated make config: don't edit | 2 | # Automatically generated make config: don't edit |
| 3 | # Linux kernel version: 2.6.27-rc4 | 3 | # Linux kernel version: 2.6.27-rc5 |
| 4 | # Mon Aug 25 15:04:00 2008 | 4 | # Wed Sep 3 17:23:09 2008 |
| 5 | # | 5 | # |
| 6 | # CONFIG_64BIT is not set | 6 | # CONFIG_64BIT is not set |
| 7 | CONFIG_X86_32=y | 7 | CONFIG_X86_32=y |
| @@ -202,7 +202,7 @@ CONFIG_SCHED_NO_NO_OMIT_FRAME_POINTER=y | |||
| 202 | # CONFIG_M586 is not set | 202 | # CONFIG_M586 is not set |
| 203 | # CONFIG_M586TSC is not set | 203 | # CONFIG_M586TSC is not set |
| 204 | # CONFIG_M586MMX is not set | 204 | # CONFIG_M586MMX is not set |
| 205 | # CONFIG_M686 is not set | 205 | CONFIG_M686=y |
| 206 | # CONFIG_MPENTIUMII is not set | 206 | # CONFIG_MPENTIUMII is not set |
| 207 | # CONFIG_MPENTIUMIII is not set | 207 | # CONFIG_MPENTIUMIII is not set |
| 208 | # CONFIG_MPENTIUMM is not set | 208 | # CONFIG_MPENTIUMM is not set |
| @@ -221,13 +221,14 @@ CONFIG_SCHED_NO_NO_OMIT_FRAME_POINTER=y | |||
| 221 | # CONFIG_MVIAC3_2 is not set | 221 | # CONFIG_MVIAC3_2 is not set |
| 222 | # CONFIG_MVIAC7 is not set | 222 | # CONFIG_MVIAC7 is not set |
| 223 | # CONFIG_MPSC is not set | 223 | # CONFIG_MPSC is not set |
| 224 | CONFIG_MCORE2=y | 224 | # CONFIG_MCORE2 is not set |
| 225 | # CONFIG_GENERIC_CPU is not set | 225 | # CONFIG_GENERIC_CPU is not set |
| 226 | CONFIG_X86_GENERIC=y | 226 | CONFIG_X86_GENERIC=y |
| 227 | CONFIG_X86_CPU=y | 227 | CONFIG_X86_CPU=y |
| 228 | CONFIG_X86_CMPXCHG=y | 228 | CONFIG_X86_CMPXCHG=y |
| 229 | CONFIG_X86_L1_CACHE_SHIFT=7 | 229 | CONFIG_X86_L1_CACHE_SHIFT=7 |
| 230 | CONFIG_X86_XADD=y | 230 | CONFIG_X86_XADD=y |
| 231 | # CONFIG_X86_PPRO_FENCE is not set | ||
| 231 | CONFIG_X86_WP_WORKS_OK=y | 232 | CONFIG_X86_WP_WORKS_OK=y |
| 232 | CONFIG_X86_INVLPG=y | 233 | CONFIG_X86_INVLPG=y |
| 233 | CONFIG_X86_BSWAP=y | 234 | CONFIG_X86_BSWAP=y |
| @@ -235,14 +236,15 @@ CONFIG_X86_POPAD_OK=y | |||
| 235 | CONFIG_X86_INTEL_USERCOPY=y | 236 | CONFIG_X86_INTEL_USERCOPY=y |
| 236 | CONFIG_X86_USE_PPRO_CHECKSUM=y | 237 | CONFIG_X86_USE_PPRO_CHECKSUM=y |
| 237 | CONFIG_X86_TSC=y | 238 | CONFIG_X86_TSC=y |
| 239 | CONFIG_X86_CMOV=y | ||
| 238 | CONFIG_X86_MINIMUM_CPU_FAMILY=4 | 240 | CONFIG_X86_MINIMUM_CPU_FAMILY=4 |
| 239 | CONFIG_X86_DEBUGCTLMSR=y | 241 | CONFIG_X86_DEBUGCTLMSR=y |
| 240 | CONFIG_HPET_TIMER=y | 242 | CONFIG_HPET_TIMER=y |
| 241 | CONFIG_HPET_EMULATE_RTC=y | 243 | CONFIG_HPET_EMULATE_RTC=y |
| 242 | CONFIG_DMI=y | 244 | CONFIG_DMI=y |
| 243 | # CONFIG_IOMMU_HELPER is not set | 245 | # CONFIG_IOMMU_HELPER is not set |
| 244 | CONFIG_NR_CPUS=4 | 246 | CONFIG_NR_CPUS=64 |
| 245 | # CONFIG_SCHED_SMT is not set | 247 | CONFIG_SCHED_SMT=y |
| 246 | CONFIG_SCHED_MC=y | 248 | CONFIG_SCHED_MC=y |
| 247 | # CONFIG_PREEMPT_NONE is not set | 249 | # CONFIG_PREEMPT_NONE is not set |
| 248 | CONFIG_PREEMPT_VOLUNTARY=y | 250 | CONFIG_PREEMPT_VOLUNTARY=y |
| @@ -254,7 +256,8 @@ CONFIG_VM86=y | |||
| 254 | # CONFIG_TOSHIBA is not set | 256 | # CONFIG_TOSHIBA is not set |
| 255 | # CONFIG_I8K is not set | 257 | # CONFIG_I8K is not set |
| 256 | CONFIG_X86_REBOOTFIXUPS=y | 258 | CONFIG_X86_REBOOTFIXUPS=y |
| 257 | # CONFIG_MICROCODE is not set | 259 | CONFIG_MICROCODE=y |
| 260 | CONFIG_MICROCODE_OLD_INTERFACE=y | ||
| 258 | CONFIG_X86_MSR=y | 261 | CONFIG_X86_MSR=y |
| 259 | CONFIG_X86_CPUID=y | 262 | CONFIG_X86_CPUID=y |
| 260 | # CONFIG_NOHIGHMEM is not set | 263 | # CONFIG_NOHIGHMEM is not set |
| @@ -2115,7 +2118,7 @@ CONFIG_IO_DELAY_0X80=y | |||
| 2115 | CONFIG_DEFAULT_IO_DELAY_TYPE=0 | 2118 | CONFIG_DEFAULT_IO_DELAY_TYPE=0 |
| 2116 | CONFIG_DEBUG_BOOT_PARAMS=y | 2119 | CONFIG_DEBUG_BOOT_PARAMS=y |
| 2117 | # CONFIG_CPA_DEBUG is not set | 2120 | # CONFIG_CPA_DEBUG is not set |
| 2118 | # CONFIG_OPTIMIZE_INLINING is not set | 2121 | CONFIG_OPTIMIZE_INLINING=y |
| 2119 | 2122 | ||
| 2120 | # | 2123 | # |
| 2121 | # Security options | 2124 | # Security options |
diff --git a/arch/x86/configs/x86_64_defconfig b/arch/x86/configs/x86_64_defconfig index 678c8acefe04..e620ea6e2a7a 100644 --- a/arch/x86/configs/x86_64_defconfig +++ b/arch/x86/configs/x86_64_defconfig | |||
| @@ -1,7 +1,7 @@ | |||
| 1 | # | 1 | # |
| 2 | # Automatically generated make config: don't edit | 2 | # Automatically generated make config: don't edit |
| 3 | # Linux kernel version: 2.6.27-rc4 | 3 | # Linux kernel version: 2.6.27-rc5 |
| 4 | # Mon Aug 25 14:40:46 2008 | 4 | # Wed Sep 3 17:13:39 2008 |
| 5 | # | 5 | # |
| 6 | CONFIG_64BIT=y | 6 | CONFIG_64BIT=y |
| 7 | # CONFIG_X86_32 is not set | 7 | # CONFIG_X86_32 is not set |
| @@ -218,17 +218,14 @@ CONFIG_X86_PC=y | |||
| 218 | # CONFIG_MVIAC3_2 is not set | 218 | # CONFIG_MVIAC3_2 is not set |
| 219 | # CONFIG_MVIAC7 is not set | 219 | # CONFIG_MVIAC7 is not set |
| 220 | # CONFIG_MPSC is not set | 220 | # CONFIG_MPSC is not set |
| 221 | CONFIG_MCORE2=y | 221 | # CONFIG_MCORE2 is not set |
| 222 | # CONFIG_GENERIC_CPU is not set | 222 | CONFIG_GENERIC_CPU=y |
| 223 | CONFIG_X86_CPU=y | 223 | CONFIG_X86_CPU=y |
| 224 | CONFIG_X86_L1_CACHE_BYTES=64 | 224 | CONFIG_X86_L1_CACHE_BYTES=128 |
| 225 | CONFIG_X86_INTERNODE_CACHE_BYTES=64 | 225 | CONFIG_X86_INTERNODE_CACHE_BYTES=128 |
| 226 | CONFIG_X86_CMPXCHG=y | 226 | CONFIG_X86_CMPXCHG=y |
| 227 | CONFIG_X86_L1_CACHE_SHIFT=6 | 227 | CONFIG_X86_L1_CACHE_SHIFT=7 |
| 228 | CONFIG_X86_WP_WORKS_OK=y | 228 | CONFIG_X86_WP_WORKS_OK=y |
| 229 | CONFIG_X86_INTEL_USERCOPY=y | ||
| 230 | CONFIG_X86_USE_PPRO_CHECKSUM=y | ||
| 231 | CONFIG_X86_P6_NOP=y | ||
| 232 | CONFIG_X86_TSC=y | 229 | CONFIG_X86_TSC=y |
| 233 | CONFIG_X86_CMPXCHG64=y | 230 | CONFIG_X86_CMPXCHG64=y |
| 234 | CONFIG_X86_CMOV=y | 231 | CONFIG_X86_CMOV=y |
| @@ -243,9 +240,8 @@ CONFIG_CALGARY_IOMMU_ENABLED_BY_DEFAULT=y | |||
| 243 | CONFIG_AMD_IOMMU=y | 240 | CONFIG_AMD_IOMMU=y |
| 244 | CONFIG_SWIOTLB=y | 241 | CONFIG_SWIOTLB=y |
| 245 | CONFIG_IOMMU_HELPER=y | 242 | CONFIG_IOMMU_HELPER=y |
| 246 | # CONFIG_MAXSMP is not set | 243 | CONFIG_NR_CPUS=64 |
| 247 | CONFIG_NR_CPUS=4 | 244 | CONFIG_SCHED_SMT=y |
| 248 | # CONFIG_SCHED_SMT is not set | ||
| 249 | CONFIG_SCHED_MC=y | 245 | CONFIG_SCHED_MC=y |
| 250 | # CONFIG_PREEMPT_NONE is not set | 246 | # CONFIG_PREEMPT_NONE is not set |
| 251 | CONFIG_PREEMPT_VOLUNTARY=y | 247 | CONFIG_PREEMPT_VOLUNTARY=y |
| @@ -254,7 +250,8 @@ CONFIG_X86_LOCAL_APIC=y | |||
| 254 | CONFIG_X86_IO_APIC=y | 250 | CONFIG_X86_IO_APIC=y |
| 255 | # CONFIG_X86_MCE is not set | 251 | # CONFIG_X86_MCE is not set |
| 256 | # CONFIG_I8K is not set | 252 | # CONFIG_I8K is not set |
| 257 | # CONFIG_MICROCODE is not set | 253 | CONFIG_MICROCODE=y |
| 254 | CONFIG_MICROCODE_OLD_INTERFACE=y | ||
| 258 | CONFIG_X86_MSR=y | 255 | CONFIG_X86_MSR=y |
| 259 | CONFIG_X86_CPUID=y | 256 | CONFIG_X86_CPUID=y |
| 260 | CONFIG_NUMA=y | 257 | CONFIG_NUMA=y |
| @@ -290,7 +287,7 @@ CONFIG_BOUNCE=y | |||
| 290 | CONFIG_VIRT_TO_BUS=y | 287 | CONFIG_VIRT_TO_BUS=y |
| 291 | CONFIG_MTRR=y | 288 | CONFIG_MTRR=y |
| 292 | # CONFIG_MTRR_SANITIZER is not set | 289 | # CONFIG_MTRR_SANITIZER is not set |
| 293 | # CONFIG_X86_PAT is not set | 290 | CONFIG_X86_PAT=y |
| 294 | CONFIG_EFI=y | 291 | CONFIG_EFI=y |
| 295 | CONFIG_SECCOMP=y | 292 | CONFIG_SECCOMP=y |
| 296 | # CONFIG_HZ_100 is not set | 293 | # CONFIG_HZ_100 is not set |
| @@ -2089,7 +2086,7 @@ CONFIG_IO_DELAY_0X80=y | |||
| 2089 | CONFIG_DEFAULT_IO_DELAY_TYPE=0 | 2086 | CONFIG_DEFAULT_IO_DELAY_TYPE=0 |
| 2090 | CONFIG_DEBUG_BOOT_PARAMS=y | 2087 | CONFIG_DEBUG_BOOT_PARAMS=y |
| 2091 | # CONFIG_CPA_DEBUG is not set | 2088 | # CONFIG_CPA_DEBUG is not set |
| 2092 | # CONFIG_OPTIMIZE_INLINING is not set | 2089 | CONFIG_OPTIMIZE_INLINING=y |
| 2093 | 2090 | ||
| 2094 | # | 2091 | # |
| 2095 | # Security options | 2092 | # Security options |
diff --git a/arch/x86/ia32/ia32_aout.c b/arch/x86/ia32/ia32_aout.c index a0e1dbe67dc1..127ec3f07214 100644 --- a/arch/x86/ia32/ia32_aout.c +++ b/arch/x86/ia32/ia32_aout.c | |||
| @@ -85,8 +85,10 @@ static void dump_thread32(struct pt_regs *regs, struct user32 *dump) | |||
| 85 | dump->regs.ax = regs->ax; | 85 | dump->regs.ax = regs->ax; |
| 86 | dump->regs.ds = current->thread.ds; | 86 | dump->regs.ds = current->thread.ds; |
| 87 | dump->regs.es = current->thread.es; | 87 | dump->regs.es = current->thread.es; |
| 88 | asm("movl %%fs,%0" : "=r" (fs)); dump->regs.fs = fs; | 88 | savesegment(fs, fs); |
| 89 | asm("movl %%gs,%0" : "=r" (gs)); dump->regs.gs = gs; | 89 | dump->regs.fs = fs; |
| 90 | savesegment(gs, gs); | ||
| 91 | dump->regs.gs = gs; | ||
| 90 | dump->regs.orig_ax = regs->orig_ax; | 92 | dump->regs.orig_ax = regs->orig_ax; |
| 91 | dump->regs.ip = regs->ip; | 93 | dump->regs.ip = regs->ip; |
| 92 | dump->regs.cs = regs->cs; | 94 | dump->regs.cs = regs->cs; |
| @@ -430,8 +432,9 @@ beyond_if: | |||
| 430 | current->mm->start_stack = | 432 | current->mm->start_stack = |
| 431 | (unsigned long)create_aout_tables((char __user *)bprm->p, bprm); | 433 | (unsigned long)create_aout_tables((char __user *)bprm->p, bprm); |
| 432 | /* start thread */ | 434 | /* start thread */ |
| 433 | asm volatile("movl %0,%%fs" :: "r" (0)); \ | 435 | loadsegment(fs, 0); |
| 434 | asm volatile("movl %0,%%es; movl %0,%%ds": :"r" (__USER32_DS)); | 436 | loadsegment(ds, __USER32_DS); |
| 437 | loadsegment(es, __USER32_DS); | ||
| 435 | load_gs_index(0); | 438 | load_gs_index(0); |
| 436 | (regs)->ip = ex.a_entry; | 439 | (regs)->ip = ex.a_entry; |
| 437 | (regs)->sp = current->mm->start_stack; | 440 | (regs)->sp = current->mm->start_stack; |
diff --git a/arch/x86/ia32/ia32_signal.c b/arch/x86/ia32/ia32_signal.c index 20af4c79579a..f1a2ac777faf 100644 --- a/arch/x86/ia32/ia32_signal.c +++ b/arch/x86/ia32/ia32_signal.c | |||
| @@ -206,7 +206,7 @@ struct rt_sigframe | |||
| 206 | { unsigned int cur; \ | 206 | { unsigned int cur; \ |
| 207 | unsigned short pre; \ | 207 | unsigned short pre; \ |
| 208 | err |= __get_user(pre, &sc->seg); \ | 208 | err |= __get_user(pre, &sc->seg); \ |
| 209 | asm volatile("movl %%" #seg ",%0" : "=r" (cur)); \ | 209 | savesegment(seg, cur); \ |
| 210 | pre |= mask; \ | 210 | pre |= mask; \ |
| 211 | if (pre != cur) loadsegment(seg, pre); } | 211 | if (pre != cur) loadsegment(seg, pre); } |
| 212 | 212 | ||
| @@ -235,7 +235,7 @@ static int ia32_restore_sigcontext(struct pt_regs *regs, | |||
| 235 | */ | 235 | */ |
| 236 | err |= __get_user(gs, &sc->gs); | 236 | err |= __get_user(gs, &sc->gs); |
| 237 | gs |= 3; | 237 | gs |= 3; |
| 238 | asm("movl %%gs,%0" : "=r" (oldgs)); | 238 | savesegment(gs, oldgs); |
| 239 | if (gs != oldgs) | 239 | if (gs != oldgs) |
| 240 | load_gs_index(gs); | 240 | load_gs_index(gs); |
| 241 | 241 | ||
| @@ -355,14 +355,13 @@ static int ia32_setup_sigcontext(struct sigcontext_ia32 __user *sc, | |||
| 355 | { | 355 | { |
| 356 | int tmp, err = 0; | 356 | int tmp, err = 0; |
| 357 | 357 | ||
| 358 | tmp = 0; | 358 | savesegment(gs, tmp); |
| 359 | __asm__("movl %%gs,%0" : "=r"(tmp): "0"(tmp)); | ||
| 360 | err |= __put_user(tmp, (unsigned int __user *)&sc->gs); | 359 | err |= __put_user(tmp, (unsigned int __user *)&sc->gs); |
| 361 | __asm__("movl %%fs,%0" : "=r"(tmp): "0"(tmp)); | 360 | savesegment(fs, tmp); |
| 362 | err |= __put_user(tmp, (unsigned int __user *)&sc->fs); | 361 | err |= __put_user(tmp, (unsigned int __user *)&sc->fs); |
| 363 | __asm__("movl %%ds,%0" : "=r"(tmp): "0"(tmp)); | 362 | savesegment(ds, tmp); |
| 364 | err |= __put_user(tmp, (unsigned int __user *)&sc->ds); | 363 | err |= __put_user(tmp, (unsigned int __user *)&sc->ds); |
| 365 | __asm__("movl %%es,%0" : "=r"(tmp): "0"(tmp)); | 364 | savesegment(es, tmp); |
| 366 | err |= __put_user(tmp, (unsigned int __user *)&sc->es); | 365 | err |= __put_user(tmp, (unsigned int __user *)&sc->es); |
| 367 | 366 | ||
| 368 | err |= __put_user((u32)regs->di, &sc->di); | 367 | err |= __put_user((u32)regs->di, &sc->di); |
| @@ -498,8 +497,8 @@ int ia32_setup_frame(int sig, struct k_sigaction *ka, | |||
| 498 | regs->dx = 0; | 497 | regs->dx = 0; |
| 499 | regs->cx = 0; | 498 | regs->cx = 0; |
| 500 | 499 | ||
| 501 | asm volatile("movl %0,%%ds" :: "r" (__USER32_DS)); | 500 | loadsegment(ds, __USER32_DS); |
| 502 | asm volatile("movl %0,%%es" :: "r" (__USER32_DS)); | 501 | loadsegment(es, __USER32_DS); |
| 503 | 502 | ||
| 504 | regs->cs = __USER32_CS; | 503 | regs->cs = __USER32_CS; |
| 505 | regs->ss = __USER32_DS; | 504 | regs->ss = __USER32_DS; |
| @@ -591,8 +590,8 @@ int ia32_setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info, | |||
| 591 | regs->dx = (unsigned long) &frame->info; | 590 | regs->dx = (unsigned long) &frame->info; |
| 592 | regs->cx = (unsigned long) &frame->uc; | 591 | regs->cx = (unsigned long) &frame->uc; |
| 593 | 592 | ||
| 594 | asm volatile("movl %0,%%ds" :: "r" (__USER32_DS)); | 593 | loadsegment(ds, __USER32_DS); |
| 595 | asm volatile("movl %0,%%es" :: "r" (__USER32_DS)); | 594 | loadsegment(es, __USER32_DS); |
| 596 | 595 | ||
| 597 | regs->cs = __USER32_CS; | 596 | regs->cs = __USER32_CS; |
| 598 | regs->ss = __USER32_DS; | 597 | regs->ss = __USER32_DS; |
diff --git a/arch/x86/ia32/sys_ia32.c b/arch/x86/ia32/sys_ia32.c index d3c64088b981..beda4232ce69 100644 --- a/arch/x86/ia32/sys_ia32.c +++ b/arch/x86/ia32/sys_ia32.c | |||
| @@ -556,15 +556,6 @@ asmlinkage long sys32_rt_sigqueueinfo(int pid, int sig, | |||
| 556 | return ret; | 556 | return ret; |
| 557 | } | 557 | } |
| 558 | 558 | ||
| 559 | /* These are here just in case some old ia32 binary calls it. */ | ||
| 560 | asmlinkage long sys32_pause(void) | ||
| 561 | { | ||
| 562 | current->state = TASK_INTERRUPTIBLE; | ||
| 563 | schedule(); | ||
| 564 | return -ERESTARTNOHAND; | ||
| 565 | } | ||
| 566 | |||
| 567 | |||
| 568 | #ifdef CONFIG_SYSCTL_SYSCALL | 559 | #ifdef CONFIG_SYSCTL_SYSCALL |
| 569 | struct sysctl_ia32 { | 560 | struct sysctl_ia32 { |
| 570 | unsigned int name; | 561 | unsigned int name; |
diff --git a/arch/x86/kernel/acpi/boot.c b/arch/x86/kernel/acpi/boot.c index bfd10fd211cd..e5032d7b391d 100644 --- a/arch/x86/kernel/acpi/boot.c +++ b/arch/x86/kernel/acpi/boot.c | |||
| @@ -58,7 +58,6 @@ EXPORT_SYMBOL(acpi_disabled); | |||
| 58 | #ifdef CONFIG_X86_64 | 58 | #ifdef CONFIG_X86_64 |
| 59 | 59 | ||
| 60 | #include <asm/proto.h> | 60 | #include <asm/proto.h> |
| 61 | #include <asm/genapic.h> | ||
| 62 | 61 | ||
| 63 | #else /* X86 */ | 62 | #else /* X86 */ |
| 64 | 63 | ||
| @@ -97,8 +96,6 @@ static u64 acpi_lapic_addr __initdata = APIC_DEFAULT_PHYS_BASE; | |||
| 97 | #warning ACPI uses CMPXCHG, i486 and later hardware | 96 | #warning ACPI uses CMPXCHG, i486 and later hardware |
| 98 | #endif | 97 | #endif |
| 99 | 98 | ||
| 100 | static int acpi_mcfg_64bit_base_addr __initdata = FALSE; | ||
| 101 | |||
| 102 | /* -------------------------------------------------------------------------- | 99 | /* -------------------------------------------------------------------------- |
| 103 | Boot-time Configuration | 100 | Boot-time Configuration |
| 104 | -------------------------------------------------------------------------- */ | 101 | -------------------------------------------------------------------------- */ |
| @@ -160,6 +157,8 @@ char *__init __acpi_map_table(unsigned long phys, unsigned long size) | |||
| 160 | struct acpi_mcfg_allocation *pci_mmcfg_config; | 157 | struct acpi_mcfg_allocation *pci_mmcfg_config; |
| 161 | int pci_mmcfg_config_num; | 158 | int pci_mmcfg_config_num; |
| 162 | 159 | ||
| 160 | static int acpi_mcfg_64bit_base_addr __initdata = FALSE; | ||
| 161 | |||
| 163 | static int __init acpi_mcfg_oem_check(struct acpi_table_mcfg *mcfg) | 162 | static int __init acpi_mcfg_oem_check(struct acpi_table_mcfg *mcfg) |
| 164 | { | 163 | { |
| 165 | if (!strcmp(mcfg->header.oem_id, "SGI")) | 164 | if (!strcmp(mcfg->header.oem_id, "SGI")) |
diff --git a/arch/x86/kernel/alternative.c b/arch/x86/kernel/alternative.c index 65a0c1b48696..fb04e49776ba 100644 --- a/arch/x86/kernel/alternative.c +++ b/arch/x86/kernel/alternative.c | |||
| @@ -231,25 +231,25 @@ static void alternatives_smp_lock(u8 **start, u8 **end, u8 *text, u8 *text_end) | |||
| 231 | continue; | 231 | continue; |
| 232 | if (*ptr > text_end) | 232 | if (*ptr > text_end) |
| 233 | continue; | 233 | continue; |
| 234 | text_poke(*ptr, ((unsigned char []){0xf0}), 1); /* add lock prefix */ | 234 | /* turn DS segment override prefix into lock prefix */ |
| 235 | text_poke(*ptr, ((unsigned char []){0xf0}), 1); | ||
| 235 | }; | 236 | }; |
| 236 | } | 237 | } |
| 237 | 238 | ||
| 238 | static void alternatives_smp_unlock(u8 **start, u8 **end, u8 *text, u8 *text_end) | 239 | static void alternatives_smp_unlock(u8 **start, u8 **end, u8 *text, u8 *text_end) |
| 239 | { | 240 | { |
| 240 | u8 **ptr; | 241 | u8 **ptr; |
| 241 | char insn[1]; | ||
| 242 | 242 | ||
| 243 | if (noreplace_smp) | 243 | if (noreplace_smp) |
| 244 | return; | 244 | return; |
| 245 | 245 | ||
| 246 | add_nops(insn, 1); | ||
| 247 | for (ptr = start; ptr < end; ptr++) { | 246 | for (ptr = start; ptr < end; ptr++) { |
| 248 | if (*ptr < text) | 247 | if (*ptr < text) |
| 249 | continue; | 248 | continue; |
| 250 | if (*ptr > text_end) | 249 | if (*ptr > text_end) |
| 251 | continue; | 250 | continue; |
| 252 | text_poke(*ptr, insn, 1); | 251 | /* turn lock prefix into DS segment override prefix */ |
| 252 | text_poke(*ptr, ((unsigned char []){0x3E}), 1); | ||
| 253 | }; | 253 | }; |
| 254 | } | 254 | } |
| 255 | 255 | ||
diff --git a/arch/x86/kernel/aperture_64.c b/arch/x86/kernel/aperture_64.c index 44e21826db11..9a32b37ee2ee 100644 --- a/arch/x86/kernel/aperture_64.c +++ b/arch/x86/kernel/aperture_64.c | |||
| @@ -455,11 +455,11 @@ out: | |||
| 455 | force_iommu || | 455 | force_iommu || |
| 456 | valid_agp || | 456 | valid_agp || |
| 457 | fallback_aper_force) { | 457 | fallback_aper_force) { |
| 458 | printk(KERN_ERR | 458 | printk(KERN_INFO |
| 459 | "Your BIOS doesn't leave a aperture memory hole\n"); | 459 | "Your BIOS doesn't leave a aperture memory hole\n"); |
| 460 | printk(KERN_ERR | 460 | printk(KERN_INFO |
| 461 | "Please enable the IOMMU option in the BIOS setup\n"); | 461 | "Please enable the IOMMU option in the BIOS setup\n"); |
| 462 | printk(KERN_ERR | 462 | printk(KERN_INFO |
| 463 | "This costs you %d MB of RAM\n", | 463 | "This costs you %d MB of RAM\n", |
| 464 | 32 << fallback_aper_order); | 464 | 32 << fallback_aper_order); |
| 465 | 465 | ||
diff --git a/arch/x86/kernel/apm_32.c b/arch/x86/kernel/apm_32.c index 732d1f4e10ee..5145a6e72bbb 100644 --- a/arch/x86/kernel/apm_32.c +++ b/arch/x86/kernel/apm_32.c | |||
| @@ -228,7 +228,6 @@ | |||
| 228 | #include <linux/suspend.h> | 228 | #include <linux/suspend.h> |
| 229 | #include <linux/kthread.h> | 229 | #include <linux/kthread.h> |
| 230 | #include <linux/jiffies.h> | 230 | #include <linux/jiffies.h> |
| 231 | #include <linux/smp_lock.h> | ||
| 232 | 231 | ||
| 233 | #include <asm/system.h> | 232 | #include <asm/system.h> |
| 234 | #include <asm/uaccess.h> | 233 | #include <asm/uaccess.h> |
diff --git a/arch/x86/kernel/bios_uv.c b/arch/x86/kernel/bios_uv.c index c639bd55391c..fdd585f9c53d 100644 --- a/arch/x86/kernel/bios_uv.c +++ b/arch/x86/kernel/bios_uv.c | |||
| @@ -25,11 +25,11 @@ x86_bios_strerror(long status) | |||
| 25 | { | 25 | { |
| 26 | const char *str; | 26 | const char *str; |
| 27 | switch (status) { | 27 | switch (status) { |
| 28 | case 0: str = "Call completed without error"; break; | 28 | case 0: str = "Call completed without error"; break; |
| 29 | case -1: str = "Not implemented"; break; | 29 | case -1: str = "Not implemented"; break; |
| 30 | case -2: str = "Invalid argument"; break; | 30 | case -2: str = "Invalid argument"; break; |
| 31 | case -3: str = "Call completed with error"; break; | 31 | case -3: str = "Call completed with error"; break; |
| 32 | default: str = "Unknown BIOS status code"; break; | 32 | default: str = "Unknown BIOS status code"; break; |
| 33 | } | 33 | } |
| 34 | return str; | 34 | return str; |
| 35 | } | 35 | } |
diff --git a/arch/x86/kernel/cpu/common_64.c b/arch/x86/kernel/cpu/common_64.c index a11f5d4477cd..305b465889b0 100644 --- a/arch/x86/kernel/cpu/common_64.c +++ b/arch/x86/kernel/cpu/common_64.c | |||
| @@ -430,6 +430,49 @@ static __init int setup_noclflush(char *arg) | |||
| 430 | } | 430 | } |
| 431 | __setup("noclflush", setup_noclflush); | 431 | __setup("noclflush", setup_noclflush); |
| 432 | 432 | ||
| 433 | struct msr_range { | ||
| 434 | unsigned min; | ||
| 435 | unsigned max; | ||
| 436 | }; | ||
| 437 | |||
| 438 | static struct msr_range msr_range_array[] __cpuinitdata = { | ||
| 439 | { 0x00000000, 0x00000418}, | ||
| 440 | { 0xc0000000, 0xc000040b}, | ||
| 441 | { 0xc0010000, 0xc0010142}, | ||
| 442 | { 0xc0011000, 0xc001103b}, | ||
| 443 | }; | ||
| 444 | |||
| 445 | static void __cpuinit print_cpu_msr(void) | ||
| 446 | { | ||
| 447 | unsigned index; | ||
| 448 | u64 val; | ||
| 449 | int i; | ||
| 450 | unsigned index_min, index_max; | ||
| 451 | |||
| 452 | for (i = 0; i < ARRAY_SIZE(msr_range_array); i++) { | ||
| 453 | index_min = msr_range_array[i].min; | ||
| 454 | index_max = msr_range_array[i].max; | ||
| 455 | for (index = index_min; index < index_max; index++) { | ||
| 456 | if (rdmsrl_amd_safe(index, &val)) | ||
| 457 | continue; | ||
| 458 | printk(KERN_INFO " MSR%08x: %016llx\n", index, val); | ||
| 459 | } | ||
| 460 | } | ||
| 461 | } | ||
| 462 | |||
| 463 | static int show_msr __cpuinitdata; | ||
| 464 | static __init int setup_show_msr(char *arg) | ||
| 465 | { | ||
| 466 | int num; | ||
| 467 | |||
| 468 | get_option(&arg, &num); | ||
| 469 | |||
| 470 | if (num > 0) | ||
| 471 | show_msr = num; | ||
| 472 | return 1; | ||
| 473 | } | ||
| 474 | __setup("show_msr=", setup_show_msr); | ||
| 475 | |||
| 433 | void __cpuinit print_cpu_info(struct cpuinfo_x86 *c) | 476 | void __cpuinit print_cpu_info(struct cpuinfo_x86 *c) |
| 434 | { | 477 | { |
| 435 | if (c->x86_model_id[0]) | 478 | if (c->x86_model_id[0]) |
| @@ -439,6 +482,14 @@ void __cpuinit print_cpu_info(struct cpuinfo_x86 *c) | |||
| 439 | printk(KERN_CONT " stepping %02x\n", c->x86_mask); | 482 | printk(KERN_CONT " stepping %02x\n", c->x86_mask); |
| 440 | else | 483 | else |
| 441 | printk(KERN_CONT "\n"); | 484 | printk(KERN_CONT "\n"); |
| 485 | |||
| 486 | #ifdef CONFIG_SMP | ||
| 487 | if (c->cpu_index < show_msr) | ||
| 488 | print_cpu_msr(); | ||
| 489 | #else | ||
| 490 | if (show_msr) | ||
| 491 | print_cpu_msr(); | ||
| 492 | #endif | ||
| 442 | } | 493 | } |
| 443 | 494 | ||
| 444 | static __init int setup_disablecpuid(char *arg) | 495 | static __init int setup_disablecpuid(char *arg) |
diff --git a/arch/x86/kernel/cpu/mtrr/generic.c b/arch/x86/kernel/cpu/mtrr/generic.c index cb7d3b6a80eb..4e8d77f01eeb 100644 --- a/arch/x86/kernel/cpu/mtrr/generic.c +++ b/arch/x86/kernel/cpu/mtrr/generic.c | |||
| @@ -401,12 +401,7 @@ static void generic_get_mtrr(unsigned int reg, unsigned long *base, | |||
| 401 | tmp |= ~((1<<(hi - 1)) - 1); | 401 | tmp |= ~((1<<(hi - 1)) - 1); |
| 402 | 402 | ||
| 403 | if (tmp != mask_lo) { | 403 | if (tmp != mask_lo) { |
| 404 | static int once = 1; | 404 | WARN_ONCE(1, KERN_INFO "mtrr: your BIOS has set up an incorrect mask, fixing it up.\n"); |
| 405 | |||
| 406 | if (once) { | ||
| 407 | printk(KERN_INFO "mtrr: your BIOS has set up an incorrect mask, fixing it up.\n"); | ||
| 408 | once = 0; | ||
| 409 | } | ||
| 410 | mask_lo = tmp; | 405 | mask_lo = tmp; |
| 411 | } | 406 | } |
| 412 | } | 407 | } |
diff --git a/arch/x86/kernel/cpu/mtrr/if.c b/arch/x86/kernel/cpu/mtrr/if.c index 84c480bb3715..4c4214690dd1 100644 --- a/arch/x86/kernel/cpu/mtrr/if.c +++ b/arch/x86/kernel/cpu/mtrr/if.c | |||
| @@ -405,9 +405,9 @@ static int mtrr_seq_show(struct seq_file *seq, void *offset) | |||
| 405 | } | 405 | } |
| 406 | /* RED-PEN: base can be > 32bit */ | 406 | /* RED-PEN: base can be > 32bit */ |
| 407 | len += seq_printf(seq, | 407 | len += seq_printf(seq, |
| 408 | "reg%02i: base=0x%05lx000 (%4luMB), size=%4lu%cB: %s, count=%d\n", | 408 | "reg%02i: base=0x%06lx000 (%5luMB), size=%5lu%cB, count=%d: %s\n", |
| 409 | i, base, base >> (20 - PAGE_SHIFT), size, factor, | 409 | i, base, base >> (20 - PAGE_SHIFT), size, factor, |
| 410 | mtrr_attrib_to_str(type), mtrr_usage_table[i]); | 410 | mtrr_usage_table[i], mtrr_attrib_to_str(type)); |
| 411 | } | 411 | } |
| 412 | } | 412 | } |
| 413 | return 0; | 413 | return 0; |
diff --git a/arch/x86/kernel/cpu/mtrr/main.c b/arch/x86/kernel/cpu/mtrr/main.c index 5df16d818371..c78c04821ea1 100644 --- a/arch/x86/kernel/cpu/mtrr/main.c +++ b/arch/x86/kernel/cpu/mtrr/main.c | |||
| @@ -759,7 +759,8 @@ x86_get_mtrr_mem_range(struct res_range *range, int nr_range, | |||
| 759 | /* take out UC ranges */ | 759 | /* take out UC ranges */ |
| 760 | for (i = 0; i < num_var_ranges; i++) { | 760 | for (i = 0; i < num_var_ranges; i++) { |
| 761 | type = range_state[i].type; | 761 | type = range_state[i].type; |
| 762 | if (type != MTRR_TYPE_UNCACHABLE) | 762 | if (type != MTRR_TYPE_UNCACHABLE && |
| 763 | type != MTRR_TYPE_WRPROT) | ||
| 763 | continue; | 764 | continue; |
| 764 | size = range_state[i].size_pfn; | 765 | size = range_state[i].size_pfn; |
| 765 | if (!size) | 766 | if (!size) |
| @@ -836,6 +837,13 @@ static int __init enable_mtrr_cleanup_setup(char *str) | |||
| 836 | } | 837 | } |
| 837 | early_param("enable_mtrr_cleanup", enable_mtrr_cleanup_setup); | 838 | early_param("enable_mtrr_cleanup", enable_mtrr_cleanup_setup); |
| 838 | 839 | ||
| 840 | static int __init mtrr_cleanup_debug_setup(char *str) | ||
| 841 | { | ||
| 842 | debug_print = 1; | ||
| 843 | return 0; | ||
| 844 | } | ||
| 845 | early_param("mtrr_cleanup_debug", mtrr_cleanup_debug_setup); | ||
| 846 | |||
| 839 | struct var_mtrr_state { | 847 | struct var_mtrr_state { |
| 840 | unsigned long range_startk; | 848 | unsigned long range_startk; |
| 841 | unsigned long range_sizek; | 849 | unsigned long range_sizek; |
| @@ -898,6 +906,27 @@ set_var_mtrr_all(unsigned int address_bits) | |||
| 898 | } | 906 | } |
| 899 | } | 907 | } |
| 900 | 908 | ||
| 909 | static unsigned long to_size_factor(unsigned long sizek, char *factorp) | ||
| 910 | { | ||
| 911 | char factor; | ||
| 912 | unsigned long base = sizek; | ||
| 913 | |||
| 914 | if (base & ((1<<10) - 1)) { | ||
| 915 | /* not MB alignment */ | ||
| 916 | factor = 'K'; | ||
| 917 | } else if (base & ((1<<20) - 1)){ | ||
| 918 | factor = 'M'; | ||
| 919 | base >>= 10; | ||
| 920 | } else { | ||
| 921 | factor = 'G'; | ||
| 922 | base >>= 20; | ||
| 923 | } | ||
| 924 | |||
| 925 | *factorp = factor; | ||
| 926 | |||
| 927 | return base; | ||
| 928 | } | ||
| 929 | |||
| 901 | static unsigned int __init | 930 | static unsigned int __init |
| 902 | range_to_mtrr(unsigned int reg, unsigned long range_startk, | 931 | range_to_mtrr(unsigned int reg, unsigned long range_startk, |
| 903 | unsigned long range_sizek, unsigned char type) | 932 | unsigned long range_sizek, unsigned char type) |
| @@ -919,13 +948,21 @@ range_to_mtrr(unsigned int reg, unsigned long range_startk, | |||
| 919 | align = max_align; | 948 | align = max_align; |
| 920 | 949 | ||
| 921 | sizek = 1 << align; | 950 | sizek = 1 << align; |
| 922 | if (debug_print) | 951 | if (debug_print) { |
| 952 | char start_factor = 'K', size_factor = 'K'; | ||
| 953 | unsigned long start_base, size_base; | ||
| 954 | |||
| 955 | start_base = to_size_factor(range_startk, &start_factor), | ||
| 956 | size_base = to_size_factor(sizek, &size_factor), | ||
| 957 | |||
| 923 | printk(KERN_DEBUG "Setting variable MTRR %d, " | 958 | printk(KERN_DEBUG "Setting variable MTRR %d, " |
| 924 | "base: %ldMB, range: %ldMB, type %s\n", | 959 | "base: %ld%cB, range: %ld%cB, type %s\n", |
| 925 | reg, range_startk >> 10, sizek >> 10, | 960 | reg, start_base, start_factor, |
| 961 | size_base, size_factor, | ||
| 926 | (type == MTRR_TYPE_UNCACHABLE)?"UC": | 962 | (type == MTRR_TYPE_UNCACHABLE)?"UC": |
| 927 | ((type == MTRR_TYPE_WRBACK)?"WB":"Other") | 963 | ((type == MTRR_TYPE_WRBACK)?"WB":"Other") |
| 928 | ); | 964 | ); |
| 965 | } | ||
| 929 | save_var_mtrr(reg++, range_startk, sizek, type); | 966 | save_var_mtrr(reg++, range_startk, sizek, type); |
| 930 | range_startk += sizek; | 967 | range_startk += sizek; |
| 931 | range_sizek -= sizek; | 968 | range_sizek -= sizek; |
| @@ -970,6 +1007,8 @@ range_to_mtrr_with_hole(struct var_mtrr_state *state, unsigned long basek, | |||
| 970 | /* try to append some small hole */ | 1007 | /* try to append some small hole */ |
| 971 | range0_basek = state->range_startk; | 1008 | range0_basek = state->range_startk; |
| 972 | range0_sizek = ALIGN(state->range_sizek, chunk_sizek); | 1009 | range0_sizek = ALIGN(state->range_sizek, chunk_sizek); |
| 1010 | |||
| 1011 | /* no increase */ | ||
| 973 | if (range0_sizek == state->range_sizek) { | 1012 | if (range0_sizek == state->range_sizek) { |
| 974 | if (debug_print) | 1013 | if (debug_print) |
| 975 | printk(KERN_DEBUG "rangeX: %016lx - %016lx\n", | 1014 | printk(KERN_DEBUG "rangeX: %016lx - %016lx\n", |
| @@ -980,13 +1019,40 @@ range_to_mtrr_with_hole(struct var_mtrr_state *state, unsigned long basek, | |||
| 980 | return 0; | 1019 | return 0; |
| 981 | } | 1020 | } |
| 982 | 1021 | ||
| 983 | range0_sizek -= chunk_sizek; | 1022 | /* only cut back, when it is not the last */ |
| 984 | if (range0_sizek && sizek) { | 1023 | if (sizek) { |
| 985 | while (range0_basek + range0_sizek > (basek + sizek)) { | 1024 | while (range0_basek + range0_sizek > (basek + sizek)) { |
| 986 | range0_sizek -= chunk_sizek; | 1025 | if (range0_sizek >= chunk_sizek) |
| 987 | if (!range0_sizek) | 1026 | range0_sizek -= chunk_sizek; |
| 988 | break; | 1027 | else |
| 989 | } | 1028 | range0_sizek = 0; |
| 1029 | |||
| 1030 | if (!range0_sizek) | ||
| 1031 | break; | ||
| 1032 | } | ||
| 1033 | } | ||
| 1034 | |||
| 1035 | second_try: | ||
| 1036 | range_basek = range0_basek + range0_sizek; | ||
| 1037 | |||
| 1038 | /* one hole in the middle */ | ||
| 1039 | if (range_basek > basek && range_basek <= (basek + sizek)) | ||
| 1040 | second_sizek = range_basek - basek; | ||
| 1041 | |||
| 1042 | if (range0_sizek > state->range_sizek) { | ||
| 1043 | |||
| 1044 | /* one hole in middle or at end */ | ||
| 1045 | hole_sizek = range0_sizek - state->range_sizek - second_sizek; | ||
| 1046 | |||
| 1047 | /* hole size should be less than half of range0 size */ | ||
| 1048 | if (hole_sizek >= (range0_sizek >> 1) && | ||
| 1049 | range0_sizek >= chunk_sizek) { | ||
| 1050 | range0_sizek -= chunk_sizek; | ||
| 1051 | second_sizek = 0; | ||
| 1052 | hole_sizek = 0; | ||
| 1053 | |||
| 1054 | goto second_try; | ||
| 1055 | } | ||
| 990 | } | 1056 | } |
| 991 | 1057 | ||
| 992 | if (range0_sizek) { | 1058 | if (range0_sizek) { |
| @@ -996,50 +1062,28 @@ range_to_mtrr_with_hole(struct var_mtrr_state *state, unsigned long basek, | |||
| 996 | (range0_basek + range0_sizek)<<10); | 1062 | (range0_basek + range0_sizek)<<10); |
| 997 | state->reg = range_to_mtrr(state->reg, range0_basek, | 1063 | state->reg = range_to_mtrr(state->reg, range0_basek, |
| 998 | range0_sizek, MTRR_TYPE_WRBACK); | 1064 | range0_sizek, MTRR_TYPE_WRBACK); |
| 999 | |||
| 1000 | } | ||
| 1001 | |||
| 1002 | range_basek = range0_basek + range0_sizek; | ||
| 1003 | range_sizek = chunk_sizek; | ||
| 1004 | |||
| 1005 | if (range_basek + range_sizek > basek && | ||
| 1006 | range_basek + range_sizek <= (basek + sizek)) { | ||
| 1007 | /* one hole */ | ||
| 1008 | second_basek = basek; | ||
| 1009 | second_sizek = range_basek + range_sizek - basek; | ||
| 1010 | } | 1065 | } |
| 1011 | 1066 | ||
| 1012 | /* if last piece, only could one hole near end */ | 1067 | if (range0_sizek < state->range_sizek) { |
| 1013 | if ((second_basek || !basek) && | 1068 | /* need to handle left over */ |
| 1014 | range_sizek - (state->range_sizek - range0_sizek) - second_sizek < | ||
| 1015 | (chunk_sizek >> 1)) { | ||
| 1016 | /* | ||
| 1017 | * one hole in middle (second_sizek is 0) or at end | ||
| 1018 | * (second_sizek is 0 ) | ||
| 1019 | */ | ||
| 1020 | hole_sizek = range_sizek - (state->range_sizek - range0_sizek) | ||
| 1021 | - second_sizek; | ||
| 1022 | hole_basek = range_basek + range_sizek - hole_sizek | ||
| 1023 | - second_sizek; | ||
| 1024 | } else { | ||
| 1025 | /* fallback for big hole, or several holes */ | ||
| 1026 | range_sizek = state->range_sizek - range0_sizek; | 1069 | range_sizek = state->range_sizek - range0_sizek; |
| 1027 | second_basek = 0; | 1070 | |
| 1028 | second_sizek = 0; | 1071 | if (debug_print) |
| 1072 | printk(KERN_DEBUG "range: %016lx - %016lx\n", | ||
| 1073 | range_basek<<10, | ||
| 1074 | (range_basek + range_sizek)<<10); | ||
| 1075 | state->reg = range_to_mtrr(state->reg, range_basek, | ||
| 1076 | range_sizek, MTRR_TYPE_WRBACK); | ||
| 1029 | } | 1077 | } |
| 1030 | 1078 | ||
| 1031 | if (debug_print) | ||
| 1032 | printk(KERN_DEBUG "range: %016lx - %016lx\n", range_basek<<10, | ||
| 1033 | (range_basek + range_sizek)<<10); | ||
| 1034 | state->reg = range_to_mtrr(state->reg, range_basek, range_sizek, | ||
| 1035 | MTRR_TYPE_WRBACK); | ||
| 1036 | if (hole_sizek) { | 1079 | if (hole_sizek) { |
| 1080 | hole_basek = range_basek - hole_sizek - second_sizek; | ||
| 1037 | if (debug_print) | 1081 | if (debug_print) |
| 1038 | printk(KERN_DEBUG "hole: %016lx - %016lx\n", | 1082 | printk(KERN_DEBUG "hole: %016lx - %016lx\n", |
| 1039 | hole_basek<<10, (hole_basek + hole_sizek)<<10); | 1083 | hole_basek<<10, |
| 1040 | state->reg = range_to_mtrr(state->reg, hole_basek, hole_sizek, | 1084 | (hole_basek + hole_sizek)<<10); |
| 1041 | MTRR_TYPE_UNCACHABLE); | 1085 | state->reg = range_to_mtrr(state->reg, hole_basek, |
| 1042 | 1086 | hole_sizek, MTRR_TYPE_UNCACHABLE); | |
| 1043 | } | 1087 | } |
| 1044 | 1088 | ||
| 1045 | return second_sizek; | 1089 | return second_sizek; |
| @@ -1154,11 +1198,11 @@ struct mtrr_cleanup_result { | |||
| 1154 | }; | 1198 | }; |
| 1155 | 1199 | ||
| 1156 | /* | 1200 | /* |
| 1157 | * gran_size: 1M, 2M, ..., 2G | 1201 | * gran_size: 64K, 128K, 256K, 512K, 1M, 2M, ..., 2G |
| 1158 | * chunk size: gran_size, ..., 4G | 1202 | * chunk size: gran_size, ..., 2G |
| 1159 | * so we need (2+13)*6 | 1203 | * so we need (1+16)*8 |
| 1160 | */ | 1204 | */ |
| 1161 | #define NUM_RESULT 90 | 1205 | #define NUM_RESULT 136 |
| 1162 | #define PSHIFT (PAGE_SHIFT - 10) | 1206 | #define PSHIFT (PAGE_SHIFT - 10) |
| 1163 | 1207 | ||
| 1164 | static struct mtrr_cleanup_result __initdata result[NUM_RESULT]; | 1208 | static struct mtrr_cleanup_result __initdata result[NUM_RESULT]; |
| @@ -1168,13 +1212,14 @@ static unsigned long __initdata min_loss_pfn[RANGE_NUM]; | |||
| 1168 | static int __init mtrr_cleanup(unsigned address_bits) | 1212 | static int __init mtrr_cleanup(unsigned address_bits) |
| 1169 | { | 1213 | { |
| 1170 | unsigned long extra_remove_base, extra_remove_size; | 1214 | unsigned long extra_remove_base, extra_remove_size; |
| 1171 | unsigned long i, base, size, def, dummy; | 1215 | unsigned long base, size, def, dummy; |
| 1172 | mtrr_type type; | 1216 | mtrr_type type; |
| 1173 | int nr_range, nr_range_new; | 1217 | int nr_range, nr_range_new; |
| 1174 | u64 chunk_size, gran_size; | 1218 | u64 chunk_size, gran_size; |
| 1175 | unsigned long range_sums, range_sums_new; | 1219 | unsigned long range_sums, range_sums_new; |
| 1176 | int index_good; | 1220 | int index_good; |
| 1177 | int num_reg_good; | 1221 | int num_reg_good; |
| 1222 | int i; | ||
| 1178 | 1223 | ||
| 1179 | /* extra one for all 0 */ | 1224 | /* extra one for all 0 */ |
| 1180 | int num[MTRR_NUM_TYPES + 1]; | 1225 | int num[MTRR_NUM_TYPES + 1]; |
| @@ -1204,6 +1249,8 @@ static int __init mtrr_cleanup(unsigned address_bits) | |||
| 1204 | continue; | 1249 | continue; |
| 1205 | if (!size) | 1250 | if (!size) |
| 1206 | type = MTRR_NUM_TYPES; | 1251 | type = MTRR_NUM_TYPES; |
| 1252 | if (type == MTRR_TYPE_WRPROT) | ||
| 1253 | type = MTRR_TYPE_UNCACHABLE; | ||
| 1207 | num[type]++; | 1254 | num[type]++; |
| 1208 | } | 1255 | } |
| 1209 | 1256 | ||
| @@ -1216,23 +1263,57 @@ static int __init mtrr_cleanup(unsigned address_bits) | |||
| 1216 | num_var_ranges - num[MTRR_NUM_TYPES]) | 1263 | num_var_ranges - num[MTRR_NUM_TYPES]) |
| 1217 | return 0; | 1264 | return 0; |
| 1218 | 1265 | ||
| 1266 | /* print original var MTRRs at first, for debugging: */ | ||
| 1267 | printk(KERN_DEBUG "original variable MTRRs\n"); | ||
| 1268 | for (i = 0; i < num_var_ranges; i++) { | ||
| 1269 | char start_factor = 'K', size_factor = 'K'; | ||
| 1270 | unsigned long start_base, size_base; | ||
| 1271 | |||
| 1272 | size_base = range_state[i].size_pfn << (PAGE_SHIFT - 10); | ||
| 1273 | if (!size_base) | ||
| 1274 | continue; | ||
| 1275 | |||
| 1276 | size_base = to_size_factor(size_base, &size_factor), | ||
| 1277 | start_base = range_state[i].base_pfn << (PAGE_SHIFT - 10); | ||
| 1278 | start_base = to_size_factor(start_base, &start_factor), | ||
| 1279 | type = range_state[i].type; | ||
| 1280 | |||
| 1281 | printk(KERN_DEBUG "reg %d, base: %ld%cB, range: %ld%cB, type %s\n", | ||
| 1282 | i, start_base, start_factor, | ||
| 1283 | size_base, size_factor, | ||
| 1284 | (type == MTRR_TYPE_UNCACHABLE) ? "UC" : | ||
| 1285 | ((type == MTRR_TYPE_WRPROT) ? "WP" : | ||
| 1286 | ((type == MTRR_TYPE_WRBACK) ? "WB" : "Other")) | ||
| 1287 | ); | ||
| 1288 | } | ||
| 1289 | |||
| 1219 | memset(range, 0, sizeof(range)); | 1290 | memset(range, 0, sizeof(range)); |
| 1220 | extra_remove_size = 0; | 1291 | extra_remove_size = 0; |
| 1221 | if (mtrr_tom2) { | 1292 | extra_remove_base = 1 << (32 - PAGE_SHIFT); |
| 1222 | extra_remove_base = 1 << (32 - PAGE_SHIFT); | 1293 | if (mtrr_tom2) |
| 1223 | extra_remove_size = | 1294 | extra_remove_size = |
| 1224 | (mtrr_tom2 >> PAGE_SHIFT) - extra_remove_base; | 1295 | (mtrr_tom2 >> PAGE_SHIFT) - extra_remove_base; |
| 1225 | } | ||
| 1226 | nr_range = x86_get_mtrr_mem_range(range, 0, extra_remove_base, | 1296 | nr_range = x86_get_mtrr_mem_range(range, 0, extra_remove_base, |
| 1227 | extra_remove_size); | 1297 | extra_remove_size); |
| 1298 | /* | ||
| 1299 | * [0, 1M) should always be coverred by var mtrr with WB | ||
| 1300 | * and fixed mtrrs should take effective before var mtrr for it | ||
| 1301 | */ | ||
| 1302 | nr_range = add_range_with_merge(range, nr_range, 0, | ||
| 1303 | (1ULL<<(20 - PAGE_SHIFT)) - 1); | ||
| 1304 | /* sort the ranges */ | ||
| 1305 | sort(range, nr_range, sizeof(struct res_range), cmp_range, NULL); | ||
| 1306 | |||
| 1228 | range_sums = sum_ranges(range, nr_range); | 1307 | range_sums = sum_ranges(range, nr_range); |
| 1229 | printk(KERN_INFO "total RAM coverred: %ldM\n", | 1308 | printk(KERN_INFO "total RAM coverred: %ldM\n", |
| 1230 | range_sums >> (20 - PAGE_SHIFT)); | 1309 | range_sums >> (20 - PAGE_SHIFT)); |
| 1231 | 1310 | ||
| 1232 | if (mtrr_chunk_size && mtrr_gran_size) { | 1311 | if (mtrr_chunk_size && mtrr_gran_size) { |
| 1233 | int num_reg; | 1312 | int num_reg; |
| 1313 | char gran_factor, chunk_factor, lose_factor; | ||
| 1314 | unsigned long gran_base, chunk_base, lose_base; | ||
| 1234 | 1315 | ||
| 1235 | debug_print = 1; | 1316 | debug_print++; |
| 1236 | /* convert ranges to var ranges state */ | 1317 | /* convert ranges to var ranges state */ |
| 1237 | num_reg = x86_setup_var_mtrrs(range, nr_range, mtrr_chunk_size, | 1318 | num_reg = x86_setup_var_mtrrs(range, nr_range, mtrr_chunk_size, |
| 1238 | mtrr_gran_size); | 1319 | mtrr_gran_size); |
| @@ -1256,34 +1337,48 @@ static int __init mtrr_cleanup(unsigned address_bits) | |||
| 1256 | result[i].lose_cover_sizek = | 1337 | result[i].lose_cover_sizek = |
| 1257 | (range_sums - range_sums_new) << PSHIFT; | 1338 | (range_sums - range_sums_new) << PSHIFT; |
| 1258 | 1339 | ||
| 1259 | printk(KERN_INFO "%sgran_size: %ldM \tchunk_size: %ldM \t", | 1340 | gran_base = to_size_factor(result[i].gran_sizek, &gran_factor), |
| 1260 | result[i].bad?"*BAD*":" ", result[i].gran_sizek >> 10, | 1341 | chunk_base = to_size_factor(result[i].chunk_sizek, &chunk_factor), |
| 1261 | result[i].chunk_sizek >> 10); | 1342 | lose_base = to_size_factor(result[i].lose_cover_sizek, &lose_factor), |
| 1262 | printk(KERN_CONT "num_reg: %d \tlose cover RAM: %s%ldM \n", | 1343 | printk(KERN_INFO "%sgran_size: %ld%c \tchunk_size: %ld%c \t", |
| 1344 | result[i].bad?"*BAD*":" ", | ||
| 1345 | gran_base, gran_factor, chunk_base, chunk_factor); | ||
| 1346 | printk(KERN_CONT "num_reg: %d \tlose cover RAM: %s%ld%c\n", | ||
| 1263 | result[i].num_reg, result[i].bad?"-":"", | 1347 | result[i].num_reg, result[i].bad?"-":"", |
| 1264 | result[i].lose_cover_sizek >> 10); | 1348 | lose_base, lose_factor); |
| 1265 | if (!result[i].bad) { | 1349 | if (!result[i].bad) { |
| 1266 | set_var_mtrr_all(address_bits); | 1350 | set_var_mtrr_all(address_bits); |
| 1267 | return 1; | 1351 | return 1; |
| 1268 | } | 1352 | } |
| 1269 | printk(KERN_INFO "invalid mtrr_gran_size or mtrr_chunk_size, " | 1353 | printk(KERN_INFO "invalid mtrr_gran_size or mtrr_chunk_size, " |
| 1270 | "will find optimal one\n"); | 1354 | "will find optimal one\n"); |
| 1271 | debug_print = 0; | 1355 | debug_print--; |
| 1272 | memset(result, 0, sizeof(result[0])); | 1356 | memset(result, 0, sizeof(result[0])); |
| 1273 | } | 1357 | } |
| 1274 | 1358 | ||
| 1275 | i = 0; | 1359 | i = 0; |
| 1276 | memset(min_loss_pfn, 0xff, sizeof(min_loss_pfn)); | 1360 | memset(min_loss_pfn, 0xff, sizeof(min_loss_pfn)); |
| 1277 | memset(result, 0, sizeof(result)); | 1361 | memset(result, 0, sizeof(result)); |
| 1278 | for (gran_size = (1ULL<<20); gran_size < (1ULL<<32); gran_size <<= 1) { | 1362 | for (gran_size = (1ULL<<16); gran_size < (1ULL<<32); gran_size <<= 1) { |
| 1279 | for (chunk_size = gran_size; chunk_size < (1ULL<<33); | 1363 | char gran_factor; |
| 1364 | unsigned long gran_base; | ||
| 1365 | |||
| 1366 | if (debug_print) | ||
| 1367 | gran_base = to_size_factor(gran_size >> 10, &gran_factor); | ||
| 1368 | |||
| 1369 | for (chunk_size = gran_size; chunk_size < (1ULL<<32); | ||
| 1280 | chunk_size <<= 1) { | 1370 | chunk_size <<= 1) { |
| 1281 | int num_reg; | 1371 | int num_reg; |
| 1282 | 1372 | ||
| 1283 | if (debug_print) | 1373 | if (debug_print) { |
| 1284 | printk(KERN_INFO | 1374 | char chunk_factor; |
| 1285 | "\ngran_size: %lldM chunk_size_size: %lldM\n", | 1375 | unsigned long chunk_base; |
| 1286 | gran_size >> 20, chunk_size >> 20); | 1376 | |
| 1377 | chunk_base = to_size_factor(chunk_size>>10, &chunk_factor), | ||
| 1378 | printk(KERN_INFO "\n"); | ||
| 1379 | printk(KERN_INFO "gran_size: %ld%c chunk_size: %ld%c \n", | ||
| 1380 | gran_base, gran_factor, chunk_base, chunk_factor); | ||
| 1381 | } | ||
| 1287 | if (i >= NUM_RESULT) | 1382 | if (i >= NUM_RESULT) |
| 1288 | continue; | 1383 | continue; |
| 1289 | 1384 | ||
| @@ -1326,12 +1421,18 @@ static int __init mtrr_cleanup(unsigned address_bits) | |||
| 1326 | 1421 | ||
| 1327 | /* print out all */ | 1422 | /* print out all */ |
| 1328 | for (i = 0; i < NUM_RESULT; i++) { | 1423 | for (i = 0; i < NUM_RESULT; i++) { |
| 1329 | printk(KERN_INFO "%sgran_size: %ldM \tchunk_size: %ldM \t", | 1424 | char gran_factor, chunk_factor, lose_factor; |
| 1330 | result[i].bad?"*BAD* ":" ", result[i].gran_sizek >> 10, | 1425 | unsigned long gran_base, chunk_base, lose_base; |
| 1331 | result[i].chunk_sizek >> 10); | 1426 | |
| 1332 | printk(KERN_CONT "num_reg: %d \tlose RAM: %s%ldM\n", | 1427 | gran_base = to_size_factor(result[i].gran_sizek, &gran_factor), |
| 1333 | result[i].num_reg, result[i].bad?"-":"", | 1428 | chunk_base = to_size_factor(result[i].chunk_sizek, &chunk_factor), |
| 1334 | result[i].lose_cover_sizek >> 10); | 1429 | lose_base = to_size_factor(result[i].lose_cover_sizek, &lose_factor), |
| 1430 | printk(KERN_INFO "%sgran_size: %ld%c \tchunk_size: %ld%c \t", | ||
| 1431 | result[i].bad?"*BAD*":" ", | ||
| 1432 | gran_base, gran_factor, chunk_base, chunk_factor); | ||
| 1433 | printk(KERN_CONT "num_reg: %d \tlose cover RAM: %s%ld%c\n", | ||
| 1434 | result[i].num_reg, result[i].bad?"-":"", | ||
| 1435 | lose_base, lose_factor); | ||
| 1335 | } | 1436 | } |
| 1336 | 1437 | ||
| 1337 | /* try to find the optimal index */ | 1438 | /* try to find the optimal index */ |
| @@ -1339,10 +1440,8 @@ static int __init mtrr_cleanup(unsigned address_bits) | |||
| 1339 | nr_mtrr_spare_reg = num_var_ranges - 1; | 1440 | nr_mtrr_spare_reg = num_var_ranges - 1; |
| 1340 | num_reg_good = -1; | 1441 | num_reg_good = -1; |
| 1341 | for (i = num_var_ranges - nr_mtrr_spare_reg; i > 0; i--) { | 1442 | for (i = num_var_ranges - nr_mtrr_spare_reg; i > 0; i--) { |
| 1342 | if (!min_loss_pfn[i]) { | 1443 | if (!min_loss_pfn[i]) |
| 1343 | num_reg_good = i; | 1444 | num_reg_good = i; |
| 1344 | break; | ||
| 1345 | } | ||
| 1346 | } | 1445 | } |
| 1347 | 1446 | ||
| 1348 | index_good = -1; | 1447 | index_good = -1; |
| @@ -1358,21 +1457,26 @@ static int __init mtrr_cleanup(unsigned address_bits) | |||
| 1358 | } | 1457 | } |
| 1359 | 1458 | ||
| 1360 | if (index_good != -1) { | 1459 | if (index_good != -1) { |
| 1460 | char gran_factor, chunk_factor, lose_factor; | ||
| 1461 | unsigned long gran_base, chunk_base, lose_base; | ||
| 1462 | |||
| 1361 | printk(KERN_INFO "Found optimal setting for mtrr clean up\n"); | 1463 | printk(KERN_INFO "Found optimal setting for mtrr clean up\n"); |
| 1362 | i = index_good; | 1464 | i = index_good; |
| 1363 | printk(KERN_INFO "gran_size: %ldM \tchunk_size: %ldM \t", | 1465 | gran_base = to_size_factor(result[i].gran_sizek, &gran_factor), |
| 1364 | result[i].gran_sizek >> 10, | 1466 | chunk_base = to_size_factor(result[i].chunk_sizek, &chunk_factor), |
| 1365 | result[i].chunk_sizek >> 10); | 1467 | lose_base = to_size_factor(result[i].lose_cover_sizek, &lose_factor), |
| 1366 | printk(KERN_CONT "num_reg: %d \tlose RAM: %ldM\n", | 1468 | printk(KERN_INFO "gran_size: %ld%c \tchunk_size: %ld%c \t", |
| 1367 | result[i].num_reg, | 1469 | gran_base, gran_factor, chunk_base, chunk_factor); |
| 1368 | result[i].lose_cover_sizek >> 10); | 1470 | printk(KERN_CONT "num_reg: %d \tlose RAM: %ld%c\n", |
| 1471 | result[i].num_reg, lose_base, lose_factor); | ||
| 1369 | /* convert ranges to var ranges state */ | 1472 | /* convert ranges to var ranges state */ |
| 1370 | chunk_size = result[i].chunk_sizek; | 1473 | chunk_size = result[i].chunk_sizek; |
| 1371 | chunk_size <<= 10; | 1474 | chunk_size <<= 10; |
| 1372 | gran_size = result[i].gran_sizek; | 1475 | gran_size = result[i].gran_sizek; |
| 1373 | gran_size <<= 10; | 1476 | gran_size <<= 10; |
| 1374 | debug_print = 1; | 1477 | debug_print++; |
| 1375 | x86_setup_var_mtrrs(range, nr_range, chunk_size, gran_size); | 1478 | x86_setup_var_mtrrs(range, nr_range, chunk_size, gran_size); |
| 1479 | debug_print--; | ||
| 1376 | set_var_mtrr_all(address_bits); | 1480 | set_var_mtrr_all(address_bits); |
| 1377 | return 1; | 1481 | return 1; |
| 1378 | } | 1482 | } |
diff --git a/arch/x86/kernel/cpu/perfctr-watchdog.c b/arch/x86/kernel/cpu/perfctr-watchdog.c index 05cc22dbd4ff..6bff382094f5 100644 --- a/arch/x86/kernel/cpu/perfctr-watchdog.c +++ b/arch/x86/kernel/cpu/perfctr-watchdog.c | |||
| @@ -295,13 +295,19 @@ static int setup_k7_watchdog(unsigned nmi_hz) | |||
| 295 | /* setup the timer */ | 295 | /* setup the timer */ |
| 296 | wrmsr(evntsel_msr, evntsel, 0); | 296 | wrmsr(evntsel_msr, evntsel, 0); |
| 297 | write_watchdog_counter(perfctr_msr, "K7_PERFCTR0",nmi_hz); | 297 | write_watchdog_counter(perfctr_msr, "K7_PERFCTR0",nmi_hz); |
| 298 | apic_write(APIC_LVTPC, APIC_DM_NMI); | ||
| 299 | evntsel |= K7_EVNTSEL_ENABLE; | ||
| 300 | wrmsr(evntsel_msr, evntsel, 0); | ||
| 301 | 298 | ||
| 299 | /* initialize the wd struct before enabling */ | ||
| 302 | wd->perfctr_msr = perfctr_msr; | 300 | wd->perfctr_msr = perfctr_msr; |
| 303 | wd->evntsel_msr = evntsel_msr; | 301 | wd->evntsel_msr = evntsel_msr; |
| 304 | wd->cccr_msr = 0; /* unused */ | 302 | wd->cccr_msr = 0; /* unused */ |
| 303 | |||
| 304 | /* ok, everything is initialized, announce that we're set */ | ||
| 305 | cpu_nmi_set_wd_enabled(); | ||
| 306 | |||
| 307 | apic_write(APIC_LVTPC, APIC_DM_NMI); | ||
| 308 | evntsel |= K7_EVNTSEL_ENABLE; | ||
| 309 | wrmsr(evntsel_msr, evntsel, 0); | ||
| 310 | |||
| 305 | return 1; | 311 | return 1; |
| 306 | } | 312 | } |
| 307 | 313 | ||
| @@ -379,13 +385,19 @@ static int setup_p6_watchdog(unsigned nmi_hz) | |||
| 379 | wrmsr(evntsel_msr, evntsel, 0); | 385 | wrmsr(evntsel_msr, evntsel, 0); |
| 380 | nmi_hz = adjust_for_32bit_ctr(nmi_hz); | 386 | nmi_hz = adjust_for_32bit_ctr(nmi_hz); |
| 381 | write_watchdog_counter32(perfctr_msr, "P6_PERFCTR0",nmi_hz); | 387 | write_watchdog_counter32(perfctr_msr, "P6_PERFCTR0",nmi_hz); |
| 382 | apic_write(APIC_LVTPC, APIC_DM_NMI); | ||
| 383 | evntsel |= P6_EVNTSEL0_ENABLE; | ||
| 384 | wrmsr(evntsel_msr, evntsel, 0); | ||
| 385 | 388 | ||
| 389 | /* initialize the wd struct before enabling */ | ||
| 386 | wd->perfctr_msr = perfctr_msr; | 390 | wd->perfctr_msr = perfctr_msr; |
| 387 | wd->evntsel_msr = evntsel_msr; | 391 | wd->evntsel_msr = evntsel_msr; |
| 388 | wd->cccr_msr = 0; /* unused */ | 392 | wd->cccr_msr = 0; /* unused */ |
| 393 | |||
| 394 | /* ok, everything is initialized, announce that we're set */ | ||
| 395 | cpu_nmi_set_wd_enabled(); | ||
| 396 | |||
| 397 | apic_write(APIC_LVTPC, APIC_DM_NMI); | ||
| 398 | evntsel |= P6_EVNTSEL0_ENABLE; | ||
| 399 | wrmsr(evntsel_msr, evntsel, 0); | ||
| 400 | |||
| 389 | return 1; | 401 | return 1; |
| 390 | } | 402 | } |
| 391 | 403 | ||
| @@ -432,6 +444,27 @@ static const struct wd_ops p6_wd_ops = { | |||
| 432 | #define P4_CCCR_ENABLE (1 << 12) | 444 | #define P4_CCCR_ENABLE (1 << 12) |
| 433 | #define P4_CCCR_OVF (1 << 31) | 445 | #define P4_CCCR_OVF (1 << 31) |
| 434 | 446 | ||
| 447 | #define P4_CONTROLS 18 | ||
| 448 | static unsigned int p4_controls[18] = { | ||
| 449 | MSR_P4_BPU_CCCR0, | ||
| 450 | MSR_P4_BPU_CCCR1, | ||
| 451 | MSR_P4_BPU_CCCR2, | ||
| 452 | MSR_P4_BPU_CCCR3, | ||
| 453 | MSR_P4_MS_CCCR0, | ||
| 454 | MSR_P4_MS_CCCR1, | ||
| 455 | MSR_P4_MS_CCCR2, | ||
| 456 | MSR_P4_MS_CCCR3, | ||
| 457 | MSR_P4_FLAME_CCCR0, | ||
| 458 | MSR_P4_FLAME_CCCR1, | ||
| 459 | MSR_P4_FLAME_CCCR2, | ||
| 460 | MSR_P4_FLAME_CCCR3, | ||
| 461 | MSR_P4_IQ_CCCR0, | ||
| 462 | MSR_P4_IQ_CCCR1, | ||
| 463 | MSR_P4_IQ_CCCR2, | ||
| 464 | MSR_P4_IQ_CCCR3, | ||
| 465 | MSR_P4_IQ_CCCR4, | ||
| 466 | MSR_P4_IQ_CCCR5, | ||
| 467 | }; | ||
| 435 | /* | 468 | /* |
| 436 | * Set up IQ_COUNTER0 to behave like a clock, by having IQ_CCCR0 filter | 469 | * Set up IQ_COUNTER0 to behave like a clock, by having IQ_CCCR0 filter |
| 437 | * CRU_ESCR0 (with any non-null event selector) through a complemented | 470 | * CRU_ESCR0 (with any non-null event selector) through a complemented |
| @@ -473,6 +506,26 @@ static int setup_p4_watchdog(unsigned nmi_hz) | |||
| 473 | evntsel_msr = MSR_P4_CRU_ESCR0; | 506 | evntsel_msr = MSR_P4_CRU_ESCR0; |
| 474 | cccr_msr = MSR_P4_IQ_CCCR0; | 507 | cccr_msr = MSR_P4_IQ_CCCR0; |
| 475 | cccr_val = P4_CCCR_OVF_PMI0 | P4_CCCR_ESCR_SELECT(4); | 508 | cccr_val = P4_CCCR_OVF_PMI0 | P4_CCCR_ESCR_SELECT(4); |
| 509 | |||
| 510 | /* | ||
| 511 | * If we're on the kdump kernel or other situation, we may | ||
| 512 | * still have other performance counter registers set to | ||
| 513 | * interrupt and they'll keep interrupting forever because | ||
| 514 | * of the P4_CCCR_OVF quirk. So we need to ACK all the | ||
| 515 | * pending interrupts and disable all the registers here, | ||
| 516 | * before reenabling the NMI delivery. Refer to p4_rearm() | ||
| 517 | * about the P4_CCCR_OVF quirk. | ||
| 518 | */ | ||
| 519 | if (reset_devices) { | ||
| 520 | unsigned int low, high; | ||
| 521 | int i; | ||
| 522 | |||
| 523 | for (i = 0; i < P4_CONTROLS; i++) { | ||
| 524 | rdmsr(p4_controls[i], low, high); | ||
| 525 | low &= ~(P4_CCCR_ENABLE | P4_CCCR_OVF); | ||
| 526 | wrmsr(p4_controls[i], low, high); | ||
| 527 | } | ||
| 528 | } | ||
| 476 | } else { | 529 | } else { |
| 477 | /* logical cpu 1 */ | 530 | /* logical cpu 1 */ |
| 478 | perfctr_msr = MSR_P4_IQ_PERFCTR1; | 531 | perfctr_msr = MSR_P4_IQ_PERFCTR1; |
| @@ -499,12 +552,17 @@ static int setup_p4_watchdog(unsigned nmi_hz) | |||
| 499 | wrmsr(evntsel_msr, evntsel, 0); | 552 | wrmsr(evntsel_msr, evntsel, 0); |
| 500 | wrmsr(cccr_msr, cccr_val, 0); | 553 | wrmsr(cccr_msr, cccr_val, 0); |
| 501 | write_watchdog_counter(perfctr_msr, "P4_IQ_COUNTER0", nmi_hz); | 554 | write_watchdog_counter(perfctr_msr, "P4_IQ_COUNTER0", nmi_hz); |
| 502 | apic_write(APIC_LVTPC, APIC_DM_NMI); | 555 | |
| 503 | cccr_val |= P4_CCCR_ENABLE; | ||
| 504 | wrmsr(cccr_msr, cccr_val, 0); | ||
| 505 | wd->perfctr_msr = perfctr_msr; | 556 | wd->perfctr_msr = perfctr_msr; |
| 506 | wd->evntsel_msr = evntsel_msr; | 557 | wd->evntsel_msr = evntsel_msr; |
| 507 | wd->cccr_msr = cccr_msr; | 558 | wd->cccr_msr = cccr_msr; |
| 559 | |||
| 560 | /* ok, everything is initialized, announce that we're set */ | ||
| 561 | cpu_nmi_set_wd_enabled(); | ||
| 562 | |||
| 563 | apic_write(APIC_LVTPC, APIC_DM_NMI); | ||
| 564 | cccr_val |= P4_CCCR_ENABLE; | ||
| 565 | wrmsr(cccr_msr, cccr_val, 0); | ||
| 508 | return 1; | 566 | return 1; |
| 509 | } | 567 | } |
| 510 | 568 | ||
| @@ -620,13 +678,17 @@ static int setup_intel_arch_watchdog(unsigned nmi_hz) | |||
| 620 | wrmsr(evntsel_msr, evntsel, 0); | 678 | wrmsr(evntsel_msr, evntsel, 0); |
| 621 | nmi_hz = adjust_for_32bit_ctr(nmi_hz); | 679 | nmi_hz = adjust_for_32bit_ctr(nmi_hz); |
| 622 | write_watchdog_counter32(perfctr_msr, "INTEL_ARCH_PERFCTR0", nmi_hz); | 680 | write_watchdog_counter32(perfctr_msr, "INTEL_ARCH_PERFCTR0", nmi_hz); |
| 623 | apic_write(APIC_LVTPC, APIC_DM_NMI); | ||
| 624 | evntsel |= ARCH_PERFMON_EVENTSEL0_ENABLE; | ||
| 625 | wrmsr(evntsel_msr, evntsel, 0); | ||
| 626 | 681 | ||
| 627 | wd->perfctr_msr = perfctr_msr; | 682 | wd->perfctr_msr = perfctr_msr; |
| 628 | wd->evntsel_msr = evntsel_msr; | 683 | wd->evntsel_msr = evntsel_msr; |
| 629 | wd->cccr_msr = 0; /* unused */ | 684 | wd->cccr_msr = 0; /* unused */ |
| 685 | |||
| 686 | /* ok, everything is initialized, announce that we're set */ | ||
| 687 | cpu_nmi_set_wd_enabled(); | ||
| 688 | |||
| 689 | apic_write(APIC_LVTPC, APIC_DM_NMI); | ||
| 690 | evntsel |= ARCH_PERFMON_EVENTSEL0_ENABLE; | ||
| 691 | wrmsr(evntsel_msr, evntsel, 0); | ||
| 630 | intel_arch_wd_ops.checkbit = 1ULL << (eax.split.bit_width - 1); | 692 | intel_arch_wd_ops.checkbit = 1ULL << (eax.split.bit_width - 1); |
| 631 | return 1; | 693 | return 1; |
| 632 | } | 694 | } |
diff --git a/arch/x86/kernel/cpuid.c b/arch/x86/kernel/cpuid.c index 8e9cd6a8ec12..6a44d6465991 100644 --- a/arch/x86/kernel/cpuid.c +++ b/arch/x86/kernel/cpuid.c | |||
| @@ -36,7 +36,6 @@ | |||
| 36 | #include <linux/smp_lock.h> | 36 | #include <linux/smp_lock.h> |
| 37 | #include <linux/major.h> | 37 | #include <linux/major.h> |
| 38 | #include <linux/fs.h> | 38 | #include <linux/fs.h> |
| 39 | #include <linux/smp_lock.h> | ||
| 40 | #include <linux/device.h> | 39 | #include <linux/device.h> |
| 41 | #include <linux/cpu.h> | 40 | #include <linux/cpu.h> |
| 42 | #include <linux/notifier.h> | 41 | #include <linux/notifier.h> |
diff --git a/arch/x86/kernel/crash_dump_64.c b/arch/x86/kernel/crash_dump_64.c index 15e6c6bc4a46..e90a60ef10c2 100644 --- a/arch/x86/kernel/crash_dump_64.c +++ b/arch/x86/kernel/crash_dump_64.c | |||
| @@ -7,9 +7,8 @@ | |||
| 7 | 7 | ||
| 8 | #include <linux/errno.h> | 8 | #include <linux/errno.h> |
| 9 | #include <linux/crash_dump.h> | 9 | #include <linux/crash_dump.h> |
| 10 | 10 | #include <linux/uaccess.h> | |
| 11 | #include <asm/uaccess.h> | 11 | #include <linux/io.h> |
| 12 | #include <asm/io.h> | ||
| 13 | 12 | ||
| 14 | /** | 13 | /** |
| 15 | * copy_oldmem_page - copy one page from "oldmem" | 14 | * copy_oldmem_page - copy one page from "oldmem" |
| @@ -25,7 +24,7 @@ | |||
| 25 | * in the current kernel. We stitch up a pte, similar to kmap_atomic. | 24 | * in the current kernel. We stitch up a pte, similar to kmap_atomic. |
| 26 | */ | 25 | */ |
| 27 | ssize_t copy_oldmem_page(unsigned long pfn, char *buf, | 26 | ssize_t copy_oldmem_page(unsigned long pfn, char *buf, |
| 28 | size_t csize, unsigned long offset, int userbuf) | 27 | size_t csize, unsigned long offset, int userbuf) |
| 29 | { | 28 | { |
| 30 | void *vaddr; | 29 | void *vaddr; |
| 31 | 30 | ||
| @@ -33,14 +32,16 @@ ssize_t copy_oldmem_page(unsigned long pfn, char *buf, | |||
| 33 | return 0; | 32 | return 0; |
| 34 | 33 | ||
| 35 | vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); | 34 | vaddr = ioremap(pfn << PAGE_SHIFT, PAGE_SIZE); |
| 35 | if (!vaddr) | ||
| 36 | return -ENOMEM; | ||
| 36 | 37 | ||
| 37 | if (userbuf) { | 38 | if (userbuf) { |
| 38 | if (copy_to_user(buf, (vaddr + offset), csize)) { | 39 | if (copy_to_user(buf, vaddr + offset, csize)) { |
| 39 | iounmap(vaddr); | 40 | iounmap(vaddr); |
| 40 | return -EFAULT; | 41 | return -EFAULT; |
| 41 | } | 42 | } |
| 42 | } else | 43 | } else |
| 43 | memcpy(buf, (vaddr + offset), csize); | 44 | memcpy(buf, vaddr + offset, csize); |
| 44 | 45 | ||
| 45 | iounmap(vaddr); | 46 | iounmap(vaddr); |
| 46 | return csize; | 47 | return csize; |
diff --git a/arch/x86/kernel/efi.c b/arch/x86/kernel/efi.c index 06cc8d4254b1..945a31cdd81f 100644 --- a/arch/x86/kernel/efi.c +++ b/arch/x86/kernel/efi.c | |||
| @@ -414,9 +414,11 @@ void __init efi_init(void) | |||
| 414 | if (memmap.map == NULL) | 414 | if (memmap.map == NULL) |
| 415 | printk(KERN_ERR "Could not map the EFI memory map!\n"); | 415 | printk(KERN_ERR "Could not map the EFI memory map!\n"); |
| 416 | memmap.map_end = memmap.map + (memmap.nr_map * memmap.desc_size); | 416 | memmap.map_end = memmap.map + (memmap.nr_map * memmap.desc_size); |
| 417 | |||
| 417 | if (memmap.desc_size != sizeof(efi_memory_desc_t)) | 418 | if (memmap.desc_size != sizeof(efi_memory_desc_t)) |
| 418 | printk(KERN_WARNING "Kernel-defined memdesc" | 419 | printk(KERN_WARNING |
| 419 | "doesn't match the one from EFI!\n"); | 420 | "Kernel-defined memdesc doesn't match the one from EFI!\n"); |
| 421 | |||
| 420 | if (add_efi_memmap) | 422 | if (add_efi_memmap) |
| 421 | do_add_efi_memmap(); | 423 | do_add_efi_memmap(); |
| 422 | 424 | ||
diff --git a/arch/x86/kernel/entry_64.S b/arch/x86/kernel/entry_64.S index 89434d439605..cf3a0b2d0059 100644 --- a/arch/x86/kernel/entry_64.S +++ b/arch/x86/kernel/entry_64.S | |||
| @@ -275,9 +275,9 @@ ENTRY(native_usergs_sysret64) | |||
| 275 | ENTRY(ret_from_fork) | 275 | ENTRY(ret_from_fork) |
| 276 | CFI_DEFAULT_STACK | 276 | CFI_DEFAULT_STACK |
| 277 | push kernel_eflags(%rip) | 277 | push kernel_eflags(%rip) |
| 278 | CFI_ADJUST_CFA_OFFSET 4 | 278 | CFI_ADJUST_CFA_OFFSET 8 |
| 279 | popf # reset kernel eflags | 279 | popf # reset kernel eflags |
| 280 | CFI_ADJUST_CFA_OFFSET -4 | 280 | CFI_ADJUST_CFA_OFFSET -8 |
| 281 | call schedule_tail | 281 | call schedule_tail |
| 282 | GET_THREAD_INFO(%rcx) | 282 | GET_THREAD_INFO(%rcx) |
| 283 | testl $(_TIF_SYSCALL_TRACE|_TIF_SYSCALL_AUDIT),TI_flags(%rcx) | 283 | testl $(_TIF_SYSCALL_TRACE|_TIF_SYSCALL_AUDIT),TI_flags(%rcx) |
diff --git a/arch/x86/kernel/head64.c b/arch/x86/kernel/head64.c index 9bfc4d72fb2e..d16084f90649 100644 --- a/arch/x86/kernel/head64.c +++ b/arch/x86/kernel/head64.c | |||
| @@ -108,12 +108,11 @@ void __init x86_64_start_kernel(char * real_mode_data) | |||
| 108 | } | 108 | } |
| 109 | load_idt((const struct desc_ptr *)&idt_descr); | 109 | load_idt((const struct desc_ptr *)&idt_descr); |
| 110 | 110 | ||
| 111 | early_printk("Kernel alive\n"); | 111 | if (console_loglevel == 10) |
| 112 | early_printk("Kernel alive\n"); | ||
| 112 | 113 | ||
| 113 | x86_64_init_pda(); | 114 | x86_64_init_pda(); |
| 114 | 115 | ||
| 115 | early_printk("Kernel really alive\n"); | ||
| 116 | |||
| 117 | x86_64_start_reservations(real_mode_data); | 116 | x86_64_start_reservations(real_mode_data); |
| 118 | } | 117 | } |
| 119 | 118 | ||
diff --git a/arch/x86/kernel/irq_32.c b/arch/x86/kernel/irq_32.c index 1cf8c1fcc088..b71e02d42f4f 100644 --- a/arch/x86/kernel/irq_32.c +++ b/arch/x86/kernel/irq_32.c | |||
| @@ -325,7 +325,7 @@ skip: | |||
| 325 | for_each_online_cpu(j) | 325 | for_each_online_cpu(j) |
| 326 | seq_printf(p, "%10u ", | 326 | seq_printf(p, "%10u ", |
| 327 | per_cpu(irq_stat,j).irq_call_count); | 327 | per_cpu(irq_stat,j).irq_call_count); |
| 328 | seq_printf(p, " function call interrupts\n"); | 328 | seq_printf(p, " Function call interrupts\n"); |
| 329 | seq_printf(p, "TLB: "); | 329 | seq_printf(p, "TLB: "); |
| 330 | for_each_online_cpu(j) | 330 | for_each_online_cpu(j) |
| 331 | seq_printf(p, "%10u ", | 331 | seq_printf(p, "%10u ", |
diff --git a/arch/x86/kernel/irq_64.c b/arch/x86/kernel/irq_64.c index 1f78b238d8d2..f065fe9071b9 100644 --- a/arch/x86/kernel/irq_64.c +++ b/arch/x86/kernel/irq_64.c | |||
| @@ -129,7 +129,7 @@ skip: | |||
| 129 | seq_printf(p, "CAL: "); | 129 | seq_printf(p, "CAL: "); |
| 130 | for_each_online_cpu(j) | 130 | for_each_online_cpu(j) |
| 131 | seq_printf(p, "%10u ", cpu_pda(j)->irq_call_count); | 131 | seq_printf(p, "%10u ", cpu_pda(j)->irq_call_count); |
| 132 | seq_printf(p, " function call interrupts\n"); | 132 | seq_printf(p, " Function call interrupts\n"); |
| 133 | seq_printf(p, "TLB: "); | 133 | seq_printf(p, "TLB: "); |
| 134 | for_each_online_cpu(j) | 134 | for_each_online_cpu(j) |
| 135 | seq_printf(p, "%10u ", cpu_pda(j)->irq_tlb_count); | 135 | seq_printf(p, "%10u ", cpu_pda(j)->irq_tlb_count); |
diff --git a/arch/x86/kernel/kvm.c b/arch/x86/kernel/kvm.c index 8b7a3cf37d2b..478bca986eca 100644 --- a/arch/x86/kernel/kvm.c +++ b/arch/x86/kernel/kvm.c | |||
| @@ -178,7 +178,7 @@ static void kvm_flush_tlb(void) | |||
| 178 | kvm_deferred_mmu_op(&ftlb, sizeof ftlb); | 178 | kvm_deferred_mmu_op(&ftlb, sizeof ftlb); |
| 179 | } | 179 | } |
| 180 | 180 | ||
| 181 | static void kvm_release_pt(u32 pfn) | 181 | static void kvm_release_pt(unsigned long pfn) |
| 182 | { | 182 | { |
| 183 | struct kvm_mmu_op_release_pt rpt = { | 183 | struct kvm_mmu_op_release_pt rpt = { |
| 184 | .header.op = KVM_MMU_OP_RELEASE_PT, | 184 | .header.op = KVM_MMU_OP_RELEASE_PT, |
diff --git a/arch/x86/kernel/nmi.c b/arch/x86/kernel/nmi.c index abb78a2cc4ad..2c97f07f1c2c 100644 --- a/arch/x86/kernel/nmi.c +++ b/arch/x86/kernel/nmi.c | |||
| @@ -299,6 +299,15 @@ void acpi_nmi_disable(void) | |||
| 299 | on_each_cpu(__acpi_nmi_disable, NULL, 1); | 299 | on_each_cpu(__acpi_nmi_disable, NULL, 1); |
| 300 | } | 300 | } |
| 301 | 301 | ||
| 302 | /* | ||
| 303 | * This function is called as soon the LAPIC NMI watchdog driver has everything | ||
| 304 | * in place and it's ready to check if the NMIs belong to the NMI watchdog | ||
| 305 | */ | ||
| 306 | void cpu_nmi_set_wd_enabled(void) | ||
| 307 | { | ||
| 308 | __get_cpu_var(wd_enabled) = 1; | ||
| 309 | } | ||
| 310 | |||
| 302 | void setup_apic_nmi_watchdog(void *unused) | 311 | void setup_apic_nmi_watchdog(void *unused) |
| 303 | { | 312 | { |
| 304 | if (__get_cpu_var(wd_enabled)) | 313 | if (__get_cpu_var(wd_enabled)) |
| @@ -311,8 +320,6 @@ void setup_apic_nmi_watchdog(void *unused) | |||
| 311 | 320 | ||
| 312 | switch (nmi_watchdog) { | 321 | switch (nmi_watchdog) { |
| 313 | case NMI_LOCAL_APIC: | 322 | case NMI_LOCAL_APIC: |
| 314 | /* enable it before to avoid race with handler */ | ||
| 315 | __get_cpu_var(wd_enabled) = 1; | ||
| 316 | if (lapic_watchdog_init(nmi_hz) < 0) { | 323 | if (lapic_watchdog_init(nmi_hz) < 0) { |
| 317 | __get_cpu_var(wd_enabled) = 0; | 324 | __get_cpu_var(wd_enabled) = 0; |
| 318 | return; | 325 | return; |
diff --git a/arch/x86/kernel/olpc.c b/arch/x86/kernel/olpc.c index 3e6672274807..7a13fac63a1f 100644 --- a/arch/x86/kernel/olpc.c +++ b/arch/x86/kernel/olpc.c | |||
| @@ -190,12 +190,12 @@ EXPORT_SYMBOL_GPL(olpc_ec_cmd); | |||
| 190 | static void __init platform_detect(void) | 190 | static void __init platform_detect(void) |
| 191 | { | 191 | { |
| 192 | size_t propsize; | 192 | size_t propsize; |
| 193 | u32 rev; | 193 | __be32 rev; |
| 194 | 194 | ||
| 195 | if (ofw("getprop", 4, 1, NULL, "board-revision-int", &rev, 4, | 195 | if (ofw("getprop", 4, 1, NULL, "board-revision-int", &rev, 4, |
| 196 | &propsize) || propsize != 4) { | 196 | &propsize) || propsize != 4) { |
| 197 | printk(KERN_ERR "ofw: getprop call failed!\n"); | 197 | printk(KERN_ERR "ofw: getprop call failed!\n"); |
| 198 | rev = 0; | 198 | rev = cpu_to_be32(0); |
| 199 | } | 199 | } |
| 200 | olpc_platform_info.boardrev = be32_to_cpu(rev); | 200 | olpc_platform_info.boardrev = be32_to_cpu(rev); |
| 201 | } | 201 | } |
| @@ -203,7 +203,7 @@ static void __init platform_detect(void) | |||
| 203 | static void __init platform_detect(void) | 203 | static void __init platform_detect(void) |
| 204 | { | 204 | { |
| 205 | /* stopgap until OFW support is added to the kernel */ | 205 | /* stopgap until OFW support is added to the kernel */ |
| 206 | olpc_platform_info.boardrev = be32_to_cpu(0xc2); | 206 | olpc_platform_info.boardrev = 0xc2; |
| 207 | } | 207 | } |
| 208 | #endif | 208 | #endif |
| 209 | 209 | ||
diff --git a/arch/x86/kernel/paravirt.c b/arch/x86/kernel/paravirt.c index 300da17e61cb..e2f43768723a 100644 --- a/arch/x86/kernel/paravirt.c +++ b/arch/x86/kernel/paravirt.c | |||
| @@ -330,6 +330,7 @@ struct pv_cpu_ops pv_cpu_ops = { | |||
| 330 | #endif | 330 | #endif |
| 331 | .wbinvd = native_wbinvd, | 331 | .wbinvd = native_wbinvd, |
| 332 | .read_msr = native_read_msr_safe, | 332 | .read_msr = native_read_msr_safe, |
| 333 | .read_msr_amd = native_read_msr_amd_safe, | ||
| 333 | .write_msr = native_write_msr_safe, | 334 | .write_msr = native_write_msr_safe, |
| 334 | .read_tsc = native_read_tsc, | 335 | .read_tsc = native_read_tsc, |
| 335 | .read_pmc = native_read_pmc, | 336 | .read_pmc = native_read_pmc, |
diff --git a/arch/x86/kernel/paravirt_patch_32.c b/arch/x86/kernel/paravirt_patch_32.c index 58262218781b..9fe644f4861d 100644 --- a/arch/x86/kernel/paravirt_patch_32.c +++ b/arch/x86/kernel/paravirt_patch_32.c | |||
| @@ -23,7 +23,7 @@ unsigned native_patch(u8 type, u16 clobbers, void *ibuf, | |||
| 23 | start = start_##ops##_##x; \ | 23 | start = start_##ops##_##x; \ |
| 24 | end = end_##ops##_##x; \ | 24 | end = end_##ops##_##x; \ |
| 25 | goto patch_site | 25 | goto patch_site |
| 26 | switch(type) { | 26 | switch (type) { |
| 27 | PATCH_SITE(pv_irq_ops, irq_disable); | 27 | PATCH_SITE(pv_irq_ops, irq_disable); |
| 28 | PATCH_SITE(pv_irq_ops, irq_enable); | 28 | PATCH_SITE(pv_irq_ops, irq_enable); |
| 29 | PATCH_SITE(pv_irq_ops, restore_fl); | 29 | PATCH_SITE(pv_irq_ops, restore_fl); |
diff --git a/arch/x86/kernel/pci-dma.c b/arch/x86/kernel/pci-dma.c index 87d4d6964ec2..f704cb51ff82 100644 --- a/arch/x86/kernel/pci-dma.c +++ b/arch/x86/kernel/pci-dma.c | |||
| @@ -82,7 +82,7 @@ void __init dma32_reserve_bootmem(void) | |||
| 82 | * using 512M as goal | 82 | * using 512M as goal |
| 83 | */ | 83 | */ |
| 84 | align = 64ULL<<20; | 84 | align = 64ULL<<20; |
| 85 | size = round_up(dma32_bootmem_size, align); | 85 | size = roundup(dma32_bootmem_size, align); |
| 86 | dma32_bootmem_ptr = __alloc_bootmem_nopanic(size, align, | 86 | dma32_bootmem_ptr = __alloc_bootmem_nopanic(size, align, |
| 87 | 512ULL<<20); | 87 | 512ULL<<20); |
| 88 | if (dma32_bootmem_ptr) | 88 | if (dma32_bootmem_ptr) |
diff --git a/arch/x86/kernel/pci-gart_64.c b/arch/x86/kernel/pci-gart_64.c index 49285f8fd4d5..1a895a582534 100644 --- a/arch/x86/kernel/pci-gart_64.c +++ b/arch/x86/kernel/pci-gart_64.c | |||
| @@ -82,7 +82,8 @@ AGPEXTERN __u32 *agp_gatt_table; | |||
| 82 | static unsigned long next_bit; /* protected by iommu_bitmap_lock */ | 82 | static unsigned long next_bit; /* protected by iommu_bitmap_lock */ |
| 83 | static int need_flush; /* global flush state. set for each gart wrap */ | 83 | static int need_flush; /* global flush state. set for each gart wrap */ |
| 84 | 84 | ||
| 85 | static unsigned long alloc_iommu(struct device *dev, int size) | 85 | static unsigned long alloc_iommu(struct device *dev, int size, |
| 86 | unsigned long align_mask) | ||
| 86 | { | 87 | { |
| 87 | unsigned long offset, flags; | 88 | unsigned long offset, flags; |
| 88 | unsigned long boundary_size; | 89 | unsigned long boundary_size; |
| @@ -90,16 +91,17 @@ static unsigned long alloc_iommu(struct device *dev, int size) | |||
| 90 | 91 | ||
| 91 | base_index = ALIGN(iommu_bus_base & dma_get_seg_boundary(dev), | 92 | base_index = ALIGN(iommu_bus_base & dma_get_seg_boundary(dev), |
| 92 | PAGE_SIZE) >> PAGE_SHIFT; | 93 | PAGE_SIZE) >> PAGE_SHIFT; |
| 93 | boundary_size = ALIGN(dma_get_seg_boundary(dev) + 1, | 94 | boundary_size = ALIGN((unsigned long long)dma_get_seg_boundary(dev) + 1, |
| 94 | PAGE_SIZE) >> PAGE_SHIFT; | 95 | PAGE_SIZE) >> PAGE_SHIFT; |
| 95 | 96 | ||
| 96 | spin_lock_irqsave(&iommu_bitmap_lock, flags); | 97 | spin_lock_irqsave(&iommu_bitmap_lock, flags); |
| 97 | offset = iommu_area_alloc(iommu_gart_bitmap, iommu_pages, next_bit, | 98 | offset = iommu_area_alloc(iommu_gart_bitmap, iommu_pages, next_bit, |
| 98 | size, base_index, boundary_size, 0); | 99 | size, base_index, boundary_size, align_mask); |
| 99 | if (offset == -1) { | 100 | if (offset == -1) { |
| 100 | need_flush = 1; | 101 | need_flush = 1; |
| 101 | offset = iommu_area_alloc(iommu_gart_bitmap, iommu_pages, 0, | 102 | offset = iommu_area_alloc(iommu_gart_bitmap, iommu_pages, 0, |
| 102 | size, base_index, boundary_size, 0); | 103 | size, base_index, boundary_size, |
| 104 | align_mask); | ||
| 103 | } | 105 | } |
| 104 | if (offset != -1) { | 106 | if (offset != -1) { |
| 105 | next_bit = offset+size; | 107 | next_bit = offset+size; |
| @@ -236,10 +238,10 @@ nonforced_iommu(struct device *dev, unsigned long addr, size_t size) | |||
| 236 | * Caller needs to check if the iommu is needed and flush. | 238 | * Caller needs to check if the iommu is needed and flush. |
| 237 | */ | 239 | */ |
| 238 | static dma_addr_t dma_map_area(struct device *dev, dma_addr_t phys_mem, | 240 | static dma_addr_t dma_map_area(struct device *dev, dma_addr_t phys_mem, |
| 239 | size_t size, int dir) | 241 | size_t size, int dir, unsigned long align_mask) |
| 240 | { | 242 | { |
| 241 | unsigned long npages = iommu_num_pages(phys_mem, size); | 243 | unsigned long npages = iommu_num_pages(phys_mem, size); |
| 242 | unsigned long iommu_page = alloc_iommu(dev, npages); | 244 | unsigned long iommu_page = alloc_iommu(dev, npages, align_mask); |
| 243 | int i; | 245 | int i; |
| 244 | 246 | ||
| 245 | if (iommu_page == -1) { | 247 | if (iommu_page == -1) { |
| @@ -262,7 +264,11 @@ static dma_addr_t dma_map_area(struct device *dev, dma_addr_t phys_mem, | |||
| 262 | static dma_addr_t | 264 | static dma_addr_t |
| 263 | gart_map_simple(struct device *dev, phys_addr_t paddr, size_t size, int dir) | 265 | gart_map_simple(struct device *dev, phys_addr_t paddr, size_t size, int dir) |
| 264 | { | 266 | { |
| 265 | dma_addr_t map = dma_map_area(dev, paddr, size, dir); | 267 | dma_addr_t map; |
| 268 | unsigned long align_mask; | ||
| 269 | |||
| 270 | align_mask = (1UL << get_order(size)) - 1; | ||
| 271 | map = dma_map_area(dev, paddr, size, dir, align_mask); | ||
| 266 | 272 | ||
| 267 | flush_gart(); | 273 | flush_gart(); |
| 268 | 274 | ||
| @@ -281,7 +287,8 @@ gart_map_single(struct device *dev, phys_addr_t paddr, size_t size, int dir) | |||
| 281 | if (!need_iommu(dev, paddr, size)) | 287 | if (!need_iommu(dev, paddr, size)) |
| 282 | return paddr; | 288 | return paddr; |
| 283 | 289 | ||
| 284 | bus = gart_map_simple(dev, paddr, size, dir); | 290 | bus = dma_map_area(dev, paddr, size, dir, 0); |
| 291 | flush_gart(); | ||
| 285 | 292 | ||
| 286 | return bus; | 293 | return bus; |
| 287 | } | 294 | } |
| @@ -340,7 +347,7 @@ static int dma_map_sg_nonforce(struct device *dev, struct scatterlist *sg, | |||
| 340 | unsigned long addr = sg_phys(s); | 347 | unsigned long addr = sg_phys(s); |
| 341 | 348 | ||
| 342 | if (nonforced_iommu(dev, addr, s->length)) { | 349 | if (nonforced_iommu(dev, addr, s->length)) { |
| 343 | addr = dma_map_area(dev, addr, s->length, dir); | 350 | addr = dma_map_area(dev, addr, s->length, dir, 0); |
| 344 | if (addr == bad_dma_address) { | 351 | if (addr == bad_dma_address) { |
| 345 | if (i > 0) | 352 | if (i > 0) |
| 346 | gart_unmap_sg(dev, sg, i, dir); | 353 | gart_unmap_sg(dev, sg, i, dir); |
| @@ -362,7 +369,7 @@ static int __dma_map_cont(struct device *dev, struct scatterlist *start, | |||
| 362 | int nelems, struct scatterlist *sout, | 369 | int nelems, struct scatterlist *sout, |
| 363 | unsigned long pages) | 370 | unsigned long pages) |
| 364 | { | 371 | { |
| 365 | unsigned long iommu_start = alloc_iommu(dev, pages); | 372 | unsigned long iommu_start = alloc_iommu(dev, pages, 0); |
| 366 | unsigned long iommu_page = iommu_start; | 373 | unsigned long iommu_page = iommu_start; |
| 367 | struct scatterlist *s; | 374 | struct scatterlist *s; |
| 368 | int i; | 375 | int i; |
| @@ -626,7 +633,6 @@ static __init int init_k8_gatt(struct agp_kern_info *info) | |||
| 626 | struct pci_dev *dev; | 633 | struct pci_dev *dev; |
| 627 | void *gatt; | 634 | void *gatt; |
| 628 | int i, error; | 635 | int i, error; |
| 629 | unsigned long start_pfn, end_pfn; | ||
| 630 | 636 | ||
| 631 | printk(KERN_INFO "PCI-DMA: Disabling AGP.\n"); | 637 | printk(KERN_INFO "PCI-DMA: Disabling AGP.\n"); |
| 632 | aper_size = aper_base = info->aper_size = 0; | 638 | aper_size = aper_base = info->aper_size = 0; |
| @@ -672,12 +678,6 @@ static __init int init_k8_gatt(struct agp_kern_info *info) | |||
| 672 | printk(KERN_INFO "PCI-DMA: aperture base @ %x size %u KB\n", | 678 | printk(KERN_INFO "PCI-DMA: aperture base @ %x size %u KB\n", |
| 673 | aper_base, aper_size>>10); | 679 | aper_base, aper_size>>10); |
| 674 | 680 | ||
| 675 | /* need to map that range */ | ||
| 676 | end_pfn = (aper_base>>PAGE_SHIFT) + (aper_size>>PAGE_SHIFT); | ||
| 677 | if (end_pfn > max_low_pfn_mapped) { | ||
| 678 | start_pfn = (aper_base>>PAGE_SHIFT); | ||
| 679 | init_memory_mapping(start_pfn<<PAGE_SHIFT, end_pfn<<PAGE_SHIFT); | ||
| 680 | } | ||
| 681 | return 0; | 681 | return 0; |
| 682 | 682 | ||
| 683 | nommu: | 683 | nommu: |
| @@ -727,7 +727,8 @@ void __init gart_iommu_init(void) | |||
| 727 | { | 727 | { |
| 728 | struct agp_kern_info info; | 728 | struct agp_kern_info info; |
| 729 | unsigned long iommu_start; | 729 | unsigned long iommu_start; |
| 730 | unsigned long aper_size; | 730 | unsigned long aper_base, aper_size; |
| 731 | unsigned long start_pfn, end_pfn; | ||
| 731 | unsigned long scratch; | 732 | unsigned long scratch; |
| 732 | long i; | 733 | long i; |
| 733 | 734 | ||
| @@ -765,8 +766,16 @@ void __init gart_iommu_init(void) | |||
| 765 | return; | 766 | return; |
| 766 | } | 767 | } |
| 767 | 768 | ||
| 769 | /* need to map that range */ | ||
| 770 | aper_size = info.aper_size << 20; | ||
| 771 | aper_base = info.aper_base; | ||
| 772 | end_pfn = (aper_base>>PAGE_SHIFT) + (aper_size>>PAGE_SHIFT); | ||
| 773 | if (end_pfn > max_low_pfn_mapped) { | ||
| 774 | start_pfn = (aper_base>>PAGE_SHIFT); | ||
| 775 | init_memory_mapping(start_pfn<<PAGE_SHIFT, end_pfn<<PAGE_SHIFT); | ||
| 776 | } | ||
| 777 | |||
| 768 | printk(KERN_INFO "PCI-DMA: using GART IOMMU.\n"); | 778 | printk(KERN_INFO "PCI-DMA: using GART IOMMU.\n"); |
| 769 | aper_size = info.aper_size * 1024 * 1024; | ||
| 770 | iommu_size = check_iommu_size(info.aper_base, aper_size); | 779 | iommu_size = check_iommu_size(info.aper_base, aper_size); |
| 771 | iommu_pages = iommu_size >> PAGE_SHIFT; | 780 | iommu_pages = iommu_size >> PAGE_SHIFT; |
| 772 | 781 | ||
diff --git a/arch/x86/kernel/pcspeaker.c b/arch/x86/kernel/pcspeaker.c index bc1f2d3ea277..a311ffcaad16 100644 --- a/arch/x86/kernel/pcspeaker.c +++ b/arch/x86/kernel/pcspeaker.c | |||
| @@ -1,20 +1,13 @@ | |||
| 1 | #include <linux/platform_device.h> | 1 | #include <linux/platform_device.h> |
| 2 | #include <linux/errno.h> | 2 | #include <linux/err.h> |
| 3 | #include <linux/init.h> | 3 | #include <linux/init.h> |
| 4 | 4 | ||
| 5 | static __init int add_pcspkr(void) | 5 | static __init int add_pcspkr(void) |
| 6 | { | 6 | { |
| 7 | struct platform_device *pd; | 7 | struct platform_device *pd; |
| 8 | int ret; | ||
| 9 | 8 | ||
| 10 | pd = platform_device_alloc("pcspkr", -1); | 9 | pd = platform_device_register_simple("pcspkr", -1, NULL, 0); |
| 11 | if (!pd) | ||
| 12 | return -ENOMEM; | ||
| 13 | 10 | ||
| 14 | ret = platform_device_add(pd); | 11 | return IS_ERR(pd) ? PTR_ERR(pd) : 0; |
| 15 | if (ret) | ||
| 16 | platform_device_put(pd); | ||
| 17 | |||
| 18 | return ret; | ||
| 19 | } | 12 | } |
| 20 | device_initcall(add_pcspkr); | 13 | device_initcall(add_pcspkr); |
diff --git a/arch/x86/kernel/process.c b/arch/x86/kernel/process.c index 876e91890777..ec7a2ba9bce8 100644 --- a/arch/x86/kernel/process.c +++ b/arch/x86/kernel/process.c | |||
| @@ -185,7 +185,8 @@ static void mwait_idle(void) | |||
| 185 | static void poll_idle(void) | 185 | static void poll_idle(void) |
| 186 | { | 186 | { |
| 187 | local_irq_enable(); | 187 | local_irq_enable(); |
| 188 | cpu_relax(); | 188 | while (!need_resched()) |
| 189 | cpu_relax(); | ||
| 189 | } | 190 | } |
| 190 | 191 | ||
| 191 | /* | 192 | /* |
diff --git a/arch/x86/kernel/process_32.c b/arch/x86/kernel/process_32.c index c8609dea443f..205188db9626 100644 --- a/arch/x86/kernel/process_32.c +++ b/arch/x86/kernel/process_32.c | |||
| @@ -37,6 +37,7 @@ | |||
| 37 | #include <linux/tick.h> | 37 | #include <linux/tick.h> |
| 38 | #include <linux/percpu.h> | 38 | #include <linux/percpu.h> |
| 39 | #include <linux/prctl.h> | 39 | #include <linux/prctl.h> |
| 40 | #include <linux/dmi.h> | ||
| 40 | 41 | ||
| 41 | #include <asm/uaccess.h> | 42 | #include <asm/uaccess.h> |
| 42 | #include <asm/pgtable.h> | 43 | #include <asm/pgtable.h> |
| @@ -163,6 +164,7 @@ void __show_registers(struct pt_regs *regs, int all) | |||
| 163 | unsigned long d0, d1, d2, d3, d6, d7; | 164 | unsigned long d0, d1, d2, d3, d6, d7; |
| 164 | unsigned long sp; | 165 | unsigned long sp; |
| 165 | unsigned short ss, gs; | 166 | unsigned short ss, gs; |
| 167 | const char *board; | ||
| 166 | 168 | ||
| 167 | if (user_mode_vm(regs)) { | 169 | if (user_mode_vm(regs)) { |
| 168 | sp = regs->sp; | 170 | sp = regs->sp; |
| @@ -175,11 +177,15 @@ void __show_registers(struct pt_regs *regs, int all) | |||
| 175 | } | 177 | } |
| 176 | 178 | ||
| 177 | printk("\n"); | 179 | printk("\n"); |
| 178 | printk("Pid: %d, comm: %s %s (%s %.*s)\n", | 180 | |
| 181 | board = dmi_get_system_info(DMI_PRODUCT_NAME); | ||
| 182 | if (!board) | ||
| 183 | board = ""; | ||
| 184 | printk("Pid: %d, comm: %s %s (%s %.*s) %s\n", | ||
| 179 | task_pid_nr(current), current->comm, | 185 | task_pid_nr(current), current->comm, |
| 180 | print_tainted(), init_utsname()->release, | 186 | print_tainted(), init_utsname()->release, |
| 181 | (int)strcspn(init_utsname()->version, " "), | 187 | (int)strcspn(init_utsname()->version, " "), |
| 182 | init_utsname()->version); | 188 | init_utsname()->version, board); |
| 183 | 189 | ||
| 184 | printk("EIP: %04x:[<%08lx>] EFLAGS: %08lx CPU: %d\n", | 190 | printk("EIP: %04x:[<%08lx>] EFLAGS: %08lx CPU: %d\n", |
| 185 | (u16)regs->cs, regs->ip, regs->flags, | 191 | (u16)regs->cs, regs->ip, regs->flags, |
diff --git a/arch/x86/kernel/process_64.c b/arch/x86/kernel/process_64.c index 79e3e173ab40..2a8ccb9238b4 100644 --- a/arch/x86/kernel/process_64.c +++ b/arch/x86/kernel/process_64.c | |||
| @@ -37,11 +37,11 @@ | |||
| 37 | #include <linux/kdebug.h> | 37 | #include <linux/kdebug.h> |
| 38 | #include <linux/tick.h> | 38 | #include <linux/tick.h> |
| 39 | #include <linux/prctl.h> | 39 | #include <linux/prctl.h> |
| 40 | #include <linux/uaccess.h> | ||
| 41 | #include <linux/io.h> | ||
| 40 | 42 | ||
| 41 | #include <asm/uaccess.h> | ||
| 42 | #include <asm/pgtable.h> | 43 | #include <asm/pgtable.h> |
| 43 | #include <asm/system.h> | 44 | #include <asm/system.h> |
| 44 | #include <asm/io.h> | ||
| 45 | #include <asm/processor.h> | 45 | #include <asm/processor.h> |
| 46 | #include <asm/i387.h> | 46 | #include <asm/i387.h> |
| 47 | #include <asm/mmu_context.h> | 47 | #include <asm/mmu_context.h> |
| @@ -89,7 +89,7 @@ void exit_idle(void) | |||
| 89 | #ifdef CONFIG_HOTPLUG_CPU | 89 | #ifdef CONFIG_HOTPLUG_CPU |
| 90 | DECLARE_PER_CPU(int, cpu_state); | 90 | DECLARE_PER_CPU(int, cpu_state); |
| 91 | 91 | ||
| 92 | #include <asm/nmi.h> | 92 | #include <linux/nmi.h> |
| 93 | /* We halt the CPU with physical CPU hotplug */ | 93 | /* We halt the CPU with physical CPU hotplug */ |
| 94 | static inline void play_dead(void) | 94 | static inline void play_dead(void) |
| 95 | { | 95 | { |
| @@ -154,7 +154,7 @@ void cpu_idle(void) | |||
| 154 | } | 154 | } |
| 155 | 155 | ||
| 156 | /* Prints also some state that isn't saved in the pt_regs */ | 156 | /* Prints also some state that isn't saved in the pt_regs */ |
| 157 | void __show_regs(struct pt_regs * regs) | 157 | void __show_regs(struct pt_regs *regs) |
| 158 | { | 158 | { |
| 159 | unsigned long cr0 = 0L, cr2 = 0L, cr3 = 0L, cr4 = 0L, fs, gs, shadowgs; | 159 | unsigned long cr0 = 0L, cr2 = 0L, cr3 = 0L, cr4 = 0L, fs, gs, shadowgs; |
| 160 | unsigned long d0, d1, d2, d3, d6, d7; | 160 | unsigned long d0, d1, d2, d3, d6, d7; |
| @@ -163,59 +163,61 @@ void __show_regs(struct pt_regs * regs) | |||
| 163 | 163 | ||
| 164 | printk("\n"); | 164 | printk("\n"); |
| 165 | print_modules(); | 165 | print_modules(); |
| 166 | printk("Pid: %d, comm: %.20s %s %s %.*s\n", | 166 | printk(KERN_INFO "Pid: %d, comm: %.20s %s %s %.*s\n", |
| 167 | current->pid, current->comm, print_tainted(), | 167 | current->pid, current->comm, print_tainted(), |
| 168 | init_utsname()->release, | 168 | init_utsname()->release, |
| 169 | (int)strcspn(init_utsname()->version, " "), | 169 | (int)strcspn(init_utsname()->version, " "), |
| 170 | init_utsname()->version); | 170 | init_utsname()->version); |
| 171 | printk("RIP: %04lx:[<%016lx>] ", regs->cs & 0xffff, regs->ip); | 171 | printk(KERN_INFO "RIP: %04lx:[<%016lx>] ", regs->cs & 0xffff, regs->ip); |
| 172 | printk_address(regs->ip, 1); | 172 | printk_address(regs->ip, 1); |
| 173 | printk("RSP: %04lx:%016lx EFLAGS: %08lx\n", regs->ss, regs->sp, | 173 | printk(KERN_INFO "RSP: %04lx:%016lx EFLAGS: %08lx\n", regs->ss, |
| 174 | regs->flags); | 174 | regs->sp, regs->flags); |
| 175 | printk("RAX: %016lx RBX: %016lx RCX: %016lx\n", | 175 | printk(KERN_INFO "RAX: %016lx RBX: %016lx RCX: %016lx\n", |
| 176 | regs->ax, regs->bx, regs->cx); | 176 | regs->ax, regs->bx, regs->cx); |
| 177 | printk("RDX: %016lx RSI: %016lx RDI: %016lx\n", | 177 | printk(KERN_INFO "RDX: %016lx RSI: %016lx RDI: %016lx\n", |
| 178 | regs->dx, regs->si, regs->di); | 178 | regs->dx, regs->si, regs->di); |
| 179 | printk("RBP: %016lx R08: %016lx R09: %016lx\n", | 179 | printk(KERN_INFO "RBP: %016lx R08: %016lx R09: %016lx\n", |
| 180 | regs->bp, regs->r8, regs->r9); | 180 | regs->bp, regs->r8, regs->r9); |
| 181 | printk("R10: %016lx R11: %016lx R12: %016lx\n", | 181 | printk(KERN_INFO "R10: %016lx R11: %016lx R12: %016lx\n", |
| 182 | regs->r10, regs->r11, regs->r12); | 182 | regs->r10, regs->r11, regs->r12); |
| 183 | printk("R13: %016lx R14: %016lx R15: %016lx\n", | 183 | printk(KERN_INFO "R13: %016lx R14: %016lx R15: %016lx\n", |
| 184 | regs->r13, regs->r14, regs->r15); | 184 | regs->r13, regs->r14, regs->r15); |
| 185 | 185 | ||
| 186 | asm("movl %%ds,%0" : "=r" (ds)); | 186 | asm("movl %%ds,%0" : "=r" (ds)); |
| 187 | asm("movl %%cs,%0" : "=r" (cs)); | 187 | asm("movl %%cs,%0" : "=r" (cs)); |
| 188 | asm("movl %%es,%0" : "=r" (es)); | 188 | asm("movl %%es,%0" : "=r" (es)); |
| 189 | asm("movl %%fs,%0" : "=r" (fsindex)); | 189 | asm("movl %%fs,%0" : "=r" (fsindex)); |
| 190 | asm("movl %%gs,%0" : "=r" (gsindex)); | 190 | asm("movl %%gs,%0" : "=r" (gsindex)); |
| 191 | 191 | ||
| 192 | rdmsrl(MSR_FS_BASE, fs); | 192 | rdmsrl(MSR_FS_BASE, fs); |
| 193 | rdmsrl(MSR_GS_BASE, gs); | 193 | rdmsrl(MSR_GS_BASE, gs); |
| 194 | rdmsrl(MSR_KERNEL_GS_BASE, shadowgs); | 194 | rdmsrl(MSR_KERNEL_GS_BASE, shadowgs); |
| 195 | 195 | ||
| 196 | cr0 = read_cr0(); | 196 | cr0 = read_cr0(); |
| 197 | cr2 = read_cr2(); | 197 | cr2 = read_cr2(); |
| 198 | cr3 = read_cr3(); | 198 | cr3 = read_cr3(); |
| 199 | cr4 = read_cr4(); | 199 | cr4 = read_cr4(); |
| 200 | 200 | ||
| 201 | printk("FS: %016lx(%04x) GS:%016lx(%04x) knlGS:%016lx\n", | 201 | printk(KERN_INFO "FS: %016lx(%04x) GS:%016lx(%04x) knlGS:%016lx\n", |
| 202 | fs,fsindex,gs,gsindex,shadowgs); | 202 | fs, fsindex, gs, gsindex, shadowgs); |
| 203 | printk("CS: %04x DS: %04x ES: %04x CR0: %016lx\n", cs, ds, es, cr0); | 203 | printk(KERN_INFO "CS: %04x DS: %04x ES: %04x CR0: %016lx\n", cs, ds, |
| 204 | printk("CR2: %016lx CR3: %016lx CR4: %016lx\n", cr2, cr3, cr4); | 204 | es, cr0); |
| 205 | printk(KERN_INFO "CR2: %016lx CR3: %016lx CR4: %016lx\n", cr2, cr3, | ||
| 206 | cr4); | ||
| 205 | 207 | ||
| 206 | get_debugreg(d0, 0); | 208 | get_debugreg(d0, 0); |
| 207 | get_debugreg(d1, 1); | 209 | get_debugreg(d1, 1); |
| 208 | get_debugreg(d2, 2); | 210 | get_debugreg(d2, 2); |
| 209 | printk("DR0: %016lx DR1: %016lx DR2: %016lx\n", d0, d1, d2); | 211 | printk(KERN_INFO "DR0: %016lx DR1: %016lx DR2: %016lx\n", d0, d1, d2); |
| 210 | get_debugreg(d3, 3); | 212 | get_debugreg(d3, 3); |
| 211 | get_debugreg(d6, 6); | 213 | get_debugreg(d6, 6); |
| 212 | get_debugreg(d7, 7); | 214 | get_debugreg(d7, 7); |
| 213 | printk("DR3: %016lx DR6: %016lx DR7: %016lx\n", d3, d6, d7); | 215 | printk(KERN_INFO "DR3: %016lx DR6: %016lx DR7: %016lx\n", d3, d6, d7); |
| 214 | } | 216 | } |
| 215 | 217 | ||
| 216 | void show_regs(struct pt_regs *regs) | 218 | void show_regs(struct pt_regs *regs) |
| 217 | { | 219 | { |
| 218 | printk("CPU %d:", smp_processor_id()); | 220 | printk(KERN_INFO "CPU %d:", smp_processor_id()); |
| 219 | __show_regs(regs); | 221 | __show_regs(regs); |
| 220 | show_trace(NULL, regs, (void *)(regs + 1), regs->bp); | 222 | show_trace(NULL, regs, (void *)(regs + 1), regs->bp); |
| 221 | } | 223 | } |
| @@ -324,10 +326,10 @@ void prepare_to_copy(struct task_struct *tsk) | |||
| 324 | 326 | ||
| 325 | int copy_thread(int nr, unsigned long clone_flags, unsigned long sp, | 327 | int copy_thread(int nr, unsigned long clone_flags, unsigned long sp, |
| 326 | unsigned long unused, | 328 | unsigned long unused, |
| 327 | struct task_struct * p, struct pt_regs * regs) | 329 | struct task_struct *p, struct pt_regs *regs) |
| 328 | { | 330 | { |
| 329 | int err; | 331 | int err; |
| 330 | struct pt_regs * childregs; | 332 | struct pt_regs *childregs; |
| 331 | struct task_struct *me = current; | 333 | struct task_struct *me = current; |
| 332 | 334 | ||
| 333 | childregs = ((struct pt_regs *) | 335 | childregs = ((struct pt_regs *) |
| @@ -372,10 +374,10 @@ int copy_thread(int nr, unsigned long clone_flags, unsigned long sp, | |||
| 372 | if (test_thread_flag(TIF_IA32)) | 374 | if (test_thread_flag(TIF_IA32)) |
| 373 | err = do_set_thread_area(p, -1, | 375 | err = do_set_thread_area(p, -1, |
| 374 | (struct user_desc __user *)childregs->si, 0); | 376 | (struct user_desc __user *)childregs->si, 0); |
| 375 | else | 377 | else |
| 376 | #endif | 378 | #endif |
| 377 | err = do_arch_prctl(p, ARCH_SET_FS, childregs->r8); | 379 | err = do_arch_prctl(p, ARCH_SET_FS, childregs->r8); |
| 378 | if (err) | 380 | if (err) |
| 379 | goto out; | 381 | goto out; |
| 380 | } | 382 | } |
| 381 | err = 0; | 383 | err = 0; |
| @@ -568,7 +570,7 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) | |||
| 568 | unsigned fsindex, gsindex; | 570 | unsigned fsindex, gsindex; |
| 569 | 571 | ||
| 570 | /* we're going to use this soon, after a few expensive things */ | 572 | /* we're going to use this soon, after a few expensive things */ |
| 571 | if (next_p->fpu_counter>5) | 573 | if (next_p->fpu_counter > 5) |
| 572 | prefetch(next->xstate); | 574 | prefetch(next->xstate); |
| 573 | 575 | ||
| 574 | /* | 576 | /* |
| @@ -576,13 +578,13 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) | |||
| 576 | */ | 578 | */ |
| 577 | load_sp0(tss, next); | 579 | load_sp0(tss, next); |
| 578 | 580 | ||
| 579 | /* | 581 | /* |
| 580 | * Switch DS and ES. | 582 | * Switch DS and ES. |
| 581 | * This won't pick up thread selector changes, but I guess that is ok. | 583 | * This won't pick up thread selector changes, but I guess that is ok. |
| 582 | */ | 584 | */ |
| 583 | savesegment(es, prev->es); | 585 | savesegment(es, prev->es); |
| 584 | if (unlikely(next->es | prev->es)) | 586 | if (unlikely(next->es | prev->es)) |
| 585 | loadsegment(es, next->es); | 587 | loadsegment(es, next->es); |
| 586 | 588 | ||
| 587 | savesegment(ds, prev->ds); | 589 | savesegment(ds, prev->ds); |
| 588 | if (unlikely(next->ds | prev->ds)) | 590 | if (unlikely(next->ds | prev->ds)) |
| @@ -608,7 +610,7 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) | |||
| 608 | */ | 610 | */ |
| 609 | arch_leave_lazy_cpu_mode(); | 611 | arch_leave_lazy_cpu_mode(); |
| 610 | 612 | ||
| 611 | /* | 613 | /* |
| 612 | * Switch FS and GS. | 614 | * Switch FS and GS. |
| 613 | * | 615 | * |
| 614 | * Segment register != 0 always requires a reload. Also | 616 | * Segment register != 0 always requires a reload. Also |
| @@ -617,13 +619,13 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) | |||
| 617 | */ | 619 | */ |
| 618 | if (unlikely(fsindex | next->fsindex | prev->fs)) { | 620 | if (unlikely(fsindex | next->fsindex | prev->fs)) { |
| 619 | loadsegment(fs, next->fsindex); | 621 | loadsegment(fs, next->fsindex); |
| 620 | /* | 622 | /* |
| 621 | * Check if the user used a selector != 0; if yes | 623 | * Check if the user used a selector != 0; if yes |
| 622 | * clear 64bit base, since overloaded base is always | 624 | * clear 64bit base, since overloaded base is always |
| 623 | * mapped to the Null selector | 625 | * mapped to the Null selector |
| 624 | */ | 626 | */ |
| 625 | if (fsindex) | 627 | if (fsindex) |
| 626 | prev->fs = 0; | 628 | prev->fs = 0; |
| 627 | } | 629 | } |
| 628 | /* when next process has a 64bit base use it */ | 630 | /* when next process has a 64bit base use it */ |
| 629 | if (next->fs) | 631 | if (next->fs) |
| @@ -633,7 +635,7 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) | |||
| 633 | if (unlikely(gsindex | next->gsindex | prev->gs)) { | 635 | if (unlikely(gsindex | next->gsindex | prev->gs)) { |
| 634 | load_gs_index(next->gsindex); | 636 | load_gs_index(next->gsindex); |
| 635 | if (gsindex) | 637 | if (gsindex) |
| 636 | prev->gs = 0; | 638 | prev->gs = 0; |
| 637 | } | 639 | } |
| 638 | if (next->gs) | 640 | if (next->gs) |
| 639 | wrmsrl(MSR_KERNEL_GS_BASE, next->gs); | 641 | wrmsrl(MSR_KERNEL_GS_BASE, next->gs); |
| @@ -642,12 +644,12 @@ __switch_to(struct task_struct *prev_p, struct task_struct *next_p) | |||
| 642 | /* Must be after DS reload */ | 644 | /* Must be after DS reload */ |
| 643 | unlazy_fpu(prev_p); | 645 | unlazy_fpu(prev_p); |
| 644 | 646 | ||
| 645 | /* | 647 | /* |
| 646 | * Switch the PDA and FPU contexts. | 648 | * Switch the PDA and FPU contexts. |
| 647 | */ | 649 | */ |
| 648 | prev->usersp = read_pda(oldrsp); | 650 | prev->usersp = read_pda(oldrsp); |
| 649 | write_pda(oldrsp, next->usersp); | 651 | write_pda(oldrsp, next->usersp); |
| 650 | write_pda(pcurrent, next_p); | 652 | write_pda(pcurrent, next_p); |
| 651 | 653 | ||
| 652 | write_pda(kernelstack, | 654 | write_pda(kernelstack, |
| 653 | (unsigned long)task_stack_page(next_p) + | 655 | (unsigned long)task_stack_page(next_p) + |
| @@ -688,7 +690,7 @@ long sys_execve(char __user *name, char __user * __user *argv, | |||
| 688 | char __user * __user *envp, struct pt_regs *regs) | 690 | char __user * __user *envp, struct pt_regs *regs) |
| 689 | { | 691 | { |
| 690 | long error; | 692 | long error; |
| 691 | char * filename; | 693 | char *filename; |
| 692 | 694 | ||
| 693 | filename = getname(name); | 695 | filename = getname(name); |
| 694 | error = PTR_ERR(filename); | 696 | error = PTR_ERR(filename); |
| @@ -746,55 +748,55 @@ asmlinkage long sys_vfork(struct pt_regs *regs) | |||
| 746 | unsigned long get_wchan(struct task_struct *p) | 748 | unsigned long get_wchan(struct task_struct *p) |
| 747 | { | 749 | { |
| 748 | unsigned long stack; | 750 | unsigned long stack; |
| 749 | u64 fp,ip; | 751 | u64 fp, ip; |
| 750 | int count = 0; | 752 | int count = 0; |
| 751 | 753 | ||
| 752 | if (!p || p == current || p->state==TASK_RUNNING) | 754 | if (!p || p == current || p->state == TASK_RUNNING) |
| 753 | return 0; | 755 | return 0; |
| 754 | stack = (unsigned long)task_stack_page(p); | 756 | stack = (unsigned long)task_stack_page(p); |
| 755 | if (p->thread.sp < stack || p->thread.sp > stack+THREAD_SIZE) | 757 | if (p->thread.sp < stack || p->thread.sp > stack+THREAD_SIZE) |
| 756 | return 0; | 758 | return 0; |
| 757 | fp = *(u64 *)(p->thread.sp); | 759 | fp = *(u64 *)(p->thread.sp); |
| 758 | do { | 760 | do { |
| 759 | if (fp < (unsigned long)stack || | 761 | if (fp < (unsigned long)stack || |
| 760 | fp > (unsigned long)stack+THREAD_SIZE) | 762 | fp > (unsigned long)stack+THREAD_SIZE) |
| 761 | return 0; | 763 | return 0; |
| 762 | ip = *(u64 *)(fp+8); | 764 | ip = *(u64 *)(fp+8); |
| 763 | if (!in_sched_functions(ip)) | 765 | if (!in_sched_functions(ip)) |
| 764 | return ip; | 766 | return ip; |
| 765 | fp = *(u64 *)fp; | 767 | fp = *(u64 *)fp; |
| 766 | } while (count++ < 16); | 768 | } while (count++ < 16); |
| 767 | return 0; | 769 | return 0; |
| 768 | } | 770 | } |
| 769 | 771 | ||
| 770 | long do_arch_prctl(struct task_struct *task, int code, unsigned long addr) | 772 | long do_arch_prctl(struct task_struct *task, int code, unsigned long addr) |
| 771 | { | 773 | { |
| 772 | int ret = 0; | 774 | int ret = 0; |
| 773 | int doit = task == current; | 775 | int doit = task == current; |
| 774 | int cpu; | 776 | int cpu; |
| 775 | 777 | ||
| 776 | switch (code) { | 778 | switch (code) { |
| 777 | case ARCH_SET_GS: | 779 | case ARCH_SET_GS: |
| 778 | if (addr >= TASK_SIZE_OF(task)) | 780 | if (addr >= TASK_SIZE_OF(task)) |
| 779 | return -EPERM; | 781 | return -EPERM; |
| 780 | cpu = get_cpu(); | 782 | cpu = get_cpu(); |
| 781 | /* handle small bases via the GDT because that's faster to | 783 | /* handle small bases via the GDT because that's faster to |
| 782 | switch. */ | 784 | switch. */ |
| 783 | if (addr <= 0xffffffff) { | 785 | if (addr <= 0xffffffff) { |
| 784 | set_32bit_tls(task, GS_TLS, addr); | 786 | set_32bit_tls(task, GS_TLS, addr); |
| 785 | if (doit) { | 787 | if (doit) { |
| 786 | load_TLS(&task->thread, cpu); | 788 | load_TLS(&task->thread, cpu); |
| 787 | load_gs_index(GS_TLS_SEL); | 789 | load_gs_index(GS_TLS_SEL); |
| 788 | } | 790 | } |
| 789 | task->thread.gsindex = GS_TLS_SEL; | 791 | task->thread.gsindex = GS_TLS_SEL; |
| 790 | task->thread.gs = 0; | 792 | task->thread.gs = 0; |
| 791 | } else { | 793 | } else { |
| 792 | task->thread.gsindex = 0; | 794 | task->thread.gsindex = 0; |
| 793 | task->thread.gs = addr; | 795 | task->thread.gs = addr; |
| 794 | if (doit) { | 796 | if (doit) { |
| 795 | load_gs_index(0); | 797 | load_gs_index(0); |
| 796 | ret = checking_wrmsrl(MSR_KERNEL_GS_BASE, addr); | 798 | ret = checking_wrmsrl(MSR_KERNEL_GS_BASE, addr); |
| 797 | } | 799 | } |
| 798 | } | 800 | } |
| 799 | put_cpu(); | 801 | put_cpu(); |
| 800 | break; | 802 | break; |
| @@ -848,8 +850,7 @@ long do_arch_prctl(struct task_struct *task, int code, unsigned long addr) | |||
| 848 | rdmsrl(MSR_KERNEL_GS_BASE, base); | 850 | rdmsrl(MSR_KERNEL_GS_BASE, base); |
| 849 | else | 851 | else |
| 850 | base = task->thread.gs; | 852 | base = task->thread.gs; |
| 851 | } | 853 | } else |
| 852 | else | ||
| 853 | base = task->thread.gs; | 854 | base = task->thread.gs; |
| 854 | ret = put_user(base, (unsigned long __user *)addr); | 855 | ret = put_user(base, (unsigned long __user *)addr); |
| 855 | break; | 856 | break; |
diff --git a/arch/x86/kernel/reboot.c b/arch/x86/kernel/reboot.c index 724adfc63cb9..f4c93f1cfc19 100644 --- a/arch/x86/kernel/reboot.c +++ b/arch/x86/kernel/reboot.c | |||
| @@ -29,7 +29,11 @@ EXPORT_SYMBOL(pm_power_off); | |||
| 29 | 29 | ||
| 30 | static const struct desc_ptr no_idt = {}; | 30 | static const struct desc_ptr no_idt = {}; |
| 31 | static int reboot_mode; | 31 | static int reboot_mode; |
| 32 | enum reboot_type reboot_type = BOOT_KBD; | 32 | /* |
| 33 | * Keyboard reset and triple fault may result in INIT, not RESET, which | ||
| 34 | * doesn't work when we're in vmx root mode. Try ACPI first. | ||
| 35 | */ | ||
| 36 | enum reboot_type reboot_type = BOOT_ACPI; | ||
| 33 | int reboot_force; | 37 | int reboot_force; |
| 34 | 38 | ||
| 35 | #if defined(CONFIG_X86_32) && defined(CONFIG_SMP) | 39 | #if defined(CONFIG_X86_32) && defined(CONFIG_SMP) |
diff --git a/arch/x86/kernel/setup.c b/arch/x86/kernel/setup.c index 9838f2539dfc..141efab52400 100644 --- a/arch/x86/kernel/setup.c +++ b/arch/x86/kernel/setup.c | |||
| @@ -223,6 +223,9 @@ unsigned long saved_video_mode; | |||
| 223 | #define RAMDISK_LOAD_FLAG 0x4000 | 223 | #define RAMDISK_LOAD_FLAG 0x4000 |
| 224 | 224 | ||
| 225 | static char __initdata command_line[COMMAND_LINE_SIZE]; | 225 | static char __initdata command_line[COMMAND_LINE_SIZE]; |
| 226 | #ifdef CONFIG_CMDLINE_BOOL | ||
| 227 | static char __initdata builtin_cmdline[COMMAND_LINE_SIZE] = CONFIG_CMDLINE; | ||
| 228 | #endif | ||
| 226 | 229 | ||
| 227 | #if defined(CONFIG_EDD) || defined(CONFIG_EDD_MODULE) | 230 | #if defined(CONFIG_EDD) || defined(CONFIG_EDD_MODULE) |
| 228 | struct edd edd; | 231 | struct edd edd; |
| @@ -665,6 +668,19 @@ void __init setup_arch(char **cmdline_p) | |||
| 665 | bss_resource.start = virt_to_phys(&__bss_start); | 668 | bss_resource.start = virt_to_phys(&__bss_start); |
| 666 | bss_resource.end = virt_to_phys(&__bss_stop)-1; | 669 | bss_resource.end = virt_to_phys(&__bss_stop)-1; |
| 667 | 670 | ||
| 671 | #ifdef CONFIG_CMDLINE_BOOL | ||
| 672 | #ifdef CONFIG_CMDLINE_OVERRIDE | ||
| 673 | strlcpy(boot_command_line, builtin_cmdline, COMMAND_LINE_SIZE); | ||
| 674 | #else | ||
| 675 | if (builtin_cmdline[0]) { | ||
| 676 | /* append boot loader cmdline to builtin */ | ||
| 677 | strlcat(builtin_cmdline, " ", COMMAND_LINE_SIZE); | ||
| 678 | strlcat(builtin_cmdline, boot_command_line, COMMAND_LINE_SIZE); | ||
| 679 | strlcpy(boot_command_line, builtin_cmdline, COMMAND_LINE_SIZE); | ||
| 680 | } | ||
| 681 | #endif | ||
| 682 | #endif | ||
| 683 | |||
| 668 | strlcpy(command_line, boot_command_line, COMMAND_LINE_SIZE); | 684 | strlcpy(command_line, boot_command_line, COMMAND_LINE_SIZE); |
| 669 | *cmdline_p = command_line; | 685 | *cmdline_p = command_line; |
| 670 | 686 | ||
diff --git a/arch/x86/kernel/setup_percpu.c b/arch/x86/kernel/setup_percpu.c index 76e305e064f9..0e67f72d9316 100644 --- a/arch/x86/kernel/setup_percpu.c +++ b/arch/x86/kernel/setup_percpu.c | |||
| @@ -162,9 +162,16 @@ void __init setup_per_cpu_areas(void) | |||
| 162 | printk(KERN_INFO | 162 | printk(KERN_INFO |
| 163 | "cpu %d has no node %d or node-local memory\n", | 163 | "cpu %d has no node %d or node-local memory\n", |
| 164 | cpu, node); | 164 | cpu, node); |
| 165 | if (ptr) | ||
| 166 | printk(KERN_DEBUG "per cpu data for cpu%d at %016lx\n", | ||
| 167 | cpu, __pa(ptr)); | ||
| 165 | } | 168 | } |
| 166 | else | 169 | else { |
| 167 | ptr = alloc_bootmem_pages_node(NODE_DATA(node), size); | 170 | ptr = alloc_bootmem_pages_node(NODE_DATA(node), size); |
| 171 | if (ptr) | ||
| 172 | printk(KERN_DEBUG "per cpu data for cpu%d on node%d at %016lx\n", | ||
| 173 | cpu, node, __pa(ptr)); | ||
| 174 | } | ||
| 168 | #endif | 175 | #endif |
| 169 | per_cpu_offset(cpu) = ptr - __per_cpu_start; | 176 | per_cpu_offset(cpu) = ptr - __per_cpu_start; |
| 170 | memcpy(ptr, __per_cpu_start, __per_cpu_end - __per_cpu_start); | 177 | memcpy(ptr, __per_cpu_start, __per_cpu_end - __per_cpu_start); |
diff --git a/arch/x86/kernel/sigframe.h b/arch/x86/kernel/sigframe.h index 72bbb519d2dc..8b4956e800ac 100644 --- a/arch/x86/kernel/sigframe.h +++ b/arch/x86/kernel/sigframe.h | |||
| @@ -24,4 +24,9 @@ struct rt_sigframe { | |||
| 24 | struct ucontext uc; | 24 | struct ucontext uc; |
| 25 | struct siginfo info; | 25 | struct siginfo info; |
| 26 | }; | 26 | }; |
| 27 | |||
| 28 | int ia32_setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info, | ||
| 29 | sigset_t *set, struct pt_regs *regs); | ||
| 30 | int ia32_setup_frame(int sig, struct k_sigaction *ka, | ||
| 31 | sigset_t *set, struct pt_regs *regs); | ||
| 27 | #endif | 32 | #endif |
diff --git a/arch/x86/kernel/signal_64.c b/arch/x86/kernel/signal_64.c index 4d32487805ef..694aa888bb19 100644 --- a/arch/x86/kernel/signal_64.c +++ b/arch/x86/kernel/signal_64.c | |||
| @@ -20,9 +20,10 @@ | |||
| 20 | #include <linux/stddef.h> | 20 | #include <linux/stddef.h> |
| 21 | #include <linux/personality.h> | 21 | #include <linux/personality.h> |
| 22 | #include <linux/compiler.h> | 22 | #include <linux/compiler.h> |
| 23 | #include <linux/uaccess.h> | ||
| 24 | |||
| 23 | #include <asm/processor.h> | 25 | #include <asm/processor.h> |
| 24 | #include <asm/ucontext.h> | 26 | #include <asm/ucontext.h> |
| 25 | #include <asm/uaccess.h> | ||
| 26 | #include <asm/i387.h> | 27 | #include <asm/i387.h> |
| 27 | #include <asm/proto.h> | 28 | #include <asm/proto.h> |
| 28 | #include <asm/ia32_unistd.h> | 29 | #include <asm/ia32_unistd.h> |
| @@ -44,11 +45,6 @@ | |||
| 44 | # define FIX_EFLAGS __FIX_EFLAGS | 45 | # define FIX_EFLAGS __FIX_EFLAGS |
| 45 | #endif | 46 | #endif |
| 46 | 47 | ||
| 47 | int ia32_setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info, | ||
| 48 | sigset_t *set, struct pt_regs * regs); | ||
| 49 | int ia32_setup_frame(int sig, struct k_sigaction *ka, | ||
| 50 | sigset_t *set, struct pt_regs * regs); | ||
| 51 | |||
| 52 | asmlinkage long | 48 | asmlinkage long |
| 53 | sys_sigaltstack(const stack_t __user *uss, stack_t __user *uoss, | 49 | sys_sigaltstack(const stack_t __user *uss, stack_t __user *uoss, |
| 54 | struct pt_regs *regs) | 50 | struct pt_regs *regs) |
| @@ -131,7 +127,7 @@ restore_sigcontext(struct pt_regs *regs, struct sigcontext __user *sc, | |||
| 131 | /* Always make any pending restarted system calls return -EINTR */ | 127 | /* Always make any pending restarted system calls return -EINTR */ |
| 132 | current_thread_info()->restart_block.fn = do_no_restart_syscall; | 128 | current_thread_info()->restart_block.fn = do_no_restart_syscall; |
| 133 | 129 | ||
| 134 | #define COPY(x) err |= __get_user(regs->x, &sc->x) | 130 | #define COPY(x) (err |= __get_user(regs->x, &sc->x)) |
| 135 | 131 | ||
| 136 | COPY(di); COPY(si); COPY(bp); COPY(sp); COPY(bx); | 132 | COPY(di); COPY(si); COPY(bp); COPY(sp); COPY(bx); |
| 137 | COPY(dx); COPY(cx); COPY(ip); | 133 | COPY(dx); COPY(cx); COPY(ip); |
| @@ -161,7 +157,7 @@ restore_sigcontext(struct pt_regs *regs, struct sigcontext __user *sc, | |||
| 161 | } | 157 | } |
| 162 | 158 | ||
| 163 | { | 159 | { |
| 164 | struct _fpstate __user * buf; | 160 | struct _fpstate __user *buf; |
| 165 | err |= __get_user(buf, &sc->fpstate); | 161 | err |= __get_user(buf, &sc->fpstate); |
| 166 | 162 | ||
| 167 | if (buf) { | 163 | if (buf) { |
| @@ -201,7 +197,7 @@ asmlinkage long sys_rt_sigreturn(struct pt_regs *regs) | |||
| 201 | current->blocked = set; | 197 | current->blocked = set; |
| 202 | recalc_sigpending(); | 198 | recalc_sigpending(); |
| 203 | spin_unlock_irq(¤t->sighand->siglock); | 199 | spin_unlock_irq(¤t->sighand->siglock); |
| 204 | 200 | ||
| 205 | if (restore_sigcontext(regs, &frame->uc.uc_mcontext, &ax)) | 201 | if (restore_sigcontext(regs, &frame->uc.uc_mcontext, &ax)) |
| 206 | goto badframe; | 202 | goto badframe; |
| 207 | 203 | ||
| @@ -211,16 +207,17 @@ asmlinkage long sys_rt_sigreturn(struct pt_regs *regs) | |||
| 211 | return ax; | 207 | return ax; |
| 212 | 208 | ||
| 213 | badframe: | 209 | badframe: |
| 214 | signal_fault(regs,frame,"sigreturn"); | 210 | signal_fault(regs, frame, "sigreturn"); |
| 215 | return 0; | 211 | return 0; |
| 216 | } | 212 | } |
| 217 | 213 | ||
| 218 | /* | 214 | /* |
| 219 | * Set up a signal frame. | 215 | * Set up a signal frame. |
| 220 | */ | 216 | */ |
| 221 | 217 | ||
| 222 | static inline int | 218 | static inline int |
| 223 | setup_sigcontext(struct sigcontext __user *sc, struct pt_regs *regs, unsigned long mask, struct task_struct *me) | 219 | setup_sigcontext(struct sigcontext __user *sc, struct pt_regs *regs, |
| 220 | unsigned long mask, struct task_struct *me) | ||
| 224 | { | 221 | { |
| 225 | int err = 0; | 222 | int err = 0; |
| 226 | 223 | ||
| @@ -276,35 +273,35 @@ get_stack(struct k_sigaction *ka, struct pt_regs *regs, unsigned long size) | |||
| 276 | } | 273 | } |
| 277 | 274 | ||
| 278 | static int setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info, | 275 | static int setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info, |
| 279 | sigset_t *set, struct pt_regs * regs) | 276 | sigset_t *set, struct pt_regs *regs) |
| 280 | { | 277 | { |
| 281 | struct rt_sigframe __user *frame; | 278 | struct rt_sigframe __user *frame; |
| 282 | struct _fpstate __user *fp = NULL; | 279 | struct _fpstate __user *fp = NULL; |
| 283 | int err = 0; | 280 | int err = 0; |
| 284 | struct task_struct *me = current; | 281 | struct task_struct *me = current; |
| 285 | 282 | ||
| 286 | if (used_math()) { | 283 | if (used_math()) { |
| 287 | fp = get_stack(ka, regs, sizeof(struct _fpstate)); | 284 | fp = get_stack(ka, regs, sizeof(struct _fpstate)); |
| 288 | frame = (void __user *)round_down( | 285 | frame = (void __user *)round_down( |
| 289 | (unsigned long)fp - sizeof(struct rt_sigframe), 16) - 8; | 286 | (unsigned long)fp - sizeof(struct rt_sigframe), 16) - 8; |
| 290 | 287 | ||
| 291 | if (!access_ok(VERIFY_WRITE, fp, sizeof(struct _fpstate))) | 288 | if (!access_ok(VERIFY_WRITE, fp, sizeof(struct _fpstate))) |
| 292 | goto give_sigsegv; | 289 | goto give_sigsegv; |
| 293 | 290 | ||
| 294 | if (save_i387(fp) < 0) | 291 | if (save_i387(fp) < 0) |
| 295 | err |= -1; | 292 | err |= -1; |
| 296 | } else | 293 | } else |
| 297 | frame = get_stack(ka, regs, sizeof(struct rt_sigframe)) - 8; | 294 | frame = get_stack(ka, regs, sizeof(struct rt_sigframe)) - 8; |
| 298 | 295 | ||
| 299 | if (!access_ok(VERIFY_WRITE, frame, sizeof(*frame))) | 296 | if (!access_ok(VERIFY_WRITE, frame, sizeof(*frame))) |
| 300 | goto give_sigsegv; | 297 | goto give_sigsegv; |
| 301 | 298 | ||
| 302 | if (ka->sa.sa_flags & SA_SIGINFO) { | 299 | if (ka->sa.sa_flags & SA_SIGINFO) { |
| 303 | err |= copy_siginfo_to_user(&frame->info, info); | 300 | err |= copy_siginfo_to_user(&frame->info, info); |
| 304 | if (err) | 301 | if (err) |
| 305 | goto give_sigsegv; | 302 | goto give_sigsegv; |
| 306 | } | 303 | } |
| 307 | 304 | ||
| 308 | /* Create the ucontext. */ | 305 | /* Create the ucontext. */ |
| 309 | err |= __put_user(0, &frame->uc.uc_flags); | 306 | err |= __put_user(0, &frame->uc.uc_flags); |
| 310 | err |= __put_user(0, &frame->uc.uc_link); | 307 | err |= __put_user(0, &frame->uc.uc_link); |
| @@ -314,9 +311,9 @@ static int setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info, | |||
| 314 | err |= __put_user(me->sas_ss_size, &frame->uc.uc_stack.ss_size); | 311 | err |= __put_user(me->sas_ss_size, &frame->uc.uc_stack.ss_size); |
| 315 | err |= setup_sigcontext(&frame->uc.uc_mcontext, regs, set->sig[0], me); | 312 | err |= setup_sigcontext(&frame->uc.uc_mcontext, regs, set->sig[0], me); |
| 316 | err |= __put_user(fp, &frame->uc.uc_mcontext.fpstate); | 313 | err |= __put_user(fp, &frame->uc.uc_mcontext.fpstate); |
| 317 | if (sizeof(*set) == 16) { | 314 | if (sizeof(*set) == 16) { |
| 318 | __put_user(set->sig[0], &frame->uc.uc_sigmask.sig[0]); | 315 | __put_user(set->sig[0], &frame->uc.uc_sigmask.sig[0]); |
| 319 | __put_user(set->sig[1], &frame->uc.uc_sigmask.sig[1]); | 316 | __put_user(set->sig[1], &frame->uc.uc_sigmask.sig[1]); |
| 320 | } else | 317 | } else |
| 321 | err |= __copy_to_user(&frame->uc.uc_sigmask, set, sizeof(*set)); | 318 | err |= __copy_to_user(&frame->uc.uc_sigmask, set, sizeof(*set)); |
| 322 | 319 | ||
| @@ -327,7 +324,7 @@ static int setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info, | |||
| 327 | err |= __put_user(ka->sa.sa_restorer, &frame->pretcode); | 324 | err |= __put_user(ka->sa.sa_restorer, &frame->pretcode); |
| 328 | } else { | 325 | } else { |
| 329 | /* could use a vstub here */ | 326 | /* could use a vstub here */ |
| 330 | goto give_sigsegv; | 327 | goto give_sigsegv; |
| 331 | } | 328 | } |
| 332 | 329 | ||
| 333 | if (err) | 330 | if (err) |
| @@ -335,7 +332,7 @@ static int setup_rt_frame(int sig, struct k_sigaction *ka, siginfo_t *info, | |||
| 335 | 332 | ||
| 336 | /* Set up registers for signal handler */ | 333 | /* Set up registers for signal handler */ |
| 337 | regs->di = sig; | 334 | regs->di = sig; |
| 338 | /* In case the signal handler was declared without prototypes */ | 335 | /* In case the signal handler was declared without prototypes */ |
| 339 | regs->ax = 0; | 336 | regs->ax = 0; |
| 340 | 337 | ||
| 341 | /* This also works for non SA_SIGINFO handlers because they expect the | 338 | /* This also works for non SA_SIGINFO handlers because they expect the |
| @@ -359,7 +356,7 @@ give_sigsegv: | |||
| 359 | 356 | ||
| 360 | /* | 357 | /* |
| 361 | * OK, we're invoking a handler | 358 | * OK, we're invoking a handler |
| 362 | */ | 359 | */ |
| 363 | 360 | ||
| 364 | static int | 361 | static int |
| 365 | handle_signal(unsigned long sig, siginfo_t *info, struct k_sigaction *ka, | 362 | handle_signal(unsigned long sig, siginfo_t *info, struct k_sigaction *ka, |
| @@ -403,7 +400,7 @@ handle_signal(unsigned long sig, siginfo_t *info, struct k_sigaction *ka, | |||
| 403 | ret = ia32_setup_rt_frame(sig, ka, info, oldset, regs); | 400 | ret = ia32_setup_rt_frame(sig, ka, info, oldset, regs); |
| 404 | else | 401 | else |
| 405 | ret = ia32_setup_frame(sig, ka, oldset, regs); | 402 | ret = ia32_setup_frame(sig, ka, oldset, regs); |
| 406 | } else | 403 | } else |
| 407 | #endif | 404 | #endif |
| 408 | ret = setup_rt_frame(sig, ka, info, oldset, regs); | 405 | ret = setup_rt_frame(sig, ka, info, oldset, regs); |
| 409 | 406 | ||
| @@ -429,9 +426,9 @@ handle_signal(unsigned long sig, siginfo_t *info, struct k_sigaction *ka, | |||
| 429 | regs->flags &= ~X86_EFLAGS_TF; | 426 | regs->flags &= ~X86_EFLAGS_TF; |
| 430 | 427 | ||
| 431 | spin_lock_irq(¤t->sighand->siglock); | 428 | spin_lock_irq(¤t->sighand->siglock); |
| 432 | sigorsets(¤t->blocked,¤t->blocked,&ka->sa.sa_mask); | 429 | sigorsets(¤t->blocked, ¤t->blocked, &ka->sa.sa_mask); |
| 433 | if (!(ka->sa.sa_flags & SA_NODEFER)) | 430 | if (!(ka->sa.sa_flags & SA_NODEFER)) |
| 434 | sigaddset(¤t->blocked,sig); | 431 | sigaddset(¤t->blocked, sig); |
| 435 | recalc_sigpending(); | 432 | recalc_sigpending(); |
| 436 | spin_unlock_irq(¤t->sighand->siglock); | 433 | spin_unlock_irq(¤t->sighand->siglock); |
| 437 | 434 | ||
| @@ -541,14 +538,15 @@ void do_notify_resume(struct pt_regs *regs, void *unused, | |||
| 541 | } | 538 | } |
| 542 | 539 | ||
| 543 | void signal_fault(struct pt_regs *regs, void __user *frame, char *where) | 540 | void signal_fault(struct pt_regs *regs, void __user *frame, char *where) |
| 544 | { | 541 | { |
| 545 | struct task_struct *me = current; | 542 | struct task_struct *me = current; |
| 546 | if (show_unhandled_signals && printk_ratelimit()) { | 543 | if (show_unhandled_signals && printk_ratelimit()) { |
| 547 | printk("%s[%d] bad frame in %s frame:%p ip:%lx sp:%lx orax:%lx", | 544 | printk("%s[%d] bad frame in %s frame:%p ip:%lx sp:%lx orax:%lx", |
| 548 | me->comm,me->pid,where,frame,regs->ip,regs->sp,regs->orig_ax); | 545 | me->comm, me->pid, where, frame, regs->ip, |
| 546 | regs->sp, regs->orig_ax); | ||
| 549 | print_vma_addr(" in ", regs->ip); | 547 | print_vma_addr(" in ", regs->ip); |
| 550 | printk("\n"); | 548 | printk("\n"); |
| 551 | } | 549 | } |
| 552 | 550 | ||
| 553 | force_sig(SIGSEGV, me); | 551 | force_sig(SIGSEGV, me); |
| 554 | } | 552 | } |
diff --git a/arch/x86/kernel/smpboot.c b/arch/x86/kernel/smpboot.c index a66c93550a0d..45531e3ba194 100644 --- a/arch/x86/kernel/smpboot.c +++ b/arch/x86/kernel/smpboot.c | |||
| @@ -1313,16 +1313,13 @@ __init void prefill_possible_map(void) | |||
| 1313 | if (!num_processors) | 1313 | if (!num_processors) |
| 1314 | num_processors = 1; | 1314 | num_processors = 1; |
| 1315 | 1315 | ||
| 1316 | #ifdef CONFIG_HOTPLUG_CPU | ||
| 1317 | if (additional_cpus == -1) { | 1316 | if (additional_cpus == -1) { |
| 1318 | if (disabled_cpus > 0) | 1317 | if (disabled_cpus > 0) |
| 1319 | additional_cpus = disabled_cpus; | 1318 | additional_cpus = disabled_cpus; |
| 1320 | else | 1319 | else |
| 1321 | additional_cpus = 0; | 1320 | additional_cpus = 0; |
| 1322 | } | 1321 | } |
| 1323 | #else | 1322 | |
| 1324 | additional_cpus = 0; | ||
| 1325 | #endif | ||
| 1326 | possible = num_processors + additional_cpus; | 1323 | possible = num_processors + additional_cpus; |
| 1327 | if (possible > NR_CPUS) | 1324 | if (possible > NR_CPUS) |
| 1328 | possible = NR_CPUS; | 1325 | possible = NR_CPUS; |
diff --git a/arch/x86/kernel/sys_x86_64.c b/arch/x86/kernel/sys_x86_64.c index c9288c883e20..6bc211accf08 100644 --- a/arch/x86/kernel/sys_x86_64.c +++ b/arch/x86/kernel/sys_x86_64.c | |||
| @@ -13,16 +13,17 @@ | |||
| 13 | #include <linux/utsname.h> | 13 | #include <linux/utsname.h> |
| 14 | #include <linux/personality.h> | 14 | #include <linux/personality.h> |
| 15 | #include <linux/random.h> | 15 | #include <linux/random.h> |
| 16 | #include <linux/uaccess.h> | ||
| 16 | 17 | ||
| 17 | #include <asm/uaccess.h> | ||
| 18 | #include <asm/ia32.h> | 18 | #include <asm/ia32.h> |
| 19 | #include <asm/syscalls.h> | 19 | #include <asm/syscalls.h> |
| 20 | 20 | ||
| 21 | asmlinkage long sys_mmap(unsigned long addr, unsigned long len, unsigned long prot, unsigned long flags, | 21 | asmlinkage long sys_mmap(unsigned long addr, unsigned long len, |
| 22 | unsigned long fd, unsigned long off) | 22 | unsigned long prot, unsigned long flags, |
| 23 | unsigned long fd, unsigned long off) | ||
| 23 | { | 24 | { |
| 24 | long error; | 25 | long error; |
| 25 | struct file * file; | 26 | struct file *file; |
| 26 | 27 | ||
| 27 | error = -EINVAL; | 28 | error = -EINVAL; |
| 28 | if (off & ~PAGE_MASK) | 29 | if (off & ~PAGE_MASK) |
| @@ -57,9 +58,9 @@ static void find_start_end(unsigned long flags, unsigned long *begin, | |||
| 57 | unmapped base down for this case. This can give | 58 | unmapped base down for this case. This can give |
| 58 | conflicts with the heap, but we assume that glibc | 59 | conflicts with the heap, but we assume that glibc |
| 59 | malloc knows how to fall back to mmap. Give it 1GB | 60 | malloc knows how to fall back to mmap. Give it 1GB |
| 60 | of playground for now. -AK */ | 61 | of playground for now. -AK */ |
| 61 | *begin = 0x40000000; | 62 | *begin = 0x40000000; |
| 62 | *end = 0x80000000; | 63 | *end = 0x80000000; |
| 63 | if (current->flags & PF_RANDOMIZE) { | 64 | if (current->flags & PF_RANDOMIZE) { |
| 64 | new_begin = randomize_range(*begin, *begin + 0x02000000, 0); | 65 | new_begin = randomize_range(*begin, *begin + 0x02000000, 0); |
| 65 | if (new_begin) | 66 | if (new_begin) |
| @@ -67,9 +68,9 @@ static void find_start_end(unsigned long flags, unsigned long *begin, | |||
| 67 | } | 68 | } |
| 68 | } else { | 69 | } else { |
| 69 | *begin = TASK_UNMAPPED_BASE; | 70 | *begin = TASK_UNMAPPED_BASE; |
| 70 | *end = TASK_SIZE; | 71 | *end = TASK_SIZE; |
| 71 | } | 72 | } |
| 72 | } | 73 | } |
| 73 | 74 | ||
| 74 | unsigned long | 75 | unsigned long |
| 75 | arch_get_unmapped_area(struct file *filp, unsigned long addr, | 76 | arch_get_unmapped_area(struct file *filp, unsigned long addr, |
| @@ -79,11 +80,11 @@ arch_get_unmapped_area(struct file *filp, unsigned long addr, | |||
| 79 | struct vm_area_struct *vma; | 80 | struct vm_area_struct *vma; |
| 80 | unsigned long start_addr; | 81 | unsigned long start_addr; |
| 81 | unsigned long begin, end; | 82 | unsigned long begin, end; |
| 82 | 83 | ||
| 83 | if (flags & MAP_FIXED) | 84 | if (flags & MAP_FIXED) |
| 84 | return addr; | 85 | return addr; |
| 85 | 86 | ||
| 86 | find_start_end(flags, &begin, &end); | 87 | find_start_end(flags, &begin, &end); |
| 87 | 88 | ||
| 88 | if (len > end) | 89 | if (len > end) |
| 89 | return -ENOMEM; | 90 | return -ENOMEM; |
| @@ -97,12 +98,12 @@ arch_get_unmapped_area(struct file *filp, unsigned long addr, | |||
| 97 | } | 98 | } |
| 98 | if (((flags & MAP_32BIT) || test_thread_flag(TIF_IA32)) | 99 | if (((flags & MAP_32BIT) || test_thread_flag(TIF_IA32)) |
| 99 | && len <= mm->cached_hole_size) { | 100 | && len <= mm->cached_hole_size) { |
| 100 | mm->cached_hole_size = 0; | 101 | mm->cached_hole_size = 0; |
| 101 | mm->free_area_cache = begin; | 102 | mm->free_area_cache = begin; |
| 102 | } | 103 | } |
| 103 | addr = mm->free_area_cache; | 104 | addr = mm->free_area_cache; |
| 104 | if (addr < begin) | 105 | if (addr < begin) |
| 105 | addr = begin; | 106 | addr = begin; |
| 106 | start_addr = addr; | 107 | start_addr = addr; |
| 107 | 108 | ||
| 108 | full_search: | 109 | full_search: |
| @@ -128,7 +129,7 @@ full_search: | |||
| 128 | return addr; | 129 | return addr; |
| 129 | } | 130 | } |
| 130 | if (addr + mm->cached_hole_size < vma->vm_start) | 131 | if (addr + mm->cached_hole_size < vma->vm_start) |
| 131 | mm->cached_hole_size = vma->vm_start - addr; | 132 | mm->cached_hole_size = vma->vm_start - addr; |
| 132 | 133 | ||
| 133 | addr = vma->vm_end; | 134 | addr = vma->vm_end; |
| 134 | } | 135 | } |
| @@ -178,7 +179,7 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, | |||
| 178 | vma = find_vma(mm, addr-len); | 179 | vma = find_vma(mm, addr-len); |
| 179 | if (!vma || addr <= vma->vm_start) | 180 | if (!vma || addr <= vma->vm_start) |
| 180 | /* remember the address as a hint for next time */ | 181 | /* remember the address as a hint for next time */ |
| 181 | return (mm->free_area_cache = addr-len); | 182 | return mm->free_area_cache = addr-len; |
| 182 | } | 183 | } |
| 183 | 184 | ||
| 184 | if (mm->mmap_base < len) | 185 | if (mm->mmap_base < len) |
| @@ -195,7 +196,7 @@ arch_get_unmapped_area_topdown(struct file *filp, const unsigned long addr0, | |||
| 195 | vma = find_vma(mm, addr); | 196 | vma = find_vma(mm, addr); |
| 196 | if (!vma || addr+len <= vma->vm_start) | 197 | if (!vma || addr+len <= vma->vm_start) |
| 197 | /* remember the address as a hint for next time */ | 198 | /* remember the address as a hint for next time */ |
| 198 | return (mm->free_area_cache = addr); | 199 | return mm->free_area_cache = addr; |
| 199 | 200 | ||
| 200 | /* remember the largest hole we saw so far */ | 201 | /* remember the largest hole we saw so far */ |
| 201 | if (addr + mm->cached_hole_size < vma->vm_start) | 202 | if (addr + mm->cached_hole_size < vma->vm_start) |
| @@ -225,13 +226,13 @@ bottomup: | |||
| 225 | } | 226 | } |
| 226 | 227 | ||
| 227 | 228 | ||
| 228 | asmlinkage long sys_uname(struct new_utsname __user * name) | 229 | asmlinkage long sys_uname(struct new_utsname __user *name) |
| 229 | { | 230 | { |
| 230 | int err; | 231 | int err; |
| 231 | down_read(&uts_sem); | 232 | down_read(&uts_sem); |
| 232 | err = copy_to_user(name, utsname(), sizeof (*name)); | 233 | err = copy_to_user(name, utsname(), sizeof(*name)); |
| 233 | up_read(&uts_sem); | 234 | up_read(&uts_sem); |
| 234 | if (personality(current->personality) == PER_LINUX32) | 235 | if (personality(current->personality) == PER_LINUX32) |
| 235 | err |= copy_to_user(&name->machine, "i686", 5); | 236 | err |= copy_to_user(&name->machine, "i686", 5); |
| 236 | return err ? -EFAULT : 0; | 237 | return err ? -EFAULT : 0; |
| 237 | } | 238 | } |
diff --git a/arch/x86/kernel/traps_64.c b/arch/x86/kernel/traps_64.c index 513caaca7115..7a31f104bef9 100644 --- a/arch/x86/kernel/traps_64.c +++ b/arch/x86/kernel/traps_64.c | |||
| @@ -32,6 +32,8 @@ | |||
| 32 | #include <linux/bug.h> | 32 | #include <linux/bug.h> |
| 33 | #include <linux/nmi.h> | 33 | #include <linux/nmi.h> |
| 34 | #include <linux/mm.h> | 34 | #include <linux/mm.h> |
| 35 | #include <linux/smp.h> | ||
| 36 | #include <linux/io.h> | ||
| 35 | 37 | ||
| 36 | #if defined(CONFIG_EDAC) | 38 | #if defined(CONFIG_EDAC) |
| 37 | #include <linux/edac.h> | 39 | #include <linux/edac.h> |
| @@ -45,9 +47,6 @@ | |||
| 45 | #include <asm/unwind.h> | 47 | #include <asm/unwind.h> |
| 46 | #include <asm/desc.h> | 48 | #include <asm/desc.h> |
| 47 | #include <asm/i387.h> | 49 | #include <asm/i387.h> |
| 48 | #include <asm/nmi.h> | ||
| 49 | #include <asm/smp.h> | ||
| 50 | #include <asm/io.h> | ||
| 51 | #include <asm/pgalloc.h> | 50 | #include <asm/pgalloc.h> |
| 52 | #include <asm/proto.h> | 51 | #include <asm/proto.h> |
| 53 | #include <asm/pda.h> | 52 | #include <asm/pda.h> |
| @@ -85,7 +84,8 @@ static inline void preempt_conditional_cli(struct pt_regs *regs) | |||
| 85 | 84 | ||
| 86 | void printk_address(unsigned long address, int reliable) | 85 | void printk_address(unsigned long address, int reliable) |
| 87 | { | 86 | { |
| 88 | printk(" [<%016lx>] %s%pS\n", address, reliable ? "": "? ", (void *) address); | 87 | printk(" [<%016lx>] %s%pS\n", |
| 88 | address, reliable ? "" : "? ", (void *) address); | ||
| 89 | } | 89 | } |
| 90 | 90 | ||
| 91 | static unsigned long *in_exception_stack(unsigned cpu, unsigned long stack, | 91 | static unsigned long *in_exception_stack(unsigned cpu, unsigned long stack, |
| @@ -98,7 +98,8 @@ static unsigned long *in_exception_stack(unsigned cpu, unsigned long stack, | |||
| 98 | [STACKFAULT_STACK - 1] = "#SS", | 98 | [STACKFAULT_STACK - 1] = "#SS", |
| 99 | [MCE_STACK - 1] = "#MC", | 99 | [MCE_STACK - 1] = "#MC", |
| 100 | #if DEBUG_STKSZ > EXCEPTION_STKSZ | 100 | #if DEBUG_STKSZ > EXCEPTION_STKSZ |
| 101 | [N_EXCEPTION_STACKS ... N_EXCEPTION_STACKS + DEBUG_STKSZ / EXCEPTION_STKSZ - 2] = "#DB[?]" | 101 | [N_EXCEPTION_STACKS ... |
| 102 | N_EXCEPTION_STACKS + DEBUG_STKSZ / EXCEPTION_STKSZ - 2] = "#DB[?]" | ||
| 102 | #endif | 103 | #endif |
| 103 | }; | 104 | }; |
| 104 | unsigned k; | 105 | unsigned k; |
| @@ -163,7 +164,7 @@ static unsigned long *in_exception_stack(unsigned cpu, unsigned long stack, | |||
| 163 | } | 164 | } |
| 164 | 165 | ||
| 165 | /* | 166 | /* |
| 166 | * x86-64 can have up to three kernel stacks: | 167 | * x86-64 can have up to three kernel stacks: |
| 167 | * process stack | 168 | * process stack |
| 168 | * interrupt stack | 169 | * interrupt stack |
| 169 | * severe exception (double fault, nmi, stack fault, debug, mce) hardware stack | 170 | * severe exception (double fault, nmi, stack fault, debug, mce) hardware stack |
| @@ -219,7 +220,7 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs, | |||
| 219 | const struct stacktrace_ops *ops, void *data) | 220 | const struct stacktrace_ops *ops, void *data) |
| 220 | { | 221 | { |
| 221 | const unsigned cpu = get_cpu(); | 222 | const unsigned cpu = get_cpu(); |
| 222 | unsigned long *irqstack_end = (unsigned long*)cpu_pda(cpu)->irqstackptr; | 223 | unsigned long *irqstack_end = (unsigned long *)cpu_pda(cpu)->irqstackptr; |
| 223 | unsigned used = 0; | 224 | unsigned used = 0; |
| 224 | struct thread_info *tinfo; | 225 | struct thread_info *tinfo; |
| 225 | 226 | ||
| @@ -237,7 +238,7 @@ void dump_trace(struct task_struct *task, struct pt_regs *regs, | |||
| 237 | if (!bp) { | 238 | if (!bp) { |
| 238 | if (task == current) { | 239 | if (task == current) { |
| 239 | /* Grab bp right from our regs */ | 240 | /* Grab bp right from our regs */ |
| 240 | asm("movq %%rbp, %0" : "=r" (bp) :); | 241 | asm("movq %%rbp, %0" : "=r" (bp) : ); |
| 241 | } else { | 242 | } else { |
| 242 | /* bp is the last reg pushed by switch_to */ | 243 | /* bp is the last reg pushed by switch_to */ |
| 243 | bp = *(unsigned long *) task->thread.sp; | 244 | bp = *(unsigned long *) task->thread.sp; |
| @@ -339,9 +340,8 @@ static void | |||
| 339 | show_trace_log_lvl(struct task_struct *task, struct pt_regs *regs, | 340 | show_trace_log_lvl(struct task_struct *task, struct pt_regs *regs, |
| 340 | unsigned long *stack, unsigned long bp, char *log_lvl) | 341 | unsigned long *stack, unsigned long bp, char *log_lvl) |
| 341 | { | 342 | { |
| 342 | printk("\nCall Trace:\n"); | 343 | printk("Call Trace:\n"); |
| 343 | dump_trace(task, regs, stack, bp, &print_trace_ops, log_lvl); | 344 | dump_trace(task, regs, stack, bp, &print_trace_ops, log_lvl); |
| 344 | printk("\n"); | ||
| 345 | } | 345 | } |
| 346 | 346 | ||
| 347 | void show_trace(struct task_struct *task, struct pt_regs *regs, | 347 | void show_trace(struct task_struct *task, struct pt_regs *regs, |
| @@ -357,11 +357,15 @@ show_stack_log_lvl(struct task_struct *task, struct pt_regs *regs, | |||
| 357 | unsigned long *stack; | 357 | unsigned long *stack; |
| 358 | int i; | 358 | int i; |
| 359 | const int cpu = smp_processor_id(); | 359 | const int cpu = smp_processor_id(); |
| 360 | unsigned long *irqstack_end = (unsigned long *) (cpu_pda(cpu)->irqstackptr); | 360 | unsigned long *irqstack_end = |
| 361 | unsigned long *irqstack = (unsigned long *) (cpu_pda(cpu)->irqstackptr - IRQSTACKSIZE); | 361 | (unsigned long *) (cpu_pda(cpu)->irqstackptr); |
| 362 | unsigned long *irqstack = | ||
| 363 | (unsigned long *) (cpu_pda(cpu)->irqstackptr - IRQSTACKSIZE); | ||
| 362 | 364 | ||
| 363 | // debugging aid: "show_stack(NULL, NULL);" prints the | 365 | /* |
| 364 | // back trace for this cpu. | 366 | * debugging aid: "show_stack(NULL, NULL);" prints the |
| 367 | * back trace for this cpu. | ||
| 368 | */ | ||
| 365 | 369 | ||
| 366 | if (sp == NULL) { | 370 | if (sp == NULL) { |
| 367 | if (task) | 371 | if (task) |
| @@ -386,6 +390,7 @@ show_stack_log_lvl(struct task_struct *task, struct pt_regs *regs, | |||
| 386 | printk(" %016lx", *stack++); | 390 | printk(" %016lx", *stack++); |
| 387 | touch_nmi_watchdog(); | 391 | touch_nmi_watchdog(); |
| 388 | } | 392 | } |
| 393 | printk("\n"); | ||
| 389 | show_trace_log_lvl(task, regs, sp, bp, log_lvl); | 394 | show_trace_log_lvl(task, regs, sp, bp, log_lvl); |
| 390 | } | 395 | } |
| 391 | 396 | ||
| @@ -404,7 +409,7 @@ void dump_stack(void) | |||
| 404 | 409 | ||
| 405 | #ifdef CONFIG_FRAME_POINTER | 410 | #ifdef CONFIG_FRAME_POINTER |
| 406 | if (!bp) | 411 | if (!bp) |
| 407 | asm("movq %%rbp, %0" : "=r" (bp):); | 412 | asm("movq %%rbp, %0" : "=r" (bp) : ); |
| 408 | #endif | 413 | #endif |
| 409 | 414 | ||
| 410 | printk("Pid: %d, comm: %.20s %s %s %.*s\n", | 415 | printk("Pid: %d, comm: %.20s %s %s %.*s\n", |
| @@ -414,7 +419,6 @@ void dump_stack(void) | |||
| 414 | init_utsname()->version); | 419 | init_utsname()->version); |
| 415 | show_trace(NULL, NULL, &stack, bp); | 420 | show_trace(NULL, NULL, &stack, bp); |
| 416 | } | 421 | } |
| 417 | |||
| 418 | EXPORT_SYMBOL(dump_stack); | 422 | EXPORT_SYMBOL(dump_stack); |
| 419 | 423 | ||
| 420 | void show_registers(struct pt_regs *regs) | 424 | void show_registers(struct pt_regs *regs) |
| @@ -443,7 +447,6 @@ void show_registers(struct pt_regs *regs) | |||
| 443 | printk("Stack: "); | 447 | printk("Stack: "); |
| 444 | show_stack_log_lvl(NULL, regs, (unsigned long *)sp, | 448 | show_stack_log_lvl(NULL, regs, (unsigned long *)sp, |
| 445 | regs->bp, ""); | 449 | regs->bp, ""); |
| 446 | printk("\n"); | ||
| 447 | 450 | ||
| 448 | printk(KERN_EMERG "Code: "); | 451 | printk(KERN_EMERG "Code: "); |
| 449 | 452 | ||
| @@ -493,7 +496,7 @@ unsigned __kprobes long oops_begin(void) | |||
| 493 | raw_local_irq_save(flags); | 496 | raw_local_irq_save(flags); |
| 494 | cpu = smp_processor_id(); | 497 | cpu = smp_processor_id(); |
| 495 | if (!__raw_spin_trylock(&die_lock)) { | 498 | if (!__raw_spin_trylock(&die_lock)) { |
| 496 | if (cpu == die_owner) | 499 | if (cpu == die_owner) |
| 497 | /* nested oops. should stop eventually */; | 500 | /* nested oops. should stop eventually */; |
| 498 | else | 501 | else |
| 499 | __raw_spin_lock(&die_lock); | 502 | __raw_spin_lock(&die_lock); |
| @@ -638,7 +641,7 @@ kernel_trap: | |||
| 638 | } | 641 | } |
| 639 | 642 | ||
| 640 | #define DO_ERROR(trapnr, signr, str, name) \ | 643 | #define DO_ERROR(trapnr, signr, str, name) \ |
| 641 | asmlinkage void do_##name(struct pt_regs * regs, long error_code) \ | 644 | asmlinkage void do_##name(struct pt_regs *regs, long error_code) \ |
| 642 | { \ | 645 | { \ |
| 643 | if (notify_die(DIE_TRAP, str, regs, error_code, trapnr, signr) \ | 646 | if (notify_die(DIE_TRAP, str, regs, error_code, trapnr, signr) \ |
| 644 | == NOTIFY_STOP) \ | 647 | == NOTIFY_STOP) \ |
| @@ -648,7 +651,7 @@ asmlinkage void do_##name(struct pt_regs * regs, long error_code) \ | |||
| 648 | } | 651 | } |
| 649 | 652 | ||
| 650 | #define DO_ERROR_INFO(trapnr, signr, str, name, sicode, siaddr) \ | 653 | #define DO_ERROR_INFO(trapnr, signr, str, name, sicode, siaddr) \ |
| 651 | asmlinkage void do_##name(struct pt_regs * regs, long error_code) \ | 654 | asmlinkage void do_##name(struct pt_regs *regs, long error_code) \ |
| 652 | { \ | 655 | { \ |
| 653 | siginfo_t info; \ | 656 | siginfo_t info; \ |
| 654 | info.si_signo = signr; \ | 657 | info.si_signo = signr; \ |
| @@ -683,7 +686,7 @@ asmlinkage void do_stack_segment(struct pt_regs *regs, long error_code) | |||
| 683 | preempt_conditional_cli(regs); | 686 | preempt_conditional_cli(regs); |
| 684 | } | 687 | } |
| 685 | 688 | ||
| 686 | asmlinkage void do_double_fault(struct pt_regs * regs, long error_code) | 689 | asmlinkage void do_double_fault(struct pt_regs *regs, long error_code) |
| 687 | { | 690 | { |
| 688 | static const char str[] = "double fault"; | 691 | static const char str[] = "double fault"; |
| 689 | struct task_struct *tsk = current; | 692 | struct task_struct *tsk = current; |
| @@ -778,9 +781,10 @@ io_check_error(unsigned char reason, struct pt_regs *regs) | |||
| 778 | } | 781 | } |
| 779 | 782 | ||
| 780 | static notrace __kprobes void | 783 | static notrace __kprobes void |
| 781 | unknown_nmi_error(unsigned char reason, struct pt_regs * regs) | 784 | unknown_nmi_error(unsigned char reason, struct pt_regs *regs) |
| 782 | { | 785 | { |
| 783 | if (notify_die(DIE_NMIUNKNOWN, "nmi", regs, reason, 2, SIGINT) == NOTIFY_STOP) | 786 | if (notify_die(DIE_NMIUNKNOWN, "nmi", regs, reason, 2, SIGINT) == |
| 787 | NOTIFY_STOP) | ||
| 784 | return; | 788 | return; |
| 785 | printk(KERN_EMERG "Uhhuh. NMI received for unknown reason %02x.\n", | 789 | printk(KERN_EMERG "Uhhuh. NMI received for unknown reason %02x.\n", |
| 786 | reason); | 790 | reason); |
| @@ -882,7 +886,7 @@ asmlinkage __kprobes struct pt_regs *sync_regs(struct pt_regs *eregs) | |||
| 882 | else if (user_mode(eregs)) | 886 | else if (user_mode(eregs)) |
| 883 | regs = task_pt_regs(current); | 887 | regs = task_pt_regs(current); |
| 884 | /* Exception from kernel and interrupts are enabled. Move to | 888 | /* Exception from kernel and interrupts are enabled. Move to |
| 885 | kernel process stack. */ | 889 | kernel process stack. */ |
| 886 | else if (eregs->flags & X86_EFLAGS_IF) | 890 | else if (eregs->flags & X86_EFLAGS_IF) |
| 887 | regs = (struct pt_regs *)(eregs->sp -= sizeof(struct pt_regs)); | 891 | regs = (struct pt_regs *)(eregs->sp -= sizeof(struct pt_regs)); |
| 888 | if (eregs != regs) | 892 | if (eregs != regs) |
| @@ -891,7 +895,7 @@ asmlinkage __kprobes struct pt_regs *sync_regs(struct pt_regs *eregs) | |||
| 891 | } | 895 | } |
| 892 | 896 | ||
| 893 | /* runs on IST stack. */ | 897 | /* runs on IST stack. */ |
| 894 | asmlinkage void __kprobes do_debug(struct pt_regs * regs, | 898 | asmlinkage void __kprobes do_debug(struct pt_regs *regs, |
| 895 | unsigned long error_code) | 899 | unsigned long error_code) |
| 896 | { | 900 | { |
| 897 | struct task_struct *tsk = current; | 901 | struct task_struct *tsk = current; |
| @@ -1035,7 +1039,7 @@ asmlinkage void do_coprocessor_error(struct pt_regs *regs) | |||
| 1035 | 1039 | ||
| 1036 | asmlinkage void bad_intr(void) | 1040 | asmlinkage void bad_intr(void) |
| 1037 | { | 1041 | { |
| 1038 | printk("bad interrupt"); | 1042 | printk("bad interrupt"); |
| 1039 | } | 1043 | } |
| 1040 | 1044 | ||
| 1041 | asmlinkage void do_simd_coprocessor_error(struct pt_regs *regs) | 1045 | asmlinkage void do_simd_coprocessor_error(struct pt_regs *regs) |
| @@ -1047,7 +1051,7 @@ asmlinkage void do_simd_coprocessor_error(struct pt_regs *regs) | |||
| 1047 | 1051 | ||
| 1048 | conditional_sti(regs); | 1052 | conditional_sti(regs); |
| 1049 | if (!user_mode(regs) && | 1053 | if (!user_mode(regs) && |
| 1050 | kernel_math_error(regs, "kernel simd math error", 19)) | 1054 | kernel_math_error(regs, "kernel simd math error", 19)) |
| 1051 | return; | 1055 | return; |
| 1052 | 1056 | ||
| 1053 | /* | 1057 | /* |
| @@ -1092,7 +1096,7 @@ asmlinkage void do_simd_coprocessor_error(struct pt_regs *regs) | |||
| 1092 | force_sig_info(SIGFPE, &info, task); | 1096 | force_sig_info(SIGFPE, &info, task); |
| 1093 | } | 1097 | } |
| 1094 | 1098 | ||
| 1095 | asmlinkage void do_spurious_interrupt_bug(struct pt_regs * regs) | 1099 | asmlinkage void do_spurious_interrupt_bug(struct pt_regs *regs) |
| 1096 | { | 1100 | { |
| 1097 | } | 1101 | } |
| 1098 | 1102 | ||
| @@ -1149,8 +1153,10 @@ void __init trap_init(void) | |||
| 1149 | set_intr_gate(0, ÷_error); | 1153 | set_intr_gate(0, ÷_error); |
| 1150 | set_intr_gate_ist(1, &debug, DEBUG_STACK); | 1154 | set_intr_gate_ist(1, &debug, DEBUG_STACK); |
| 1151 | set_intr_gate_ist(2, &nmi, NMI_STACK); | 1155 | set_intr_gate_ist(2, &nmi, NMI_STACK); |
| 1152 | set_system_gate_ist(3, &int3, DEBUG_STACK); /* int3 can be called from all */ | 1156 | /* int3 can be called from all */ |
| 1153 | set_system_gate(4, &overflow); /* int4 can be called from all */ | 1157 | set_system_gate_ist(3, &int3, DEBUG_STACK); |
| 1158 | /* int4 can be called from all */ | ||
| 1159 | set_system_gate(4, &overflow); | ||
| 1154 | set_intr_gate(5, &bounds); | 1160 | set_intr_gate(5, &bounds); |
| 1155 | set_intr_gate(6, &invalid_op); | 1161 | set_intr_gate(6, &invalid_op); |
| 1156 | set_intr_gate(7, &device_not_available); | 1162 | set_intr_gate(7, &device_not_available); |
diff --git a/arch/x86/kernel/tsc.c b/arch/x86/kernel/tsc.c index 8f98e9de1b82..161bb850fc47 100644 --- a/arch/x86/kernel/tsc.c +++ b/arch/x86/kernel/tsc.c | |||
| @@ -104,7 +104,7 @@ __setup("notsc", notsc_setup); | |||
| 104 | /* | 104 | /* |
| 105 | * Read TSC and the reference counters. Take care of SMI disturbance | 105 | * Read TSC and the reference counters. Take care of SMI disturbance |
| 106 | */ | 106 | */ |
| 107 | static u64 tsc_read_refs(u64 *pm, u64 *hpet) | 107 | static u64 tsc_read_refs(u64 *p, int hpet) |
| 108 | { | 108 | { |
| 109 | u64 t1, t2; | 109 | u64 t1, t2; |
| 110 | int i; | 110 | int i; |
| @@ -112,9 +112,9 @@ static u64 tsc_read_refs(u64 *pm, u64 *hpet) | |||
| 112 | for (i = 0; i < MAX_RETRIES; i++) { | 112 | for (i = 0; i < MAX_RETRIES; i++) { |
| 113 | t1 = get_cycles(); | 113 | t1 = get_cycles(); |
| 114 | if (hpet) | 114 | if (hpet) |
| 115 | *hpet = hpet_readl(HPET_COUNTER) & 0xFFFFFFFF; | 115 | *p = hpet_readl(HPET_COUNTER) & 0xFFFFFFFF; |
| 116 | else | 116 | else |
| 117 | *pm = acpi_pm_read_early(); | 117 | *p = acpi_pm_read_early(); |
| 118 | t2 = get_cycles(); | 118 | t2 = get_cycles(); |
| 119 | if ((t2 - t1) < SMI_TRESHOLD) | 119 | if ((t2 - t1) < SMI_TRESHOLD) |
| 120 | return t2; | 120 | return t2; |
| @@ -123,13 +123,59 @@ static u64 tsc_read_refs(u64 *pm, u64 *hpet) | |||
| 123 | } | 123 | } |
| 124 | 124 | ||
| 125 | /* | 125 | /* |
| 126 | * Calculate the TSC frequency from HPET reference | ||
| 127 | */ | ||
| 128 | static unsigned long calc_hpet_ref(u64 deltatsc, u64 hpet1, u64 hpet2) | ||
| 129 | { | ||
| 130 | u64 tmp; | ||
| 131 | |||
| 132 | if (hpet2 < hpet1) | ||
| 133 | hpet2 += 0x100000000ULL; | ||
| 134 | hpet2 -= hpet1; | ||
| 135 | tmp = ((u64)hpet2 * hpet_readl(HPET_PERIOD)); | ||
| 136 | do_div(tmp, 1000000); | ||
| 137 | do_div(deltatsc, tmp); | ||
| 138 | |||
| 139 | return (unsigned long) deltatsc; | ||
| 140 | } | ||
| 141 | |||
| 142 | /* | ||
| 143 | * Calculate the TSC frequency from PMTimer reference | ||
| 144 | */ | ||
| 145 | static unsigned long calc_pmtimer_ref(u64 deltatsc, u64 pm1, u64 pm2) | ||
| 146 | { | ||
| 147 | u64 tmp; | ||
| 148 | |||
| 149 | if (!pm1 && !pm2) | ||
| 150 | return ULONG_MAX; | ||
| 151 | |||
| 152 | if (pm2 < pm1) | ||
| 153 | pm2 += (u64)ACPI_PM_OVRRUN; | ||
| 154 | pm2 -= pm1; | ||
| 155 | tmp = pm2 * 1000000000LL; | ||
| 156 | do_div(tmp, PMTMR_TICKS_PER_SEC); | ||
| 157 | do_div(deltatsc, tmp); | ||
| 158 | |||
| 159 | return (unsigned long) deltatsc; | ||
| 160 | } | ||
| 161 | |||
| 162 | #define CAL_MS 10 | ||
| 163 | #define CAL_LATCH (CLOCK_TICK_RATE / (1000 / CAL_MS)) | ||
| 164 | #define CAL_PIT_LOOPS 1000 | ||
| 165 | |||
| 166 | #define CAL2_MS 50 | ||
| 167 | #define CAL2_LATCH (CLOCK_TICK_RATE / (1000 / CAL2_MS)) | ||
| 168 | #define CAL2_PIT_LOOPS 5000 | ||
| 169 | |||
| 170 | |||
| 171 | /* | ||
| 126 | * Try to calibrate the TSC against the Programmable | 172 | * Try to calibrate the TSC against the Programmable |
| 127 | * Interrupt Timer and return the frequency of the TSC | 173 | * Interrupt Timer and return the frequency of the TSC |
| 128 | * in kHz. | 174 | * in kHz. |
| 129 | * | 175 | * |
| 130 | * Return ULONG_MAX on failure to calibrate. | 176 | * Return ULONG_MAX on failure to calibrate. |
| 131 | */ | 177 | */ |
| 132 | static unsigned long pit_calibrate_tsc(void) | 178 | static unsigned long pit_calibrate_tsc(u32 latch, unsigned long ms, int loopmin) |
| 133 | { | 179 | { |
| 134 | u64 tsc, t1, t2, delta; | 180 | u64 tsc, t1, t2, delta; |
| 135 | unsigned long tscmin, tscmax; | 181 | unsigned long tscmin, tscmax; |
| @@ -144,8 +190,8 @@ static unsigned long pit_calibrate_tsc(void) | |||
| 144 | * (LSB then MSB) to begin countdown. | 190 | * (LSB then MSB) to begin countdown. |
| 145 | */ | 191 | */ |
| 146 | outb(0xb0, 0x43); | 192 | outb(0xb0, 0x43); |
| 147 | outb((CLOCK_TICK_RATE / (1000 / 50)) & 0xff, 0x42); | 193 | outb(latch & 0xff, 0x42); |
| 148 | outb((CLOCK_TICK_RATE / (1000 / 50)) >> 8, 0x42); | 194 | outb(latch >> 8, 0x42); |
| 149 | 195 | ||
| 150 | tsc = t1 = t2 = get_cycles(); | 196 | tsc = t1 = t2 = get_cycles(); |
| 151 | 197 | ||
| @@ -166,31 +212,154 @@ static unsigned long pit_calibrate_tsc(void) | |||
| 166 | /* | 212 | /* |
| 167 | * Sanity checks: | 213 | * Sanity checks: |
| 168 | * | 214 | * |
| 169 | * If we were not able to read the PIT more than 5000 | 215 | * If we were not able to read the PIT more than loopmin |
| 170 | * times, then we have been hit by a massive SMI | 216 | * times, then we have been hit by a massive SMI |
| 171 | * | 217 | * |
| 172 | * If the maximum is 10 times larger than the minimum, | 218 | * If the maximum is 10 times larger than the minimum, |
| 173 | * then we got hit by an SMI as well. | 219 | * then we got hit by an SMI as well. |
| 174 | */ | 220 | */ |
| 175 | if (pitcnt < 5000 || tscmax > 10 * tscmin) | 221 | if (pitcnt < loopmin || tscmax > 10 * tscmin) |
| 176 | return ULONG_MAX; | 222 | return ULONG_MAX; |
| 177 | 223 | ||
| 178 | /* Calculate the PIT value */ | 224 | /* Calculate the PIT value */ |
| 179 | delta = t2 - t1; | 225 | delta = t2 - t1; |
| 180 | do_div(delta, 50); | 226 | do_div(delta, ms); |
| 181 | return delta; | 227 | return delta; |
| 182 | } | 228 | } |
| 183 | 229 | ||
| 230 | /* | ||
| 231 | * This reads the current MSB of the PIT counter, and | ||
| 232 | * checks if we are running on sufficiently fast and | ||
| 233 | * non-virtualized hardware. | ||
| 234 | * | ||
| 235 | * Our expectations are: | ||
| 236 | * | ||
| 237 | * - the PIT is running at roughly 1.19MHz | ||
| 238 | * | ||
| 239 | * - each IO is going to take about 1us on real hardware, | ||
| 240 | * but we allow it to be much faster (by a factor of 10) or | ||
| 241 | * _slightly_ slower (ie we allow up to a 2us read+counter | ||
| 242 | * update - anything else implies a unacceptably slow CPU | ||
| 243 | * or PIT for the fast calibration to work. | ||
| 244 | * | ||
| 245 | * - with 256 PIT ticks to read the value, we have 214us to | ||
| 246 | * see the same MSB (and overhead like doing a single TSC | ||
| 247 | * read per MSB value etc). | ||
| 248 | * | ||
| 249 | * - We're doing 2 reads per loop (LSB, MSB), and we expect | ||
| 250 | * them each to take about a microsecond on real hardware. | ||
| 251 | * So we expect a count value of around 100. But we'll be | ||
| 252 | * generous, and accept anything over 50. | ||
| 253 | * | ||
| 254 | * - if the PIT is stuck, and we see *many* more reads, we | ||
| 255 | * return early (and the next caller of pit_expect_msb() | ||
| 256 | * then consider it a failure when they don't see the | ||
| 257 | * next expected value). | ||
| 258 | * | ||
| 259 | * These expectations mean that we know that we have seen the | ||
| 260 | * transition from one expected value to another with a fairly | ||
| 261 | * high accuracy, and we didn't miss any events. We can thus | ||
| 262 | * use the TSC value at the transitions to calculate a pretty | ||
| 263 | * good value for the TSC frequencty. | ||
| 264 | */ | ||
| 265 | static inline int pit_expect_msb(unsigned char val) | ||
| 266 | { | ||
| 267 | int count = 0; | ||
| 268 | |||
| 269 | for (count = 0; count < 50000; count++) { | ||
| 270 | /* Ignore LSB */ | ||
| 271 | inb(0x42); | ||
| 272 | if (inb(0x42) != val) | ||
| 273 | break; | ||
| 274 | } | ||
| 275 | return count > 50; | ||
| 276 | } | ||
| 277 | |||
| 278 | /* | ||
| 279 | * How many MSB values do we want to see? We aim for a | ||
| 280 | * 15ms calibration, which assuming a 2us counter read | ||
| 281 | * error should give us roughly 150 ppm precision for | ||
| 282 | * the calibration. | ||
| 283 | */ | ||
| 284 | #define QUICK_PIT_MS 15 | ||
| 285 | #define QUICK_PIT_ITERATIONS (QUICK_PIT_MS * PIT_TICK_RATE / 1000 / 256) | ||
| 286 | |||
| 287 | static unsigned long quick_pit_calibrate(void) | ||
| 288 | { | ||
| 289 | /* Set the Gate high, disable speaker */ | ||
| 290 | outb((inb(0x61) & ~0x02) | 0x01, 0x61); | ||
| 291 | |||
| 292 | /* | ||
| 293 | * Counter 2, mode 0 (one-shot), binary count | ||
| 294 | * | ||
| 295 | * NOTE! Mode 2 decrements by two (and then the | ||
| 296 | * output is flipped each time, giving the same | ||
| 297 | * final output frequency as a decrement-by-one), | ||
| 298 | * so mode 0 is much better when looking at the | ||
| 299 | * individual counts. | ||
| 300 | */ | ||
| 301 | outb(0xb0, 0x43); | ||
| 302 | |||
| 303 | /* Start at 0xffff */ | ||
| 304 | outb(0xff, 0x42); | ||
| 305 | outb(0xff, 0x42); | ||
| 306 | |||
| 307 | if (pit_expect_msb(0xff)) { | ||
| 308 | int i; | ||
| 309 | u64 t1, t2, delta; | ||
| 310 | unsigned char expect = 0xfe; | ||
| 311 | |||
| 312 | t1 = get_cycles(); | ||
| 313 | for (i = 0; i < QUICK_PIT_ITERATIONS; i++, expect--) { | ||
| 314 | if (!pit_expect_msb(expect)) | ||
| 315 | goto failed; | ||
| 316 | } | ||
| 317 | t2 = get_cycles(); | ||
| 318 | |||
| 319 | /* | ||
| 320 | * Make sure we can rely on the second TSC timestamp: | ||
| 321 | */ | ||
| 322 | if (!pit_expect_msb(expect)) | ||
| 323 | goto failed; | ||
| 324 | |||
| 325 | /* | ||
| 326 | * Ok, if we get here, then we've seen the | ||
| 327 | * MSB of the PIT decrement QUICK_PIT_ITERATIONS | ||
| 328 | * times, and each MSB had many hits, so we never | ||
| 329 | * had any sudden jumps. | ||
| 330 | * | ||
| 331 | * As a result, we can depend on there not being | ||
| 332 | * any odd delays anywhere, and the TSC reads are | ||
| 333 | * reliable. | ||
| 334 | * | ||
| 335 | * kHz = ticks / time-in-seconds / 1000; | ||
| 336 | * kHz = (t2 - t1) / (QPI * 256 / PIT_TICK_RATE) / 1000 | ||
| 337 | * kHz = ((t2 - t1) * PIT_TICK_RATE) / (QPI * 256 * 1000) | ||
| 338 | */ | ||
| 339 | delta = (t2 - t1)*PIT_TICK_RATE; | ||
| 340 | do_div(delta, QUICK_PIT_ITERATIONS*256*1000); | ||
| 341 | printk("Fast TSC calibration using PIT\n"); | ||
| 342 | return delta; | ||
| 343 | } | ||
| 344 | failed: | ||
| 345 | return 0; | ||
| 346 | } | ||
| 184 | 347 | ||
| 185 | /** | 348 | /** |
| 186 | * native_calibrate_tsc - calibrate the tsc on boot | 349 | * native_calibrate_tsc - calibrate the tsc on boot |
| 187 | */ | 350 | */ |
| 188 | unsigned long native_calibrate_tsc(void) | 351 | unsigned long native_calibrate_tsc(void) |
| 189 | { | 352 | { |
| 190 | u64 tsc1, tsc2, delta, pm1, pm2, hpet1, hpet2; | 353 | u64 tsc1, tsc2, delta, ref1, ref2; |
| 191 | unsigned long tsc_pit_min = ULONG_MAX, tsc_ref_min = ULONG_MAX; | 354 | unsigned long tsc_pit_min = ULONG_MAX, tsc_ref_min = ULONG_MAX; |
| 192 | unsigned long flags; | 355 | unsigned long flags, latch, ms, fast_calibrate; |
| 193 | int hpet = is_hpet_enabled(), i; | 356 | int hpet = is_hpet_enabled(), i, loopmin; |
| 357 | |||
| 358 | local_irq_save(flags); | ||
| 359 | fast_calibrate = quick_pit_calibrate(); | ||
| 360 | local_irq_restore(flags); | ||
| 361 | if (fast_calibrate) | ||
| 362 | return fast_calibrate; | ||
| 194 | 363 | ||
| 195 | /* | 364 | /* |
| 196 | * Run 5 calibration loops to get the lowest frequency value | 365 | * Run 5 calibration loops to get the lowest frequency value |
| @@ -216,7 +385,13 @@ unsigned long native_calibrate_tsc(void) | |||
| 216 | * calibration delay loop as we have to wait for a certain | 385 | * calibration delay loop as we have to wait for a certain |
| 217 | * amount of time anyway. | 386 | * amount of time anyway. |
| 218 | */ | 387 | */ |
| 219 | for (i = 0; i < 5; i++) { | 388 | |
| 389 | /* Preset PIT loop values */ | ||
| 390 | latch = CAL_LATCH; | ||
| 391 | ms = CAL_MS; | ||
| 392 | loopmin = CAL_PIT_LOOPS; | ||
| 393 | |||
| 394 | for (i = 0; i < 3; i++) { | ||
| 220 | unsigned long tsc_pit_khz; | 395 | unsigned long tsc_pit_khz; |
| 221 | 396 | ||
| 222 | /* | 397 | /* |
| @@ -226,16 +401,16 @@ unsigned long native_calibrate_tsc(void) | |||
| 226 | * read the end value. | 401 | * read the end value. |
| 227 | */ | 402 | */ |
| 228 | local_irq_save(flags); | 403 | local_irq_save(flags); |
| 229 | tsc1 = tsc_read_refs(&pm1, hpet ? &hpet1 : NULL); | 404 | tsc1 = tsc_read_refs(&ref1, hpet); |
| 230 | tsc_pit_khz = pit_calibrate_tsc(); | 405 | tsc_pit_khz = pit_calibrate_tsc(latch, ms, loopmin); |
| 231 | tsc2 = tsc_read_refs(&pm2, hpet ? &hpet2 : NULL); | 406 | tsc2 = tsc_read_refs(&ref2, hpet); |
| 232 | local_irq_restore(flags); | 407 | local_irq_restore(flags); |
| 233 | 408 | ||
| 234 | /* Pick the lowest PIT TSC calibration so far */ | 409 | /* Pick the lowest PIT TSC calibration so far */ |
| 235 | tsc_pit_min = min(tsc_pit_min, tsc_pit_khz); | 410 | tsc_pit_min = min(tsc_pit_min, tsc_pit_khz); |
| 236 | 411 | ||
| 237 | /* hpet or pmtimer available ? */ | 412 | /* hpet or pmtimer available ? */ |
| 238 | if (!hpet && !pm1 && !pm2) | 413 | if (!hpet && !ref1 && !ref2) |
| 239 | continue; | 414 | continue; |
| 240 | 415 | ||
| 241 | /* Check, whether the sampling was disturbed by an SMI */ | 416 | /* Check, whether the sampling was disturbed by an SMI */ |
| @@ -243,23 +418,41 @@ unsigned long native_calibrate_tsc(void) | |||
| 243 | continue; | 418 | continue; |
| 244 | 419 | ||
| 245 | tsc2 = (tsc2 - tsc1) * 1000000LL; | 420 | tsc2 = (tsc2 - tsc1) * 1000000LL; |
| 421 | if (hpet) | ||
| 422 | tsc2 = calc_hpet_ref(tsc2, ref1, ref2); | ||
| 423 | else | ||
| 424 | tsc2 = calc_pmtimer_ref(tsc2, ref1, ref2); | ||
| 246 | 425 | ||
| 247 | if (hpet) { | 426 | tsc_ref_min = min(tsc_ref_min, (unsigned long) tsc2); |
| 248 | if (hpet2 < hpet1) | 427 | |
| 249 | hpet2 += 0x100000000ULL; | 428 | /* Check the reference deviation */ |
| 250 | hpet2 -= hpet1; | 429 | delta = ((u64) tsc_pit_min) * 100; |
| 251 | tsc1 = ((u64)hpet2 * hpet_readl(HPET_PERIOD)); | 430 | do_div(delta, tsc_ref_min); |
| 252 | do_div(tsc1, 1000000); | 431 | |
| 253 | } else { | 432 | /* |
| 254 | if (pm2 < pm1) | 433 | * If both calibration results are inside a 10% window |
| 255 | pm2 += (u64)ACPI_PM_OVRRUN; | 434 | * then we can be sure, that the calibration |
| 256 | pm2 -= pm1; | 435 | * succeeded. We break out of the loop right away. We |
| 257 | tsc1 = pm2 * 1000000000LL; | 436 | * use the reference value, as it is more precise. |
| 258 | do_div(tsc1, PMTMR_TICKS_PER_SEC); | 437 | */ |
| 438 | if (delta >= 90 && delta <= 110) { | ||
| 439 | printk(KERN_INFO | ||
| 440 | "TSC: PIT calibration matches %s. %d loops\n", | ||
| 441 | hpet ? "HPET" : "PMTIMER", i + 1); | ||
| 442 | return tsc_ref_min; | ||
| 259 | } | 443 | } |
| 260 | 444 | ||
| 261 | do_div(tsc2, tsc1); | 445 | /* |
| 262 | tsc_ref_min = min(tsc_ref_min, (unsigned long) tsc2); | 446 | * Check whether PIT failed more than once. This |
| 447 | * happens in virtualized environments. We need to | ||
| 448 | * give the virtual PC a slightly longer timeframe for | ||
| 449 | * the HPET/PMTIMER to make the result precise. | ||
| 450 | */ | ||
| 451 | if (i == 1 && tsc_pit_min == ULONG_MAX) { | ||
| 452 | latch = CAL2_LATCH; | ||
| 453 | ms = CAL2_MS; | ||
| 454 | loopmin = CAL2_PIT_LOOPS; | ||
| 455 | } | ||
| 263 | } | 456 | } |
| 264 | 457 | ||
| 265 | /* | 458 | /* |
| @@ -270,7 +463,7 @@ unsigned long native_calibrate_tsc(void) | |||
| 270 | printk(KERN_WARNING "TSC: Unable to calibrate against PIT\n"); | 463 | printk(KERN_WARNING "TSC: Unable to calibrate against PIT\n"); |
| 271 | 464 | ||
| 272 | /* We don't have an alternative source, disable TSC */ | 465 | /* We don't have an alternative source, disable TSC */ |
| 273 | if (!hpet && !pm1 && !pm2) { | 466 | if (!hpet && !ref1 && !ref2) { |
| 274 | printk("TSC: No reference (HPET/PMTIMER) available\n"); | 467 | printk("TSC: No reference (HPET/PMTIMER) available\n"); |
| 275 | return 0; | 468 | return 0; |
| 276 | } | 469 | } |
| @@ -278,7 +471,7 @@ unsigned long native_calibrate_tsc(void) | |||
| 278 | /* The alternative source failed as well, disable TSC */ | 471 | /* The alternative source failed as well, disable TSC */ |
| 279 | if (tsc_ref_min == ULONG_MAX) { | 472 | if (tsc_ref_min == ULONG_MAX) { |
| 280 | printk(KERN_WARNING "TSC: HPET/PMTIMER calibration " | 473 | printk(KERN_WARNING "TSC: HPET/PMTIMER calibration " |
| 281 | "failed due to SMI disturbance.\n"); | 474 | "failed.\n"); |
| 282 | return 0; | 475 | return 0; |
| 283 | } | 476 | } |
| 284 | 477 | ||
| @@ -290,44 +483,25 @@ unsigned long native_calibrate_tsc(void) | |||
| 290 | } | 483 | } |
| 291 | 484 | ||
| 292 | /* We don't have an alternative source, use the PIT calibration value */ | 485 | /* We don't have an alternative source, use the PIT calibration value */ |
| 293 | if (!hpet && !pm1 && !pm2) { | 486 | if (!hpet && !ref1 && !ref2) { |
| 294 | printk(KERN_INFO "TSC: Using PIT calibration value\n"); | 487 | printk(KERN_INFO "TSC: Using PIT calibration value\n"); |
| 295 | return tsc_pit_min; | 488 | return tsc_pit_min; |
| 296 | } | 489 | } |
| 297 | 490 | ||
| 298 | /* The alternative source failed, use the PIT calibration value */ | 491 | /* The alternative source failed, use the PIT calibration value */ |
| 299 | if (tsc_ref_min == ULONG_MAX) { | 492 | if (tsc_ref_min == ULONG_MAX) { |
| 300 | printk(KERN_WARNING "TSC: HPET/PMTIMER calibration failed due " | 493 | printk(KERN_WARNING "TSC: HPET/PMTIMER calibration failed. " |
| 301 | "to SMI disturbance. Using PIT calibration\n"); | 494 | "Using PIT calibration\n"); |
| 302 | return tsc_pit_min; | 495 | return tsc_pit_min; |
| 303 | } | 496 | } |
| 304 | 497 | ||
| 305 | /* Check the reference deviation */ | ||
| 306 | delta = ((u64) tsc_pit_min) * 100; | ||
| 307 | do_div(delta, tsc_ref_min); | ||
| 308 | |||
| 309 | /* | ||
| 310 | * If both calibration results are inside a 5% window, the we | ||
| 311 | * use the lower frequency of those as it is probably the | ||
| 312 | * closest estimate. | ||
| 313 | */ | ||
| 314 | if (delta >= 95 && delta <= 105) { | ||
| 315 | printk(KERN_INFO "TSC: PIT calibration confirmed by %s.\n", | ||
| 316 | hpet ? "HPET" : "PMTIMER"); | ||
| 317 | printk(KERN_INFO "TSC: using %s calibration value\n", | ||
| 318 | tsc_pit_min <= tsc_ref_min ? "PIT" : | ||
| 319 | hpet ? "HPET" : "PMTIMER"); | ||
| 320 | return tsc_pit_min <= tsc_ref_min ? tsc_pit_min : tsc_ref_min; | ||
| 321 | } | ||
| 322 | |||
| 323 | printk(KERN_WARNING "TSC: PIT calibration deviates from %s: %lu %lu.\n", | ||
| 324 | hpet ? "HPET" : "PMTIMER", tsc_pit_min, tsc_ref_min); | ||
| 325 | |||
| 326 | /* | 498 | /* |
| 327 | * The calibration values differ too much. In doubt, we use | 499 | * The calibration values differ too much. In doubt, we use |
| 328 | * the PIT value as we know that there are PMTIMERs around | 500 | * the PIT value as we know that there are PMTIMERs around |
| 329 | * running at double speed. | 501 | * running at double speed. At least we let the user know: |
| 330 | */ | 502 | */ |
| 503 | printk(KERN_WARNING "TSC: PIT calibration deviates from %s: %lu %lu.\n", | ||
| 504 | hpet ? "HPET" : "PMTIMER", tsc_pit_min, tsc_ref_min); | ||
| 331 | printk(KERN_INFO "TSC: Using PIT calibration value\n"); | 505 | printk(KERN_INFO "TSC: Using PIT calibration value\n"); |
| 332 | return tsc_pit_min; | 506 | return tsc_pit_min; |
| 333 | } | 507 | } |
diff --git a/arch/x86/kernel/visws_quirks.c b/arch/x86/kernel/visws_quirks.c index 594ef47f0a63..61a97e616f70 100644 --- a/arch/x86/kernel/visws_quirks.c +++ b/arch/x86/kernel/visws_quirks.c | |||
| @@ -25,45 +25,31 @@ | |||
| 25 | #include <asm/visws/cobalt.h> | 25 | #include <asm/visws/cobalt.h> |
| 26 | #include <asm/visws/piix4.h> | 26 | #include <asm/visws/piix4.h> |
| 27 | #include <asm/arch_hooks.h> | 27 | #include <asm/arch_hooks.h> |
| 28 | #include <asm/io_apic.h> | ||
| 28 | #include <asm/fixmap.h> | 29 | #include <asm/fixmap.h> |
| 29 | #include <asm/reboot.h> | 30 | #include <asm/reboot.h> |
| 30 | #include <asm/setup.h> | 31 | #include <asm/setup.h> |
| 31 | #include <asm/e820.h> | 32 | #include <asm/e820.h> |
| 32 | #include <asm/smp.h> | ||
| 33 | #include <asm/io.h> | 33 | #include <asm/io.h> |
| 34 | 34 | ||
| 35 | #include <mach_ipi.h> | 35 | #include <mach_ipi.h> |
| 36 | 36 | ||
| 37 | #include "mach_apic.h" | 37 | #include "mach_apic.h" |
| 38 | 38 | ||
| 39 | #include <linux/init.h> | ||
| 40 | #include <linux/smp.h> | ||
| 41 | |||
| 42 | #include <linux/kernel_stat.h> | 39 | #include <linux/kernel_stat.h> |
| 43 | #include <linux/interrupt.h> | ||
| 44 | #include <linux/init.h> | ||
| 45 | 40 | ||
| 46 | #include <asm/io.h> | ||
| 47 | #include <asm/apic.h> | ||
| 48 | #include <asm/i8259.h> | 41 | #include <asm/i8259.h> |
| 49 | #include <asm/irq_vectors.h> | 42 | #include <asm/irq_vectors.h> |
| 50 | #include <asm/visws/cobalt.h> | ||
| 51 | #include <asm/visws/lithium.h> | 43 | #include <asm/visws/lithium.h> |
| 52 | #include <asm/visws/piix4.h> | ||
| 53 | 44 | ||
| 54 | #include <linux/sched.h> | 45 | #include <linux/sched.h> |
| 55 | #include <linux/kernel.h> | 46 | #include <linux/kernel.h> |
| 56 | #include <linux/init.h> | ||
| 57 | #include <linux/pci.h> | 47 | #include <linux/pci.h> |
| 58 | #include <linux/pci_ids.h> | 48 | #include <linux/pci_ids.h> |
| 59 | 49 | ||
| 60 | extern int no_broadcast; | 50 | extern int no_broadcast; |
| 61 | 51 | ||
| 62 | #include <asm/io.h> | ||
| 63 | #include <asm/apic.h> | 52 | #include <asm/apic.h> |
| 64 | #include <asm/arch_hooks.h> | ||
| 65 | #include <asm/visws/cobalt.h> | ||
| 66 | #include <asm/visws/lithium.h> | ||
| 67 | 53 | ||
| 68 | char visws_board_type = -1; | 54 | char visws_board_type = -1; |
| 69 | char visws_board_rev = -1; | 55 | char visws_board_rev = -1; |
diff --git a/arch/x86/kernel/vmi_32.c b/arch/x86/kernel/vmi_32.c index edfb09f30479..8c9ad02af5a2 100644 --- a/arch/x86/kernel/vmi_32.c +++ b/arch/x86/kernel/vmi_32.c | |||
| @@ -393,13 +393,13 @@ static void *vmi_kmap_atomic_pte(struct page *page, enum km_type type) | |||
| 393 | } | 393 | } |
| 394 | #endif | 394 | #endif |
| 395 | 395 | ||
| 396 | static void vmi_allocate_pte(struct mm_struct *mm, u32 pfn) | 396 | static void vmi_allocate_pte(struct mm_struct *mm, unsigned long pfn) |
| 397 | { | 397 | { |
| 398 | vmi_set_page_type(pfn, VMI_PAGE_L1); | 398 | vmi_set_page_type(pfn, VMI_PAGE_L1); |
| 399 | vmi_ops.allocate_page(pfn, VMI_PAGE_L1, 0, 0, 0); | 399 | vmi_ops.allocate_page(pfn, VMI_PAGE_L1, 0, 0, 0); |
| 400 | } | 400 | } |
| 401 | 401 | ||
| 402 | static void vmi_allocate_pmd(struct mm_struct *mm, u32 pfn) | 402 | static void vmi_allocate_pmd(struct mm_struct *mm, unsigned long pfn) |
| 403 | { | 403 | { |
| 404 | /* | 404 | /* |
| 405 | * This call comes in very early, before mem_map is setup. | 405 | * This call comes in very early, before mem_map is setup. |
| @@ -410,20 +410,20 @@ static void vmi_allocate_pmd(struct mm_struct *mm, u32 pfn) | |||
| 410 | vmi_ops.allocate_page(pfn, VMI_PAGE_L2, 0, 0, 0); | 410 | vmi_ops.allocate_page(pfn, VMI_PAGE_L2, 0, 0, 0); |
| 411 | } | 411 | } |
| 412 | 412 | ||
| 413 | static void vmi_allocate_pmd_clone(u32 pfn, u32 clonepfn, u32 start, u32 count) | 413 | static void vmi_allocate_pmd_clone(unsigned long pfn, unsigned long clonepfn, unsigned long start, unsigned long count) |
| 414 | { | 414 | { |
| 415 | vmi_set_page_type(pfn, VMI_PAGE_L2 | VMI_PAGE_CLONE); | 415 | vmi_set_page_type(pfn, VMI_PAGE_L2 | VMI_PAGE_CLONE); |
| 416 | vmi_check_page_type(clonepfn, VMI_PAGE_L2); | 416 | vmi_check_page_type(clonepfn, VMI_PAGE_L2); |
| 417 | vmi_ops.allocate_page(pfn, VMI_PAGE_L2 | VMI_PAGE_CLONE, clonepfn, start, count); | 417 | vmi_ops.allocate_page(pfn, VMI_PAGE_L2 | VMI_PAGE_CLONE, clonepfn, start, count); |
| 418 | } | 418 | } |
| 419 | 419 | ||
| 420 | static void vmi_release_pte(u32 pfn) | 420 | static void vmi_release_pte(unsigned long pfn) |
| 421 | { | 421 | { |
| 422 | vmi_ops.release_page(pfn, VMI_PAGE_L1); | 422 | vmi_ops.release_page(pfn, VMI_PAGE_L1); |
| 423 | vmi_set_page_type(pfn, VMI_PAGE_NORMAL); | 423 | vmi_set_page_type(pfn, VMI_PAGE_NORMAL); |
| 424 | } | 424 | } |
| 425 | 425 | ||
| 426 | static void vmi_release_pmd(u32 pfn) | 426 | static void vmi_release_pmd(unsigned long pfn) |
| 427 | { | 427 | { |
| 428 | vmi_ops.release_page(pfn, VMI_PAGE_L2); | 428 | vmi_ops.release_page(pfn, VMI_PAGE_L2); |
| 429 | vmi_set_page_type(pfn, VMI_PAGE_NORMAL); | 429 | vmi_set_page_type(pfn, VMI_PAGE_NORMAL); |
diff --git a/arch/x86/lib/msr-on-cpu.c b/arch/x86/lib/msr-on-cpu.c index 01b868ba82f8..321cf720dbb6 100644 --- a/arch/x86/lib/msr-on-cpu.c +++ b/arch/x86/lib/msr-on-cpu.c | |||
| @@ -16,37 +16,46 @@ static void __rdmsr_on_cpu(void *info) | |||
| 16 | rdmsr(rv->msr_no, rv->l, rv->h); | 16 | rdmsr(rv->msr_no, rv->l, rv->h); |
| 17 | } | 17 | } |
| 18 | 18 | ||
| 19 | static void __rdmsr_safe_on_cpu(void *info) | 19 | static void __wrmsr_on_cpu(void *info) |
| 20 | { | 20 | { |
| 21 | struct msr_info *rv = info; | 21 | struct msr_info *rv = info; |
| 22 | 22 | ||
| 23 | rv->err = rdmsr_safe(rv->msr_no, &rv->l, &rv->h); | 23 | wrmsr(rv->msr_no, rv->l, rv->h); |
| 24 | } | 24 | } |
| 25 | 25 | ||
| 26 | static int _rdmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h, int safe) | 26 | int rdmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h) |
| 27 | { | 27 | { |
| 28 | int err = 0; | 28 | int err; |
| 29 | struct msr_info rv; | 29 | struct msr_info rv; |
| 30 | 30 | ||
| 31 | rv.msr_no = msr_no; | 31 | rv.msr_no = msr_no; |
| 32 | if (safe) { | 32 | err = smp_call_function_single(cpu, __rdmsr_on_cpu, &rv, 1); |
| 33 | err = smp_call_function_single(cpu, __rdmsr_safe_on_cpu, | ||
| 34 | &rv, 1); | ||
| 35 | err = err ? err : rv.err; | ||
| 36 | } else { | ||
| 37 | err = smp_call_function_single(cpu, __rdmsr_on_cpu, &rv, 1); | ||
| 38 | } | ||
| 39 | *l = rv.l; | 33 | *l = rv.l; |
| 40 | *h = rv.h; | 34 | *h = rv.h; |
| 41 | 35 | ||
| 42 | return err; | 36 | return err; |
| 43 | } | 37 | } |
| 44 | 38 | ||
| 45 | static void __wrmsr_on_cpu(void *info) | 39 | int wrmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h) |
| 40 | { | ||
| 41 | int err; | ||
| 42 | struct msr_info rv; | ||
| 43 | |||
| 44 | rv.msr_no = msr_no; | ||
| 45 | rv.l = l; | ||
| 46 | rv.h = h; | ||
| 47 | err = smp_call_function_single(cpu, __wrmsr_on_cpu, &rv, 1); | ||
| 48 | |||
| 49 | return err; | ||
| 50 | } | ||
| 51 | |||
| 52 | /* These "safe" variants are slower and should be used when the target MSR | ||
| 53 | may not actually exist. */ | ||
| 54 | static void __rdmsr_safe_on_cpu(void *info) | ||
| 46 | { | 55 | { |
| 47 | struct msr_info *rv = info; | 56 | struct msr_info *rv = info; |
| 48 | 57 | ||
| 49 | wrmsr(rv->msr_no, rv->l, rv->h); | 58 | rv->err = rdmsr_safe(rv->msr_no, &rv->l, &rv->h); |
| 50 | } | 59 | } |
| 51 | 60 | ||
| 52 | static void __wrmsr_safe_on_cpu(void *info) | 61 | static void __wrmsr_safe_on_cpu(void *info) |
| @@ -56,45 +65,30 @@ static void __wrmsr_safe_on_cpu(void *info) | |||
| 56 | rv->err = wrmsr_safe(rv->msr_no, rv->l, rv->h); | 65 | rv->err = wrmsr_safe(rv->msr_no, rv->l, rv->h); |
| 57 | } | 66 | } |
| 58 | 67 | ||
| 59 | static int _wrmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h, int safe) | 68 | int rdmsr_safe_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h) |
| 60 | { | 69 | { |
| 61 | int err = 0; | 70 | int err; |
| 62 | struct msr_info rv; | 71 | struct msr_info rv; |
| 63 | 72 | ||
| 64 | rv.msr_no = msr_no; | 73 | rv.msr_no = msr_no; |
| 65 | rv.l = l; | 74 | err = smp_call_function_single(cpu, __rdmsr_safe_on_cpu, &rv, 1); |
| 66 | rv.h = h; | 75 | *l = rv.l; |
| 67 | if (safe) { | 76 | *h = rv.h; |
| 68 | err = smp_call_function_single(cpu, __wrmsr_safe_on_cpu, | ||
| 69 | &rv, 1); | ||
| 70 | err = err ? err : rv.err; | ||
| 71 | } else { | ||
| 72 | err = smp_call_function_single(cpu, __wrmsr_on_cpu, &rv, 1); | ||
| 73 | } | ||
| 74 | |||
| 75 | return err; | ||
| 76 | } | ||
| 77 | 77 | ||
| 78 | int wrmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h) | 78 | return err ? err : rv.err; |
| 79 | { | ||
| 80 | return _wrmsr_on_cpu(cpu, msr_no, l, h, 0); | ||
| 81 | } | 79 | } |
| 82 | 80 | ||
| 83 | int rdmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h) | ||
| 84 | { | ||
| 85 | return _rdmsr_on_cpu(cpu, msr_no, l, h, 0); | ||
| 86 | } | ||
| 87 | |||
| 88 | /* These "safe" variants are slower and should be used when the target MSR | ||
| 89 | may not actually exist. */ | ||
| 90 | int wrmsr_safe_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h) | 81 | int wrmsr_safe_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h) |
| 91 | { | 82 | { |
| 92 | return _wrmsr_on_cpu(cpu, msr_no, l, h, 1); | 83 | int err; |
| 93 | } | 84 | struct msr_info rv; |
| 94 | 85 | ||
| 95 | int rdmsr_safe_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h) | 86 | rv.msr_no = msr_no; |
| 96 | { | 87 | rv.l = l; |
| 97 | return _rdmsr_on_cpu(cpu, msr_no, l, h, 1); | 88 | rv.h = h; |
| 89 | err = smp_call_function_single(cpu, __wrmsr_safe_on_cpu, &rv, 1); | ||
| 90 | |||
| 91 | return err ? err : rv.err; | ||
| 98 | } | 92 | } |
| 99 | 93 | ||
| 100 | EXPORT_SYMBOL(rdmsr_on_cpu); | 94 | EXPORT_SYMBOL(rdmsr_on_cpu); |
diff --git a/arch/x86/lib/string_32.c b/arch/x86/lib/string_32.c index 94972e7c094d..82004d2bf05e 100644 --- a/arch/x86/lib/string_32.c +++ b/arch/x86/lib/string_32.c | |||
| @@ -22,7 +22,7 @@ char *strcpy(char *dest, const char *src) | |||
| 22 | "testb %%al,%%al\n\t" | 22 | "testb %%al,%%al\n\t" |
| 23 | "jne 1b" | 23 | "jne 1b" |
| 24 | : "=&S" (d0), "=&D" (d1), "=&a" (d2) | 24 | : "=&S" (d0), "=&D" (d1), "=&a" (d2) |
| 25 | :"0" (src), "1" (dest) : "memory"); | 25 | : "0" (src), "1" (dest) : "memory"); |
| 26 | return dest; | 26 | return dest; |
| 27 | } | 27 | } |
| 28 | EXPORT_SYMBOL(strcpy); | 28 | EXPORT_SYMBOL(strcpy); |
| @@ -42,7 +42,7 @@ char *strncpy(char *dest, const char *src, size_t count) | |||
| 42 | "stosb\n" | 42 | "stosb\n" |
| 43 | "2:" | 43 | "2:" |
| 44 | : "=&S" (d0), "=&D" (d1), "=&c" (d2), "=&a" (d3) | 44 | : "=&S" (d0), "=&D" (d1), "=&c" (d2), "=&a" (d3) |
| 45 | :"0" (src), "1" (dest), "2" (count) : "memory"); | 45 | : "0" (src), "1" (dest), "2" (count) : "memory"); |
| 46 | return dest; | 46 | return dest; |
| 47 | } | 47 | } |
| 48 | EXPORT_SYMBOL(strncpy); | 48 | EXPORT_SYMBOL(strncpy); |
| @@ -60,7 +60,7 @@ char *strcat(char *dest, const char *src) | |||
| 60 | "testb %%al,%%al\n\t" | 60 | "testb %%al,%%al\n\t" |
| 61 | "jne 1b" | 61 | "jne 1b" |
| 62 | : "=&S" (d0), "=&D" (d1), "=&a" (d2), "=&c" (d3) | 62 | : "=&S" (d0), "=&D" (d1), "=&a" (d2), "=&c" (d3) |
| 63 | : "0" (src), "1" (dest), "2" (0), "3" (0xffffffffu): "memory"); | 63 | : "0" (src), "1" (dest), "2" (0), "3" (0xffffffffu) : "memory"); |
| 64 | return dest; | 64 | return dest; |
| 65 | } | 65 | } |
| 66 | EXPORT_SYMBOL(strcat); | 66 | EXPORT_SYMBOL(strcat); |
| @@ -105,9 +105,9 @@ int strcmp(const char *cs, const char *ct) | |||
| 105 | "2:\tsbbl %%eax,%%eax\n\t" | 105 | "2:\tsbbl %%eax,%%eax\n\t" |
| 106 | "orb $1,%%al\n" | 106 | "orb $1,%%al\n" |
| 107 | "3:" | 107 | "3:" |
| 108 | :"=a" (res), "=&S" (d0), "=&D" (d1) | 108 | : "=a" (res), "=&S" (d0), "=&D" (d1) |
| 109 | :"1" (cs), "2" (ct) | 109 | : "1" (cs), "2" (ct) |
| 110 | :"memory"); | 110 | : "memory"); |
| 111 | return res; | 111 | return res; |
| 112 | } | 112 | } |
| 113 | EXPORT_SYMBOL(strcmp); | 113 | EXPORT_SYMBOL(strcmp); |
| @@ -130,9 +130,9 @@ int strncmp(const char *cs, const char *ct, size_t count) | |||
| 130 | "3:\tsbbl %%eax,%%eax\n\t" | 130 | "3:\tsbbl %%eax,%%eax\n\t" |
| 131 | "orb $1,%%al\n" | 131 | "orb $1,%%al\n" |
| 132 | "4:" | 132 | "4:" |
| 133 | :"=a" (res), "=&S" (d0), "=&D" (d1), "=&c" (d2) | 133 | : "=a" (res), "=&S" (d0), "=&D" (d1), "=&c" (d2) |
| 134 | :"1" (cs), "2" (ct), "3" (count) | 134 | : "1" (cs), "2" (ct), "3" (count) |
| 135 | :"memory"); | 135 | : "memory"); |
| 136 | return res; | 136 | return res; |
| 137 | } | 137 | } |
| 138 | EXPORT_SYMBOL(strncmp); | 138 | EXPORT_SYMBOL(strncmp); |
| @@ -152,9 +152,9 @@ char *strchr(const char *s, int c) | |||
| 152 | "movl $1,%1\n" | 152 | "movl $1,%1\n" |
| 153 | "2:\tmovl %1,%0\n\t" | 153 | "2:\tmovl %1,%0\n\t" |
| 154 | "decl %0" | 154 | "decl %0" |
| 155 | :"=a" (res), "=&S" (d0) | 155 | : "=a" (res), "=&S" (d0) |
| 156 | :"1" (s), "0" (c) | 156 | : "1" (s), "0" (c) |
| 157 | :"memory"); | 157 | : "memory"); |
| 158 | return res; | 158 | return res; |
| 159 | } | 159 | } |
| 160 | EXPORT_SYMBOL(strchr); | 160 | EXPORT_SYMBOL(strchr); |
| @@ -169,9 +169,9 @@ size_t strlen(const char *s) | |||
| 169 | "scasb\n\t" | 169 | "scasb\n\t" |
| 170 | "notl %0\n\t" | 170 | "notl %0\n\t" |
| 171 | "decl %0" | 171 | "decl %0" |
| 172 | :"=c" (res), "=&D" (d0) | 172 | : "=c" (res), "=&D" (d0) |
| 173 | :"1" (s), "a" (0), "0" (0xffffffffu) | 173 | : "1" (s), "a" (0), "0" (0xffffffffu) |
| 174 | :"memory"); | 174 | : "memory"); |
| 175 | return res; | 175 | return res; |
| 176 | } | 176 | } |
| 177 | EXPORT_SYMBOL(strlen); | 177 | EXPORT_SYMBOL(strlen); |
| @@ -189,9 +189,9 @@ void *memchr(const void *cs, int c, size_t count) | |||
| 189 | "je 1f\n\t" | 189 | "je 1f\n\t" |
| 190 | "movl $1,%0\n" | 190 | "movl $1,%0\n" |
| 191 | "1:\tdecl %0" | 191 | "1:\tdecl %0" |
| 192 | :"=D" (res), "=&c" (d0) | 192 | : "=D" (res), "=&c" (d0) |
| 193 | :"a" (c), "0" (cs), "1" (count) | 193 | : "a" (c), "0" (cs), "1" (count) |
| 194 | :"memory"); | 194 | : "memory"); |
| 195 | return res; | 195 | return res; |
| 196 | } | 196 | } |
| 197 | EXPORT_SYMBOL(memchr); | 197 | EXPORT_SYMBOL(memchr); |
| @@ -228,9 +228,9 @@ size_t strnlen(const char *s, size_t count) | |||
| 228 | "cmpl $-1,%1\n\t" | 228 | "cmpl $-1,%1\n\t" |
| 229 | "jne 1b\n" | 229 | "jne 1b\n" |
| 230 | "3:\tsubl %2,%0" | 230 | "3:\tsubl %2,%0" |
| 231 | :"=a" (res), "=&d" (d0) | 231 | : "=a" (res), "=&d" (d0) |
| 232 | :"c" (s), "1" (count) | 232 | : "c" (s), "1" (count) |
| 233 | :"memory"); | 233 | : "memory"); |
| 234 | return res; | 234 | return res; |
| 235 | } | 235 | } |
| 236 | EXPORT_SYMBOL(strnlen); | 236 | EXPORT_SYMBOL(strnlen); |
diff --git a/arch/x86/lib/strstr_32.c b/arch/x86/lib/strstr_32.c index 42e8a50303f3..8e2d55f754bf 100644 --- a/arch/x86/lib/strstr_32.c +++ b/arch/x86/lib/strstr_32.c | |||
| @@ -23,9 +23,9 @@ __asm__ __volatile__( | |||
| 23 | "jne 1b\n\t" | 23 | "jne 1b\n\t" |
| 24 | "xorl %%eax,%%eax\n\t" | 24 | "xorl %%eax,%%eax\n\t" |
| 25 | "2:" | 25 | "2:" |
| 26 | :"=a" (__res), "=&c" (d0), "=&S" (d1) | 26 | : "=a" (__res), "=&c" (d0), "=&S" (d1) |
| 27 | :"0" (0), "1" (0xffffffff), "2" (cs), "g" (ct) | 27 | : "0" (0), "1" (0xffffffff), "2" (cs), "g" (ct) |
| 28 | :"dx", "di"); | 28 | : "dx", "di"); |
| 29 | return __res; | 29 | return __res; |
| 30 | } | 30 | } |
| 31 | 31 | ||
diff --git a/arch/x86/mm/discontig_32.c b/arch/x86/mm/discontig_32.c index 62fa440678d8..847c164725f4 100644 --- a/arch/x86/mm/discontig_32.c +++ b/arch/x86/mm/discontig_32.c | |||
| @@ -328,7 +328,7 @@ void __init initmem_init(unsigned long start_pfn, | |||
| 328 | 328 | ||
| 329 | get_memcfg_numa(); | 329 | get_memcfg_numa(); |
| 330 | 330 | ||
| 331 | kva_pages = round_up(calculate_numa_remap_pages(), PTRS_PER_PTE); | 331 | kva_pages = roundup(calculate_numa_remap_pages(), PTRS_PER_PTE); |
| 332 | 332 | ||
| 333 | kva_target_pfn = round_down(max_low_pfn - kva_pages, PTRS_PER_PTE); | 333 | kva_target_pfn = round_down(max_low_pfn - kva_pages, PTRS_PER_PTE); |
| 334 | do { | 334 | do { |
diff --git a/arch/x86/mm/dump_pagetables.c b/arch/x86/mm/dump_pagetables.c index a20d1fa64b4e..e7277cbcfb40 100644 --- a/arch/x86/mm/dump_pagetables.c +++ b/arch/x86/mm/dump_pagetables.c | |||
| @@ -148,8 +148,8 @@ static void note_page(struct seq_file *m, struct pg_state *st, | |||
| 148 | * we have now. "break" is either changing perms, levels or | 148 | * we have now. "break" is either changing perms, levels or |
| 149 | * address space marker. | 149 | * address space marker. |
| 150 | */ | 150 | */ |
| 151 | prot = pgprot_val(new_prot) & ~(PTE_PFN_MASK); | 151 | prot = pgprot_val(new_prot) & PTE_FLAGS_MASK; |
| 152 | cur = pgprot_val(st->current_prot) & ~(PTE_PFN_MASK); | 152 | cur = pgprot_val(st->current_prot) & PTE_FLAGS_MASK; |
| 153 | 153 | ||
| 154 | if (!st->level) { | 154 | if (!st->level) { |
| 155 | /* First entry */ | 155 | /* First entry */ |
diff --git a/arch/x86/mm/init_64.c b/arch/x86/mm/init_64.c index d3746efb060d..770536ebf7e9 100644 --- a/arch/x86/mm/init_64.c +++ b/arch/x86/mm/init_64.c | |||
| @@ -225,7 +225,7 @@ void __init init_extra_mapping_uc(unsigned long phys, unsigned long size) | |||
| 225 | void __init cleanup_highmap(void) | 225 | void __init cleanup_highmap(void) |
| 226 | { | 226 | { |
| 227 | unsigned long vaddr = __START_KERNEL_map; | 227 | unsigned long vaddr = __START_KERNEL_map; |
| 228 | unsigned long end = round_up((unsigned long)_end, PMD_SIZE) - 1; | 228 | unsigned long end = roundup((unsigned long)_end, PMD_SIZE) - 1; |
| 229 | pmd_t *pmd = level2_kernel_pgt; | 229 | pmd_t *pmd = level2_kernel_pgt; |
| 230 | pmd_t *last_pmd = pmd + PTRS_PER_PMD; | 230 | pmd_t *last_pmd = pmd + PTRS_PER_PMD; |
| 231 | 231 | ||
| @@ -451,14 +451,14 @@ static void __init find_early_table_space(unsigned long end) | |||
| 451 | unsigned long puds, pmds, ptes, tables, start; | 451 | unsigned long puds, pmds, ptes, tables, start; |
| 452 | 452 | ||
| 453 | puds = (end + PUD_SIZE - 1) >> PUD_SHIFT; | 453 | puds = (end + PUD_SIZE - 1) >> PUD_SHIFT; |
| 454 | tables = round_up(puds * sizeof(pud_t), PAGE_SIZE); | 454 | tables = roundup(puds * sizeof(pud_t), PAGE_SIZE); |
| 455 | if (direct_gbpages) { | 455 | if (direct_gbpages) { |
| 456 | unsigned long extra; | 456 | unsigned long extra; |
| 457 | extra = end - ((end>>PUD_SHIFT) << PUD_SHIFT); | 457 | extra = end - ((end>>PUD_SHIFT) << PUD_SHIFT); |
| 458 | pmds = (extra + PMD_SIZE - 1) >> PMD_SHIFT; | 458 | pmds = (extra + PMD_SIZE - 1) >> PMD_SHIFT; |
| 459 | } else | 459 | } else |
| 460 | pmds = (end + PMD_SIZE - 1) >> PMD_SHIFT; | 460 | pmds = (end + PMD_SIZE - 1) >> PMD_SHIFT; |
| 461 | tables += round_up(pmds * sizeof(pmd_t), PAGE_SIZE); | 461 | tables += roundup(pmds * sizeof(pmd_t), PAGE_SIZE); |
| 462 | 462 | ||
| 463 | if (cpu_has_pse) { | 463 | if (cpu_has_pse) { |
| 464 | unsigned long extra; | 464 | unsigned long extra; |
| @@ -466,7 +466,7 @@ static void __init find_early_table_space(unsigned long end) | |||
| 466 | ptes = (extra + PAGE_SIZE - 1) >> PAGE_SHIFT; | 466 | ptes = (extra + PAGE_SIZE - 1) >> PAGE_SHIFT; |
| 467 | } else | 467 | } else |
| 468 | ptes = (end + PAGE_SIZE - 1) >> PAGE_SHIFT; | 468 | ptes = (end + PAGE_SIZE - 1) >> PAGE_SHIFT; |
| 469 | tables += round_up(ptes * sizeof(pte_t), PAGE_SIZE); | 469 | tables += roundup(ptes * sizeof(pte_t), PAGE_SIZE); |
| 470 | 470 | ||
| 471 | /* | 471 | /* |
| 472 | * RED-PEN putting page tables only on node 0 could | 472 | * RED-PEN putting page tables only on node 0 could |
diff --git a/arch/x86/mm/numa_64.c b/arch/x86/mm/numa_64.c index a4dd793d6003..cebcbf152d46 100644 --- a/arch/x86/mm/numa_64.c +++ b/arch/x86/mm/numa_64.c | |||
| @@ -79,7 +79,7 @@ static int __init allocate_cachealigned_memnodemap(void) | |||
| 79 | return 0; | 79 | return 0; |
| 80 | 80 | ||
| 81 | addr = 0x8000; | 81 | addr = 0x8000; |
| 82 | nodemap_size = round_up(sizeof(s16) * memnodemapsize, L1_CACHE_BYTES); | 82 | nodemap_size = roundup(sizeof(s16) * memnodemapsize, L1_CACHE_BYTES); |
| 83 | nodemap_addr = find_e820_area(addr, max_pfn<<PAGE_SHIFT, | 83 | nodemap_addr = find_e820_area(addr, max_pfn<<PAGE_SHIFT, |
| 84 | nodemap_size, L1_CACHE_BYTES); | 84 | nodemap_size, L1_CACHE_BYTES); |
| 85 | if (nodemap_addr == -1UL) { | 85 | if (nodemap_addr == -1UL) { |
| @@ -176,10 +176,10 @@ void __init setup_node_bootmem(int nodeid, unsigned long start, | |||
| 176 | unsigned long start_pfn, last_pfn, bootmap_pages, bootmap_size; | 176 | unsigned long start_pfn, last_pfn, bootmap_pages, bootmap_size; |
| 177 | unsigned long bootmap_start, nodedata_phys; | 177 | unsigned long bootmap_start, nodedata_phys; |
| 178 | void *bootmap; | 178 | void *bootmap; |
| 179 | const int pgdat_size = round_up(sizeof(pg_data_t), PAGE_SIZE); | 179 | const int pgdat_size = roundup(sizeof(pg_data_t), PAGE_SIZE); |
| 180 | int nid; | 180 | int nid; |
| 181 | 181 | ||
| 182 | start = round_up(start, ZONE_ALIGN); | 182 | start = roundup(start, ZONE_ALIGN); |
| 183 | 183 | ||
| 184 | printk(KERN_INFO "Bootmem setup node %d %016lx-%016lx\n", nodeid, | 184 | printk(KERN_INFO "Bootmem setup node %d %016lx-%016lx\n", nodeid, |
| 185 | start, end); | 185 | start, end); |
| @@ -210,9 +210,9 @@ void __init setup_node_bootmem(int nodeid, unsigned long start, | |||
| 210 | bootmap_pages = bootmem_bootmap_pages(last_pfn - start_pfn); | 210 | bootmap_pages = bootmem_bootmap_pages(last_pfn - start_pfn); |
| 211 | nid = phys_to_nid(nodedata_phys); | 211 | nid = phys_to_nid(nodedata_phys); |
| 212 | if (nid == nodeid) | 212 | if (nid == nodeid) |
| 213 | bootmap_start = round_up(nodedata_phys + pgdat_size, PAGE_SIZE); | 213 | bootmap_start = roundup(nodedata_phys + pgdat_size, PAGE_SIZE); |
| 214 | else | 214 | else |
| 215 | bootmap_start = round_up(start, PAGE_SIZE); | 215 | bootmap_start = roundup(start, PAGE_SIZE); |
| 216 | /* | 216 | /* |
| 217 | * SMP_CACHE_BYTES could be enough, but init_bootmem_node like | 217 | * SMP_CACHE_BYTES could be enough, but init_bootmem_node like |
| 218 | * to use that to align to PAGE_SIZE | 218 | * to use that to align to PAGE_SIZE |
diff --git a/arch/x86/mm/pageattr.c b/arch/x86/mm/pageattr.c index 43e2f8483e4f..898fad617abe 100644 --- a/arch/x86/mm/pageattr.c +++ b/arch/x86/mm/pageattr.c | |||
| @@ -84,7 +84,7 @@ static inline unsigned long highmap_start_pfn(void) | |||
| 84 | 84 | ||
| 85 | static inline unsigned long highmap_end_pfn(void) | 85 | static inline unsigned long highmap_end_pfn(void) |
| 86 | { | 86 | { |
| 87 | return __pa(round_up((unsigned long)_end, PMD_SIZE)) >> PAGE_SHIFT; | 87 | return __pa(roundup((unsigned long)_end, PMD_SIZE)) >> PAGE_SHIFT; |
| 88 | } | 88 | } |
| 89 | 89 | ||
| 90 | #endif | 90 | #endif |
| @@ -906,11 +906,13 @@ int set_memory_ro(unsigned long addr, int numpages) | |||
| 906 | { | 906 | { |
| 907 | return change_page_attr_clear(addr, numpages, __pgprot(_PAGE_RW)); | 907 | return change_page_attr_clear(addr, numpages, __pgprot(_PAGE_RW)); |
| 908 | } | 908 | } |
| 909 | EXPORT_SYMBOL_GPL(set_memory_ro); | ||
| 909 | 910 | ||
| 910 | int set_memory_rw(unsigned long addr, int numpages) | 911 | int set_memory_rw(unsigned long addr, int numpages) |
| 911 | { | 912 | { |
| 912 | return change_page_attr_set(addr, numpages, __pgprot(_PAGE_RW)); | 913 | return change_page_attr_set(addr, numpages, __pgprot(_PAGE_RW)); |
| 913 | } | 914 | } |
| 915 | EXPORT_SYMBOL_GPL(set_memory_rw); | ||
| 914 | 916 | ||
| 915 | int set_memory_np(unsigned long addr, int numpages) | 917 | int set_memory_np(unsigned long addr, int numpages) |
| 916 | { | 918 | { |
diff --git a/arch/x86/mm/pgtable.c b/arch/x86/mm/pgtable.c index d50302774fe2..86f2ffc43c3d 100644 --- a/arch/x86/mm/pgtable.c +++ b/arch/x86/mm/pgtable.c | |||
| @@ -63,10 +63,8 @@ static inline void pgd_list_del(pgd_t *pgd) | |||
| 63 | #define UNSHARED_PTRS_PER_PGD \ | 63 | #define UNSHARED_PTRS_PER_PGD \ |
| 64 | (SHARED_KERNEL_PMD ? KERNEL_PGD_BOUNDARY : PTRS_PER_PGD) | 64 | (SHARED_KERNEL_PMD ? KERNEL_PGD_BOUNDARY : PTRS_PER_PGD) |
| 65 | 65 | ||
| 66 | static void pgd_ctor(void *p) | 66 | static void pgd_ctor(pgd_t *pgd) |
| 67 | { | 67 | { |
| 68 | pgd_t *pgd = p; | ||
| 69 | |||
| 70 | /* If the pgd points to a shared pagetable level (either the | 68 | /* If the pgd points to a shared pagetable level (either the |
| 71 | ptes in non-PAE, or shared PMD in PAE), then just copy the | 69 | ptes in non-PAE, or shared PMD in PAE), then just copy the |
| 72 | references from swapper_pg_dir. */ | 70 | references from swapper_pg_dir. */ |
| @@ -87,7 +85,7 @@ static void pgd_ctor(void *p) | |||
| 87 | pgd_list_add(pgd); | 85 | pgd_list_add(pgd); |
| 88 | } | 86 | } |
| 89 | 87 | ||
| 90 | static void pgd_dtor(void *pgd) | 88 | static void pgd_dtor(pgd_t *pgd) |
| 91 | { | 89 | { |
| 92 | unsigned long flags; /* can be called from interrupt context */ | 90 | unsigned long flags; /* can be called from interrupt context */ |
| 93 | 91 | ||
diff --git a/arch/x86/mm/pgtable_32.c b/arch/x86/mm/pgtable_32.c index cab0abbd1ebe..0951db9ee519 100644 --- a/arch/x86/mm/pgtable_32.c +++ b/arch/x86/mm/pgtable_32.c | |||
| @@ -123,7 +123,8 @@ static int __init parse_vmalloc(char *arg) | |||
| 123 | if (!arg) | 123 | if (!arg) |
| 124 | return -EINVAL; | 124 | return -EINVAL; |
| 125 | 125 | ||
| 126 | __VMALLOC_RESERVE = memparse(arg, &arg); | 126 | /* Add VMALLOC_OFFSET to the parsed value due to vm area guard hole*/ |
| 127 | __VMALLOC_RESERVE = memparse(arg, &arg) + VMALLOC_OFFSET; | ||
| 127 | return 0; | 128 | return 0; |
| 128 | } | 129 | } |
| 129 | early_param("vmalloc", parse_vmalloc); | 130 | early_param("vmalloc", parse_vmalloc); |
diff --git a/arch/x86/oprofile/op_model_p4.c b/arch/x86/oprofile/op_model_p4.c index 56b4757a1f47..43ac5af338d8 100644 --- a/arch/x86/oprofile/op_model_p4.c +++ b/arch/x86/oprofile/op_model_p4.c | |||
| @@ -10,11 +10,12 @@ | |||
| 10 | 10 | ||
| 11 | #include <linux/oprofile.h> | 11 | #include <linux/oprofile.h> |
| 12 | #include <linux/smp.h> | 12 | #include <linux/smp.h> |
| 13 | #include <linux/ptrace.h> | ||
| 14 | #include <linux/nmi.h> | ||
| 13 | #include <asm/msr.h> | 15 | #include <asm/msr.h> |
| 14 | #include <asm/ptrace.h> | ||
| 15 | #include <asm/fixmap.h> | 16 | #include <asm/fixmap.h> |
| 16 | #include <asm/apic.h> | 17 | #include <asm/apic.h> |
| 17 | #include <asm/nmi.h> | 18 | |
| 18 | 19 | ||
| 19 | #include "op_x86_model.h" | 20 | #include "op_x86_model.h" |
| 20 | #include "op_counter.h" | 21 | #include "op_counter.h" |
| @@ -40,7 +41,7 @@ static unsigned int num_controls = NUM_CONTROLS_NON_HT; | |||
| 40 | static inline void setup_num_counters(void) | 41 | static inline void setup_num_counters(void) |
| 41 | { | 42 | { |
| 42 | #ifdef CONFIG_SMP | 43 | #ifdef CONFIG_SMP |
| 43 | if (smp_num_siblings == 2){ | 44 | if (smp_num_siblings == 2) { |
| 44 | num_counters = NUM_COUNTERS_HT2; | 45 | num_counters = NUM_COUNTERS_HT2; |
| 45 | num_controls = NUM_CONTROLS_HT2; | 46 | num_controls = NUM_CONTROLS_HT2; |
| 46 | } | 47 | } |
| @@ -86,7 +87,7 @@ struct p4_event_binding { | |||
| 86 | #define CTR_FLAME_2 (1 << 6) | 87 | #define CTR_FLAME_2 (1 << 6) |
| 87 | #define CTR_IQ_5 (1 << 7) | 88 | #define CTR_IQ_5 (1 << 7) |
| 88 | 89 | ||
| 89 | static struct p4_counter_binding p4_counters [NUM_COUNTERS_NON_HT] = { | 90 | static struct p4_counter_binding p4_counters[NUM_COUNTERS_NON_HT] = { |
| 90 | { CTR_BPU_0, MSR_P4_BPU_PERFCTR0, MSR_P4_BPU_CCCR0 }, | 91 | { CTR_BPU_0, MSR_P4_BPU_PERFCTR0, MSR_P4_BPU_CCCR0 }, |
| 91 | { CTR_MS_0, MSR_P4_MS_PERFCTR0, MSR_P4_MS_CCCR0 }, | 92 | { CTR_MS_0, MSR_P4_MS_PERFCTR0, MSR_P4_MS_CCCR0 }, |
| 92 | { CTR_FLAME_0, MSR_P4_FLAME_PERFCTR0, MSR_P4_FLAME_CCCR0 }, | 93 | { CTR_FLAME_0, MSR_P4_FLAME_PERFCTR0, MSR_P4_FLAME_CCCR0 }, |
| @@ -97,32 +98,32 @@ static struct p4_counter_binding p4_counters [NUM_COUNTERS_NON_HT] = { | |||
| 97 | { CTR_IQ_5, MSR_P4_IQ_PERFCTR5, MSR_P4_IQ_CCCR5 } | 98 | { CTR_IQ_5, MSR_P4_IQ_PERFCTR5, MSR_P4_IQ_CCCR5 } |
| 98 | }; | 99 | }; |
| 99 | 100 | ||
| 100 | #define NUM_UNUSED_CCCRS NUM_CCCRS_NON_HT - NUM_COUNTERS_NON_HT | 101 | #define NUM_UNUSED_CCCRS (NUM_CCCRS_NON_HT - NUM_COUNTERS_NON_HT) |
| 101 | 102 | ||
| 102 | /* p4 event codes in libop/op_event.h are indices into this table. */ | 103 | /* p4 event codes in libop/op_event.h are indices into this table. */ |
| 103 | 104 | ||
| 104 | static struct p4_event_binding p4_events[NUM_EVENTS] = { | 105 | static struct p4_event_binding p4_events[NUM_EVENTS] = { |
| 105 | 106 | ||
| 106 | { /* BRANCH_RETIRED */ | 107 | { /* BRANCH_RETIRED */ |
| 107 | 0x05, 0x06, | 108 | 0x05, 0x06, |
| 108 | { {CTR_IQ_4, MSR_P4_CRU_ESCR2}, | 109 | { {CTR_IQ_4, MSR_P4_CRU_ESCR2}, |
| 109 | {CTR_IQ_5, MSR_P4_CRU_ESCR3} } | 110 | {CTR_IQ_5, MSR_P4_CRU_ESCR3} } |
| 110 | }, | 111 | }, |
| 111 | 112 | ||
| 112 | { /* MISPRED_BRANCH_RETIRED */ | 113 | { /* MISPRED_BRANCH_RETIRED */ |
| 113 | 0x04, 0x03, | 114 | 0x04, 0x03, |
| 114 | { { CTR_IQ_4, MSR_P4_CRU_ESCR0}, | 115 | { { CTR_IQ_4, MSR_P4_CRU_ESCR0}, |
| 115 | { CTR_IQ_5, MSR_P4_CRU_ESCR1} } | 116 | { CTR_IQ_5, MSR_P4_CRU_ESCR1} } |
| 116 | }, | 117 | }, |
| 117 | 118 | ||
| 118 | { /* TC_DELIVER_MODE */ | 119 | { /* TC_DELIVER_MODE */ |
| 119 | 0x01, 0x01, | 120 | 0x01, 0x01, |
| 120 | { { CTR_MS_0, MSR_P4_TC_ESCR0}, | 121 | { { CTR_MS_0, MSR_P4_TC_ESCR0}, |
| 121 | { CTR_MS_2, MSR_P4_TC_ESCR1} } | 122 | { CTR_MS_2, MSR_P4_TC_ESCR1} } |
| 122 | }, | 123 | }, |
| 123 | 124 | ||
| 124 | { /* BPU_FETCH_REQUEST */ | 125 | { /* BPU_FETCH_REQUEST */ |
| 125 | 0x00, 0x03, | 126 | 0x00, 0x03, |
| 126 | { { CTR_BPU_0, MSR_P4_BPU_ESCR0}, | 127 | { { CTR_BPU_0, MSR_P4_BPU_ESCR0}, |
| 127 | { CTR_BPU_2, MSR_P4_BPU_ESCR1} } | 128 | { CTR_BPU_2, MSR_P4_BPU_ESCR1} } |
| 128 | }, | 129 | }, |
| @@ -146,7 +147,7 @@ static struct p4_event_binding p4_events[NUM_EVENTS] = { | |||
| 146 | }, | 147 | }, |
| 147 | 148 | ||
| 148 | { /* LOAD_PORT_REPLAY */ | 149 | { /* LOAD_PORT_REPLAY */ |
| 149 | 0x02, 0x04, | 150 | 0x02, 0x04, |
| 150 | { { CTR_FLAME_0, MSR_P4_SAAT_ESCR0}, | 151 | { { CTR_FLAME_0, MSR_P4_SAAT_ESCR0}, |
| 151 | { CTR_FLAME_2, MSR_P4_SAAT_ESCR1} } | 152 | { CTR_FLAME_2, MSR_P4_SAAT_ESCR1} } |
| 152 | }, | 153 | }, |
| @@ -170,43 +171,43 @@ static struct p4_event_binding p4_events[NUM_EVENTS] = { | |||
| 170 | }, | 171 | }, |
| 171 | 172 | ||
| 172 | { /* BSQ_CACHE_REFERENCE */ | 173 | { /* BSQ_CACHE_REFERENCE */ |
| 173 | 0x07, 0x0c, | 174 | 0x07, 0x0c, |
| 174 | { { CTR_BPU_0, MSR_P4_BSU_ESCR0}, | 175 | { { CTR_BPU_0, MSR_P4_BSU_ESCR0}, |
| 175 | { CTR_BPU_2, MSR_P4_BSU_ESCR1} } | 176 | { CTR_BPU_2, MSR_P4_BSU_ESCR1} } |
| 176 | }, | 177 | }, |
| 177 | 178 | ||
| 178 | { /* IOQ_ALLOCATION */ | 179 | { /* IOQ_ALLOCATION */ |
| 179 | 0x06, 0x03, | 180 | 0x06, 0x03, |
| 180 | { { CTR_BPU_0, MSR_P4_FSB_ESCR0}, | 181 | { { CTR_BPU_0, MSR_P4_FSB_ESCR0}, |
| 181 | { 0, 0 } } | 182 | { 0, 0 } } |
| 182 | }, | 183 | }, |
| 183 | 184 | ||
| 184 | { /* IOQ_ACTIVE_ENTRIES */ | 185 | { /* IOQ_ACTIVE_ENTRIES */ |
| 185 | 0x06, 0x1a, | 186 | 0x06, 0x1a, |
| 186 | { { CTR_BPU_2, MSR_P4_FSB_ESCR1}, | 187 | { { CTR_BPU_2, MSR_P4_FSB_ESCR1}, |
| 187 | { 0, 0 } } | 188 | { 0, 0 } } |
| 188 | }, | 189 | }, |
| 189 | 190 | ||
| 190 | { /* FSB_DATA_ACTIVITY */ | 191 | { /* FSB_DATA_ACTIVITY */ |
| 191 | 0x06, 0x17, | 192 | 0x06, 0x17, |
| 192 | { { CTR_BPU_0, MSR_P4_FSB_ESCR0}, | 193 | { { CTR_BPU_0, MSR_P4_FSB_ESCR0}, |
| 193 | { CTR_BPU_2, MSR_P4_FSB_ESCR1} } | 194 | { CTR_BPU_2, MSR_P4_FSB_ESCR1} } |
| 194 | }, | 195 | }, |
| 195 | 196 | ||
| 196 | { /* BSQ_ALLOCATION */ | 197 | { /* BSQ_ALLOCATION */ |
| 197 | 0x07, 0x05, | 198 | 0x07, 0x05, |
| 198 | { { CTR_BPU_0, MSR_P4_BSU_ESCR0}, | 199 | { { CTR_BPU_0, MSR_P4_BSU_ESCR0}, |
| 199 | { 0, 0 } } | 200 | { 0, 0 } } |
| 200 | }, | 201 | }, |
| 201 | 202 | ||
| 202 | { /* BSQ_ACTIVE_ENTRIES */ | 203 | { /* BSQ_ACTIVE_ENTRIES */ |
| 203 | 0x07, 0x06, | 204 | 0x07, 0x06, |
| 204 | { { CTR_BPU_2, MSR_P4_BSU_ESCR1 /* guess */}, | 205 | { { CTR_BPU_2, MSR_P4_BSU_ESCR1 /* guess */}, |
| 205 | { 0, 0 } } | 206 | { 0, 0 } } |
| 206 | }, | 207 | }, |
| 207 | 208 | ||
| 208 | { /* X87_ASSIST */ | 209 | { /* X87_ASSIST */ |
| 209 | 0x05, 0x03, | 210 | 0x05, 0x03, |
| 210 | { { CTR_IQ_4, MSR_P4_CRU_ESCR2}, | 211 | { { CTR_IQ_4, MSR_P4_CRU_ESCR2}, |
| 211 | { CTR_IQ_5, MSR_P4_CRU_ESCR3} } | 212 | { CTR_IQ_5, MSR_P4_CRU_ESCR3} } |
| 212 | }, | 213 | }, |
| @@ -216,21 +217,21 @@ static struct p4_event_binding p4_events[NUM_EVENTS] = { | |||
| 216 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, | 217 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, |
| 217 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } | 218 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } |
| 218 | }, | 219 | }, |
| 219 | 220 | ||
| 220 | { /* PACKED_SP_UOP */ | 221 | { /* PACKED_SP_UOP */ |
| 221 | 0x01, 0x08, | 222 | 0x01, 0x08, |
| 222 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, | 223 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, |
| 223 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } | 224 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } |
| 224 | }, | 225 | }, |
| 225 | 226 | ||
| 226 | { /* PACKED_DP_UOP */ | 227 | { /* PACKED_DP_UOP */ |
| 227 | 0x01, 0x0c, | 228 | 0x01, 0x0c, |
| 228 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, | 229 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, |
| 229 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } | 230 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } |
| 230 | }, | 231 | }, |
| 231 | 232 | ||
| 232 | { /* SCALAR_SP_UOP */ | 233 | { /* SCALAR_SP_UOP */ |
| 233 | 0x01, 0x0a, | 234 | 0x01, 0x0a, |
| 234 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, | 235 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, |
| 235 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } | 236 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } |
| 236 | }, | 237 | }, |
| @@ -242,31 +243,31 @@ static struct p4_event_binding p4_events[NUM_EVENTS] = { | |||
| 242 | }, | 243 | }, |
| 243 | 244 | ||
| 244 | { /* 64BIT_MMX_UOP */ | 245 | { /* 64BIT_MMX_UOP */ |
| 245 | 0x01, 0x02, | 246 | 0x01, 0x02, |
| 246 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, | 247 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, |
| 247 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } | 248 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } |
| 248 | }, | 249 | }, |
| 249 | 250 | ||
| 250 | { /* 128BIT_MMX_UOP */ | 251 | { /* 128BIT_MMX_UOP */ |
| 251 | 0x01, 0x1a, | 252 | 0x01, 0x1a, |
| 252 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, | 253 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, |
| 253 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } | 254 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } |
| 254 | }, | 255 | }, |
| 255 | 256 | ||
| 256 | { /* X87_FP_UOP */ | 257 | { /* X87_FP_UOP */ |
| 257 | 0x01, 0x04, | 258 | 0x01, 0x04, |
| 258 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, | 259 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, |
| 259 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } | 260 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } |
| 260 | }, | 261 | }, |
| 261 | 262 | ||
| 262 | { /* X87_SIMD_MOVES_UOP */ | 263 | { /* X87_SIMD_MOVES_UOP */ |
| 263 | 0x01, 0x2e, | 264 | 0x01, 0x2e, |
| 264 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, | 265 | { { CTR_FLAME_0, MSR_P4_FIRM_ESCR0}, |
| 265 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } | 266 | { CTR_FLAME_2, MSR_P4_FIRM_ESCR1} } |
| 266 | }, | 267 | }, |
| 267 | 268 | ||
| 268 | { /* MACHINE_CLEAR */ | 269 | { /* MACHINE_CLEAR */ |
| 269 | 0x05, 0x02, | 270 | 0x05, 0x02, |
| 270 | { { CTR_IQ_4, MSR_P4_CRU_ESCR2}, | 271 | { { CTR_IQ_4, MSR_P4_CRU_ESCR2}, |
| 271 | { CTR_IQ_5, MSR_P4_CRU_ESCR3} } | 272 | { CTR_IQ_5, MSR_P4_CRU_ESCR3} } |
| 272 | }, | 273 | }, |
| @@ -276,9 +277,9 @@ static struct p4_event_binding p4_events[NUM_EVENTS] = { | |||
| 276 | { { CTR_BPU_0, MSR_P4_FSB_ESCR0}, | 277 | { { CTR_BPU_0, MSR_P4_FSB_ESCR0}, |
| 277 | { CTR_BPU_2, MSR_P4_FSB_ESCR1} } | 278 | { CTR_BPU_2, MSR_P4_FSB_ESCR1} } |
| 278 | }, | 279 | }, |
| 279 | 280 | ||
| 280 | { /* TC_MS_XFER */ | 281 | { /* TC_MS_XFER */ |
| 281 | 0x00, 0x05, | 282 | 0x00, 0x05, |
| 282 | { { CTR_MS_0, MSR_P4_MS_ESCR0}, | 283 | { { CTR_MS_0, MSR_P4_MS_ESCR0}, |
| 283 | { CTR_MS_2, MSR_P4_MS_ESCR1} } | 284 | { CTR_MS_2, MSR_P4_MS_ESCR1} } |
| 284 | }, | 285 | }, |
| @@ -308,7 +309,7 @@ static struct p4_event_binding p4_events[NUM_EVENTS] = { | |||
| 308 | }, | 309 | }, |
| 309 | 310 | ||
| 310 | { /* INSTR_RETIRED */ | 311 | { /* INSTR_RETIRED */ |
| 311 | 0x04, 0x02, | 312 | 0x04, 0x02, |
| 312 | { { CTR_IQ_4, MSR_P4_CRU_ESCR0}, | 313 | { { CTR_IQ_4, MSR_P4_CRU_ESCR0}, |
| 313 | { CTR_IQ_5, MSR_P4_CRU_ESCR1} } | 314 | { CTR_IQ_5, MSR_P4_CRU_ESCR1} } |
| 314 | }, | 315 | }, |
| @@ -319,14 +320,14 @@ static struct p4_event_binding p4_events[NUM_EVENTS] = { | |||
| 319 | { CTR_IQ_5, MSR_P4_CRU_ESCR1} } | 320 | { CTR_IQ_5, MSR_P4_CRU_ESCR1} } |
| 320 | }, | 321 | }, |
| 321 | 322 | ||
| 322 | { /* UOP_TYPE */ | 323 | { /* UOP_TYPE */ |
| 323 | 0x02, 0x02, | 324 | 0x02, 0x02, |
| 324 | { { CTR_IQ_4, MSR_P4_RAT_ESCR0}, | 325 | { { CTR_IQ_4, MSR_P4_RAT_ESCR0}, |
| 325 | { CTR_IQ_5, MSR_P4_RAT_ESCR1} } | 326 | { CTR_IQ_5, MSR_P4_RAT_ESCR1} } |
| 326 | }, | 327 | }, |
| 327 | 328 | ||
| 328 | { /* RETIRED_MISPRED_BRANCH_TYPE */ | 329 | { /* RETIRED_MISPRED_BRANCH_TYPE */ |
| 329 | 0x02, 0x05, | 330 | 0x02, 0x05, |
| 330 | { { CTR_MS_0, MSR_P4_TBPU_ESCR0}, | 331 | { { CTR_MS_0, MSR_P4_TBPU_ESCR0}, |
| 331 | { CTR_MS_2, MSR_P4_TBPU_ESCR1} } | 332 | { CTR_MS_2, MSR_P4_TBPU_ESCR1} } |
| 332 | }, | 333 | }, |
| @@ -349,8 +350,8 @@ static struct p4_event_binding p4_events[NUM_EVENTS] = { | |||
| 349 | #define ESCR_SET_OS_1(escr, os) ((escr) |= (((os) & 1) << 1)) | 350 | #define ESCR_SET_OS_1(escr, os) ((escr) |= (((os) & 1) << 1)) |
| 350 | #define ESCR_SET_EVENT_SELECT(escr, sel) ((escr) |= (((sel) & 0x3f) << 25)) | 351 | #define ESCR_SET_EVENT_SELECT(escr, sel) ((escr) |= (((sel) & 0x3f) << 25)) |
| 351 | #define ESCR_SET_EVENT_MASK(escr, mask) ((escr) |= (((mask) & 0xffff) << 9)) | 352 | #define ESCR_SET_EVENT_MASK(escr, mask) ((escr) |= (((mask) & 0xffff) << 9)) |
| 352 | #define ESCR_READ(escr,high,ev,i) do {rdmsr(ev->bindings[(i)].escr_address, (escr), (high));} while (0) | 353 | #define ESCR_READ(escr, high, ev, i) do {rdmsr(ev->bindings[(i)].escr_address, (escr), (high)); } while (0) |
| 353 | #define ESCR_WRITE(escr,high,ev,i) do {wrmsr(ev->bindings[(i)].escr_address, (escr), (high));} while (0) | 354 | #define ESCR_WRITE(escr, high, ev, i) do {wrmsr(ev->bindings[(i)].escr_address, (escr), (high)); } while (0) |
| 354 | 355 | ||
| 355 | #define CCCR_RESERVED_BITS 0x38030FFF | 356 | #define CCCR_RESERVED_BITS 0x38030FFF |
| 356 | #define CCCR_CLEAR(cccr) ((cccr) &= CCCR_RESERVED_BITS) | 357 | #define CCCR_CLEAR(cccr) ((cccr) &= CCCR_RESERVED_BITS) |
| @@ -360,15 +361,15 @@ static struct p4_event_binding p4_events[NUM_EVENTS] = { | |||
| 360 | #define CCCR_SET_PMI_OVF_1(cccr) ((cccr) |= (1<<27)) | 361 | #define CCCR_SET_PMI_OVF_1(cccr) ((cccr) |= (1<<27)) |
| 361 | #define CCCR_SET_ENABLE(cccr) ((cccr) |= (1<<12)) | 362 | #define CCCR_SET_ENABLE(cccr) ((cccr) |= (1<<12)) |
| 362 | #define CCCR_SET_DISABLE(cccr) ((cccr) &= ~(1<<12)) | 363 | #define CCCR_SET_DISABLE(cccr) ((cccr) &= ~(1<<12)) |
| 363 | #define CCCR_READ(low, high, i) do {rdmsr(p4_counters[(i)].cccr_address, (low), (high));} while (0) | 364 | #define CCCR_READ(low, high, i) do {rdmsr(p4_counters[(i)].cccr_address, (low), (high)); } while (0) |
| 364 | #define CCCR_WRITE(low, high, i) do {wrmsr(p4_counters[(i)].cccr_address, (low), (high));} while (0) | 365 | #define CCCR_WRITE(low, high, i) do {wrmsr(p4_counters[(i)].cccr_address, (low), (high)); } while (0) |
| 365 | #define CCCR_OVF_P(cccr) ((cccr) & (1U<<31)) | 366 | #define CCCR_OVF_P(cccr) ((cccr) & (1U<<31)) |
| 366 | #define CCCR_CLEAR_OVF(cccr) ((cccr) &= (~(1U<<31))) | 367 | #define CCCR_CLEAR_OVF(cccr) ((cccr) &= (~(1U<<31))) |
| 367 | 368 | ||
| 368 | #define CTRL_IS_RESERVED(msrs,c) (msrs->controls[(c)].addr ? 1 : 0) | 369 | #define CTRL_IS_RESERVED(msrs, c) (msrs->controls[(c)].addr ? 1 : 0) |
| 369 | #define CTR_IS_RESERVED(msrs,c) (msrs->counters[(c)].addr ? 1 : 0) | 370 | #define CTR_IS_RESERVED(msrs, c) (msrs->counters[(c)].addr ? 1 : 0) |
| 370 | #define CTR_READ(l,h,i) do {rdmsr(p4_counters[(i)].counter_address, (l), (h));} while (0) | 371 | #define CTR_READ(l, h, i) do {rdmsr(p4_counters[(i)].counter_address, (l), (h)); } while (0) |
| 371 | #define CTR_WRITE(l,i) do {wrmsr(p4_counters[(i)].counter_address, -(u32)(l), -1);} while (0) | 372 | #define CTR_WRITE(l, i) do {wrmsr(p4_counters[(i)].counter_address, -(u32)(l), -1); } while (0) |
| 372 | #define CTR_OVERFLOW_P(ctr) (!((ctr) & 0x80000000)) | 373 | #define CTR_OVERFLOW_P(ctr) (!((ctr) & 0x80000000)) |
| 373 | 374 | ||
| 374 | 375 | ||
| @@ -380,7 +381,7 @@ static unsigned int get_stagger(void) | |||
| 380 | #ifdef CONFIG_SMP | 381 | #ifdef CONFIG_SMP |
| 381 | int cpu = smp_processor_id(); | 382 | int cpu = smp_processor_id(); |
| 382 | return (cpu != first_cpu(per_cpu(cpu_sibling_map, cpu))); | 383 | return (cpu != first_cpu(per_cpu(cpu_sibling_map, cpu))); |
| 383 | #endif | 384 | #endif |
| 384 | return 0; | 385 | return 0; |
| 385 | } | 386 | } |
| 386 | 387 | ||
| @@ -395,25 +396,23 @@ static unsigned long reset_value[NUM_COUNTERS_NON_HT]; | |||
| 395 | 396 | ||
| 396 | static void p4_fill_in_addresses(struct op_msrs * const msrs) | 397 | static void p4_fill_in_addresses(struct op_msrs * const msrs) |
| 397 | { | 398 | { |
| 398 | unsigned int i; | 399 | unsigned int i; |
| 399 | unsigned int addr, cccraddr, stag; | 400 | unsigned int addr, cccraddr, stag; |
| 400 | 401 | ||
| 401 | setup_num_counters(); | 402 | setup_num_counters(); |
| 402 | stag = get_stagger(); | 403 | stag = get_stagger(); |
| 403 | 404 | ||
| 404 | /* initialize some registers */ | 405 | /* initialize some registers */ |
| 405 | for (i = 0; i < num_counters; ++i) { | 406 | for (i = 0; i < num_counters; ++i) |
| 406 | msrs->counters[i].addr = 0; | 407 | msrs->counters[i].addr = 0; |
| 407 | } | 408 | for (i = 0; i < num_controls; ++i) |
| 408 | for (i = 0; i < num_controls; ++i) { | ||
| 409 | msrs->controls[i].addr = 0; | 409 | msrs->controls[i].addr = 0; |
| 410 | } | 410 | |
| 411 | |||
| 412 | /* the counter & cccr registers we pay attention to */ | 411 | /* the counter & cccr registers we pay attention to */ |
| 413 | for (i = 0; i < num_counters; ++i) { | 412 | for (i = 0; i < num_counters; ++i) { |
| 414 | addr = p4_counters[VIRT_CTR(stag, i)].counter_address; | 413 | addr = p4_counters[VIRT_CTR(stag, i)].counter_address; |
| 415 | cccraddr = p4_counters[VIRT_CTR(stag, i)].cccr_address; | 414 | cccraddr = p4_counters[VIRT_CTR(stag, i)].cccr_address; |
| 416 | if (reserve_perfctr_nmi(addr)){ | 415 | if (reserve_perfctr_nmi(addr)) { |
| 417 | msrs->counters[i].addr = addr; | 416 | msrs->counters[i].addr = addr; |
| 418 | msrs->controls[i].addr = cccraddr; | 417 | msrs->controls[i].addr = cccraddr; |
| 419 | } | 418 | } |
| @@ -447,22 +446,22 @@ static void p4_fill_in_addresses(struct op_msrs * const msrs) | |||
| 447 | if (reserve_evntsel_nmi(addr)) | 446 | if (reserve_evntsel_nmi(addr)) |
| 448 | msrs->controls[i].addr = addr; | 447 | msrs->controls[i].addr = addr; |
| 449 | } | 448 | } |
| 450 | 449 | ||
| 451 | for (addr = MSR_P4_MS_ESCR0 + stag; | 450 | for (addr = MSR_P4_MS_ESCR0 + stag; |
| 452 | addr <= MSR_P4_TC_ESCR1; ++i, addr += addr_increment()) { | 451 | addr <= MSR_P4_TC_ESCR1; ++i, addr += addr_increment()) { |
| 453 | if (reserve_evntsel_nmi(addr)) | 452 | if (reserve_evntsel_nmi(addr)) |
| 454 | msrs->controls[i].addr = addr; | 453 | msrs->controls[i].addr = addr; |
| 455 | } | 454 | } |
| 456 | 455 | ||
| 457 | for (addr = MSR_P4_IX_ESCR0 + stag; | 456 | for (addr = MSR_P4_IX_ESCR0 + stag; |
| 458 | addr <= MSR_P4_CRU_ESCR3; ++i, addr += addr_increment()) { | 457 | addr <= MSR_P4_CRU_ESCR3; ++i, addr += addr_increment()) { |
| 459 | if (reserve_evntsel_nmi(addr)) | 458 | if (reserve_evntsel_nmi(addr)) |
| 460 | msrs->controls[i].addr = addr; | 459 | msrs->controls[i].addr = addr; |
| 461 | } | 460 | } |
| 462 | 461 | ||
| 463 | /* there are 2 remaining non-contiguously located ESCRs */ | 462 | /* there are 2 remaining non-contiguously located ESCRs */ |
| 464 | 463 | ||
| 465 | if (num_counters == NUM_COUNTERS_NON_HT) { | 464 | if (num_counters == NUM_COUNTERS_NON_HT) { |
| 466 | /* standard non-HT CPUs handle both remaining ESCRs*/ | 465 | /* standard non-HT CPUs handle both remaining ESCRs*/ |
| 467 | if (reserve_evntsel_nmi(MSR_P4_CRU_ESCR5)) | 466 | if (reserve_evntsel_nmi(MSR_P4_CRU_ESCR5)) |
| 468 | msrs->controls[i++].addr = MSR_P4_CRU_ESCR5; | 467 | msrs->controls[i++].addr = MSR_P4_CRU_ESCR5; |
| @@ -498,20 +497,20 @@ static void pmc_setup_one_p4_counter(unsigned int ctr) | |||
| 498 | unsigned int stag; | 497 | unsigned int stag; |
| 499 | 498 | ||
| 500 | stag = get_stagger(); | 499 | stag = get_stagger(); |
| 501 | 500 | ||
| 502 | /* convert from counter *number* to counter *bit* */ | 501 | /* convert from counter *number* to counter *bit* */ |
| 503 | counter_bit = 1 << VIRT_CTR(stag, ctr); | 502 | counter_bit = 1 << VIRT_CTR(stag, ctr); |
| 504 | 503 | ||
| 505 | /* find our event binding structure. */ | 504 | /* find our event binding structure. */ |
| 506 | if (counter_config[ctr].event <= 0 || counter_config[ctr].event > NUM_EVENTS) { | 505 | if (counter_config[ctr].event <= 0 || counter_config[ctr].event > NUM_EVENTS) { |
| 507 | printk(KERN_ERR | 506 | printk(KERN_ERR |
| 508 | "oprofile: P4 event code 0x%lx out of range\n", | 507 | "oprofile: P4 event code 0x%lx out of range\n", |
| 509 | counter_config[ctr].event); | 508 | counter_config[ctr].event); |
| 510 | return; | 509 | return; |
| 511 | } | 510 | } |
| 512 | 511 | ||
| 513 | ev = &(p4_events[counter_config[ctr].event - 1]); | 512 | ev = &(p4_events[counter_config[ctr].event - 1]); |
| 514 | 513 | ||
| 515 | for (i = 0; i < maxbind; i++) { | 514 | for (i = 0; i < maxbind; i++) { |
| 516 | if (ev->bindings[i].virt_counter & counter_bit) { | 515 | if (ev->bindings[i].virt_counter & counter_bit) { |
| 517 | 516 | ||
| @@ -526,25 +525,24 @@ static void pmc_setup_one_p4_counter(unsigned int ctr) | |||
| 526 | ESCR_SET_OS_1(escr, counter_config[ctr].kernel); | 525 | ESCR_SET_OS_1(escr, counter_config[ctr].kernel); |
| 527 | } | 526 | } |
| 528 | ESCR_SET_EVENT_SELECT(escr, ev->event_select); | 527 | ESCR_SET_EVENT_SELECT(escr, ev->event_select); |
| 529 | ESCR_SET_EVENT_MASK(escr, counter_config[ctr].unit_mask); | 528 | ESCR_SET_EVENT_MASK(escr, counter_config[ctr].unit_mask); |
| 530 | ESCR_WRITE(escr, high, ev, i); | 529 | ESCR_WRITE(escr, high, ev, i); |
| 531 | 530 | ||
| 532 | /* modify CCCR */ | 531 | /* modify CCCR */ |
| 533 | CCCR_READ(cccr, high, VIRT_CTR(stag, ctr)); | 532 | CCCR_READ(cccr, high, VIRT_CTR(stag, ctr)); |
| 534 | CCCR_CLEAR(cccr); | 533 | CCCR_CLEAR(cccr); |
| 535 | CCCR_SET_REQUIRED_BITS(cccr); | 534 | CCCR_SET_REQUIRED_BITS(cccr); |
| 536 | CCCR_SET_ESCR_SELECT(cccr, ev->escr_select); | 535 | CCCR_SET_ESCR_SELECT(cccr, ev->escr_select); |
| 537 | if (stag == 0) { | 536 | if (stag == 0) |
| 538 | CCCR_SET_PMI_OVF_0(cccr); | 537 | CCCR_SET_PMI_OVF_0(cccr); |
| 539 | } else { | 538 | else |
| 540 | CCCR_SET_PMI_OVF_1(cccr); | 539 | CCCR_SET_PMI_OVF_1(cccr); |
| 541 | } | ||
| 542 | CCCR_WRITE(cccr, high, VIRT_CTR(stag, ctr)); | 540 | CCCR_WRITE(cccr, high, VIRT_CTR(stag, ctr)); |
| 543 | return; | 541 | return; |
| 544 | } | 542 | } |
| 545 | } | 543 | } |
| 546 | 544 | ||
| 547 | printk(KERN_ERR | 545 | printk(KERN_ERR |
| 548 | "oprofile: P4 event code 0x%lx no binding, stag %d ctr %d\n", | 546 | "oprofile: P4 event code 0x%lx no binding, stag %d ctr %d\n", |
| 549 | counter_config[ctr].event, stag, ctr); | 547 | counter_config[ctr].event, stag, ctr); |
| 550 | } | 548 | } |
| @@ -559,14 +557,14 @@ static void p4_setup_ctrs(struct op_msrs const * const msrs) | |||
| 559 | stag = get_stagger(); | 557 | stag = get_stagger(); |
| 560 | 558 | ||
| 561 | rdmsr(MSR_IA32_MISC_ENABLE, low, high); | 559 | rdmsr(MSR_IA32_MISC_ENABLE, low, high); |
| 562 | if (! MISC_PMC_ENABLED_P(low)) { | 560 | if (!MISC_PMC_ENABLED_P(low)) { |
| 563 | printk(KERN_ERR "oprofile: P4 PMC not available\n"); | 561 | printk(KERN_ERR "oprofile: P4 PMC not available\n"); |
| 564 | return; | 562 | return; |
| 565 | } | 563 | } |
| 566 | 564 | ||
| 567 | /* clear the cccrs we will use */ | 565 | /* clear the cccrs we will use */ |
| 568 | for (i = 0 ; i < num_counters ; i++) { | 566 | for (i = 0 ; i < num_counters ; i++) { |
| 569 | if (unlikely(!CTRL_IS_RESERVED(msrs,i))) | 567 | if (unlikely(!CTRL_IS_RESERVED(msrs, i))) |
| 570 | continue; | 568 | continue; |
| 571 | rdmsr(p4_counters[VIRT_CTR(stag, i)].cccr_address, low, high); | 569 | rdmsr(p4_counters[VIRT_CTR(stag, i)].cccr_address, low, high); |
| 572 | CCCR_CLEAR(low); | 570 | CCCR_CLEAR(low); |
| @@ -576,14 +574,14 @@ static void p4_setup_ctrs(struct op_msrs const * const msrs) | |||
| 576 | 574 | ||
| 577 | /* clear all escrs (including those outside our concern) */ | 575 | /* clear all escrs (including those outside our concern) */ |
| 578 | for (i = num_counters; i < num_controls; i++) { | 576 | for (i = num_counters; i < num_controls; i++) { |
| 579 | if (unlikely(!CTRL_IS_RESERVED(msrs,i))) | 577 | if (unlikely(!CTRL_IS_RESERVED(msrs, i))) |
| 580 | continue; | 578 | continue; |
| 581 | wrmsr(msrs->controls[i].addr, 0, 0); | 579 | wrmsr(msrs->controls[i].addr, 0, 0); |
| 582 | } | 580 | } |
| 583 | 581 | ||
| 584 | /* setup all counters */ | 582 | /* setup all counters */ |
| 585 | for (i = 0 ; i < num_counters ; ++i) { | 583 | for (i = 0 ; i < num_counters ; ++i) { |
| 586 | if ((counter_config[i].enabled) && (CTRL_IS_RESERVED(msrs,i))) { | 584 | if ((counter_config[i].enabled) && (CTRL_IS_RESERVED(msrs, i))) { |
| 587 | reset_value[i] = counter_config[i].count; | 585 | reset_value[i] = counter_config[i].count; |
| 588 | pmc_setup_one_p4_counter(i); | 586 | pmc_setup_one_p4_counter(i); |
| 589 | CTR_WRITE(counter_config[i].count, VIRT_CTR(stag, i)); | 587 | CTR_WRITE(counter_config[i].count, VIRT_CTR(stag, i)); |
| @@ -603,11 +601,11 @@ static int p4_check_ctrs(struct pt_regs * const regs, | |||
| 603 | stag = get_stagger(); | 601 | stag = get_stagger(); |
| 604 | 602 | ||
| 605 | for (i = 0; i < num_counters; ++i) { | 603 | for (i = 0; i < num_counters; ++i) { |
| 606 | 604 | ||
| 607 | if (!reset_value[i]) | 605 | if (!reset_value[i]) |
| 608 | continue; | 606 | continue; |
| 609 | 607 | ||
| 610 | /* | 608 | /* |
| 611 | * there is some eccentricity in the hardware which | 609 | * there is some eccentricity in the hardware which |
| 612 | * requires that we perform 2 extra corrections: | 610 | * requires that we perform 2 extra corrections: |
| 613 | * | 611 | * |
| @@ -616,24 +614,24 @@ static int p4_check_ctrs(struct pt_regs * const regs, | |||
| 616 | * | 614 | * |
| 617 | * - write the counter back twice to ensure it gets | 615 | * - write the counter back twice to ensure it gets |
| 618 | * updated properly. | 616 | * updated properly. |
| 619 | * | 617 | * |
| 620 | * the former seems to be related to extra NMIs happening | 618 | * the former seems to be related to extra NMIs happening |
| 621 | * during the current NMI; the latter is reported as errata | 619 | * during the current NMI; the latter is reported as errata |
| 622 | * N15 in intel doc 249199-029, pentium 4 specification | 620 | * N15 in intel doc 249199-029, pentium 4 specification |
| 623 | * update, though their suggested work-around does not | 621 | * update, though their suggested work-around does not |
| 624 | * appear to solve the problem. | 622 | * appear to solve the problem. |
| 625 | */ | 623 | */ |
| 626 | 624 | ||
| 627 | real = VIRT_CTR(stag, i); | 625 | real = VIRT_CTR(stag, i); |
| 628 | 626 | ||
| 629 | CCCR_READ(low, high, real); | 627 | CCCR_READ(low, high, real); |
| 630 | CTR_READ(ctr, high, real); | 628 | CTR_READ(ctr, high, real); |
| 631 | if (CCCR_OVF_P(low) || CTR_OVERFLOW_P(ctr)) { | 629 | if (CCCR_OVF_P(low) || CTR_OVERFLOW_P(ctr)) { |
| 632 | oprofile_add_sample(regs, i); | 630 | oprofile_add_sample(regs, i); |
| 633 | CTR_WRITE(reset_value[i], real); | 631 | CTR_WRITE(reset_value[i], real); |
| 634 | CCCR_CLEAR_OVF(low); | 632 | CCCR_CLEAR_OVF(low); |
| 635 | CCCR_WRITE(low, high, real); | 633 | CCCR_WRITE(low, high, real); |
| 636 | CTR_WRITE(reset_value[i], real); | 634 | CTR_WRITE(reset_value[i], real); |
| 637 | } | 635 | } |
| 638 | } | 636 | } |
| 639 | 637 | ||
| @@ -683,15 +681,16 @@ static void p4_shutdown(struct op_msrs const * const msrs) | |||
| 683 | int i; | 681 | int i; |
| 684 | 682 | ||
| 685 | for (i = 0 ; i < num_counters ; ++i) { | 683 | for (i = 0 ; i < num_counters ; ++i) { |
| 686 | if (CTR_IS_RESERVED(msrs,i)) | 684 | if (CTR_IS_RESERVED(msrs, i)) |
| 687 | release_perfctr_nmi(msrs->counters[i].addr); | 685 | release_perfctr_nmi(msrs->counters[i].addr); |
| 688 | } | 686 | } |
| 689 | /* some of the control registers are specially reserved in | 687 | /* |
| 688 | * some of the control registers are specially reserved in | ||
| 690 | * conjunction with the counter registers (hence the starting offset). | 689 | * conjunction with the counter registers (hence the starting offset). |
| 691 | * This saves a few bits. | 690 | * This saves a few bits. |
| 692 | */ | 691 | */ |
| 693 | for (i = num_counters ; i < num_controls ; ++i) { | 692 | for (i = num_counters ; i < num_controls ; ++i) { |
| 694 | if (CTRL_IS_RESERVED(msrs,i)) | 693 | if (CTRL_IS_RESERVED(msrs, i)) |
| 695 | release_evntsel_nmi(msrs->controls[i].addr); | 694 | release_evntsel_nmi(msrs->controls[i].addr); |
| 696 | } | 695 | } |
| 697 | } | 696 | } |
diff --git a/arch/x86/pci/amd_bus.c b/arch/x86/pci/amd_bus.c index 6a0fca78c362..22e057665e55 100644 --- a/arch/x86/pci/amd_bus.c +++ b/arch/x86/pci/amd_bus.c | |||
| @@ -580,7 +580,7 @@ static int __cpuinit amd_cpu_notify(struct notifier_block *self, | |||
| 580 | unsigned long action, void *hcpu) | 580 | unsigned long action, void *hcpu) |
| 581 | { | 581 | { |
| 582 | int cpu = (long)hcpu; | 582 | int cpu = (long)hcpu; |
| 583 | switch(action) { | 583 | switch (action) { |
| 584 | case CPU_ONLINE: | 584 | case CPU_ONLINE: |
| 585 | case CPU_ONLINE_FROZEN: | 585 | case CPU_ONLINE_FROZEN: |
| 586 | smp_call_function_single(cpu, enable_pci_io_ecs, NULL, 0); | 586 | smp_call_function_single(cpu, enable_pci_io_ecs, NULL, 0); |
diff --git a/arch/x86/pci/irq.c b/arch/x86/pci/irq.c index 8e077185e185..006599db0dc7 100644 --- a/arch/x86/pci/irq.c +++ b/arch/x86/pci/irq.c | |||
| @@ -1043,35 +1043,44 @@ static void __init pcibios_fixup_irqs(void) | |||
| 1043 | if (io_apic_assign_pci_irqs) { | 1043 | if (io_apic_assign_pci_irqs) { |
| 1044 | int irq; | 1044 | int irq; |
| 1045 | 1045 | ||
| 1046 | if (pin) { | 1046 | if (!pin) |
| 1047 | /* | 1047 | continue; |
| 1048 | * interrupt pins are numbered starting | 1048 | |
| 1049 | * from 1 | 1049 | /* |
| 1050 | */ | 1050 | * interrupt pins are numbered starting from 1 |
| 1051 | pin--; | 1051 | */ |
| 1052 | irq = IO_APIC_get_PCI_irq_vector(dev->bus->number, | 1052 | pin--; |
| 1053 | PCI_SLOT(dev->devfn), pin); | 1053 | irq = IO_APIC_get_PCI_irq_vector(dev->bus->number, |
| 1054 | /* | 1054 | PCI_SLOT(dev->devfn), pin); |
| 1055 | * Busses behind bridges are typically not listed in the MP-table. | 1055 | /* |
| 1056 | * In this case we have to look up the IRQ based on the parent bus, | 1056 | * Busses behind bridges are typically not listed in the |
| 1057 | * parent slot, and pin number. The SMP code detects such bridged | 1057 | * MP-table. In this case we have to look up the IRQ |
| 1058 | * busses itself so we should get into this branch reliably. | 1058 | * based on the parent bus, parent slot, and pin number. |
| 1059 | */ | 1059 | * The SMP code detects such bridged busses itself so we |
| 1060 | if (irq < 0 && dev->bus->parent) { /* go back to the bridge */ | 1060 | * should get into this branch reliably. |
| 1061 | struct pci_dev *bridge = dev->bus->self; | 1061 | */ |
| 1062 | 1062 | if (irq < 0 && dev->bus->parent) { | |
| 1063 | pin = (pin + PCI_SLOT(dev->devfn)) % 4; | 1063 | /* go back to the bridge */ |
| 1064 | irq = IO_APIC_get_PCI_irq_vector(bridge->bus->number, | 1064 | struct pci_dev *bridge = dev->bus->self; |
| 1065 | PCI_SLOT(bridge->devfn), pin); | 1065 | int bus; |
| 1066 | if (irq >= 0) | 1066 | |
| 1067 | dev_warn(&dev->dev, "using bridge %s INT %c to get IRQ %d\n", | 1067 | pin = (pin + PCI_SLOT(dev->devfn)) % 4; |
| 1068 | pci_name(bridge), | 1068 | bus = bridge->bus->number; |
| 1069 | 'A' + pin, irq); | 1069 | irq = IO_APIC_get_PCI_irq_vector(bus, |
| 1070 | } | 1070 | PCI_SLOT(bridge->devfn), pin); |
| 1071 | if (irq >= 0) { | 1071 | if (irq >= 0) |
| 1072 | dev_info(&dev->dev, "PCI->APIC IRQ transform: INT %c -> IRQ %d\n", 'A' + pin, irq); | 1072 | dev_warn(&dev->dev, |
| 1073 | dev->irq = irq; | 1073 | "using bridge %s INT %c to " |
| 1074 | } | 1074 | "get IRQ %d\n", |
| 1075 | pci_name(bridge), | ||
| 1076 | 'A' + pin, irq); | ||
| 1077 | } | ||
| 1078 | if (irq >= 0) { | ||
| 1079 | dev_info(&dev->dev, | ||
| 1080 | "PCI->APIC IRQ transform: INT %c " | ||
| 1081 | "-> IRQ %d\n", | ||
| 1082 | 'A' + pin, irq); | ||
| 1083 | dev->irq = irq; | ||
| 1075 | } | 1084 | } |
| 1076 | } | 1085 | } |
| 1077 | #endif | 1086 | #endif |
diff --git a/arch/x86/power/hibernate_asm_32.S b/arch/x86/power/hibernate_asm_32.S index 4fc7e872c85e..d1e9b53f9d33 100644 --- a/arch/x86/power/hibernate_asm_32.S +++ b/arch/x86/power/hibernate_asm_32.S | |||
| @@ -1,5 +1,3 @@ | |||
| 1 | .text | ||
| 2 | |||
| 3 | /* | 1 | /* |
| 4 | * This may not use any stack, nor any variable that is not "NoSave": | 2 | * This may not use any stack, nor any variable that is not "NoSave": |
| 5 | * | 3 | * |
| @@ -12,17 +10,18 @@ | |||
| 12 | #include <asm/segment.h> | 10 | #include <asm/segment.h> |
| 13 | #include <asm/page.h> | 11 | #include <asm/page.h> |
| 14 | #include <asm/asm-offsets.h> | 12 | #include <asm/asm-offsets.h> |
| 13 | #include <asm/processor-flags.h> | ||
| 15 | 14 | ||
| 16 | .text | 15 | .text |
| 17 | 16 | ||
| 18 | ENTRY(swsusp_arch_suspend) | 17 | ENTRY(swsusp_arch_suspend) |
| 19 | |||
| 20 | movl %esp, saved_context_esp | 18 | movl %esp, saved_context_esp |
| 21 | movl %ebx, saved_context_ebx | 19 | movl %ebx, saved_context_ebx |
| 22 | movl %ebp, saved_context_ebp | 20 | movl %ebp, saved_context_ebp |
| 23 | movl %esi, saved_context_esi | 21 | movl %esi, saved_context_esi |
| 24 | movl %edi, saved_context_edi | 22 | movl %edi, saved_context_edi |
| 25 | pushfl ; popl saved_context_eflags | 23 | pushfl |
| 24 | popl saved_context_eflags | ||
| 26 | 25 | ||
| 27 | call swsusp_save | 26 | call swsusp_save |
| 28 | ret | 27 | ret |
| @@ -59,7 +58,7 @@ done: | |||
| 59 | movl mmu_cr4_features, %ecx | 58 | movl mmu_cr4_features, %ecx |
| 60 | jecxz 1f # cr4 Pentium and higher, skip if zero | 59 | jecxz 1f # cr4 Pentium and higher, skip if zero |
| 61 | movl %ecx, %edx | 60 | movl %ecx, %edx |
| 62 | andl $~(1<<7), %edx; # PGE | 61 | andl $~(X86_CR4_PGE), %edx |
| 63 | movl %edx, %cr4; # turn off PGE | 62 | movl %edx, %cr4; # turn off PGE |
| 64 | 1: | 63 | 1: |
| 65 | movl %cr3, %eax; # flush TLB | 64 | movl %cr3, %eax; # flush TLB |
| @@ -74,7 +73,8 @@ done: | |||
| 74 | movl saved_context_esi, %esi | 73 | movl saved_context_esi, %esi |
| 75 | movl saved_context_edi, %edi | 74 | movl saved_context_edi, %edi |
| 76 | 75 | ||
| 77 | pushl saved_context_eflags ; popfl | 76 | pushl saved_context_eflags |
| 77 | popfl | ||
| 78 | 78 | ||
| 79 | xorl %eax, %eax | 79 | xorl %eax, %eax |
| 80 | 80 | ||
diff --git a/arch/x86/xen/enlighten.c b/arch/x86/xen/enlighten.c index a4e201b47f64..7dcd321a0508 100644 --- a/arch/x86/xen/enlighten.c +++ b/arch/x86/xen/enlighten.c | |||
| @@ -812,7 +812,7 @@ static int xen_write_msr_safe(unsigned int msr, unsigned low, unsigned high) | |||
| 812 | 812 | ||
| 813 | /* Early in boot, while setting up the initial pagetable, assume | 813 | /* Early in boot, while setting up the initial pagetable, assume |
| 814 | everything is pinned. */ | 814 | everything is pinned. */ |
| 815 | static __init void xen_alloc_pte_init(struct mm_struct *mm, u32 pfn) | 815 | static __init void xen_alloc_pte_init(struct mm_struct *mm, unsigned long pfn) |
| 816 | { | 816 | { |
| 817 | #ifdef CONFIG_FLATMEM | 817 | #ifdef CONFIG_FLATMEM |
| 818 | BUG_ON(mem_map); /* should only be used early */ | 818 | BUG_ON(mem_map); /* should only be used early */ |
| @@ -822,7 +822,7 @@ static __init void xen_alloc_pte_init(struct mm_struct *mm, u32 pfn) | |||
| 822 | 822 | ||
| 823 | /* Early release_pte assumes that all pts are pinned, since there's | 823 | /* Early release_pte assumes that all pts are pinned, since there's |
| 824 | only init_mm and anything attached to that is pinned. */ | 824 | only init_mm and anything attached to that is pinned. */ |
| 825 | static void xen_release_pte_init(u32 pfn) | 825 | static void xen_release_pte_init(unsigned long pfn) |
| 826 | { | 826 | { |
| 827 | make_lowmem_page_readwrite(__va(PFN_PHYS(pfn))); | 827 | make_lowmem_page_readwrite(__va(PFN_PHYS(pfn))); |
| 828 | } | 828 | } |
| @@ -838,7 +838,7 @@ static void pin_pagetable_pfn(unsigned cmd, unsigned long pfn) | |||
| 838 | 838 | ||
| 839 | /* This needs to make sure the new pte page is pinned iff its being | 839 | /* This needs to make sure the new pte page is pinned iff its being |
| 840 | attached to a pinned pagetable. */ | 840 | attached to a pinned pagetable. */ |
| 841 | static void xen_alloc_ptpage(struct mm_struct *mm, u32 pfn, unsigned level) | 841 | static void xen_alloc_ptpage(struct mm_struct *mm, unsigned long pfn, unsigned level) |
| 842 | { | 842 | { |
| 843 | struct page *page = pfn_to_page(pfn); | 843 | struct page *page = pfn_to_page(pfn); |
| 844 | 844 | ||
| @@ -856,12 +856,12 @@ static void xen_alloc_ptpage(struct mm_struct *mm, u32 pfn, unsigned level) | |||
| 856 | } | 856 | } |
| 857 | } | 857 | } |
| 858 | 858 | ||
| 859 | static void xen_alloc_pte(struct mm_struct *mm, u32 pfn) | 859 | static void xen_alloc_pte(struct mm_struct *mm, unsigned long pfn) |
| 860 | { | 860 | { |
| 861 | xen_alloc_ptpage(mm, pfn, PT_PTE); | 861 | xen_alloc_ptpage(mm, pfn, PT_PTE); |
| 862 | } | 862 | } |
| 863 | 863 | ||
| 864 | static void xen_alloc_pmd(struct mm_struct *mm, u32 pfn) | 864 | static void xen_alloc_pmd(struct mm_struct *mm, unsigned long pfn) |
| 865 | { | 865 | { |
| 866 | xen_alloc_ptpage(mm, pfn, PT_PMD); | 866 | xen_alloc_ptpage(mm, pfn, PT_PMD); |
| 867 | } | 867 | } |
| @@ -909,7 +909,7 @@ static void xen_pgd_free(struct mm_struct *mm, pgd_t *pgd) | |||
| 909 | } | 909 | } |
| 910 | 910 | ||
| 911 | /* This should never happen until we're OK to use struct page */ | 911 | /* This should never happen until we're OK to use struct page */ |
| 912 | static void xen_release_ptpage(u32 pfn, unsigned level) | 912 | static void xen_release_ptpage(unsigned long pfn, unsigned level) |
| 913 | { | 913 | { |
| 914 | struct page *page = pfn_to_page(pfn); | 914 | struct page *page = pfn_to_page(pfn); |
| 915 | 915 | ||
| @@ -923,23 +923,23 @@ static void xen_release_ptpage(u32 pfn, unsigned level) | |||
| 923 | } | 923 | } |
| 924 | } | 924 | } |
| 925 | 925 | ||
| 926 | static void xen_release_pte(u32 pfn) | 926 | static void xen_release_pte(unsigned long pfn) |
| 927 | { | 927 | { |
| 928 | xen_release_ptpage(pfn, PT_PTE); | 928 | xen_release_ptpage(pfn, PT_PTE); |
| 929 | } | 929 | } |
| 930 | 930 | ||
| 931 | static void xen_release_pmd(u32 pfn) | 931 | static void xen_release_pmd(unsigned long pfn) |
| 932 | { | 932 | { |
| 933 | xen_release_ptpage(pfn, PT_PMD); | 933 | xen_release_ptpage(pfn, PT_PMD); |
| 934 | } | 934 | } |
| 935 | 935 | ||
| 936 | #if PAGETABLE_LEVELS == 4 | 936 | #if PAGETABLE_LEVELS == 4 |
| 937 | static void xen_alloc_pud(struct mm_struct *mm, u32 pfn) | 937 | static void xen_alloc_pud(struct mm_struct *mm, unsigned long pfn) |
| 938 | { | 938 | { |
| 939 | xen_alloc_ptpage(mm, pfn, PT_PUD); | 939 | xen_alloc_ptpage(mm, pfn, PT_PUD); |
| 940 | } | 940 | } |
| 941 | 941 | ||
| 942 | static void xen_release_pud(u32 pfn) | 942 | static void xen_release_pud(unsigned long pfn) |
| 943 | { | 943 | { |
| 944 | xen_release_ptpage(pfn, PT_PUD); | 944 | xen_release_ptpage(pfn, PT_PUD); |
| 945 | } | 945 | } |
diff --git a/include/asm-x86/apic.h b/include/asm-x86/apic.h index 4a59f0d6e38c..65590c9aecd4 100644 --- a/include/asm-x86/apic.h +++ b/include/asm-x86/apic.h | |||
| @@ -81,9 +81,7 @@ extern int get_physical_broadcast(void); | |||
| 81 | static inline void ack_APIC_irq(void) | 81 | static inline void ack_APIC_irq(void) |
| 82 | { | 82 | { |
| 83 | /* | 83 | /* |
| 84 | * ack_APIC_irq() actually gets compiled as a single instruction: | 84 | * ack_APIC_irq() actually gets compiled as a single instruction |
| 85 | * - a single rmw on Pentium/82489DX | ||
| 86 | * - a single write on P6+ cores (CONFIG_X86_GOOD_APIC) | ||
| 87 | * ... yummie. | 85 | * ... yummie. |
| 88 | */ | 86 | */ |
| 89 | 87 | ||
diff --git a/include/asm-x86/asm.h b/include/asm-x86/asm.h index 2439ae49e8ac..e1355f44d7c3 100644 --- a/include/asm-x86/asm.h +++ b/include/asm-x86/asm.h | |||
| @@ -20,17 +20,22 @@ | |||
| 20 | 20 | ||
| 21 | #define _ASM_PTR __ASM_SEL(.long, .quad) | 21 | #define _ASM_PTR __ASM_SEL(.long, .quad) |
| 22 | #define _ASM_ALIGN __ASM_SEL(.balign 4, .balign 8) | 22 | #define _ASM_ALIGN __ASM_SEL(.balign 4, .balign 8) |
| 23 | #define _ASM_MOV_UL __ASM_SIZE(mov) | ||
| 24 | 23 | ||
| 24 | #define _ASM_MOV __ASM_SIZE(mov) | ||
| 25 | #define _ASM_INC __ASM_SIZE(inc) | 25 | #define _ASM_INC __ASM_SIZE(inc) |
| 26 | #define _ASM_DEC __ASM_SIZE(dec) | 26 | #define _ASM_DEC __ASM_SIZE(dec) |
| 27 | #define _ASM_ADD __ASM_SIZE(add) | 27 | #define _ASM_ADD __ASM_SIZE(add) |
| 28 | #define _ASM_SUB __ASM_SIZE(sub) | 28 | #define _ASM_SUB __ASM_SIZE(sub) |
| 29 | #define _ASM_XADD __ASM_SIZE(xadd) | 29 | #define _ASM_XADD __ASM_SIZE(xadd) |
| 30 | |||
| 30 | #define _ASM_AX __ASM_REG(ax) | 31 | #define _ASM_AX __ASM_REG(ax) |
| 31 | #define _ASM_BX __ASM_REG(bx) | 32 | #define _ASM_BX __ASM_REG(bx) |
| 32 | #define _ASM_CX __ASM_REG(cx) | 33 | #define _ASM_CX __ASM_REG(cx) |
| 33 | #define _ASM_DX __ASM_REG(dx) | 34 | #define _ASM_DX __ASM_REG(dx) |
| 35 | #define _ASM_SP __ASM_REG(sp) | ||
| 36 | #define _ASM_BP __ASM_REG(bp) | ||
| 37 | #define _ASM_SI __ASM_REG(si) | ||
| 38 | #define _ASM_DI __ASM_REG(di) | ||
| 34 | 39 | ||
| 35 | /* Exception table entry */ | 40 | /* Exception table entry */ |
| 36 | # define _ASM_EXTABLE(from,to) \ | 41 | # define _ASM_EXTABLE(from,to) \ |
diff --git a/include/asm-x86/elf.h b/include/asm-x86/elf.h index cd678b2d6a74..5c4745bec906 100644 --- a/include/asm-x86/elf.h +++ b/include/asm-x86/elf.h | |||
| @@ -148,8 +148,9 @@ do { \ | |||
| 148 | 148 | ||
| 149 | static inline void start_ia32_thread(struct pt_regs *regs, u32 ip, u32 sp) | 149 | static inline void start_ia32_thread(struct pt_regs *regs, u32 ip, u32 sp) |
| 150 | { | 150 | { |
| 151 | asm volatile("movl %0,%%fs" :: "r" (0)); | 151 | loadsegment(fs, 0); |
| 152 | asm volatile("movl %0,%%es; movl %0,%%ds" : : "r" (__USER32_DS)); | 152 | loadsegment(ds, __USER32_DS); |
| 153 | loadsegment(es, __USER32_DS); | ||
| 153 | load_gs_index(0); | 154 | load_gs_index(0); |
| 154 | regs->ip = ip; | 155 | regs->ip = ip; |
| 155 | regs->sp = sp; | 156 | regs->sp = sp; |
diff --git a/include/asm-x86/futex.h b/include/asm-x86/futex.h index 45dc24d84186..06b924ef6fa5 100644 --- a/include/asm-x86/futex.h +++ b/include/asm-x86/futex.h | |||
| @@ -25,7 +25,7 @@ | |||
| 25 | asm volatile("1:\tmovl %2, %0\n" \ | 25 | asm volatile("1:\tmovl %2, %0\n" \ |
| 26 | "\tmovl\t%0, %3\n" \ | 26 | "\tmovl\t%0, %3\n" \ |
| 27 | "\t" insn "\n" \ | 27 | "\t" insn "\n" \ |
| 28 | "2:\tlock; cmpxchgl %3, %2\n" \ | 28 | "2:\t" LOCK_PREFIX "cmpxchgl %3, %2\n" \ |
| 29 | "\tjnz\t1b\n" \ | 29 | "\tjnz\t1b\n" \ |
| 30 | "3:\t.section .fixup,\"ax\"\n" \ | 30 | "3:\t.section .fixup,\"ax\"\n" \ |
| 31 | "4:\tmov\t%5, %1\n" \ | 31 | "4:\tmov\t%5, %1\n" \ |
| @@ -64,7 +64,7 @@ static inline int futex_atomic_op_inuser(int encoded_op, int __user *uaddr) | |||
| 64 | __futex_atomic_op1("xchgl %0, %2", ret, oldval, uaddr, oparg); | 64 | __futex_atomic_op1("xchgl %0, %2", ret, oldval, uaddr, oparg); |
| 65 | break; | 65 | break; |
| 66 | case FUTEX_OP_ADD: | 66 | case FUTEX_OP_ADD: |
| 67 | __futex_atomic_op1("lock; xaddl %0, %2", ret, oldval, | 67 | __futex_atomic_op1(LOCK_PREFIX "xaddl %0, %2", ret, oldval, |
| 68 | uaddr, oparg); | 68 | uaddr, oparg); |
| 69 | break; | 69 | break; |
| 70 | case FUTEX_OP_OR: | 70 | case FUTEX_OP_OR: |
| @@ -122,7 +122,7 @@ static inline int futex_atomic_cmpxchg_inatomic(int __user *uaddr, int oldval, | |||
| 122 | if (!access_ok(VERIFY_WRITE, uaddr, sizeof(int))) | 122 | if (!access_ok(VERIFY_WRITE, uaddr, sizeof(int))) |
| 123 | return -EFAULT; | 123 | return -EFAULT; |
| 124 | 124 | ||
| 125 | asm volatile("1:\tlock; cmpxchgl %3, %1\n" | 125 | asm volatile("1:\t" LOCK_PREFIX "cmpxchgl %3, %1\n" |
| 126 | "2:\t.section .fixup, \"ax\"\n" | 126 | "2:\t.section .fixup, \"ax\"\n" |
| 127 | "3:\tmov %2, %0\n" | 127 | "3:\tmov %2, %0\n" |
| 128 | "\tjmp 2b\n" | 128 | "\tjmp 2b\n" |
diff --git a/include/asm-x86/gart.h b/include/asm-x86/gart.h index 07f445844146..baa54faba892 100644 --- a/include/asm-x86/gart.h +++ b/include/asm-x86/gart.h | |||
| @@ -52,15 +52,15 @@ static inline int aperture_valid(u64 aper_base, u32 aper_size, u32 min_size) | |||
| 52 | return 0; | 52 | return 0; |
| 53 | 53 | ||
| 54 | if (aper_base + aper_size > 0x100000000ULL) { | 54 | if (aper_base + aper_size > 0x100000000ULL) { |
| 55 | printk(KERN_ERR "Aperture beyond 4GB. Ignoring.\n"); | 55 | printk(KERN_INFO "Aperture beyond 4GB. Ignoring.\n"); |
| 56 | return 0; | 56 | return 0; |
| 57 | } | 57 | } |
| 58 | if (e820_any_mapped(aper_base, aper_base + aper_size, E820_RAM)) { | 58 | if (e820_any_mapped(aper_base, aper_base + aper_size, E820_RAM)) { |
| 59 | printk(KERN_ERR "Aperture pointing to e820 RAM. Ignoring.\n"); | 59 | printk(KERN_INFO "Aperture pointing to e820 RAM. Ignoring.\n"); |
| 60 | return 0; | 60 | return 0; |
| 61 | } | 61 | } |
| 62 | if (aper_size < min_size) { | 62 | if (aper_size < min_size) { |
| 63 | printk(KERN_ERR "Aperture too small (%d MB) than (%d MB)\n", | 63 | printk(KERN_INFO "Aperture too small (%d MB) than (%d MB)\n", |
| 64 | aper_size>>20, min_size>>20); | 64 | aper_size>>20, min_size>>20); |
| 65 | return 0; | 65 | return 0; |
| 66 | } | 66 | } |
diff --git a/include/asm-x86/mach-rdc321x/gpio.h b/include/asm-x86/mach-rdc321x/gpio.h index 6184561980f2..94b6cdf532e2 100644 --- a/include/asm-x86/mach-rdc321x/gpio.h +++ b/include/asm-x86/mach-rdc321x/gpio.h | |||
| @@ -1,6 +1,8 @@ | |||
| 1 | #ifndef ASM_X86__MACH_RDC321X__GPIO_H | 1 | #ifndef ASM_X86__MACH_RDC321X__GPIO_H |
| 2 | #define ASM_X86__MACH_RDC321X__GPIO_H | 2 | #define ASM_X86__MACH_RDC321X__GPIO_H |
| 3 | 3 | ||
| 4 | #include <linux/kernel.h> | ||
| 5 | |||
| 4 | extern int rdc_gpio_get_value(unsigned gpio); | 6 | extern int rdc_gpio_get_value(unsigned gpio); |
| 5 | extern void rdc_gpio_set_value(unsigned gpio, int value); | 7 | extern void rdc_gpio_set_value(unsigned gpio, int value); |
| 6 | extern int rdc_gpio_direction_input(unsigned gpio); | 8 | extern int rdc_gpio_direction_input(unsigned gpio); |
| @@ -18,6 +20,7 @@ static inline int gpio_request(unsigned gpio, const char *label) | |||
| 18 | 20 | ||
| 19 | static inline void gpio_free(unsigned gpio) | 21 | static inline void gpio_free(unsigned gpio) |
| 20 | { | 22 | { |
| 23 | might_sleep(); | ||
| 21 | rdc_gpio_free(gpio); | 24 | rdc_gpio_free(gpio); |
| 22 | } | 25 | } |
| 23 | 26 | ||
diff --git a/include/asm-x86/mmu.h b/include/asm-x86/mmu.h index a30d7a9c8297..9d5aff14334a 100644 --- a/include/asm-x86/mmu.h +++ b/include/asm-x86/mmu.h | |||
| @@ -7,14 +7,9 @@ | |||
| 7 | /* | 7 | /* |
| 8 | * The x86 doesn't have a mmu context, but | 8 | * The x86 doesn't have a mmu context, but |
| 9 | * we put the segment information here. | 9 | * we put the segment information here. |
| 10 | * | ||
| 11 | * cpu_vm_mask is used to optimize ldt flushing. | ||
| 12 | */ | 10 | */ |
| 13 | typedef struct { | 11 | typedef struct { |
| 14 | void *ldt; | 12 | void *ldt; |
| 15 | #ifdef CONFIG_X86_64 | ||
| 16 | rwlock_t ldtlock; | ||
| 17 | #endif | ||
| 18 | int size; | 13 | int size; |
| 19 | struct mutex lock; | 14 | struct mutex lock; |
| 20 | void *vdso; | 15 | void *vdso; |
diff --git a/include/asm-x86/msr.h b/include/asm-x86/msr.h index eee83f783f6d..530af1f6389e 100644 --- a/include/asm-x86/msr.h +++ b/include/asm-x86/msr.h | |||
| @@ -63,6 +63,22 @@ static inline unsigned long long native_read_msr_safe(unsigned int msr, | |||
| 63 | return EAX_EDX_VAL(val, low, high); | 63 | return EAX_EDX_VAL(val, low, high); |
| 64 | } | 64 | } |
| 65 | 65 | ||
| 66 | static inline unsigned long long native_read_msr_amd_safe(unsigned int msr, | ||
| 67 | int *err) | ||
| 68 | { | ||
| 69 | DECLARE_ARGS(val, low, high); | ||
| 70 | |||
| 71 | asm volatile("2: rdmsr ; xor %0,%0\n" | ||
| 72 | "1:\n\t" | ||
| 73 | ".section .fixup,\"ax\"\n\t" | ||
| 74 | "3: mov %3,%0 ; jmp 1b\n\t" | ||
| 75 | ".previous\n\t" | ||
| 76 | _ASM_EXTABLE(2b, 3b) | ||
| 77 | : "=r" (*err), EAX_EDX_RET(val, low, high) | ||
| 78 | : "c" (msr), "D" (0x9c5a203a), "i" (-EFAULT)); | ||
| 79 | return EAX_EDX_VAL(val, low, high); | ||
| 80 | } | ||
| 81 | |||
| 66 | static inline void native_write_msr(unsigned int msr, | 82 | static inline void native_write_msr(unsigned int msr, |
| 67 | unsigned low, unsigned high) | 83 | unsigned low, unsigned high) |
| 68 | { | 84 | { |
| @@ -158,6 +174,13 @@ static inline int rdmsrl_safe(unsigned msr, unsigned long long *p) | |||
| 158 | *p = native_read_msr_safe(msr, &err); | 174 | *p = native_read_msr_safe(msr, &err); |
| 159 | return err; | 175 | return err; |
| 160 | } | 176 | } |
| 177 | static inline int rdmsrl_amd_safe(unsigned msr, unsigned long long *p) | ||
| 178 | { | ||
| 179 | int err; | ||
| 180 | |||
| 181 | *p = native_read_msr_amd_safe(msr, &err); | ||
| 182 | return err; | ||
| 183 | } | ||
| 161 | 184 | ||
| 162 | #define rdtscl(low) \ | 185 | #define rdtscl(low) \ |
| 163 | ((low) = (u32)native_read_tsc()) | 186 | ((low) = (u32)native_read_tsc()) |
diff --git a/include/asm-x86/nmi.h b/include/asm-x86/nmi.h index f8b76f383904..d5e715f024dc 100644 --- a/include/asm-x86/nmi.h +++ b/include/asm-x86/nmi.h | |||
| @@ -34,6 +34,7 @@ extern void stop_apic_nmi_watchdog(void *); | |||
| 34 | extern void disable_timer_nmi_watchdog(void); | 34 | extern void disable_timer_nmi_watchdog(void); |
| 35 | extern void enable_timer_nmi_watchdog(void); | 35 | extern void enable_timer_nmi_watchdog(void); |
| 36 | extern int nmi_watchdog_tick(struct pt_regs *regs, unsigned reason); | 36 | extern int nmi_watchdog_tick(struct pt_regs *regs, unsigned reason); |
| 37 | extern void cpu_nmi_set_wd_enabled(void); | ||
| 37 | 38 | ||
| 38 | extern atomic_t nmi_active; | 39 | extern atomic_t nmi_active; |
| 39 | extern unsigned int nmi_watchdog; | 40 | extern unsigned int nmi_watchdog; |
diff --git a/include/asm-x86/page_32.h b/include/asm-x86/page_32.h index f32062a821c5..72f7305682c6 100644 --- a/include/asm-x86/page_32.h +++ b/include/asm-x86/page_32.h | |||
| @@ -89,9 +89,6 @@ extern int nx_enabled; | |||
| 89 | extern unsigned int __VMALLOC_RESERVE; | 89 | extern unsigned int __VMALLOC_RESERVE; |
| 90 | extern int sysctl_legacy_va_layout; | 90 | extern int sysctl_legacy_va_layout; |
| 91 | 91 | ||
| 92 | #define VMALLOC_RESERVE ((unsigned long)__VMALLOC_RESERVE) | ||
| 93 | #define MAXMEM (-__PAGE_OFFSET - __VMALLOC_RESERVE) | ||
| 94 | |||
| 95 | extern void find_low_pfn_range(void); | 92 | extern void find_low_pfn_range(void); |
| 96 | extern unsigned long init_memory_mapping(unsigned long start, | 93 | extern unsigned long init_memory_mapping(unsigned long start, |
| 97 | unsigned long end); | 94 | unsigned long end); |
diff --git a/include/asm-x86/paravirt.h b/include/asm-x86/paravirt.h index 2e6821a0b6e7..891971f57d35 100644 --- a/include/asm-x86/paravirt.h +++ b/include/asm-x86/paravirt.h | |||
| @@ -137,6 +137,7 @@ struct pv_cpu_ops { | |||
| 137 | 137 | ||
| 138 | /* MSR, PMC and TSR operations. | 138 | /* MSR, PMC and TSR operations. |
| 139 | err = 0/-EFAULT. wrmsr returns 0/-EFAULT. */ | 139 | err = 0/-EFAULT. wrmsr returns 0/-EFAULT. */ |
| 140 | u64 (*read_msr_amd)(unsigned int msr, int *err); | ||
| 140 | u64 (*read_msr)(unsigned int msr, int *err); | 141 | u64 (*read_msr)(unsigned int msr, int *err); |
| 141 | int (*write_msr)(unsigned int msr, unsigned low, unsigned high); | 142 | int (*write_msr)(unsigned int msr, unsigned low, unsigned high); |
| 142 | 143 | ||
| @@ -257,13 +258,13 @@ struct pv_mmu_ops { | |||
| 257 | * Hooks for allocating/releasing pagetable pages when they're | 258 | * Hooks for allocating/releasing pagetable pages when they're |
| 258 | * attached to a pagetable | 259 | * attached to a pagetable |
| 259 | */ | 260 | */ |
| 260 | void (*alloc_pte)(struct mm_struct *mm, u32 pfn); | 261 | void (*alloc_pte)(struct mm_struct *mm, unsigned long pfn); |
| 261 | void (*alloc_pmd)(struct mm_struct *mm, u32 pfn); | 262 | void (*alloc_pmd)(struct mm_struct *mm, unsigned long pfn); |
| 262 | void (*alloc_pmd_clone)(u32 pfn, u32 clonepfn, u32 start, u32 count); | 263 | void (*alloc_pmd_clone)(unsigned long pfn, unsigned long clonepfn, unsigned long start, unsigned long count); |
| 263 | void (*alloc_pud)(struct mm_struct *mm, u32 pfn); | 264 | void (*alloc_pud)(struct mm_struct *mm, unsigned long pfn); |
| 264 | void (*release_pte)(u32 pfn); | 265 | void (*release_pte)(unsigned long pfn); |
| 265 | void (*release_pmd)(u32 pfn); | 266 | void (*release_pmd)(unsigned long pfn); |
| 266 | void (*release_pud)(u32 pfn); | 267 | void (*release_pud)(unsigned long pfn); |
| 267 | 268 | ||
| 268 | /* Pagetable manipulation functions */ | 269 | /* Pagetable manipulation functions */ |
| 269 | void (*set_pte)(pte_t *ptep, pte_t pteval); | 270 | void (*set_pte)(pte_t *ptep, pte_t pteval); |
| @@ -726,6 +727,10 @@ static inline u64 paravirt_read_msr(unsigned msr, int *err) | |||
| 726 | { | 727 | { |
| 727 | return PVOP_CALL2(u64, pv_cpu_ops.read_msr, msr, err); | 728 | return PVOP_CALL2(u64, pv_cpu_ops.read_msr, msr, err); |
| 728 | } | 729 | } |
| 730 | static inline u64 paravirt_read_msr_amd(unsigned msr, int *err) | ||
| 731 | { | ||
| 732 | return PVOP_CALL2(u64, pv_cpu_ops.read_msr_amd, msr, err); | ||
| 733 | } | ||
| 729 | static inline int paravirt_write_msr(unsigned msr, unsigned low, unsigned high) | 734 | static inline int paravirt_write_msr(unsigned msr, unsigned low, unsigned high) |
| 730 | { | 735 | { |
| 731 | return PVOP_CALL3(int, pv_cpu_ops.write_msr, msr, low, high); | 736 | return PVOP_CALL3(int, pv_cpu_ops.write_msr, msr, low, high); |
| @@ -771,6 +776,13 @@ static inline int rdmsrl_safe(unsigned msr, unsigned long long *p) | |||
| 771 | *p = paravirt_read_msr(msr, &err); | 776 | *p = paravirt_read_msr(msr, &err); |
| 772 | return err; | 777 | return err; |
| 773 | } | 778 | } |
| 779 | static inline int rdmsrl_amd_safe(unsigned msr, unsigned long long *p) | ||
| 780 | { | ||
| 781 | int err; | ||
| 782 | |||
| 783 | *p = paravirt_read_msr_amd(msr, &err); | ||
| 784 | return err; | ||
| 785 | } | ||
| 774 | 786 | ||
| 775 | static inline u64 paravirt_read_tsc(void) | 787 | static inline u64 paravirt_read_tsc(void) |
| 776 | { | 788 | { |
| @@ -993,35 +1005,35 @@ static inline void paravirt_pgd_free(struct mm_struct *mm, pgd_t *pgd) | |||
| 993 | PVOP_VCALL2(pv_mmu_ops.pgd_free, mm, pgd); | 1005 | PVOP_VCALL2(pv_mmu_ops.pgd_free, mm, pgd); |
| 994 | } | 1006 | } |
| 995 | 1007 | ||
| 996 | static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned pfn) | 1008 | static inline void paravirt_alloc_pte(struct mm_struct *mm, unsigned long pfn) |
| 997 | { | 1009 | { |
| 998 | PVOP_VCALL2(pv_mmu_ops.alloc_pte, mm, pfn); | 1010 | PVOP_VCALL2(pv_mmu_ops.alloc_pte, mm, pfn); |
| 999 | } | 1011 | } |
| 1000 | static inline void paravirt_release_pte(unsigned pfn) | 1012 | static inline void paravirt_release_pte(unsigned long pfn) |
| 1001 | { | 1013 | { |
| 1002 | PVOP_VCALL1(pv_mmu_ops.release_pte, pfn); | 1014 | PVOP_VCALL1(pv_mmu_ops.release_pte, pfn); |
| 1003 | } | 1015 | } |
| 1004 | 1016 | ||
| 1005 | static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned pfn) | 1017 | static inline void paravirt_alloc_pmd(struct mm_struct *mm, unsigned long pfn) |
| 1006 | { | 1018 | { |
| 1007 | PVOP_VCALL2(pv_mmu_ops.alloc_pmd, mm, pfn); | 1019 | PVOP_VCALL2(pv_mmu_ops.alloc_pmd, mm, pfn); |
| 1008 | } | 1020 | } |
| 1009 | 1021 | ||
| 1010 | static inline void paravirt_alloc_pmd_clone(unsigned pfn, unsigned clonepfn, | 1022 | static inline void paravirt_alloc_pmd_clone(unsigned long pfn, unsigned long clonepfn, |
| 1011 | unsigned start, unsigned count) | 1023 | unsigned long start, unsigned long count) |
| 1012 | { | 1024 | { |
| 1013 | PVOP_VCALL4(pv_mmu_ops.alloc_pmd_clone, pfn, clonepfn, start, count); | 1025 | PVOP_VCALL4(pv_mmu_ops.alloc_pmd_clone, pfn, clonepfn, start, count); |
| 1014 | } | 1026 | } |
| 1015 | static inline void paravirt_release_pmd(unsigned pfn) | 1027 | static inline void paravirt_release_pmd(unsigned long pfn) |
| 1016 | { | 1028 | { |
| 1017 | PVOP_VCALL1(pv_mmu_ops.release_pmd, pfn); | 1029 | PVOP_VCALL1(pv_mmu_ops.release_pmd, pfn); |
| 1018 | } | 1030 | } |
| 1019 | 1031 | ||
| 1020 | static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned pfn) | 1032 | static inline void paravirt_alloc_pud(struct mm_struct *mm, unsigned long pfn) |
| 1021 | { | 1033 | { |
| 1022 | PVOP_VCALL2(pv_mmu_ops.alloc_pud, mm, pfn); | 1034 | PVOP_VCALL2(pv_mmu_ops.alloc_pud, mm, pfn); |
| 1023 | } | 1035 | } |
| 1024 | static inline void paravirt_release_pud(unsigned pfn) | 1036 | static inline void paravirt_release_pud(unsigned long pfn) |
| 1025 | { | 1037 | { |
| 1026 | PVOP_VCALL1(pv_mmu_ops.release_pud, pfn); | 1038 | PVOP_VCALL1(pv_mmu_ops.release_pud, pfn); |
| 1027 | } | 1039 | } |
diff --git a/include/asm-x86/pgtable-2level.h b/include/asm-x86/pgtable-2level.h index 60440b191626..81762081dcd8 100644 --- a/include/asm-x86/pgtable-2level.h +++ b/include/asm-x86/pgtable-2level.h | |||
| @@ -53,9 +53,7 @@ static inline pte_t native_ptep_get_and_clear(pte_t *xp) | |||
| 53 | #define native_ptep_get_and_clear(xp) native_local_ptep_get_and_clear(xp) | 53 | #define native_ptep_get_and_clear(xp) native_local_ptep_get_and_clear(xp) |
| 54 | #endif | 54 | #endif |
| 55 | 55 | ||
| 56 | #define pte_page(x) pfn_to_page(pte_pfn(x)) | ||
| 57 | #define pte_none(x) (!(x).pte_low) | 56 | #define pte_none(x) (!(x).pte_low) |
| 58 | #define pte_pfn(x) (pte_val(x) >> PAGE_SHIFT) | ||
| 59 | 57 | ||
| 60 | /* | 58 | /* |
| 61 | * Bits 0, 6 and 7 are taken, split up the 29 bits of offset | 59 | * Bits 0, 6 and 7 are taken, split up the 29 bits of offset |
diff --git a/include/asm-x86/pgtable-3level.h b/include/asm-x86/pgtable-3level.h index e713bd5f39a6..75f4276b5ddb 100644 --- a/include/asm-x86/pgtable-3level.h +++ b/include/asm-x86/pgtable-3level.h | |||
| @@ -151,18 +151,11 @@ static inline int pte_same(pte_t a, pte_t b) | |||
| 151 | return a.pte_low == b.pte_low && a.pte_high == b.pte_high; | 151 | return a.pte_low == b.pte_low && a.pte_high == b.pte_high; |
| 152 | } | 152 | } |
| 153 | 153 | ||
| 154 | #define pte_page(x) pfn_to_page(pte_pfn(x)) | ||
| 155 | |||
| 156 | static inline int pte_none(pte_t pte) | 154 | static inline int pte_none(pte_t pte) |
| 157 | { | 155 | { |
| 158 | return !pte.pte_low && !pte.pte_high; | 156 | return !pte.pte_low && !pte.pte_high; |
| 159 | } | 157 | } |
| 160 | 158 | ||
| 161 | static inline unsigned long pte_pfn(pte_t pte) | ||
| 162 | { | ||
| 163 | return (pte_val(pte) & PTE_PFN_MASK) >> PAGE_SHIFT; | ||
| 164 | } | ||
| 165 | |||
| 166 | /* | 159 | /* |
| 167 | * Bits 0, 6 and 7 are taken in the low part of the pte, | 160 | * Bits 0, 6 and 7 are taken in the low part of the pte, |
| 168 | * put the 32 bits of offset into the high part. | 161 | * put the 32 bits of offset into the high part. |
diff --git a/include/asm-x86/pgtable.h b/include/asm-x86/pgtable.h index 57d919a2d79d..888add7b0882 100644 --- a/include/asm-x86/pgtable.h +++ b/include/asm-x86/pgtable.h | |||
| @@ -186,6 +186,13 @@ static inline int pte_special(pte_t pte) | |||
| 186 | return pte_val(pte) & _PAGE_SPECIAL; | 186 | return pte_val(pte) & _PAGE_SPECIAL; |
| 187 | } | 187 | } |
| 188 | 188 | ||
| 189 | static inline unsigned long pte_pfn(pte_t pte) | ||
| 190 | { | ||
| 191 | return (pte_val(pte) & PTE_PFN_MASK) >> PAGE_SHIFT; | ||
| 192 | } | ||
| 193 | |||
| 194 | #define pte_page(pte) pfn_to_page(pte_pfn(pte)) | ||
| 195 | |||
| 189 | static inline int pmd_large(pmd_t pte) | 196 | static inline int pmd_large(pmd_t pte) |
| 190 | { | 197 | { |
| 191 | return (pmd_val(pte) & (_PAGE_PSE | _PAGE_PRESENT)) == | 198 | return (pmd_val(pte) & (_PAGE_PSE | _PAGE_PRESENT)) == |
diff --git a/include/asm-x86/pgtable_32.h b/include/asm-x86/pgtable_32.h index 45c8235400fe..8de702dc7d62 100644 --- a/include/asm-x86/pgtable_32.h +++ b/include/asm-x86/pgtable_32.h | |||
| @@ -57,8 +57,7 @@ extern void set_pmd_pfn(unsigned long, unsigned long, pgprot_t); | |||
| 57 | * area for the same reason. ;) | 57 | * area for the same reason. ;) |
| 58 | */ | 58 | */ |
| 59 | #define VMALLOC_OFFSET (8 * 1024 * 1024) | 59 | #define VMALLOC_OFFSET (8 * 1024 * 1024) |
| 60 | #define VMALLOC_START (((unsigned long)high_memory + 2 * VMALLOC_OFFSET - 1) \ | 60 | #define VMALLOC_START ((unsigned long)high_memory + VMALLOC_OFFSET) |
| 61 | & ~(VMALLOC_OFFSET - 1)) | ||
| 62 | #ifdef CONFIG_X86_PAE | 61 | #ifdef CONFIG_X86_PAE |
| 63 | #define LAST_PKMAP 512 | 62 | #define LAST_PKMAP 512 |
| 64 | #else | 63 | #else |
| @@ -74,6 +73,8 @@ extern void set_pmd_pfn(unsigned long, unsigned long, pgprot_t); | |||
| 74 | # define VMALLOC_END (FIXADDR_START - 2 * PAGE_SIZE) | 73 | # define VMALLOC_END (FIXADDR_START - 2 * PAGE_SIZE) |
| 75 | #endif | 74 | #endif |
| 76 | 75 | ||
| 76 | #define MAXMEM (VMALLOC_END - PAGE_OFFSET - __VMALLOC_RESERVE) | ||
| 77 | |||
| 77 | /* | 78 | /* |
| 78 | * Define this if things work differently on an i386 and an i486: | 79 | * Define this if things work differently on an i386 and an i486: |
| 79 | * it will (on an i486) warn about kernel memory accesses that are | 80 | * it will (on an i486) warn about kernel memory accesses that are |
diff --git a/include/asm-x86/pgtable_64.h b/include/asm-x86/pgtable_64.h index e3dcf7a08a0b..fde9770e53d1 100644 --- a/include/asm-x86/pgtable_64.h +++ b/include/asm-x86/pgtable_64.h | |||
| @@ -175,8 +175,6 @@ static inline int pmd_bad(pmd_t pmd) | |||
| 175 | #define pte_present(x) (pte_val((x)) & (_PAGE_PRESENT | _PAGE_PROTNONE)) | 175 | #define pte_present(x) (pte_val((x)) & (_PAGE_PRESENT | _PAGE_PROTNONE)) |
| 176 | 176 | ||
| 177 | #define pages_to_mb(x) ((x) >> (20 - PAGE_SHIFT)) /* FIXME: is this right? */ | 177 | #define pages_to_mb(x) ((x) >> (20 - PAGE_SHIFT)) /* FIXME: is this right? */ |
| 178 | #define pte_page(x) pfn_to_page(pte_pfn((x))) | ||
| 179 | #define pte_pfn(x) ((pte_val((x)) & __PHYSICAL_MASK) >> PAGE_SHIFT) | ||
| 180 | 178 | ||
| 181 | /* | 179 | /* |
| 182 | * Macro to mark a page protection value as "uncacheable". | 180 | * Macro to mark a page protection value as "uncacheable". |
diff --git a/include/asm-x86/resume-trace.h b/include/asm-x86/resume-trace.h index 519a8ecbfc95..e39376d7de50 100644 --- a/include/asm-x86/resume-trace.h +++ b/include/asm-x86/resume-trace.h | |||
| @@ -7,7 +7,7 @@ | |||
| 7 | do { \ | 7 | do { \ |
| 8 | if (pm_trace_enabled) { \ | 8 | if (pm_trace_enabled) { \ |
| 9 | const void *tracedata; \ | 9 | const void *tracedata; \ |
| 10 | asm volatile(_ASM_MOV_UL " $1f,%0\n" \ | 10 | asm volatile(_ASM_MOV " $1f,%0\n" \ |
| 11 | ".section .tracedata,\"a\"\n" \ | 11 | ".section .tracedata,\"a\"\n" \ |
| 12 | "1:\t.word %c1\n\t" \ | 12 | "1:\t.word %c1\n\t" \ |
| 13 | _ASM_PTR " %c2\n" \ | 13 | _ASM_PTR " %c2\n" \ |
diff --git a/include/asm-x86/spinlock.h b/include/asm-x86/spinlock.h index 5d08fa280fdf..93adae338ac6 100644 --- a/include/asm-x86/spinlock.h +++ b/include/asm-x86/spinlock.h | |||
| @@ -97,7 +97,7 @@ static __always_inline int __ticket_spin_trylock(raw_spinlock_t *lock) | |||
| 97 | "jne 1f\n\t" | 97 | "jne 1f\n\t" |
| 98 | "movw %w0,%w1\n\t" | 98 | "movw %w0,%w1\n\t" |
| 99 | "incb %h1\n\t" | 99 | "incb %h1\n\t" |
| 100 | "lock ; cmpxchgw %w1,%2\n\t" | 100 | LOCK_PREFIX "cmpxchgw %w1,%2\n\t" |
| 101 | "1:" | 101 | "1:" |
| 102 | "sete %b1\n\t" | 102 | "sete %b1\n\t" |
| 103 | "movzbl %b1,%0\n\t" | 103 | "movzbl %b1,%0\n\t" |
| @@ -135,7 +135,7 @@ static __always_inline void __ticket_spin_lock(raw_spinlock_t *lock) | |||
| 135 | int inc = 0x00010000; | 135 | int inc = 0x00010000; |
| 136 | int tmp; | 136 | int tmp; |
| 137 | 137 | ||
| 138 | asm volatile("lock ; xaddl %0, %1\n" | 138 | asm volatile(LOCK_PREFIX "xaddl %0, %1\n" |
| 139 | "movzwl %w0, %2\n\t" | 139 | "movzwl %w0, %2\n\t" |
| 140 | "shrl $16, %0\n\t" | 140 | "shrl $16, %0\n\t" |
| 141 | "1:\t" | 141 | "1:\t" |
| @@ -162,7 +162,7 @@ static __always_inline int __ticket_spin_trylock(raw_spinlock_t *lock) | |||
| 162 | "cmpl %0,%1\n\t" | 162 | "cmpl %0,%1\n\t" |
| 163 | "jne 1f\n\t" | 163 | "jne 1f\n\t" |
| 164 | "addl $0x00010000, %1\n\t" | 164 | "addl $0x00010000, %1\n\t" |
| 165 | "lock ; cmpxchgl %1,%2\n\t" | 165 | LOCK_PREFIX "cmpxchgl %1,%2\n\t" |
| 166 | "1:" | 166 | "1:" |
| 167 | "sete %b1\n\t" | 167 | "sete %b1\n\t" |
| 168 | "movzbl %b1,%0\n\t" | 168 | "movzbl %b1,%0\n\t" |
