aboutsummaryrefslogtreecommitdiffstats
path: root/arch/x86/mm
diff options
context:
space:
mode:
Diffstat (limited to 'arch/x86/mm')
-rw-r--r--arch/x86/mm/boot_ioremap_32.c2
-rw-r--r--arch/x86/mm/discontig_32.c2
-rw-r--r--arch/x86/mm/fault_32.c4
-rw-r--r--arch/x86/mm/fault_64.c4
-rw-r--r--arch/x86/mm/srat_64.c2
5 files changed, 7 insertions, 7 deletions
diff --git a/arch/x86/mm/boot_ioremap_32.c b/arch/x86/mm/boot_ioremap_32.c
index 4de95a17a7d4..f14da2a53ece 100644
--- a/arch/x86/mm/boot_ioremap_32.c
+++ b/arch/x86/mm/boot_ioremap_32.c
@@ -10,7 +10,7 @@
10 10
11/* 11/*
12 * We need to use the 2-level pagetable functions, but CONFIG_X86_PAE 12 * We need to use the 2-level pagetable functions, but CONFIG_X86_PAE
13 * keeps that from happenning. If anyone has a better way, I'm listening. 13 * keeps that from happening. If anyone has a better way, I'm listening.
14 * 14 *
15 * boot_pte_t is defined only if this all works correctly 15 * boot_pte_t is defined only if this all works correctly
16 */ 16 */
diff --git a/arch/x86/mm/discontig_32.c b/arch/x86/mm/discontig_32.c
index 13893772cc48..fe608a45ffb6 100644
--- a/arch/x86/mm/discontig_32.c
+++ b/arch/x86/mm/discontig_32.c
@@ -273,7 +273,7 @@ unsigned long __init setup_memory(void)
273 * When mapping a NUMA machine we allocate the node_mem_map arrays 273 * When mapping a NUMA machine we allocate the node_mem_map arrays
274 * from node local memory. They are then mapped directly into KVA 274 * from node local memory. They are then mapped directly into KVA
275 * between zone normal and vmalloc space. Calculate the size of 275 * between zone normal and vmalloc space. Calculate the size of
276 * this space and use it to adjust the boundry between ZONE_NORMAL 276 * this space and use it to adjust the boundary between ZONE_NORMAL
277 * and ZONE_HIGHMEM. 277 * and ZONE_HIGHMEM.
278 */ 278 */
279 find_max_pfn(); 279 find_max_pfn();
diff --git a/arch/x86/mm/fault_32.c b/arch/x86/mm/fault_32.c
index b695d70e998c..503dfc05111b 100644
--- a/arch/x86/mm/fault_32.c
+++ b/arch/x86/mm/fault_32.c
@@ -354,7 +354,7 @@ fastcall void __kprobes do_page_fault(struct pt_regs *regs,
354 354
355 /* When running in the kernel we expect faults to occur only to 355 /* When running in the kernel we expect faults to occur only to
356 * addresses in user space. All other faults represent errors in the 356 * addresses in user space. All other faults represent errors in the
357 * kernel and should generate an OOPS. Unfortunatly, in the case of an 357 * kernel and should generate an OOPS. Unfortunately, in the case of an
358 * erroneous fault occurring in a code path which already holds mmap_sem 358 * erroneous fault occurring in a code path which already holds mmap_sem
359 * we will deadlock attempting to validate the fault against the 359 * we will deadlock attempting to validate the fault against the
360 * address space. Luckily the kernel only validly references user 360 * address space. Luckily the kernel only validly references user
@@ -362,7 +362,7 @@ fastcall void __kprobes do_page_fault(struct pt_regs *regs,
362 * exceptions table. 362 * exceptions table.
363 * 363 *
364 * As the vast majority of faults will be valid we will only perform 364 * As the vast majority of faults will be valid we will only perform
365 * the source reference check when there is a possibilty of a deadlock. 365 * the source reference check when there is a possibility of a deadlock.
366 * Attempt to lock the address space, if we cannot we then validate the 366 * Attempt to lock the address space, if we cannot we then validate the
367 * source. If this is invalid we can skip the address space check, 367 * source. If this is invalid we can skip the address space check,
368 * thus avoiding the deadlock. 368 * thus avoiding the deadlock.
diff --git a/arch/x86/mm/fault_64.c b/arch/x86/mm/fault_64.c
index 00be7f0a71b2..644b4f7ece10 100644
--- a/arch/x86/mm/fault_64.c
+++ b/arch/x86/mm/fault_64.c
@@ -373,7 +373,7 @@ asmlinkage void __kprobes do_page_fault(struct pt_regs *regs,
373 again: 373 again:
374 /* When running in the kernel we expect faults to occur only to 374 /* When running in the kernel we expect faults to occur only to
375 * addresses in user space. All other faults represent errors in the 375 * addresses in user space. All other faults represent errors in the
376 * kernel and should generate an OOPS. Unfortunatly, in the case of an 376 * kernel and should generate an OOPS. Unfortunately, in the case of an
377 * erroneous fault occurring in a code path which already holds mmap_sem 377 * erroneous fault occurring in a code path which already holds mmap_sem
378 * we will deadlock attempting to validate the fault against the 378 * we will deadlock attempting to validate the fault against the
379 * address space. Luckily the kernel only validly references user 379 * address space. Luckily the kernel only validly references user
@@ -381,7 +381,7 @@ asmlinkage void __kprobes do_page_fault(struct pt_regs *regs,
381 * exceptions table. 381 * exceptions table.
382 * 382 *
383 * As the vast majority of faults will be valid we will only perform 383 * As the vast majority of faults will be valid we will only perform
384 * the source reference check when there is a possibilty of a deadlock. 384 * the source reference check when there is a possibility of a deadlock.
385 * Attempt to lock the address space, if we cannot we then validate the 385 * Attempt to lock the address space, if we cannot we then validate the
386 * source. If this is invalid we can skip the address space check, 386 * source. If this is invalid we can skip the address space check,
387 * thus avoiding the deadlock. 387 * thus avoiding the deadlock.
diff --git a/arch/x86/mm/srat_64.c b/arch/x86/mm/srat_64.c
index 56089ccc3949..ea85172fc0cc 100644
--- a/arch/x86/mm/srat_64.c
+++ b/arch/x86/mm/srat_64.c
@@ -218,7 +218,7 @@ static inline int save_add_info(void) {return 0;}
218/* 218/*
219 * Update nodes_add and decide if to include add are in the zone. 219 * Update nodes_add and decide if to include add are in the zone.
220 * Both SPARSE and RESERVE need nodes_add infomation. 220 * Both SPARSE and RESERVE need nodes_add infomation.
221 * This code supports one contigious hot add area per node. 221 * This code supports one contiguous hot add area per node.
222 */ 222 */
223static int reserve_hotadd(int node, unsigned long start, unsigned long end) 223static int reserve_hotadd(int node, unsigned long start, unsigned long end)
224{ 224{