diff options
| author | Paul Mundt <lethal@linux-sh.org> | 2011-01-26 04:23:27 -0500 |
|---|---|---|
| committer | Paul Mundt <lethal@linux-sh.org> | 2011-01-26 04:23:27 -0500 |
| commit | 6b620478e58677bb3f4ec884abb29ef9d68c7821 (patch) | |
| tree | 7a438fc5412587f9bd9b907d4d423ea52a498006 /arch/x86/include | |
| parent | 906b17dc089f7fa87e37a9cfe6ee185efc90e0da (diff) | |
| parent | 6fb1b304255efc5c4c93874ac8c066272e257e28 (diff) | |
Merge branch 'master' of git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux-2.6 into common/serial-rework
Diffstat (limited to 'arch/x86/include')
| -rw-r--r-- | arch/x86/include/asm/cacheflush.h | 42 | ||||
| -rw-r--r-- | arch/x86/include/asm/cpu.h | 1 | ||||
| -rw-r--r-- | arch/x86/include/asm/jump_label.h | 2 | ||||
| -rw-r--r-- | arch/x86/include/asm/numa_32.h | 2 | ||||
| -rw-r--r-- | arch/x86/include/asm/numa_64.h | 1 | ||||
| -rw-r--r-- | arch/x86/include/asm/paravirt.h | 5 | ||||
| -rw-r--r-- | arch/x86/include/asm/percpu.h | 8 |
7 files changed, 12 insertions, 49 deletions
diff --git a/arch/x86/include/asm/cacheflush.h b/arch/x86/include/asm/cacheflush.h index 63e35ec9075c..62f084478f7e 100644 --- a/arch/x86/include/asm/cacheflush.h +++ b/arch/x86/include/asm/cacheflush.h | |||
| @@ -1,48 +1,8 @@ | |||
| 1 | #ifndef _ASM_X86_CACHEFLUSH_H | 1 | #ifndef _ASM_X86_CACHEFLUSH_H |
| 2 | #define _ASM_X86_CACHEFLUSH_H | 2 | #define _ASM_X86_CACHEFLUSH_H |
| 3 | 3 | ||
| 4 | /* Keep includes the same across arches. */ | ||
| 5 | #include <linux/mm.h> | ||
| 6 | |||
| 7 | /* Caches aren't brain-dead on the intel. */ | 4 | /* Caches aren't brain-dead on the intel. */ |
| 8 | static inline void flush_cache_all(void) { } | 5 | #include <asm-generic/cacheflush.h> |
| 9 | static inline void flush_cache_mm(struct mm_struct *mm) { } | ||
| 10 | static inline void flush_cache_dup_mm(struct mm_struct *mm) { } | ||
| 11 | static inline void flush_cache_range(struct vm_area_struct *vma, | ||
| 12 | unsigned long start, unsigned long end) { } | ||
| 13 | static inline void flush_cache_page(struct vm_area_struct *vma, | ||
| 14 | unsigned long vmaddr, unsigned long pfn) { } | ||
| 15 | #define ARCH_IMPLEMENTS_FLUSH_DCACHE_PAGE 0 | ||
| 16 | static inline void flush_dcache_page(struct page *page) { } | ||
| 17 | static inline void flush_dcache_mmap_lock(struct address_space *mapping) { } | ||
| 18 | static inline void flush_dcache_mmap_unlock(struct address_space *mapping) { } | ||
| 19 | static inline void flush_icache_range(unsigned long start, | ||
| 20 | unsigned long end) { } | ||
| 21 | static inline void flush_icache_page(struct vm_area_struct *vma, | ||
| 22 | struct page *page) { } | ||
| 23 | static inline void flush_icache_user_range(struct vm_area_struct *vma, | ||
| 24 | struct page *page, | ||
| 25 | unsigned long addr, | ||
| 26 | unsigned long len) { } | ||
| 27 | static inline void flush_cache_vmap(unsigned long start, unsigned long end) { } | ||
| 28 | static inline void flush_cache_vunmap(unsigned long start, | ||
| 29 | unsigned long end) { } | ||
| 30 | |||
| 31 | static inline void copy_to_user_page(struct vm_area_struct *vma, | ||
| 32 | struct page *page, unsigned long vaddr, | ||
| 33 | void *dst, const void *src, | ||
| 34 | unsigned long len) | ||
| 35 | { | ||
| 36 | memcpy(dst, src, len); | ||
| 37 | } | ||
| 38 | |||
| 39 | static inline void copy_from_user_page(struct vm_area_struct *vma, | ||
| 40 | struct page *page, unsigned long vaddr, | ||
| 41 | void *dst, const void *src, | ||
| 42 | unsigned long len) | ||
| 43 | { | ||
| 44 | memcpy(dst, src, len); | ||
| 45 | } | ||
| 46 | 6 | ||
| 47 | #ifdef CONFIG_X86_PAT | 7 | #ifdef CONFIG_X86_PAT |
| 48 | /* | 8 | /* |
diff --git a/arch/x86/include/asm/cpu.h b/arch/x86/include/asm/cpu.h index 4fab24de26b1..6e6e7558e702 100644 --- a/arch/x86/include/asm/cpu.h +++ b/arch/x86/include/asm/cpu.h | |||
| @@ -32,5 +32,6 @@ extern void arch_unregister_cpu(int); | |||
| 32 | 32 | ||
| 33 | DECLARE_PER_CPU(int, cpu_state); | 33 | DECLARE_PER_CPU(int, cpu_state); |
| 34 | 34 | ||
| 35 | int __cpuinit mwait_usable(const struct cpuinfo_x86 *); | ||
| 35 | 36 | ||
| 36 | #endif /* _ASM_X86_CPU_H */ | 37 | #endif /* _ASM_X86_CPU_H */ |
diff --git a/arch/x86/include/asm/jump_label.h b/arch/x86/include/asm/jump_label.h index f52d42e80585..574dbc22893a 100644 --- a/arch/x86/include/asm/jump_label.h +++ b/arch/x86/include/asm/jump_label.h | |||
| @@ -14,7 +14,7 @@ | |||
| 14 | do { \ | 14 | do { \ |
| 15 | asm goto("1:" \ | 15 | asm goto("1:" \ |
| 16 | JUMP_LABEL_INITIAL_NOP \ | 16 | JUMP_LABEL_INITIAL_NOP \ |
| 17 | ".pushsection __jump_table, \"a\" \n\t"\ | 17 | ".pushsection __jump_table, \"aw\" \n\t"\ |
| 18 | _ASM_PTR "1b, %l[" #label "], %c0 \n\t" \ | 18 | _ASM_PTR "1b, %l[" #label "], %c0 \n\t" \ |
| 19 | ".popsection \n\t" \ | 19 | ".popsection \n\t" \ |
| 20 | : : "i" (key) : : label); \ | 20 | : : "i" (key) : : label); \ |
diff --git a/arch/x86/include/asm/numa_32.h b/arch/x86/include/asm/numa_32.h index a37229011b56..b0ef2b449a9d 100644 --- a/arch/x86/include/asm/numa_32.h +++ b/arch/x86/include/asm/numa_32.h | |||
| @@ -1,6 +1,8 @@ | |||
| 1 | #ifndef _ASM_X86_NUMA_32_H | 1 | #ifndef _ASM_X86_NUMA_32_H |
| 2 | #define _ASM_X86_NUMA_32_H | 2 | #define _ASM_X86_NUMA_32_H |
| 3 | 3 | ||
| 4 | extern int numa_off; | ||
| 5 | |||
| 4 | extern int pxm_to_nid(int pxm); | 6 | extern int pxm_to_nid(int pxm); |
| 5 | extern void numa_remove_cpu(int cpu); | 7 | extern void numa_remove_cpu(int cpu); |
| 6 | 8 | ||
diff --git a/arch/x86/include/asm/numa_64.h b/arch/x86/include/asm/numa_64.h index 5ae87285a502..0493be39607c 100644 --- a/arch/x86/include/asm/numa_64.h +++ b/arch/x86/include/asm/numa_64.h | |||
| @@ -40,6 +40,7 @@ extern void __cpuinit numa_remove_cpu(int cpu); | |||
| 40 | #ifdef CONFIG_NUMA_EMU | 40 | #ifdef CONFIG_NUMA_EMU |
| 41 | #define FAKE_NODE_MIN_SIZE ((u64)32 << 20) | 41 | #define FAKE_NODE_MIN_SIZE ((u64)32 << 20) |
| 42 | #define FAKE_NODE_MIN_HASH_MASK (~(FAKE_NODE_MIN_SIZE - 1UL)) | 42 | #define FAKE_NODE_MIN_HASH_MASK (~(FAKE_NODE_MIN_SIZE - 1UL)) |
| 43 | void numa_emu_cmdline(char *); | ||
| 43 | #endif /* CONFIG_NUMA_EMU */ | 44 | #endif /* CONFIG_NUMA_EMU */ |
| 44 | #else | 45 | #else |
| 45 | static inline void init_cpu_to_node(void) { } | 46 | static inline void init_cpu_to_node(void) { } |
diff --git a/arch/x86/include/asm/paravirt.h b/arch/x86/include/asm/paravirt.h index 2071a8b2b32f..ebbc4d8ab170 100644 --- a/arch/x86/include/asm/paravirt.h +++ b/arch/x86/include/asm/paravirt.h | |||
| @@ -558,13 +558,12 @@ static inline void set_pte_at(struct mm_struct *mm, unsigned long addr, | |||
| 558 | static inline void set_pmd_at(struct mm_struct *mm, unsigned long addr, | 558 | static inline void set_pmd_at(struct mm_struct *mm, unsigned long addr, |
| 559 | pmd_t *pmdp, pmd_t pmd) | 559 | pmd_t *pmdp, pmd_t pmd) |
| 560 | { | 560 | { |
| 561 | #if PAGETABLE_LEVELS >= 3 | ||
| 562 | if (sizeof(pmdval_t) > sizeof(long)) | 561 | if (sizeof(pmdval_t) > sizeof(long)) |
| 563 | /* 5 arg words */ | 562 | /* 5 arg words */ |
| 564 | pv_mmu_ops.set_pmd_at(mm, addr, pmdp, pmd); | 563 | pv_mmu_ops.set_pmd_at(mm, addr, pmdp, pmd); |
| 565 | else | 564 | else |
| 566 | PVOP_VCALL4(pv_mmu_ops.set_pmd_at, mm, addr, pmdp, pmd.pmd); | 565 | PVOP_VCALL4(pv_mmu_ops.set_pmd_at, mm, addr, pmdp, |
| 567 | #endif | 566 | native_pmd_val(pmd)); |
| 568 | } | 567 | } |
| 569 | #endif | 568 | #endif |
| 570 | 569 | ||
diff --git a/arch/x86/include/asm/percpu.h b/arch/x86/include/asm/percpu.h index 8ee45167e817..3788f4649db4 100644 --- a/arch/x86/include/asm/percpu.h +++ b/arch/x86/include/asm/percpu.h | |||
| @@ -414,8 +414,6 @@ do { \ | |||
| 414 | #define this_cpu_xchg_1(pcp, nval) percpu_xchg_op(pcp, nval) | 414 | #define this_cpu_xchg_1(pcp, nval) percpu_xchg_op(pcp, nval) |
| 415 | #define this_cpu_xchg_2(pcp, nval) percpu_xchg_op(pcp, nval) | 415 | #define this_cpu_xchg_2(pcp, nval) percpu_xchg_op(pcp, nval) |
| 416 | #define this_cpu_xchg_4(pcp, nval) percpu_xchg_op(pcp, nval) | 416 | #define this_cpu_xchg_4(pcp, nval) percpu_xchg_op(pcp, nval) |
| 417 | #define this_cpu_xchg_8(pcp, nval) percpu_xchg_op(pcp, nval) | ||
| 418 | #define this_cpu_cmpxchg_8(pcp, oval, nval) percpu_cmpxchg_op(pcp, oval, nval) | ||
| 419 | 417 | ||
| 420 | #define irqsafe_cpu_add_1(pcp, val) percpu_add_op((pcp), val) | 418 | #define irqsafe_cpu_add_1(pcp, val) percpu_add_op((pcp), val) |
| 421 | #define irqsafe_cpu_add_2(pcp, val) percpu_add_op((pcp), val) | 419 | #define irqsafe_cpu_add_2(pcp, val) percpu_add_op((pcp), val) |
| @@ -432,8 +430,6 @@ do { \ | |||
| 432 | #define irqsafe_cpu_xchg_1(pcp, nval) percpu_xchg_op(pcp, nval) | 430 | #define irqsafe_cpu_xchg_1(pcp, nval) percpu_xchg_op(pcp, nval) |
| 433 | #define irqsafe_cpu_xchg_2(pcp, nval) percpu_xchg_op(pcp, nval) | 431 | #define irqsafe_cpu_xchg_2(pcp, nval) percpu_xchg_op(pcp, nval) |
| 434 | #define irqsafe_cpu_xchg_4(pcp, nval) percpu_xchg_op(pcp, nval) | 432 | #define irqsafe_cpu_xchg_4(pcp, nval) percpu_xchg_op(pcp, nval) |
| 435 | #define irqsafe_cpu_xchg_8(pcp, nval) percpu_xchg_op(pcp, nval) | ||
| 436 | #define irqsafe_cpu_cmpxchg_8(pcp, oval, nval) percpu_cmpxchg_op(pcp, oval, nval) | ||
| 437 | 433 | ||
| 438 | #ifndef CONFIG_M386 | 434 | #ifndef CONFIG_M386 |
| 439 | #define __this_cpu_add_return_1(pcp, val) percpu_add_return_op(pcp, val) | 435 | #define __this_cpu_add_return_1(pcp, val) percpu_add_return_op(pcp, val) |
| @@ -475,11 +471,15 @@ do { \ | |||
| 475 | #define this_cpu_or_8(pcp, val) percpu_to_op("or", (pcp), val) | 471 | #define this_cpu_or_8(pcp, val) percpu_to_op("or", (pcp), val) |
| 476 | #define this_cpu_xor_8(pcp, val) percpu_to_op("xor", (pcp), val) | 472 | #define this_cpu_xor_8(pcp, val) percpu_to_op("xor", (pcp), val) |
| 477 | #define this_cpu_add_return_8(pcp, val) percpu_add_return_op(pcp, val) | 473 | #define this_cpu_add_return_8(pcp, val) percpu_add_return_op(pcp, val) |
| 474 | #define this_cpu_xchg_8(pcp, nval) percpu_xchg_op(pcp, nval) | ||
| 475 | #define this_cpu_cmpxchg_8(pcp, oval, nval) percpu_cmpxchg_op(pcp, oval, nval) | ||
| 478 | 476 | ||
| 479 | #define irqsafe_cpu_add_8(pcp, val) percpu_add_op((pcp), val) | 477 | #define irqsafe_cpu_add_8(pcp, val) percpu_add_op((pcp), val) |
| 480 | #define irqsafe_cpu_and_8(pcp, val) percpu_to_op("and", (pcp), val) | 478 | #define irqsafe_cpu_and_8(pcp, val) percpu_to_op("and", (pcp), val) |
| 481 | #define irqsafe_cpu_or_8(pcp, val) percpu_to_op("or", (pcp), val) | 479 | #define irqsafe_cpu_or_8(pcp, val) percpu_to_op("or", (pcp), val) |
| 482 | #define irqsafe_cpu_xor_8(pcp, val) percpu_to_op("xor", (pcp), val) | 480 | #define irqsafe_cpu_xor_8(pcp, val) percpu_to_op("xor", (pcp), val) |
| 481 | #define irqsafe_cpu_xchg_8(pcp, nval) percpu_xchg_op(pcp, nval) | ||
| 482 | #define irqsafe_cpu_cmpxchg_8(pcp, oval, nval) percpu_cmpxchg_op(pcp, oval, nval) | ||
| 483 | #endif | 483 | #endif |
| 484 | 484 | ||
| 485 | /* This is not atomic against other CPUs -- CPU preemption needs to be off */ | 485 | /* This is not atomic against other CPUs -- CPU preemption needs to be off */ |
