diff options
Diffstat (limited to 'include')
175 files changed, 5290 insertions, 4279 deletions
diff --git a/include/asm-x86/Kbuild b/include/asm-x86/Kbuild index 3b8160a2b47e..1e3554596f72 100644 --- a/include/asm-x86/Kbuild +++ b/include/asm-x86/Kbuild | |||
| @@ -10,6 +10,7 @@ header-y += prctl.h | |||
| 10 | header-y += ptrace-abi.h | 10 | header-y += ptrace-abi.h |
| 11 | header-y += sigcontext32.h | 11 | header-y += sigcontext32.h |
| 12 | header-y += ucontext.h | 12 | header-y += ucontext.h |
| 13 | header-y += processor-flags.h | ||
| 13 | 14 | ||
| 14 | unifdef-y += e820.h | 15 | unifdef-y += e820.h |
| 15 | unifdef-y += ist.h | 16 | unifdef-y += ist.h |
diff --git a/include/asm-x86/a.out-core.h b/include/asm-x86/a.out-core.h index d2b6e11d3e97..714207a1c387 100644 --- a/include/asm-x86/a.out-core.h +++ b/include/asm-x86/a.out-core.h | |||
| @@ -29,8 +29,9 @@ static inline void aout_dump_thread(struct pt_regs *regs, struct user *dump) | |||
| 29 | dump->magic = CMAGIC; | 29 | dump->magic = CMAGIC; |
| 30 | dump->start_code = 0; | 30 | dump->start_code = 0; |
| 31 | dump->start_stack = regs->sp & ~(PAGE_SIZE - 1); | 31 | dump->start_stack = regs->sp & ~(PAGE_SIZE - 1); |
| 32 | dump->u_tsize = ((unsigned long) current->mm->end_code) >> PAGE_SHIFT; | 32 | dump->u_tsize = ((unsigned long)current->mm->end_code) >> PAGE_SHIFT; |
| 33 | dump->u_dsize = ((unsigned long) (current->mm->brk + (PAGE_SIZE-1))) >> PAGE_SHIFT; | 33 | dump->u_dsize = ((unsigned long)(current->mm->brk + (PAGE_SIZE - 1))) |
| 34 | >> PAGE_SHIFT; | ||
| 34 | dump->u_dsize -= dump->u_tsize; | 35 | dump->u_dsize -= dump->u_tsize; |
| 35 | dump->u_ssize = 0; | 36 | dump->u_ssize = 0; |
| 36 | dump->u_debugreg[0] = current->thread.debugreg0; | 37 | dump->u_debugreg[0] = current->thread.debugreg0; |
| @@ -43,7 +44,8 @@ static inline void aout_dump_thread(struct pt_regs *regs, struct user *dump) | |||
| 43 | dump->u_debugreg[7] = current->thread.debugreg7; | 44 | dump->u_debugreg[7] = current->thread.debugreg7; |
| 44 | 45 | ||
| 45 | if (dump->start_stack < TASK_SIZE) | 46 | if (dump->start_stack < TASK_SIZE) |
| 46 | dump->u_ssize = ((unsigned long) (TASK_SIZE - dump->start_stack)) >> PAGE_SHIFT; | 47 | dump->u_ssize = ((unsigned long)(TASK_SIZE - dump->start_stack)) |
| 48 | >> PAGE_SHIFT; | ||
| 47 | 49 | ||
| 48 | dump->regs.bx = regs->bx; | 50 | dump->regs.bx = regs->bx; |
| 49 | dump->regs.cx = regs->cx; | 51 | dump->regs.cx = regs->cx; |
| @@ -55,7 +57,7 @@ static inline void aout_dump_thread(struct pt_regs *regs, struct user *dump) | |||
| 55 | dump->regs.ds = (u16)regs->ds; | 57 | dump->regs.ds = (u16)regs->ds; |
| 56 | dump->regs.es = (u16)regs->es; | 58 | dump->regs.es = (u16)regs->es; |
| 57 | dump->regs.fs = (u16)regs->fs; | 59 | dump->regs.fs = (u16)regs->fs; |
| 58 | savesegment(gs,gs); | 60 | savesegment(gs, gs); |
| 59 | dump->regs.orig_ax = regs->orig_ax; | 61 | dump->regs.orig_ax = regs->orig_ax; |
| 60 | dump->regs.ip = regs->ip; | 62 | dump->regs.ip = regs->ip; |
| 61 | dump->regs.cs = (u16)regs->cs; | 63 | dump->regs.cs = (u16)regs->cs; |
| @@ -63,7 +65,7 @@ static inline void aout_dump_thread(struct pt_regs *regs, struct user *dump) | |||
| 63 | dump->regs.sp = regs->sp; | 65 | dump->regs.sp = regs->sp; |
| 64 | dump->regs.ss = (u16)regs->ss; | 66 | dump->regs.ss = (u16)regs->ss; |
| 65 | 67 | ||
| 66 | dump->u_fpvalid = dump_fpu (regs, &dump->i387); | 68 | dump->u_fpvalid = dump_fpu(regs, &dump->i387); |
| 67 | } | 69 | } |
| 68 | 70 | ||
| 69 | #endif /* CONFIG_X86_32 */ | 71 | #endif /* CONFIG_X86_32 */ |
diff --git a/include/asm-x86/acpi.h b/include/asm-x86/acpi.h index 7a72d6aa50be..14411c9de46f 100644 --- a/include/asm-x86/acpi.h +++ b/include/asm-x86/acpi.h | |||
| @@ -67,16 +67,16 @@ int __acpi_release_global_lock(unsigned int *lock); | |||
| 67 | */ | 67 | */ |
| 68 | #define ACPI_DIV_64_BY_32(n_hi, n_lo, d32, q32, r32) \ | 68 | #define ACPI_DIV_64_BY_32(n_hi, n_lo, d32, q32, r32) \ |
| 69 | asm("divl %2;" \ | 69 | asm("divl %2;" \ |
| 70 | :"=a"(q32), "=d"(r32) \ | 70 | : "=a"(q32), "=d"(r32) \ |
| 71 | :"r"(d32), \ | 71 | : "r"(d32), \ |
| 72 | "0"(n_lo), "1"(n_hi)) | 72 | "0"(n_lo), "1"(n_hi)) |
| 73 | 73 | ||
| 74 | 74 | ||
| 75 | #define ACPI_SHIFT_RIGHT_64(n_hi, n_lo) \ | 75 | #define ACPI_SHIFT_RIGHT_64(n_hi, n_lo) \ |
| 76 | asm("shrl $1,%2 ;" \ | 76 | asm("shrl $1,%2 ;" \ |
| 77 | "rcrl $1,%3;" \ | 77 | "rcrl $1,%3;" \ |
| 78 | :"=r"(n_hi), "=r"(n_lo) \ | 78 | : "=r"(n_hi), "=r"(n_lo) \ |
| 79 | :"0"(n_hi), "1"(n_lo)) | 79 | : "0"(n_hi), "1"(n_lo)) |
| 80 | 80 | ||
| 81 | #ifdef CONFIG_ACPI | 81 | #ifdef CONFIG_ACPI |
| 82 | extern int acpi_lapic; | 82 | extern int acpi_lapic; |
diff --git a/include/asm-x86/alternative.h b/include/asm-x86/alternative.h index d8bacf3c4b08..1f6a9ca10126 100644 --- a/include/asm-x86/alternative.h +++ b/include/asm-x86/alternative.h | |||
| @@ -66,8 +66,8 @@ extern void alternatives_smp_module_del(struct module *mod); | |||
| 66 | extern void alternatives_smp_switch(int smp); | 66 | extern void alternatives_smp_switch(int smp); |
| 67 | #else | 67 | #else |
| 68 | static inline void alternatives_smp_module_add(struct module *mod, char *name, | 68 | static inline void alternatives_smp_module_add(struct module *mod, char *name, |
| 69 | void *locks, void *locks_end, | 69 | void *locks, void *locks_end, |
| 70 | void *text, void *text_end) {} | 70 | void *text, void *text_end) {} |
| 71 | static inline void alternatives_smp_module_del(struct module *mod) {} | 71 | static inline void alternatives_smp_module_del(struct module *mod) {} |
| 72 | static inline void alternatives_smp_switch(int smp) {} | 72 | static inline void alternatives_smp_switch(int smp) {} |
| 73 | #endif /* CONFIG_SMP */ | 73 | #endif /* CONFIG_SMP */ |
| @@ -148,14 +148,34 @@ struct paravirt_patch_site; | |||
| 148 | void apply_paravirt(struct paravirt_patch_site *start, | 148 | void apply_paravirt(struct paravirt_patch_site *start, |
| 149 | struct paravirt_patch_site *end); | 149 | struct paravirt_patch_site *end); |
| 150 | #else | 150 | #else |
| 151 | static inline void | 151 | static inline void apply_paravirt(struct paravirt_patch_site *start, |
| 152 | apply_paravirt(struct paravirt_patch_site *start, | 152 | struct paravirt_patch_site *end) |
| 153 | struct paravirt_patch_site *end) | ||
| 154 | {} | 153 | {} |
| 155 | #define __parainstructions NULL | 154 | #define __parainstructions NULL |
| 156 | #define __parainstructions_end NULL | 155 | #define __parainstructions_end NULL |
| 157 | #endif | 156 | #endif |
| 158 | 157 | ||
| 159 | extern void text_poke(void *addr, unsigned char *opcode, int len); | 158 | extern void add_nops(void *insns, unsigned int len); |
| 159 | |||
| 160 | /* | ||
| 161 | * Clear and restore the kernel write-protection flag on the local CPU. | ||
| 162 | * Allows the kernel to edit read-only pages. | ||
| 163 | * Side-effect: any interrupt handler running between save and restore will have | ||
| 164 | * the ability to write to read-only pages. | ||
| 165 | * | ||
| 166 | * Warning: | ||
| 167 | * Code patching in the UP case is safe if NMIs and MCE handlers are stopped and | ||
| 168 | * no thread can be preempted in the instructions being modified (no iret to an | ||
| 169 | * invalid instruction possible) or if the instructions are changed from a | ||
| 170 | * consistent state to another consistent state atomically. | ||
| 171 | * More care must be taken when modifying code in the SMP case because of | ||
| 172 | * Intel's errata. | ||
| 173 | * On the local CPU you need to be protected again NMI or MCE handlers seeing an | ||
| 174 | * inconsistent instruction while you patch. | ||
| 175 | * The _early version expects the memory to already be RW. | ||
| 176 | */ | ||
| 177 | |||
| 178 | extern void *text_poke(void *addr, const void *opcode, size_t len); | ||
| 179 | extern void *text_poke_early(void *addr, const void *opcode, size_t len); | ||
| 160 | 180 | ||
| 161 | #endif /* _ASM_X86_ALTERNATIVE_H */ | 181 | #endif /* _ASM_X86_ALTERNATIVE_H */ |
diff --git a/include/asm-x86/apic.h b/include/asm-x86/apic.h index bcfc07fd3661..be9639a9a186 100644 --- a/include/asm-x86/apic.h +++ b/include/asm-x86/apic.h | |||
| @@ -44,7 +44,6 @@ extern int apic_runs_main_timer; | |||
| 44 | extern int ioapic_force; | 44 | extern int ioapic_force; |
| 45 | extern int disable_apic; | 45 | extern int disable_apic; |
| 46 | extern int disable_apic_timer; | 46 | extern int disable_apic_timer; |
| 47 | extern unsigned boot_cpu_id; | ||
| 48 | 47 | ||
| 49 | /* | 48 | /* |
| 50 | * Basic functions accessing APICs. | 49 | * Basic functions accessing APICs. |
| @@ -59,6 +58,8 @@ extern unsigned boot_cpu_id; | |||
| 59 | #define setup_secondary_clock setup_secondary_APIC_clock | 58 | #define setup_secondary_clock setup_secondary_APIC_clock |
| 60 | #endif | 59 | #endif |
| 61 | 60 | ||
| 61 | extern int is_vsmp_box(void); | ||
| 62 | |||
| 62 | static inline void native_apic_write(unsigned long reg, u32 v) | 63 | static inline void native_apic_write(unsigned long reg, u32 v) |
| 63 | { | 64 | { |
| 64 | *((volatile u32 *)(APIC_BASE + reg)) = v; | 65 | *((volatile u32 *)(APIC_BASE + reg)) = v; |
| @@ -66,7 +67,7 @@ static inline void native_apic_write(unsigned long reg, u32 v) | |||
| 66 | 67 | ||
| 67 | static inline void native_apic_write_atomic(unsigned long reg, u32 v) | 68 | static inline void native_apic_write_atomic(unsigned long reg, u32 v) |
| 68 | { | 69 | { |
| 69 | (void) xchg((u32*)(APIC_BASE + reg), v); | 70 | (void)xchg((u32 *)(APIC_BASE + reg), v); |
| 70 | } | 71 | } |
| 71 | 72 | ||
| 72 | static inline u32 native_apic_read(unsigned long reg) | 73 | static inline u32 native_apic_read(unsigned long reg) |
| @@ -123,7 +124,7 @@ extern void enable_NMI_through_LVT0(void); | |||
| 123 | * On 32bit this is mach-xxx local | 124 | * On 32bit this is mach-xxx local |
| 124 | */ | 125 | */ |
| 125 | #ifdef CONFIG_X86_64 | 126 | #ifdef CONFIG_X86_64 |
| 126 | extern void setup_apic_routing(void); | 127 | extern void early_init_lapic_mapping(void); |
| 127 | #endif | 128 | #endif |
| 128 | 129 | ||
| 129 | extern u8 setup_APIC_eilvt_mce(u8 vector, u8 msg_type, u8 mask); | 130 | extern u8 setup_APIC_eilvt_mce(u8 vector, u8 msg_type, u8 mask); |
diff --git a/include/asm-x86/apicdef.h b/include/asm-x86/apicdef.h index 550af7a6f88e..6b9008c78731 100644 --- a/include/asm-x86/apicdef.h +++ b/include/asm-x86/apicdef.h | |||
| @@ -12,17 +12,15 @@ | |||
| 12 | 12 | ||
| 13 | #define APIC_ID 0x20 | 13 | #define APIC_ID 0x20 |
| 14 | 14 | ||
| 15 | #ifdef CONFIG_X86_64 | ||
| 16 | # define APIC_ID_MASK (0xFFu<<24) | ||
| 17 | # define GET_APIC_ID(x) (((x)>>24)&0xFFu) | ||
| 18 | # define SET_APIC_ID(x) (((x)<<24)) | ||
| 19 | #endif | ||
| 20 | |||
| 21 | #define APIC_LVR 0x30 | 15 | #define APIC_LVR 0x30 |
| 22 | #define APIC_LVR_MASK 0xFF00FF | 16 | #define APIC_LVR_MASK 0xFF00FF |
| 23 | #define GET_APIC_VERSION(x) ((x)&0xFFu) | 17 | #define GET_APIC_VERSION(x) ((x) & 0xFFu) |
| 24 | #define GET_APIC_MAXLVT(x) (((x)>>16)&0xFFu) | 18 | #define GET_APIC_MAXLVT(x) (((x) >> 16) & 0xFFu) |
| 25 | #define APIC_INTEGRATED(x) ((x)&0xF0u) | 19 | #ifdef CONFIG_X86_32 |
| 20 | # define APIC_INTEGRATED(x) ((x) & 0xF0u) | ||
| 21 | #else | ||
| 22 | # define APIC_INTEGRATED(x) (1) | ||
| 23 | #endif | ||
| 26 | #define APIC_XAPIC(x) ((x) >= 0x14) | 24 | #define APIC_XAPIC(x) ((x) >= 0x14) |
| 27 | #define APIC_TASKPRI 0x80 | 25 | #define APIC_TASKPRI 0x80 |
| 28 | #define APIC_TPRI_MASK 0xFFu | 26 | #define APIC_TPRI_MASK 0xFFu |
| @@ -33,16 +31,16 @@ | |||
| 33 | #define APIC_EIO_ACK 0x0 | 31 | #define APIC_EIO_ACK 0x0 |
| 34 | #define APIC_RRR 0xC0 | 32 | #define APIC_RRR 0xC0 |
| 35 | #define APIC_LDR 0xD0 | 33 | #define APIC_LDR 0xD0 |
| 36 | #define APIC_LDR_MASK (0xFFu<<24) | 34 | #define APIC_LDR_MASK (0xFFu << 24) |
| 37 | #define GET_APIC_LOGICAL_ID(x) (((x)>>24)&0xFFu) | 35 | #define GET_APIC_LOGICAL_ID(x) (((x) >> 24) & 0xFFu) |
| 38 | #define SET_APIC_LOGICAL_ID(x) (((x)<<24)) | 36 | #define SET_APIC_LOGICAL_ID(x) (((x) << 24)) |
| 39 | #define APIC_ALL_CPUS 0xFFu | 37 | #define APIC_ALL_CPUS 0xFFu |
| 40 | #define APIC_DFR 0xE0 | 38 | #define APIC_DFR 0xE0 |
| 41 | #define APIC_DFR_CLUSTER 0x0FFFFFFFul | 39 | #define APIC_DFR_CLUSTER 0x0FFFFFFFul |
| 42 | #define APIC_DFR_FLAT 0xFFFFFFFFul | 40 | #define APIC_DFR_FLAT 0xFFFFFFFFul |
| 43 | #define APIC_SPIV 0xF0 | 41 | #define APIC_SPIV 0xF0 |
| 44 | #define APIC_SPIV_FOCUS_DISABLED (1<<9) | 42 | #define APIC_SPIV_FOCUS_DISABLED (1 << 9) |
| 45 | #define APIC_SPIV_APIC_ENABLED (1<<8) | 43 | #define APIC_SPIV_APIC_ENABLED (1 << 8) |
| 46 | #define APIC_ISR 0x100 | 44 | #define APIC_ISR 0x100 |
| 47 | #define APIC_ISR_NR 0x8 /* Number of 32 bit ISR registers. */ | 45 | #define APIC_ISR_NR 0x8 /* Number of 32 bit ISR registers. */ |
| 48 | #define APIC_TMR 0x180 | 46 | #define APIC_TMR 0x180 |
| @@ -78,27 +76,27 @@ | |||
| 78 | #define APIC_DM_EXTINT 0x00700 | 76 | #define APIC_DM_EXTINT 0x00700 |
| 79 | #define APIC_VECTOR_MASK 0x000FF | 77 | #define APIC_VECTOR_MASK 0x000FF |
| 80 | #define APIC_ICR2 0x310 | 78 | #define APIC_ICR2 0x310 |
| 81 | #define GET_APIC_DEST_FIELD(x) (((x)>>24)&0xFF) | 79 | #define GET_APIC_DEST_FIELD(x) (((x) >> 24) & 0xFF) |
| 82 | #define SET_APIC_DEST_FIELD(x) ((x)<<24) | 80 | #define SET_APIC_DEST_FIELD(x) ((x) << 24) |
| 83 | #define APIC_LVTT 0x320 | 81 | #define APIC_LVTT 0x320 |
| 84 | #define APIC_LVTTHMR 0x330 | 82 | #define APIC_LVTTHMR 0x330 |
| 85 | #define APIC_LVTPC 0x340 | 83 | #define APIC_LVTPC 0x340 |
| 86 | #define APIC_LVT0 0x350 | 84 | #define APIC_LVT0 0x350 |
| 87 | #define APIC_LVT_TIMER_BASE_MASK (0x3<<18) | 85 | #define APIC_LVT_TIMER_BASE_MASK (0x3 << 18) |
| 88 | #define GET_APIC_TIMER_BASE(x) (((x)>>18)&0x3) | 86 | #define GET_APIC_TIMER_BASE(x) (((x) >> 18) & 0x3) |
| 89 | #define SET_APIC_TIMER_BASE(x) (((x)<<18)) | 87 | #define SET_APIC_TIMER_BASE(x) (((x) << 18)) |
| 90 | #define APIC_TIMER_BASE_CLKIN 0x0 | 88 | #define APIC_TIMER_BASE_CLKIN 0x0 |
| 91 | #define APIC_TIMER_BASE_TMBASE 0x1 | 89 | #define APIC_TIMER_BASE_TMBASE 0x1 |
| 92 | #define APIC_TIMER_BASE_DIV 0x2 | 90 | #define APIC_TIMER_BASE_DIV 0x2 |
| 93 | #define APIC_LVT_TIMER_PERIODIC (1<<17) | 91 | #define APIC_LVT_TIMER_PERIODIC (1 << 17) |
| 94 | #define APIC_LVT_MASKED (1<<16) | 92 | #define APIC_LVT_MASKED (1 << 16) |
| 95 | #define APIC_LVT_LEVEL_TRIGGER (1<<15) | 93 | #define APIC_LVT_LEVEL_TRIGGER (1 << 15) |
| 96 | #define APIC_LVT_REMOTE_IRR (1<<14) | 94 | #define APIC_LVT_REMOTE_IRR (1 << 14) |
| 97 | #define APIC_INPUT_POLARITY (1<<13) | 95 | #define APIC_INPUT_POLARITY (1 << 13) |
| 98 | #define APIC_SEND_PENDING (1<<12) | 96 | #define APIC_SEND_PENDING (1 << 12) |
| 99 | #define APIC_MODE_MASK 0x700 | 97 | #define APIC_MODE_MASK 0x700 |
| 100 | #define GET_APIC_DELIVERY_MODE(x) (((x)>>8)&0x7) | 98 | #define GET_APIC_DELIVERY_MODE(x) (((x) >> 8) & 0x7) |
| 101 | #define SET_APIC_DELIVERY_MODE(x, y) (((x)&~0x700)|((y)<<8)) | 99 | #define SET_APIC_DELIVERY_MODE(x, y) (((x) & ~0x700) | ((y) << 8)) |
| 102 | #define APIC_MODE_FIXED 0x0 | 100 | #define APIC_MODE_FIXED 0x0 |
| 103 | #define APIC_MODE_NMI 0x4 | 101 | #define APIC_MODE_NMI 0x4 |
| 104 | #define APIC_MODE_EXTINT 0x7 | 102 | #define APIC_MODE_EXTINT 0x7 |
| @@ -107,7 +105,7 @@ | |||
| 107 | #define APIC_TMICT 0x380 | 105 | #define APIC_TMICT 0x380 |
| 108 | #define APIC_TMCCT 0x390 | 106 | #define APIC_TMCCT 0x390 |
| 109 | #define APIC_TDCR 0x3E0 | 107 | #define APIC_TDCR 0x3E0 |
| 110 | #define APIC_TDR_DIV_TMBASE (1<<2) | 108 | #define APIC_TDR_DIV_TMBASE (1 << 2) |
| 111 | #define APIC_TDR_DIV_1 0xB | 109 | #define APIC_TDR_DIV_1 0xB |
| 112 | #define APIC_TDR_DIV_2 0x0 | 110 | #define APIC_TDR_DIV_2 0x0 |
| 113 | #define APIC_TDR_DIV_4 0x1 | 111 | #define APIC_TDR_DIV_4 0x1 |
| @@ -117,14 +115,14 @@ | |||
| 117 | #define APIC_TDR_DIV_64 0x9 | 115 | #define APIC_TDR_DIV_64 0x9 |
| 118 | #define APIC_TDR_DIV_128 0xA | 116 | #define APIC_TDR_DIV_128 0xA |
| 119 | #define APIC_EILVT0 0x500 | 117 | #define APIC_EILVT0 0x500 |
| 120 | #define APIC_EILVT_NR_AMD_K8 1 /* Number of extended interrupts */ | 118 | #define APIC_EILVT_NR_AMD_K8 1 /* # of extended interrupts */ |
| 121 | #define APIC_EILVT_NR_AMD_10H 4 | 119 | #define APIC_EILVT_NR_AMD_10H 4 |
| 122 | #define APIC_EILVT_LVTOFF(x) (((x)>>4)&0xF) | 120 | #define APIC_EILVT_LVTOFF(x) (((x) >> 4) & 0xF) |
| 123 | #define APIC_EILVT_MSG_FIX 0x0 | 121 | #define APIC_EILVT_MSG_FIX 0x0 |
| 124 | #define APIC_EILVT_MSG_SMI 0x2 | 122 | #define APIC_EILVT_MSG_SMI 0x2 |
| 125 | #define APIC_EILVT_MSG_NMI 0x4 | 123 | #define APIC_EILVT_MSG_NMI 0x4 |
| 126 | #define APIC_EILVT_MSG_EXT 0x7 | 124 | #define APIC_EILVT_MSG_EXT 0x7 |
| 127 | #define APIC_EILVT_MASKED (1<<16) | 125 | #define APIC_EILVT_MASKED (1 << 16) |
| 128 | #define APIC_EILVT1 0x510 | 126 | #define APIC_EILVT1 0x510 |
| 129 | #define APIC_EILVT2 0x520 | 127 | #define APIC_EILVT2 0x520 |
| 130 | #define APIC_EILVT3 0x530 | 128 | #define APIC_EILVT3 0x530 |
| @@ -135,7 +133,7 @@ | |||
| 135 | # define MAX_IO_APICS 64 | 133 | # define MAX_IO_APICS 64 |
| 136 | #else | 134 | #else |
| 137 | # define MAX_IO_APICS 128 | 135 | # define MAX_IO_APICS 128 |
| 138 | # define MAX_LOCAL_APIC 256 | 136 | # define MAX_LOCAL_APIC 32768 |
| 139 | #endif | 137 | #endif |
| 140 | 138 | ||
| 141 | /* | 139 | /* |
| @@ -408,6 +406,9 @@ struct local_apic { | |||
| 408 | 406 | ||
| 409 | #undef u32 | 407 | #undef u32 |
| 410 | 408 | ||
| 411 | #define BAD_APICID 0xFFu | 409 | #ifdef CONFIG_X86_32 |
| 412 | 410 | #define BAD_APICID 0xFFu | |
| 411 | #else | ||
| 412 | #define BAD_APICID 0xFFFFu | ||
| 413 | #endif | ||
| 413 | #endif | 414 | #endif |
diff --git a/include/asm-x86/atomic_32.h b/include/asm-x86/atomic_32.h index 437aac801711..21a4825148c0 100644 --- a/include/asm-x86/atomic_32.h +++ b/include/asm-x86/atomic_32.h | |||
| @@ -15,138 +15,133 @@ | |||
| 15 | * on us. We need to use _exactly_ the address the user gave us, | 15 | * on us. We need to use _exactly_ the address the user gave us, |
| 16 | * not some alias that contains the same information. | 16 | * not some alias that contains the same information. |
| 17 | */ | 17 | */ |
| 18 | typedef struct { int counter; } atomic_t; | 18 | typedef struct { |
| 19 | int counter; | ||
| 20 | } atomic_t; | ||
| 19 | 21 | ||
| 20 | #define ATOMIC_INIT(i) { (i) } | 22 | #define ATOMIC_INIT(i) { (i) } |
| 21 | 23 | ||
| 22 | /** | 24 | /** |
| 23 | * atomic_read - read atomic variable | 25 | * atomic_read - read atomic variable |
| 24 | * @v: pointer of type atomic_t | 26 | * @v: pointer of type atomic_t |
| 25 | * | 27 | * |
| 26 | * Atomically reads the value of @v. | 28 | * Atomically reads the value of @v. |
| 27 | */ | 29 | */ |
| 28 | #define atomic_read(v) ((v)->counter) | 30 | #define atomic_read(v) ((v)->counter) |
| 29 | 31 | ||
| 30 | /** | 32 | /** |
| 31 | * atomic_set - set atomic variable | 33 | * atomic_set - set atomic variable |
| 32 | * @v: pointer of type atomic_t | 34 | * @v: pointer of type atomic_t |
| 33 | * @i: required value | 35 | * @i: required value |
| 34 | * | 36 | * |
| 35 | * Atomically sets the value of @v to @i. | 37 | * Atomically sets the value of @v to @i. |
| 36 | */ | 38 | */ |
| 37 | #define atomic_set(v,i) (((v)->counter) = (i)) | 39 | #define atomic_set(v, i) (((v)->counter) = (i)) |
| 38 | 40 | ||
| 39 | /** | 41 | /** |
| 40 | * atomic_add - add integer to atomic variable | 42 | * atomic_add - add integer to atomic variable |
| 41 | * @i: integer value to add | 43 | * @i: integer value to add |
| 42 | * @v: pointer of type atomic_t | 44 | * @v: pointer of type atomic_t |
| 43 | * | 45 | * |
| 44 | * Atomically adds @i to @v. | 46 | * Atomically adds @i to @v. |
| 45 | */ | 47 | */ |
| 46 | static __inline__ void atomic_add(int i, atomic_t *v) | 48 | static inline void atomic_add(int i, atomic_t *v) |
| 47 | { | 49 | { |
| 48 | __asm__ __volatile__( | 50 | asm volatile(LOCK_PREFIX "addl %1,%0" |
| 49 | LOCK_PREFIX "addl %1,%0" | 51 | : "+m" (v->counter) |
| 50 | :"+m" (v->counter) | 52 | : "ir" (i)); |
| 51 | :"ir" (i)); | ||
| 52 | } | 53 | } |
| 53 | 54 | ||
| 54 | /** | 55 | /** |
| 55 | * atomic_sub - subtract integer from atomic variable | 56 | * atomic_sub - subtract integer from atomic variable |
| 56 | * @i: integer value to subtract | 57 | * @i: integer value to subtract |
| 57 | * @v: pointer of type atomic_t | 58 | * @v: pointer of type atomic_t |
| 58 | * | 59 | * |
| 59 | * Atomically subtracts @i from @v. | 60 | * Atomically subtracts @i from @v. |
| 60 | */ | 61 | */ |
| 61 | static __inline__ void atomic_sub(int i, atomic_t *v) | 62 | static inline void atomic_sub(int i, atomic_t *v) |
| 62 | { | 63 | { |
| 63 | __asm__ __volatile__( | 64 | asm volatile(LOCK_PREFIX "subl %1,%0" |
| 64 | LOCK_PREFIX "subl %1,%0" | 65 | : "+m" (v->counter) |
| 65 | :"+m" (v->counter) | 66 | : "ir" (i)); |
| 66 | :"ir" (i)); | ||
| 67 | } | 67 | } |
| 68 | 68 | ||
| 69 | /** | 69 | /** |
| 70 | * atomic_sub_and_test - subtract value from variable and test result | 70 | * atomic_sub_and_test - subtract value from variable and test result |
| 71 | * @i: integer value to subtract | 71 | * @i: integer value to subtract |
| 72 | * @v: pointer of type atomic_t | 72 | * @v: pointer of type atomic_t |
| 73 | * | 73 | * |
| 74 | * Atomically subtracts @i from @v and returns | 74 | * Atomically subtracts @i from @v and returns |
| 75 | * true if the result is zero, or false for all | 75 | * true if the result is zero, or false for all |
| 76 | * other cases. | 76 | * other cases. |
| 77 | */ | 77 | */ |
| 78 | static __inline__ int atomic_sub_and_test(int i, atomic_t *v) | 78 | static inline int atomic_sub_and_test(int i, atomic_t *v) |
| 79 | { | 79 | { |
| 80 | unsigned char c; | 80 | unsigned char c; |
| 81 | 81 | ||
| 82 | __asm__ __volatile__( | 82 | asm volatile(LOCK_PREFIX "subl %2,%0; sete %1" |
| 83 | LOCK_PREFIX "subl %2,%0; sete %1" | 83 | : "+m" (v->counter), "=qm" (c) |
| 84 | :"+m" (v->counter), "=qm" (c) | 84 | : "ir" (i) : "memory"); |
| 85 | :"ir" (i) : "memory"); | ||
| 86 | return c; | 85 | return c; |
| 87 | } | 86 | } |
| 88 | 87 | ||
| 89 | /** | 88 | /** |
| 90 | * atomic_inc - increment atomic variable | 89 | * atomic_inc - increment atomic variable |
| 91 | * @v: pointer of type atomic_t | 90 | * @v: pointer of type atomic_t |
| 92 | * | 91 | * |
| 93 | * Atomically increments @v by 1. | 92 | * Atomically increments @v by 1. |
| 94 | */ | 93 | */ |
| 95 | static __inline__ void atomic_inc(atomic_t *v) | 94 | static inline void atomic_inc(atomic_t *v) |
| 96 | { | 95 | { |
| 97 | __asm__ __volatile__( | 96 | asm volatile(LOCK_PREFIX "incl %0" |
| 98 | LOCK_PREFIX "incl %0" | 97 | : "+m" (v->counter)); |
| 99 | :"+m" (v->counter)); | ||
| 100 | } | 98 | } |
| 101 | 99 | ||
| 102 | /** | 100 | /** |
| 103 | * atomic_dec - decrement atomic variable | 101 | * atomic_dec - decrement atomic variable |
| 104 | * @v: pointer of type atomic_t | 102 | * @v: pointer of type atomic_t |
| 105 | * | 103 | * |
| 106 | * Atomically decrements @v by 1. | 104 | * Atomically decrements @v by 1. |
| 107 | */ | 105 | */ |
| 108 | static __inline__ void atomic_dec(atomic_t *v) | 106 | static inline void atomic_dec(atomic_t *v) |
| 109 | { | 107 | { |
| 110 | __asm__ __volatile__( | 108 | asm volatile(LOCK_PREFIX "decl %0" |
| 111 | LOCK_PREFIX "decl %0" | 109 | : "+m" (v->counter)); |
| 112 | :"+m" (v->counter)); | ||
| 113 | } | 110 | } |
| 114 | 111 | ||
| 115 | /** | 112 | /** |
| 116 | * atomic_dec_and_test - decrement and test | 113 | * atomic_dec_and_test - decrement and test |
| 117 | * @v: pointer of type atomic_t | 114 | * @v: pointer of type atomic_t |
| 118 | * | 115 | * |
| 119 | * Atomically decrements @v by 1 and | 116 | * Atomically decrements @v by 1 and |
| 120 | * returns true if the result is 0, or false for all other | 117 | * returns true if the result is 0, or false for all other |
| 121 | * cases. | 118 | * cases. |
| 122 | */ | 119 | */ |
| 123 | static __inline__ int atomic_dec_and_test(atomic_t *v) | 120 | static inline int atomic_dec_and_test(atomic_t *v) |
| 124 | { | 121 | { |
| 125 | unsigned char c; | 122 | unsigned char c; |
| 126 | 123 | ||
| 127 | __asm__ __volatile__( | 124 | asm volatile(LOCK_PREFIX "decl %0; sete %1" |
| 128 | LOCK_PREFIX "decl %0; sete %1" | 125 | : "+m" (v->counter), "=qm" (c) |
| 129 | :"+m" (v->counter), "=qm" (c) | 126 | : : "memory"); |
| 130 | : : "memory"); | ||
| 131 | return c != 0; | 127 | return c != 0; |
| 132 | } | 128 | } |
| 133 | 129 | ||
| 134 | /** | 130 | /** |
| 135 | * atomic_inc_and_test - increment and test | 131 | * atomic_inc_and_test - increment and test |
| 136 | * @v: pointer of type atomic_t | 132 | * @v: pointer of type atomic_t |
| 137 | * | 133 | * |
| 138 | * Atomically increments @v by 1 | 134 | * Atomically increments @v by 1 |
| 139 | * and returns true if the result is zero, or false for all | 135 | * and returns true if the result is zero, or false for all |
| 140 | * other cases. | 136 | * other cases. |
| 141 | */ | 137 | */ |
| 142 | static __inline__ int atomic_inc_and_test(atomic_t *v) | 138 | static inline int atomic_inc_and_test(atomic_t *v) |
| 143 | { | 139 | { |
| 144 | unsigned char c; | 140 | unsigned char c; |
| 145 | 141 | ||
| 146 | __asm__ __volatile__( | 142 | asm volatile(LOCK_PREFIX "incl %0; sete %1" |
| 147 | LOCK_PREFIX "incl %0; sete %1" | 143 | : "+m" (v->counter), "=qm" (c) |
| 148 | :"+m" (v->counter), "=qm" (c) | 144 | : : "memory"); |
| 149 | : : "memory"); | ||
| 150 | return c != 0; | 145 | return c != 0; |
| 151 | } | 146 | } |
| 152 | 147 | ||
| @@ -154,19 +149,18 @@ static __inline__ int atomic_inc_and_test(atomic_t *v) | |||
| 154 | * atomic_add_negative - add and test if negative | 149 | * atomic_add_negative - add and test if negative |
| 155 | * @v: pointer of type atomic_t | 150 | * @v: pointer of type atomic_t |
| 156 | * @i: integer value to add | 151 | * @i: integer value to add |
| 157 | * | 152 | * |
| 158 | * Atomically adds @i to @v and returns true | 153 | * Atomically adds @i to @v and returns true |
| 159 | * if the result is negative, or false when | 154 | * if the result is negative, or false when |
| 160 | * result is greater than or equal to zero. | 155 | * result is greater than or equal to zero. |
| 161 | */ | 156 | */ |
| 162 | static __inline__ int atomic_add_negative(int i, atomic_t *v) | 157 | static inline int atomic_add_negative(int i, atomic_t *v) |
| 163 | { | 158 | { |
| 164 | unsigned char c; | 159 | unsigned char c; |
| 165 | 160 | ||
| 166 | __asm__ __volatile__( | 161 | asm volatile(LOCK_PREFIX "addl %2,%0; sets %1" |
| 167 | LOCK_PREFIX "addl %2,%0; sets %1" | 162 | : "+m" (v->counter), "=qm" (c) |
| 168 | :"+m" (v->counter), "=qm" (c) | 163 | : "ir" (i) : "memory"); |
| 169 | :"ir" (i) : "memory"); | ||
| 170 | return c; | 164 | return c; |
| 171 | } | 165 | } |
| 172 | 166 | ||
| @@ -177,20 +171,19 @@ static __inline__ int atomic_add_negative(int i, atomic_t *v) | |||
| 177 | * | 171 | * |
| 178 | * Atomically adds @i to @v and returns @i + @v | 172 | * Atomically adds @i to @v and returns @i + @v |
| 179 | */ | 173 | */ |
| 180 | static __inline__ int atomic_add_return(int i, atomic_t *v) | 174 | static inline int atomic_add_return(int i, atomic_t *v) |
| 181 | { | 175 | { |
| 182 | int __i; | 176 | int __i; |
| 183 | #ifdef CONFIG_M386 | 177 | #ifdef CONFIG_M386 |
| 184 | unsigned long flags; | 178 | unsigned long flags; |
| 185 | if(unlikely(boot_cpu_data.x86 <= 3)) | 179 | if (unlikely(boot_cpu_data.x86 <= 3)) |
| 186 | goto no_xadd; | 180 | goto no_xadd; |
| 187 | #endif | 181 | #endif |
| 188 | /* Modern 486+ processor */ | 182 | /* Modern 486+ processor */ |
| 189 | __i = i; | 183 | __i = i; |
| 190 | __asm__ __volatile__( | 184 | asm volatile(LOCK_PREFIX "xaddl %0, %1" |
| 191 | LOCK_PREFIX "xaddl %0, %1" | 185 | : "+r" (i), "+m" (v->counter) |
| 192 | :"+r" (i), "+m" (v->counter) | 186 | : : "memory"); |
| 193 | : : "memory"); | ||
| 194 | return i + __i; | 187 | return i + __i; |
| 195 | 188 | ||
| 196 | #ifdef CONFIG_M386 | 189 | #ifdef CONFIG_M386 |
| @@ -210,9 +203,9 @@ no_xadd: /* Legacy 386 processor */ | |||
| 210 | * | 203 | * |
| 211 | * Atomically subtracts @i from @v and returns @v - @i | 204 | * Atomically subtracts @i from @v and returns @v - @i |
| 212 | */ | 205 | */ |
| 213 | static __inline__ int atomic_sub_return(int i, atomic_t *v) | 206 | static inline int atomic_sub_return(int i, atomic_t *v) |
| 214 | { | 207 | { |
| 215 | return atomic_add_return(-i,v); | 208 | return atomic_add_return(-i, v); |
| 216 | } | 209 | } |
| 217 | 210 | ||
| 218 | #define atomic_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), (old), (new))) | 211 | #define atomic_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), (old), (new))) |
| @@ -227,7 +220,7 @@ static __inline__ int atomic_sub_return(int i, atomic_t *v) | |||
| 227 | * Atomically adds @a to @v, so long as @v was not already @u. | 220 | * Atomically adds @a to @v, so long as @v was not already @u. |
| 228 | * Returns non-zero if @v was not @u, and zero otherwise. | 221 | * Returns non-zero if @v was not @u, and zero otherwise. |
| 229 | */ | 222 | */ |
| 230 | static __inline__ int atomic_add_unless(atomic_t *v, int a, int u) | 223 | static inline int atomic_add_unless(atomic_t *v, int a, int u) |
| 231 | { | 224 | { |
| 232 | int c, old; | 225 | int c, old; |
| 233 | c = atomic_read(v); | 226 | c = atomic_read(v); |
| @@ -244,17 +237,17 @@ static __inline__ int atomic_add_unless(atomic_t *v, int a, int u) | |||
| 244 | 237 | ||
| 245 | #define atomic_inc_not_zero(v) atomic_add_unless((v), 1, 0) | 238 | #define atomic_inc_not_zero(v) atomic_add_unless((v), 1, 0) |
| 246 | 239 | ||
| 247 | #define atomic_inc_return(v) (atomic_add_return(1,v)) | 240 | #define atomic_inc_return(v) (atomic_add_return(1, v)) |
| 248 | #define atomic_dec_return(v) (atomic_sub_return(1,v)) | 241 | #define atomic_dec_return(v) (atomic_sub_return(1, v)) |
| 249 | 242 | ||
| 250 | /* These are x86-specific, used by some header files */ | 243 | /* These are x86-specific, used by some header files */ |
| 251 | #define atomic_clear_mask(mask, addr) \ | 244 | #define atomic_clear_mask(mask, addr) \ |
| 252 | __asm__ __volatile__(LOCK_PREFIX "andl %0,%1" \ | 245 | asm volatile(LOCK_PREFIX "andl %0,%1" \ |
| 253 | : : "r" (~(mask)),"m" (*addr) : "memory") | 246 | : : "r" (~(mask)), "m" (*(addr)) : "memory") |
| 254 | 247 | ||
| 255 | #define atomic_set_mask(mask, addr) \ | 248 | #define atomic_set_mask(mask, addr) \ |
| 256 | __asm__ __volatile__(LOCK_PREFIX "orl %0,%1" \ | 249 | asm volatile(LOCK_PREFIX "orl %0,%1" \ |
| 257 | : : "r" (mask),"m" (*(addr)) : "memory") | 250 | : : "r" (mask), "m" (*(addr)) : "memory") |
| 258 | 251 | ||
| 259 | /* Atomic operations are already serializing on x86 */ | 252 | /* Atomic operations are already serializing on x86 */ |
| 260 | #define smp_mb__before_atomic_dec() barrier() | 253 | #define smp_mb__before_atomic_dec() barrier() |
diff --git a/include/asm-x86/atomic_64.h b/include/asm-x86/atomic_64.h index 2d20a7a19f62..3e0cd7d38335 100644 --- a/include/asm-x86/atomic_64.h +++ b/include/asm-x86/atomic_64.h | |||
| @@ -22,140 +22,135 @@ | |||
| 22 | * on us. We need to use _exactly_ the address the user gave us, | 22 | * on us. We need to use _exactly_ the address the user gave us, |
| 23 | * not some alias that contains the same information. | 23 | * not some alias that contains the same information. |
| 24 | */ | 24 | */ |
| 25 | typedef struct { int counter; } atomic_t; | 25 | typedef struct { |
| 26 | int counter; | ||
| 27 | } atomic_t; | ||
| 26 | 28 | ||
| 27 | #define ATOMIC_INIT(i) { (i) } | 29 | #define ATOMIC_INIT(i) { (i) } |
| 28 | 30 | ||
| 29 | /** | 31 | /** |
| 30 | * atomic_read - read atomic variable | 32 | * atomic_read - read atomic variable |
| 31 | * @v: pointer of type atomic_t | 33 | * @v: pointer of type atomic_t |
| 32 | * | 34 | * |
| 33 | * Atomically reads the value of @v. | 35 | * Atomically reads the value of @v. |
| 34 | */ | 36 | */ |
| 35 | #define atomic_read(v) ((v)->counter) | 37 | #define atomic_read(v) ((v)->counter) |
| 36 | 38 | ||
| 37 | /** | 39 | /** |
| 38 | * atomic_set - set atomic variable | 40 | * atomic_set - set atomic variable |
| 39 | * @v: pointer of type atomic_t | 41 | * @v: pointer of type atomic_t |
| 40 | * @i: required value | 42 | * @i: required value |
| 41 | * | 43 | * |
| 42 | * Atomically sets the value of @v to @i. | 44 | * Atomically sets the value of @v to @i. |
| 43 | */ | 45 | */ |
| 44 | #define atomic_set(v,i) (((v)->counter) = (i)) | 46 | #define atomic_set(v, i) (((v)->counter) = (i)) |
| 45 | 47 | ||
| 46 | /** | 48 | /** |
| 47 | * atomic_add - add integer to atomic variable | 49 | * atomic_add - add integer to atomic variable |
| 48 | * @i: integer value to add | 50 | * @i: integer value to add |
| 49 | * @v: pointer of type atomic_t | 51 | * @v: pointer of type atomic_t |
| 50 | * | 52 | * |
| 51 | * Atomically adds @i to @v. | 53 | * Atomically adds @i to @v. |
| 52 | */ | 54 | */ |
| 53 | static __inline__ void atomic_add(int i, atomic_t *v) | 55 | static inline void atomic_add(int i, atomic_t *v) |
| 54 | { | 56 | { |
| 55 | __asm__ __volatile__( | 57 | asm volatile(LOCK_PREFIX "addl %1,%0" |
| 56 | LOCK_PREFIX "addl %1,%0" | 58 | : "=m" (v->counter) |
| 57 | :"=m" (v->counter) | 59 | : "ir" (i), "m" (v->counter)); |
| 58 | :"ir" (i), "m" (v->counter)); | ||
| 59 | } | 60 | } |
| 60 | 61 | ||
| 61 | /** | 62 | /** |
| 62 | * atomic_sub - subtract the atomic variable | 63 | * atomic_sub - subtract the atomic variable |
| 63 | * @i: integer value to subtract | 64 | * @i: integer value to subtract |
| 64 | * @v: pointer of type atomic_t | 65 | * @v: pointer of type atomic_t |
| 65 | * | 66 | * |
| 66 | * Atomically subtracts @i from @v. | 67 | * Atomically subtracts @i from @v. |
| 67 | */ | 68 | */ |
| 68 | static __inline__ void atomic_sub(int i, atomic_t *v) | 69 | static inline void atomic_sub(int i, atomic_t *v) |
| 69 | { | 70 | { |
| 70 | __asm__ __volatile__( | 71 | asm volatile(LOCK_PREFIX "subl %1,%0" |
| 71 | LOCK_PREFIX "subl %1,%0" | 72 | : "=m" (v->counter) |
| 72 | :"=m" (v->counter) | 73 | : "ir" (i), "m" (v->counter)); |
| 73 | :"ir" (i), "m" (v->counter)); | ||
| 74 | } | 74 | } |
| 75 | 75 | ||
| 76 | /** | 76 | /** |
| 77 | * atomic_sub_and_test - subtract value from variable and test result | 77 | * atomic_sub_and_test - subtract value from variable and test result |
| 78 | * @i: integer value to subtract | 78 | * @i: integer value to subtract |
| 79 | * @v: pointer of type atomic_t | 79 | * @v: pointer of type atomic_t |
| 80 | * | 80 | * |
| 81 | * Atomically subtracts @i from @v and returns | 81 | * Atomically subtracts @i from @v and returns |
| 82 | * true if the result is zero, or false for all | 82 | * true if the result is zero, or false for all |
| 83 | * other cases. | 83 | * other cases. |
| 84 | */ | 84 | */ |
| 85 | static __inline__ int atomic_sub_and_test(int i, atomic_t *v) | 85 | static inline int atomic_sub_and_test(int i, atomic_t *v) |
| 86 | { | 86 | { |
| 87 | unsigned char c; | 87 | unsigned char c; |
| 88 | 88 | ||
| 89 | __asm__ __volatile__( | 89 | asm volatile(LOCK_PREFIX "subl %2,%0; sete %1" |
| 90 | LOCK_PREFIX "subl %2,%0; sete %1" | 90 | : "=m" (v->counter), "=qm" (c) |
| 91 | :"=m" (v->counter), "=qm" (c) | 91 | : "ir" (i), "m" (v->counter) : "memory"); |
| 92 | :"ir" (i), "m" (v->counter) : "memory"); | ||
| 93 | return c; | 92 | return c; |
| 94 | } | 93 | } |
| 95 | 94 | ||
| 96 | /** | 95 | /** |
| 97 | * atomic_inc - increment atomic variable | 96 | * atomic_inc - increment atomic variable |
| 98 | * @v: pointer of type atomic_t | 97 | * @v: pointer of type atomic_t |
| 99 | * | 98 | * |
| 100 | * Atomically increments @v by 1. | 99 | * Atomically increments @v by 1. |
| 101 | */ | 100 | */ |
| 102 | static __inline__ void atomic_inc(atomic_t *v) | 101 | static inline void atomic_inc(atomic_t *v) |
| 103 | { | 102 | { |
| 104 | __asm__ __volatile__( | 103 | asm volatile(LOCK_PREFIX "incl %0" |
| 105 | LOCK_PREFIX "incl %0" | 104 | : "=m" (v->counter) |
| 106 | :"=m" (v->counter) | 105 | : "m" (v->counter)); |
| 107 | :"m" (v->counter)); | ||
| 108 | } | 106 | } |
| 109 | 107 | ||
| 110 | /** | 108 | /** |
| 111 | * atomic_dec - decrement atomic variable | 109 | * atomic_dec - decrement atomic variable |
| 112 | * @v: pointer of type atomic_t | 110 | * @v: pointer of type atomic_t |
| 113 | * | 111 | * |
| 114 | * Atomically decrements @v by 1. | 112 | * Atomically decrements @v by 1. |
| 115 | */ | 113 | */ |
| 116 | static __inline__ void atomic_dec(atomic_t *v) | 114 | static inline void atomic_dec(atomic_t *v) |
| 117 | { | 115 | { |
| 118 | __asm__ __volatile__( | 116 | asm volatile(LOCK_PREFIX "decl %0" |
| 119 | LOCK_PREFIX "decl %0" | 117 | : "=m" (v->counter) |
| 120 | :"=m" (v->counter) | 118 | : "m" (v->counter)); |
| 121 | :"m" (v->counter)); | ||
| 122 | } | 119 | } |
| 123 | 120 | ||
| 124 | /** | 121 | /** |
| 125 | * atomic_dec_and_test - decrement and test | 122 | * atomic_dec_and_test - decrement and test |
| 126 | * @v: pointer of type atomic_t | 123 | * @v: pointer of type atomic_t |
| 127 | * | 124 | * |
| 128 | * Atomically decrements @v by 1 and | 125 | * Atomically decrements @v by 1 and |
| 129 | * returns true if the result is 0, or false for all other | 126 | * returns true if the result is 0, or false for all other |
| 130 | * cases. | 127 | * cases. |
| 131 | */ | 128 | */ |
| 132 | static __inline__ int atomic_dec_and_test(atomic_t *v) | 129 | static inline int atomic_dec_and_test(atomic_t *v) |
| 133 | { | 130 | { |
| 134 | unsigned char c; | 131 | unsigned char c; |
| 135 | 132 | ||
| 136 | __asm__ __volatile__( | 133 | asm volatile(LOCK_PREFIX "decl %0; sete %1" |
| 137 | LOCK_PREFIX "decl %0; sete %1" | 134 | : "=m" (v->counter), "=qm" (c) |
| 138 | :"=m" (v->counter), "=qm" (c) | 135 | : "m" (v->counter) : "memory"); |
| 139 | :"m" (v->counter) : "memory"); | ||
| 140 | return c != 0; | 136 | return c != 0; |
| 141 | } | 137 | } |
| 142 | 138 | ||
| 143 | /** | 139 | /** |
| 144 | * atomic_inc_and_test - increment and test | 140 | * atomic_inc_and_test - increment and test |
| 145 | * @v: pointer of type atomic_t | 141 | * @v: pointer of type atomic_t |
| 146 | * | 142 | * |
| 147 | * Atomically increments @v by 1 | 143 | * Atomically increments @v by 1 |
| 148 | * and returns true if the result is zero, or false for all | 144 | * and returns true if the result is zero, or false for all |
| 149 | * other cases. | 145 | * other cases. |
| 150 | */ | 146 | */ |
| 151 | static __inline__ int atomic_inc_and_test(atomic_t *v) | 147 | static inline int atomic_inc_and_test(atomic_t *v) |
| 152 | { | 148 | { |
| 153 | unsigned char c; | 149 | unsigned char c; |
| 154 | 150 | ||
| 155 | __asm__ __volatile__( | 151 | asm volatile(LOCK_PREFIX "incl %0; sete %1" |
| 156 | LOCK_PREFIX "incl %0; sete %1" | 152 | : "=m" (v->counter), "=qm" (c) |
| 157 | :"=m" (v->counter), "=qm" (c) | 153 | : "m" (v->counter) : "memory"); |
| 158 | :"m" (v->counter) : "memory"); | ||
| 159 | return c != 0; | 154 | return c != 0; |
| 160 | } | 155 | } |
| 161 | 156 | ||
| @@ -163,19 +158,18 @@ static __inline__ int atomic_inc_and_test(atomic_t *v) | |||
| 163 | * atomic_add_negative - add and test if negative | 158 | * atomic_add_negative - add and test if negative |
| 164 | * @i: integer value to add | 159 | * @i: integer value to add |
| 165 | * @v: pointer of type atomic_t | 160 | * @v: pointer of type atomic_t |
| 166 | * | 161 | * |
| 167 | * Atomically adds @i to @v and returns true | 162 | * Atomically adds @i to @v and returns true |
| 168 | * if the result is negative, or false when | 163 | * if the result is negative, or false when |
| 169 | * result is greater than or equal to zero. | 164 | * result is greater than or equal to zero. |
| 170 | */ | 165 | */ |
| 171 | static __inline__ int atomic_add_negative(int i, atomic_t *v) | 166 | static inline int atomic_add_negative(int i, atomic_t *v) |
| 172 | { | 167 | { |
| 173 | unsigned char c; | 168 | unsigned char c; |
| 174 | 169 | ||
| 175 | __asm__ __volatile__( | 170 | asm volatile(LOCK_PREFIX "addl %2,%0; sets %1" |
| 176 | LOCK_PREFIX "addl %2,%0; sets %1" | 171 | : "=m" (v->counter), "=qm" (c) |
| 177 | :"=m" (v->counter), "=qm" (c) | 172 | : "ir" (i), "m" (v->counter) : "memory"); |
| 178 | :"ir" (i), "m" (v->counter) : "memory"); | ||
| 179 | return c; | 173 | return c; |
| 180 | } | 174 | } |
| 181 | 175 | ||
| @@ -186,27 +180,28 @@ static __inline__ int atomic_add_negative(int i, atomic_t *v) | |||
| 186 | * | 180 | * |
| 187 | * Atomically adds @i to @v and returns @i + @v | 181 | * Atomically adds @i to @v and returns @i + @v |
| 188 | */ | 182 | */ |
| 189 | static __inline__ int atomic_add_return(int i, atomic_t *v) | 183 | static inline int atomic_add_return(int i, atomic_t *v) |
| 190 | { | 184 | { |
| 191 | int __i = i; | 185 | int __i = i; |
| 192 | __asm__ __volatile__( | 186 | asm volatile(LOCK_PREFIX "xaddl %0, %1" |
| 193 | LOCK_PREFIX "xaddl %0, %1" | 187 | : "+r" (i), "+m" (v->counter) |
| 194 | :"+r" (i), "+m" (v->counter) | 188 | : : "memory"); |
| 195 | : : "memory"); | ||
| 196 | return i + __i; | 189 | return i + __i; |
| 197 | } | 190 | } |
| 198 | 191 | ||
| 199 | static __inline__ int atomic_sub_return(int i, atomic_t *v) | 192 | static inline int atomic_sub_return(int i, atomic_t *v) |
| 200 | { | 193 | { |
| 201 | return atomic_add_return(-i,v); | 194 | return atomic_add_return(-i, v); |
| 202 | } | 195 | } |
| 203 | 196 | ||
| 204 | #define atomic_inc_return(v) (atomic_add_return(1,v)) | 197 | #define atomic_inc_return(v) (atomic_add_return(1, v)) |
| 205 | #define atomic_dec_return(v) (atomic_sub_return(1,v)) | 198 | #define atomic_dec_return(v) (atomic_sub_return(1, v)) |
| 206 | 199 | ||
| 207 | /* An 64bit atomic type */ | 200 | /* An 64bit atomic type */ |
| 208 | 201 | ||
| 209 | typedef struct { long counter; } atomic64_t; | 202 | typedef struct { |
| 203 | long counter; | ||
| 204 | } atomic64_t; | ||
| 210 | 205 | ||
| 211 | #define ATOMIC64_INIT(i) { (i) } | 206 | #define ATOMIC64_INIT(i) { (i) } |
| 212 | 207 | ||
| @@ -226,7 +221,7 @@ typedef struct { long counter; } atomic64_t; | |||
| 226 | * | 221 | * |
| 227 | * Atomically sets the value of @v to @i. | 222 | * Atomically sets the value of @v to @i. |
| 228 | */ | 223 | */ |
| 229 | #define atomic64_set(v,i) (((v)->counter) = (i)) | 224 | #define atomic64_set(v, i) (((v)->counter) = (i)) |
| 230 | 225 | ||
| 231 | /** | 226 | /** |
| 232 | * atomic64_add - add integer to atomic64 variable | 227 | * atomic64_add - add integer to atomic64 variable |
| @@ -235,12 +230,11 @@ typedef struct { long counter; } atomic64_t; | |||
| 235 | * | 230 | * |
| 236 | * Atomically adds @i to @v. | 231 | * Atomically adds @i to @v. |
| 237 | */ | 232 | */ |
| 238 | static __inline__ void atomic64_add(long i, atomic64_t *v) | 233 | static inline void atomic64_add(long i, atomic64_t *v) |
| 239 | { | 234 | { |
| 240 | __asm__ __volatile__( | 235 | asm volatile(LOCK_PREFIX "addq %1,%0" |
| 241 | LOCK_PREFIX "addq %1,%0" | 236 | : "=m" (v->counter) |
| 242 | :"=m" (v->counter) | 237 | : "ir" (i), "m" (v->counter)); |
| 243 | :"ir" (i), "m" (v->counter)); | ||
| 244 | } | 238 | } |
| 245 | 239 | ||
| 246 | /** | 240 | /** |
| @@ -250,12 +244,11 @@ static __inline__ void atomic64_add(long i, atomic64_t *v) | |||
| 250 | * | 244 | * |
| 251 | * Atomically subtracts @i from @v. | 245 | * Atomically subtracts @i from @v. |
| 252 | */ | 246 | */ |
| 253 | static __inline__ void atomic64_sub(long i, atomic64_t *v) | 247 | static inline void atomic64_sub(long i, atomic64_t *v) |
| 254 | { | 248 | { |
| 255 | __asm__ __volatile__( | 249 | asm volatile(LOCK_PREFIX "subq %1,%0" |
| 256 | LOCK_PREFIX "subq %1,%0" | 250 | : "=m" (v->counter) |
| 257 | :"=m" (v->counter) | 251 | : "ir" (i), "m" (v->counter)); |
| 258 | :"ir" (i), "m" (v->counter)); | ||
| 259 | } | 252 | } |
| 260 | 253 | ||
| 261 | /** | 254 | /** |
| @@ -267,14 +260,13 @@ static __inline__ void atomic64_sub(long i, atomic64_t *v) | |||
| 267 | * true if the result is zero, or false for all | 260 | * true if the result is zero, or false for all |
| 268 | * other cases. | 261 | * other cases. |
| 269 | */ | 262 | */ |
| 270 | static __inline__ int atomic64_sub_and_test(long i, atomic64_t *v) | 263 | static inline int atomic64_sub_and_test(long i, atomic64_t *v) |
| 271 | { | 264 | { |
| 272 | unsigned char c; | 265 | unsigned char c; |
| 273 | 266 | ||
| 274 | __asm__ __volatile__( | 267 | asm volatile(LOCK_PREFIX "subq %2,%0; sete %1" |
| 275 | LOCK_PREFIX "subq %2,%0; sete %1" | 268 | : "=m" (v->counter), "=qm" (c) |
| 276 | :"=m" (v->counter), "=qm" (c) | 269 | : "ir" (i), "m" (v->counter) : "memory"); |
| 277 | :"ir" (i), "m" (v->counter) : "memory"); | ||
| 278 | return c; | 270 | return c; |
| 279 | } | 271 | } |
| 280 | 272 | ||
| @@ -284,12 +276,11 @@ static __inline__ int atomic64_sub_and_test(long i, atomic64_t *v) | |||
| 284 | * | 276 | * |
| 285 | * Atomically increments @v by 1. | 277 | * Atomically increments @v by 1. |
| 286 | */ | 278 | */ |
| 287 | static __inline__ void atomic64_inc(atomic64_t *v) | 279 | static inline void atomic64_inc(atomic64_t *v) |
| 288 | { | 280 | { |
| 289 | __asm__ __volatile__( | 281 | asm volatile(LOCK_PREFIX "incq %0" |
| 290 | LOCK_PREFIX "incq %0" | 282 | : "=m" (v->counter) |
| 291 | :"=m" (v->counter) | 283 | : "m" (v->counter)); |
| 292 | :"m" (v->counter)); | ||
| 293 | } | 284 | } |
| 294 | 285 | ||
| 295 | /** | 286 | /** |
| @@ -298,12 +289,11 @@ static __inline__ void atomic64_inc(atomic64_t *v) | |||
| 298 | * | 289 | * |
| 299 | * Atomically decrements @v by 1. | 290 | * Atomically decrements @v by 1. |
| 300 | */ | 291 | */ |
| 301 | static __inline__ void atomic64_dec(atomic64_t *v) | 292 | static inline void atomic64_dec(atomic64_t *v) |
| 302 | { | 293 | { |
| 303 | __asm__ __volatile__( | 294 | asm volatile(LOCK_PREFIX "decq %0" |
| 304 | LOCK_PREFIX "decq %0" | 295 | : "=m" (v->counter) |
| 305 | :"=m" (v->counter) | 296 | : "m" (v->counter)); |
| 306 | :"m" (v->counter)); | ||
| 307 | } | 297 | } |
| 308 | 298 | ||
| 309 | /** | 299 | /** |
| @@ -314,14 +304,13 @@ static __inline__ void atomic64_dec(atomic64_t *v) | |||
| 314 | * returns true if the result is 0, or false for all other | 304 | * returns true if the result is 0, or false for all other |
| 315 | * cases. | 305 | * cases. |
| 316 | */ | 306 | */ |
| 317 | static __inline__ int atomic64_dec_and_test(atomic64_t *v) | 307 | static inline int atomic64_dec_and_test(atomic64_t *v) |
| 318 | { | 308 | { |
| 319 | unsigned char c; | 309 | unsigned char c; |
| 320 | 310 | ||
| 321 | __asm__ __volatile__( | 311 | asm volatile(LOCK_PREFIX "decq %0; sete %1" |
| 322 | LOCK_PREFIX "decq %0; sete %1" | 312 | : "=m" (v->counter), "=qm" (c) |
| 323 | :"=m" (v->counter), "=qm" (c) | 313 | : "m" (v->counter) : "memory"); |
| 324 | :"m" (v->counter) : "memory"); | ||
| 325 | return c != 0; | 314 | return c != 0; |
| 326 | } | 315 | } |
| 327 | 316 | ||
| @@ -333,14 +322,13 @@ static __inline__ int atomic64_dec_and_test(atomic64_t *v) | |||
| 333 | * and returns true if the result is zero, or false for all | 322 | * and returns true if the result is zero, or false for all |
| 334 | * other cases. | 323 | * other cases. |
| 335 | */ | 324 | */ |
| 336 | static __inline__ int atomic64_inc_and_test(atomic64_t *v) | 325 | static inline int atomic64_inc_and_test(atomic64_t *v) |
| 337 | { | 326 | { |
| 338 | unsigned char c; | 327 | unsigned char c; |
| 339 | 328 | ||
| 340 | __asm__ __volatile__( | 329 | asm volatile(LOCK_PREFIX "incq %0; sete %1" |
| 341 | LOCK_PREFIX "incq %0; sete %1" | 330 | : "=m" (v->counter), "=qm" (c) |
| 342 | :"=m" (v->counter), "=qm" (c) | 331 | : "m" (v->counter) : "memory"); |
| 343 | :"m" (v->counter) : "memory"); | ||
| 344 | return c != 0; | 332 | return c != 0; |
| 345 | } | 333 | } |
| 346 | 334 | ||
| @@ -353,14 +341,13 @@ static __inline__ int atomic64_inc_and_test(atomic64_t *v) | |||
| 353 | * if the result is negative, or false when | 341 | * if the result is negative, or false when |
| 354 | * result is greater than or equal to zero. | 342 | * result is greater than or equal to zero. |
| 355 | */ | 343 | */ |
| 356 | static __inline__ int atomic64_add_negative(long i, atomic64_t *v) | 344 | static inline int atomic64_add_negative(long i, atomic64_t *v) |
| 357 | { | 345 | { |
| 358 | unsigned char c; | 346 | unsigned char c; |
| 359 | 347 | ||
| 360 | __asm__ __volatile__( | 348 | asm volatile(LOCK_PREFIX "addq %2,%0; sets %1" |
| 361 | LOCK_PREFIX "addq %2,%0; sets %1" | 349 | : "=m" (v->counter), "=qm" (c) |
| 362 | :"=m" (v->counter), "=qm" (c) | 350 | : "ir" (i), "m" (v->counter) : "memory"); |
| 363 | :"ir" (i), "m" (v->counter) : "memory"); | ||
| 364 | return c; | 351 | return c; |
| 365 | } | 352 | } |
| 366 | 353 | ||
| @@ -371,29 +358,28 @@ static __inline__ int atomic64_add_negative(long i, atomic64_t *v) | |||
| 371 | * | 358 | * |
| 372 | * Atomically adds @i to @v and returns @i + @v | 359 | * Atomically adds @i to @v and returns @i + @v |
| 373 | */ | 360 | */ |
| 374 | static __inline__ long atomic64_add_return(long i, atomic64_t *v) | 361 | static inline long atomic64_add_return(long i, atomic64_t *v) |
| 375 | { | 362 | { |
| 376 | long __i = i; | 363 | long __i = i; |
| 377 | __asm__ __volatile__( | 364 | asm volatile(LOCK_PREFIX "xaddq %0, %1;" |
| 378 | LOCK_PREFIX "xaddq %0, %1;" | 365 | : "+r" (i), "+m" (v->counter) |
| 379 | :"+r" (i), "+m" (v->counter) | 366 | : : "memory"); |
| 380 | : : "memory"); | ||
| 381 | return i + __i; | 367 | return i + __i; |
| 382 | } | 368 | } |
| 383 | 369 | ||
| 384 | static __inline__ long atomic64_sub_return(long i, atomic64_t *v) | 370 | static inline long atomic64_sub_return(long i, atomic64_t *v) |
| 385 | { | 371 | { |
| 386 | return atomic64_add_return(-i,v); | 372 | return atomic64_add_return(-i, v); |
| 387 | } | 373 | } |
| 388 | 374 | ||
| 389 | #define atomic64_inc_return(v) (atomic64_add_return(1,v)) | 375 | #define atomic64_inc_return(v) (atomic64_add_return(1, (v))) |
| 390 | #define atomic64_dec_return(v) (atomic64_sub_return(1,v)) | 376 | #define atomic64_dec_return(v) (atomic64_sub_return(1, (v))) |
| 391 | 377 | ||
| 392 | #define atomic64_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), old, new)) | 378 | #define atomic64_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), (old), (new))) |
| 393 | #define atomic64_xchg(v, new) (xchg(&((v)->counter), new)) | 379 | #define atomic64_xchg(v, new) (xchg(&((v)->counter), new)) |
| 394 | 380 | ||
| 395 | #define atomic_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), old, new)) | 381 | #define atomic_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), (old), (new))) |
| 396 | #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) | 382 | #define atomic_xchg(v, new) (xchg(&((v)->counter), (new))) |
| 397 | 383 | ||
| 398 | /** | 384 | /** |
| 399 | * atomic_add_unless - add unless the number is a given value | 385 | * atomic_add_unless - add unless the number is a given value |
| @@ -404,7 +390,7 @@ static __inline__ long atomic64_sub_return(long i, atomic64_t *v) | |||
| 404 | * Atomically adds @a to @v, so long as it was not @u. | 390 | * Atomically adds @a to @v, so long as it was not @u. |
| 405 | * Returns non-zero if @v was not @u, and zero otherwise. | 391 | * Returns non-zero if @v was not @u, and zero otherwise. |
| 406 | */ | 392 | */ |
| 407 | static __inline__ int atomic_add_unless(atomic_t *v, int a, int u) | 393 | static inline int atomic_add_unless(atomic_t *v, int a, int u) |
| 408 | { | 394 | { |
| 409 | int c, old; | 395 | int c, old; |
| 410 | c = atomic_read(v); | 396 | c = atomic_read(v); |
| @@ -430,7 +416,7 @@ static __inline__ int atomic_add_unless(atomic_t *v, int a, int u) | |||
| 430 | * Atomically adds @a to @v, so long as it was not @u. | 416 | * Atomically adds @a to @v, so long as it was not @u. |
| 431 | * Returns non-zero if @v was not @u, and zero otherwise. | 417 | * Returns non-zero if @v was not @u, and zero otherwise. |
| 432 | */ | 418 | */ |
| 433 | static __inline__ int atomic64_add_unless(atomic64_t *v, long a, long u) | 419 | static inline int atomic64_add_unless(atomic64_t *v, long a, long u) |
| 434 | { | 420 | { |
| 435 | long c, old; | 421 | long c, old; |
| 436 | c = atomic64_read(v); | 422 | c = atomic64_read(v); |
| @@ -448,13 +434,14 @@ static __inline__ int atomic64_add_unless(atomic64_t *v, long a, long u) | |||
| 448 | #define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1, 0) | 434 | #define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1, 0) |
| 449 | 435 | ||
| 450 | /* These are x86-specific, used by some header files */ | 436 | /* These are x86-specific, used by some header files */ |
| 451 | #define atomic_clear_mask(mask, addr) \ | 437 | #define atomic_clear_mask(mask, addr) \ |
| 452 | __asm__ __volatile__(LOCK_PREFIX "andl %0,%1" \ | 438 | asm volatile(LOCK_PREFIX "andl %0,%1" \ |
| 453 | : : "r" (~(mask)),"m" (*addr) : "memory") | 439 | : : "r" (~(mask)), "m" (*(addr)) : "memory") |
| 454 | 440 | ||
| 455 | #define atomic_set_mask(mask, addr) \ | 441 | #define atomic_set_mask(mask, addr) \ |
| 456 | __asm__ __volatile__(LOCK_PREFIX "orl %0,%1" \ | 442 | asm volatile(LOCK_PREFIX "orl %0,%1" \ |
| 457 | : : "r" ((unsigned)mask),"m" (*(addr)) : "memory") | 443 | : : "r" ((unsigned)(mask)), "m" (*(addr)) \ |
| 444 | : "memory") | ||
| 458 | 445 | ||
| 459 | /* Atomic operations are already serializing on x86 */ | 446 | /* Atomic operations are already serializing on x86 */ |
| 460 | #define smp_mb__before_atomic_dec() barrier() | 447 | #define smp_mb__before_atomic_dec() barrier() |
diff --git a/include/asm-x86/mach-default/bios_ebda.h b/include/asm-x86/bios_ebda.h index 9cbd9a668af8..9cbd9a668af8 100644 --- a/include/asm-x86/mach-default/bios_ebda.h +++ b/include/asm-x86/bios_ebda.h | |||
diff --git a/include/asm-x86/bitops.h b/include/asm-x86/bitops.h index 1a23ce1a5697..1ae7b270a1ef 100644 --- a/include/asm-x86/bitops.h +++ b/include/asm-x86/bitops.h | |||
| @@ -23,10 +23,13 @@ | |||
| 23 | #if __GNUC__ < 4 || (__GNUC__ == 4 && __GNUC_MINOR__ < 1) | 23 | #if __GNUC__ < 4 || (__GNUC__ == 4 && __GNUC_MINOR__ < 1) |
| 24 | /* Technically wrong, but this avoids compilation errors on some gcc | 24 | /* Technically wrong, but this avoids compilation errors on some gcc |
| 25 | versions. */ | 25 | versions. */ |
| 26 | #define ADDR "=m" (*(volatile long *) addr) | 26 | #define ADDR "=m" (*(volatile long *)addr) |
| 27 | #define BIT_ADDR "=m" (((volatile int *)addr)[nr >> 5]) | ||
| 27 | #else | 28 | #else |
| 28 | #define ADDR "+m" (*(volatile long *) addr) | 29 | #define ADDR "+m" (*(volatile long *) addr) |
| 30 | #define BIT_ADDR "+m" (((volatile int *)addr)[nr >> 5]) | ||
| 29 | #endif | 31 | #endif |
| 32 | #define BASE_ADDR "m" (*(volatile int *)addr) | ||
| 30 | 33 | ||
| 31 | /** | 34 | /** |
| 32 | * set_bit - Atomically set a bit in memory | 35 | * set_bit - Atomically set a bit in memory |
| @@ -45,9 +48,7 @@ | |||
| 45 | */ | 48 | */ |
| 46 | static inline void set_bit(int nr, volatile void *addr) | 49 | static inline void set_bit(int nr, volatile void *addr) |
| 47 | { | 50 | { |
| 48 | asm volatile(LOCK_PREFIX "bts %1,%0" | 51 | asm volatile(LOCK_PREFIX "bts %1,%0" : ADDR : "Ir" (nr) : "memory"); |
| 49 | : ADDR | ||
| 50 | : "Ir" (nr) : "memory"); | ||
| 51 | } | 52 | } |
| 52 | 53 | ||
| 53 | /** | 54 | /** |
| @@ -79,9 +80,7 @@ static inline void __set_bit(int nr, volatile void *addr) | |||
| 79 | */ | 80 | */ |
| 80 | static inline void clear_bit(int nr, volatile void *addr) | 81 | static inline void clear_bit(int nr, volatile void *addr) |
| 81 | { | 82 | { |
| 82 | asm volatile(LOCK_PREFIX "btr %1,%0" | 83 | asm volatile(LOCK_PREFIX "btr %1,%2" : BIT_ADDR : "Ir" (nr), BASE_ADDR); |
| 83 | : ADDR | ||
| 84 | : "Ir" (nr)); | ||
| 85 | } | 84 | } |
| 86 | 85 | ||
| 87 | /* | 86 | /* |
| @@ -100,7 +99,7 @@ static inline void clear_bit_unlock(unsigned nr, volatile void *addr) | |||
| 100 | 99 | ||
| 101 | static inline void __clear_bit(int nr, volatile void *addr) | 100 | static inline void __clear_bit(int nr, volatile void *addr) |
| 102 | { | 101 | { |
| 103 | asm volatile("btr %1,%0" : ADDR : "Ir" (nr)); | 102 | asm volatile("btr %1,%2" : BIT_ADDR : "Ir" (nr), BASE_ADDR); |
| 104 | } | 103 | } |
| 105 | 104 | ||
| 106 | /* | 105 | /* |
| @@ -135,7 +134,7 @@ static inline void __clear_bit_unlock(unsigned nr, volatile void *addr) | |||
| 135 | */ | 134 | */ |
| 136 | static inline void __change_bit(int nr, volatile void *addr) | 135 | static inline void __change_bit(int nr, volatile void *addr) |
| 137 | { | 136 | { |
| 138 | asm volatile("btc %1,%0" : ADDR : "Ir" (nr)); | 137 | asm volatile("btc %1,%2" : BIT_ADDR : "Ir" (nr), BASE_ADDR); |
| 139 | } | 138 | } |
| 140 | 139 | ||
| 141 | /** | 140 | /** |
| @@ -149,8 +148,7 @@ static inline void __change_bit(int nr, volatile void *addr) | |||
| 149 | */ | 148 | */ |
| 150 | static inline void change_bit(int nr, volatile void *addr) | 149 | static inline void change_bit(int nr, volatile void *addr) |
| 151 | { | 150 | { |
| 152 | asm volatile(LOCK_PREFIX "btc %1,%0" | 151 | asm volatile(LOCK_PREFIX "btc %1,%2" : BIT_ADDR : "Ir" (nr), BASE_ADDR); |
| 153 | : ADDR : "Ir" (nr)); | ||
| 154 | } | 152 | } |
| 155 | 153 | ||
| 156 | /** | 154 | /** |
| @@ -166,9 +164,7 @@ static inline int test_and_set_bit(int nr, volatile void *addr) | |||
| 166 | int oldbit; | 164 | int oldbit; |
| 167 | 165 | ||
| 168 | asm volatile(LOCK_PREFIX "bts %2,%1\n\t" | 166 | asm volatile(LOCK_PREFIX "bts %2,%1\n\t" |
| 169 | "sbb %0,%0" | 167 | "sbb %0,%0" : "=r" (oldbit), ADDR : "Ir" (nr) : "memory"); |
| 170 | : "=r" (oldbit), ADDR | ||
| 171 | : "Ir" (nr) : "memory"); | ||
| 172 | 168 | ||
| 173 | return oldbit; | 169 | return oldbit; |
| 174 | } | 170 | } |
| @@ -198,10 +194,9 @@ static inline int __test_and_set_bit(int nr, volatile void *addr) | |||
| 198 | { | 194 | { |
| 199 | int oldbit; | 195 | int oldbit; |
| 200 | 196 | ||
| 201 | asm("bts %2,%1\n\t" | 197 | asm volatile("bts %2,%3\n\t" |
| 202 | "sbb %0,%0" | 198 | "sbb %0,%0" |
| 203 | : "=r" (oldbit), ADDR | 199 | : "=r" (oldbit), BIT_ADDR : "Ir" (nr), BASE_ADDR); |
| 204 | : "Ir" (nr)); | ||
| 205 | return oldbit; | 200 | return oldbit; |
| 206 | } | 201 | } |
| 207 | 202 | ||
| @@ -219,8 +214,7 @@ static inline int test_and_clear_bit(int nr, volatile void *addr) | |||
| 219 | 214 | ||
| 220 | asm volatile(LOCK_PREFIX "btr %2,%1\n\t" | 215 | asm volatile(LOCK_PREFIX "btr %2,%1\n\t" |
| 221 | "sbb %0,%0" | 216 | "sbb %0,%0" |
| 222 | : "=r" (oldbit), ADDR | 217 | : "=r" (oldbit), ADDR : "Ir" (nr) : "memory"); |
| 223 | : "Ir" (nr) : "memory"); | ||
| 224 | 218 | ||
| 225 | return oldbit; | 219 | return oldbit; |
| 226 | } | 220 | } |
| @@ -238,10 +232,9 @@ static inline int __test_and_clear_bit(int nr, volatile void *addr) | |||
| 238 | { | 232 | { |
| 239 | int oldbit; | 233 | int oldbit; |
| 240 | 234 | ||
| 241 | asm volatile("btr %2,%1\n\t" | 235 | asm volatile("btr %2,%3\n\t" |
| 242 | "sbb %0,%0" | 236 | "sbb %0,%0" |
| 243 | : "=r" (oldbit), ADDR | 237 | : "=r" (oldbit), BIT_ADDR : "Ir" (nr), BASE_ADDR); |
| 244 | : "Ir" (nr)); | ||
| 245 | return oldbit; | 238 | return oldbit; |
| 246 | } | 239 | } |
| 247 | 240 | ||
| @@ -250,10 +243,9 @@ static inline int __test_and_change_bit(int nr, volatile void *addr) | |||
| 250 | { | 243 | { |
| 251 | int oldbit; | 244 | int oldbit; |
| 252 | 245 | ||
| 253 | asm volatile("btc %2,%1\n\t" | 246 | asm volatile("btc %2,%3\n\t" |
| 254 | "sbb %0,%0" | 247 | "sbb %0,%0" |
| 255 | : "=r" (oldbit), ADDR | 248 | : "=r" (oldbit), BIT_ADDR : "Ir" (nr), BASE_ADDR); |
| 256 | : "Ir" (nr) : "memory"); | ||
| 257 | 249 | ||
| 258 | return oldbit; | 250 | return oldbit; |
| 259 | } | 251 | } |
| @@ -272,8 +264,7 @@ static inline int test_and_change_bit(int nr, volatile void *addr) | |||
| 272 | 264 | ||
| 273 | asm volatile(LOCK_PREFIX "btc %2,%1\n\t" | 265 | asm volatile(LOCK_PREFIX "btc %2,%1\n\t" |
| 274 | "sbb %0,%0" | 266 | "sbb %0,%0" |
| 275 | : "=r" (oldbit), ADDR | 267 | : "=r" (oldbit), ADDR : "Ir" (nr) : "memory"); |
| 276 | : "Ir" (nr) : "memory"); | ||
| 277 | 268 | ||
| 278 | return oldbit; | 269 | return oldbit; |
| 279 | } | 270 | } |
| @@ -288,10 +279,11 @@ static inline int variable_test_bit(int nr, volatile const void *addr) | |||
| 288 | { | 279 | { |
| 289 | int oldbit; | 280 | int oldbit; |
| 290 | 281 | ||
| 291 | asm volatile("bt %2,%1\n\t" | 282 | asm volatile("bt %2,%3\n\t" |
| 292 | "sbb %0,%0" | 283 | "sbb %0,%0" |
| 293 | : "=r" (oldbit) | 284 | : "=r" (oldbit) |
| 294 | : "m" (*(unsigned long *)addr), "Ir" (nr)); | 285 | : "m" (((volatile const int *)addr)[nr >> 5]), |
| 286 | "Ir" (nr), BASE_ADDR); | ||
| 295 | 287 | ||
| 296 | return oldbit; | 288 | return oldbit; |
| 297 | } | 289 | } |
| @@ -310,6 +302,8 @@ static int test_bit(int nr, const volatile unsigned long *addr); | |||
| 310 | constant_test_bit((nr),(addr)) : \ | 302 | constant_test_bit((nr),(addr)) : \ |
| 311 | variable_test_bit((nr),(addr))) | 303 | variable_test_bit((nr),(addr))) |
| 312 | 304 | ||
| 305 | #undef BASE_ADDR | ||
| 306 | #undef BIT_ADDR | ||
| 313 | #undef ADDR | 307 | #undef ADDR |
| 314 | 308 | ||
| 315 | #ifdef CONFIG_X86_32 | 309 | #ifdef CONFIG_X86_32 |
diff --git a/include/asm-x86/bitops_32.h b/include/asm-x86/bitops_32.h index e4d75fcf9c03..2513a81f82aa 100644 --- a/include/asm-x86/bitops_32.h +++ b/include/asm-x86/bitops_32.h | |||
| @@ -20,20 +20,22 @@ static inline int find_first_zero_bit(const unsigned long *addr, unsigned size) | |||
| 20 | 20 | ||
| 21 | if (!size) | 21 | if (!size) |
| 22 | return 0; | 22 | return 0; |
| 23 | /* This looks at memory. Mark it volatile to tell gcc not to move it around */ | 23 | /* This looks at memory. |
| 24 | __asm__ __volatile__( | 24 | * Mark it volatile to tell gcc not to move it around |
| 25 | "movl $-1,%%eax\n\t" | 25 | */ |
| 26 | "xorl %%edx,%%edx\n\t" | 26 | asm volatile("movl $-1,%%eax\n\t" |
| 27 | "repe; scasl\n\t" | 27 | "xorl %%edx,%%edx\n\t" |
| 28 | "je 1f\n\t" | 28 | "repe; scasl\n\t" |
| 29 | "xorl -4(%%edi),%%eax\n\t" | 29 | "je 1f\n\t" |
| 30 | "subl $4,%%edi\n\t" | 30 | "xorl -4(%%edi),%%eax\n\t" |
| 31 | "bsfl %%eax,%%edx\n" | 31 | "subl $4,%%edi\n\t" |
| 32 | "1:\tsubl %%ebx,%%edi\n\t" | 32 | "bsfl %%eax,%%edx\n" |
| 33 | "shll $3,%%edi\n\t" | 33 | "1:\tsubl %%ebx,%%edi\n\t" |
| 34 | "addl %%edi,%%edx" | 34 | "shll $3,%%edi\n\t" |
| 35 | :"=d" (res), "=&c" (d0), "=&D" (d1), "=&a" (d2) | 35 | "addl %%edi,%%edx" |
| 36 | :"1" ((size + 31) >> 5), "2" (addr), "b" (addr) : "memory"); | 36 | : "=d" (res), "=&c" (d0), "=&D" (d1), "=&a" (d2) |
| 37 | : "1" ((size + 31) >> 5), "2" (addr), | ||
| 38 | "b" (addr) : "memory"); | ||
| 37 | return res; | 39 | return res; |
| 38 | } | 40 | } |
| 39 | 41 | ||
| @@ -75,7 +77,7 @@ static inline unsigned find_first_bit(const unsigned long *addr, unsigned size) | |||
| 75 | unsigned long val = *addr++; | 77 | unsigned long val = *addr++; |
| 76 | if (val) | 78 | if (val) |
| 77 | return __ffs(val) + x; | 79 | return __ffs(val) + x; |
| 78 | x += (sizeof(*addr)<<3); | 80 | x += sizeof(*addr) << 3; |
| 79 | } | 81 | } |
| 80 | return x; | 82 | return x; |
| 81 | } | 83 | } |
| @@ -152,10 +154,10 @@ static inline int fls(int x) | |||
| 152 | 154 | ||
| 153 | #include <asm-generic/bitops/ext2-non-atomic.h> | 155 | #include <asm-generic/bitops/ext2-non-atomic.h> |
| 154 | 156 | ||
| 155 | #define ext2_set_bit_atomic(lock, nr, addr) \ | 157 | #define ext2_set_bit_atomic(lock, nr, addr) \ |
| 156 | test_and_set_bit((nr), (unsigned long *)addr) | 158 | test_and_set_bit((nr), (unsigned long *)(addr)) |
| 157 | #define ext2_clear_bit_atomic(lock, nr, addr) \ | 159 | #define ext2_clear_bit_atomic(lock, nr, addr) \ |
| 158 | test_and_clear_bit((nr), (unsigned long *)addr) | 160 | test_and_clear_bit((nr), (unsigned long *)(addr)) |
| 159 | 161 | ||
| 160 | #include <asm-generic/bitops/minix.h> | 162 | #include <asm-generic/bitops/minix.h> |
| 161 | 163 | ||
diff --git a/include/asm-x86/bitops_64.h b/include/asm-x86/bitops_64.h index aaf15194d536..365f8207ea59 100644 --- a/include/asm-x86/bitops_64.h +++ b/include/asm-x86/bitops_64.h | |||
| @@ -17,35 +17,35 @@ static inline long __scanbit(unsigned long val, unsigned long max) | |||
| 17 | return val; | 17 | return val; |
| 18 | } | 18 | } |
| 19 | 19 | ||
| 20 | #define find_first_bit(addr,size) \ | ||
| 21 | ((__builtin_constant_p(size) && (size) <= BITS_PER_LONG ? \ | ||
| 22 | (__scanbit(*(unsigned long *)addr,(size))) : \ | ||
| 23 | find_first_bit(addr,size))) | ||
| 24 | |||
| 25 | #define find_next_bit(addr,size,off) \ | 20 | #define find_next_bit(addr,size,off) \ |
| 26 | ((__builtin_constant_p(size) && (size) <= BITS_PER_LONG ? \ | 21 | ((__builtin_constant_p(size) && (size) <= BITS_PER_LONG ? \ |
| 27 | ((off) + (__scanbit((*(unsigned long *)addr) >> (off),(size)-(off)))) : \ | 22 | ((off) + (__scanbit((*(unsigned long *)addr) >> (off),(size)-(off)))) : \ |
| 28 | find_next_bit(addr,size,off))) | 23 | find_next_bit(addr,size,off))) |
| 29 | 24 | ||
| 30 | #define find_first_zero_bit(addr,size) \ | ||
| 31 | ((__builtin_constant_p(size) && (size) <= BITS_PER_LONG ? \ | ||
| 32 | (__scanbit(~*(unsigned long *)addr,(size))) : \ | ||
| 33 | find_first_zero_bit(addr,size))) | ||
| 34 | |||
| 35 | #define find_next_zero_bit(addr,size,off) \ | 25 | #define find_next_zero_bit(addr,size,off) \ |
| 36 | ((__builtin_constant_p(size) && (size) <= BITS_PER_LONG ? \ | 26 | ((__builtin_constant_p(size) && (size) <= BITS_PER_LONG ? \ |
| 37 | ((off)+(__scanbit(~(((*(unsigned long *)addr)) >> (off)),(size)-(off)))) : \ | 27 | ((off)+(__scanbit(~(((*(unsigned long *)addr)) >> (off)),(size)-(off)))) : \ |
| 38 | find_next_zero_bit(addr,size,off))) | 28 | find_next_zero_bit(addr,size,off))) |
| 39 | 29 | ||
| 40 | static inline void set_bit_string(unsigned long *bitmap, unsigned long i, | 30 | #define find_first_bit(addr, size) \ |
| 41 | int len) | 31 | ((__builtin_constant_p((size)) && (size) <= BITS_PER_LONG \ |
| 42 | { | 32 | ? (__scanbit(*(unsigned long *)(addr), (size))) \ |
| 43 | unsigned long end = i + len; | 33 | : find_first_bit((addr), (size)))) |
| 34 | |||
| 35 | #define find_first_zero_bit(addr, size) \ | ||
| 36 | ((__builtin_constant_p((size)) && (size) <= BITS_PER_LONG \ | ||
| 37 | ? (__scanbit(~*(unsigned long *)(addr), (size))) \ | ||
| 38 | : find_first_zero_bit((addr), (size)))) | ||
| 39 | |||
| 40 | static inline void set_bit_string(unsigned long *bitmap, unsigned long i, | ||
| 41 | int len) | ||
| 42 | { | ||
| 43 | unsigned long end = i + len; | ||
| 44 | while (i < end) { | 44 | while (i < end) { |
| 45 | __set_bit(i, bitmap); | 45 | __set_bit(i, bitmap); |
| 46 | i++; | 46 | i++; |
| 47 | } | 47 | } |
| 48 | } | 48 | } |
| 49 | 49 | ||
| 50 | /** | 50 | /** |
| 51 | * ffz - find first zero in word. | 51 | * ffz - find first zero in word. |
| @@ -150,10 +150,10 @@ static inline int fls(int x) | |||
| 150 | 150 | ||
| 151 | #include <asm-generic/bitops/ext2-non-atomic.h> | 151 | #include <asm-generic/bitops/ext2-non-atomic.h> |
| 152 | 152 | ||
| 153 | #define ext2_set_bit_atomic(lock,nr,addr) \ | 153 | #define ext2_set_bit_atomic(lock, nr, addr) \ |
| 154 | test_and_set_bit((nr),(unsigned long*)addr) | 154 | test_and_set_bit((nr), (unsigned long *)(addr)) |
| 155 | #define ext2_clear_bit_atomic(lock,nr,addr) \ | 155 | #define ext2_clear_bit_atomic(lock, nr, addr) \ |
| 156 | test_and_clear_bit((nr),(unsigned long*)addr) | 156 | test_and_clear_bit((nr), (unsigned long *)(addr)) |
| 157 | 157 | ||
| 158 | #include <asm-generic/bitops/minix.h> | 158 | #include <asm-generic/bitops/minix.h> |
| 159 | 159 | ||
diff --git a/include/asm-x86/bug.h b/include/asm-x86/bug.h index 8d477a201392..b69aa64b82a4 100644 --- a/include/asm-x86/bug.h +++ b/include/asm-x86/bug.h | |||
| @@ -12,25 +12,25 @@ | |||
| 12 | # define __BUG_C0 "2:\t.quad 1b, %c0\n" | 12 | # define __BUG_C0 "2:\t.quad 1b, %c0\n" |
| 13 | #endif | 13 | #endif |
| 14 | 14 | ||
| 15 | #define BUG() \ | 15 | #define BUG() \ |
| 16 | do { \ | 16 | do { \ |
| 17 | asm volatile("1:\tud2\n" \ | 17 | asm volatile("1:\tud2\n" \ |
| 18 | ".pushsection __bug_table,\"a\"\n" \ | 18 | ".pushsection __bug_table,\"a\"\n" \ |
| 19 | __BUG_C0 \ | 19 | __BUG_C0 \ |
| 20 | "\t.word %c1, 0\n" \ | 20 | "\t.word %c1, 0\n" \ |
| 21 | "\t.org 2b+%c2\n" \ | 21 | "\t.org 2b+%c2\n" \ |
| 22 | ".popsection" \ | 22 | ".popsection" \ |
| 23 | : : "i" (__FILE__), "i" (__LINE__), \ | 23 | : : "i" (__FILE__), "i" (__LINE__), \ |
| 24 | "i" (sizeof(struct bug_entry))); \ | 24 | "i" (sizeof(struct bug_entry))); \ |
| 25 | for(;;) ; \ | 25 | for (;;) ; \ |
| 26 | } while(0) | 26 | } while (0) |
| 27 | 27 | ||
| 28 | #else | 28 | #else |
| 29 | #define BUG() \ | 29 | #define BUG() \ |
| 30 | do { \ | 30 | do { \ |
| 31 | asm volatile("ud2"); \ | 31 | asm volatile("ud2"); \ |
| 32 | for(;;) ; \ | 32 | for (;;) ; \ |
| 33 | } while(0) | 33 | } while (0) |
| 34 | #endif | 34 | #endif |
| 35 | 35 | ||
| 36 | #endif /* !CONFIG_BUG */ | 36 | #endif /* !CONFIG_BUG */ |
diff --git a/include/asm-x86/byteorder.h b/include/asm-x86/byteorder.h index fe2f2e5d51ba..e02ae2d89acf 100644 --- a/include/asm-x86/byteorder.h +++ b/include/asm-x86/byteorder.h | |||
| @@ -8,50 +8,59 @@ | |||
| 8 | 8 | ||
| 9 | #ifdef __i386__ | 9 | #ifdef __i386__ |
| 10 | 10 | ||
| 11 | static __inline__ __attribute_const__ __u32 ___arch__swab32(__u32 x) | 11 | static inline __attribute_const__ __u32 ___arch__swab32(__u32 x) |
| 12 | { | 12 | { |
| 13 | #ifdef CONFIG_X86_BSWAP | 13 | #ifdef CONFIG_X86_BSWAP |
| 14 | __asm__("bswap %0" : "=r" (x) : "0" (x)); | 14 | asm("bswap %0" : "=r" (x) : "0" (x)); |
| 15 | #else | 15 | #else |
| 16 | __asm__("xchgb %b0,%h0\n\t" /* swap lower bytes */ | 16 | asm("xchgb %b0,%h0\n\t" /* swap lower bytes */ |
| 17 | "rorl $16,%0\n\t" /* swap words */ | 17 | "rorl $16,%0\n\t" /* swap words */ |
| 18 | "xchgb %b0,%h0" /* swap higher bytes */ | 18 | "xchgb %b0,%h0" /* swap higher bytes */ |
| 19 | :"=q" (x) | 19 | : "=q" (x) |
| 20 | : "0" (x)); | 20 | : "0" (x)); |
| 21 | #endif | 21 | #endif |
| 22 | return x; | 22 | return x; |
| 23 | } | 23 | } |
| 24 | 24 | ||
| 25 | static __inline__ __attribute_const__ __u64 ___arch__swab64(__u64 val) | 25 | static inline __attribute_const__ __u64 ___arch__swab64(__u64 val) |
| 26 | { | 26 | { |
| 27 | union { | 27 | union { |
| 28 | struct { __u32 a,b; } s; | 28 | struct { |
| 29 | __u32 a; | ||
| 30 | __u32 b; | ||
| 31 | } s; | ||
| 29 | __u64 u; | 32 | __u64 u; |
| 30 | } v; | 33 | } v; |
| 31 | v.u = val; | 34 | v.u = val; |
| 32 | #ifdef CONFIG_X86_BSWAP | 35 | #ifdef CONFIG_X86_BSWAP |
| 33 | __asm__("bswapl %0 ; bswapl %1 ; xchgl %0,%1" | 36 | asm("bswapl %0 ; bswapl %1 ; xchgl %0,%1" |
| 34 | : "=r" (v.s.a), "=r" (v.s.b) | 37 | : "=r" (v.s.a), "=r" (v.s.b) |
| 35 | : "0" (v.s.a), "1" (v.s.b)); | 38 | : "0" (v.s.a), "1" (v.s.b)); |
| 36 | #else | 39 | #else |
| 37 | v.s.a = ___arch__swab32(v.s.a); | 40 | v.s.a = ___arch__swab32(v.s.a); |
| 38 | v.s.b = ___arch__swab32(v.s.b); | 41 | v.s.b = ___arch__swab32(v.s.b); |
| 39 | __asm__("xchgl %0,%1" : "=r" (v.s.a), "=r" (v.s.b) : "0" (v.s.a), "1" (v.s.b)); | 42 | asm("xchgl %0,%1" |
| 43 | : "=r" (v.s.a), "=r" (v.s.b) | ||
| 44 | : "0" (v.s.a), "1" (v.s.b)); | ||
| 40 | #endif | 45 | #endif |
| 41 | return v.u; | 46 | return v.u; |
| 42 | } | 47 | } |
| 43 | 48 | ||
| 44 | #else /* __i386__ */ | 49 | #else /* __i386__ */ |
| 45 | 50 | ||
| 46 | static __inline__ __attribute_const__ __u64 ___arch__swab64(__u64 x) | 51 | static inline __attribute_const__ __u64 ___arch__swab64(__u64 x) |
| 47 | { | 52 | { |
| 48 | __asm__("bswapq %0" : "=r" (x) : "0" (x)); | 53 | asm("bswapq %0" |
| 54 | : "=r" (x) | ||
| 55 | : "0" (x)); | ||
| 49 | return x; | 56 | return x; |
| 50 | } | 57 | } |
| 51 | 58 | ||
| 52 | static __inline__ __attribute_const__ __u32 ___arch__swab32(__u32 x) | 59 | static inline __attribute_const__ __u32 ___arch__swab32(__u32 x) |
| 53 | { | 60 | { |
| 54 | __asm__("bswapl %0" : "=r" (x) : "0" (x)); | 61 | asm("bswapl %0" |
| 62 | : "=r" (x) | ||
| 63 | : "0" (x)); | ||
| 55 | return x; | 64 | return x; |
| 56 | } | 65 | } |
| 57 | 66 | ||
diff --git a/include/asm-x86/cacheflush.h b/include/asm-x86/cacheflush.h index 5396c212d8c0..f4c0ab50d2c2 100644 --- a/include/asm-x86/cacheflush.h +++ b/include/asm-x86/cacheflush.h | |||
| @@ -14,33 +14,85 @@ | |||
| 14 | #define flush_dcache_mmap_lock(mapping) do { } while (0) | 14 | #define flush_dcache_mmap_lock(mapping) do { } while (0) |
| 15 | #define flush_dcache_mmap_unlock(mapping) do { } while (0) | 15 | #define flush_dcache_mmap_unlock(mapping) do { } while (0) |
| 16 | #define flush_icache_range(start, end) do { } while (0) | 16 | #define flush_icache_range(start, end) do { } while (0) |
| 17 | #define flush_icache_page(vma,pg) do { } while (0) | 17 | #define flush_icache_page(vma, pg) do { } while (0) |
| 18 | #define flush_icache_user_range(vma,pg,adr,len) do { } while (0) | 18 | #define flush_icache_user_range(vma, pg, adr, len) do { } while (0) |
| 19 | #define flush_cache_vmap(start, end) do { } while (0) | 19 | #define flush_cache_vmap(start, end) do { } while (0) |
| 20 | #define flush_cache_vunmap(start, end) do { } while (0) | 20 | #define flush_cache_vunmap(start, end) do { } while (0) |
| 21 | 21 | ||
| 22 | #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ | 22 | #define copy_to_user_page(vma, page, vaddr, dst, src, len) \ |
| 23 | memcpy(dst, src, len) | 23 | memcpy((dst), (src), (len)) |
| 24 | #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ | 24 | #define copy_from_user_page(vma, page, vaddr, dst, src, len) \ |
| 25 | memcpy(dst, src, len) | 25 | memcpy((dst), (src), (len)) |
| 26 | 26 | ||
| 27 | int __deprecated_for_modules change_page_attr(struct page *page, int numpages, | ||
| 28 | pgprot_t prot); | ||
| 29 | 27 | ||
| 30 | int set_pages_uc(struct page *page, int numpages); | 28 | /* |
| 31 | int set_pages_wb(struct page *page, int numpages); | 29 | * The set_memory_* API can be used to change various attributes of a virtual |
| 32 | int set_pages_x(struct page *page, int numpages); | 30 | * address range. The attributes include: |
| 33 | int set_pages_nx(struct page *page, int numpages); | 31 | * Cachability : UnCached, WriteCombining, WriteBack |
| 34 | int set_pages_ro(struct page *page, int numpages); | 32 | * Executability : eXeutable, NoteXecutable |
| 35 | int set_pages_rw(struct page *page, int numpages); | 33 | * Read/Write : ReadOnly, ReadWrite |
| 34 | * Presence : NotPresent | ||
| 35 | * | ||
| 36 | * Within a catagory, the attributes are mutually exclusive. | ||
| 37 | * | ||
| 38 | * The implementation of this API will take care of various aspects that | ||
| 39 | * are associated with changing such attributes, such as: | ||
| 40 | * - Flushing TLBs | ||
| 41 | * - Flushing CPU caches | ||
| 42 | * - Making sure aliases of the memory behind the mapping don't violate | ||
| 43 | * coherency rules as defined by the CPU in the system. | ||
| 44 | * | ||
| 45 | * What this API does not do: | ||
| 46 | * - Provide exclusion between various callers - including callers that | ||
| 47 | * operation on other mappings of the same physical page | ||
| 48 | * - Restore default attributes when a page is freed | ||
| 49 | * - Guarantee that mappings other than the requested one are | ||
| 50 | * in any state, other than that these do not violate rules for | ||
| 51 | * the CPU you have. Do not depend on any effects on other mappings, | ||
| 52 | * CPUs other than the one you have may have more relaxed rules. | ||
| 53 | * The caller is required to take care of these. | ||
| 54 | */ | ||
| 36 | 55 | ||
| 56 | int _set_memory_uc(unsigned long addr, int numpages); | ||
| 57 | int _set_memory_wc(unsigned long addr, int numpages); | ||
| 58 | int _set_memory_wb(unsigned long addr, int numpages); | ||
| 37 | int set_memory_uc(unsigned long addr, int numpages); | 59 | int set_memory_uc(unsigned long addr, int numpages); |
| 60 | int set_memory_wc(unsigned long addr, int numpages); | ||
| 38 | int set_memory_wb(unsigned long addr, int numpages); | 61 | int set_memory_wb(unsigned long addr, int numpages); |
| 39 | int set_memory_x(unsigned long addr, int numpages); | 62 | int set_memory_x(unsigned long addr, int numpages); |
| 40 | int set_memory_nx(unsigned long addr, int numpages); | 63 | int set_memory_nx(unsigned long addr, int numpages); |
| 41 | int set_memory_ro(unsigned long addr, int numpages); | 64 | int set_memory_ro(unsigned long addr, int numpages); |
| 42 | int set_memory_rw(unsigned long addr, int numpages); | 65 | int set_memory_rw(unsigned long addr, int numpages); |
| 43 | int set_memory_np(unsigned long addr, int numpages); | 66 | int set_memory_np(unsigned long addr, int numpages); |
| 67 | int set_memory_4k(unsigned long addr, int numpages); | ||
| 68 | |||
| 69 | /* | ||
| 70 | * For legacy compatibility with the old APIs, a few functions | ||
| 71 | * are provided that work on a "struct page". | ||
| 72 | * These functions operate ONLY on the 1:1 kernel mapping of the | ||
| 73 | * memory that the struct page represents, and internally just | ||
| 74 | * call the set_memory_* function. See the description of the | ||
| 75 | * set_memory_* function for more details on conventions. | ||
| 76 | * | ||
| 77 | * These APIs should be considered *deprecated* and are likely going to | ||
| 78 | * be removed in the future. | ||
| 79 | * The reason for this is the implicit operation on the 1:1 mapping only, | ||
| 80 | * making this not a generally useful API. | ||
| 81 | * | ||
| 82 | * Specifically, many users of the old APIs had a virtual address, | ||
| 83 | * called virt_to_page() or vmalloc_to_page() on that address to | ||
| 84 | * get a struct page* that the old API required. | ||
| 85 | * To convert these cases, use set_memory_*() on the original | ||
| 86 | * virtual address, do not use these functions. | ||
| 87 | */ | ||
| 88 | |||
| 89 | int set_pages_uc(struct page *page, int numpages); | ||
| 90 | int set_pages_wb(struct page *page, int numpages); | ||
| 91 | int set_pages_x(struct page *page, int numpages); | ||
| 92 | int set_pages_nx(struct page *page, int numpages); | ||
| 93 | int set_pages_ro(struct page *page, int numpages); | ||
| 94 | int set_pages_rw(struct page *page, int numpages); | ||
| 95 | |||
| 44 | 96 | ||
| 45 | void clflush_cache_range(void *addr, unsigned int size); | 97 | void clflush_cache_range(void *addr, unsigned int size); |
| 46 | 98 | ||
diff --git a/include/asm-x86/checksum_32.h b/include/asm-x86/checksum_32.h index 75194abbe8ee..52bbb0d8c4c1 100644 --- a/include/asm-x86/checksum_32.h +++ b/include/asm-x86/checksum_32.h | |||
| @@ -28,7 +28,8 @@ asmlinkage __wsum csum_partial(const void *buff, int len, __wsum sum); | |||
| 28 | */ | 28 | */ |
| 29 | 29 | ||
| 30 | asmlinkage __wsum csum_partial_copy_generic(const void *src, void *dst, | 30 | asmlinkage __wsum csum_partial_copy_generic(const void *src, void *dst, |
| 31 | int len, __wsum sum, int *src_err_ptr, int *dst_err_ptr); | 31 | int len, __wsum sum, |
| 32 | int *src_err_ptr, int *dst_err_ptr); | ||
| 32 | 33 | ||
| 33 | /* | 34 | /* |
| 34 | * Note: when you get a NULL pointer exception here this means someone | 35 | * Note: when you get a NULL pointer exception here this means someone |
| @@ -37,20 +38,20 @@ asmlinkage __wsum csum_partial_copy_generic(const void *src, void *dst, | |||
| 37 | * If you use these functions directly please don't forget the | 38 | * If you use these functions directly please don't forget the |
| 38 | * access_ok(). | 39 | * access_ok(). |
| 39 | */ | 40 | */ |
| 40 | static __inline__ | 41 | static inline __wsum csum_partial_copy_nocheck(const void *src, void *dst, |
| 41 | __wsum csum_partial_copy_nocheck (const void *src, void *dst, | 42 | int len, __wsum sum) |
| 42 | int len, __wsum sum) | ||
| 43 | { | 43 | { |
| 44 | return csum_partial_copy_generic ( src, dst, len, sum, NULL, NULL); | 44 | return csum_partial_copy_generic(src, dst, len, sum, NULL, NULL); |
| 45 | } | 45 | } |
| 46 | 46 | ||
| 47 | static __inline__ | 47 | static inline __wsum csum_partial_copy_from_user(const void __user *src, |
| 48 | __wsum csum_partial_copy_from_user(const void __user *src, void *dst, | 48 | void *dst, |
| 49 | int len, __wsum sum, int *err_ptr) | 49 | int len, __wsum sum, |
| 50 | int *err_ptr) | ||
| 50 | { | 51 | { |
| 51 | might_sleep(); | 52 | might_sleep(); |
| 52 | return csum_partial_copy_generic((__force void *)src, dst, | 53 | return csum_partial_copy_generic((__force void *)src, dst, |
| 53 | len, sum, err_ptr, NULL); | 54 | len, sum, err_ptr, NULL); |
| 54 | } | 55 | } |
| 55 | 56 | ||
| 56 | /* | 57 | /* |
| @@ -64,30 +65,29 @@ static inline __sum16 ip_fast_csum(const void *iph, unsigned int ihl) | |||
| 64 | { | 65 | { |
| 65 | unsigned int sum; | 66 | unsigned int sum; |
| 66 | 67 | ||
| 67 | __asm__ __volatile__( | 68 | asm volatile("movl (%1), %0 ;\n" |
| 68 | "movl (%1), %0 ;\n" | 69 | "subl $4, %2 ;\n" |
| 69 | "subl $4, %2 ;\n" | 70 | "jbe 2f ;\n" |
| 70 | "jbe 2f ;\n" | 71 | "addl 4(%1), %0 ;\n" |
| 71 | "addl 4(%1), %0 ;\n" | 72 | "adcl 8(%1), %0 ;\n" |
| 72 | "adcl 8(%1), %0 ;\n" | 73 | "adcl 12(%1), %0;\n" |
| 73 | "adcl 12(%1), %0 ;\n" | 74 | "1: adcl 16(%1), %0 ;\n" |
| 74 | "1: adcl 16(%1), %0 ;\n" | 75 | "lea 4(%1), %1 ;\n" |
| 75 | "lea 4(%1), %1 ;\n" | 76 | "decl %2 ;\n" |
| 76 | "decl %2 ;\n" | 77 | "jne 1b ;\n" |
| 77 | "jne 1b ;\n" | 78 | "adcl $0, %0 ;\n" |
| 78 | "adcl $0, %0 ;\n" | 79 | "movl %0, %2 ;\n" |
| 79 | "movl %0, %2 ;\n" | 80 | "shrl $16, %0 ;\n" |
| 80 | "shrl $16, %0 ;\n" | 81 | "addw %w2, %w0 ;\n" |
| 81 | "addw %w2, %w0 ;\n" | 82 | "adcl $0, %0 ;\n" |
| 82 | "adcl $0, %0 ;\n" | 83 | "notl %0 ;\n" |
| 83 | "notl %0 ;\n" | 84 | "2: ;\n" |
| 84 | "2: ;\n" | ||
| 85 | /* Since the input registers which are loaded with iph and ihl | 85 | /* Since the input registers which are loaded with iph and ihl |
| 86 | are modified, we must also specify them as outputs, or gcc | 86 | are modified, we must also specify them as outputs, or gcc |
| 87 | will assume they contain their original values. */ | 87 | will assume they contain their original values. */ |
| 88 | : "=r" (sum), "=r" (iph), "=r" (ihl) | 88 | : "=r" (sum), "=r" (iph), "=r" (ihl) |
| 89 | : "1" (iph), "2" (ihl) | 89 | : "1" (iph), "2" (ihl) |
| 90 | : "memory"); | 90 | : "memory"); |
| 91 | return (__force __sum16)sum; | 91 | return (__force __sum16)sum; |
| 92 | } | 92 | } |
| 93 | 93 | ||
| @@ -97,29 +97,27 @@ static inline __sum16 ip_fast_csum(const void *iph, unsigned int ihl) | |||
| 97 | 97 | ||
| 98 | static inline __sum16 csum_fold(__wsum sum) | 98 | static inline __sum16 csum_fold(__wsum sum) |
| 99 | { | 99 | { |
| 100 | __asm__( | 100 | asm("addl %1, %0 ;\n" |
| 101 | "addl %1, %0 ;\n" | 101 | "adcl $0xffff, %0 ;\n" |
| 102 | "adcl $0xffff, %0 ;\n" | 102 | : "=r" (sum) |
| 103 | : "=r" (sum) | 103 | : "r" ((__force u32)sum << 16), |
| 104 | : "r" ((__force u32)sum << 16), | 104 | "0" ((__force u32)sum & 0xffff0000)); |
| 105 | "0" ((__force u32)sum & 0xffff0000) | ||
| 106 | ); | ||
| 107 | return (__force __sum16)(~(__force u32)sum >> 16); | 105 | return (__force __sum16)(~(__force u32)sum >> 16); |
| 108 | } | 106 | } |
| 109 | 107 | ||
| 110 | static inline __wsum csum_tcpudp_nofold(__be32 saddr, __be32 daddr, | 108 | static inline __wsum csum_tcpudp_nofold(__be32 saddr, __be32 daddr, |
| 111 | unsigned short len, | 109 | unsigned short len, |
| 112 | unsigned short proto, | 110 | unsigned short proto, |
| 113 | __wsum sum) | 111 | __wsum sum) |
| 114 | { | 112 | { |
| 115 | __asm__( | 113 | asm("addl %1, %0 ;\n" |
| 116 | "addl %1, %0 ;\n" | 114 | "adcl %2, %0 ;\n" |
| 117 | "adcl %2, %0 ;\n" | 115 | "adcl %3, %0 ;\n" |
| 118 | "adcl %3, %0 ;\n" | 116 | "adcl $0, %0 ;\n" |
| 119 | "adcl $0, %0 ;\n" | 117 | : "=r" (sum) |
| 120 | : "=r" (sum) | 118 | : "g" (daddr), "g"(saddr), |
| 121 | : "g" (daddr), "g"(saddr), "g"((len + proto) << 8), "0"(sum)); | 119 | "g" ((len + proto) << 8), "0" (sum)); |
| 122 | return sum; | 120 | return sum; |
| 123 | } | 121 | } |
| 124 | 122 | ||
| 125 | /* | 123 | /* |
| @@ -127,11 +125,11 @@ static inline __wsum csum_tcpudp_nofold(__be32 saddr, __be32 daddr, | |||
| 127 | * returns a 16-bit checksum, already complemented | 125 | * returns a 16-bit checksum, already complemented |
| 128 | */ | 126 | */ |
| 129 | static inline __sum16 csum_tcpudp_magic(__be32 saddr, __be32 daddr, | 127 | static inline __sum16 csum_tcpudp_magic(__be32 saddr, __be32 daddr, |
| 130 | unsigned short len, | 128 | unsigned short len, |
| 131 | unsigned short proto, | 129 | unsigned short proto, |
| 132 | __wsum sum) | 130 | __wsum sum) |
| 133 | { | 131 | { |
| 134 | return csum_fold(csum_tcpudp_nofold(saddr,daddr,len,proto,sum)); | 132 | return csum_fold(csum_tcpudp_nofold(saddr, daddr, len, proto, sum)); |
| 135 | } | 133 | } |
| 136 | 134 | ||
| 137 | /* | 135 | /* |
| @@ -141,30 +139,29 @@ static inline __sum16 csum_tcpudp_magic(__be32 saddr, __be32 daddr, | |||
| 141 | 139 | ||
| 142 | static inline __sum16 ip_compute_csum(const void *buff, int len) | 140 | static inline __sum16 ip_compute_csum(const void *buff, int len) |
| 143 | { | 141 | { |
| 144 | return csum_fold (csum_partial(buff, len, 0)); | 142 | return csum_fold(csum_partial(buff, len, 0)); |
| 145 | } | 143 | } |
| 146 | 144 | ||
| 147 | #define _HAVE_ARCH_IPV6_CSUM | 145 | #define _HAVE_ARCH_IPV6_CSUM |
| 148 | static __inline__ __sum16 csum_ipv6_magic(const struct in6_addr *saddr, | 146 | static inline __sum16 csum_ipv6_magic(const struct in6_addr *saddr, |
| 149 | const struct in6_addr *daddr, | 147 | const struct in6_addr *daddr, |
| 150 | __u32 len, unsigned short proto, | 148 | __u32 len, unsigned short proto, |
| 151 | __wsum sum) | 149 | __wsum sum) |
| 152 | { | 150 | { |
| 153 | __asm__( | 151 | asm("addl 0(%1), %0 ;\n" |
| 154 | "addl 0(%1), %0 ;\n" | 152 | "adcl 4(%1), %0 ;\n" |
| 155 | "adcl 4(%1), %0 ;\n" | 153 | "adcl 8(%1), %0 ;\n" |
| 156 | "adcl 8(%1), %0 ;\n" | 154 | "adcl 12(%1), %0 ;\n" |
| 157 | "adcl 12(%1), %0 ;\n" | 155 | "adcl 0(%2), %0 ;\n" |
| 158 | "adcl 0(%2), %0 ;\n" | 156 | "adcl 4(%2), %0 ;\n" |
| 159 | "adcl 4(%2), %0 ;\n" | 157 | "adcl 8(%2), %0 ;\n" |
| 160 | "adcl 8(%2), %0 ;\n" | 158 | "adcl 12(%2), %0 ;\n" |
| 161 | "adcl 12(%2), %0 ;\n" | 159 | "adcl %3, %0 ;\n" |
| 162 | "adcl %3, %0 ;\n" | 160 | "adcl %4, %0 ;\n" |
| 163 | "adcl %4, %0 ;\n" | 161 | "adcl $0, %0 ;\n" |
| 164 | "adcl $0, %0 ;\n" | 162 | : "=&r" (sum) |
| 165 | : "=&r" (sum) | 163 | : "r" (saddr), "r" (daddr), |
| 166 | : "r" (saddr), "r" (daddr), | 164 | "r" (htonl(len)), "r" (htonl(proto)), "0" (sum)); |
| 167 | "r"(htonl(len)), "r"(htonl(proto)), "0"(sum)); | ||
| 168 | 165 | ||
| 169 | return csum_fold(sum); | 166 | return csum_fold(sum); |
| 170 | } | 167 | } |
| @@ -173,14 +170,15 @@ static __inline__ __sum16 csum_ipv6_magic(const struct in6_addr *saddr, | |||
| 173 | * Copy and checksum to user | 170 | * Copy and checksum to user |
| 174 | */ | 171 | */ |
| 175 | #define HAVE_CSUM_COPY_USER | 172 | #define HAVE_CSUM_COPY_USER |
| 176 | static __inline__ __wsum csum_and_copy_to_user(const void *src, | 173 | static inline __wsum csum_and_copy_to_user(const void *src, |
| 177 | void __user *dst, | 174 | void __user *dst, |
| 178 | int len, __wsum sum, | 175 | int len, __wsum sum, |
| 179 | int *err_ptr) | 176 | int *err_ptr) |
| 180 | { | 177 | { |
| 181 | might_sleep(); | 178 | might_sleep(); |
| 182 | if (access_ok(VERIFY_WRITE, dst, len)) | 179 | if (access_ok(VERIFY_WRITE, dst, len)) |
| 183 | return csum_partial_copy_generic(src, (__force void *)dst, len, sum, NULL, err_ptr); | 180 | return csum_partial_copy_generic(src, (__force void *)dst, |
| 181 | len, sum, NULL, err_ptr); | ||
| 184 | 182 | ||
| 185 | if (len) | 183 | if (len) |
| 186 | *err_ptr = -EFAULT; | 184 | *err_ptr = -EFAULT; |
diff --git a/include/asm-x86/checksum_64.h b/include/asm-x86/checksum_64.h index e5f79997decc..8bd861cc5267 100644 --- a/include/asm-x86/checksum_64.h +++ b/include/asm-x86/checksum_64.h | |||
| @@ -1,33 +1,31 @@ | |||
| 1 | #ifndef _X86_64_CHECKSUM_H | 1 | #ifndef _X86_64_CHECKSUM_H |
| 2 | #define _X86_64_CHECKSUM_H | 2 | #define _X86_64_CHECKSUM_H |
| 3 | 3 | ||
| 4 | /* | 4 | /* |
| 5 | * Checksums for x86-64 | 5 | * Checksums for x86-64 |
| 6 | * Copyright 2002 by Andi Kleen, SuSE Labs | 6 | * Copyright 2002 by Andi Kleen, SuSE Labs |
| 7 | * with some code from asm-x86/checksum.h | 7 | * with some code from asm-x86/checksum.h |
| 8 | */ | 8 | */ |
| 9 | 9 | ||
| 10 | #include <linux/compiler.h> | 10 | #include <linux/compiler.h> |
| 11 | #include <asm/uaccess.h> | 11 | #include <asm/uaccess.h> |
| 12 | #include <asm/byteorder.h> | 12 | #include <asm/byteorder.h> |
| 13 | 13 | ||
| 14 | /** | 14 | /** |
| 15 | * csum_fold - Fold and invert a 32bit checksum. | 15 | * csum_fold - Fold and invert a 32bit checksum. |
| 16 | * sum: 32bit unfolded sum | 16 | * sum: 32bit unfolded sum |
| 17 | * | 17 | * |
| 18 | * Fold a 32bit running checksum to 16bit and invert it. This is usually | 18 | * Fold a 32bit running checksum to 16bit and invert it. This is usually |
| 19 | * the last step before putting a checksum into a packet. | 19 | * the last step before putting a checksum into a packet. |
| 20 | * Make sure not to mix with 64bit checksums. | 20 | * Make sure not to mix with 64bit checksums. |
| 21 | */ | 21 | */ |
| 22 | static inline __sum16 csum_fold(__wsum sum) | 22 | static inline __sum16 csum_fold(__wsum sum) |
| 23 | { | 23 | { |
| 24 | __asm__( | 24 | asm(" addl %1,%0\n" |
| 25 | " addl %1,%0\n" | 25 | " adcl $0xffff,%0" |
| 26 | " adcl $0xffff,%0" | 26 | : "=r" (sum) |
| 27 | : "=r" (sum) | 27 | : "r" ((__force u32)sum << 16), |
| 28 | : "r" ((__force u32)sum << 16), | 28 | "0" ((__force u32)sum & 0xffff0000)); |
| 29 | "0" ((__force u32)sum & 0xffff0000) | ||
| 30 | ); | ||
| 31 | return (__force __sum16)(~(__force u32)sum >> 16); | 29 | return (__force __sum16)(~(__force u32)sum >> 16); |
| 32 | } | 30 | } |
| 33 | 31 | ||
| @@ -43,46 +41,46 @@ static inline __sum16 csum_fold(__wsum sum) | |||
| 43 | * ip_fast_csum - Compute the IPv4 header checksum efficiently. | 41 | * ip_fast_csum - Compute the IPv4 header checksum efficiently. |
| 44 | * iph: ipv4 header | 42 | * iph: ipv4 header |
| 45 | * ihl: length of header / 4 | 43 | * ihl: length of header / 4 |
| 46 | */ | 44 | */ |
| 47 | static inline __sum16 ip_fast_csum(const void *iph, unsigned int ihl) | 45 | static inline __sum16 ip_fast_csum(const void *iph, unsigned int ihl) |
| 48 | { | 46 | { |
| 49 | unsigned int sum; | 47 | unsigned int sum; |
| 50 | 48 | ||
| 51 | asm( " movl (%1), %0\n" | 49 | asm(" movl (%1), %0\n" |
| 52 | " subl $4, %2\n" | 50 | " subl $4, %2\n" |
| 53 | " jbe 2f\n" | 51 | " jbe 2f\n" |
| 54 | " addl 4(%1), %0\n" | 52 | " addl 4(%1), %0\n" |
| 55 | " adcl 8(%1), %0\n" | 53 | " adcl 8(%1), %0\n" |
| 56 | " adcl 12(%1), %0\n" | 54 | " adcl 12(%1), %0\n" |
| 57 | "1: adcl 16(%1), %0\n" | 55 | "1: adcl 16(%1), %0\n" |
| 58 | " lea 4(%1), %1\n" | 56 | " lea 4(%1), %1\n" |
| 59 | " decl %2\n" | 57 | " decl %2\n" |
| 60 | " jne 1b\n" | 58 | " jne 1b\n" |
| 61 | " adcl $0, %0\n" | 59 | " adcl $0, %0\n" |
| 62 | " movl %0, %2\n" | 60 | " movl %0, %2\n" |
| 63 | " shrl $16, %0\n" | 61 | " shrl $16, %0\n" |
| 64 | " addw %w2, %w0\n" | 62 | " addw %w2, %w0\n" |
| 65 | " adcl $0, %0\n" | 63 | " adcl $0, %0\n" |
| 66 | " notl %0\n" | 64 | " notl %0\n" |
| 67 | "2:" | 65 | "2:" |
| 68 | /* Since the input registers which are loaded with iph and ihl | 66 | /* Since the input registers which are loaded with iph and ihl |
| 69 | are modified, we must also specify them as outputs, or gcc | 67 | are modified, we must also specify them as outputs, or gcc |
| 70 | will assume they contain their original values. */ | 68 | will assume they contain their original values. */ |
| 71 | : "=r" (sum), "=r" (iph), "=r" (ihl) | 69 | : "=r" (sum), "=r" (iph), "=r" (ihl) |
| 72 | : "1" (iph), "2" (ihl) | 70 | : "1" (iph), "2" (ihl) |
| 73 | : "memory"); | 71 | : "memory"); |
| 74 | return (__force __sum16)sum; | 72 | return (__force __sum16)sum; |
| 75 | } | 73 | } |
| 76 | 74 | ||
| 77 | /** | 75 | /** |
| 78 | * csum_tcpup_nofold - Compute an IPv4 pseudo header checksum. | 76 | * csum_tcpup_nofold - Compute an IPv4 pseudo header checksum. |
| 79 | * @saddr: source address | 77 | * @saddr: source address |
| 80 | * @daddr: destination address | 78 | * @daddr: destination address |
| 81 | * @len: length of packet | 79 | * @len: length of packet |
| 82 | * @proto: ip protocol of packet | 80 | * @proto: ip protocol of packet |
| 83 | * @sum: initial sum to be added in (32bit unfolded) | 81 | * @sum: initial sum to be added in (32bit unfolded) |
| 84 | * | 82 | * |
| 85 | * Returns the pseudo header checksum the input data. Result is | 83 | * Returns the pseudo header checksum the input data. Result is |
| 86 | * 32bit unfolded. | 84 | * 32bit unfolded. |
| 87 | */ | 85 | */ |
| 88 | static inline __wsum | 86 | static inline __wsum |
| @@ -93,32 +91,32 @@ csum_tcpudp_nofold(__be32 saddr, __be32 daddr, unsigned short len, | |||
| 93 | " adcl %2, %0\n" | 91 | " adcl %2, %0\n" |
| 94 | " adcl %3, %0\n" | 92 | " adcl %3, %0\n" |
| 95 | " adcl $0, %0\n" | 93 | " adcl $0, %0\n" |
| 96 | : "=r" (sum) | 94 | : "=r" (sum) |
| 97 | : "g" (daddr), "g" (saddr), | 95 | : "g" (daddr), "g" (saddr), |
| 98 | "g" ((len + proto)<<8), "0" (sum)); | 96 | "g" ((len + proto)<<8), "0" (sum)); |
| 99 | return sum; | 97 | return sum; |
| 100 | } | 98 | } |
| 101 | 99 | ||
| 102 | 100 | ||
| 103 | /** | 101 | /** |
| 104 | * csum_tcpup_magic - Compute an IPv4 pseudo header checksum. | 102 | * csum_tcpup_magic - Compute an IPv4 pseudo header checksum. |
| 105 | * @saddr: source address | 103 | * @saddr: source address |
| 106 | * @daddr: destination address | 104 | * @daddr: destination address |
| 107 | * @len: length of packet | 105 | * @len: length of packet |
| 108 | * @proto: ip protocol of packet | 106 | * @proto: ip protocol of packet |
| 109 | * @sum: initial sum to be added in (32bit unfolded) | 107 | * @sum: initial sum to be added in (32bit unfolded) |
| 110 | * | 108 | * |
| 111 | * Returns the 16bit pseudo header checksum the input data already | 109 | * Returns the 16bit pseudo header checksum the input data already |
| 112 | * complemented and ready to be filled in. | 110 | * complemented and ready to be filled in. |
| 113 | */ | 111 | */ |
| 114 | static inline __sum16 | 112 | static inline __sum16 csum_tcpudp_magic(__be32 saddr, __be32 daddr, |
| 115 | csum_tcpudp_magic(__be32 saddr, __be32 daddr, | 113 | unsigned short len, |
| 116 | unsigned short len, unsigned short proto, __wsum sum) | 114 | unsigned short proto, __wsum sum) |
| 117 | { | 115 | { |
| 118 | return csum_fold(csum_tcpudp_nofold(saddr,daddr,len,proto,sum)); | 116 | return csum_fold(csum_tcpudp_nofold(saddr, daddr, len, proto, sum)); |
| 119 | } | 117 | } |
| 120 | 118 | ||
| 121 | /** | 119 | /** |
| 122 | * csum_partial - Compute an internet checksum. | 120 | * csum_partial - Compute an internet checksum. |
| 123 | * @buff: buffer to be checksummed | 121 | * @buff: buffer to be checksummed |
| 124 | * @len: length of buffer. | 122 | * @len: length of buffer. |
| @@ -127,7 +125,7 @@ csum_tcpudp_magic(__be32 saddr, __be32 daddr, | |||
| 127 | * Returns the 32bit unfolded internet checksum of the buffer. | 125 | * Returns the 32bit unfolded internet checksum of the buffer. |
| 128 | * Before filling it in it needs to be csum_fold()'ed. | 126 | * Before filling it in it needs to be csum_fold()'ed. |
| 129 | * buff should be aligned to a 64bit boundary if possible. | 127 | * buff should be aligned to a 64bit boundary if possible. |
| 130 | */ | 128 | */ |
| 131 | extern __wsum csum_partial(const void *buff, int len, __wsum sum); | 129 | extern __wsum csum_partial(const void *buff, int len, __wsum sum); |
| 132 | 130 | ||
| 133 | #define _HAVE_ARCH_COPY_AND_CSUM_FROM_USER 1 | 131 | #define _HAVE_ARCH_COPY_AND_CSUM_FROM_USER 1 |
| @@ -136,23 +134,22 @@ extern __wsum csum_partial(const void *buff, int len, __wsum sum); | |||
| 136 | 134 | ||
| 137 | /* Do not call this directly. Use the wrappers below */ | 135 | /* Do not call this directly. Use the wrappers below */ |
| 138 | extern __wsum csum_partial_copy_generic(const void *src, const void *dst, | 136 | extern __wsum csum_partial_copy_generic(const void *src, const void *dst, |
| 139 | int len, | 137 | int len, __wsum sum, |
| 140 | __wsum sum, | 138 | int *src_err_ptr, int *dst_err_ptr); |
| 141 | int *src_err_ptr, int *dst_err_ptr); | ||
| 142 | 139 | ||
| 143 | 140 | ||
| 144 | extern __wsum csum_partial_copy_from_user(const void __user *src, void *dst, | 141 | extern __wsum csum_partial_copy_from_user(const void __user *src, void *dst, |
| 145 | int len, __wsum isum, int *errp); | 142 | int len, __wsum isum, int *errp); |
| 146 | extern __wsum csum_partial_copy_to_user(const void *src, void __user *dst, | 143 | extern __wsum csum_partial_copy_to_user(const void *src, void __user *dst, |
| 147 | int len, __wsum isum, int *errp); | 144 | int len, __wsum isum, int *errp); |
| 148 | extern __wsum csum_partial_copy_nocheck(const void *src, void *dst, int len, | 145 | extern __wsum csum_partial_copy_nocheck(const void *src, void *dst, |
| 149 | __wsum sum); | 146 | int len, __wsum sum); |
| 150 | 147 | ||
| 151 | /* Old names. To be removed. */ | 148 | /* Old names. To be removed. */ |
| 152 | #define csum_and_copy_to_user csum_partial_copy_to_user | 149 | #define csum_and_copy_to_user csum_partial_copy_to_user |
| 153 | #define csum_and_copy_from_user csum_partial_copy_from_user | 150 | #define csum_and_copy_from_user csum_partial_copy_from_user |
| 154 | 151 | ||
| 155 | /** | 152 | /** |
| 156 | * ip_compute_csum - Compute an 16bit IP checksum. | 153 | * ip_compute_csum - Compute an 16bit IP checksum. |
| 157 | * @buff: buffer address. | 154 | * @buff: buffer address. |
| 158 | * @len: length of buffer. | 155 | * @len: length of buffer. |
| @@ -170,7 +167,7 @@ extern __sum16 ip_compute_csum(const void *buff, int len); | |||
| 170 | * @proto: protocol of packet | 167 | * @proto: protocol of packet |
| 171 | * @sum: initial sum (32bit unfolded) to be added in | 168 | * @sum: initial sum (32bit unfolded) to be added in |
| 172 | * | 169 | * |
| 173 | * Computes an IPv6 pseudo header checksum. This sum is added the checksum | 170 | * Computes an IPv6 pseudo header checksum. This sum is added the checksum |
| 174 | * into UDP/TCP packets and contains some link layer information. | 171 | * into UDP/TCP packets and contains some link layer information. |
| 175 | * Returns the unfolded 32bit checksum. | 172 | * Returns the unfolded 32bit checksum. |
| 176 | */ | 173 | */ |
| @@ -185,11 +182,10 @@ csum_ipv6_magic(const struct in6_addr *saddr, const struct in6_addr *daddr, | |||
| 185 | static inline unsigned add32_with_carry(unsigned a, unsigned b) | 182 | static inline unsigned add32_with_carry(unsigned a, unsigned b) |
| 186 | { | 183 | { |
| 187 | asm("addl %2,%0\n\t" | 184 | asm("addl %2,%0\n\t" |
| 188 | "adcl $0,%0" | 185 | "adcl $0,%0" |
| 189 | : "=r" (a) | 186 | : "=r" (a) |
| 190 | : "0" (a), "r" (b)); | 187 | : "0" (a), "r" (b)); |
| 191 | return a; | 188 | return a; |
| 192 | } | 189 | } |
| 193 | 190 | ||
| 194 | #endif | 191 | #endif |
| 195 | |||
diff --git a/include/asm-x86/cmpxchg_32.h b/include/asm-x86/cmpxchg_32.h index 959fad00dff5..bf5a69d1329e 100644 --- a/include/asm-x86/cmpxchg_32.h +++ b/include/asm-x86/cmpxchg_32.h | |||
| @@ -8,9 +8,12 @@ | |||
| 8 | * you need to test for the feature in boot_cpu_data. | 8 | * you need to test for the feature in boot_cpu_data. |
| 9 | */ | 9 | */ |
| 10 | 10 | ||
| 11 | #define xchg(ptr,v) ((__typeof__(*(ptr)))__xchg((unsigned long)(v),(ptr),sizeof(*(ptr)))) | 11 | #define xchg(ptr, v) \ |
| 12 | ((__typeof__(*(ptr)))__xchg((unsigned long)(v), (ptr), sizeof(*(ptr)))) | ||
| 12 | 13 | ||
| 13 | struct __xchg_dummy { unsigned long a[100]; }; | 14 | struct __xchg_dummy { |
| 15 | unsigned long a[100]; | ||
| 16 | }; | ||
| 14 | #define __xg(x) ((struct __xchg_dummy *)(x)) | 17 | #define __xg(x) ((struct __xchg_dummy *)(x)) |
| 15 | 18 | ||
| 16 | /* | 19 | /* |
| @@ -27,72 +30,74 @@ struct __xchg_dummy { unsigned long a[100]; }; | |||
| 27 | * of the instruction set reference 24319102.pdf. We need | 30 | * of the instruction set reference 24319102.pdf. We need |
| 28 | * the reader side to see the coherent 64bit value. | 31 | * the reader side to see the coherent 64bit value. |
| 29 | */ | 32 | */ |
| 30 | static inline void __set_64bit (unsigned long long * ptr, | 33 | static inline void __set_64bit(unsigned long long *ptr, |
| 31 | unsigned int low, unsigned int high) | 34 | unsigned int low, unsigned int high) |
| 32 | { | 35 | { |
| 33 | __asm__ __volatile__ ( | 36 | asm volatile("\n1:\t" |
| 34 | "\n1:\t" | 37 | "movl (%0), %%eax\n\t" |
| 35 | "movl (%0), %%eax\n\t" | 38 | "movl 4(%0), %%edx\n\t" |
| 36 | "movl 4(%0), %%edx\n\t" | 39 | LOCK_PREFIX "cmpxchg8b (%0)\n\t" |
| 37 | LOCK_PREFIX "cmpxchg8b (%0)\n\t" | 40 | "jnz 1b" |
| 38 | "jnz 1b" | 41 | : /* no outputs */ |
| 39 | : /* no outputs */ | 42 | : "D"(ptr), |
| 40 | : "D"(ptr), | 43 | "b"(low), |
| 41 | "b"(low), | 44 | "c"(high) |
| 42 | "c"(high) | 45 | : "ax", "dx", "memory"); |
| 43 | : "ax","dx","memory"); | ||
| 44 | } | 46 | } |
| 45 | 47 | ||
| 46 | static inline void __set_64bit_constant (unsigned long long *ptr, | 48 | static inline void __set_64bit_constant(unsigned long long *ptr, |
| 47 | unsigned long long value) | 49 | unsigned long long value) |
| 48 | { | 50 | { |
| 49 | __set_64bit(ptr,(unsigned int)(value), (unsigned int)((value)>>32ULL)); | 51 | __set_64bit(ptr, (unsigned int)value, (unsigned int)(value >> 32)); |
| 50 | } | 52 | } |
| 51 | #define ll_low(x) *(((unsigned int*)&(x))+0) | ||
| 52 | #define ll_high(x) *(((unsigned int*)&(x))+1) | ||
| 53 | 53 | ||
| 54 | static inline void __set_64bit_var (unsigned long long *ptr, | 54 | #define ll_low(x) *(((unsigned int *)&(x)) + 0) |
| 55 | unsigned long long value) | 55 | #define ll_high(x) *(((unsigned int *)&(x)) + 1) |
| 56 | |||
| 57 | static inline void __set_64bit_var(unsigned long long *ptr, | ||
| 58 | unsigned long long value) | ||
| 56 | { | 59 | { |
| 57 | __set_64bit(ptr,ll_low(value), ll_high(value)); | 60 | __set_64bit(ptr, ll_low(value), ll_high(value)); |
| 58 | } | 61 | } |
| 59 | 62 | ||
| 60 | #define set_64bit(ptr,value) \ | 63 | #define set_64bit(ptr, value) \ |
| 61 | (__builtin_constant_p(value) ? \ | 64 | (__builtin_constant_p((value)) \ |
| 62 | __set_64bit_constant(ptr, value) : \ | 65 | ? __set_64bit_constant((ptr), (value)) \ |
| 63 | __set_64bit_var(ptr, value) ) | 66 | : __set_64bit_var((ptr), (value))) |
| 64 | 67 | ||
| 65 | #define _set_64bit(ptr,value) \ | 68 | #define _set_64bit(ptr, value) \ |
| 66 | (__builtin_constant_p(value) ? \ | 69 | (__builtin_constant_p(value) \ |
| 67 | __set_64bit(ptr, (unsigned int)(value), (unsigned int)((value)>>32ULL) ) : \ | 70 | ? __set_64bit(ptr, (unsigned int)(value), \ |
| 68 | __set_64bit(ptr, ll_low(value), ll_high(value)) ) | 71 | (unsigned int)((value) >> 32)) \ |
| 72 | : __set_64bit(ptr, ll_low((value)), ll_high((value)))) | ||
| 69 | 73 | ||
| 70 | /* | 74 | /* |
| 71 | * Note: no "lock" prefix even on SMP: xchg always implies lock anyway | 75 | * Note: no "lock" prefix even on SMP: xchg always implies lock anyway |
| 72 | * Note 2: xchg has side effect, so that attribute volatile is necessary, | 76 | * Note 2: xchg has side effect, so that attribute volatile is necessary, |
| 73 | * but generally the primitive is invalid, *ptr is output argument. --ANK | 77 | * but generally the primitive is invalid, *ptr is output argument. --ANK |
| 74 | */ | 78 | */ |
| 75 | static inline unsigned long __xchg(unsigned long x, volatile void * ptr, int size) | 79 | static inline unsigned long __xchg(unsigned long x, volatile void *ptr, |
| 80 | int size) | ||
| 76 | { | 81 | { |
| 77 | switch (size) { | 82 | switch (size) { |
| 78 | case 1: | 83 | case 1: |
| 79 | __asm__ __volatile__("xchgb %b0,%1" | 84 | asm volatile("xchgb %b0,%1" |
| 80 | :"=q" (x) | 85 | : "=q" (x) |
| 81 | :"m" (*__xg(ptr)), "0" (x) | 86 | : "m" (*__xg(ptr)), "0" (x) |
| 82 | :"memory"); | 87 | : "memory"); |
| 83 | break; | 88 | break; |
| 84 | case 2: | 89 | case 2: |
| 85 | __asm__ __volatile__("xchgw %w0,%1" | 90 | asm volatile("xchgw %w0,%1" |
| 86 | :"=r" (x) | 91 | : "=r" (x) |
| 87 | :"m" (*__xg(ptr)), "0" (x) | 92 | : "m" (*__xg(ptr)), "0" (x) |
| 88 | :"memory"); | 93 | : "memory"); |
| 89 | break; | 94 | break; |
| 90 | case 4: | 95 | case 4: |
| 91 | __asm__ __volatile__("xchgl %0,%1" | 96 | asm volatile("xchgl %0,%1" |
| 92 | :"=r" (x) | 97 | : "=r" (x) |
| 93 | :"m" (*__xg(ptr)), "0" (x) | 98 | : "m" (*__xg(ptr)), "0" (x) |
| 94 | :"memory"); | 99 | : "memory"); |
| 95 | break; | 100 | break; |
| 96 | } | 101 | } |
| 97 | return x; | 102 | return x; |
| 98 | } | 103 | } |
| @@ -105,24 +110,27 @@ static inline unsigned long __xchg(unsigned long x, volatile void * ptr, int siz | |||
| 105 | 110 | ||
| 106 | #ifdef CONFIG_X86_CMPXCHG | 111 | #ifdef CONFIG_X86_CMPXCHG |
| 107 | #define __HAVE_ARCH_CMPXCHG 1 | 112 | #define __HAVE_ARCH_CMPXCHG 1 |
| 108 | #define cmpxchg(ptr, o, n) \ | 113 | #define cmpxchg(ptr, o, n) \ |
| 109 | ((__typeof__(*(ptr)))__cmpxchg((ptr), (unsigned long)(o), \ | 114 | ((__typeof__(*(ptr)))__cmpxchg((ptr), (unsigned long)(o), \ |
| 110 | (unsigned long)(n), sizeof(*(ptr)))) | 115 | (unsigned long)(n), \ |
| 111 | #define sync_cmpxchg(ptr, o, n) \ | 116 | sizeof(*(ptr)))) |
| 112 | ((__typeof__(*(ptr)))__sync_cmpxchg((ptr), (unsigned long)(o), \ | 117 | #define sync_cmpxchg(ptr, o, n) \ |
| 113 | (unsigned long)(n), sizeof(*(ptr)))) | 118 | ((__typeof__(*(ptr)))__sync_cmpxchg((ptr), (unsigned long)(o), \ |
| 114 | #define cmpxchg_local(ptr, o, n) \ | 119 | (unsigned long)(n), \ |
| 115 | ((__typeof__(*(ptr)))__cmpxchg_local((ptr), (unsigned long)(o), \ | 120 | sizeof(*(ptr)))) |
| 116 | (unsigned long)(n), sizeof(*(ptr)))) | 121 | #define cmpxchg_local(ptr, o, n) \ |
| 122 | ((__typeof__(*(ptr)))__cmpxchg_local((ptr), (unsigned long)(o), \ | ||
| 123 | (unsigned long)(n), \ | ||
| 124 | sizeof(*(ptr)))) | ||
| 117 | #endif | 125 | #endif |
| 118 | 126 | ||
| 119 | #ifdef CONFIG_X86_CMPXCHG64 | 127 | #ifdef CONFIG_X86_CMPXCHG64 |
| 120 | #define cmpxchg64(ptr, o, n) \ | 128 | #define cmpxchg64(ptr, o, n) \ |
| 121 | ((__typeof__(*(ptr)))__cmpxchg64((ptr), (unsigned long long)(o), \ | 129 | ((__typeof__(*(ptr)))__cmpxchg64((ptr), (unsigned long long)(o), \ |
| 122 | (unsigned long long)(n))) | 130 | (unsigned long long)(n))) |
| 123 | #define cmpxchg64_local(ptr, o, n) \ | 131 | #define cmpxchg64_local(ptr, o, n) \ |
| 124 | ((__typeof__(*(ptr)))__cmpxchg64_local((ptr), (unsigned long long)(o),\ | 132 | ((__typeof__(*(ptr)))__cmpxchg64_local((ptr), (unsigned long long)(o), \ |
| 125 | (unsigned long long)(n))) | 133 | (unsigned long long)(n))) |
| 126 | #endif | 134 | #endif |
| 127 | 135 | ||
| 128 | static inline unsigned long __cmpxchg(volatile void *ptr, unsigned long old, | 136 | static inline unsigned long __cmpxchg(volatile void *ptr, unsigned long old, |
| @@ -131,22 +139,22 @@ static inline unsigned long __cmpxchg(volatile void *ptr, unsigned long old, | |||
| 131 | unsigned long prev; | 139 | unsigned long prev; |
| 132 | switch (size) { | 140 | switch (size) { |
| 133 | case 1: | 141 | case 1: |
| 134 | __asm__ __volatile__(LOCK_PREFIX "cmpxchgb %b1,%2" | 142 | asm volatile(LOCK_PREFIX "cmpxchgb %b1,%2" |
| 135 | : "=a"(prev) | 143 | : "=a"(prev) |
| 136 | : "q"(new), "m"(*__xg(ptr)), "0"(old) | 144 | : "q"(new), "m"(*__xg(ptr)), "0"(old) |
| 137 | : "memory"); | 145 | : "memory"); |
| 138 | return prev; | 146 | return prev; |
| 139 | case 2: | 147 | case 2: |
| 140 | __asm__ __volatile__(LOCK_PREFIX "cmpxchgw %w1,%2" | 148 | asm volatile(LOCK_PREFIX "cmpxchgw %w1,%2" |
| 141 | : "=a"(prev) | 149 | : "=a"(prev) |
| 142 | : "r"(new), "m"(*__xg(ptr)), "0"(old) | 150 | : "r"(new), "m"(*__xg(ptr)), "0"(old) |
| 143 | : "memory"); | 151 | : "memory"); |
| 144 | return prev; | 152 | return prev; |
| 145 | case 4: | 153 | case 4: |
| 146 | __asm__ __volatile__(LOCK_PREFIX "cmpxchgl %1,%2" | 154 | asm volatile(LOCK_PREFIX "cmpxchgl %1,%2" |
| 147 | : "=a"(prev) | 155 | : "=a"(prev) |
| 148 | : "r"(new), "m"(*__xg(ptr)), "0"(old) | 156 | : "r"(new), "m"(*__xg(ptr)), "0"(old) |
| 149 | : "memory"); | 157 | : "memory"); |
| 150 | return prev; | 158 | return prev; |
| 151 | } | 159 | } |
| 152 | return old; | 160 | return old; |
| @@ -158,85 +166,88 @@ static inline unsigned long __cmpxchg(volatile void *ptr, unsigned long old, | |||
| 158 | * isn't. | 166 | * isn't. |
| 159 | */ | 167 | */ |
| 160 | static inline unsigned long __sync_cmpxchg(volatile void *ptr, | 168 | static inline unsigned long __sync_cmpxchg(volatile void *ptr, |
| 161 | unsigned long old, | 169 | unsigned long old, |
| 162 | unsigned long new, int size) | 170 | unsigned long new, int size) |
| 163 | { | 171 | { |
| 164 | unsigned long prev; | 172 | unsigned long prev; |
| 165 | switch (size) { | 173 | switch (size) { |
| 166 | case 1: | 174 | case 1: |
| 167 | __asm__ __volatile__("lock; cmpxchgb %b1,%2" | 175 | asm volatile("lock; cmpxchgb %b1,%2" |
| 168 | : "=a"(prev) | 176 | : "=a"(prev) |
| 169 | : "q"(new), "m"(*__xg(ptr)), "0"(old) | 177 | : "q"(new), "m"(*__xg(ptr)), "0"(old) |
| 170 | : "memory"); | 178 | : "memory"); |
| 171 | return prev; | 179 | return prev; |
| 172 | case 2: | 180 | case 2: |
| 173 | __asm__ __volatile__("lock; cmpxchgw %w1,%2" | 181 | asm volatile("lock; cmpxchgw %w1,%2" |
| 174 | : "=a"(prev) | 182 | : "=a"(prev) |
| 175 | : "r"(new), "m"(*__xg(ptr)), "0"(old) | 183 | : "r"(new), "m"(*__xg(ptr)), "0"(old) |
| 176 | : "memory"); | 184 | : "memory"); |
| 177 | return prev; | 185 | return prev; |
| 178 | case 4: | 186 | case 4: |
| 179 | __asm__ __volatile__("lock; cmpxchgl %1,%2" | 187 | asm volatile("lock; cmpxchgl %1,%2" |
| 180 | : "=a"(prev) | 188 | : "=a"(prev) |
| 181 | : "r"(new), "m"(*__xg(ptr)), "0"(old) | 189 | : "r"(new), "m"(*__xg(ptr)), "0"(old) |
| 182 | : "memory"); | 190 | : "memory"); |
| 183 | return prev; | 191 | return prev; |
| 184 | } | 192 | } |
| 185 | return old; | 193 | return old; |
| 186 | } | 194 | } |
| 187 | 195 | ||
| 188 | static inline unsigned long __cmpxchg_local(volatile void *ptr, | 196 | static inline unsigned long __cmpxchg_local(volatile void *ptr, |
| 189 | unsigned long old, unsigned long new, int size) | 197 | unsigned long old, |
| 198 | unsigned long new, int size) | ||
| 190 | { | 199 | { |
| 191 | unsigned long prev; | 200 | unsigned long prev; |
| 192 | switch (size) { | 201 | switch (size) { |
| 193 | case 1: | 202 | case 1: |
| 194 | __asm__ __volatile__("cmpxchgb %b1,%2" | 203 | asm volatile("cmpxchgb %b1,%2" |
| 195 | : "=a"(prev) | 204 | : "=a"(prev) |
| 196 | : "q"(new), "m"(*__xg(ptr)), "0"(old) | 205 | : "q"(new), "m"(*__xg(ptr)), "0"(old) |
| 197 | : "memory"); | 206 | : "memory"); |
| 198 | return prev; | 207 | return prev; |
| 199 | case 2: | 208 | case 2: |
| 200 | __asm__ __volatile__("cmpxchgw %w1,%2" | 209 | asm volatile("cmpxchgw %w1,%2" |
| 201 | : "=a"(prev) | 210 | : "=a"(prev) |
| 202 | : "r"(new), "m"(*__xg(ptr)), "0"(old) | 211 | : "r"(new), "m"(*__xg(ptr)), "0"(old) |
| 203 | : "memory"); | 212 | : "memory"); |
| 204 | return prev; | 213 | return prev; |
| 205 | case 4: | 214 | case 4: |
| 206 | __asm__ __volatile__("cmpxchgl %1,%2" | 215 | asm volatile("cmpxchgl %1,%2" |
| 207 | : "=a"(prev) | 216 | : "=a"(prev) |
| 208 | : "r"(new), "m"(*__xg(ptr)), "0"(old) | 217 | : "r"(new), "m"(*__xg(ptr)), "0"(old) |
| 209 | : "memory"); | 218 | : "memory"); |
| 210 | return prev; | 219 | return prev; |
| 211 | } | 220 | } |
| 212 | return old; | 221 | return old; |
| 213 | } | 222 | } |
| 214 | 223 | ||
| 215 | static inline unsigned long long __cmpxchg64(volatile void *ptr, | 224 | static inline unsigned long long __cmpxchg64(volatile void *ptr, |
| 216 | unsigned long long old, unsigned long long new) | 225 | unsigned long long old, |
| 226 | unsigned long long new) | ||
| 217 | { | 227 | { |
| 218 | unsigned long long prev; | 228 | unsigned long long prev; |
| 219 | __asm__ __volatile__(LOCK_PREFIX "cmpxchg8b %3" | 229 | asm volatile(LOCK_PREFIX "cmpxchg8b %3" |
| 220 | : "=A"(prev) | 230 | : "=A"(prev) |
| 221 | : "b"((unsigned long)new), | 231 | : "b"((unsigned long)new), |
| 222 | "c"((unsigned long)(new >> 32)), | 232 | "c"((unsigned long)(new >> 32)), |
| 223 | "m"(*__xg(ptr)), | 233 | "m"(*__xg(ptr)), |
| 224 | "0"(old) | 234 | "0"(old) |
| 225 | : "memory"); | 235 | : "memory"); |
| 226 | return prev; | 236 | return prev; |
| 227 | } | 237 | } |
| 228 | 238 | ||
| 229 | static inline unsigned long long __cmpxchg64_local(volatile void *ptr, | 239 | static inline unsigned long long __cmpxchg64_local(volatile void *ptr, |
| 230 | unsigned long long old, unsigned long long new) | 240 | unsigned long long old, |
| 241 | unsigned long long new) | ||
| 231 | { | 242 | { |
| 232 | unsigned long long prev; | 243 | unsigned long long prev; |
| 233 | __asm__ __volatile__("cmpxchg8b %3" | 244 | asm volatile("cmpxchg8b %3" |
| 234 | : "=A"(prev) | 245 | : "=A"(prev) |
| 235 | : "b"((unsigned long)new), | 246 | : "b"((unsigned long)new), |
| 236 | "c"((unsigned long)(new >> 32)), | 247 | "c"((unsigned long)(new >> 32)), |
| 237 | "m"(*__xg(ptr)), | 248 | "m"(*__xg(ptr)), |
| 238 | "0"(old) | 249 | "0"(old) |
| 239 | : "memory"); | 250 | : "memory"); |
| 240 | return prev; | 251 | return prev; |
| 241 | } | 252 | } |
| 242 | 253 | ||
| @@ -252,7 +263,7 @@ extern unsigned long cmpxchg_386_u16(volatile void *, u16, u16); | |||
| 252 | extern unsigned long cmpxchg_386_u32(volatile void *, u32, u32); | 263 | extern unsigned long cmpxchg_386_u32(volatile void *, u32, u32); |
| 253 | 264 | ||
| 254 | static inline unsigned long cmpxchg_386(volatile void *ptr, unsigned long old, | 265 | static inline unsigned long cmpxchg_386(volatile void *ptr, unsigned long old, |
| 255 | unsigned long new, int size) | 266 | unsigned long new, int size) |
| 256 | { | 267 | { |
| 257 | switch (size) { | 268 | switch (size) { |
| 258 | case 1: | 269 | case 1: |
diff --git a/include/asm-x86/cmpxchg_64.h b/include/asm-x86/cmpxchg_64.h index 56f5b41e071c..d9b26b9a28cf 100644 --- a/include/asm-x86/cmpxchg_64.h +++ b/include/asm-x86/cmpxchg_64.h | |||
| @@ -3,7 +3,8 @@ | |||
| 3 | 3 | ||
| 4 | #include <asm/alternative.h> /* Provides LOCK_PREFIX */ | 4 | #include <asm/alternative.h> /* Provides LOCK_PREFIX */ |
| 5 | 5 | ||
| 6 | #define xchg(ptr,v) ((__typeof__(*(ptr)))__xchg((unsigned long)(v),(ptr),sizeof(*(ptr)))) | 6 | #define xchg(ptr, v) ((__typeof__(*(ptr)))__xchg((unsigned long)(v), \ |
| 7 | (ptr), sizeof(*(ptr)))) | ||
| 7 | 8 | ||
| 8 | #define __xg(x) ((volatile long *)(x)) | 9 | #define __xg(x) ((volatile long *)(x)) |
| 9 | 10 | ||
| @@ -19,33 +20,34 @@ static inline void set_64bit(volatile unsigned long *ptr, unsigned long val) | |||
| 19 | * Note 2: xchg has side effect, so that attribute volatile is necessary, | 20 | * Note 2: xchg has side effect, so that attribute volatile is necessary, |
| 20 | * but generally the primitive is invalid, *ptr is output argument. --ANK | 21 | * but generally the primitive is invalid, *ptr is output argument. --ANK |
| 21 | */ | 22 | */ |
| 22 | static inline unsigned long __xchg(unsigned long x, volatile void * ptr, int size) | 23 | static inline unsigned long __xchg(unsigned long x, volatile void *ptr, |
| 24 | int size) | ||
| 23 | { | 25 | { |
| 24 | switch (size) { | 26 | switch (size) { |
| 25 | case 1: | 27 | case 1: |
| 26 | __asm__ __volatile__("xchgb %b0,%1" | 28 | asm volatile("xchgb %b0,%1" |
| 27 | :"=q" (x) | 29 | : "=q" (x) |
| 28 | :"m" (*__xg(ptr)), "0" (x) | 30 | : "m" (*__xg(ptr)), "0" (x) |
| 29 | :"memory"); | 31 | : "memory"); |
| 30 | break; | 32 | break; |
| 31 | case 2: | 33 | case 2: |
| 32 | __asm__ __volatile__("xchgw %w0,%1" | 34 | asm volatile("xchgw %w0,%1" |
| 33 | :"=r" (x) | 35 | : "=r" (x) |
| 34 | :"m" (*__xg(ptr)), "0" (x) | 36 | : "m" (*__xg(ptr)), "0" (x) |
| 35 | :"memory"); | 37 | : "memory"); |
| 36 | break; | 38 | break; |
| 37 | case 4: | 39 | case 4: |
| 38 | __asm__ __volatile__("xchgl %k0,%1" | 40 | asm volatile("xchgl %k0,%1" |
| 39 | :"=r" (x) | 41 | : "=r" (x) |
| 40 | :"m" (*__xg(ptr)), "0" (x) | 42 | : "m" (*__xg(ptr)), "0" (x) |
| 41 | :"memory"); | 43 | : "memory"); |
| 42 | break; | 44 | break; |
| 43 | case 8: | 45 | case 8: |
| 44 | __asm__ __volatile__("xchgq %0,%1" | 46 | asm volatile("xchgq %0,%1" |
| 45 | :"=r" (x) | 47 | : "=r" (x) |
| 46 | :"m" (*__xg(ptr)), "0" (x) | 48 | : "m" (*__xg(ptr)), "0" (x) |
| 47 | :"memory"); | 49 | : "memory"); |
| 48 | break; | 50 | break; |
| 49 | } | 51 | } |
| 50 | return x; | 52 | return x; |
| 51 | } | 53 | } |
| @@ -64,61 +66,62 @@ static inline unsigned long __cmpxchg(volatile void *ptr, unsigned long old, | |||
| 64 | unsigned long prev; | 66 | unsigned long prev; |
| 65 | switch (size) { | 67 | switch (size) { |
| 66 | case 1: | 68 | case 1: |
| 67 | __asm__ __volatile__(LOCK_PREFIX "cmpxchgb %b1,%2" | 69 | asm volatile(LOCK_PREFIX "cmpxchgb %b1,%2" |
| 68 | : "=a"(prev) | 70 | : "=a"(prev) |
| 69 | : "q"(new), "m"(*__xg(ptr)), "0"(old) | 71 | : "q"(new), "m"(*__xg(ptr)), "0"(old) |
| 70 | : "memory"); | 72 | : "memory"); |
| 71 | return prev; | 73 | return prev; |
| 72 | case 2: | 74 | case 2: |
| 73 | __asm__ __volatile__(LOCK_PREFIX "cmpxchgw %w1,%2" | 75 | asm volatile(LOCK_PREFIX "cmpxchgw %w1,%2" |
| 74 | : "=a"(prev) | 76 | : "=a"(prev) |
| 75 | : "r"(new), "m"(*__xg(ptr)), "0"(old) | 77 | : "r"(new), "m"(*__xg(ptr)), "0"(old) |
| 76 | : "memory"); | 78 | : "memory"); |
| 77 | return prev; | 79 | return prev; |
| 78 | case 4: | 80 | case 4: |
| 79 | __asm__ __volatile__(LOCK_PREFIX "cmpxchgl %k1,%2" | 81 | asm volatile(LOCK_PREFIX "cmpxchgl %k1,%2" |
| 80 | : "=a"(prev) | 82 | : "=a"(prev) |
| 81 | : "r"(new), "m"(*__xg(ptr)), "0"(old) | 83 | : "r"(new), "m"(*__xg(ptr)), "0"(old) |
| 82 | : "memory"); | 84 | : "memory"); |
| 83 | return prev; | 85 | return prev; |
| 84 | case 8: | 86 | case 8: |
| 85 | __asm__ __volatile__(LOCK_PREFIX "cmpxchgq %1,%2" | 87 | asm volatile(LOCK_PREFIX "cmpxchgq %1,%2" |
| 86 | : "=a"(prev) | 88 | : "=a"(prev) |
| 87 | : "r"(new), "m"(*__xg(ptr)), "0"(old) | 89 | : "r"(new), "m"(*__xg(ptr)), "0"(old) |
| 88 | : "memory"); | 90 | : "memory"); |
| 89 | return prev; | 91 | return prev; |
| 90 | } | 92 | } |
| 91 | return old; | 93 | return old; |
| 92 | } | 94 | } |
| 93 | 95 | ||
| 94 | static inline unsigned long __cmpxchg_local(volatile void *ptr, | 96 | static inline unsigned long __cmpxchg_local(volatile void *ptr, |
| 95 | unsigned long old, unsigned long new, int size) | 97 | unsigned long old, |
| 98 | unsigned long new, int size) | ||
| 96 | { | 99 | { |
| 97 | unsigned long prev; | 100 | unsigned long prev; |
| 98 | switch (size) { | 101 | switch (size) { |
| 99 | case 1: | 102 | case 1: |
| 100 | __asm__ __volatile__("cmpxchgb %b1,%2" | 103 | asm volatile("cmpxchgb %b1,%2" |
| 101 | : "=a"(prev) | 104 | : "=a"(prev) |
| 102 | : "q"(new), "m"(*__xg(ptr)), "0"(old) | 105 | : "q"(new), "m"(*__xg(ptr)), "0"(old) |
| 103 | : "memory"); | 106 | : "memory"); |
| 104 | return prev; | 107 | return prev; |
| 105 | case 2: | 108 | case 2: |
| 106 | __asm__ __volatile__("cmpxchgw %w1,%2" | 109 | asm volatile("cmpxchgw %w1,%2" |
| 107 | : "=a"(prev) | 110 | : "=a"(prev) |
| 108 | : "r"(new), "m"(*__xg(ptr)), "0"(old) | 111 | : "r"(new), "m"(*__xg(ptr)), "0"(old) |
| 109 | : "memory"); | 112 | : "memory"); |
| 110 | return prev; | 113 | return prev; |
| 111 | case 4: | 114 | case 4: |
| 112 | __asm__ __volatile__("cmpxchgl %k1,%2" | 115 | asm volatile("cmpxchgl %k1,%2" |
| 113 | : "=a"(prev) | 116 | : "=a"(prev) |
| 114 | : "r"(new), "m"(*__xg(ptr)), "0"(old) | 117 | : "r"(new), "m"(*__xg(ptr)), "0"(old) |
| 115 | : "memory"); | 118 | : "memory"); |
| 116 | return prev; | 119 | return prev; |
| 117 | case 8: | 120 | case 8: |
| 118 | __asm__ __volatile__("cmpxchgq %1,%2" | 121 | asm volatile("cmpxchgq %1,%2" |
| 119 | : "=a"(prev) | 122 | : "=a"(prev) |
| 120 | : "r"(new), "m"(*__xg(ptr)), "0"(old) | 123 | : "r"(new), "m"(*__xg(ptr)), "0"(old) |
| 121 | : "memory"); | 124 | : "memory"); |
| 122 | return prev; | 125 | return prev; |
| 123 | } | 126 | } |
| 124 | return old; | 127 | return old; |
| @@ -126,19 +129,20 @@ static inline unsigned long __cmpxchg_local(volatile void *ptr, | |||
| 126 | 129 | ||
| 127 | #define cmpxchg(ptr, o, n) \ | 130 | #define cmpxchg(ptr, o, n) \ |
| 128 | ((__typeof__(*(ptr)))__cmpxchg((ptr), (unsigned long)(o), \ | 131 | ((__typeof__(*(ptr)))__cmpxchg((ptr), (unsigned long)(o), \ |
| 129 | (unsigned long)(n), sizeof(*(ptr)))) | 132 | (unsigned long)(n), sizeof(*(ptr)))) |
| 130 | #define cmpxchg64(ptr, o, n) \ | 133 | #define cmpxchg64(ptr, o, n) \ |
| 131 | ({ \ | 134 | ({ \ |
| 132 | BUILD_BUG_ON(sizeof(*(ptr)) != 8); \ | 135 | BUILD_BUG_ON(sizeof(*(ptr)) != 8); \ |
| 133 | cmpxchg((ptr), (o), (n)); \ | 136 | cmpxchg((ptr), (o), (n)); \ |
| 134 | }) | 137 | }) |
| 135 | #define cmpxchg_local(ptr, o, n) \ | 138 | #define cmpxchg_local(ptr, o, n) \ |
| 136 | ((__typeof__(*(ptr)))__cmpxchg_local((ptr), (unsigned long)(o), \ | 139 | ((__typeof__(*(ptr)))__cmpxchg_local((ptr), (unsigned long)(o), \ |
| 137 | (unsigned long)(n), sizeof(*(ptr)))) | 140 | (unsigned long)(n), \ |
| 141 | sizeof(*(ptr)))) | ||
| 138 | #define cmpxchg64_local(ptr, o, n) \ | 142 | #define cmpxchg64_local(ptr, o, n) \ |
| 139 | ({ \ | 143 | ({ \ |
| 140 | BUILD_BUG_ON(sizeof(*(ptr)) != 8); \ | 144 | BUILD_BUG_ON(sizeof(*(ptr)) != 8); \ |
| 141 | cmpxchg_local((ptr), (o), (n)); \ | 145 | cmpxchg_local((ptr), (o), (n)); \ |
| 142 | }) | 146 | }) |
| 143 | 147 | ||
| 144 | #endif | 148 | #endif |
diff --git a/include/asm-x86/compat.h b/include/asm-x86/compat.h index d3e8f3e87ee8..1793ac317a30 100644 --- a/include/asm-x86/compat.h +++ b/include/asm-x86/compat.h | |||
| @@ -204,7 +204,7 @@ static inline compat_uptr_t ptr_to_compat(void __user *uptr) | |||
| 204 | return (u32)(unsigned long)uptr; | 204 | return (u32)(unsigned long)uptr; |
| 205 | } | 205 | } |
| 206 | 206 | ||
| 207 | static __inline__ void __user *compat_alloc_user_space(long len) | 207 | static inline void __user *compat_alloc_user_space(long len) |
| 208 | { | 208 | { |
| 209 | struct pt_regs *regs = task_pt_regs(current); | 209 | struct pt_regs *regs = task_pt_regs(current); |
| 210 | return (void __user *)regs->sp - len; | 210 | return (void __user *)regs->sp - len; |
diff --git a/include/asm-x86/cpufeature.h b/include/asm-x86/cpufeature.h index 065e92966c7c..0d609c837a41 100644 --- a/include/asm-x86/cpufeature.h +++ b/include/asm-x86/cpufeature.h | |||
| @@ -120,6 +120,9 @@ | |||
| 120 | extern const char * const x86_cap_flags[NCAPINTS*32]; | 120 | extern const char * const x86_cap_flags[NCAPINTS*32]; |
| 121 | extern const char * const x86_power_flags[32]; | 121 | extern const char * const x86_power_flags[32]; |
| 122 | 122 | ||
| 123 | #define test_cpu_cap(c, bit) \ | ||
| 124 | test_bit(bit, (unsigned long *)((c)->x86_capability)) | ||
| 125 | |||
| 123 | #define cpu_has(c, bit) \ | 126 | #define cpu_has(c, bit) \ |
| 124 | (__builtin_constant_p(bit) && \ | 127 | (__builtin_constant_p(bit) && \ |
| 125 | ( (((bit)>>5)==0 && (1UL<<((bit)&31) & REQUIRED_MASK0)) || \ | 128 | ( (((bit)>>5)==0 && (1UL<<((bit)&31) & REQUIRED_MASK0)) || \ |
| @@ -131,7 +134,8 @@ extern const char * const x86_power_flags[32]; | |||
| 131 | (((bit)>>5)==6 && (1UL<<((bit)&31) & REQUIRED_MASK6)) || \ | 134 | (((bit)>>5)==6 && (1UL<<((bit)&31) & REQUIRED_MASK6)) || \ |
| 132 | (((bit)>>5)==7 && (1UL<<((bit)&31) & REQUIRED_MASK7)) ) \ | 135 | (((bit)>>5)==7 && (1UL<<((bit)&31) & REQUIRED_MASK7)) ) \ |
| 133 | ? 1 : \ | 136 | ? 1 : \ |
| 134 | test_bit(bit, (unsigned long *)((c)->x86_capability))) | 137 | test_cpu_cap(c, bit)) |
| 138 | |||
| 135 | #define boot_cpu_has(bit) cpu_has(&boot_cpu_data, bit) | 139 | #define boot_cpu_has(bit) cpu_has(&boot_cpu_data, bit) |
| 136 | 140 | ||
| 137 | #define set_cpu_cap(c, bit) set_bit(bit, (unsigned long *)((c)->x86_capability)) | 141 | #define set_cpu_cap(c, bit) set_bit(bit, (unsigned long *)((c)->x86_capability)) |
| @@ -181,6 +185,8 @@ extern const char * const x86_power_flags[32]; | |||
| 181 | #define cpu_has_clflush boot_cpu_has(X86_FEATURE_CLFLSH) | 185 | #define cpu_has_clflush boot_cpu_has(X86_FEATURE_CLFLSH) |
| 182 | #define cpu_has_bts boot_cpu_has(X86_FEATURE_BTS) | 186 | #define cpu_has_bts boot_cpu_has(X86_FEATURE_BTS) |
| 183 | #define cpu_has_gbpages boot_cpu_has(X86_FEATURE_GBPAGES) | 187 | #define cpu_has_gbpages boot_cpu_has(X86_FEATURE_GBPAGES) |
| 188 | #define cpu_has_arch_perfmon boot_cpu_has(X86_FEATURE_ARCH_PERFMON) | ||
| 189 | #define cpu_has_pat boot_cpu_has(X86_FEATURE_PAT) | ||
| 184 | 190 | ||
| 185 | #if defined(CONFIG_X86_INVLPG) || defined(CONFIG_X86_64) | 191 | #if defined(CONFIG_X86_INVLPG) || defined(CONFIG_X86_64) |
| 186 | # define cpu_has_invlpg 1 | 192 | # define cpu_has_invlpg 1 |
diff --git a/include/asm-x86/current_32.h b/include/asm-x86/current_32.h index d35248539912..5af9bdb97a16 100644 --- a/include/asm-x86/current_32.h +++ b/include/asm-x86/current_32.h | |||
| @@ -11,7 +11,7 @@ static __always_inline struct task_struct *get_current(void) | |||
| 11 | { | 11 | { |
| 12 | return x86_read_percpu(current_task); | 12 | return x86_read_percpu(current_task); |
| 13 | } | 13 | } |
| 14 | 14 | ||
| 15 | #define current get_current() | 15 | #define current get_current() |
| 16 | 16 | ||
| 17 | #endif /* !(_I386_CURRENT_H) */ | 17 | #endif /* !(_I386_CURRENT_H) */ |
diff --git a/include/asm-x86/current_64.h b/include/asm-x86/current_64.h index bc8adecee66d..2d368ede2fc1 100644 --- a/include/asm-x86/current_64.h +++ b/include/asm-x86/current_64.h | |||
| @@ -1,23 +1,23 @@ | |||
| 1 | #ifndef _X86_64_CURRENT_H | 1 | #ifndef _X86_64_CURRENT_H |
| 2 | #define _X86_64_CURRENT_H | 2 | #define _X86_64_CURRENT_H |
| 3 | 3 | ||
| 4 | #if !defined(__ASSEMBLY__) | 4 | #if !defined(__ASSEMBLY__) |
| 5 | struct task_struct; | 5 | struct task_struct; |
| 6 | 6 | ||
| 7 | #include <asm/pda.h> | 7 | #include <asm/pda.h> |
| 8 | 8 | ||
| 9 | static inline struct task_struct *get_current(void) | 9 | static inline struct task_struct *get_current(void) |
| 10 | { | 10 | { |
| 11 | struct task_struct *t = read_pda(pcurrent); | 11 | struct task_struct *t = read_pda(pcurrent); |
| 12 | return t; | 12 | return t; |
| 13 | } | 13 | } |
| 14 | 14 | ||
| 15 | #define current get_current() | 15 | #define current get_current() |
| 16 | 16 | ||
| 17 | #else | 17 | #else |
| 18 | 18 | ||
| 19 | #ifndef ASM_OFFSET_H | 19 | #ifndef ASM_OFFSET_H |
| 20 | #include <asm/asm-offsets.h> | 20 | #include <asm/asm-offsets.h> |
| 21 | #endif | 21 | #endif |
| 22 | 22 | ||
| 23 | #define GET_CURRENT(reg) movq %gs:(pda_pcurrent),reg | 23 | #define GET_CURRENT(reg) movq %gs:(pda_pcurrent),reg |
diff --git a/include/asm-x86/desc.h b/include/asm-x86/desc.h index 5b6a05d3a771..268a012bcd79 100644 --- a/include/asm-x86/desc.h +++ b/include/asm-x86/desc.h | |||
| @@ -62,8 +62,8 @@ static inline struct desc_struct *get_cpu_gdt_table(unsigned int cpu) | |||
| 62 | } | 62 | } |
| 63 | 63 | ||
| 64 | static inline void pack_gate(gate_desc *gate, unsigned char type, | 64 | static inline void pack_gate(gate_desc *gate, unsigned char type, |
| 65 | unsigned long base, unsigned dpl, unsigned flags, unsigned short seg) | 65 | unsigned long base, unsigned dpl, unsigned flags, |
| 66 | 66 | unsigned short seg) | |
| 67 | { | 67 | { |
| 68 | gate->a = (seg << 16) | (base & 0xffff); | 68 | gate->a = (seg << 16) | (base & 0xffff); |
| 69 | gate->b = (base & 0xffff0000) | | 69 | gate->b = (base & 0xffff0000) | |
| @@ -84,22 +84,23 @@ static inline int desc_empty(const void *ptr) | |||
| 84 | #define load_TR_desc() native_load_tr_desc() | 84 | #define load_TR_desc() native_load_tr_desc() |
| 85 | #define load_gdt(dtr) native_load_gdt(dtr) | 85 | #define load_gdt(dtr) native_load_gdt(dtr) |
| 86 | #define load_idt(dtr) native_load_idt(dtr) | 86 | #define load_idt(dtr) native_load_idt(dtr) |
| 87 | #define load_tr(tr) __asm__ __volatile("ltr %0"::"m" (tr)) | 87 | #define load_tr(tr) asm volatile("ltr %0"::"m" (tr)) |
| 88 | #define load_ldt(ldt) __asm__ __volatile("lldt %0"::"m" (ldt)) | 88 | #define load_ldt(ldt) asm volatile("lldt %0"::"m" (ldt)) |
| 89 | 89 | ||
| 90 | #define store_gdt(dtr) native_store_gdt(dtr) | 90 | #define store_gdt(dtr) native_store_gdt(dtr) |
| 91 | #define store_idt(dtr) native_store_idt(dtr) | 91 | #define store_idt(dtr) native_store_idt(dtr) |
| 92 | #define store_tr(tr) (tr = native_store_tr()) | 92 | #define store_tr(tr) (tr = native_store_tr()) |
| 93 | #define store_ldt(ldt) __asm__ ("sldt %0":"=m" (ldt)) | 93 | #define store_ldt(ldt) asm("sldt %0":"=m" (ldt)) |
| 94 | 94 | ||
| 95 | #define load_TLS(t, cpu) native_load_tls(t, cpu) | 95 | #define load_TLS(t, cpu) native_load_tls(t, cpu) |
| 96 | #define set_ldt native_set_ldt | 96 | #define set_ldt native_set_ldt |
| 97 | 97 | ||
| 98 | #define write_ldt_entry(dt, entry, desc) \ | 98 | #define write_ldt_entry(dt, entry, desc) \ |
| 99 | native_write_ldt_entry(dt, entry, desc) | 99 | native_write_ldt_entry(dt, entry, desc) |
| 100 | #define write_gdt_entry(dt, entry, desc, type) \ | 100 | #define write_gdt_entry(dt, entry, desc, type) \ |
| 101 | native_write_gdt_entry(dt, entry, desc, type) | 101 | native_write_gdt_entry(dt, entry, desc, type) |
| 102 | #define write_idt_entry(dt, entry, g) native_write_idt_entry(dt, entry, g) | 102 | #define write_idt_entry(dt, entry, g) \ |
| 103 | native_write_idt_entry(dt, entry, g) | ||
| 103 | #endif | 104 | #endif |
| 104 | 105 | ||
| 105 | static inline void native_write_idt_entry(gate_desc *idt, int entry, | 106 | static inline void native_write_idt_entry(gate_desc *idt, int entry, |
| @@ -138,8 +139,8 @@ static inline void pack_descriptor(struct desc_struct *desc, unsigned long base, | |||
| 138 | { | 139 | { |
| 139 | desc->a = ((base & 0xffff) << 16) | (limit & 0xffff); | 140 | desc->a = ((base & 0xffff) << 16) | (limit & 0xffff); |
| 140 | desc->b = (base & 0xff000000) | ((base & 0xff0000) >> 16) | | 141 | desc->b = (base & 0xff000000) | ((base & 0xff0000) >> 16) | |
| 141 | (limit & 0x000f0000) | ((type & 0xff) << 8) | | 142 | (limit & 0x000f0000) | ((type & 0xff) << 8) | |
| 142 | ((flags & 0xf) << 20); | 143 | ((flags & 0xf) << 20); |
| 143 | desc->p = 1; | 144 | desc->p = 1; |
| 144 | } | 145 | } |
| 145 | 146 | ||
| @@ -159,7 +160,6 @@ static inline void set_tssldt_descriptor(void *d, unsigned long addr, | |||
| 159 | desc->base2 = (PTR_MIDDLE(addr) >> 8) & 0xFF; | 160 | desc->base2 = (PTR_MIDDLE(addr) >> 8) & 0xFF; |
| 160 | desc->base3 = PTR_HIGH(addr); | 161 | desc->base3 = PTR_HIGH(addr); |
| 161 | #else | 162 | #else |
| 162 | |||
| 163 | pack_descriptor((struct desc_struct *)d, addr, size, 0x80 | type, 0); | 163 | pack_descriptor((struct desc_struct *)d, addr, size, 0x80 | type, 0); |
| 164 | #endif | 164 | #endif |
| 165 | } | 165 | } |
| @@ -177,7 +177,8 @@ static inline void __set_tss_desc(unsigned cpu, unsigned int entry, void *addr) | |||
| 177 | * last valid byte | 177 | * last valid byte |
| 178 | */ | 178 | */ |
| 179 | set_tssldt_descriptor(&tss, (unsigned long)addr, DESC_TSS, | 179 | set_tssldt_descriptor(&tss, (unsigned long)addr, DESC_TSS, |
| 180 | IO_BITMAP_OFFSET + IO_BITMAP_BYTES + sizeof(unsigned long) - 1); | 180 | IO_BITMAP_OFFSET + IO_BITMAP_BYTES + |
| 181 | sizeof(unsigned long) - 1); | ||
| 181 | write_gdt_entry(d, entry, &tss, DESC_TSS); | 182 | write_gdt_entry(d, entry, &tss, DESC_TSS); |
| 182 | } | 183 | } |
| 183 | 184 | ||
| @@ -186,7 +187,7 @@ static inline void __set_tss_desc(unsigned cpu, unsigned int entry, void *addr) | |||
| 186 | static inline void native_set_ldt(const void *addr, unsigned int entries) | 187 | static inline void native_set_ldt(const void *addr, unsigned int entries) |
| 187 | { | 188 | { |
| 188 | if (likely(entries == 0)) | 189 | if (likely(entries == 0)) |
| 189 | __asm__ __volatile__("lldt %w0"::"q" (0)); | 190 | asm volatile("lldt %w0"::"q" (0)); |
| 190 | else { | 191 | else { |
| 191 | unsigned cpu = smp_processor_id(); | 192 | unsigned cpu = smp_processor_id(); |
| 192 | ldt_desc ldt; | 193 | ldt_desc ldt; |
| @@ -195,7 +196,7 @@ static inline void native_set_ldt(const void *addr, unsigned int entries) | |||
| 195 | DESC_LDT, entries * sizeof(ldt) - 1); | 196 | DESC_LDT, entries * sizeof(ldt) - 1); |
| 196 | write_gdt_entry(get_cpu_gdt_table(cpu), GDT_ENTRY_LDT, | 197 | write_gdt_entry(get_cpu_gdt_table(cpu), GDT_ENTRY_LDT, |
| 197 | &ldt, DESC_LDT); | 198 | &ldt, DESC_LDT); |
| 198 | __asm__ __volatile__("lldt %w0"::"q" (GDT_ENTRY_LDT*8)); | 199 | asm volatile("lldt %w0"::"q" (GDT_ENTRY_LDT*8)); |
| 199 | } | 200 | } |
| 200 | } | 201 | } |
| 201 | 202 | ||
| @@ -240,15 +241,15 @@ static inline void native_load_tls(struct thread_struct *t, unsigned int cpu) | |||
| 240 | gdt[GDT_ENTRY_TLS_MIN + i] = t->tls_array[i]; | 241 | gdt[GDT_ENTRY_TLS_MIN + i] = t->tls_array[i]; |
| 241 | } | 242 | } |
| 242 | 243 | ||
| 243 | #define _LDT_empty(info) (\ | 244 | #define _LDT_empty(info) \ |
| 244 | (info)->base_addr == 0 && \ | 245 | ((info)->base_addr == 0 && \ |
| 245 | (info)->limit == 0 && \ | 246 | (info)->limit == 0 && \ |
| 246 | (info)->contents == 0 && \ | 247 | (info)->contents == 0 && \ |
| 247 | (info)->read_exec_only == 1 && \ | 248 | (info)->read_exec_only == 1 && \ |
| 248 | (info)->seg_32bit == 0 && \ | 249 | (info)->seg_32bit == 0 && \ |
| 249 | (info)->limit_in_pages == 0 && \ | 250 | (info)->limit_in_pages == 0 && \ |
| 250 | (info)->seg_not_present == 1 && \ | 251 | (info)->seg_not_present == 1 && \ |
| 251 | (info)->useable == 0) | 252 | (info)->useable == 0) |
| 252 | 253 | ||
| 253 | #ifdef CONFIG_X86_64 | 254 | #ifdef CONFIG_X86_64 |
| 254 | #define LDT_empty(info) (_LDT_empty(info) && ((info)->lm == 0)) | 255 | #define LDT_empty(info) (_LDT_empty(info) && ((info)->lm == 0)) |
| @@ -287,7 +288,7 @@ static inline unsigned long get_desc_limit(const struct desc_struct *desc) | |||
| 287 | } | 288 | } |
| 288 | 289 | ||
| 289 | static inline void _set_gate(int gate, unsigned type, void *addr, | 290 | static inline void _set_gate(int gate, unsigned type, void *addr, |
| 290 | unsigned dpl, unsigned ist, unsigned seg) | 291 | unsigned dpl, unsigned ist, unsigned seg) |
| 291 | { | 292 | { |
| 292 | gate_desc s; | 293 | gate_desc s; |
| 293 | pack_gate(&s, type, (unsigned long)addr, dpl, ist, seg); | 294 | pack_gate(&s, type, (unsigned long)addr, dpl, ist, seg); |
| @@ -370,10 +371,10 @@ static inline void set_system_gate_ist(int n, void *addr, unsigned ist) | |||
| 370 | * Will read the base address of GDT_ENTRY_ESPFIX_SS and put it into %eax. | 371 | * Will read the base address of GDT_ENTRY_ESPFIX_SS and put it into %eax. |
| 371 | */ | 372 | */ |
| 372 | #define GET_DESC_BASE(idx, gdt, base, lo_w, lo_b, hi_b) \ | 373 | #define GET_DESC_BASE(idx, gdt, base, lo_w, lo_b, hi_b) \ |
| 373 | movb idx*8+4(gdt), lo_b; \ | 374 | movb idx * 8 + 4(gdt), lo_b; \ |
| 374 | movb idx*8+7(gdt), hi_b; \ | 375 | movb idx * 8 + 7(gdt), hi_b; \ |
| 375 | shll $16, base; \ | 376 | shll $16, base; \ |
| 376 | movw idx*8+2(gdt), lo_w; | 377 | movw idx * 8 + 2(gdt), lo_w; |
| 377 | 378 | ||
| 378 | 379 | ||
| 379 | #endif /* __ASSEMBLY__ */ | 380 | #endif /* __ASSEMBLY__ */ |
diff --git a/include/asm-x86/desc_defs.h b/include/asm-x86/desc_defs.h index e33f078b3e54..eccb4ea1f918 100644 --- a/include/asm-x86/desc_defs.h +++ b/include/asm-x86/desc_defs.h | |||
| @@ -18,17 +18,19 @@ | |||
| 18 | * incrementally. We keep the signature as a struct, rather than an union, | 18 | * incrementally. We keep the signature as a struct, rather than an union, |
| 19 | * so we can get rid of it transparently in the future -- glommer | 19 | * so we can get rid of it transparently in the future -- glommer |
| 20 | */ | 20 | */ |
| 21 | // 8 byte segment descriptor | 21 | /* 8 byte segment descriptor */ |
| 22 | struct desc_struct { | 22 | struct desc_struct { |
| 23 | union { | 23 | union { |
| 24 | struct { unsigned int a, b; }; | 24 | struct { |
| 25 | unsigned int a; | ||
| 26 | unsigned int b; | ||
| 27 | }; | ||
| 25 | struct { | 28 | struct { |
| 26 | u16 limit0; | 29 | u16 limit0; |
| 27 | u16 base0; | 30 | u16 base0; |
| 28 | unsigned base1: 8, type: 4, s: 1, dpl: 2, p: 1; | 31 | unsigned base1: 8, type: 4, s: 1, dpl: 2, p: 1; |
| 29 | unsigned limit: 4, avl: 1, l: 1, d: 1, g: 1, base2: 8; | 32 | unsigned limit: 4, avl: 1, l: 1, d: 1, g: 1, base2: 8; |
| 30 | }; | 33 | }; |
| 31 | |||
| 32 | }; | 34 | }; |
| 33 | } __attribute__((packed)); | 35 | } __attribute__((packed)); |
| 34 | 36 | ||
| @@ -39,7 +41,7 @@ enum { | |||
| 39 | GATE_TASK = 0x5, | 41 | GATE_TASK = 0x5, |
| 40 | }; | 42 | }; |
| 41 | 43 | ||
| 42 | // 16byte gate | 44 | /* 16byte gate */ |
| 43 | struct gate_struct64 { | 45 | struct gate_struct64 { |
| 44 | u16 offset_low; | 46 | u16 offset_low; |
| 45 | u16 segment; | 47 | u16 segment; |
| @@ -56,10 +58,10 @@ struct gate_struct64 { | |||
| 56 | enum { | 58 | enum { |
| 57 | DESC_TSS = 0x9, | 59 | DESC_TSS = 0x9, |
| 58 | DESC_LDT = 0x2, | 60 | DESC_LDT = 0x2, |
| 59 | DESCTYPE_S = 0x10, /* !system */ | 61 | DESCTYPE_S = 0x10, /* !system */ |
| 60 | }; | 62 | }; |
| 61 | 63 | ||
| 62 | // LDT or TSS descriptor in the GDT. 16 bytes. | 64 | /* LDT or TSS descriptor in the GDT. 16 bytes. */ |
| 63 | struct ldttss_desc64 { | 65 | struct ldttss_desc64 { |
| 64 | u16 limit0; | 66 | u16 limit0; |
| 65 | u16 base0; | 67 | u16 base0; |
| @@ -84,7 +86,6 @@ struct desc_ptr { | |||
| 84 | unsigned long address; | 86 | unsigned long address; |
| 85 | } __attribute__((packed)) ; | 87 | } __attribute__((packed)) ; |
| 86 | 88 | ||
| 87 | |||
| 88 | #endif /* !__ASSEMBLY__ */ | 89 | #endif /* !__ASSEMBLY__ */ |
| 89 | 90 | ||
| 90 | #endif | 91 | #endif |
diff --git a/include/asm-x86/div64.h b/include/asm-x86/div64.h index e98d16e7a37a..0dbf8bf3ef0a 100644 --- a/include/asm-x86/div64.h +++ b/include/asm-x86/div64.h | |||
| @@ -17,18 +17,20 @@ | |||
| 17 | * This ends up being the most efficient "calling | 17 | * This ends up being the most efficient "calling |
| 18 | * convention" on x86. | 18 | * convention" on x86. |
| 19 | */ | 19 | */ |
| 20 | #define do_div(n,base) ({ \ | 20 | #define do_div(n, base) \ |
| 21 | unsigned long __upper, __low, __high, __mod, __base; \ | 21 | ({ \ |
| 22 | __base = (base); \ | 22 | unsigned long __upper, __low, __high, __mod, __base; \ |
| 23 | asm("":"=a" (__low), "=d" (__high):"A" (n)); \ | 23 | __base = (base); \ |
| 24 | __upper = __high; \ | 24 | asm("":"=a" (__low), "=d" (__high) : "A" (n)); \ |
| 25 | if (__high) { \ | 25 | __upper = __high; \ |
| 26 | __upper = __high % (__base); \ | 26 | if (__high) { \ |
| 27 | __high = __high / (__base); \ | 27 | __upper = __high % (__base); \ |
| 28 | } \ | 28 | __high = __high / (__base); \ |
| 29 | asm("divl %2":"=a" (__low), "=d" (__mod):"rm" (__base), "0" (__low), "1" (__upper)); \ | 29 | } \ |
| 30 | asm("":"=A" (n):"a" (__low),"d" (__high)); \ | 30 | asm("divl %2":"=a" (__low), "=d" (__mod) \ |
| 31 | __mod; \ | 31 | : "rm" (__base), "0" (__low), "1" (__upper)); \ |
| 32 | asm("":"=A" (n) : "a" (__low), "d" (__high)); \ | ||
| 33 | __mod; \ | ||
| 32 | }) | 34 | }) |
| 33 | 35 | ||
| 34 | /* | 36 | /* |
| @@ -37,14 +39,13 @@ | |||
| 37 | * | 39 | * |
| 38 | * Warning, this will do an exception if X overflows. | 40 | * Warning, this will do an exception if X overflows. |
| 39 | */ | 41 | */ |
| 40 | #define div_long_long_rem(a,b,c) div_ll_X_l_rem(a,b,c) | 42 | #define div_long_long_rem(a, b, c) div_ll_X_l_rem(a, b, c) |
| 41 | 43 | ||
| 42 | static inline long | 44 | static inline long div_ll_X_l_rem(long long divs, long div, long *rem) |
| 43 | div_ll_X_l_rem(long long divs, long div, long *rem) | ||
| 44 | { | 45 | { |
| 45 | long dum2; | 46 | long dum2; |
| 46 | __asm__("divl %2":"=a"(dum2), "=d"(*rem) | 47 | asm("divl %2":"=a"(dum2), "=d"(*rem) |
| 47 | : "rm"(div), "A"(divs)); | 48 | : "rm"(div), "A"(divs)); |
| 48 | 49 | ||
| 49 | return dum2; | 50 | return dum2; |
| 50 | 51 | ||
diff --git a/include/asm-x86/dma.h b/include/asm-x86/dma.h index e9733ce89880..ca1098a7e580 100644 --- a/include/asm-x86/dma.h +++ b/include/asm-x86/dma.h | |||
| @@ -12,7 +12,6 @@ | |||
| 12 | #include <asm/io.h> /* need byte IO */ | 12 | #include <asm/io.h> /* need byte IO */ |
| 13 | #include <linux/delay.h> | 13 | #include <linux/delay.h> |
| 14 | 14 | ||
| 15 | |||
| 16 | #ifdef HAVE_REALLY_SLOW_DMA_CONTROLLER | 15 | #ifdef HAVE_REALLY_SLOW_DMA_CONTROLLER |
| 17 | #define dma_outb outb_p | 16 | #define dma_outb outb_p |
| 18 | #else | 17 | #else |
| @@ -74,15 +73,15 @@ | |||
| 74 | #ifdef CONFIG_X86_32 | 73 | #ifdef CONFIG_X86_32 |
| 75 | 74 | ||
| 76 | /* The maximum address that we can perform a DMA transfer to on this platform */ | 75 | /* The maximum address that we can perform a DMA transfer to on this platform */ |
| 77 | #define MAX_DMA_ADDRESS (PAGE_OFFSET+0x1000000) | 76 | #define MAX_DMA_ADDRESS (PAGE_OFFSET + 0x1000000) |
| 78 | 77 | ||
| 79 | #else | 78 | #else |
| 80 | 79 | ||
| 81 | /* 16MB ISA DMA zone */ | 80 | /* 16MB ISA DMA zone */ |
| 82 | #define MAX_DMA_PFN ((16*1024*1024) >> PAGE_SHIFT) | 81 | #define MAX_DMA_PFN ((16 * 1024 * 1024) >> PAGE_SHIFT) |
| 83 | 82 | ||
| 84 | /* 4GB broken PCI/AGP hardware bus master zone */ | 83 | /* 4GB broken PCI/AGP hardware bus master zone */ |
| 85 | #define MAX_DMA32_PFN ((4UL*1024*1024*1024) >> PAGE_SHIFT) | 84 | #define MAX_DMA32_PFN ((4UL * 1024 * 1024 * 1024) >> PAGE_SHIFT) |
| 86 | 85 | ||
| 87 | /* Compat define for old dma zone */ | 86 | /* Compat define for old dma zone */ |
| 88 | #define MAX_DMA_ADDRESS ((unsigned long)__va(MAX_DMA_PFN << PAGE_SHIFT)) | 87 | #define MAX_DMA_ADDRESS ((unsigned long)__va(MAX_DMA_PFN << PAGE_SHIFT)) |
| @@ -154,20 +153,20 @@ | |||
| 154 | 153 | ||
| 155 | extern spinlock_t dma_spin_lock; | 154 | extern spinlock_t dma_spin_lock; |
| 156 | 155 | ||
| 157 | static __inline__ unsigned long claim_dma_lock(void) | 156 | static inline unsigned long claim_dma_lock(void) |
| 158 | { | 157 | { |
| 159 | unsigned long flags; | 158 | unsigned long flags; |
| 160 | spin_lock_irqsave(&dma_spin_lock, flags); | 159 | spin_lock_irqsave(&dma_spin_lock, flags); |
| 161 | return flags; | 160 | return flags; |
| 162 | } | 161 | } |
| 163 | 162 | ||
| 164 | static __inline__ void release_dma_lock(unsigned long flags) | 163 | static inline void release_dma_lock(unsigned long flags) |
| 165 | { | 164 | { |
| 166 | spin_unlock_irqrestore(&dma_spin_lock, flags); | 165 | spin_unlock_irqrestore(&dma_spin_lock, flags); |
| 167 | } | 166 | } |
| 168 | 167 | ||
| 169 | /* enable/disable a specific DMA channel */ | 168 | /* enable/disable a specific DMA channel */ |
| 170 | static __inline__ void enable_dma(unsigned int dmanr) | 169 | static inline void enable_dma(unsigned int dmanr) |
| 171 | { | 170 | { |
| 172 | if (dmanr <= 3) | 171 | if (dmanr <= 3) |
| 173 | dma_outb(dmanr, DMA1_MASK_REG); | 172 | dma_outb(dmanr, DMA1_MASK_REG); |
| @@ -175,7 +174,7 @@ static __inline__ void enable_dma(unsigned int dmanr) | |||
| 175 | dma_outb(dmanr & 3, DMA2_MASK_REG); | 174 | dma_outb(dmanr & 3, DMA2_MASK_REG); |
| 176 | } | 175 | } |
| 177 | 176 | ||
| 178 | static __inline__ void disable_dma(unsigned int dmanr) | 177 | static inline void disable_dma(unsigned int dmanr) |
| 179 | { | 178 | { |
| 180 | if (dmanr <= 3) | 179 | if (dmanr <= 3) |
| 181 | dma_outb(dmanr | 4, DMA1_MASK_REG); | 180 | dma_outb(dmanr | 4, DMA1_MASK_REG); |
| @@ -190,7 +189,7 @@ static __inline__ void disable_dma(unsigned int dmanr) | |||
| 190 | * --- In order to do that, the DMA routines below should --- | 189 | * --- In order to do that, the DMA routines below should --- |
| 191 | * --- only be used while holding the DMA lock ! --- | 190 | * --- only be used while holding the DMA lock ! --- |
| 192 | */ | 191 | */ |
| 193 | static __inline__ void clear_dma_ff(unsigned int dmanr) | 192 | static inline void clear_dma_ff(unsigned int dmanr) |
| 194 | { | 193 | { |
| 195 | if (dmanr <= 3) | 194 | if (dmanr <= 3) |
| 196 | dma_outb(0, DMA1_CLEAR_FF_REG); | 195 | dma_outb(0, DMA1_CLEAR_FF_REG); |
| @@ -199,7 +198,7 @@ static __inline__ void clear_dma_ff(unsigned int dmanr) | |||
| 199 | } | 198 | } |
| 200 | 199 | ||
| 201 | /* set mode (above) for a specific DMA channel */ | 200 | /* set mode (above) for a specific DMA channel */ |
| 202 | static __inline__ void set_dma_mode(unsigned int dmanr, char mode) | 201 | static inline void set_dma_mode(unsigned int dmanr, char mode) |
| 203 | { | 202 | { |
| 204 | if (dmanr <= 3) | 203 | if (dmanr <= 3) |
| 205 | dma_outb(mode | dmanr, DMA1_MODE_REG); | 204 | dma_outb(mode | dmanr, DMA1_MODE_REG); |
| @@ -212,7 +211,7 @@ static __inline__ void set_dma_mode(unsigned int dmanr, char mode) | |||
| 212 | * the lower 16 bits of the DMA current address register, but a 64k boundary | 211 | * the lower 16 bits of the DMA current address register, but a 64k boundary |
| 213 | * may have been crossed. | 212 | * may have been crossed. |
| 214 | */ | 213 | */ |
| 215 | static __inline__ void set_dma_page(unsigned int dmanr, char pagenr) | 214 | static inline void set_dma_page(unsigned int dmanr, char pagenr) |
| 216 | { | 215 | { |
| 217 | switch (dmanr) { | 216 | switch (dmanr) { |
| 218 | case 0: | 217 | case 0: |
| @@ -243,15 +242,15 @@ static __inline__ void set_dma_page(unsigned int dmanr, char pagenr) | |||
| 243 | /* Set transfer address & page bits for specific DMA channel. | 242 | /* Set transfer address & page bits for specific DMA channel. |
| 244 | * Assumes dma flipflop is clear. | 243 | * Assumes dma flipflop is clear. |
| 245 | */ | 244 | */ |
| 246 | static __inline__ void set_dma_addr(unsigned int dmanr, unsigned int a) | 245 | static inline void set_dma_addr(unsigned int dmanr, unsigned int a) |
| 247 | { | 246 | { |
| 248 | set_dma_page(dmanr, a>>16); | 247 | set_dma_page(dmanr, a>>16); |
| 249 | if (dmanr <= 3) { | 248 | if (dmanr <= 3) { |
| 250 | dma_outb(a & 0xff, ((dmanr & 3) << 1) + IO_DMA1_BASE); | 249 | dma_outb(a & 0xff, ((dmanr & 3) << 1) + IO_DMA1_BASE); |
| 251 | dma_outb((a >> 8) & 0xff, ((dmanr & 3) << 1) + IO_DMA1_BASE); | 250 | dma_outb((a >> 8) & 0xff, ((dmanr & 3) << 1) + IO_DMA1_BASE); |
| 252 | } else { | 251 | } else { |
| 253 | dma_outb((a >> 1) & 0xff, ((dmanr & 3) << 2) + IO_DMA2_BASE); | 252 | dma_outb((a >> 1) & 0xff, ((dmanr & 3) << 2) + IO_DMA2_BASE); |
| 254 | dma_outb((a >> 9) & 0xff, ((dmanr & 3) << 2) + IO_DMA2_BASE); | 253 | dma_outb((a >> 9) & 0xff, ((dmanr & 3) << 2) + IO_DMA2_BASE); |
| 255 | } | 254 | } |
| 256 | } | 255 | } |
| 257 | 256 | ||
| @@ -264,18 +263,18 @@ static __inline__ void set_dma_addr(unsigned int dmanr, unsigned int a) | |||
| 264 | * Assumes dma flip-flop is clear. | 263 | * Assumes dma flip-flop is clear. |
| 265 | * NOTE 2: "count" represents _bytes_ and must be even for channels 5-7. | 264 | * NOTE 2: "count" represents _bytes_ and must be even for channels 5-7. |
| 266 | */ | 265 | */ |
| 267 | static __inline__ void set_dma_count(unsigned int dmanr, unsigned int count) | 266 | static inline void set_dma_count(unsigned int dmanr, unsigned int count) |
| 268 | { | 267 | { |
| 269 | count--; | 268 | count--; |
| 270 | if (dmanr <= 3) { | 269 | if (dmanr <= 3) { |
| 271 | dma_outb(count & 0xff, ((dmanr & 3) << 1) + 1 + IO_DMA1_BASE); | 270 | dma_outb(count & 0xff, ((dmanr & 3) << 1) + 1 + IO_DMA1_BASE); |
| 272 | dma_outb((count >> 8) & 0xff, | 271 | dma_outb((count >> 8) & 0xff, |
| 273 | ((dmanr & 3) << 1) + 1 + IO_DMA1_BASE); | 272 | ((dmanr & 3) << 1) + 1 + IO_DMA1_BASE); |
| 274 | } else { | 273 | } else { |
| 275 | dma_outb((count >> 1) & 0xff, | 274 | dma_outb((count >> 1) & 0xff, |
| 276 | ((dmanr & 3) << 2) + 2 + IO_DMA2_BASE); | 275 | ((dmanr & 3) << 2) + 2 + IO_DMA2_BASE); |
| 277 | dma_outb((count >> 9) & 0xff, | 276 | dma_outb((count >> 9) & 0xff, |
| 278 | ((dmanr & 3) << 2) + 2 + IO_DMA2_BASE); | 277 | ((dmanr & 3) << 2) + 2 + IO_DMA2_BASE); |
| 279 | } | 278 | } |
| 280 | } | 279 | } |
| 281 | 280 | ||
| @@ -288,7 +287,7 @@ static __inline__ void set_dma_count(unsigned int dmanr, unsigned int count) | |||
| 288 | * | 287 | * |
| 289 | * Assumes DMA flip-flop is clear. | 288 | * Assumes DMA flip-flop is clear. |
| 290 | */ | 289 | */ |
| 291 | static __inline__ int get_dma_residue(unsigned int dmanr) | 290 | static inline int get_dma_residue(unsigned int dmanr) |
| 292 | { | 291 | { |
| 293 | unsigned int io_port; | 292 | unsigned int io_port; |
| 294 | /* using short to get 16-bit wrap around */ | 293 | /* using short to get 16-bit wrap around */ |
diff --git a/include/asm-x86/dwarf2_64.h b/include/asm-x86/dwarf2_64.h index eedc08526b0b..c950519a264d 100644 --- a/include/asm-x86/dwarf2_64.h +++ b/include/asm-x86/dwarf2_64.h | |||
| @@ -1,16 +1,15 @@ | |||
| 1 | #ifndef _DWARF2_H | 1 | #ifndef _DWARF2_H |
| 2 | #define _DWARF2_H 1 | 2 | #define _DWARF2_H 1 |
| 3 | 3 | ||
| 4 | |||
| 5 | #ifndef __ASSEMBLY__ | 4 | #ifndef __ASSEMBLY__ |
| 6 | #warning "asm/dwarf2.h should be only included in pure assembly files" | 5 | #warning "asm/dwarf2.h should be only included in pure assembly files" |
| 7 | #endif | 6 | #endif |
| 8 | 7 | ||
| 9 | /* | 8 | /* |
| 10 | Macros for dwarf2 CFI unwind table entries. | 9 | Macros for dwarf2 CFI unwind table entries. |
| 11 | See "as.info" for details on these pseudo ops. Unfortunately | 10 | See "as.info" for details on these pseudo ops. Unfortunately |
| 12 | they are only supported in very new binutils, so define them | 11 | they are only supported in very new binutils, so define them |
| 13 | away for older version. | 12 | away for older version. |
| 14 | */ | 13 | */ |
| 15 | 14 | ||
| 16 | #ifdef CONFIG_AS_CFI | 15 | #ifdef CONFIG_AS_CFI |
diff --git a/include/asm-x86/e820_32.h b/include/asm-x86/e820_32.h index e7207a6de3e0..43b1a8bd4b34 100644 --- a/include/asm-x86/e820_32.h +++ b/include/asm-x86/e820_32.h | |||
| @@ -34,8 +34,8 @@ extern void e820_register_memory(void); | |||
| 34 | extern void limit_regions(unsigned long long size); | 34 | extern void limit_regions(unsigned long long size); |
| 35 | extern void print_memory_map(char *who); | 35 | extern void print_memory_map(char *who); |
| 36 | extern void init_iomem_resources(struct resource *code_resource, | 36 | extern void init_iomem_resources(struct resource *code_resource, |
| 37 | struct resource *data_resource, | 37 | struct resource *data_resource, |
| 38 | struct resource *bss_resource); | 38 | struct resource *bss_resource); |
| 39 | 39 | ||
| 40 | #if defined(CONFIG_PM) && defined(CONFIG_HIBERNATION) | 40 | #if defined(CONFIG_PM) && defined(CONFIG_HIBERNATION) |
| 41 | extern void e820_mark_nosave_regions(void); | 41 | extern void e820_mark_nosave_regions(void); |
diff --git a/include/asm-x86/e820_64.h b/include/asm-x86/e820_64.h index 22ede73ae724..f478c57eb060 100644 --- a/include/asm-x86/e820_64.h +++ b/include/asm-x86/e820_64.h | |||
| @@ -14,20 +14,24 @@ | |||
| 14 | #include <linux/ioport.h> | 14 | #include <linux/ioport.h> |
| 15 | 15 | ||
| 16 | #ifndef __ASSEMBLY__ | 16 | #ifndef __ASSEMBLY__ |
| 17 | extern unsigned long find_e820_area(unsigned long start, unsigned long end, | 17 | extern unsigned long find_e820_area(unsigned long start, unsigned long end, |
| 18 | unsigned size, unsigned long align); | 18 | unsigned long size, unsigned long align); |
| 19 | extern void add_memory_region(unsigned long start, unsigned long size, | 19 | extern unsigned long find_e820_area_size(unsigned long start, |
| 20 | unsigned long *sizep, | ||
| 21 | unsigned long align); | ||
| 22 | extern void add_memory_region(unsigned long start, unsigned long size, | ||
| 20 | int type); | 23 | int type); |
| 21 | extern void update_memory_range(u64 start, u64 size, unsigned old_type, | 24 | extern void update_memory_range(u64 start, u64 size, unsigned old_type, |
| 22 | unsigned new_type); | 25 | unsigned new_type); |
| 23 | extern void setup_memory_region(void); | 26 | extern void setup_memory_region(void); |
| 24 | extern void contig_e820_setup(void); | 27 | extern void contig_e820_setup(void); |
| 25 | extern unsigned long e820_end_of_ram(void); | 28 | extern unsigned long e820_end_of_ram(void); |
| 26 | extern void e820_reserve_resources(struct resource *code_resource, | 29 | extern void e820_reserve_resources(void); |
| 27 | struct resource *data_resource, struct resource *bss_resource); | ||
| 28 | extern void e820_mark_nosave_regions(void); | 30 | extern void e820_mark_nosave_regions(void); |
| 29 | extern int e820_any_mapped(unsigned long start, unsigned long end, unsigned type); | 31 | extern int e820_any_mapped(unsigned long start, unsigned long end, |
| 30 | extern int e820_all_mapped(unsigned long start, unsigned long end, unsigned type); | 32 | unsigned type); |
| 33 | extern int e820_all_mapped(unsigned long start, unsigned long end, | ||
| 34 | unsigned type); | ||
| 31 | extern int e820_any_non_reserved(unsigned long start, unsigned long end); | 35 | extern int e820_any_non_reserved(unsigned long start, unsigned long end); |
| 32 | extern int is_memory_any_valid(unsigned long start, unsigned long end); | 36 | extern int is_memory_any_valid(unsigned long start, unsigned long end); |
| 33 | extern int e820_all_non_reserved(unsigned long start, unsigned long end); | 37 | extern int e820_all_non_reserved(unsigned long start, unsigned long end); |
| @@ -35,8 +39,8 @@ extern int is_memory_all_valid(unsigned long start, unsigned long end); | |||
| 35 | extern unsigned long e820_hole_size(unsigned long start, unsigned long end); | 39 | extern unsigned long e820_hole_size(unsigned long start, unsigned long end); |
| 36 | 40 | ||
| 37 | extern void e820_setup_gap(void); | 41 | extern void e820_setup_gap(void); |
| 38 | extern void e820_register_active_regions(int nid, | 42 | extern void e820_register_active_regions(int nid, unsigned long start_pfn, |
| 39 | unsigned long start_pfn, unsigned long end_pfn); | 43 | unsigned long end_pfn); |
| 40 | 44 | ||
| 41 | extern void finish_e820_parsing(void); | 45 | extern void finish_e820_parsing(void); |
| 42 | 46 | ||
diff --git a/include/asm-x86/edac.h b/include/asm-x86/edac.h index cf3200a745ad..a8088f63a30e 100644 --- a/include/asm-x86/edac.h +++ b/include/asm-x86/edac.h | |||
| @@ -3,7 +3,7 @@ | |||
| 3 | 3 | ||
| 4 | /* ECC atomic, DMA, SMP and interrupt safe scrub function */ | 4 | /* ECC atomic, DMA, SMP and interrupt safe scrub function */ |
| 5 | 5 | ||
| 6 | static __inline__ void atomic_scrub(void *va, u32 size) | 6 | static inline void atomic_scrub(void *va, u32 size) |
| 7 | { | 7 | { |
| 8 | u32 i, *virt_addr = va; | 8 | u32 i, *virt_addr = va; |
| 9 | 9 | ||
| @@ -12,7 +12,7 @@ static __inline__ void atomic_scrub(void *va, u32 size) | |||
| 12 | * are interrupt, DMA and SMP safe. | 12 | * are interrupt, DMA and SMP safe. |
| 13 | */ | 13 | */ |
| 14 | for (i = 0; i < size / 4; i++, virt_addr++) | 14 | for (i = 0; i < size / 4; i++, virt_addr++) |
| 15 | __asm__ __volatile__("lock; addl $0, %0"::"m"(*virt_addr)); | 15 | asm volatile("lock; addl $0, %0"::"m" (*virt_addr)); |
| 16 | } | 16 | } |
| 17 | 17 | ||
| 18 | #endif | 18 | #endif |
diff --git a/include/asm-x86/efi.h b/include/asm-x86/efi.h index ea9734b74aca..d53004b855cc 100644 --- a/include/asm-x86/efi.h +++ b/include/asm-x86/efi.h | |||
| @@ -20,7 +20,7 @@ extern unsigned long asmlinkage efi_call_phys(void *, ...); | |||
| 20 | */ | 20 | */ |
| 21 | 21 | ||
| 22 | #define efi_call_virt(f, args...) \ | 22 | #define efi_call_virt(f, args...) \ |
| 23 | ((efi_##f##_t __attribute__((regparm(0)))*)efi.systab->runtime->f)(args) | 23 | ((efi_##f##_t __attribute__((regparm(0)))*)efi.systab->runtime->f)(args) |
| 24 | 24 | ||
| 25 | #define efi_call_virt0(f) efi_call_virt(f) | 25 | #define efi_call_virt0(f) efi_call_virt(f) |
| 26 | #define efi_call_virt1(f, a1) efi_call_virt(f, a1) | 26 | #define efi_call_virt1(f, a1) efi_call_virt(f, a1) |
diff --git a/include/asm-x86/elf.h b/include/asm-x86/elf.h index fb62f9941e38..8f232dc5b5fe 100644 --- a/include/asm-x86/elf.h +++ b/include/asm-x86/elf.h | |||
| @@ -11,7 +11,7 @@ | |||
| 11 | 11 | ||
| 12 | typedef unsigned long elf_greg_t; | 12 | typedef unsigned long elf_greg_t; |
| 13 | 13 | ||
| 14 | #define ELF_NGREG (sizeof (struct user_regs_struct) / sizeof(elf_greg_t)) | 14 | #define ELF_NGREG (sizeof(struct user_regs_struct) / sizeof(elf_greg_t)) |
| 15 | typedef elf_greg_t elf_gregset_t[ELF_NGREG]; | 15 | typedef elf_greg_t elf_gregset_t[ELF_NGREG]; |
| 16 | 16 | ||
| 17 | typedef struct user_i387_struct elf_fpregset_t; | 17 | typedef struct user_i387_struct elf_fpregset_t; |
| @@ -82,8 +82,9 @@ extern unsigned int vdso_enabled; | |||
| 82 | #define elf_check_arch_ia32(x) \ | 82 | #define elf_check_arch_ia32(x) \ |
| 83 | (((x)->e_machine == EM_386) || ((x)->e_machine == EM_486)) | 83 | (((x)->e_machine == EM_386) || ((x)->e_machine == EM_486)) |
| 84 | 84 | ||
| 85 | #ifdef CONFIG_X86_32 | ||
| 86 | #include <asm/processor.h> | 85 | #include <asm/processor.h> |
| 86 | |||
| 87 | #ifdef CONFIG_X86_32 | ||
| 87 | #include <asm/system.h> /* for savesegment */ | 88 | #include <asm/system.h> /* for savesegment */ |
| 88 | #include <asm/desc.h> | 89 | #include <asm/desc.h> |
| 89 | 90 | ||
| @@ -99,10 +100,11 @@ extern unsigned int vdso_enabled; | |||
| 99 | We might as well make sure everything else is cleared too (except for %esp), | 100 | We might as well make sure everything else is cleared too (except for %esp), |
| 100 | just to make things more deterministic. | 101 | just to make things more deterministic. |
| 101 | */ | 102 | */ |
| 102 | #define ELF_PLAT_INIT(_r, load_addr) do { \ | 103 | #define ELF_PLAT_INIT(_r, load_addr) \ |
| 103 | _r->bx = 0; _r->cx = 0; _r->dx = 0; \ | 104 | do { \ |
| 104 | _r->si = 0; _r->di = 0; _r->bp = 0; \ | 105 | _r->bx = 0; _r->cx = 0; _r->dx = 0; \ |
| 105 | _r->ax = 0; \ | 106 | _r->si = 0; _r->di = 0; _r->bp = 0; \ |
| 107 | _r->ax = 0; \ | ||
| 106 | } while (0) | 108 | } while (0) |
| 107 | 109 | ||
| 108 | /* | 110 | /* |
| @@ -110,24 +112,25 @@ extern unsigned int vdso_enabled; | |||
| 110 | * now struct_user_regs, they are different) | 112 | * now struct_user_regs, they are different) |
| 111 | */ | 113 | */ |
| 112 | 114 | ||
| 113 | #define ELF_CORE_COPY_REGS(pr_reg, regs) do { \ | 115 | #define ELF_CORE_COPY_REGS(pr_reg, regs) \ |
| 114 | pr_reg[0] = regs->bx; \ | 116 | do { \ |
| 115 | pr_reg[1] = regs->cx; \ | 117 | pr_reg[0] = regs->bx; \ |
| 116 | pr_reg[2] = regs->dx; \ | 118 | pr_reg[1] = regs->cx; \ |
| 117 | pr_reg[3] = regs->si; \ | 119 | pr_reg[2] = regs->dx; \ |
| 118 | pr_reg[4] = regs->di; \ | 120 | pr_reg[3] = regs->si; \ |
| 119 | pr_reg[5] = regs->bp; \ | 121 | pr_reg[4] = regs->di; \ |
| 120 | pr_reg[6] = regs->ax; \ | 122 | pr_reg[5] = regs->bp; \ |
| 121 | pr_reg[7] = regs->ds & 0xffff; \ | 123 | pr_reg[6] = regs->ax; \ |
| 122 | pr_reg[8] = regs->es & 0xffff; \ | 124 | pr_reg[7] = regs->ds & 0xffff; \ |
| 123 | pr_reg[9] = regs->fs & 0xffff; \ | 125 | pr_reg[8] = regs->es & 0xffff; \ |
| 124 | savesegment(gs, pr_reg[10]); \ | 126 | pr_reg[9] = regs->fs & 0xffff; \ |
| 125 | pr_reg[11] = regs->orig_ax; \ | 127 | savesegment(gs, pr_reg[10]); \ |
| 126 | pr_reg[12] = regs->ip; \ | 128 | pr_reg[11] = regs->orig_ax; \ |
| 127 | pr_reg[13] = regs->cs & 0xffff; \ | 129 | pr_reg[12] = regs->ip; \ |
| 128 | pr_reg[14] = regs->flags; \ | 130 | pr_reg[13] = regs->cs & 0xffff; \ |
| 129 | pr_reg[15] = regs->sp; \ | 131 | pr_reg[14] = regs->flags; \ |
| 130 | pr_reg[16] = regs->ss & 0xffff; \ | 132 | pr_reg[15] = regs->sp; \ |
| 133 | pr_reg[16] = regs->ss & 0xffff; \ | ||
| 131 | } while (0); | 134 | } while (0); |
| 132 | 135 | ||
| 133 | #define ELF_PLATFORM (utsname()->machine) | 136 | #define ELF_PLATFORM (utsname()->machine) |
| @@ -135,12 +138,10 @@ extern unsigned int vdso_enabled; | |||
| 135 | 138 | ||
| 136 | #else /* CONFIG_X86_32 */ | 139 | #else /* CONFIG_X86_32 */ |
| 137 | 140 | ||
| 138 | #include <asm/processor.h> | ||
| 139 | |||
| 140 | /* | 141 | /* |
| 141 | * This is used to ensure we don't load something for the wrong architecture. | 142 | * This is used to ensure we don't load something for the wrong architecture. |
| 142 | */ | 143 | */ |
| 143 | #define elf_check_arch(x) \ | 144 | #define elf_check_arch(x) \ |
| 144 | ((x)->e_machine == EM_X86_64) | 145 | ((x)->e_machine == EM_X86_64) |
| 145 | 146 | ||
| 146 | #define compat_elf_check_arch(x) elf_check_arch_ia32(x) | 147 | #define compat_elf_check_arch(x) elf_check_arch_ia32(x) |
| @@ -169,24 +170,30 @@ static inline void elf_common_init(struct thread_struct *t, | |||
| 169 | t->ds = t->es = ds; | 170 | t->ds = t->es = ds; |
| 170 | } | 171 | } |
| 171 | 172 | ||
| 172 | #define ELF_PLAT_INIT(_r, load_addr) do { \ | 173 | #define ELF_PLAT_INIT(_r, load_addr) \ |
| 173 | elf_common_init(¤t->thread, _r, 0); \ | 174 | do { \ |
| 174 | clear_thread_flag(TIF_IA32); \ | 175 | elf_common_init(¤t->thread, _r, 0); \ |
| 176 | clear_thread_flag(TIF_IA32); \ | ||
| 175 | } while (0) | 177 | } while (0) |
| 176 | 178 | ||
| 177 | #define COMPAT_ELF_PLAT_INIT(regs, load_addr) \ | 179 | #define COMPAT_ELF_PLAT_INIT(regs, load_addr) \ |
| 178 | elf_common_init(¤t->thread, regs, __USER_DS) | 180 | elf_common_init(¤t->thread, regs, __USER_DS) |
| 179 | #define compat_start_thread(regs, ip, sp) do { \ | 181 | |
| 180 | start_ia32_thread(regs, ip, sp); \ | 182 | #define compat_start_thread(regs, ip, sp) \ |
| 181 | set_fs(USER_DS); \ | 183 | do { \ |
| 182 | } while (0) | 184 | start_ia32_thread(regs, ip, sp); \ |
| 183 | #define COMPAT_SET_PERSONALITY(ex, ibcs2) do { \ | 185 | set_fs(USER_DS); \ |
| 184 | if (test_thread_flag(TIF_IA32)) \ | 186 | } while (0) |
| 185 | clear_thread_flag(TIF_ABI_PENDING); \ | 187 | |
| 186 | else \ | 188 | #define COMPAT_SET_PERSONALITY(ex, ibcs2) \ |
| 187 | set_thread_flag(TIF_ABI_PENDING); \ | 189 | do { \ |
| 188 | current->personality |= force_personality32; \ | 190 | if (test_thread_flag(TIF_IA32)) \ |
| 189 | } while (0) | 191 | clear_thread_flag(TIF_ABI_PENDING); \ |
| 192 | else \ | ||
| 193 | set_thread_flag(TIF_ABI_PENDING); \ | ||
| 194 | current->personality |= force_personality32; \ | ||
| 195 | } while (0) | ||
| 196 | |||
| 190 | #define COMPAT_ELF_PLATFORM ("i686") | 197 | #define COMPAT_ELF_PLATFORM ("i686") |
| 191 | 198 | ||
| 192 | /* | 199 | /* |
| @@ -195,7 +202,8 @@ static inline void elf_common_init(struct thread_struct *t, | |||
| 195 | * getting dumped. | 202 | * getting dumped. |
| 196 | */ | 203 | */ |
| 197 | 204 | ||
| 198 | #define ELF_CORE_COPY_REGS(pr_reg, regs) do { \ | 205 | #define ELF_CORE_COPY_REGS(pr_reg, regs) \ |
| 206 | do { \ | ||
| 199 | unsigned v; \ | 207 | unsigned v; \ |
| 200 | (pr_reg)[0] = (regs)->r15; \ | 208 | (pr_reg)[0] = (regs)->r15; \ |
| 201 | (pr_reg)[1] = (regs)->r14; \ | 209 | (pr_reg)[1] = (regs)->r14; \ |
| @@ -269,10 +277,12 @@ extern int force_personality32; | |||
| 269 | 277 | ||
| 270 | struct task_struct; | 278 | struct task_struct; |
| 271 | 279 | ||
| 272 | #define ARCH_DLINFO_IA32(vdso_enabled) \ | 280 | #define ARCH_DLINFO_IA32(vdso_enabled) \ |
| 273 | do if (vdso_enabled) { \ | 281 | do { \ |
| 282 | if (vdso_enabled) { \ | ||
| 274 | NEW_AUX_ENT(AT_SYSINFO, VDSO_ENTRY); \ | 283 | NEW_AUX_ENT(AT_SYSINFO, VDSO_ENTRY); \ |
| 275 | NEW_AUX_ENT(AT_SYSINFO_EHDR, VDSO_CURRENT_BASE); \ | 284 | NEW_AUX_ENT(AT_SYSINFO_EHDR, VDSO_CURRENT_BASE); \ |
| 285 | } \ | ||
| 276 | } while (0) | 286 | } while (0) |
| 277 | 287 | ||
| 278 | #ifdef CONFIG_X86_32 | 288 | #ifdef CONFIG_X86_32 |
| @@ -290,9 +300,11 @@ do if (vdso_enabled) { \ | |||
| 290 | /* 1GB for 64bit, 8MB for 32bit */ | 300 | /* 1GB for 64bit, 8MB for 32bit */ |
| 291 | #define STACK_RND_MASK (test_thread_flag(TIF_IA32) ? 0x7ff : 0x3fffff) | 301 | #define STACK_RND_MASK (test_thread_flag(TIF_IA32) ? 0x7ff : 0x3fffff) |
| 292 | 302 | ||
| 293 | #define ARCH_DLINFO \ | 303 | #define ARCH_DLINFO \ |
| 294 | do if (vdso_enabled) { \ | 304 | do { \ |
| 295 | NEW_AUX_ENT(AT_SYSINFO_EHDR,(unsigned long)current->mm->context.vdso);\ | 305 | if (vdso_enabled) \ |
| 306 | NEW_AUX_ENT(AT_SYSINFO_EHDR, \ | ||
| 307 | (unsigned long)current->mm->context.vdso); \ | ||
| 296 | } while (0) | 308 | } while (0) |
| 297 | 309 | ||
| 298 | #define AT_SYSINFO 32 | 310 | #define AT_SYSINFO 32 |
| @@ -305,8 +317,8 @@ do if (vdso_enabled) { \ | |||
| 305 | 317 | ||
| 306 | #define VDSO_CURRENT_BASE ((unsigned long)current->mm->context.vdso) | 318 | #define VDSO_CURRENT_BASE ((unsigned long)current->mm->context.vdso) |
| 307 | 319 | ||
| 308 | #define VDSO_ENTRY \ | 320 | #define VDSO_ENTRY \ |
| 309 | ((unsigned long) VDSO32_SYMBOL(VDSO_CURRENT_BASE, vsyscall)) | 321 | ((unsigned long)VDSO32_SYMBOL(VDSO_CURRENT_BASE, vsyscall)) |
| 310 | 322 | ||
| 311 | struct linux_binprm; | 323 | struct linux_binprm; |
| 312 | 324 | ||
diff --git a/include/asm-x86/fixmap_32.h b/include/asm-x86/fixmap_32.h index a7404d50686b..eb1665125c44 100644 --- a/include/asm-x86/fixmap_32.h +++ b/include/asm-x86/fixmap_32.h | |||
| @@ -99,8 +99,7 @@ enum fixed_addresses { | |||
| 99 | */ | 99 | */ |
| 100 | #define NR_FIX_BTMAPS 64 | 100 | #define NR_FIX_BTMAPS 64 |
| 101 | #define FIX_BTMAPS_NESTING 4 | 101 | #define FIX_BTMAPS_NESTING 4 |
| 102 | FIX_BTMAP_END = | 102 | FIX_BTMAP_END = __end_of_permanent_fixed_addresses + 512 - |
| 103 | __end_of_permanent_fixed_addresses + 512 - | ||
| 104 | (__end_of_permanent_fixed_addresses & 511), | 103 | (__end_of_permanent_fixed_addresses & 511), |
| 105 | FIX_BTMAP_BEGIN = FIX_BTMAP_END + NR_FIX_BTMAPS*FIX_BTMAPS_NESTING - 1, | 104 | FIX_BTMAP_BEGIN = FIX_BTMAP_END + NR_FIX_BTMAPS*FIX_BTMAPS_NESTING - 1, |
| 106 | FIX_WP_TEST, | 105 | FIX_WP_TEST, |
| @@ -110,20 +109,20 @@ enum fixed_addresses { | |||
| 110 | __end_of_fixed_addresses | 109 | __end_of_fixed_addresses |
| 111 | }; | 110 | }; |
| 112 | 111 | ||
| 113 | extern void __set_fixmap (enum fixed_addresses idx, | 112 | extern void __set_fixmap(enum fixed_addresses idx, |
| 114 | unsigned long phys, pgprot_t flags); | 113 | unsigned long phys, pgprot_t flags); |
| 115 | extern void reserve_top_address(unsigned long reserve); | 114 | extern void reserve_top_address(unsigned long reserve); |
| 116 | 115 | ||
| 117 | #define set_fixmap(idx, phys) \ | 116 | #define set_fixmap(idx, phys) \ |
| 118 | __set_fixmap(idx, phys, PAGE_KERNEL) | 117 | __set_fixmap(idx, phys, PAGE_KERNEL) |
| 119 | /* | 118 | /* |
| 120 | * Some hardware wants to get fixmapped without caching. | 119 | * Some hardware wants to get fixmapped without caching. |
| 121 | */ | 120 | */ |
| 122 | #define set_fixmap_nocache(idx, phys) \ | 121 | #define set_fixmap_nocache(idx, phys) \ |
| 123 | __set_fixmap(idx, phys, PAGE_KERNEL_NOCACHE) | 122 | __set_fixmap(idx, phys, PAGE_KERNEL_NOCACHE) |
| 124 | 123 | ||
| 125 | #define clear_fixmap(idx) \ | 124 | #define clear_fixmap(idx) \ |
| 126 | __set_fixmap(idx, 0, __pgprot(0)) | 125 | __set_fixmap(idx, 0, __pgprot(0)) |
| 127 | 126 | ||
| 128 | #define FIXADDR_TOP ((unsigned long)__FIXADDR_TOP) | 127 | #define FIXADDR_TOP ((unsigned long)__FIXADDR_TOP) |
| 129 | 128 | ||
| @@ -156,7 +155,7 @@ static __always_inline unsigned long fix_to_virt(const unsigned int idx) | |||
| 156 | if (idx >= __end_of_fixed_addresses) | 155 | if (idx >= __end_of_fixed_addresses) |
| 157 | __this_fixmap_does_not_exist(); | 156 | __this_fixmap_does_not_exist(); |
| 158 | 157 | ||
| 159 | return __fix_to_virt(idx); | 158 | return __fix_to_virt(idx); |
| 160 | } | 159 | } |
| 161 | 160 | ||
| 162 | static inline unsigned long virt_to_fix(const unsigned long vaddr) | 161 | static inline unsigned long virt_to_fix(const unsigned long vaddr) |
diff --git a/include/asm-x86/fixmap_64.h b/include/asm-x86/fixmap_64.h index 70ddb21e6458..f3d76858c0e6 100644 --- a/include/asm-x86/fixmap_64.h +++ b/include/asm-x86/fixmap_64.h | |||
| @@ -34,32 +34,34 @@ | |||
| 34 | 34 | ||
| 35 | enum fixed_addresses { | 35 | enum fixed_addresses { |
| 36 | VSYSCALL_LAST_PAGE, | 36 | VSYSCALL_LAST_PAGE, |
| 37 | VSYSCALL_FIRST_PAGE = VSYSCALL_LAST_PAGE + ((VSYSCALL_END-VSYSCALL_START) >> PAGE_SHIFT) - 1, | 37 | VSYSCALL_FIRST_PAGE = VSYSCALL_LAST_PAGE |
| 38 | + ((VSYSCALL_END-VSYSCALL_START) >> PAGE_SHIFT) - 1, | ||
| 38 | VSYSCALL_HPET, | 39 | VSYSCALL_HPET, |
| 39 | FIX_DBGP_BASE, | 40 | FIX_DBGP_BASE, |
| 40 | FIX_EARLYCON_MEM_BASE, | 41 | FIX_EARLYCON_MEM_BASE, |
| 41 | FIX_HPET_BASE, | 42 | FIX_HPET_BASE, |
| 42 | FIX_APIC_BASE, /* local (CPU) APIC) -- required for SMP or not */ | 43 | FIX_APIC_BASE, /* local (CPU) APIC) -- required for SMP or not */ |
| 43 | FIX_IO_APIC_BASE_0, | 44 | FIX_IO_APIC_BASE_0, |
| 44 | FIX_IO_APIC_BASE_END = FIX_IO_APIC_BASE_0 + MAX_IO_APICS-1, | 45 | FIX_IO_APIC_BASE_END = FIX_IO_APIC_BASE_0 + MAX_IO_APICS - 1, |
| 45 | FIX_EFI_IO_MAP_LAST_PAGE, | 46 | FIX_EFI_IO_MAP_LAST_PAGE, |
| 46 | FIX_EFI_IO_MAP_FIRST_PAGE = FIX_EFI_IO_MAP_LAST_PAGE+MAX_EFI_IO_PAGES-1, | 47 | FIX_EFI_IO_MAP_FIRST_PAGE = FIX_EFI_IO_MAP_LAST_PAGE |
| 48 | + MAX_EFI_IO_PAGES - 1, | ||
| 47 | #ifdef CONFIG_PROVIDE_OHCI1394_DMA_INIT | 49 | #ifdef CONFIG_PROVIDE_OHCI1394_DMA_INIT |
| 48 | FIX_OHCI1394_BASE, | 50 | FIX_OHCI1394_BASE, |
| 49 | #endif | 51 | #endif |
| 50 | __end_of_fixed_addresses | 52 | __end_of_fixed_addresses |
| 51 | }; | 53 | }; |
| 52 | 54 | ||
| 53 | extern void __set_fixmap (enum fixed_addresses idx, | 55 | extern void __set_fixmap(enum fixed_addresses idx, |
| 54 | unsigned long phys, pgprot_t flags); | 56 | unsigned long phys, pgprot_t flags); |
| 55 | 57 | ||
| 56 | #define set_fixmap(idx, phys) \ | 58 | #define set_fixmap(idx, phys) \ |
| 57 | __set_fixmap(idx, phys, PAGE_KERNEL) | 59 | __set_fixmap(idx, phys, PAGE_KERNEL) |
| 58 | /* | 60 | /* |
| 59 | * Some hardware wants to get fixmapped without caching. | 61 | * Some hardware wants to get fixmapped without caching. |
| 60 | */ | 62 | */ |
| 61 | #define set_fixmap_nocache(idx, phys) \ | 63 | #define set_fixmap_nocache(idx, phys) \ |
| 62 | __set_fixmap(idx, phys, PAGE_KERNEL_NOCACHE) | 64 | __set_fixmap(idx, phys, PAGE_KERNEL_NOCACHE) |
| 63 | 65 | ||
| 64 | #define FIXADDR_TOP (VSYSCALL_END-PAGE_SIZE) | 66 | #define FIXADDR_TOP (VSYSCALL_END-PAGE_SIZE) |
| 65 | #define FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT) | 67 | #define FIXADDR_SIZE (__end_of_fixed_addresses << PAGE_SHIFT) |
diff --git a/include/asm-x86/floppy.h b/include/asm-x86/floppy.h index a48d7153c097..dbe82a5c5eac 100644 --- a/include/asm-x86/floppy.h +++ b/include/asm-x86/floppy.h | |||
| @@ -20,20 +20,21 @@ | |||
| 20 | * driver otherwise. It doesn't matter much for performance anyway, as most | 20 | * driver otherwise. It doesn't matter much for performance anyway, as most |
| 21 | * floppy accesses go through the track buffer. | 21 | * floppy accesses go through the track buffer. |
| 22 | */ | 22 | */ |
| 23 | #define _CROSS_64KB(a,s,vdma) \ | 23 | #define _CROSS_64KB(a, s, vdma) \ |
| 24 | (!(vdma) && ((unsigned long)(a)/K_64 != ((unsigned long)(a) + (s) - 1) / K_64)) | 24 | (!(vdma) && \ |
| 25 | ((unsigned long)(a)/K_64 != ((unsigned long)(a) + (s) - 1) / K_64)) | ||
| 25 | 26 | ||
| 26 | #define CROSS_64KB(a,s) _CROSS_64KB(a,s,use_virtual_dma & 1) | 27 | #define CROSS_64KB(a, s) _CROSS_64KB(a, s, use_virtual_dma & 1) |
| 27 | 28 | ||
| 28 | 29 | ||
| 29 | #define SW fd_routine[use_virtual_dma&1] | 30 | #define SW fd_routine[use_virtual_dma & 1] |
| 30 | #define CSW fd_routine[can_use_virtual_dma & 1] | 31 | #define CSW fd_routine[can_use_virtual_dma & 1] |
| 31 | 32 | ||
| 32 | 33 | ||
| 33 | #define fd_inb(port) inb_p(port) | 34 | #define fd_inb(port) inb_p(port) |
| 34 | #define fd_outb(value,port) outb_p(value,port) | 35 | #define fd_outb(value, port) outb_p(value, port) |
| 35 | 36 | ||
| 36 | #define fd_request_dma() CSW._request_dma(FLOPPY_DMA,"floppy") | 37 | #define fd_request_dma() CSW._request_dma(FLOPPY_DMA, "floppy") |
| 37 | #define fd_free_dma() CSW._free_dma(FLOPPY_DMA) | 38 | #define fd_free_dma() CSW._free_dma(FLOPPY_DMA) |
| 38 | #define fd_enable_irq() enable_irq(FLOPPY_IRQ) | 39 | #define fd_enable_irq() enable_irq(FLOPPY_IRQ) |
| 39 | #define fd_disable_irq() disable_irq(FLOPPY_IRQ) | 40 | #define fd_disable_irq() disable_irq(FLOPPY_IRQ) |
| @@ -52,64 +53,64 @@ static int doing_pdma; | |||
| 52 | 53 | ||
| 53 | static irqreturn_t floppy_hardint(int irq, void *dev_id) | 54 | static irqreturn_t floppy_hardint(int irq, void *dev_id) |
| 54 | { | 55 | { |
| 55 | register unsigned char st; | 56 | unsigned char st; |
| 56 | 57 | ||
| 57 | #undef TRACE_FLPY_INT | 58 | #undef TRACE_FLPY_INT |
| 58 | 59 | ||
| 59 | #ifdef TRACE_FLPY_INT | 60 | #ifdef TRACE_FLPY_INT |
| 60 | static int calls=0; | 61 | static int calls; |
| 61 | static int bytes=0; | 62 | static int bytes; |
| 62 | static int dma_wait=0; | 63 | static int dma_wait; |
| 63 | #endif | 64 | #endif |
| 64 | if (!doing_pdma) | 65 | if (!doing_pdma) |
| 65 | return floppy_interrupt(irq, dev_id); | 66 | return floppy_interrupt(irq, dev_id); |
| 66 | 67 | ||
| 67 | #ifdef TRACE_FLPY_INT | 68 | #ifdef TRACE_FLPY_INT |
| 68 | if(!calls) | 69 | if (!calls) |
| 69 | bytes = virtual_dma_count; | 70 | bytes = virtual_dma_count; |
| 70 | #endif | 71 | #endif |
| 71 | 72 | ||
| 72 | { | 73 | { |
| 73 | register int lcount; | 74 | int lcount; |
| 74 | register char *lptr; | 75 | char *lptr; |
| 75 | 76 | ||
| 76 | st = 1; | 77 | st = 1; |
| 77 | for(lcount=virtual_dma_count, lptr=virtual_dma_addr; | 78 | for (lcount = virtual_dma_count, lptr = virtual_dma_addr; |
| 78 | lcount; lcount--, lptr++) { | 79 | lcount; lcount--, lptr++) { |
| 79 | st=inb(virtual_dma_port+4) & 0xa0 ; | 80 | st = inb(virtual_dma_port + 4) & 0xa0; |
| 80 | if(st != 0xa0) | 81 | if (st != 0xa0) |
| 81 | break; | 82 | break; |
| 82 | if(virtual_dma_mode) | 83 | if (virtual_dma_mode) |
| 83 | outb_p(*lptr, virtual_dma_port+5); | 84 | outb_p(*lptr, virtual_dma_port + 5); |
| 84 | else | 85 | else |
| 85 | *lptr = inb_p(virtual_dma_port+5); | 86 | *lptr = inb_p(virtual_dma_port + 5); |
| 86 | } | 87 | } |
| 87 | virtual_dma_count = lcount; | 88 | virtual_dma_count = lcount; |
| 88 | virtual_dma_addr = lptr; | 89 | virtual_dma_addr = lptr; |
| 89 | st = inb(virtual_dma_port+4); | 90 | st = inb(virtual_dma_port + 4); |
| 90 | } | 91 | } |
| 91 | 92 | ||
| 92 | #ifdef TRACE_FLPY_INT | 93 | #ifdef TRACE_FLPY_INT |
| 93 | calls++; | 94 | calls++; |
| 94 | #endif | 95 | #endif |
| 95 | if(st == 0x20) | 96 | if (st == 0x20) |
| 96 | return IRQ_HANDLED; | 97 | return IRQ_HANDLED; |
| 97 | if(!(st & 0x20)) { | 98 | if (!(st & 0x20)) { |
| 98 | virtual_dma_residue += virtual_dma_count; | 99 | virtual_dma_residue += virtual_dma_count; |
| 99 | virtual_dma_count=0; | 100 | virtual_dma_count = 0; |
| 100 | #ifdef TRACE_FLPY_INT | 101 | #ifdef TRACE_FLPY_INT |
| 101 | printk("count=%x, residue=%x calls=%d bytes=%d dma_wait=%d\n", | 102 | printk("count=%x, residue=%x calls=%d bytes=%d dma_wait=%d\n", |
| 102 | virtual_dma_count, virtual_dma_residue, calls, bytes, | 103 | virtual_dma_count, virtual_dma_residue, calls, bytes, |
| 103 | dma_wait); | 104 | dma_wait); |
| 104 | calls = 0; | 105 | calls = 0; |
| 105 | dma_wait=0; | 106 | dma_wait = 0; |
| 106 | #endif | 107 | #endif |
| 107 | doing_pdma = 0; | 108 | doing_pdma = 0; |
| 108 | floppy_interrupt(irq, dev_id); | 109 | floppy_interrupt(irq, dev_id); |
| 109 | return IRQ_HANDLED; | 110 | return IRQ_HANDLED; |
| 110 | } | 111 | } |
| 111 | #ifdef TRACE_FLPY_INT | 112 | #ifdef TRACE_FLPY_INT |
| 112 | if(!virtual_dma_count) | 113 | if (!virtual_dma_count) |
| 113 | dma_wait++; | 114 | dma_wait++; |
| 114 | #endif | 115 | #endif |
| 115 | return IRQ_HANDLED; | 116 | return IRQ_HANDLED; |
| @@ -117,14 +118,14 @@ static irqreturn_t floppy_hardint(int irq, void *dev_id) | |||
| 117 | 118 | ||
| 118 | static void fd_disable_dma(void) | 119 | static void fd_disable_dma(void) |
| 119 | { | 120 | { |
| 120 | if(! (can_use_virtual_dma & 1)) | 121 | if (!(can_use_virtual_dma & 1)) |
| 121 | disable_dma(FLOPPY_DMA); | 122 | disable_dma(FLOPPY_DMA); |
| 122 | doing_pdma = 0; | 123 | doing_pdma = 0; |
| 123 | virtual_dma_residue += virtual_dma_count; | 124 | virtual_dma_residue += virtual_dma_count; |
| 124 | virtual_dma_count=0; | 125 | virtual_dma_count = 0; |
| 125 | } | 126 | } |
| 126 | 127 | ||
| 127 | static int vdma_request_dma(unsigned int dmanr, const char * device_id) | 128 | static int vdma_request_dma(unsigned int dmanr, const char *device_id) |
| 128 | { | 129 | { |
| 129 | return 0; | 130 | return 0; |
| 130 | } | 131 | } |
| @@ -142,7 +143,7 @@ static int vdma_get_dma_residue(unsigned int dummy) | |||
| 142 | 143 | ||
| 143 | static int fd_request_irq(void) | 144 | static int fd_request_irq(void) |
| 144 | { | 145 | { |
| 145 | if(can_use_virtual_dma) | 146 | if (can_use_virtual_dma) |
| 146 | return request_irq(FLOPPY_IRQ, floppy_hardint, | 147 | return request_irq(FLOPPY_IRQ, floppy_hardint, |
| 147 | IRQF_DISABLED, "floppy", NULL); | 148 | IRQF_DISABLED, "floppy", NULL); |
| 148 | else | 149 | else |
| @@ -152,13 +153,13 @@ static int fd_request_irq(void) | |||
| 152 | 153 | ||
| 153 | static unsigned long dma_mem_alloc(unsigned long size) | 154 | static unsigned long dma_mem_alloc(unsigned long size) |
| 154 | { | 155 | { |
| 155 | return __get_dma_pages(GFP_KERNEL|__GFP_NORETRY,get_order(size)); | 156 | return __get_dma_pages(GFP_KERNEL|__GFP_NORETRY, get_order(size)); |
| 156 | } | 157 | } |
| 157 | 158 | ||
| 158 | 159 | ||
| 159 | static unsigned long vdma_mem_alloc(unsigned long size) | 160 | static unsigned long vdma_mem_alloc(unsigned long size) |
| 160 | { | 161 | { |
| 161 | return (unsigned long) vmalloc(size); | 162 | return (unsigned long)vmalloc(size); |
| 162 | 163 | ||
| 163 | } | 164 | } |
| 164 | 165 | ||
| @@ -166,7 +167,7 @@ static unsigned long vdma_mem_alloc(unsigned long size) | |||
| 166 | 167 | ||
| 167 | static void _fd_dma_mem_free(unsigned long addr, unsigned long size) | 168 | static void _fd_dma_mem_free(unsigned long addr, unsigned long size) |
| 168 | { | 169 | { |
| 169 | if((unsigned long) addr >= (unsigned long) high_memory) | 170 | if ((unsigned long)addr >= (unsigned long)high_memory) |
| 170 | vfree((void *)addr); | 171 | vfree((void *)addr); |
| 171 | else | 172 | else |
| 172 | free_pages(addr, get_order(size)); | 173 | free_pages(addr, get_order(size)); |
| @@ -176,10 +177,10 @@ static void _fd_dma_mem_free(unsigned long addr, unsigned long size) | |||
| 176 | 177 | ||
| 177 | static void _fd_chose_dma_mode(char *addr, unsigned long size) | 178 | static void _fd_chose_dma_mode(char *addr, unsigned long size) |
| 178 | { | 179 | { |
| 179 | if(can_use_virtual_dma == 2) { | 180 | if (can_use_virtual_dma == 2) { |
| 180 | if((unsigned long) addr >= (unsigned long) high_memory || | 181 | if ((unsigned long)addr >= (unsigned long)high_memory || |
| 181 | isa_virt_to_bus(addr) >= 0x1000000 || | 182 | isa_virt_to_bus(addr) >= 0x1000000 || |
| 182 | _CROSS_64KB(addr, size, 0)) | 183 | _CROSS_64KB(addr, size, 0)) |
| 183 | use_virtual_dma = 1; | 184 | use_virtual_dma = 1; |
| 184 | else | 185 | else |
| 185 | use_virtual_dma = 0; | 186 | use_virtual_dma = 0; |
| @@ -195,7 +196,7 @@ static int vdma_dma_setup(char *addr, unsigned long size, int mode, int io) | |||
| 195 | { | 196 | { |
| 196 | doing_pdma = 1; | 197 | doing_pdma = 1; |
| 197 | virtual_dma_port = io; | 198 | virtual_dma_port = io; |
| 198 | virtual_dma_mode = (mode == DMA_MODE_WRITE); | 199 | virtual_dma_mode = (mode == DMA_MODE_WRITE); |
| 199 | virtual_dma_addr = addr; | 200 | virtual_dma_addr = addr; |
| 200 | virtual_dma_count = size; | 201 | virtual_dma_count = size; |
| 201 | virtual_dma_residue = 0; | 202 | virtual_dma_residue = 0; |
| @@ -213,18 +214,18 @@ static int hard_dma_setup(char *addr, unsigned long size, int mode, int io) | |||
| 213 | /* actual, physical DMA */ | 214 | /* actual, physical DMA */ |
| 214 | doing_pdma = 0; | 215 | doing_pdma = 0; |
| 215 | clear_dma_ff(FLOPPY_DMA); | 216 | clear_dma_ff(FLOPPY_DMA); |
| 216 | set_dma_mode(FLOPPY_DMA,mode); | 217 | set_dma_mode(FLOPPY_DMA, mode); |
| 217 | set_dma_addr(FLOPPY_DMA,isa_virt_to_bus(addr)); | 218 | set_dma_addr(FLOPPY_DMA, isa_virt_to_bus(addr)); |
| 218 | set_dma_count(FLOPPY_DMA,size); | 219 | set_dma_count(FLOPPY_DMA, size); |
| 219 | enable_dma(FLOPPY_DMA); | 220 | enable_dma(FLOPPY_DMA); |
| 220 | return 0; | 221 | return 0; |
| 221 | } | 222 | } |
| 222 | 223 | ||
| 223 | static struct fd_routine_l { | 224 | static struct fd_routine_l { |
| 224 | int (*_request_dma)(unsigned int dmanr, const char * device_id); | 225 | int (*_request_dma)(unsigned int dmanr, const char *device_id); |
| 225 | void (*_free_dma)(unsigned int dmanr); | 226 | void (*_free_dma)(unsigned int dmanr); |
| 226 | int (*_get_dma_residue)(unsigned int dummy); | 227 | int (*_get_dma_residue)(unsigned int dummy); |
| 227 | unsigned long (*_dma_mem_alloc) (unsigned long size); | 228 | unsigned long (*_dma_mem_alloc)(unsigned long size); |
| 228 | int (*_dma_setup)(char *addr, unsigned long size, int mode, int io); | 229 | int (*_dma_setup)(char *addr, unsigned long size, int mode, int io); |
| 229 | } fd_routine[] = { | 230 | } fd_routine[] = { |
| 230 | { | 231 | { |
| @@ -252,7 +253,8 @@ static int FDC2 = -1; | |||
| 252 | * is needed to prevent corrupted CMOS RAM in case "insmod floppy" | 253 | * is needed to prevent corrupted CMOS RAM in case "insmod floppy" |
| 253 | * coincides with another rtc CMOS user. Paul G. | 254 | * coincides with another rtc CMOS user. Paul G. |
| 254 | */ | 255 | */ |
| 255 | #define FLOPPY0_TYPE ({ \ | 256 | #define FLOPPY0_TYPE \ |
| 257 | ({ \ | ||
| 256 | unsigned long flags; \ | 258 | unsigned long flags; \ |
| 257 | unsigned char val; \ | 259 | unsigned char val; \ |
| 258 | spin_lock_irqsave(&rtc_lock, flags); \ | 260 | spin_lock_irqsave(&rtc_lock, flags); \ |
| @@ -261,7 +263,8 @@ static int FDC2 = -1; | |||
| 261 | val; \ | 263 | val; \ |
| 262 | }) | 264 | }) |
| 263 | 265 | ||
| 264 | #define FLOPPY1_TYPE ({ \ | 266 | #define FLOPPY1_TYPE \ |
| 267 | ({ \ | ||
| 265 | unsigned long flags; \ | 268 | unsigned long flags; \ |
| 266 | unsigned char val; \ | 269 | unsigned char val; \ |
| 267 | spin_lock_irqsave(&rtc_lock, flags); \ | 270 | spin_lock_irqsave(&rtc_lock, flags); \ |
diff --git a/include/asm-x86/futex.h b/include/asm-x86/futex.h index c9952ea9f698..ac0fbf24d722 100644 --- a/include/asm-x86/futex.h +++ b/include/asm-x86/futex.h | |||
| @@ -12,35 +12,32 @@ | |||
| 12 | #include <asm/uaccess.h> | 12 | #include <asm/uaccess.h> |
| 13 | 13 | ||
| 14 | #define __futex_atomic_op1(insn, ret, oldval, uaddr, oparg) \ | 14 | #define __futex_atomic_op1(insn, ret, oldval, uaddr, oparg) \ |
| 15 | __asm__ __volatile( \ | 15 | asm volatile("1:\t" insn "\n" \ |
| 16 | "1: " insn "\n" \ | 16 | "2:\t.section .fixup,\"ax\"\n" \ |
| 17 | "2: .section .fixup,\"ax\"\n \ | 17 | "3:\tmov\t%3, %1\n" \ |
| 18 | 3: mov %3, %1\n \ | 18 | "\tjmp\t2b\n" \ |
| 19 | jmp 2b\n \ | 19 | "\t.previous\n" \ |
| 20 | .previous\n" \ | 20 | _ASM_EXTABLE(1b, 3b) \ |
| 21 | _ASM_EXTABLE(1b,3b) \ | 21 | : "=r" (oldval), "=r" (ret), "+m" (*uaddr) \ |
| 22 | : "=r" (oldval), "=r" (ret), "+m" (*uaddr) \ | 22 | : "i" (-EFAULT), "0" (oparg), "1" (0)) |
| 23 | : "i" (-EFAULT), "0" (oparg), "1" (0)) | ||
| 24 | 23 | ||
| 25 | #define __futex_atomic_op2(insn, ret, oldval, uaddr, oparg) \ | 24 | #define __futex_atomic_op2(insn, ret, oldval, uaddr, oparg) \ |
| 26 | __asm__ __volatile( \ | 25 | asm volatile("1:\tmovl %2, %0\n" \ |
| 27 | "1: movl %2, %0\n \ | 26 | "\tmovl\t%0, %3\n" \ |
| 28 | movl %0, %3\n" \ | 27 | "\t" insn "\n" \ |
| 29 | insn "\n" \ | 28 | "2:\tlock; cmpxchgl %3, %2\n" \ |
| 30 | "2: lock; cmpxchgl %3, %2\n \ | 29 | "\tjnz\t1b\n" \ |
| 31 | jnz 1b\n \ | 30 | "3:\t.section .fixup,\"ax\"\n" \ |
| 32 | 3: .section .fixup,\"ax\"\n \ | 31 | "4:\tmov\t%5, %1\n" \ |
| 33 | 4: mov %5, %1\n \ | 32 | "\tjmp\t3b\n" \ |
| 34 | jmp 3b\n \ | 33 | "\t.previous\n" \ |
| 35 | .previous\n" \ | 34 | _ASM_EXTABLE(1b, 4b) \ |
| 36 | _ASM_EXTABLE(1b,4b) \ | 35 | _ASM_EXTABLE(2b, 4b) \ |
| 37 | _ASM_EXTABLE(2b,4b) \ | 36 | : "=&a" (oldval), "=&r" (ret), \ |
| 38 | : "=&a" (oldval), "=&r" (ret), "+m" (*uaddr), \ | 37 | "+m" (*uaddr), "=&r" (tem) \ |
| 39 | "=&r" (tem) \ | 38 | : "r" (oparg), "i" (-EFAULT), "1" (0)) |
| 40 | : "r" (oparg), "i" (-EFAULT), "1" (0)) | 39 | |
| 41 | 40 | static inline int futex_atomic_op_inuser(int encoded_op, int __user *uaddr) | |
| 42 | static inline int | ||
| 43 | futex_atomic_op_inuser(int encoded_op, int __user *uaddr) | ||
| 44 | { | 41 | { |
| 45 | int op = (encoded_op >> 28) & 7; | 42 | int op = (encoded_op >> 28) & 7; |
| 46 | int cmp = (encoded_op >> 24) & 15; | 43 | int cmp = (encoded_op >> 24) & 15; |
| @@ -87,20 +84,33 @@ futex_atomic_op_inuser(int encoded_op, int __user *uaddr) | |||
| 87 | 84 | ||
| 88 | if (!ret) { | 85 | if (!ret) { |
| 89 | switch (cmp) { | 86 | switch (cmp) { |
| 90 | case FUTEX_OP_CMP_EQ: ret = (oldval == cmparg); break; | 87 | case FUTEX_OP_CMP_EQ: |
| 91 | case FUTEX_OP_CMP_NE: ret = (oldval != cmparg); break; | 88 | ret = (oldval == cmparg); |
| 92 | case FUTEX_OP_CMP_LT: ret = (oldval < cmparg); break; | 89 | break; |
| 93 | case FUTEX_OP_CMP_GE: ret = (oldval >= cmparg); break; | 90 | case FUTEX_OP_CMP_NE: |
| 94 | case FUTEX_OP_CMP_LE: ret = (oldval <= cmparg); break; | 91 | ret = (oldval != cmparg); |
| 95 | case FUTEX_OP_CMP_GT: ret = (oldval > cmparg); break; | 92 | break; |
| 96 | default: ret = -ENOSYS; | 93 | case FUTEX_OP_CMP_LT: |
| 94 | ret = (oldval < cmparg); | ||
| 95 | break; | ||
| 96 | case FUTEX_OP_CMP_GE: | ||
| 97 | ret = (oldval >= cmparg); | ||
| 98 | break; | ||
| 99 | case FUTEX_OP_CMP_LE: | ||
| 100 | ret = (oldval <= cmparg); | ||
| 101 | break; | ||
| 102 | case FUTEX_OP_CMP_GT: | ||
| 103 | ret = (oldval > cmparg); | ||
| 104 | break; | ||
| 105 | default: | ||
| 106 | ret = -ENOSYS; | ||
| 97 | } | 107 | } |
| 98 | } | 108 | } |
| 99 | return ret; | 109 | return ret; |
| 100 | } | 110 | } |
| 101 | 111 | ||
| 102 | static inline int | 112 | static inline int futex_atomic_cmpxchg_inatomic(int __user *uaddr, int oldval, |
| 103 | futex_atomic_cmpxchg_inatomic(int __user *uaddr, int oldval, int newval) | 113 | int newval) |
| 104 | { | 114 | { |
| 105 | 115 | ||
| 106 | #if defined(CONFIG_X86_32) && !defined(CONFIG_X86_BSWAP) | 116 | #if defined(CONFIG_X86_32) && !defined(CONFIG_X86_BSWAP) |
| @@ -112,16 +122,15 @@ futex_atomic_cmpxchg_inatomic(int __user *uaddr, int oldval, int newval) | |||
| 112 | if (!access_ok(VERIFY_WRITE, uaddr, sizeof(int))) | 122 | if (!access_ok(VERIFY_WRITE, uaddr, sizeof(int))) |
| 113 | return -EFAULT; | 123 | return -EFAULT; |
| 114 | 124 | ||
| 115 | __asm__ __volatile__( | 125 | asm volatile("1:\tlock; cmpxchgl %3, %1\n" |
| 116 | "1: lock; cmpxchgl %3, %1 \n" | 126 | "2:\t.section .fixup, \"ax\"\n" |
| 117 | "2: .section .fixup, \"ax\" \n" | 127 | "3:\tmov %2, %0\n" |
| 118 | "3: mov %2, %0 \n" | 128 | "\tjmp 2b\n" |
| 119 | " jmp 2b \n" | 129 | "\t.previous\n" |
| 120 | " .previous \n" | 130 | _ASM_EXTABLE(1b, 3b) |
| 121 | _ASM_EXTABLE(1b,3b) | 131 | : "=a" (oldval), "+m" (*uaddr) |
| 122 | : "=a" (oldval), "+m" (*uaddr) | 132 | : "i" (-EFAULT), "r" (newval), "0" (oldval) |
| 123 | : "i" (-EFAULT), "r" (newval), "0" (oldval) | 133 | : "memory" |
| 124 | : "memory" | ||
| 125 | ); | 134 | ); |
| 126 | 135 | ||
| 127 | return oldval; | 136 | return oldval; |
diff --git a/include/asm-x86/genapic_32.h b/include/asm-x86/genapic_32.h index 33e3ffe1766c..f1b96932746b 100644 --- a/include/asm-x86/genapic_32.h +++ b/include/asm-x86/genapic_32.h | |||
| @@ -14,23 +14,22 @@ | |||
| 14 | * Copyright 2003 Andi Kleen, SuSE Labs. | 14 | * Copyright 2003 Andi Kleen, SuSE Labs. |
| 15 | */ | 15 | */ |
| 16 | 16 | ||
| 17 | struct mpc_config_translation; | ||
| 18 | struct mpc_config_bus; | 17 | struct mpc_config_bus; |
| 19 | struct mp_config_table; | 18 | struct mp_config_table; |
| 20 | struct mpc_config_processor; | 19 | struct mpc_config_processor; |
| 21 | 20 | ||
| 22 | struct genapic { | 21 | struct genapic { |
| 23 | char *name; | 22 | char *name; |
| 24 | int (*probe)(void); | 23 | int (*probe)(void); |
| 25 | 24 | ||
| 26 | int (*apic_id_registered)(void); | 25 | int (*apic_id_registered)(void); |
| 27 | cpumask_t (*target_cpus)(void); | 26 | cpumask_t (*target_cpus)(void); |
| 28 | int int_delivery_mode; | 27 | int int_delivery_mode; |
| 29 | int int_dest_mode; | 28 | int int_dest_mode; |
| 30 | int ESR_DISABLE; | 29 | int ESR_DISABLE; |
| 31 | int apic_destination_logical; | 30 | int apic_destination_logical; |
| 32 | unsigned long (*check_apicid_used)(physid_mask_t bitmap, int apicid); | 31 | unsigned long (*check_apicid_used)(physid_mask_t bitmap, int apicid); |
| 33 | unsigned long (*check_apicid_present)(int apicid); | 32 | unsigned long (*check_apicid_present)(int apicid); |
| 34 | int no_balance_irq; | 33 | int no_balance_irq; |
| 35 | int no_ioapic_check; | 34 | int no_ioapic_check; |
| 36 | void (*init_apic_ldr)(void); | 35 | void (*init_apic_ldr)(void); |
| @@ -38,28 +37,21 @@ struct genapic { | |||
| 38 | 37 | ||
| 39 | void (*setup_apic_routing)(void); | 38 | void (*setup_apic_routing)(void); |
| 40 | int (*multi_timer_check)(int apic, int irq); | 39 | int (*multi_timer_check)(int apic, int irq); |
| 41 | int (*apicid_to_node)(int logical_apicid); | 40 | int (*apicid_to_node)(int logical_apicid); |
| 42 | int (*cpu_to_logical_apicid)(int cpu); | 41 | int (*cpu_to_logical_apicid)(int cpu); |
| 43 | int (*cpu_present_to_apicid)(int mps_cpu); | 42 | int (*cpu_present_to_apicid)(int mps_cpu); |
| 44 | physid_mask_t (*apicid_to_cpu_present)(int phys_apicid); | 43 | physid_mask_t (*apicid_to_cpu_present)(int phys_apicid); |
| 45 | int (*mpc_apic_id)(struct mpc_config_processor *m, | 44 | void (*setup_portio_remap)(void); |
| 46 | struct mpc_config_translation *t); | ||
| 47 | void (*setup_portio_remap)(void); | ||
| 48 | int (*check_phys_apicid_present)(int boot_cpu_physical_apicid); | 45 | int (*check_phys_apicid_present)(int boot_cpu_physical_apicid); |
| 49 | void (*enable_apic_mode)(void); | 46 | void (*enable_apic_mode)(void); |
| 50 | u32 (*phys_pkg_id)(u32 cpuid_apic, int index_msb); | 47 | u32 (*phys_pkg_id)(u32 cpuid_apic, int index_msb); |
| 51 | 48 | ||
| 52 | /* mpparse */ | 49 | /* mpparse */ |
| 53 | void (*mpc_oem_bus_info)(struct mpc_config_bus *, char *, | ||
| 54 | struct mpc_config_translation *); | ||
| 55 | void (*mpc_oem_pci_bus)(struct mpc_config_bus *, | ||
| 56 | struct mpc_config_translation *); | ||
| 57 | |||
| 58 | /* When one of the next two hooks returns 1 the genapic | 50 | /* When one of the next two hooks returns 1 the genapic |
| 59 | is switched to this. Essentially they are additional probe | 51 | is switched to this. Essentially they are additional probe |
| 60 | functions. */ | 52 | functions. */ |
| 61 | int (*mps_oem_check)(struct mp_config_table *mpc, char *oem, | 53 | int (*mps_oem_check)(struct mp_config_table *mpc, char *oem, |
| 62 | char *productid); | 54 | char *productid); |
| 63 | int (*acpi_madt_oem_check)(char *oem_id, char *oem_table_id); | 55 | int (*acpi_madt_oem_check)(char *oem_id, char *oem_table_id); |
| 64 | 56 | ||
| 65 | unsigned (*get_apic_id)(unsigned long x); | 57 | unsigned (*get_apic_id)(unsigned long x); |
| @@ -72,7 +64,7 @@ struct genapic { | |||
| 72 | void (*send_IPI_allbutself)(int vector); | 64 | void (*send_IPI_allbutself)(int vector); |
| 73 | void (*send_IPI_all)(int vector); | 65 | void (*send_IPI_all)(int vector); |
| 74 | #endif | 66 | #endif |
| 75 | }; | 67 | }; |
| 76 | 68 | ||
| 77 | #define APICFUNC(x) .x = x, | 69 | #define APICFUNC(x) .x = x, |
| 78 | 70 | ||
| @@ -85,43 +77,46 @@ struct genapic { | |||
| 85 | #define IPIFUNC(x) | 77 | #define IPIFUNC(x) |
| 86 | #endif | 78 | #endif |
| 87 | 79 | ||
| 88 | #define APIC_INIT(aname, aprobe) { \ | 80 | #define APIC_INIT(aname, aprobe) \ |
| 89 | .name = aname, \ | 81 | { \ |
| 90 | .probe = aprobe, \ | 82 | .name = aname, \ |
| 91 | .int_delivery_mode = INT_DELIVERY_MODE, \ | 83 | .probe = aprobe, \ |
| 92 | .int_dest_mode = INT_DEST_MODE, \ | 84 | .int_delivery_mode = INT_DELIVERY_MODE, \ |
| 93 | .no_balance_irq = NO_BALANCE_IRQ, \ | 85 | .int_dest_mode = INT_DEST_MODE, \ |
| 94 | .ESR_DISABLE = esr_disable, \ | 86 | .no_balance_irq = NO_BALANCE_IRQ, \ |
| 95 | .apic_destination_logical = APIC_DEST_LOGICAL, \ | 87 | .ESR_DISABLE = esr_disable, \ |
| 96 | APICFUNC(apic_id_registered) \ | 88 | .apic_destination_logical = APIC_DEST_LOGICAL, \ |
| 97 | APICFUNC(target_cpus) \ | 89 | APICFUNC(apic_id_registered) \ |
| 98 | APICFUNC(check_apicid_used) \ | 90 | APICFUNC(target_cpus) \ |
| 99 | APICFUNC(check_apicid_present) \ | 91 | APICFUNC(check_apicid_used) \ |
| 100 | APICFUNC(init_apic_ldr) \ | 92 | APICFUNC(check_apicid_present) \ |
| 101 | APICFUNC(ioapic_phys_id_map) \ | 93 | APICFUNC(init_apic_ldr) \ |
| 102 | APICFUNC(setup_apic_routing) \ | 94 | APICFUNC(ioapic_phys_id_map) \ |
| 103 | APICFUNC(multi_timer_check) \ | 95 | APICFUNC(setup_apic_routing) \ |
| 104 | APICFUNC(apicid_to_node) \ | 96 | APICFUNC(multi_timer_check) \ |
| 105 | APICFUNC(cpu_to_logical_apicid) \ | 97 | APICFUNC(apicid_to_node) \ |
| 106 | APICFUNC(cpu_present_to_apicid) \ | 98 | APICFUNC(cpu_to_logical_apicid) \ |
| 107 | APICFUNC(apicid_to_cpu_present) \ | 99 | APICFUNC(cpu_present_to_apicid) \ |
| 108 | APICFUNC(mpc_apic_id) \ | 100 | APICFUNC(apicid_to_cpu_present) \ |
| 109 | APICFUNC(setup_portio_remap) \ | 101 | APICFUNC(setup_portio_remap) \ |
| 110 | APICFUNC(check_phys_apicid_present) \ | 102 | APICFUNC(check_phys_apicid_present) \ |
| 111 | APICFUNC(mpc_oem_bus_info) \ | 103 | APICFUNC(mps_oem_check) \ |
| 112 | APICFUNC(mpc_oem_pci_bus) \ | 104 | APICFUNC(get_apic_id) \ |
| 113 | APICFUNC(mps_oem_check) \ | 105 | .apic_id_mask = APIC_ID_MASK, \ |
| 114 | APICFUNC(get_apic_id) \ | 106 | APICFUNC(cpu_mask_to_apicid) \ |
| 115 | .apic_id_mask = APIC_ID_MASK, \ | 107 | APICFUNC(acpi_madt_oem_check) \ |
| 116 | APICFUNC(cpu_mask_to_apicid) \ | 108 | IPIFUNC(send_IPI_mask) \ |
| 117 | APICFUNC(acpi_madt_oem_check) \ | 109 | IPIFUNC(send_IPI_allbutself) \ |
| 118 | IPIFUNC(send_IPI_mask) \ | 110 | IPIFUNC(send_IPI_all) \ |
| 119 | IPIFUNC(send_IPI_allbutself) \ | 111 | APICFUNC(enable_apic_mode) \ |
| 120 | IPIFUNC(send_IPI_all) \ | 112 | APICFUNC(phys_pkg_id) \ |
| 121 | APICFUNC(enable_apic_mode) \ | 113 | } |
| 122 | APICFUNC(phys_pkg_id) \ | ||
| 123 | } | ||
| 124 | 114 | ||
| 125 | extern struct genapic *genapic; | 115 | extern struct genapic *genapic; |
| 126 | 116 | ||
| 117 | enum uv_system_type {UV_NONE, UV_LEGACY_APIC, UV_X2APIC, UV_NON_UNIQUE_APIC}; | ||
| 118 | #define get_uv_system_type() UV_NONE | ||
| 119 | #define is_uv_system() 0 | ||
| 120 | |||
| 121 | |||
| 127 | #endif | 122 | #endif |
diff --git a/include/asm-x86/genapic_64.h b/include/asm-x86/genapic_64.h index d7e516ccbaa4..1de931b263ce 100644 --- a/include/asm-x86/genapic_64.h +++ b/include/asm-x86/genapic_64.h | |||
| @@ -33,5 +33,15 @@ extern struct genapic *genapic; | |||
| 33 | 33 | ||
| 34 | extern struct genapic apic_flat; | 34 | extern struct genapic apic_flat; |
| 35 | extern struct genapic apic_physflat; | 35 | extern struct genapic apic_physflat; |
| 36 | extern int acpi_madt_oem_check(char *, char *); | ||
| 37 | |||
| 38 | enum uv_system_type {UV_NONE, UV_LEGACY_APIC, UV_X2APIC, UV_NON_UNIQUE_APIC}; | ||
| 39 | extern enum uv_system_type get_uv_system_type(void); | ||
| 40 | extern int is_uv_system(void); | ||
| 41 | |||
| 42 | extern struct genapic apic_x2apic_uv_x; | ||
| 43 | DECLARE_PER_CPU(int, x2apic_extra_bits); | ||
| 44 | extern void uv_cpu_init(void); | ||
| 45 | extern int uv_wakeup_secondary(int phys_apicid, unsigned int start_rip); | ||
| 36 | 46 | ||
| 37 | #endif | 47 | #endif |
diff --git a/include/asm-x86/geode.h b/include/asm-x86/geode.h index 9e7280092a48..9870cc1f2f8f 100644 --- a/include/asm-x86/geode.h +++ b/include/asm-x86/geode.h | |||
| @@ -167,7 +167,7 @@ static inline int is_geode(void) | |||
| 167 | /* MFGPTs */ | 167 | /* MFGPTs */ |
| 168 | 168 | ||
| 169 | #define MFGPT_MAX_TIMERS 8 | 169 | #define MFGPT_MAX_TIMERS 8 |
| 170 | #define MFGPT_TIMER_ANY -1 | 170 | #define MFGPT_TIMER_ANY (-1) |
| 171 | 171 | ||
| 172 | #define MFGPT_DOMAIN_WORKING 1 | 172 | #define MFGPT_DOMAIN_WORKING 1 |
| 173 | #define MFGPT_DOMAIN_STANDBY 2 | 173 | #define MFGPT_DOMAIN_STANDBY 2 |
diff --git a/include/asm-x86/highmem.h b/include/asm-x86/highmem.h index 479767c9195f..e153f3b44774 100644 --- a/include/asm-x86/highmem.h +++ b/include/asm-x86/highmem.h | |||
| @@ -8,7 +8,7 @@ | |||
| 8 | * Gerhard.Wichert@pdb.siemens.de | 8 | * Gerhard.Wichert@pdb.siemens.de |
| 9 | * | 9 | * |
| 10 | * | 10 | * |
| 11 | * Redesigned the x86 32-bit VM architecture to deal with | 11 | * Redesigned the x86 32-bit VM architecture to deal with |
| 12 | * up to 16 Terabyte physical memory. With current x86 CPUs | 12 | * up to 16 Terabyte physical memory. With current x86 CPUs |
| 13 | * we now support up to 64 Gigabytes physical RAM. | 13 | * we now support up to 64 Gigabytes physical RAM. |
| 14 | * | 14 | * |
diff --git a/include/asm-x86/hw_irq_64.h b/include/asm-x86/hw_irq_64.h index 312a58d6dac6..0062ef390f67 100644 --- a/include/asm-x86/hw_irq_64.h +++ b/include/asm-x86/hw_irq_64.h | |||
| @@ -36,7 +36,7 @@ | |||
| 36 | * cleanup after irq migration. | 36 | * cleanup after irq migration. |
| 37 | */ | 37 | */ |
| 38 | #define IRQ_MOVE_CLEANUP_VECTOR FIRST_EXTERNAL_VECTOR | 38 | #define IRQ_MOVE_CLEANUP_VECTOR FIRST_EXTERNAL_VECTOR |
| 39 | 39 | ||
| 40 | /* | 40 | /* |
| 41 | * Vectors 0x30-0x3f are used for ISA interrupts. | 41 | * Vectors 0x30-0x3f are used for ISA interrupts. |
| 42 | */ | 42 | */ |
| @@ -159,13 +159,12 @@ extern atomic_t irq_mis_count; | |||
| 159 | * SMP has a few special interrupts for IPI messages | 159 | * SMP has a few special interrupts for IPI messages |
| 160 | */ | 160 | */ |
| 161 | 161 | ||
| 162 | #define BUILD_IRQ(nr) \ | 162 | #define BUILD_IRQ(nr) \ |
| 163 | asmlinkage void IRQ_NAME(nr); \ | 163 | asmlinkage void IRQ_NAME(nr); \ |
| 164 | __asm__( \ | 164 | asm("\n.p2align\n" \ |
| 165 | "\n.p2align\n" \ | 165 | "IRQ" #nr "_interrupt:\n\t" \ |
| 166 | "IRQ" #nr "_interrupt:\n\t" \ | 166 | "push $~(" #nr ") ; " \ |
| 167 | "push $~(" #nr ") ; " \ | 167 | "jmp common_interrupt"); |
| 168 | "jmp common_interrupt"); | ||
| 169 | 168 | ||
| 170 | #define platform_legacy_irq(irq) ((irq) < 16) | 169 | #define platform_legacy_irq(irq) ((irq) < 16) |
| 171 | 170 | ||
diff --git a/include/asm-x86/hypertransport.h b/include/asm-x86/hypertransport.h index c16c6ff4bdd7..d2bbd238b3e1 100644 --- a/include/asm-x86/hypertransport.h +++ b/include/asm-x86/hypertransport.h | |||
| @@ -8,12 +8,14 @@ | |||
| 8 | #define HT_IRQ_LOW_BASE 0xf8000000 | 8 | #define HT_IRQ_LOW_BASE 0xf8000000 |
| 9 | 9 | ||
| 10 | #define HT_IRQ_LOW_VECTOR_SHIFT 16 | 10 | #define HT_IRQ_LOW_VECTOR_SHIFT 16 |
| 11 | #define HT_IRQ_LOW_VECTOR_MASK 0x00ff0000 | 11 | #define HT_IRQ_LOW_VECTOR_MASK 0x00ff0000 |
| 12 | #define HT_IRQ_LOW_VECTOR(v) (((v) << HT_IRQ_LOW_VECTOR_SHIFT) & HT_IRQ_LOW_VECTOR_MASK) | 12 | #define HT_IRQ_LOW_VECTOR(v) \ |
| 13 | (((v) << HT_IRQ_LOW_VECTOR_SHIFT) & HT_IRQ_LOW_VECTOR_MASK) | ||
| 13 | 14 | ||
| 14 | #define HT_IRQ_LOW_DEST_ID_SHIFT 8 | 15 | #define HT_IRQ_LOW_DEST_ID_SHIFT 8 |
| 15 | #define HT_IRQ_LOW_DEST_ID_MASK 0x0000ff00 | 16 | #define HT_IRQ_LOW_DEST_ID_MASK 0x0000ff00 |
| 16 | #define HT_IRQ_LOW_DEST_ID(v) (((v) << HT_IRQ_LOW_DEST_ID_SHIFT) & HT_IRQ_LOW_DEST_ID_MASK) | 17 | #define HT_IRQ_LOW_DEST_ID(v) \ |
| 18 | (((v) << HT_IRQ_LOW_DEST_ID_SHIFT) & HT_IRQ_LOW_DEST_ID_MASK) | ||
| 17 | 19 | ||
| 18 | #define HT_IRQ_LOW_DM_PHYSICAL 0x0000000 | 20 | #define HT_IRQ_LOW_DM_PHYSICAL 0x0000000 |
| 19 | #define HT_IRQ_LOW_DM_LOGICAL 0x0000040 | 21 | #define HT_IRQ_LOW_DM_LOGICAL 0x0000040 |
| @@ -36,7 +38,8 @@ | |||
| 36 | 38 | ||
| 37 | 39 | ||
| 38 | #define HT_IRQ_HIGH_DEST_ID_SHIFT 0 | 40 | #define HT_IRQ_HIGH_DEST_ID_SHIFT 0 |
| 39 | #define HT_IRQ_HIGH_DEST_ID_MASK 0x00ffffff | 41 | #define HT_IRQ_HIGH_DEST_ID_MASK 0x00ffffff |
| 40 | #define HT_IRQ_HIGH_DEST_ID(v) ((((v) >> 8) << HT_IRQ_HIGH_DEST_ID_SHIFT) & HT_IRQ_HIGH_DEST_ID_MASK) | 42 | #define HT_IRQ_HIGH_DEST_ID(v) \ |
| 43 | ((((v) >> 8) << HT_IRQ_HIGH_DEST_ID_SHIFT) & HT_IRQ_HIGH_DEST_ID_MASK) | ||
| 41 | 44 | ||
| 42 | #endif /* ASM_HYPERTRANSPORT_H */ | 45 | #endif /* ASM_HYPERTRANSPORT_H */ |
diff --git a/include/asm-x86/i387.h b/include/asm-x86/i387.h index f377b76b2f34..54522b814f1c 100644 --- a/include/asm-x86/i387.h +++ b/include/asm-x86/i387.h | |||
| @@ -41,7 +41,7 @@ static inline void tolerant_fwait(void) | |||
| 41 | { | 41 | { |
| 42 | asm volatile("1: fwait\n" | 42 | asm volatile("1: fwait\n" |
| 43 | "2:\n" | 43 | "2:\n" |
| 44 | _ASM_EXTABLE(1b,2b)); | 44 | _ASM_EXTABLE(1b, 2b)); |
| 45 | } | 45 | } |
| 46 | 46 | ||
| 47 | static inline int restore_fpu_checking(struct i387_fxsave_struct *fx) | 47 | static inline int restore_fpu_checking(struct i387_fxsave_struct *fx) |
| @@ -54,7 +54,7 @@ static inline int restore_fpu_checking(struct i387_fxsave_struct *fx) | |||
| 54 | "3: movl $-1,%[err]\n" | 54 | "3: movl $-1,%[err]\n" |
| 55 | " jmp 2b\n" | 55 | " jmp 2b\n" |
| 56 | ".previous\n" | 56 | ".previous\n" |
| 57 | _ASM_EXTABLE(1b,3b) | 57 | _ASM_EXTABLE(1b, 3b) |
| 58 | : [err] "=r" (err) | 58 | : [err] "=r" (err) |
| 59 | #if 0 /* See comment in __save_init_fpu() below. */ | 59 | #if 0 /* See comment in __save_init_fpu() below. */ |
| 60 | : [fx] "r" (fx), "m" (*fx), "0" (0)); | 60 | : [fx] "r" (fx), "m" (*fx), "0" (0)); |
| @@ -76,11 +76,11 @@ static inline int restore_fpu_checking(struct i387_fxsave_struct *fx) | |||
| 76 | static inline void clear_fpu_state(struct i387_fxsave_struct *fx) | 76 | static inline void clear_fpu_state(struct i387_fxsave_struct *fx) |
| 77 | { | 77 | { |
| 78 | if (unlikely(fx->swd & X87_FSW_ES)) | 78 | if (unlikely(fx->swd & X87_FSW_ES)) |
| 79 | asm volatile("fnclex"); | 79 | asm volatile("fnclex"); |
| 80 | alternative_input(ASM_NOP8 ASM_NOP2, | 80 | alternative_input(ASM_NOP8 ASM_NOP2, |
| 81 | " emms\n" /* clear stack tags */ | 81 | " emms\n" /* clear stack tags */ |
| 82 | " fildl %%gs:0", /* load to clear state */ | 82 | " fildl %%gs:0", /* load to clear state */ |
| 83 | X86_FEATURE_FXSAVE_LEAK); | 83 | X86_FEATURE_FXSAVE_LEAK); |
| 84 | } | 84 | } |
| 85 | 85 | ||
| 86 | static inline int save_i387_checking(struct i387_fxsave_struct __user *fx) | 86 | static inline int save_i387_checking(struct i387_fxsave_struct __user *fx) |
| @@ -93,14 +93,15 @@ static inline int save_i387_checking(struct i387_fxsave_struct __user *fx) | |||
| 93 | "3: movl $-1,%[err]\n" | 93 | "3: movl $-1,%[err]\n" |
| 94 | " jmp 2b\n" | 94 | " jmp 2b\n" |
| 95 | ".previous\n" | 95 | ".previous\n" |
| 96 | _ASM_EXTABLE(1b,3b) | 96 | _ASM_EXTABLE(1b, 3b) |
| 97 | : [err] "=r" (err), "=m" (*fx) | 97 | : [err] "=r" (err), "=m" (*fx) |
| 98 | #if 0 /* See comment in __fxsave_clear() below. */ | 98 | #if 0 /* See comment in __fxsave_clear() below. */ |
| 99 | : [fx] "r" (fx), "0" (0)); | 99 | : [fx] "r" (fx), "0" (0)); |
| 100 | #else | 100 | #else |
| 101 | : [fx] "cdaSDb" (fx), "0" (0)); | 101 | : [fx] "cdaSDb" (fx), "0" (0)); |
| 102 | #endif | 102 | #endif |
| 103 | if (unlikely(err) && __clear_user(fx, sizeof(struct i387_fxsave_struct))) | 103 | if (unlikely(err) && |
| 104 | __clear_user(fx, sizeof(struct i387_fxsave_struct))) | ||
| 104 | err = -EFAULT; | 105 | err = -EFAULT; |
| 105 | /* No need to clear here because the caller clears USED_MATH */ | 106 | /* No need to clear here because the caller clears USED_MATH */ |
| 106 | return err; | 107 | return err; |
| @@ -156,8 +157,10 @@ static inline int save_i387(struct _fpstate __user *buf) | |||
| 156 | return 0; | 157 | return 0; |
| 157 | clear_used_math(); /* trigger finit */ | 158 | clear_used_math(); /* trigger finit */ |
| 158 | if (task_thread_info(tsk)->status & TS_USEDFPU) { | 159 | if (task_thread_info(tsk)->status & TS_USEDFPU) { |
| 159 | err = save_i387_checking((struct i387_fxsave_struct __user *)buf); | 160 | err = save_i387_checking((struct i387_fxsave_struct __user *) |
| 160 | if (err) return err; | 161 | buf); |
| 162 | if (err) | ||
| 163 | return err; | ||
| 161 | task_thread_info(tsk)->status &= ~TS_USEDFPU; | 164 | task_thread_info(tsk)->status &= ~TS_USEDFPU; |
| 162 | stts(); | 165 | stts(); |
| 163 | } else { | 166 | } else { |
diff --git a/include/asm-x86/i8259.h b/include/asm-x86/i8259.h index 67c319e0efc7..45d4df3e51e6 100644 --- a/include/asm-x86/i8259.h +++ b/include/asm-x86/i8259.h | |||
| @@ -1,9 +1,11 @@ | |||
| 1 | #ifndef __ASM_I8259_H__ | 1 | #ifndef __ASM_I8259_H__ |
| 2 | #define __ASM_I8259_H__ | 2 | #define __ASM_I8259_H__ |
| 3 | 3 | ||
| 4 | #include <linux/delay.h> | ||
| 5 | |||
| 4 | extern unsigned int cached_irq_mask; | 6 | extern unsigned int cached_irq_mask; |
| 5 | 7 | ||
| 6 | #define __byte(x,y) (((unsigned char *) &(y))[x]) | 8 | #define __byte(x, y) (((unsigned char *)&(y))[x]) |
| 7 | #define cached_master_mask (__byte(0, cached_irq_mask)) | 9 | #define cached_master_mask (__byte(0, cached_irq_mask)) |
| 8 | #define cached_slave_mask (__byte(1, cached_irq_mask)) | 10 | #define cached_slave_mask (__byte(1, cached_irq_mask)) |
| 9 | 11 | ||
| @@ -29,7 +31,28 @@ extern void enable_8259A_irq(unsigned int irq); | |||
| 29 | extern void disable_8259A_irq(unsigned int irq); | 31 | extern void disable_8259A_irq(unsigned int irq); |
| 30 | extern unsigned int startup_8259A_irq(unsigned int irq); | 32 | extern unsigned int startup_8259A_irq(unsigned int irq); |
| 31 | 33 | ||
| 32 | #define inb_pic inb_p | 34 | /* the PIC may need a careful delay on some platforms, hence specific calls */ |
| 33 | #define outb_pic outb_p | 35 | static inline unsigned char inb_pic(unsigned int port) |
| 36 | { | ||
| 37 | unsigned char value = inb(port); | ||
| 38 | |||
| 39 | /* | ||
| 40 | * delay for some accesses to PIC on motherboard or in chipset | ||
| 41 | * must be at least one microsecond, so be safe here: | ||
| 42 | */ | ||
| 43 | udelay(2); | ||
| 44 | |||
| 45 | return value; | ||
| 46 | } | ||
| 47 | |||
| 48 | static inline void outb_pic(unsigned char value, unsigned int port) | ||
| 49 | { | ||
| 50 | outb(value, port); | ||
| 51 | /* | ||
| 52 | * delay for some accesses to PIC on motherboard or in chipset | ||
| 53 | * must be at least one microsecond, so be safe here: | ||
| 54 | */ | ||
| 55 | udelay(2); | ||
| 56 | } | ||
| 34 | 57 | ||
| 35 | #endif /* __ASM_I8259_H__ */ | 58 | #endif /* __ASM_I8259_H__ */ |
diff --git a/include/asm-x86/ia32.h b/include/asm-x86/ia32.h index aa9733206e29..55d3abe5276f 100644 --- a/include/asm-x86/ia32.h +++ b/include/asm-x86/ia32.h | |||
| @@ -14,19 +14,19 @@ | |||
| 14 | 14 | ||
| 15 | /* signal.h */ | 15 | /* signal.h */ |
| 16 | struct sigaction32 { | 16 | struct sigaction32 { |
| 17 | unsigned int sa_handler; /* Really a pointer, but need to deal | 17 | unsigned int sa_handler; /* Really a pointer, but need to deal |
| 18 | with 32 bits */ | 18 | with 32 bits */ |
| 19 | unsigned int sa_flags; | 19 | unsigned int sa_flags; |
| 20 | unsigned int sa_restorer; /* Another 32 bit pointer */ | 20 | unsigned int sa_restorer; /* Another 32 bit pointer */ |
| 21 | compat_sigset_t sa_mask; /* A 32 bit mask */ | 21 | compat_sigset_t sa_mask; /* A 32 bit mask */ |
| 22 | }; | 22 | }; |
| 23 | 23 | ||
| 24 | struct old_sigaction32 { | 24 | struct old_sigaction32 { |
| 25 | unsigned int sa_handler; /* Really a pointer, but need to deal | 25 | unsigned int sa_handler; /* Really a pointer, but need to deal |
| 26 | with 32 bits */ | 26 | with 32 bits */ |
| 27 | compat_old_sigset_t sa_mask; /* A 32 bit mask */ | 27 | compat_old_sigset_t sa_mask; /* A 32 bit mask */ |
| 28 | unsigned int sa_flags; | 28 | unsigned int sa_flags; |
| 29 | unsigned int sa_restorer; /* Another 32 bit pointer */ | 29 | unsigned int sa_restorer; /* Another 32 bit pointer */ |
| 30 | }; | 30 | }; |
| 31 | 31 | ||
| 32 | typedef struct sigaltstack_ia32 { | 32 | typedef struct sigaltstack_ia32 { |
| @@ -65,7 +65,7 @@ struct stat64 { | |||
| 65 | long long st_size; | 65 | long long st_size; |
| 66 | unsigned int st_blksize; | 66 | unsigned int st_blksize; |
| 67 | 67 | ||
| 68 | long long st_blocks;/* Number 512-byte blocks allocated. */ | 68 | long long st_blocks;/* Number 512-byte blocks allocated */ |
| 69 | 69 | ||
| 70 | unsigned st_atime; | 70 | unsigned st_atime; |
| 71 | unsigned st_atime_nsec; | 71 | unsigned st_atime_nsec; |
| @@ -77,13 +77,13 @@ struct stat64 { | |||
| 77 | unsigned long long st_ino; | 77 | unsigned long long st_ino; |
| 78 | } __attribute__((packed)); | 78 | } __attribute__((packed)); |
| 79 | 79 | ||
| 80 | typedef struct compat_siginfo{ | 80 | typedef struct compat_siginfo { |
| 81 | int si_signo; | 81 | int si_signo; |
| 82 | int si_errno; | 82 | int si_errno; |
| 83 | int si_code; | 83 | int si_code; |
| 84 | 84 | ||
| 85 | union { | 85 | union { |
| 86 | int _pad[((128/sizeof(int)) - 3)]; | 86 | int _pad[((128 / sizeof(int)) - 3)]; |
| 87 | 87 | ||
| 88 | /* kill() */ | 88 | /* kill() */ |
| 89 | struct { | 89 | struct { |
| @@ -129,28 +129,26 @@ typedef struct compat_siginfo{ | |||
| 129 | } _sifields; | 129 | } _sifields; |
| 130 | } compat_siginfo_t; | 130 | } compat_siginfo_t; |
| 131 | 131 | ||
| 132 | struct sigframe32 | 132 | struct sigframe32 { |
| 133 | { | 133 | u32 pretcode; |
| 134 | u32 pretcode; | 134 | int sig; |
| 135 | int sig; | 135 | struct sigcontext_ia32 sc; |
| 136 | struct sigcontext_ia32 sc; | 136 | struct _fpstate_ia32 fpstate; |
| 137 | struct _fpstate_ia32 fpstate; | 137 | unsigned int extramask[_COMPAT_NSIG_WORDS-1]; |
| 138 | unsigned int extramask[_COMPAT_NSIG_WORDS-1]; | ||
| 139 | }; | 138 | }; |
| 140 | 139 | ||
| 141 | struct rt_sigframe32 | 140 | struct rt_sigframe32 { |
| 142 | { | 141 | u32 pretcode; |
| 143 | u32 pretcode; | 142 | int sig; |
| 144 | int sig; | 143 | u32 pinfo; |
| 145 | u32 pinfo; | 144 | u32 puc; |
| 146 | u32 puc; | 145 | compat_siginfo_t info; |
| 147 | compat_siginfo_t info; | 146 | struct ucontext_ia32 uc; |
| 148 | struct ucontext_ia32 uc; | 147 | struct _fpstate_ia32 fpstate; |
| 149 | struct _fpstate_ia32 fpstate; | ||
| 150 | }; | 148 | }; |
| 151 | 149 | ||
| 152 | struct ustat32 { | 150 | struct ustat32 { |
| 153 | __u32 f_tfree; | 151 | __u32 f_tfree; |
| 154 | compat_ino_t f_tinode; | 152 | compat_ino_t f_tinode; |
| 155 | char f_fname[6]; | 153 | char f_fname[6]; |
| 156 | char f_fpack[6]; | 154 | char f_fpack[6]; |
| @@ -168,5 +166,5 @@ extern void ia32_pick_mmap_layout(struct mm_struct *mm); | |||
| 168 | #endif | 166 | #endif |
| 169 | 167 | ||
| 170 | #endif /* !CONFIG_IA32_SUPPORT */ | 168 | #endif /* !CONFIG_IA32_SUPPORT */ |
| 171 | 169 | ||
| 172 | #endif | 170 | #endif |
diff --git a/include/asm-x86/io.h b/include/asm-x86/io.h index 5a58b176dd61..7b292d386713 100644 --- a/include/asm-x86/io.h +++ b/include/asm-x86/io.h | |||
| @@ -1,5 +1,11 @@ | |||
| 1 | #define ARCH_HAS_IOREMAP_WC | ||
| 2 | |||
| 1 | #ifdef CONFIG_X86_32 | 3 | #ifdef CONFIG_X86_32 |
| 2 | # include "io_32.h" | 4 | # include "io_32.h" |
| 3 | #else | 5 | #else |
| 4 | # include "io_64.h" | 6 | # include "io_64.h" |
| 5 | #endif | 7 | #endif |
| 8 | extern int ioremap_change_attr(unsigned long vaddr, unsigned long size, | ||
| 9 | unsigned long prot_val); | ||
| 10 | extern void __iomem *ioremap_wc(unsigned long offset, unsigned long size); | ||
| 11 | |||
diff --git a/include/asm-x86/io_32.h b/include/asm-x86/io_32.h index d4d8fbd9378c..509045f5fda2 100644 --- a/include/asm-x86/io_32.h +++ b/include/asm-x86/io_32.h | |||
| @@ -65,14 +65,14 @@ | |||
| 65 | * | 65 | * |
| 66 | * The returned physical address is the physical (CPU) mapping for | 66 | * The returned physical address is the physical (CPU) mapping for |
| 67 | * the memory address given. It is only valid to use this function on | 67 | * the memory address given. It is only valid to use this function on |
| 68 | * addresses directly mapped or allocated via kmalloc. | 68 | * addresses directly mapped or allocated via kmalloc. |
| 69 | * | 69 | * |
| 70 | * This function does not give bus mappings for DMA transfers. In | 70 | * This function does not give bus mappings for DMA transfers. In |
| 71 | * almost all conceivable cases a device driver should not be using | 71 | * almost all conceivable cases a device driver should not be using |
| 72 | * this function | 72 | * this function |
| 73 | */ | 73 | */ |
| 74 | 74 | ||
| 75 | static inline unsigned long virt_to_phys(volatile void * address) | 75 | static inline unsigned long virt_to_phys(volatile void *address) |
| 76 | { | 76 | { |
| 77 | return __pa(address); | 77 | return __pa(address); |
| 78 | } | 78 | } |
| @@ -90,7 +90,7 @@ static inline unsigned long virt_to_phys(volatile void * address) | |||
| 90 | * this function | 90 | * this function |
| 91 | */ | 91 | */ |
| 92 | 92 | ||
| 93 | static inline void * phys_to_virt(unsigned long address) | 93 | static inline void *phys_to_virt(unsigned long address) |
| 94 | { | 94 | { |
| 95 | return __va(address); | 95 | return __va(address); |
| 96 | } | 96 | } |
| @@ -169,16 +169,19 @@ extern void __iomem *fix_ioremap(unsigned idx, unsigned long phys); | |||
| 169 | 169 | ||
| 170 | static inline unsigned char readb(const volatile void __iomem *addr) | 170 | static inline unsigned char readb(const volatile void __iomem *addr) |
| 171 | { | 171 | { |
| 172 | return *(volatile unsigned char __force *) addr; | 172 | return *(volatile unsigned char __force *)addr; |
| 173 | } | 173 | } |
| 174 | |||
| 174 | static inline unsigned short readw(const volatile void __iomem *addr) | 175 | static inline unsigned short readw(const volatile void __iomem *addr) |
| 175 | { | 176 | { |
| 176 | return *(volatile unsigned short __force *) addr; | 177 | return *(volatile unsigned short __force *)addr; |
| 177 | } | 178 | } |
| 179 | |||
| 178 | static inline unsigned int readl(const volatile void __iomem *addr) | 180 | static inline unsigned int readl(const volatile void __iomem *addr) |
| 179 | { | 181 | { |
| 180 | return *(volatile unsigned int __force *) addr; | 182 | return *(volatile unsigned int __force *) addr; |
| 181 | } | 183 | } |
| 184 | |||
| 182 | #define readb_relaxed(addr) readb(addr) | 185 | #define readb_relaxed(addr) readb(addr) |
| 183 | #define readw_relaxed(addr) readw(addr) | 186 | #define readw_relaxed(addr) readw(addr) |
| 184 | #define readl_relaxed(addr) readl(addr) | 187 | #define readl_relaxed(addr) readl(addr) |
| @@ -188,15 +191,17 @@ static inline unsigned int readl(const volatile void __iomem *addr) | |||
| 188 | 191 | ||
| 189 | static inline void writeb(unsigned char b, volatile void __iomem *addr) | 192 | static inline void writeb(unsigned char b, volatile void __iomem *addr) |
| 190 | { | 193 | { |
| 191 | *(volatile unsigned char __force *) addr = b; | 194 | *(volatile unsigned char __force *)addr = b; |
| 192 | } | 195 | } |
| 196 | |||
| 193 | static inline void writew(unsigned short b, volatile void __iomem *addr) | 197 | static inline void writew(unsigned short b, volatile void __iomem *addr) |
| 194 | { | 198 | { |
| 195 | *(volatile unsigned short __force *) addr = b; | 199 | *(volatile unsigned short __force *)addr = b; |
| 196 | } | 200 | } |
| 201 | |||
| 197 | static inline void writel(unsigned int b, volatile void __iomem *addr) | 202 | static inline void writel(unsigned int b, volatile void __iomem *addr) |
| 198 | { | 203 | { |
| 199 | *(volatile unsigned int __force *) addr = b; | 204 | *(volatile unsigned int __force *)addr = b; |
| 200 | } | 205 | } |
| 201 | #define __raw_writeb writeb | 206 | #define __raw_writeb writeb |
| 202 | #define __raw_writew writew | 207 | #define __raw_writew writew |
| @@ -239,12 +244,12 @@ memcpy_toio(volatile void __iomem *dst, const void *src, int count) | |||
| 239 | * 1. Out of order aware processors | 244 | * 1. Out of order aware processors |
| 240 | * 2. Accidentally out of order processors (PPro errata #51) | 245 | * 2. Accidentally out of order processors (PPro errata #51) |
| 241 | */ | 246 | */ |
| 242 | 247 | ||
| 243 | #if defined(CONFIG_X86_OOSTORE) || defined(CONFIG_X86_PPRO_FENCE) | 248 | #if defined(CONFIG_X86_OOSTORE) || defined(CONFIG_X86_PPRO_FENCE) |
| 244 | 249 | ||
| 245 | static inline void flush_write_buffers(void) | 250 | static inline void flush_write_buffers(void) |
| 246 | { | 251 | { |
| 247 | __asm__ __volatile__ ("lock; addl $0,0(%%esp)": : :"memory"); | 252 | asm volatile("lock; addl $0,0(%%esp)": : :"memory"); |
| 248 | } | 253 | } |
| 249 | 254 | ||
| 250 | #else | 255 | #else |
| @@ -264,7 +269,8 @@ extern void io_delay_init(void); | |||
| 264 | #include <asm/paravirt.h> | 269 | #include <asm/paravirt.h> |
| 265 | #else | 270 | #else |
| 266 | 271 | ||
| 267 | static inline void slow_down_io(void) { | 272 | static inline void slow_down_io(void) |
| 273 | { | ||
| 268 | native_io_delay(); | 274 | native_io_delay(); |
| 269 | #ifdef REALLY_SLOW_IO | 275 | #ifdef REALLY_SLOW_IO |
| 270 | native_io_delay(); | 276 | native_io_delay(); |
| @@ -275,51 +281,74 @@ static inline void slow_down_io(void) { | |||
| 275 | 281 | ||
| 276 | #endif | 282 | #endif |
| 277 | 283 | ||
| 278 | #define __BUILDIO(bwl,bw,type) \ | 284 | #define __BUILDIO(bwl, bw, type) \ |
| 279 | static inline void out##bwl(unsigned type value, int port) { \ | 285 | static inline void out##bwl(unsigned type value, int port) \ |
| 280 | out##bwl##_local(value, port); \ | 286 | { \ |
| 281 | } \ | 287 | out##bwl##_local(value, port); \ |
| 282 | static inline unsigned type in##bwl(int port) { \ | 288 | } \ |
| 283 | return in##bwl##_local(port); \ | 289 | \ |
| 290 | static inline unsigned type in##bwl(int port) \ | ||
| 291 | { \ | ||
| 292 | return in##bwl##_local(port); \ | ||
| 284 | } | 293 | } |
| 285 | 294 | ||
| 286 | #define BUILDIO(bwl,bw,type) \ | 295 | #define BUILDIO(bwl, bw, type) \ |
| 287 | static inline void out##bwl##_local(unsigned type value, int port) { \ | 296 | static inline void out##bwl##_local(unsigned type value, int port) \ |
| 288 | __asm__ __volatile__("out" #bwl " %" #bw "0, %w1" : : "a"(value), "Nd"(port)); \ | 297 | { \ |
| 289 | } \ | 298 | asm volatile("out" #bwl " %" #bw "0, %w1" \ |
| 290 | static inline unsigned type in##bwl##_local(int port) { \ | 299 | : : "a"(value), "Nd"(port)); \ |
| 291 | unsigned type value; \ | 300 | } \ |
| 292 | __asm__ __volatile__("in" #bwl " %w1, %" #bw "0" : "=a"(value) : "Nd"(port)); \ | 301 | \ |
| 293 | return value; \ | 302 | static inline unsigned type in##bwl##_local(int port) \ |
| 294 | } \ | 303 | { \ |
| 295 | static inline void out##bwl##_local_p(unsigned type value, int port) { \ | 304 | unsigned type value; \ |
| 296 | out##bwl##_local(value, port); \ | 305 | asm volatile("in" #bwl " %w1, %" #bw "0" \ |
| 297 | slow_down_io(); \ | 306 | : "=a"(value) : "Nd"(port)); \ |
| 298 | } \ | 307 | return value; \ |
| 299 | static inline unsigned type in##bwl##_local_p(int port) { \ | 308 | } \ |
| 300 | unsigned type value = in##bwl##_local(port); \ | 309 | \ |
| 301 | slow_down_io(); \ | 310 | static inline void out##bwl##_local_p(unsigned type value, int port) \ |
| 302 | return value; \ | 311 | { \ |
| 303 | } \ | 312 | out##bwl##_local(value, port); \ |
| 304 | __BUILDIO(bwl,bw,type) \ | 313 | slow_down_io(); \ |
| 305 | static inline void out##bwl##_p(unsigned type value, int port) { \ | 314 | } \ |
| 306 | out##bwl(value, port); \ | 315 | \ |
| 307 | slow_down_io(); \ | 316 | static inline unsigned type in##bwl##_local_p(int port) \ |
| 308 | } \ | 317 | { \ |
| 309 | static inline unsigned type in##bwl##_p(int port) { \ | 318 | unsigned type value = in##bwl##_local(port); \ |
| 310 | unsigned type value = in##bwl(port); \ | 319 | slow_down_io(); \ |
| 311 | slow_down_io(); \ | 320 | return value; \ |
| 312 | return value; \ | 321 | } \ |
| 313 | } \ | 322 | \ |
| 314 | static inline void outs##bwl(int port, const void *addr, unsigned long count) { \ | 323 | __BUILDIO(bwl, bw, type) \ |
| 315 | __asm__ __volatile__("rep; outs" #bwl : "+S"(addr), "+c"(count) : "d"(port)); \ | 324 | \ |
| 316 | } \ | 325 | static inline void out##bwl##_p(unsigned type value, int port) \ |
| 317 | static inline void ins##bwl(int port, void *addr, unsigned long count) { \ | 326 | { \ |
| 318 | __asm__ __volatile__("rep; ins" #bwl : "+D"(addr), "+c"(count) : "d"(port)); \ | 327 | out##bwl(value, port); \ |
| 328 | slow_down_io(); \ | ||
| 329 | } \ | ||
| 330 | \ | ||
| 331 | static inline unsigned type in##bwl##_p(int port) \ | ||
| 332 | { \ | ||
| 333 | unsigned type value = in##bwl(port); \ | ||
| 334 | slow_down_io(); \ | ||
| 335 | return value; \ | ||
| 336 | } \ | ||
| 337 | \ | ||
| 338 | static inline void outs##bwl(int port, const void *addr, unsigned long count) \ | ||
| 339 | { \ | ||
| 340 | asm volatile("rep; outs" #bwl \ | ||
| 341 | : "+S"(addr), "+c"(count) : "d"(port)); \ | ||
| 342 | } \ | ||
| 343 | \ | ||
| 344 | static inline void ins##bwl(int port, void *addr, unsigned long count) \ | ||
| 345 | { \ | ||
| 346 | asm volatile("rep; ins" #bwl \ | ||
| 347 | : "+D"(addr), "+c"(count) : "d"(port)); \ | ||
| 319 | } | 348 | } |
| 320 | 349 | ||
| 321 | BUILDIO(b,b,char) | 350 | BUILDIO(b, b, char) |
| 322 | BUILDIO(w,w,short) | 351 | BUILDIO(w, w, short) |
| 323 | BUILDIO(l,,int) | 352 | BUILDIO(l, , int) |
| 324 | 353 | ||
| 325 | #endif | 354 | #endif |
diff --git a/include/asm-x86/io_64.h b/include/asm-x86/io_64.h index db0be2011a3c..c2f5eef47b88 100644 --- a/include/asm-x86/io_64.h +++ b/include/asm-x86/io_64.h | |||
| @@ -58,60 +58,75 @@ static inline void slow_down_io(void) | |||
| 58 | /* | 58 | /* |
| 59 | * Talk about misusing macros.. | 59 | * Talk about misusing macros.. |
| 60 | */ | 60 | */ |
| 61 | #define __OUT1(s,x) \ | 61 | #define __OUT1(s, x) \ |
| 62 | static inline void out##s(unsigned x value, unsigned short port) { | 62 | static inline void out##s(unsigned x value, unsigned short port) { |
| 63 | 63 | ||
| 64 | #define __OUT2(s,s1,s2) \ | 64 | #define __OUT2(s, s1, s2) \ |
| 65 | __asm__ __volatile__ ("out" #s " %" s1 "0,%" s2 "1" | 65 | asm volatile ("out" #s " %" s1 "0,%" s2 "1" |
| 66 | 66 | ||
| 67 | #ifndef REALLY_SLOW_IO | 67 | #ifndef REALLY_SLOW_IO |
| 68 | #define REALLY_SLOW_IO | 68 | #define REALLY_SLOW_IO |
| 69 | #define UNSET_REALLY_SLOW_IO | 69 | #define UNSET_REALLY_SLOW_IO |
| 70 | #endif | 70 | #endif |
| 71 | 71 | ||
| 72 | #define __OUT(s,s1,x) \ | 72 | #define __OUT(s, s1, x) \ |
| 73 | __OUT1(s,x) __OUT2(s,s1,"w") : : "a" (value), "Nd" (port)); } \ | 73 | __OUT1(s, x) __OUT2(s, s1, "w") : : "a" (value), "Nd" (port)); \ |
| 74 | __OUT1(s##_p, x) __OUT2(s, s1, "w") : : "a" (value), "Nd" (port)); \ | 74 | } \ |
| 75 | slow_down_io(); } | 75 | __OUT1(s##_p, x) __OUT2(s, s1, "w") : : "a" (value), "Nd" (port)); \ |
| 76 | slow_down_io(); \ | ||
| 77 | } | ||
| 76 | 78 | ||
| 77 | #define __IN1(s) \ | 79 | #define __IN1(s) \ |
| 78 | static inline RETURN_TYPE in##s(unsigned short port) { RETURN_TYPE _v; | 80 | static inline RETURN_TYPE in##s(unsigned short port) \ |
| 81 | { \ | ||
| 82 | RETURN_TYPE _v; | ||
| 79 | 83 | ||
| 80 | #define __IN2(s,s1,s2) \ | 84 | #define __IN2(s, s1, s2) \ |
| 81 | __asm__ __volatile__ ("in" #s " %" s2 "1,%" s1 "0" | 85 | asm volatile ("in" #s " %" s2 "1,%" s1 "0" |
| 82 | 86 | ||
| 83 | #define __IN(s,s1,i...) \ | 87 | #define __IN(s, s1, i...) \ |
| 84 | __IN1(s) __IN2(s, s1, "w") : "=a" (_v) : "Nd" (port), ##i); return _v; } \ | 88 | __IN1(s) __IN2(s, s1, "w") : "=a" (_v) : "Nd" (port), ##i); \ |
| 85 | __IN1(s##_p) __IN2(s, s1, "w") : "=a" (_v) : "Nd" (port), ##i); \ | 89 | return _v; \ |
| 86 | slow_down_io(); return _v; } | 90 | } \ |
| 91 | __IN1(s##_p) __IN2(s, s1, "w") : "=a" (_v) : "Nd" (port), ##i); \ | ||
| 92 | slow_down_io(); \ | ||
| 93 | return _v; } | ||
| 87 | 94 | ||
| 88 | #ifdef UNSET_REALLY_SLOW_IO | 95 | #ifdef UNSET_REALLY_SLOW_IO |
| 89 | #undef REALLY_SLOW_IO | 96 | #undef REALLY_SLOW_IO |
| 90 | #endif | 97 | #endif |
| 91 | 98 | ||
| 92 | #define __INS(s) \ | 99 | #define __INS(s) \ |
| 93 | static inline void ins##s(unsigned short port, void * addr, unsigned long count) \ | 100 | static inline void ins##s(unsigned short port, void *addr, \ |
| 94 | { __asm__ __volatile__ ("rep ; ins" #s \ | 101 | unsigned long count) \ |
| 95 | : "=D" (addr), "=c" (count) : "d" (port),"0" (addr),"1" (count)); } | 102 | { \ |
| 103 | asm volatile ("rep ; ins" #s \ | ||
| 104 | : "=D" (addr), "=c" (count) \ | ||
| 105 | : "d" (port), "0" (addr), "1" (count)); \ | ||
| 106 | } | ||
| 96 | 107 | ||
| 97 | #define __OUTS(s) \ | 108 | #define __OUTS(s) \ |
| 98 | static inline void outs##s(unsigned short port, const void * addr, unsigned long count) \ | 109 | static inline void outs##s(unsigned short port, const void *addr, \ |
| 99 | { __asm__ __volatile__ ("rep ; outs" #s \ | 110 | unsigned long count) \ |
| 100 | : "=S" (addr), "=c" (count) : "d" (port),"0" (addr),"1" (count)); } | 111 | { \ |
| 112 | asm volatile ("rep ; outs" #s \ | ||
| 113 | : "=S" (addr), "=c" (count) \ | ||
| 114 | : "d" (port), "0" (addr), "1" (count)); \ | ||
| 115 | } | ||
| 101 | 116 | ||
| 102 | #define RETURN_TYPE unsigned char | 117 | #define RETURN_TYPE unsigned char |
| 103 | __IN(b,"") | 118 | __IN(b, "") |
| 104 | #undef RETURN_TYPE | 119 | #undef RETURN_TYPE |
| 105 | #define RETURN_TYPE unsigned short | 120 | #define RETURN_TYPE unsigned short |
| 106 | __IN(w,"") | 121 | __IN(w, "") |
| 107 | #undef RETURN_TYPE | 122 | #undef RETURN_TYPE |
| 108 | #define RETURN_TYPE unsigned int | 123 | #define RETURN_TYPE unsigned int |
| 109 | __IN(l,"") | 124 | __IN(l, "") |
| 110 | #undef RETURN_TYPE | 125 | #undef RETURN_TYPE |
| 111 | 126 | ||
| 112 | __OUT(b,"b",char) | 127 | __OUT(b, "b", char) |
| 113 | __OUT(w,"w",short) | 128 | __OUT(w, "w", short) |
| 114 | __OUT(l,,int) | 129 | __OUT(l, , int) |
| 115 | 130 | ||
| 116 | __INS(b) | 131 | __INS(b) |
| 117 | __INS(w) | 132 | __INS(w) |
| @@ -132,12 +147,12 @@ __OUTS(l) | |||
| 132 | * Change virtual addresses to physical addresses and vv. | 147 | * Change virtual addresses to physical addresses and vv. |
| 133 | * These are pretty trivial | 148 | * These are pretty trivial |
| 134 | */ | 149 | */ |
| 135 | static inline unsigned long virt_to_phys(volatile void * address) | 150 | static inline unsigned long virt_to_phys(volatile void *address) |
| 136 | { | 151 | { |
| 137 | return __pa(address); | 152 | return __pa(address); |
| 138 | } | 153 | } |
| 139 | 154 | ||
| 140 | static inline void * phys_to_virt(unsigned long address) | 155 | static inline void *phys_to_virt(unsigned long address) |
| 141 | { | 156 | { |
| 142 | return __va(address); | 157 | return __va(address); |
| 143 | } | 158 | } |
| @@ -200,18 +215,22 @@ static inline __u8 __readb(const volatile void __iomem *addr) | |||
| 200 | { | 215 | { |
| 201 | return *(__force volatile __u8 *)addr; | 216 | return *(__force volatile __u8 *)addr; |
| 202 | } | 217 | } |
| 218 | |||
| 203 | static inline __u16 __readw(const volatile void __iomem *addr) | 219 | static inline __u16 __readw(const volatile void __iomem *addr) |
| 204 | { | 220 | { |
| 205 | return *(__force volatile __u16 *)addr; | 221 | return *(__force volatile __u16 *)addr; |
| 206 | } | 222 | } |
| 223 | |||
| 207 | static __always_inline __u32 __readl(const volatile void __iomem *addr) | 224 | static __always_inline __u32 __readl(const volatile void __iomem *addr) |
| 208 | { | 225 | { |
| 209 | return *(__force volatile __u32 *)addr; | 226 | return *(__force volatile __u32 *)addr; |
| 210 | } | 227 | } |
| 228 | |||
| 211 | static inline __u64 __readq(const volatile void __iomem *addr) | 229 | static inline __u64 __readq(const volatile void __iomem *addr) |
| 212 | { | 230 | { |
| 213 | return *(__force volatile __u64 *)addr; | 231 | return *(__force volatile __u64 *)addr; |
| 214 | } | 232 | } |
| 233 | |||
| 215 | #define readb(x) __readb(x) | 234 | #define readb(x) __readb(x) |
| 216 | #define readw(x) __readw(x) | 235 | #define readw(x) __readw(x) |
| 217 | #define readl(x) __readl(x) | 236 | #define readl(x) __readl(x) |
| @@ -231,37 +250,44 @@ static inline void __writel(__u32 b, volatile void __iomem *addr) | |||
| 231 | { | 250 | { |
| 232 | *(__force volatile __u32 *)addr = b; | 251 | *(__force volatile __u32 *)addr = b; |
| 233 | } | 252 | } |
| 253 | |||
| 234 | static inline void __writeq(__u64 b, volatile void __iomem *addr) | 254 | static inline void __writeq(__u64 b, volatile void __iomem *addr) |
| 235 | { | 255 | { |
| 236 | *(__force volatile __u64 *)addr = b; | 256 | *(__force volatile __u64 *)addr = b; |
| 237 | } | 257 | } |
| 258 | |||
| 238 | static inline void __writeb(__u8 b, volatile void __iomem *addr) | 259 | static inline void __writeb(__u8 b, volatile void __iomem *addr) |
| 239 | { | 260 | { |
| 240 | *(__force volatile __u8 *)addr = b; | 261 | *(__force volatile __u8 *)addr = b; |
| 241 | } | 262 | } |
| 263 | |||
| 242 | static inline void __writew(__u16 b, volatile void __iomem *addr) | 264 | static inline void __writew(__u16 b, volatile void __iomem *addr) |
| 243 | { | 265 | { |
| 244 | *(__force volatile __u16 *)addr = b; | 266 | *(__force volatile __u16 *)addr = b; |
| 245 | } | 267 | } |
| 246 | #define writeq(val,addr) __writeq((val),(addr)) | 268 | |
| 247 | #define writel(val,addr) __writel((val),(addr)) | 269 | #define writeq(val, addr) __writeq((val), (addr)) |
| 248 | #define writew(val,addr) __writew((val),(addr)) | 270 | #define writel(val, addr) __writel((val), (addr)) |
| 249 | #define writeb(val,addr) __writeb((val),(addr)) | 271 | #define writew(val, addr) __writew((val), (addr)) |
| 272 | #define writeb(val, addr) __writeb((val), (addr)) | ||
| 250 | #define __raw_writeb writeb | 273 | #define __raw_writeb writeb |
| 251 | #define __raw_writew writew | 274 | #define __raw_writew writew |
| 252 | #define __raw_writel writel | 275 | #define __raw_writel writel |
| 253 | #define __raw_writeq writeq | 276 | #define __raw_writeq writeq |
| 254 | 277 | ||
| 255 | void __memcpy_fromio(void*,unsigned long,unsigned); | 278 | void __memcpy_fromio(void *, unsigned long, unsigned); |
| 256 | void __memcpy_toio(unsigned long,const void*,unsigned); | 279 | void __memcpy_toio(unsigned long, const void *, unsigned); |
| 257 | 280 | ||
| 258 | static inline void memcpy_fromio(void *to, const volatile void __iomem *from, unsigned len) | 281 | static inline void memcpy_fromio(void *to, const volatile void __iomem *from, |
| 282 | unsigned len) | ||
| 259 | { | 283 | { |
| 260 | __memcpy_fromio(to,(unsigned long)from,len); | 284 | __memcpy_fromio(to, (unsigned long)from, len); |
| 261 | } | 285 | } |
| 262 | static inline void memcpy_toio(volatile void __iomem *to, const void *from, unsigned len) | 286 | |
| 287 | static inline void memcpy_toio(volatile void __iomem *to, const void *from, | ||
| 288 | unsigned len) | ||
| 263 | { | 289 | { |
| 264 | __memcpy_toio((unsigned long)to,from,len); | 290 | __memcpy_toio((unsigned long)to, from, len); |
| 265 | } | 291 | } |
| 266 | 292 | ||
| 267 | void memset_io(volatile void __iomem *a, int b, size_t c); | 293 | void memset_io(volatile void __iomem *a, int b, size_t c); |
| @@ -276,7 +302,7 @@ void memset_io(volatile void __iomem *a, int b, size_t c); | |||
| 276 | */ | 302 | */ |
| 277 | #define __ISA_IO_base ((char __iomem *)(PAGE_OFFSET)) | 303 | #define __ISA_IO_base ((char __iomem *)(PAGE_OFFSET)) |
| 278 | 304 | ||
| 279 | #define flush_write_buffers() | 305 | #define flush_write_buffers() |
| 280 | 306 | ||
| 281 | extern int iommu_bio_merge; | 307 | extern int iommu_bio_merge; |
| 282 | #define BIO_VMERGE_BOUNDARY iommu_bio_merge | 308 | #define BIO_VMERGE_BOUNDARY iommu_bio_merge |
diff --git a/include/asm-x86/io_apic.h b/include/asm-x86/io_apic.h index 0f5b3fef0b08..0c9e17c73e05 100644 --- a/include/asm-x86/io_apic.h +++ b/include/asm-x86/io_apic.h | |||
| @@ -110,6 +110,13 @@ extern int nr_ioapic_registers[MAX_IO_APICS]; | |||
| 110 | * MP-BIOS irq configuration table structures: | 110 | * MP-BIOS irq configuration table structures: |
| 111 | */ | 111 | */ |
| 112 | 112 | ||
| 113 | struct mp_ioapic_routing { | ||
| 114 | int apic_id; | ||
| 115 | int gsi_base; | ||
| 116 | int gsi_end; | ||
| 117 | u32 pin_programmed[4]; | ||
| 118 | }; | ||
| 119 | |||
| 113 | /* I/O APIC entries */ | 120 | /* I/O APIC entries */ |
| 114 | extern struct mpc_config_ioapic mp_ioapics[MAX_IO_APICS]; | 121 | extern struct mpc_config_ioapic mp_ioapics[MAX_IO_APICS]; |
| 115 | 122 | ||
| @@ -146,7 +153,6 @@ extern int io_apic_get_version(int ioapic); | |||
| 146 | extern int io_apic_get_redir_entries(int ioapic); | 153 | extern int io_apic_get_redir_entries(int ioapic); |
| 147 | extern int io_apic_set_pci_routing(int ioapic, int pin, int irq, | 154 | extern int io_apic_set_pci_routing(int ioapic, int pin, int irq, |
| 148 | int edge_level, int active_high_low); | 155 | int edge_level, int active_high_low); |
| 149 | extern int timer_uses_ioapic_pin_0; | ||
| 150 | #endif /* CONFIG_ACPI */ | 156 | #endif /* CONFIG_ACPI */ |
| 151 | 157 | ||
| 152 | extern int (*ioapic_renumber_irq)(int ioapic, int irq); | 158 | extern int (*ioapic_renumber_irq)(int ioapic, int irq); |
diff --git a/include/asm-x86/ioctls.h b/include/asm-x86/ioctls.h index 93c894dc5154..c0c338bd4068 100644 --- a/include/asm-x86/ioctls.h +++ b/include/asm-x86/ioctls.h | |||
| @@ -47,12 +47,13 @@ | |||
| 47 | #define TIOCSBRK 0x5427 /* BSD compatibility */ | 47 | #define TIOCSBRK 0x5427 /* BSD compatibility */ |
| 48 | #define TIOCCBRK 0x5428 /* BSD compatibility */ | 48 | #define TIOCCBRK 0x5428 /* BSD compatibility */ |
| 49 | #define TIOCGSID 0x5429 /* Return the session ID of FD */ | 49 | #define TIOCGSID 0x5429 /* Return the session ID of FD */ |
| 50 | #define TCGETS2 _IOR('T',0x2A, struct termios2) | 50 | #define TCGETS2 _IOR('T', 0x2A, struct termios2) |
| 51 | #define TCSETS2 _IOW('T',0x2B, struct termios2) | 51 | #define TCSETS2 _IOW('T', 0x2B, struct termios2) |
| 52 | #define TCSETSW2 _IOW('T',0x2C, struct termios2) | 52 | #define TCSETSW2 _IOW('T', 0x2C, struct termios2) |
| 53 | #define TCSETSF2 _IOW('T',0x2D, struct termios2) | 53 | #define TCSETSF2 _IOW('T', 0x2D, struct termios2) |
| 54 | #define TIOCGPTN _IOR('T',0x30, unsigned int) /* Get Pty Number (of pty-mux device) */ | 54 | #define TIOCGPTN _IOR('T', 0x30, unsigned int) |
| 55 | #define TIOCSPTLCK _IOW('T',0x31, int) /* Lock/unlock Pty */ | 55 | /* Get Pty Number (of pty-mux device) */ |
| 56 | #define TIOCSPTLCK _IOW('T', 0x31, int) /* Lock/unlock Pty */ | ||
| 56 | 57 | ||
| 57 | #define FIONCLEX 0x5450 | 58 | #define FIONCLEX 0x5450 |
| 58 | #define FIOCLEX 0x5451 | 59 | #define FIOCLEX 0x5451 |
diff --git a/include/asm-x86/ipcbuf.h b/include/asm-x86/ipcbuf.h index 2adf8b39a40b..ee678fd51594 100644 --- a/include/asm-x86/ipcbuf.h +++ b/include/asm-x86/ipcbuf.h | |||
| @@ -11,8 +11,7 @@ | |||
| 11 | * - 2 miscellaneous 32-bit values | 11 | * - 2 miscellaneous 32-bit values |
| 12 | */ | 12 | */ |
| 13 | 13 | ||
| 14 | struct ipc64_perm | 14 | struct ipc64_perm { |
| 15 | { | ||
| 16 | __kernel_key_t key; | 15 | __kernel_key_t key; |
| 17 | __kernel_uid32_t uid; | 16 | __kernel_uid32_t uid; |
| 18 | __kernel_gid32_t gid; | 17 | __kernel_gid32_t gid; |
diff --git a/include/asm-x86/ipi.h b/include/asm-x86/ipi.h index 6d011bd6067d..ecc80f341f37 100644 --- a/include/asm-x86/ipi.h +++ b/include/asm-x86/ipi.h | |||
| @@ -27,7 +27,8 @@ | |||
| 27 | * We use 'broadcast', CPU->CPU IPIs and self-IPIs too. | 27 | * We use 'broadcast', CPU->CPU IPIs and self-IPIs too. |
| 28 | */ | 28 | */ |
| 29 | 29 | ||
| 30 | static inline unsigned int __prepare_ICR (unsigned int shortcut, int vector, unsigned int dest) | 30 | static inline unsigned int __prepare_ICR(unsigned int shortcut, int vector, |
| 31 | unsigned int dest) | ||
| 31 | { | 32 | { |
| 32 | unsigned int icr = shortcut | dest; | 33 | unsigned int icr = shortcut | dest; |
| 33 | 34 | ||
| @@ -42,12 +43,13 @@ static inline unsigned int __prepare_ICR (unsigned int shortcut, int vector, uns | |||
| 42 | return icr; | 43 | return icr; |
| 43 | } | 44 | } |
| 44 | 45 | ||
| 45 | static inline int __prepare_ICR2 (unsigned int mask) | 46 | static inline int __prepare_ICR2(unsigned int mask) |
| 46 | { | 47 | { |
| 47 | return SET_APIC_DEST_FIELD(mask); | 48 | return SET_APIC_DEST_FIELD(mask); |
| 48 | } | 49 | } |
| 49 | 50 | ||
| 50 | static inline void __send_IPI_shortcut(unsigned int shortcut, int vector, unsigned int dest) | 51 | static inline void __send_IPI_shortcut(unsigned int shortcut, int vector, |
| 52 | unsigned int dest) | ||
| 51 | { | 53 | { |
| 52 | /* | 54 | /* |
| 53 | * Subtle. In the case of the 'never do double writes' workaround | 55 | * Subtle. In the case of the 'never do double writes' workaround |
| @@ -78,7 +80,8 @@ static inline void __send_IPI_shortcut(unsigned int shortcut, int vector, unsign | |||
| 78 | * This is used to send an IPI with no shorthand notation (the destination is | 80 | * This is used to send an IPI with no shorthand notation (the destination is |
| 79 | * specified in bits 56 to 63 of the ICR). | 81 | * specified in bits 56 to 63 of the ICR). |
| 80 | */ | 82 | */ |
| 81 | static inline void __send_IPI_dest_field(unsigned int mask, int vector, unsigned int dest) | 83 | static inline void __send_IPI_dest_field(unsigned int mask, int vector, |
| 84 | unsigned int dest) | ||
| 82 | { | 85 | { |
| 83 | unsigned long cfg; | 86 | unsigned long cfg; |
| 84 | 87 | ||
diff --git a/include/asm-x86/irq_32.h b/include/asm-x86/irq_32.h index aca9c96e8e6b..0b79f3185243 100644 --- a/include/asm-x86/irq_32.h +++ b/include/asm-x86/irq_32.h | |||
| @@ -15,7 +15,7 @@ | |||
| 15 | #include "irq_vectors.h" | 15 | #include "irq_vectors.h" |
| 16 | #include <asm/thread_info.h> | 16 | #include <asm/thread_info.h> |
| 17 | 17 | ||
| 18 | static __inline__ int irq_canonicalize(int irq) | 18 | static inline int irq_canonicalize(int irq) |
| 19 | { | 19 | { |
| 20 | return ((irq == 2) ? 9 : irq); | 20 | return ((irq == 2) ? 9 : irq); |
| 21 | } | 21 | } |
diff --git a/include/asm-x86/irq_64.h b/include/asm-x86/irq_64.h index 5006c6e75656..083d35a62c94 100644 --- a/include/asm-x86/irq_64.h +++ b/include/asm-x86/irq_64.h | |||
| @@ -31,10 +31,10 @@ | |||
| 31 | 31 | ||
| 32 | #define FIRST_SYSTEM_VECTOR 0xef /* duplicated in hw_irq.h */ | 32 | #define FIRST_SYSTEM_VECTOR 0xef /* duplicated in hw_irq.h */ |
| 33 | 33 | ||
| 34 | #define NR_IRQS (NR_VECTORS + (32 *NR_CPUS)) | 34 | #define NR_IRQS (NR_VECTORS + (32 * NR_CPUS)) |
| 35 | #define NR_IRQ_VECTORS NR_IRQS | 35 | #define NR_IRQ_VECTORS NR_IRQS |
| 36 | 36 | ||
| 37 | static __inline__ int irq_canonicalize(int irq) | 37 | static inline int irq_canonicalize(int irq) |
| 38 | { | 38 | { |
| 39 | return ((irq == 2) ? 9 : irq); | 39 | return ((irq == 2) ? 9 : irq); |
| 40 | } | 40 | } |
diff --git a/include/asm-x86/irqflags.h b/include/asm-x86/irqflags.h index 0e2292483b35..c242527f970e 100644 --- a/include/asm-x86/irqflags.h +++ b/include/asm-x86/irqflags.h | |||
| @@ -12,25 +12,21 @@ static inline unsigned long native_save_fl(void) | |||
| 12 | { | 12 | { |
| 13 | unsigned long flags; | 13 | unsigned long flags; |
| 14 | 14 | ||
| 15 | __asm__ __volatile__( | 15 | asm volatile("# __raw_save_flags\n\t" |
| 16 | "# __raw_save_flags\n\t" | 16 | "pushf ; pop %0" |
| 17 | "pushf ; pop %0" | 17 | : "=g" (flags) |
| 18 | : "=g" (flags) | 18 | : /* no input */ |
| 19 | : /* no input */ | 19 | : "memory"); |
| 20 | : "memory" | ||
| 21 | ); | ||
| 22 | 20 | ||
| 23 | return flags; | 21 | return flags; |
| 24 | } | 22 | } |
| 25 | 23 | ||
| 26 | static inline void native_restore_fl(unsigned long flags) | 24 | static inline void native_restore_fl(unsigned long flags) |
| 27 | { | 25 | { |
| 28 | __asm__ __volatile__( | 26 | asm volatile("push %0 ; popf" |
| 29 | "push %0 ; popf" | 27 | : /* no output */ |
| 30 | : /* no output */ | 28 | :"g" (flags) |
| 31 | :"g" (flags) | 29 | :"memory", "cc"); |
| 32 | :"memory", "cc" | ||
| 33 | ); | ||
| 34 | } | 30 | } |
| 35 | 31 | ||
| 36 | static inline void native_irq_disable(void) | 32 | static inline void native_irq_disable(void) |
| @@ -70,26 +66,6 @@ static inline void raw_local_irq_restore(unsigned long flags) | |||
| 70 | native_restore_fl(flags); | 66 | native_restore_fl(flags); |
| 71 | } | 67 | } |
| 72 | 68 | ||
| 73 | #ifdef CONFIG_X86_VSMP | ||
| 74 | |||
| 75 | /* | ||
| 76 | * Interrupt control for the VSMP architecture: | ||
| 77 | */ | ||
| 78 | |||
| 79 | static inline void raw_local_irq_disable(void) | ||
| 80 | { | ||
| 81 | unsigned long flags = __raw_local_save_flags(); | ||
| 82 | raw_local_irq_restore((flags & ~X86_EFLAGS_IF) | X86_EFLAGS_AC); | ||
| 83 | } | ||
| 84 | |||
| 85 | static inline void raw_local_irq_enable(void) | ||
| 86 | { | ||
| 87 | unsigned long flags = __raw_local_save_flags(); | ||
| 88 | raw_local_irq_restore((flags | X86_EFLAGS_IF) & (~X86_EFLAGS_AC)); | ||
| 89 | } | ||
| 90 | |||
| 91 | #else | ||
| 92 | |||
| 93 | static inline void raw_local_irq_disable(void) | 69 | static inline void raw_local_irq_disable(void) |
| 94 | { | 70 | { |
| 95 | native_irq_disable(); | 71 | native_irq_disable(); |
| @@ -100,8 +76,6 @@ static inline void raw_local_irq_enable(void) | |||
| 100 | native_irq_enable(); | 76 | native_irq_enable(); |
| 101 | } | 77 | } |
| 102 | 78 | ||
| 103 | #endif | ||
| 104 | |||
| 105 | /* | 79 | /* |
| 106 | * Used in the idle loop; sti takes one instruction cycle | 80 | * Used in the idle loop; sti takes one instruction cycle |
| 107 | * to complete: | 81 | * to complete: |
| @@ -153,23 +127,16 @@ static inline unsigned long __raw_local_irq_save(void) | |||
| 153 | #endif /* CONFIG_PARAVIRT */ | 127 | #endif /* CONFIG_PARAVIRT */ |
| 154 | 128 | ||
| 155 | #ifndef __ASSEMBLY__ | 129 | #ifndef __ASSEMBLY__ |
| 156 | #define raw_local_save_flags(flags) \ | 130 | #define raw_local_save_flags(flags) \ |
| 157 | do { (flags) = __raw_local_save_flags(); } while (0) | 131 | do { (flags) = __raw_local_save_flags(); } while (0) |
| 158 | 132 | ||
| 159 | #define raw_local_irq_save(flags) \ | 133 | #define raw_local_irq_save(flags) \ |
| 160 | do { (flags) = __raw_local_irq_save(); } while (0) | 134 | do { (flags) = __raw_local_irq_save(); } while (0) |
| 161 | 135 | ||
| 162 | #ifdef CONFIG_X86_VSMP | ||
| 163 | static inline int raw_irqs_disabled_flags(unsigned long flags) | ||
| 164 | { | ||
| 165 | return !(flags & X86_EFLAGS_IF) || (flags & X86_EFLAGS_AC); | ||
| 166 | } | ||
| 167 | #else | ||
| 168 | static inline int raw_irqs_disabled_flags(unsigned long flags) | 136 | static inline int raw_irqs_disabled_flags(unsigned long flags) |
| 169 | { | 137 | { |
| 170 | return !(flags & X86_EFLAGS_IF); | 138 | return !(flags & X86_EFLAGS_IF); |
| 171 | } | 139 | } |
| 172 | #endif | ||
| 173 | 140 | ||
| 174 | static inline int raw_irqs_disabled(void) | 141 | static inline int raw_irqs_disabled(void) |
| 175 | { | 142 | { |
diff --git a/include/asm-x86/kdebug.h b/include/asm-x86/kdebug.h index 99dcbafa1511..0c4175390dab 100644 --- a/include/asm-x86/kdebug.h +++ b/include/asm-x86/kdebug.h | |||
| @@ -23,12 +23,12 @@ enum die_val { | |||
| 23 | }; | 23 | }; |
| 24 | 24 | ||
| 25 | extern void printk_address(unsigned long address, int reliable); | 25 | extern void printk_address(unsigned long address, int reliable); |
| 26 | extern void die(const char *,struct pt_regs *,long); | 26 | extern void die(const char *, struct pt_regs *,long); |
| 27 | extern int __must_check __die(const char *, struct pt_regs *, long); | 27 | extern int __must_check __die(const char *, struct pt_regs *, long); |
| 28 | extern void show_registers(struct pt_regs *regs); | 28 | extern void show_registers(struct pt_regs *regs); |
| 29 | extern void __show_registers(struct pt_regs *, int all); | 29 | extern void __show_registers(struct pt_regs *, int all); |
| 30 | extern void show_trace(struct task_struct *t, struct pt_regs *regs, | 30 | extern void show_trace(struct task_struct *t, struct pt_regs *regs, |
| 31 | unsigned long *sp, unsigned long bp); | 31 | unsigned long *sp, unsigned long bp); |
| 32 | extern void __show_regs(struct pt_regs *regs); | 32 | extern void __show_regs(struct pt_regs *regs); |
| 33 | extern void show_regs(struct pt_regs *regs); | 33 | extern void show_regs(struct pt_regs *regs); |
| 34 | extern unsigned long oops_begin(void); | 34 | extern unsigned long oops_begin(void); |
diff --git a/include/asm-x86/kexec.h b/include/asm-x86/kexec.h index c90d3c77afc2..8f855a15f64d 100644 --- a/include/asm-x86/kexec.h +++ b/include/asm-x86/kexec.h | |||
| @@ -94,10 +94,9 @@ static inline void crash_fixup_ss_esp(struct pt_regs *newregs, | |||
| 94 | { | 94 | { |
| 95 | #ifdef CONFIG_X86_32 | 95 | #ifdef CONFIG_X86_32 |
| 96 | newregs->sp = (unsigned long)&(oldregs->sp); | 96 | newregs->sp = (unsigned long)&(oldregs->sp); |
| 97 | __asm__ __volatile__( | 97 | asm volatile("xorl %%eax, %%eax\n\t" |
| 98 | "xorl %%eax, %%eax\n\t" | 98 | "movw %%ss, %%ax\n\t" |
| 99 | "movw %%ss, %%ax\n\t" | 99 | :"=a"(newregs->ss)); |
| 100 | :"=a"(newregs->ss)); | ||
| 101 | #endif | 100 | #endif |
| 102 | } | 101 | } |
| 103 | 102 | ||
| @@ -114,39 +113,39 @@ static inline void crash_setup_regs(struct pt_regs *newregs, | |||
| 114 | crash_fixup_ss_esp(newregs, oldregs); | 113 | crash_fixup_ss_esp(newregs, oldregs); |
| 115 | } else { | 114 | } else { |
| 116 | #ifdef CONFIG_X86_32 | 115 | #ifdef CONFIG_X86_32 |
| 117 | __asm__ __volatile__("movl %%ebx,%0" : "=m"(newregs->bx)); | 116 | asm volatile("movl %%ebx,%0" : "=m"(newregs->bx)); |
| 118 | __asm__ __volatile__("movl %%ecx,%0" : "=m"(newregs->cx)); | 117 | asm volatile("movl %%ecx,%0" : "=m"(newregs->cx)); |
| 119 | __asm__ __volatile__("movl %%edx,%0" : "=m"(newregs->dx)); | 118 | asm volatile("movl %%edx,%0" : "=m"(newregs->dx)); |
| 120 | __asm__ __volatile__("movl %%esi,%0" : "=m"(newregs->si)); | 119 | asm volatile("movl %%esi,%0" : "=m"(newregs->si)); |
| 121 | __asm__ __volatile__("movl %%edi,%0" : "=m"(newregs->di)); | 120 | asm volatile("movl %%edi,%0" : "=m"(newregs->di)); |
| 122 | __asm__ __volatile__("movl %%ebp,%0" : "=m"(newregs->bp)); | 121 | asm volatile("movl %%ebp,%0" : "=m"(newregs->bp)); |
| 123 | __asm__ __volatile__("movl %%eax,%0" : "=m"(newregs->ax)); | 122 | asm volatile("movl %%eax,%0" : "=m"(newregs->ax)); |
| 124 | __asm__ __volatile__("movl %%esp,%0" : "=m"(newregs->sp)); | 123 | asm volatile("movl %%esp,%0" : "=m"(newregs->sp)); |
| 125 | __asm__ __volatile__("movl %%ss, %%eax;" :"=a"(newregs->ss)); | 124 | asm volatile("movl %%ss, %%eax;" :"=a"(newregs->ss)); |
| 126 | __asm__ __volatile__("movl %%cs, %%eax;" :"=a"(newregs->cs)); | 125 | asm volatile("movl %%cs, %%eax;" :"=a"(newregs->cs)); |
| 127 | __asm__ __volatile__("movl %%ds, %%eax;" :"=a"(newregs->ds)); | 126 | asm volatile("movl %%ds, %%eax;" :"=a"(newregs->ds)); |
| 128 | __asm__ __volatile__("movl %%es, %%eax;" :"=a"(newregs->es)); | 127 | asm volatile("movl %%es, %%eax;" :"=a"(newregs->es)); |
| 129 | __asm__ __volatile__("pushfl; popl %0" :"=m"(newregs->flags)); | 128 | asm volatile("pushfl; popl %0" :"=m"(newregs->flags)); |
| 130 | #else | 129 | #else |
| 131 | __asm__ __volatile__("movq %%rbx,%0" : "=m"(newregs->bx)); | 130 | asm volatile("movq %%rbx,%0" : "=m"(newregs->bx)); |
| 132 | __asm__ __volatile__("movq %%rcx,%0" : "=m"(newregs->cx)); | 131 | asm volatile("movq %%rcx,%0" : "=m"(newregs->cx)); |
| 133 | __asm__ __volatile__("movq %%rdx,%0" : "=m"(newregs->dx)); | 132 | asm volatile("movq %%rdx,%0" : "=m"(newregs->dx)); |
| 134 | __asm__ __volatile__("movq %%rsi,%0" : "=m"(newregs->si)); | 133 | asm volatile("movq %%rsi,%0" : "=m"(newregs->si)); |
| 135 | __asm__ __volatile__("movq %%rdi,%0" : "=m"(newregs->di)); | 134 | asm volatile("movq %%rdi,%0" : "=m"(newregs->di)); |
| 136 | __asm__ __volatile__("movq %%rbp,%0" : "=m"(newregs->bp)); | 135 | asm volatile("movq %%rbp,%0" : "=m"(newregs->bp)); |
| 137 | __asm__ __volatile__("movq %%rax,%0" : "=m"(newregs->ax)); | 136 | asm volatile("movq %%rax,%0" : "=m"(newregs->ax)); |
| 138 | __asm__ __volatile__("movq %%rsp,%0" : "=m"(newregs->sp)); | 137 | asm volatile("movq %%rsp,%0" : "=m"(newregs->sp)); |
| 139 | __asm__ __volatile__("movq %%r8,%0" : "=m"(newregs->r8)); | 138 | asm volatile("movq %%r8,%0" : "=m"(newregs->r8)); |
| 140 | __asm__ __volatile__("movq %%r9,%0" : "=m"(newregs->r9)); | 139 | asm volatile("movq %%r9,%0" : "=m"(newregs->r9)); |
| 141 | __asm__ __volatile__("movq %%r10,%0" : "=m"(newregs->r10)); | 140 | asm volatile("movq %%r10,%0" : "=m"(newregs->r10)); |
| 142 | __asm__ __volatile__("movq %%r11,%0" : "=m"(newregs->r11)); | 141 | asm volatile("movq %%r11,%0" : "=m"(newregs->r11)); |
| 143 | __asm__ __volatile__("movq %%r12,%0" : "=m"(newregs->r12)); | 142 | asm volatile("movq %%r12,%0" : "=m"(newregs->r12)); |
| 144 | __asm__ __volatile__("movq %%r13,%0" : "=m"(newregs->r13)); | 143 | asm volatile("movq %%r13,%0" : "=m"(newregs->r13)); |
| 145 | __asm__ __volatile__("movq %%r14,%0" : "=m"(newregs->r14)); | 144 | asm volatile("movq %%r14,%0" : "=m"(newregs->r14)); |
| 146 | __asm__ __volatile__("movq %%r15,%0" : "=m"(newregs->r15)); | 145 | asm volatile("movq %%r15,%0" : "=m"(newregs->r15)); |
| 147 | __asm__ __volatile__("movl %%ss, %%eax;" :"=a"(newregs->ss)); | 146 | asm volatile("movl %%ss, %%eax;" :"=a"(newregs->ss)); |
| 148 | __asm__ __volatile__("movl %%cs, %%eax;" :"=a"(newregs->cs)); | 147 | asm volatile("movl %%cs, %%eax;" :"=a"(newregs->cs)); |
| 149 | __asm__ __volatile__("pushfq; popq %0" :"=m"(newregs->flags)); | 148 | asm volatile("pushfq; popq %0" :"=m"(newregs->flags)); |
| 150 | #endif | 149 | #endif |
| 151 | newregs->ip = (unsigned long)current_text_addr(); | 150 | newregs->ip = (unsigned long)current_text_addr(); |
| 152 | } | 151 | } |
diff --git a/include/asm-x86/kprobes.h b/include/asm-x86/kprobes.h index 61ad7b5d142e..54980b0b3892 100644 --- a/include/asm-x86/kprobes.h +++ b/include/asm-x86/kprobes.h | |||
| @@ -35,12 +35,12 @@ typedef u8 kprobe_opcode_t; | |||
| 35 | #define RELATIVEJUMP_INSTRUCTION 0xe9 | 35 | #define RELATIVEJUMP_INSTRUCTION 0xe9 |
| 36 | #define MAX_INSN_SIZE 16 | 36 | #define MAX_INSN_SIZE 16 |
| 37 | #define MAX_STACK_SIZE 64 | 37 | #define MAX_STACK_SIZE 64 |
| 38 | #define MIN_STACK_SIZE(ADDR) (((MAX_STACK_SIZE) < \ | 38 | #define MIN_STACK_SIZE(ADDR) \ |
| 39 | (((unsigned long)current_thread_info()) + THREAD_SIZE \ | 39 | (((MAX_STACK_SIZE) < (((unsigned long)current_thread_info()) + \ |
| 40 | - (unsigned long)(ADDR))) \ | 40 | THREAD_SIZE - (unsigned long)(ADDR))) \ |
| 41 | ? (MAX_STACK_SIZE) \ | 41 | ? (MAX_STACK_SIZE) \ |
| 42 | : (((unsigned long)current_thread_info()) + THREAD_SIZE \ | 42 | : (((unsigned long)current_thread_info()) + \ |
| 43 | - (unsigned long)(ADDR))) | 43 | THREAD_SIZE - (unsigned long)(ADDR))) |
| 44 | 44 | ||
| 45 | #define flush_insn_slot(p) do { } while (0) | 45 | #define flush_insn_slot(p) do { } while (0) |
| 46 | 46 | ||
diff --git a/include/asm-x86/kvm_host.h b/include/asm-x86/kvm_host.h index 4702b04b979a..68ee390b2844 100644 --- a/include/asm-x86/kvm_host.h +++ b/include/asm-x86/kvm_host.h | |||
| @@ -22,15 +22,16 @@ | |||
| 22 | 22 | ||
| 23 | #define CR3_PAE_RESERVED_BITS ((X86_CR3_PWT | X86_CR3_PCD) - 1) | 23 | #define CR3_PAE_RESERVED_BITS ((X86_CR3_PWT | X86_CR3_PCD) - 1) |
| 24 | #define CR3_NONPAE_RESERVED_BITS ((PAGE_SIZE-1) & ~(X86_CR3_PWT | X86_CR3_PCD)) | 24 | #define CR3_NONPAE_RESERVED_BITS ((PAGE_SIZE-1) & ~(X86_CR3_PWT | X86_CR3_PCD)) |
| 25 | #define CR3_L_MODE_RESERVED_BITS (CR3_NONPAE_RESERVED_BITS|0xFFFFFF0000000000ULL) | 25 | #define CR3_L_MODE_RESERVED_BITS (CR3_NONPAE_RESERVED_BITS | \ |
| 26 | 0xFFFFFF0000000000ULL) | ||
| 26 | 27 | ||
| 27 | #define KVM_GUEST_CR0_MASK \ | 28 | #define KVM_GUEST_CR0_MASK \ |
| 28 | (X86_CR0_PG | X86_CR0_PE | X86_CR0_WP | X86_CR0_NE \ | 29 | (X86_CR0_PG | X86_CR0_PE | X86_CR0_WP | X86_CR0_NE \ |
| 29 | | X86_CR0_NW | X86_CR0_CD) | 30 | | X86_CR0_NW | X86_CR0_CD) |
| 30 | #define KVM_VM_CR0_ALWAYS_ON \ | 31 | #define KVM_VM_CR0_ALWAYS_ON \ |
| 31 | (X86_CR0_PG | X86_CR0_PE | X86_CR0_WP | X86_CR0_NE | X86_CR0_TS \ | 32 | (X86_CR0_PG | X86_CR0_PE | X86_CR0_WP | X86_CR0_NE | X86_CR0_TS \ |
| 32 | | X86_CR0_MP) | 33 | | X86_CR0_MP) |
| 33 | #define KVM_GUEST_CR4_MASK \ | 34 | #define KVM_GUEST_CR4_MASK \ |
| 34 | (X86_CR4_VME | X86_CR4_PSE | X86_CR4_PAE | X86_CR4_PGE | X86_CR4_VMXE) | 35 | (X86_CR4_VME | X86_CR4_PSE | X86_CR4_PAE | X86_CR4_PGE | X86_CR4_VMXE) |
| 35 | #define KVM_PMODE_VM_CR4_ALWAYS_ON (X86_CR4_PAE | X86_CR4_VMXE) | 36 | #define KVM_PMODE_VM_CR4_ALWAYS_ON (X86_CR4_PAE | X86_CR4_VMXE) |
| 36 | #define KVM_RMODE_VM_CR4_ALWAYS_ON (X86_CR4_VME | X86_CR4_PAE | X86_CR4_VMXE) | 37 | #define KVM_RMODE_VM_CR4_ALWAYS_ON (X86_CR4_VME | X86_CR4_PAE | X86_CR4_VMXE) |
| @@ -133,12 +134,12 @@ struct kvm_pte_chain { | |||
| 133 | union kvm_mmu_page_role { | 134 | union kvm_mmu_page_role { |
| 134 | unsigned word; | 135 | unsigned word; |
| 135 | struct { | 136 | struct { |
| 136 | unsigned glevels : 4; | 137 | unsigned glevels:4; |
| 137 | unsigned level : 4; | 138 | unsigned level:4; |
| 138 | unsigned quadrant : 2; | 139 | unsigned quadrant:2; |
| 139 | unsigned pad_for_nice_hex_output : 6; | 140 | unsigned pad_for_nice_hex_output:6; |
| 140 | unsigned metaphysical : 1; | 141 | unsigned metaphysical:1; |
| 141 | unsigned access : 3; | 142 | unsigned access:3; |
| 142 | }; | 143 | }; |
| 143 | }; | 144 | }; |
| 144 | 145 | ||
| @@ -606,6 +607,7 @@ static inline void kvm_inject_gp(struct kvm_vcpu *vcpu, u32 error_code) | |||
| 606 | #define TSS_BASE_SIZE 0x68 | 607 | #define TSS_BASE_SIZE 0x68 |
| 607 | #define TSS_IOPB_SIZE (65536 / 8) | 608 | #define TSS_IOPB_SIZE (65536 / 8) |
| 608 | #define TSS_REDIRECTION_SIZE (256 / 8) | 609 | #define TSS_REDIRECTION_SIZE (256 / 8) |
| 609 | #define RMODE_TSS_SIZE (TSS_BASE_SIZE + TSS_REDIRECTION_SIZE + TSS_IOPB_SIZE + 1) | 610 | #define RMODE_TSS_SIZE \ |
| 611 | (TSS_BASE_SIZE + TSS_REDIRECTION_SIZE + TSS_IOPB_SIZE + 1) | ||
| 610 | 612 | ||
| 611 | #endif | 613 | #endif |
diff --git a/include/asm-x86/kvm_x86_emulate.h b/include/asm-x86/kvm_x86_emulate.h index 7db91b9bdcd4..d6337f941c98 100644 --- a/include/asm-x86/kvm_x86_emulate.h +++ b/include/asm-x86/kvm_x86_emulate.h | |||
| @@ -68,10 +68,10 @@ struct x86_emulate_ops { | |||
| 68 | * @val: [OUT] Value read from memory, zero-extended to 'u_long'. | 68 | * @val: [OUT] Value read from memory, zero-extended to 'u_long'. |
| 69 | * @bytes: [IN ] Number of bytes to read from memory. | 69 | * @bytes: [IN ] Number of bytes to read from memory. |
| 70 | */ | 70 | */ |
| 71 | int (*read_emulated) (unsigned long addr, | 71 | int (*read_emulated)(unsigned long addr, |
| 72 | void *val, | 72 | void *val, |
| 73 | unsigned int bytes, | 73 | unsigned int bytes, |
| 74 | struct kvm_vcpu *vcpu); | 74 | struct kvm_vcpu *vcpu); |
| 75 | 75 | ||
| 76 | /* | 76 | /* |
| 77 | * write_emulated: Read bytes from emulated/special memory area. | 77 | * write_emulated: Read bytes from emulated/special memory area. |
| @@ -80,10 +80,10 @@ struct x86_emulate_ops { | |||
| 80 | * required). | 80 | * required). |
| 81 | * @bytes: [IN ] Number of bytes to write to memory. | 81 | * @bytes: [IN ] Number of bytes to write to memory. |
| 82 | */ | 82 | */ |
| 83 | int (*write_emulated) (unsigned long addr, | 83 | int (*write_emulated)(unsigned long addr, |
| 84 | const void *val, | 84 | const void *val, |
| 85 | unsigned int bytes, | 85 | unsigned int bytes, |
| 86 | struct kvm_vcpu *vcpu); | 86 | struct kvm_vcpu *vcpu); |
| 87 | 87 | ||
| 88 | /* | 88 | /* |
| 89 | * cmpxchg_emulated: Emulate an atomic (LOCKed) CMPXCHG operation on an | 89 | * cmpxchg_emulated: Emulate an atomic (LOCKed) CMPXCHG operation on an |
| @@ -93,11 +93,11 @@ struct x86_emulate_ops { | |||
| 93 | * @new: [IN ] Value to write to @addr. | 93 | * @new: [IN ] Value to write to @addr. |
| 94 | * @bytes: [IN ] Number of bytes to access using CMPXCHG. | 94 | * @bytes: [IN ] Number of bytes to access using CMPXCHG. |
| 95 | */ | 95 | */ |
| 96 | int (*cmpxchg_emulated) (unsigned long addr, | 96 | int (*cmpxchg_emulated)(unsigned long addr, |
| 97 | const void *old, | 97 | const void *old, |
| 98 | const void *new, | 98 | const void *new, |
| 99 | unsigned int bytes, | 99 | unsigned int bytes, |
| 100 | struct kvm_vcpu *vcpu); | 100 | struct kvm_vcpu *vcpu); |
| 101 | 101 | ||
| 102 | }; | 102 | }; |
| 103 | 103 | ||
| @@ -143,7 +143,7 @@ struct x86_emulate_ctxt { | |||
| 143 | /* Register state before/after emulation. */ | 143 | /* Register state before/after emulation. */ |
| 144 | struct kvm_vcpu *vcpu; | 144 | struct kvm_vcpu *vcpu; |
| 145 | 145 | ||
| 146 | /* Linear faulting address (if emulating a page-faulting instruction). */ | 146 | /* Linear faulting address (if emulating a page-faulting instruction) */ |
| 147 | unsigned long eflags; | 147 | unsigned long eflags; |
| 148 | 148 | ||
| 149 | /* Emulated execution mode, represented by an X86EMUL_MODE value. */ | 149 | /* Emulated execution mode, represented by an X86EMUL_MODE value. */ |
diff --git a/include/asm-x86/lguest.h b/include/asm-x86/lguest.h index 9b17571e9bc3..be4a7247fa2b 100644 --- a/include/asm-x86/lguest.h +++ b/include/asm-x86/lguest.h | |||
| @@ -34,8 +34,7 @@ extern const char lgstart_iret[], lgend_iret[]; | |||
| 34 | extern void lguest_iret(void); | 34 | extern void lguest_iret(void); |
| 35 | extern void lguest_init(void); | 35 | extern void lguest_init(void); |
| 36 | 36 | ||
| 37 | struct lguest_regs | 37 | struct lguest_regs { |
| 38 | { | ||
| 39 | /* Manually saved part. */ | 38 | /* Manually saved part. */ |
| 40 | unsigned long eax, ebx, ecx, edx; | 39 | unsigned long eax, ebx, ecx, edx; |
| 41 | unsigned long esi, edi, ebp; | 40 | unsigned long esi, edi, ebp; |
| @@ -51,8 +50,7 @@ struct lguest_regs | |||
| 51 | }; | 50 | }; |
| 52 | 51 | ||
| 53 | /* This is a guest-specific page (mapped ro) into the guest. */ | 52 | /* This is a guest-specific page (mapped ro) into the guest. */ |
| 54 | struct lguest_ro_state | 53 | struct lguest_ro_state { |
| 55 | { | ||
| 56 | /* Host information we need to restore when we switch back. */ | 54 | /* Host information we need to restore when we switch back. */ |
| 57 | u32 host_cr3; | 55 | u32 host_cr3; |
| 58 | struct desc_ptr host_idt_desc; | 56 | struct desc_ptr host_idt_desc; |
| @@ -67,8 +65,7 @@ struct lguest_ro_state | |||
| 67 | struct desc_struct guest_gdt[GDT_ENTRIES]; | 65 | struct desc_struct guest_gdt[GDT_ENTRIES]; |
| 68 | }; | 66 | }; |
| 69 | 67 | ||
| 70 | struct lg_cpu_arch | 68 | struct lg_cpu_arch { |
| 71 | { | ||
| 72 | /* The GDT entries copied into lguest_ro_state when running. */ | 69 | /* The GDT entries copied into lguest_ro_state when running. */ |
| 73 | struct desc_struct gdt[GDT_ENTRIES]; | 70 | struct desc_struct gdt[GDT_ENTRIES]; |
| 74 | 71 | ||
| @@ -85,7 +82,7 @@ static inline void lguest_set_ts(void) | |||
| 85 | 82 | ||
| 86 | cr0 = read_cr0(); | 83 | cr0 = read_cr0(); |
| 87 | if (!(cr0 & 8)) | 84 | if (!(cr0 & 8)) |
| 88 | write_cr0(cr0|8); | 85 | write_cr0(cr0 | 8); |
| 89 | } | 86 | } |
| 90 | 87 | ||
| 91 | /* Full 4G segment descriptors, suitable for CS and DS. */ | 88 | /* Full 4G segment descriptors, suitable for CS and DS. */ |
diff --git a/include/asm-x86/lguest_hcall.h b/include/asm-x86/lguest_hcall.h index f239e7069cab..a3241f28e34a 100644 --- a/include/asm-x86/lguest_hcall.h +++ b/include/asm-x86/lguest_hcall.h | |||
| @@ -46,7 +46,7 @@ hcall(unsigned long call, | |||
| 46 | { | 46 | { |
| 47 | /* "int" is the Intel instruction to trigger a trap. */ | 47 | /* "int" is the Intel instruction to trigger a trap. */ |
| 48 | asm volatile("int $" __stringify(LGUEST_TRAP_ENTRY) | 48 | asm volatile("int $" __stringify(LGUEST_TRAP_ENTRY) |
| 49 | /* The call in %eax (aka "a") might be overwritten */ | 49 | /* The call in %eax (aka "a") might be overwritten */ |
| 50 | : "=a"(call) | 50 | : "=a"(call) |
| 51 | /* The arguments are in %eax, %edx, %ebx & %ecx */ | 51 | /* The arguments are in %eax, %edx, %ebx & %ecx */ |
| 52 | : "a"(call), "d"(arg1), "b"(arg2), "c"(arg3) | 52 | : "a"(call), "d"(arg1), "b"(arg2), "c"(arg3) |
| @@ -62,8 +62,7 @@ hcall(unsigned long call, | |||
| 62 | #define LGUEST_IRQS (NR_IRQS < 32 ? NR_IRQS: 32) | 62 | #define LGUEST_IRQS (NR_IRQS < 32 ? NR_IRQS: 32) |
| 63 | 63 | ||
| 64 | #define LHCALL_RING_SIZE 64 | 64 | #define LHCALL_RING_SIZE 64 |
| 65 | struct hcall_args | 65 | struct hcall_args { |
| 66 | { | ||
| 67 | /* These map directly onto eax, ebx, ecx, edx in struct lguest_regs */ | 66 | /* These map directly onto eax, ebx, ecx, edx in struct lguest_regs */ |
| 68 | unsigned long arg0, arg2, arg3, arg1; | 67 | unsigned long arg0, arg2, arg3, arg1; |
| 69 | }; | 68 | }; |
diff --git a/include/asm-x86/linkage.h b/include/asm-x86/linkage.h index c048353f4b85..64e444f8e85b 100644 --- a/include/asm-x86/linkage.h +++ b/include/asm-x86/linkage.h | |||
| @@ -1,6 +1,9 @@ | |||
| 1 | #ifndef __ASM_LINKAGE_H | 1 | #ifndef __ASM_LINKAGE_H |
| 2 | #define __ASM_LINKAGE_H | 2 | #define __ASM_LINKAGE_H |
| 3 | 3 | ||
| 4 | #undef notrace | ||
| 5 | #define notrace __attribute__((no_instrument_function)) | ||
| 6 | |||
| 4 | #ifdef CONFIG_X86_64 | 7 | #ifdef CONFIG_X86_64 |
| 5 | #define __ALIGN .p2align 4,,15 | 8 | #define __ALIGN .p2align 4,,15 |
| 6 | #define __ALIGN_STR ".p2align 4,,15" | 9 | #define __ALIGN_STR ".p2align 4,,15" |
diff --git a/include/asm-x86/local.h b/include/asm-x86/local.h index f852c62b3319..330a72496abd 100644 --- a/include/asm-x86/local.h +++ b/include/asm-x86/local.h | |||
| @@ -18,32 +18,28 @@ typedef struct { | |||
| 18 | 18 | ||
| 19 | static inline void local_inc(local_t *l) | 19 | static inline void local_inc(local_t *l) |
| 20 | { | 20 | { |
| 21 | __asm__ __volatile__( | 21 | asm volatile(_ASM_INC "%0" |
| 22 | _ASM_INC "%0" | 22 | : "+m" (l->a.counter)); |
| 23 | :"+m" (l->a.counter)); | ||
| 24 | } | 23 | } |
| 25 | 24 | ||
| 26 | static inline void local_dec(local_t *l) | 25 | static inline void local_dec(local_t *l) |
| 27 | { | 26 | { |
| 28 | __asm__ __volatile__( | 27 | asm volatile(_ASM_DEC "%0" |
| 29 | _ASM_DEC "%0" | 28 | : "+m" (l->a.counter)); |
| 30 | :"+m" (l->a.counter)); | ||
| 31 | } | 29 | } |
| 32 | 30 | ||
| 33 | static inline void local_add(long i, local_t *l) | 31 | static inline void local_add(long i, local_t *l) |
| 34 | { | 32 | { |
| 35 | __asm__ __volatile__( | 33 | asm volatile(_ASM_ADD "%1,%0" |
| 36 | _ASM_ADD "%1,%0" | 34 | : "+m" (l->a.counter) |
| 37 | :"+m" (l->a.counter) | 35 | : "ir" (i)); |
| 38 | :"ir" (i)); | ||
| 39 | } | 36 | } |
| 40 | 37 | ||
| 41 | static inline void local_sub(long i, local_t *l) | 38 | static inline void local_sub(long i, local_t *l) |
| 42 | { | 39 | { |
| 43 | __asm__ __volatile__( | 40 | asm volatile(_ASM_SUB "%1,%0" |
| 44 | _ASM_SUB "%1,%0" | 41 | : "+m" (l->a.counter) |
| 45 | :"+m" (l->a.counter) | 42 | : "ir" (i)); |
| 46 | :"ir" (i)); | ||
| 47 | } | 43 | } |
| 48 | 44 | ||
| 49 | /** | 45 | /** |
| @@ -59,10 +55,9 @@ static inline int local_sub_and_test(long i, local_t *l) | |||
| 59 | { | 55 | { |
| 60 | unsigned char c; | 56 | unsigned char c; |
| 61 | 57 | ||
| 62 | __asm__ __volatile__( | 58 | asm volatile(_ASM_SUB "%2,%0; sete %1" |
| 63 | _ASM_SUB "%2,%0; sete %1" | 59 | : "+m" (l->a.counter), "=qm" (c) |
| 64 | :"+m" (l->a.counter), "=qm" (c) | 60 | : "ir" (i) : "memory"); |
| 65 | :"ir" (i) : "memory"); | ||
| 66 | return c; | 61 | return c; |
| 67 | } | 62 | } |
| 68 | 63 | ||
| @@ -78,10 +73,9 @@ static inline int local_dec_and_test(local_t *l) | |||
| 78 | { | 73 | { |
| 79 | unsigned char c; | 74 | unsigned char c; |
| 80 | 75 | ||
| 81 | __asm__ __volatile__( | 76 | asm volatile(_ASM_DEC "%0; sete %1" |
| 82 | _ASM_DEC "%0; sete %1" | 77 | : "+m" (l->a.counter), "=qm" (c) |
| 83 | :"+m" (l->a.counter), "=qm" (c) | 78 | : : "memory"); |
| 84 | : : "memory"); | ||
| 85 | return c != 0; | 79 | return c != 0; |
| 86 | } | 80 | } |
| 87 | 81 | ||
| @@ -97,10 +91,9 @@ static inline int local_inc_and_test(local_t *l) | |||
| 97 | { | 91 | { |
| 98 | unsigned char c; | 92 | unsigned char c; |
| 99 | 93 | ||
| 100 | __asm__ __volatile__( | 94 | asm volatile(_ASM_INC "%0; sete %1" |
| 101 | _ASM_INC "%0; sete %1" | 95 | : "+m" (l->a.counter), "=qm" (c) |
| 102 | :"+m" (l->a.counter), "=qm" (c) | 96 | : : "memory"); |
| 103 | : : "memory"); | ||
| 104 | return c != 0; | 97 | return c != 0; |
| 105 | } | 98 | } |
| 106 | 99 | ||
| @@ -117,10 +110,9 @@ static inline int local_add_negative(long i, local_t *l) | |||
| 117 | { | 110 | { |
| 118 | unsigned char c; | 111 | unsigned char c; |
| 119 | 112 | ||
| 120 | __asm__ __volatile__( | 113 | asm volatile(_ASM_ADD "%2,%0; sets %1" |
| 121 | _ASM_ADD "%2,%0; sets %1" | 114 | : "+m" (l->a.counter), "=qm" (c) |
| 122 | :"+m" (l->a.counter), "=qm" (c) | 115 | : "ir" (i) : "memory"); |
| 123 | :"ir" (i) : "memory"); | ||
| 124 | return c; | 116 | return c; |
| 125 | } | 117 | } |
| 126 | 118 | ||
| @@ -141,10 +133,9 @@ static inline long local_add_return(long i, local_t *l) | |||
| 141 | #endif | 133 | #endif |
| 142 | /* Modern 486+ processor */ | 134 | /* Modern 486+ processor */ |
| 143 | __i = i; | 135 | __i = i; |
| 144 | __asm__ __volatile__( | 136 | asm volatile(_ASM_XADD "%0, %1;" |
| 145 | _ASM_XADD "%0, %1;" | 137 | : "+r" (i), "+m" (l->a.counter) |
| 146 | :"+r" (i), "+m" (l->a.counter) | 138 | : : "memory"); |
| 147 | : : "memory"); | ||
| 148 | return i + __i; | 139 | return i + __i; |
| 149 | 140 | ||
| 150 | #ifdef CONFIG_M386 | 141 | #ifdef CONFIG_M386 |
| @@ -182,11 +173,11 @@ static inline long local_sub_return(long i, local_t *l) | |||
| 182 | #define local_add_unless(l, a, u) \ | 173 | #define local_add_unless(l, a, u) \ |
| 183 | ({ \ | 174 | ({ \ |
| 184 | long c, old; \ | 175 | long c, old; \ |
| 185 | c = local_read(l); \ | 176 | c = local_read((l)); \ |
| 186 | for (;;) { \ | 177 | for (;;) { \ |
| 187 | if (unlikely(c == (u))) \ | 178 | if (unlikely(c == (u))) \ |
| 188 | break; \ | 179 | break; \ |
| 189 | old = local_cmpxchg((l), c, c + (a)); \ | 180 | old = local_cmpxchg((l), c, c + (a)); \ |
| 190 | if (likely(old == c)) \ | 181 | if (likely(old == c)) \ |
| 191 | break; \ | 182 | break; \ |
| 192 | c = old; \ | 183 | c = old; \ |
| @@ -214,26 +205,30 @@ static inline long local_sub_return(long i, local_t *l) | |||
| 214 | 205 | ||
| 215 | /* Need to disable preemption for the cpu local counters otherwise we could | 206 | /* Need to disable preemption for the cpu local counters otherwise we could |
| 216 | still access a variable of a previous CPU in a non atomic way. */ | 207 | still access a variable of a previous CPU in a non atomic way. */ |
| 217 | #define cpu_local_wrap_v(l) \ | 208 | #define cpu_local_wrap_v(l) \ |
| 218 | ({ local_t res__; \ | 209 | ({ \ |
| 219 | preempt_disable(); \ | 210 | local_t res__; \ |
| 220 | res__ = (l); \ | 211 | preempt_disable(); \ |
| 221 | preempt_enable(); \ | 212 | res__ = (l); \ |
| 222 | res__; }) | 213 | preempt_enable(); \ |
| 214 | res__; \ | ||
| 215 | }) | ||
| 223 | #define cpu_local_wrap(l) \ | 216 | #define cpu_local_wrap(l) \ |
| 224 | ({ preempt_disable(); \ | 217 | ({ \ |
| 225 | l; \ | 218 | preempt_disable(); \ |
| 226 | preempt_enable(); }) \ | 219 | (l); \ |
| 227 | 220 | preempt_enable(); \ | |
| 228 | #define cpu_local_read(l) cpu_local_wrap_v(local_read(&__get_cpu_var(l))) | 221 | }) \ |
| 229 | #define cpu_local_set(l, i) cpu_local_wrap(local_set(&__get_cpu_var(l), (i))) | 222 | |
| 230 | #define cpu_local_inc(l) cpu_local_wrap(local_inc(&__get_cpu_var(l))) | 223 | #define cpu_local_read(l) cpu_local_wrap_v(local_read(&__get_cpu_var((l)))) |
| 231 | #define cpu_local_dec(l) cpu_local_wrap(local_dec(&__get_cpu_var(l))) | 224 | #define cpu_local_set(l, i) cpu_local_wrap(local_set(&__get_cpu_var((l)), (i))) |
| 232 | #define cpu_local_add(i, l) cpu_local_wrap(local_add((i), &__get_cpu_var(l))) | 225 | #define cpu_local_inc(l) cpu_local_wrap(local_inc(&__get_cpu_var((l)))) |
| 233 | #define cpu_local_sub(i, l) cpu_local_wrap(local_sub((i), &__get_cpu_var(l))) | 226 | #define cpu_local_dec(l) cpu_local_wrap(local_dec(&__get_cpu_var((l)))) |
| 234 | 227 | #define cpu_local_add(i, l) cpu_local_wrap(local_add((i), &__get_cpu_var((l)))) | |
| 235 | #define __cpu_local_inc(l) cpu_local_inc(l) | 228 | #define cpu_local_sub(i, l) cpu_local_wrap(local_sub((i), &__get_cpu_var((l)))) |
| 236 | #define __cpu_local_dec(l) cpu_local_dec(l) | 229 | |
| 230 | #define __cpu_local_inc(l) cpu_local_inc((l)) | ||
| 231 | #define __cpu_local_dec(l) cpu_local_dec((l)) | ||
| 237 | #define __cpu_local_add(i, l) cpu_local_add((i), (l)) | 232 | #define __cpu_local_add(i, l) cpu_local_add((i), (l)) |
| 238 | #define __cpu_local_sub(i, l) cpu_local_sub((i), (l)) | 233 | #define __cpu_local_sub(i, l) cpu_local_sub((i), (l)) |
| 239 | 234 | ||
diff --git a/include/asm-x86/mach-bigsmp/mach_apic.h b/include/asm-x86/mach-bigsmp/mach_apic.h index 6df235e8ea91..8327907c79bf 100644 --- a/include/asm-x86/mach-bigsmp/mach_apic.h +++ b/include/asm-x86/mach-bigsmp/mach_apic.h | |||
| @@ -1,10 +1,7 @@ | |||
| 1 | #ifndef __ASM_MACH_APIC_H | 1 | #ifndef __ASM_MACH_APIC_H |
| 2 | #define __ASM_MACH_APIC_H | 2 | #define __ASM_MACH_APIC_H |
| 3 | 3 | ||
| 4 | 4 | #define xapic_phys_to_log_apicid(cpu) (per_cpu(x86_bios_cpu_apicid, cpu)) | |
| 5 | extern u8 bios_cpu_apicid[]; | ||
| 6 | |||
| 7 | #define xapic_phys_to_log_apicid(cpu) (bios_cpu_apicid[cpu]) | ||
| 8 | #define esr_disable (1) | 5 | #define esr_disable (1) |
| 9 | 6 | ||
| 10 | static inline int apic_id_registered(void) | 7 | static inline int apic_id_registered(void) |
| @@ -90,7 +87,7 @@ static inline int apicid_to_node(int logical_apicid) | |||
| 90 | static inline int cpu_present_to_apicid(int mps_cpu) | 87 | static inline int cpu_present_to_apicid(int mps_cpu) |
| 91 | { | 88 | { |
| 92 | if (mps_cpu < NR_CPUS) | 89 | if (mps_cpu < NR_CPUS) |
| 93 | return (int) bios_cpu_apicid[mps_cpu]; | 90 | return (int) per_cpu(x86_bios_cpu_apicid, mps_cpu); |
| 94 | 91 | ||
| 95 | return BAD_APICID; | 92 | return BAD_APICID; |
| 96 | } | 93 | } |
| @@ -109,17 +106,6 @@ static inline int cpu_to_logical_apicid(int cpu) | |||
| 109 | return cpu_physical_id(cpu); | 106 | return cpu_physical_id(cpu); |
| 110 | } | 107 | } |
| 111 | 108 | ||
| 112 | static inline int mpc_apic_id(struct mpc_config_processor *m, | ||
| 113 | struct mpc_config_translation *translation_record) | ||
| 114 | { | ||
| 115 | printk("Processor #%d %u:%u APIC version %d\n", | ||
| 116 | m->mpc_apicid, | ||
| 117 | (m->mpc_cpufeature & CPU_FAMILY_MASK) >> 8, | ||
| 118 | (m->mpc_cpufeature & CPU_MODEL_MASK) >> 4, | ||
| 119 | m->mpc_apicver); | ||
| 120 | return m->mpc_apicid; | ||
| 121 | } | ||
| 122 | |||
| 123 | static inline physid_mask_t ioapic_phys_id_map(physid_mask_t phys_map) | 109 | static inline physid_mask_t ioapic_phys_id_map(physid_mask_t phys_map) |
| 124 | { | 110 | { |
| 125 | /* For clustered we don't have a good way to do this yet - hack */ | 111 | /* For clustered we don't have a good way to do this yet - hack */ |
diff --git a/include/asm-x86/mach-default/mach_apic.h b/include/asm-x86/mach-default/mach_apic.h index e3c2c1012c1c..0a6634f62abe 100644 --- a/include/asm-x86/mach-default/mach_apic.h +++ b/include/asm-x86/mach-default/mach_apic.h | |||
| @@ -1,6 +1,8 @@ | |||
| 1 | #ifndef __ASM_MACH_APIC_H | 1 | #ifndef __ASM_MACH_APIC_H |
| 2 | #define __ASM_MACH_APIC_H | 2 | #define __ASM_MACH_APIC_H |
| 3 | 3 | ||
| 4 | #ifdef CONFIG_X86_LOCAL_APIC | ||
| 5 | |||
| 4 | #include <mach_apicdef.h> | 6 | #include <mach_apicdef.h> |
| 5 | #include <asm/smp.h> | 7 | #include <asm/smp.h> |
| 6 | 8 | ||
| @@ -14,24 +16,25 @@ static inline cpumask_t target_cpus(void) | |||
| 14 | return cpumask_of_cpu(0); | 16 | return cpumask_of_cpu(0); |
| 15 | #endif | 17 | #endif |
| 16 | } | 18 | } |
| 17 | #define TARGET_CPUS (target_cpus()) | ||
| 18 | 19 | ||
| 19 | #define NO_BALANCE_IRQ (0) | 20 | #define NO_BALANCE_IRQ (0) |
| 20 | #define esr_disable (0) | 21 | #define esr_disable (0) |
| 21 | 22 | ||
| 23 | #ifdef CONFIG_X86_64 | ||
| 24 | #include <asm/genapic.h> | ||
| 25 | #define INT_DELIVERY_MODE (genapic->int_delivery_mode) | ||
| 26 | #define INT_DEST_MODE (genapic->int_dest_mode) | ||
| 27 | #define TARGET_CPUS (genapic->target_cpus()) | ||
| 28 | #define apic_id_registered (genapic->apic_id_registered) | ||
| 29 | #define init_apic_ldr (genapic->init_apic_ldr) | ||
| 30 | #define cpu_mask_to_apicid (genapic->cpu_mask_to_apicid) | ||
| 31 | #define phys_pkg_id (genapic->phys_pkg_id) | ||
| 32 | #define vector_allocation_domain (genapic->vector_allocation_domain) | ||
| 33 | extern void setup_apic_routing(void); | ||
| 34 | #else | ||
| 22 | #define INT_DELIVERY_MODE dest_LowestPrio | 35 | #define INT_DELIVERY_MODE dest_LowestPrio |
| 23 | #define INT_DEST_MODE 1 /* logical delivery broadcast to all procs */ | 36 | #define INT_DEST_MODE 1 /* logical delivery broadcast to all procs */ |
| 24 | 37 | #define TARGET_CPUS (target_cpus()) | |
| 25 | static inline unsigned long check_apicid_used(physid_mask_t bitmap, int apicid) | ||
| 26 | { | ||
| 27 | return physid_isset(apicid, bitmap); | ||
| 28 | } | ||
| 29 | |||
| 30 | static inline unsigned long check_apicid_present(int bit) | ||
| 31 | { | ||
| 32 | return physid_isset(bit, phys_cpu_present_map); | ||
| 33 | } | ||
| 34 | |||
| 35 | /* | 38 | /* |
| 36 | * Set up the logical destination ID. | 39 | * Set up the logical destination ID. |
| 37 | * | 40 | * |
| @@ -49,23 +52,51 @@ static inline void init_apic_ldr(void) | |||
| 49 | apic_write_around(APIC_LDR, val); | 52 | apic_write_around(APIC_LDR, val); |
| 50 | } | 53 | } |
| 51 | 54 | ||
| 52 | static inline physid_mask_t ioapic_phys_id_map(physid_mask_t phys_map) | 55 | static inline int apic_id_registered(void) |
| 53 | { | 56 | { |
| 54 | return phys_map; | 57 | return physid_isset(GET_APIC_ID(read_apic_id()), phys_cpu_present_map); |
| 58 | } | ||
| 59 | |||
| 60 | static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask) | ||
| 61 | { | ||
| 62 | return cpus_addr(cpumask)[0]; | ||
| 63 | } | ||
| 64 | |||
| 65 | static inline u32 phys_pkg_id(u32 cpuid_apic, int index_msb) | ||
| 66 | { | ||
| 67 | return cpuid_apic >> index_msb; | ||
| 55 | } | 68 | } |
| 56 | 69 | ||
| 57 | static inline void setup_apic_routing(void) | 70 | static inline void setup_apic_routing(void) |
| 58 | { | 71 | { |
| 72 | #ifdef CONFIG_X86_IO_APIC | ||
| 59 | printk("Enabling APIC mode: %s. Using %d I/O APICs\n", | 73 | printk("Enabling APIC mode: %s. Using %d I/O APICs\n", |
| 60 | "Flat", nr_ioapics); | 74 | "Flat", nr_ioapics); |
| 75 | #endif | ||
| 61 | } | 76 | } |
| 62 | 77 | ||
| 63 | static inline int multi_timer_check(int apic, int irq) | 78 | static inline int apicid_to_node(int logical_apicid) |
| 64 | { | 79 | { |
| 65 | return 0; | 80 | return 0; |
| 66 | } | 81 | } |
| 82 | #endif | ||
| 67 | 83 | ||
| 68 | static inline int apicid_to_node(int logical_apicid) | 84 | static inline unsigned long check_apicid_used(physid_mask_t bitmap, int apicid) |
| 85 | { | ||
| 86 | return physid_isset(apicid, bitmap); | ||
| 87 | } | ||
| 88 | |||
| 89 | static inline unsigned long check_apicid_present(int bit) | ||
| 90 | { | ||
| 91 | return physid_isset(bit, phys_cpu_present_map); | ||
| 92 | } | ||
| 93 | |||
| 94 | static inline physid_mask_t ioapic_phys_id_map(physid_mask_t phys_map) | ||
| 95 | { | ||
| 96 | return phys_map; | ||
| 97 | } | ||
| 98 | |||
| 99 | static inline int multi_timer_check(int apic, int irq) | ||
| 69 | { | 100 | { |
| 70 | return 0; | 101 | return 0; |
| 71 | } | 102 | } |
| @@ -78,8 +109,13 @@ static inline int cpu_to_logical_apicid(int cpu) | |||
| 78 | 109 | ||
| 79 | static inline int cpu_present_to_apicid(int mps_cpu) | 110 | static inline int cpu_present_to_apicid(int mps_cpu) |
| 80 | { | 111 | { |
| 112 | #ifdef CONFIG_X86_64 | ||
| 113 | if (cpu_present(mps_cpu)) | ||
| 114 | return (int)per_cpu(x86_bios_cpu_apicid, mps_cpu); | ||
| 115 | #else | ||
| 81 | if (mps_cpu < get_physical_broadcast()) | 116 | if (mps_cpu < get_physical_broadcast()) |
| 82 | return mps_cpu; | 117 | return mps_cpu; |
| 118 | #endif | ||
| 83 | else | 119 | else |
| 84 | return BAD_APICID; | 120 | return BAD_APICID; |
| 85 | } | 121 | } |
| @@ -89,17 +125,6 @@ static inline physid_mask_t apicid_to_cpu_present(int phys_apicid) | |||
| 89 | return physid_mask_of_physid(phys_apicid); | 125 | return physid_mask_of_physid(phys_apicid); |
| 90 | } | 126 | } |
| 91 | 127 | ||
| 92 | static inline int mpc_apic_id(struct mpc_config_processor *m, | ||
| 93 | struct mpc_config_translation *translation_record) | ||
| 94 | { | ||
| 95 | printk("Processor #%d %u:%u APIC version %d\n", | ||
| 96 | m->mpc_apicid, | ||
| 97 | (m->mpc_cpufeature & CPU_FAMILY_MASK) >> 8, | ||
| 98 | (m->mpc_cpufeature & CPU_MODEL_MASK) >> 4, | ||
| 99 | m->mpc_apicver); | ||
| 100 | return m->mpc_apicid; | ||
| 101 | } | ||
| 102 | |||
| 103 | static inline void setup_portio_remap(void) | 128 | static inline void setup_portio_remap(void) |
| 104 | { | 129 | { |
| 105 | } | 130 | } |
| @@ -109,23 +134,9 @@ static inline int check_phys_apicid_present(int boot_cpu_physical_apicid) | |||
| 109 | return physid_isset(boot_cpu_physical_apicid, phys_cpu_present_map); | 134 | return physid_isset(boot_cpu_physical_apicid, phys_cpu_present_map); |
| 110 | } | 135 | } |
| 111 | 136 | ||
| 112 | static inline int apic_id_registered(void) | ||
| 113 | { | ||
| 114 | return physid_isset(GET_APIC_ID(apic_read(APIC_ID)), phys_cpu_present_map); | ||
| 115 | } | ||
| 116 | |||
| 117 | static inline unsigned int cpu_mask_to_apicid(cpumask_t cpumask) | ||
| 118 | { | ||
| 119 | return cpus_addr(cpumask)[0]; | ||
| 120 | } | ||
| 121 | |||
| 122 | static inline void enable_apic_mode(void) | 137 | static inline void enable_apic_mode(void) |
| 123 | { | 138 | { |
| 124 | } | 139 | } |
| 125 | 140 | ||
| 126 | static inline u32 phys_pkg_id(u32 cpuid_apic, int index_msb) | 141 | #endif /* CONFIG_X86_LOCAL_APIC */ |
| 127 | { | ||
| 128 | return cpuid_apic >> index_msb; | ||
| 129 | } | ||
| 130 | |||
| 131 | #endif /* __ASM_MACH_APIC_H */ | 142 | #endif /* __ASM_MACH_APIC_H */ |
diff --git a/include/asm-x86/mach-default/mach_apicdef.h b/include/asm-x86/mach-default/mach_apicdef.h index ae9841319094..e4b29ba37de6 100644 --- a/include/asm-x86/mach-default/mach_apicdef.h +++ b/include/asm-x86/mach-default/mach_apicdef.h | |||
| @@ -3,10 +3,14 @@ | |||
| 3 | 3 | ||
| 4 | #include <asm/apic.h> | 4 | #include <asm/apic.h> |
| 5 | 5 | ||
| 6 | #ifdef CONFIG_X86_64 | ||
| 7 | #define APIC_ID_MASK (0xFFu<<24) | ||
| 8 | #define GET_APIC_ID(x) (((x)>>24)&0xFFu) | ||
| 9 | #define SET_APIC_ID(x) (((x)<<24)) | ||
| 10 | #else | ||
| 6 | #define APIC_ID_MASK (0xF<<24) | 11 | #define APIC_ID_MASK (0xF<<24) |
| 7 | |||
| 8 | static inline unsigned get_apic_id(unsigned long x) | 12 | static inline unsigned get_apic_id(unsigned long x) |
| 9 | { | 13 | { |
| 10 | unsigned int ver = GET_APIC_VERSION(apic_read(APIC_LVR)); | 14 | unsigned int ver = GET_APIC_VERSION(apic_read(APIC_LVR)); |
| 11 | if (APIC_XAPIC(ver)) | 15 | if (APIC_XAPIC(ver)) |
| 12 | return (((x)>>24)&0xFF); | 16 | return (((x)>>24)&0xFF); |
| @@ -15,5 +19,6 @@ static inline unsigned get_apic_id(unsigned long x) | |||
| 15 | } | 19 | } |
| 16 | 20 | ||
| 17 | #define GET_APIC_ID(x) get_apic_id(x) | 21 | #define GET_APIC_ID(x) get_apic_id(x) |
| 22 | #endif | ||
| 18 | 23 | ||
| 19 | #endif | 24 | #endif |
diff --git a/include/asm-x86/mach-default/mach_ipi.h b/include/asm-x86/mach-default/mach_ipi.h index 0dba244c86db..be323364e68f 100644 --- a/include/asm-x86/mach-default/mach_ipi.h +++ b/include/asm-x86/mach-default/mach_ipi.h | |||
| @@ -9,10 +9,15 @@ void __send_IPI_shortcut(unsigned int shortcut, int vector); | |||
| 9 | 9 | ||
| 10 | extern int no_broadcast; | 10 | extern int no_broadcast; |
| 11 | 11 | ||
| 12 | #ifdef CONFIG_X86_64 | ||
| 13 | #include <asm/genapic.h> | ||
| 14 | #define send_IPI_mask (genapic->send_IPI_mask) | ||
| 15 | #else | ||
| 12 | static inline void send_IPI_mask(cpumask_t mask, int vector) | 16 | static inline void send_IPI_mask(cpumask_t mask, int vector) |
| 13 | { | 17 | { |
| 14 | send_IPI_mask_bitmask(mask, vector); | 18 | send_IPI_mask_bitmask(mask, vector); |
| 15 | } | 19 | } |
| 20 | #endif | ||
| 16 | 21 | ||
| 17 | static inline void __local_send_IPI_allbutself(int vector) | 22 | static inline void __local_send_IPI_allbutself(int vector) |
| 18 | { | 23 | { |
| @@ -33,6 +38,10 @@ static inline void __local_send_IPI_all(int vector) | |||
| 33 | __send_IPI_shortcut(APIC_DEST_ALLINC, vector); | 38 | __send_IPI_shortcut(APIC_DEST_ALLINC, vector); |
| 34 | } | 39 | } |
| 35 | 40 | ||
| 41 | #ifdef CONFIG_X86_64 | ||
| 42 | #define send_IPI_allbutself (genapic->send_IPI_allbutself) | ||
| 43 | #define send_IPI_all (genapic->send_IPI_all) | ||
| 44 | #else | ||
| 36 | static inline void send_IPI_allbutself(int vector) | 45 | static inline void send_IPI_allbutself(int vector) |
| 37 | { | 46 | { |
| 38 | /* | 47 | /* |
| @@ -50,5 +59,6 @@ static inline void send_IPI_all(int vector) | |||
| 50 | { | 59 | { |
| 51 | __local_send_IPI_all(vector); | 60 | __local_send_IPI_all(vector); |
| 52 | } | 61 | } |
| 62 | #endif | ||
| 53 | 63 | ||
| 54 | #endif /* __ASM_MACH_IPI_H */ | 64 | #endif /* __ASM_MACH_IPI_H */ |
diff --git a/include/asm-x86/mach-default/mach_mpparse.h b/include/asm-x86/mach-default/mach_mpparse.h index 1d3832482580..d14108505bb8 100644 --- a/include/asm-x86/mach-default/mach_mpparse.h +++ b/include/asm-x86/mach-default/mach_mpparse.h | |||
| @@ -1,17 +1,6 @@ | |||
| 1 | #ifndef __ASM_MACH_MPPARSE_H | 1 | #ifndef __ASM_MACH_MPPARSE_H |
| 2 | #define __ASM_MACH_MPPARSE_H | 2 | #define __ASM_MACH_MPPARSE_H |
| 3 | 3 | ||
| 4 | static inline void mpc_oem_bus_info(struct mpc_config_bus *m, char *name, | ||
| 5 | struct mpc_config_translation *translation) | ||
| 6 | { | ||
| 7 | // Dprintk("Bus #%d is %s\n", m->mpc_busid, name); | ||
| 8 | } | ||
| 9 | |||
| 10 | static inline void mpc_oem_pci_bus(struct mpc_config_bus *m, | ||
| 11 | struct mpc_config_translation *translation) | ||
| 12 | { | ||
| 13 | } | ||
| 14 | |||
| 15 | static inline int mps_oem_check(struct mp_config_table *mpc, char *oem, | 4 | static inline int mps_oem_check(struct mp_config_table *mpc, char *oem, |
| 16 | char *productid) | 5 | char *productid) |
| 17 | { | 6 | { |
diff --git a/include/asm-x86/mach-default/mach_reboot.h b/include/asm-x86/mach-default/mach_reboot.h deleted file mode 100644 index 6adee6a97dec..000000000000 --- a/include/asm-x86/mach-default/mach_reboot.h +++ /dev/null | |||
| @@ -1,61 +0,0 @@ | |||
| 1 | /* | ||
| 2 | * arch/i386/mach-generic/mach_reboot.h | ||
| 3 | * | ||
| 4 | * Machine specific reboot functions for generic. | ||
| 5 | * Split out from reboot.c by Osamu Tomita <tomita@cinet.co.jp> | ||
| 6 | */ | ||
| 7 | #ifndef _MACH_REBOOT_H | ||
| 8 | #define _MACH_REBOOT_H | ||
| 9 | |||
| 10 | static inline void kb_wait(void) | ||
| 11 | { | ||
| 12 | int i; | ||
| 13 | |||
| 14 | for (i = 0; i < 0x10000; i++) | ||
| 15 | if ((inb_p(0x64) & 0x02) == 0) | ||
| 16 | break; | ||
| 17 | } | ||
| 18 | |||
| 19 | static inline void mach_reboot(void) | ||
| 20 | { | ||
| 21 | int i; | ||
| 22 | |||
| 23 | /* old method, works on most machines */ | ||
| 24 | for (i = 0; i < 10; i++) { | ||
| 25 | kb_wait(); | ||
| 26 | udelay(50); | ||
| 27 | outb(0xfe, 0x64); /* pulse reset low */ | ||
| 28 | udelay(50); | ||
| 29 | } | ||
| 30 | |||
| 31 | /* New method: sets the "System flag" which, when set, indicates | ||
| 32 | * successful completion of the keyboard controller self-test (Basic | ||
| 33 | * Assurance Test, BAT). This is needed for some machines with no | ||
| 34 | * keyboard plugged in. This read-modify-write sequence sets only the | ||
| 35 | * system flag | ||
| 36 | */ | ||
| 37 | for (i = 0; i < 10; i++) { | ||
| 38 | int cmd; | ||
| 39 | |||
| 40 | outb(0x20, 0x64); /* read Controller Command Byte */ | ||
| 41 | udelay(50); | ||
| 42 | kb_wait(); | ||
| 43 | udelay(50); | ||
| 44 | cmd = inb(0x60); | ||
| 45 | udelay(50); | ||
| 46 | kb_wait(); | ||
| 47 | udelay(50); | ||
| 48 | outb(0x60, 0x64); /* write Controller Command Byte */ | ||
| 49 | udelay(50); | ||
| 50 | kb_wait(); | ||
| 51 | udelay(50); | ||
| 52 | outb(cmd | 0x14, 0x60); /* set "System flag" and "Keyboard Disabled" */ | ||
| 53 | udelay(50); | ||
| 54 | kb_wait(); | ||
| 55 | udelay(50); | ||
| 56 | outb(0xfe, 0x64); /* pulse reset low */ | ||
| 57 | udelay(50); | ||
| 58 | } | ||
| 59 | } | ||
| 60 | |||
| 61 | #endif /* !_MACH_REBOOT_H */ | ||
diff --git a/include/asm-x86/mach-default/smpboot_hooks.h b/include/asm-x86/mach-default/smpboot_hooks.h index 7f45f6311059..3ff2c5bff93a 100644 --- a/include/asm-x86/mach-default/smpboot_hooks.h +++ b/include/asm-x86/mach-default/smpboot_hooks.h | |||
| @@ -41,4 +41,11 @@ static inline void smpboot_setup_io_apic(void) | |||
| 41 | */ | 41 | */ |
| 42 | if (!skip_ioapic_setup && nr_ioapics) | 42 | if (!skip_ioapic_setup && nr_ioapics) |
| 43 | setup_IO_APIC(); | 43 | setup_IO_APIC(); |
| 44 | else | ||
| 45 | nr_ioapics = 0; | ||
| 46 | } | ||
| 47 | |||
| 48 | static inline void smpboot_clear_io_apic(void) | ||
| 49 | { | ||
| 50 | nr_ioapics = 0; | ||
| 44 | } | 51 | } |
diff --git a/include/asm-x86/mach-es7000/mach_apic.h b/include/asm-x86/mach-es7000/mach_apic.h index d23011fdf454..fbc8ad256f5a 100644 --- a/include/asm-x86/mach-es7000/mach_apic.h +++ b/include/asm-x86/mach-es7000/mach_apic.h | |||
| @@ -1,9 +1,7 @@ | |||
| 1 | #ifndef __ASM_MACH_APIC_H | 1 | #ifndef __ASM_MACH_APIC_H |
| 2 | #define __ASM_MACH_APIC_H | 2 | #define __ASM_MACH_APIC_H |
| 3 | 3 | ||
| 4 | extern u8 bios_cpu_apicid[]; | 4 | #define xapic_phys_to_log_apicid(cpu) per_cpu(x86_bios_cpu_apicid, cpu) |
| 5 | |||
| 6 | #define xapic_phys_to_log_apicid(cpu) (bios_cpu_apicid[cpu]) | ||
| 7 | #define esr_disable (1) | 5 | #define esr_disable (1) |
| 8 | 6 | ||
| 9 | static inline int apic_id_registered(void) | 7 | static inline int apic_id_registered(void) |
| @@ -80,7 +78,7 @@ extern void enable_apic_mode(void); | |||
| 80 | extern int apic_version [MAX_APICS]; | 78 | extern int apic_version [MAX_APICS]; |
| 81 | static inline void setup_apic_routing(void) | 79 | static inline void setup_apic_routing(void) |
| 82 | { | 80 | { |
| 83 | int apic = bios_cpu_apicid[smp_processor_id()]; | 81 | int apic = per_cpu(x86_bios_cpu_apicid, smp_processor_id()); |
| 84 | printk("Enabling APIC mode: %s. Using %d I/O APICs, target cpus %lx\n", | 82 | printk("Enabling APIC mode: %s. Using %d I/O APICs, target cpus %lx\n", |
| 85 | (apic_version[apic] == 0x14) ? | 83 | (apic_version[apic] == 0x14) ? |
| 86 | "Physical Cluster" : "Logical Cluster", nr_ioapics, cpus_addr(TARGET_CPUS)[0]); | 84 | "Physical Cluster" : "Logical Cluster", nr_ioapics, cpus_addr(TARGET_CPUS)[0]); |
| @@ -102,7 +100,7 @@ static inline int cpu_present_to_apicid(int mps_cpu) | |||
| 102 | if (!mps_cpu) | 100 | if (!mps_cpu) |
| 103 | return boot_cpu_physical_apicid; | 101 | return boot_cpu_physical_apicid; |
| 104 | else if (mps_cpu < NR_CPUS) | 102 | else if (mps_cpu < NR_CPUS) |
| 105 | return (int) bios_cpu_apicid[mps_cpu]; | 103 | return (int) per_cpu(x86_bios_cpu_apicid, mps_cpu); |
| 106 | else | 104 | else |
| 107 | return BAD_APICID; | 105 | return BAD_APICID; |
| 108 | } | 106 | } |
| @@ -129,16 +127,6 @@ static inline int cpu_to_logical_apicid(int cpu) | |||
| 129 | #endif | 127 | #endif |
| 130 | } | 128 | } |
| 131 | 129 | ||
| 132 | static inline int mpc_apic_id(struct mpc_config_processor *m, struct mpc_config_translation *unused) | ||
| 133 | { | ||
| 134 | printk("Processor #%d %u:%u APIC version %d\n", | ||
| 135 | m->mpc_apicid, | ||
| 136 | (m->mpc_cpufeature & CPU_FAMILY_MASK) >> 8, | ||
| 137 | (m->mpc_cpufeature & CPU_MODEL_MASK) >> 4, | ||
| 138 | m->mpc_apicver); | ||
| 139 | return (m->mpc_apicid); | ||
| 140 | } | ||
| 141 | |||
| 142 | static inline physid_mask_t ioapic_phys_id_map(physid_mask_t phys_map) | 130 | static inline physid_mask_t ioapic_phys_id_map(physid_mask_t phys_map) |
| 143 | { | 131 | { |
| 144 | /* For clustered we don't have a good way to do this yet - hack */ | 132 | /* For clustered we don't have a good way to do this yet - hack */ |
| @@ -153,7 +141,7 @@ static inline void setup_portio_remap(void) | |||
| 153 | extern unsigned int boot_cpu_physical_apicid; | 141 | extern unsigned int boot_cpu_physical_apicid; |
| 154 | static inline int check_phys_apicid_present(int cpu_physical_apicid) | 142 | static inline int check_phys_apicid_present(int cpu_physical_apicid) |
| 155 | { | 143 | { |
| 156 | boot_cpu_physical_apicid = GET_APIC_ID(apic_read(APIC_ID)); | 144 | boot_cpu_physical_apicid = GET_APIC_ID(read_apic_id()); |
| 157 | return (1); | 145 | return (1); |
| 158 | } | 146 | } |
| 159 | 147 | ||
diff --git a/include/asm-x86/mach-es7000/mach_mpparse.h b/include/asm-x86/mach-es7000/mach_mpparse.h index 52ee75cd0fe1..ef26d3523625 100644 --- a/include/asm-x86/mach-es7000/mach_mpparse.h +++ b/include/asm-x86/mach-es7000/mach_mpparse.h | |||
| @@ -3,17 +3,6 @@ | |||
| 3 | 3 | ||
| 4 | #include <linux/acpi.h> | 4 | #include <linux/acpi.h> |
| 5 | 5 | ||
| 6 | static inline void mpc_oem_bus_info(struct mpc_config_bus *m, char *name, | ||
| 7 | struct mpc_config_translation *translation) | ||
| 8 | { | ||
| 9 | Dprintk("Bus #%d is %s\n", m->mpc_busid, name); | ||
| 10 | } | ||
| 11 | |||
| 12 | static inline void mpc_oem_pci_bus(struct mpc_config_bus *m, | ||
| 13 | struct mpc_config_translation *translation) | ||
| 14 | { | ||
| 15 | } | ||
| 16 | |||
| 17 | extern int parse_unisys_oem (char *oemptr); | 6 | extern int parse_unisys_oem (char *oemptr); |
| 18 | extern int find_unisys_acpi_oem_table(unsigned long *oem_addr); | 7 | extern int find_unisys_acpi_oem_table(unsigned long *oem_addr); |
| 19 | extern void setup_unisys(void); | 8 | extern void setup_unisys(void); |
diff --git a/include/asm-x86/mach-generic/mach_apic.h b/include/asm-x86/mach-generic/mach_apic.h index a236e7021528..6eff343e1233 100644 --- a/include/asm-x86/mach-generic/mach_apic.h +++ b/include/asm-x86/mach-generic/mach_apic.h | |||
| @@ -19,7 +19,6 @@ | |||
| 19 | #define cpu_to_logical_apicid (genapic->cpu_to_logical_apicid) | 19 | #define cpu_to_logical_apicid (genapic->cpu_to_logical_apicid) |
| 20 | #define cpu_present_to_apicid (genapic->cpu_present_to_apicid) | 20 | #define cpu_present_to_apicid (genapic->cpu_present_to_apicid) |
| 21 | #define apicid_to_cpu_present (genapic->apicid_to_cpu_present) | 21 | #define apicid_to_cpu_present (genapic->apicid_to_cpu_present) |
| 22 | #define mpc_apic_id (genapic->mpc_apic_id) | ||
| 23 | #define setup_portio_remap (genapic->setup_portio_remap) | 22 | #define setup_portio_remap (genapic->setup_portio_remap) |
| 24 | #define check_apicid_present (genapic->check_apicid_present) | 23 | #define check_apicid_present (genapic->check_apicid_present) |
| 25 | #define check_phys_apicid_present (genapic->check_phys_apicid_present) | 24 | #define check_phys_apicid_present (genapic->check_phys_apicid_present) |
diff --git a/include/asm-x86/mach-generic/mach_mpparse.h b/include/asm-x86/mach-generic/mach_mpparse.h index dbd9fce54f4d..0d0b5ba2e9d1 100644 --- a/include/asm-x86/mach-generic/mach_mpparse.h +++ b/include/asm-x86/mach-generic/mach_mpparse.h | |||
| @@ -1,11 +1,6 @@ | |||
| 1 | #ifndef _MACH_MPPARSE_H | 1 | #ifndef _MACH_MPPARSE_H |
| 2 | #define _MACH_MPPARSE_H 1 | 2 | #define _MACH_MPPARSE_H 1 |
| 3 | 3 | ||
| 4 | #include <asm/genapic.h> | ||
| 5 | |||
| 6 | #define mpc_oem_bus_info (genapic->mpc_oem_bus_info) | ||
| 7 | #define mpc_oem_pci_bus (genapic->mpc_oem_pci_bus) | ||
| 8 | |||
| 9 | int mps_oem_check(struct mp_config_table *mpc, char *oem, char *productid); | 4 | int mps_oem_check(struct mp_config_table *mpc, char *oem, char *productid); |
| 10 | int acpi_madt_oem_check(char *oem_id, char *oem_table_id); | 5 | int acpi_madt_oem_check(char *oem_id, char *oem_table_id); |
| 11 | 6 | ||
diff --git a/include/asm-x86/mach-numaq/mach_apic.h b/include/asm-x86/mach-numaq/mach_apic.h index 3b637fac890b..75a56e5afbe7 100644 --- a/include/asm-x86/mach-numaq/mach_apic.h +++ b/include/asm-x86/mach-numaq/mach_apic.h | |||
| @@ -95,6 +95,16 @@ static inline physid_mask_t apicid_to_cpu_present(int logical_apicid) | |||
| 95 | return physid_mask_of_physid(cpu + 4*node); | 95 | return physid_mask_of_physid(cpu + 4*node); |
| 96 | } | 96 | } |
| 97 | 97 | ||
| 98 | struct mpc_config_translation { | ||
| 99 | unsigned char mpc_type; | ||
| 100 | unsigned char trans_len; | ||
| 101 | unsigned char trans_type; | ||
| 102 | unsigned char trans_quad; | ||
| 103 | unsigned char trans_global; | ||
| 104 | unsigned char trans_local; | ||
| 105 | unsigned short trans_reserved; | ||
| 106 | }; | ||
| 107 | |||
| 98 | static inline int mpc_apic_id(struct mpc_config_processor *m, | 108 | static inline int mpc_apic_id(struct mpc_config_processor *m, |
| 99 | struct mpc_config_translation *translation_record) | 109 | struct mpc_config_translation *translation_record) |
| 100 | { | 110 | { |
diff --git a/include/asm-x86/mach-numaq/mach_mpparse.h b/include/asm-x86/mach-numaq/mach_mpparse.h index 51bbac8fc0c2..459b12401187 100644 --- a/include/asm-x86/mach-numaq/mach_mpparse.h +++ b/include/asm-x86/mach-numaq/mach_mpparse.h | |||
| @@ -1,25 +1,10 @@ | |||
| 1 | #ifndef __ASM_MACH_MPPARSE_H | 1 | #ifndef __ASM_MACH_MPPARSE_H |
| 2 | #define __ASM_MACH_MPPARSE_H | 2 | #define __ASM_MACH_MPPARSE_H |
| 3 | 3 | ||
| 4 | static inline void mpc_oem_bus_info(struct mpc_config_bus *m, char *name, | 4 | extern void mpc_oem_bus_info(struct mpc_config_bus *m, char *name, |
| 5 | struct mpc_config_translation *translation) | 5 | struct mpc_config_translation *translation); |
| 6 | { | 6 | extern void mpc_oem_pci_bus(struct mpc_config_bus *m, |
| 7 | int quad = translation->trans_quad; | 7 | struct mpc_config_translation *translation); |
| 8 | int local = translation->trans_local; | ||
| 9 | |||
| 10 | mp_bus_id_to_node[m->mpc_busid] = quad; | ||
| 11 | mp_bus_id_to_local[m->mpc_busid] = local; | ||
| 12 | printk("Bus #%d is %s (node %d)\n", m->mpc_busid, name, quad); | ||
| 13 | } | ||
| 14 | |||
| 15 | static inline void mpc_oem_pci_bus(struct mpc_config_bus *m, | ||
| 16 | struct mpc_config_translation *translation) | ||
| 17 | { | ||
| 18 | int quad = translation->trans_quad; | ||
| 19 | int local = translation->trans_local; | ||
| 20 | |||
| 21 | quad_local_to_mp_bus_id[quad][local] = m->mpc_busid; | ||
| 22 | } | ||
| 23 | 8 | ||
| 24 | /* Hook from generic ACPI tables.c */ | 9 | /* Hook from generic ACPI tables.c */ |
| 25 | static inline void acpi_madt_oem_check(char *oem_id, char *oem_table_id) | 10 | static inline void acpi_madt_oem_check(char *oem_id, char *oem_table_id) |
diff --git a/include/asm-x86/mach-summit/mach_apic.h b/include/asm-x86/mach-summit/mach_apic.h index 062c97f6100b..1f76c2e70232 100644 --- a/include/asm-x86/mach-summit/mach_apic.h +++ b/include/asm-x86/mach-summit/mach_apic.h | |||
| @@ -40,7 +40,6 @@ static inline unsigned long check_apicid_present(int bit) | |||
| 40 | 40 | ||
| 41 | #define apicid_cluster(apicid) ((apicid) & XAPIC_DEST_CLUSTER_MASK) | 41 | #define apicid_cluster(apicid) ((apicid) & XAPIC_DEST_CLUSTER_MASK) |
| 42 | 42 | ||
| 43 | extern u8 bios_cpu_apicid[]; | ||
| 44 | extern u8 cpu_2_logical_apicid[]; | 43 | extern u8 cpu_2_logical_apicid[]; |
| 45 | 44 | ||
| 46 | static inline void init_apic_ldr(void) | 45 | static inline void init_apic_ldr(void) |
| @@ -110,7 +109,7 @@ static inline int cpu_to_logical_apicid(int cpu) | |||
| 110 | static inline int cpu_present_to_apicid(int mps_cpu) | 109 | static inline int cpu_present_to_apicid(int mps_cpu) |
| 111 | { | 110 | { |
| 112 | if (mps_cpu < NR_CPUS) | 111 | if (mps_cpu < NR_CPUS) |
| 113 | return (int)bios_cpu_apicid[mps_cpu]; | 112 | return (int)per_cpu(x86_bios_cpu_apicid, mps_cpu); |
| 114 | else | 113 | else |
| 115 | return BAD_APICID; | 114 | return BAD_APICID; |
| 116 | } | 115 | } |
| @@ -126,17 +125,6 @@ static inline physid_mask_t apicid_to_cpu_present(int apicid) | |||
| 126 | return physid_mask_of_physid(0); | 125 | return physid_mask_of_physid(0); |
| 127 | } | 126 | } |
| 128 | 127 | ||
| 129 | static inline int mpc_apic_id(struct mpc_config_processor *m, | ||
| 130 | struct mpc_config_translation *translation_record) | ||
| 131 | { | ||
| 132 | printk("Processor #%d %u:%u APIC version %d\n", | ||
| 133 | m->mpc_apicid, | ||
| 134 | (m->mpc_cpufeature & CPU_FAMILY_MASK) >> 8, | ||
| 135 | (m->mpc_cpufeature & CPU_MODEL_MASK) >> 4, | ||
| 136 | m->mpc_apicver); | ||
| 137 | return m->mpc_apicid; | ||
| 138 | } | ||
| 139 | |||
| 140 | static inline void setup_portio_remap(void) | 128 | static inline void setup_portio_remap(void) |
| 141 | { | 129 | { |
| 142 | } | 130 | } |
diff --git a/include/asm-x86/mach-summit/mach_mpparse.h b/include/asm-x86/mach-summit/mach_mpparse.h index c2520539d934..fdf591701339 100644 --- a/include/asm-x86/mach-summit/mach_mpparse.h +++ b/include/asm-x86/mach-summit/mach_mpparse.h | |||
| @@ -12,17 +12,6 @@ extern void setup_summit(void); | |||
| 12 | #define setup_summit() {} | 12 | #define setup_summit() {} |
| 13 | #endif | 13 | #endif |
| 14 | 14 | ||
| 15 | static inline void mpc_oem_bus_info(struct mpc_config_bus *m, char *name, | ||
| 16 | struct mpc_config_translation *translation) | ||
| 17 | { | ||
| 18 | Dprintk("Bus #%d is %s\n", m->mpc_busid, name); | ||
| 19 | } | ||
| 20 | |||
| 21 | static inline void mpc_oem_pci_bus(struct mpc_config_bus *m, | ||
| 22 | struct mpc_config_translation *translation) | ||
| 23 | { | ||
| 24 | } | ||
| 25 | |||
| 26 | static inline int mps_oem_check(struct mp_config_table *mpc, char *oem, | 15 | static inline int mps_oem_check(struct mp_config_table *mpc, char *oem, |
| 27 | char *productid) | 16 | char *productid) |
| 28 | { | 17 | { |
diff --git a/include/asm-x86/mach-visws/mach_apic.h b/include/asm-x86/mach-visws/mach_apic.h index efac6f0d139f..a9ef33a8a995 100644 --- a/include/asm-x86/mach-visws/mach_apic.h +++ b/include/asm-x86/mach-visws/mach_apic.h | |||
| @@ -23,7 +23,7 @@ | |||
| 23 | 23 | ||
| 24 | static inline int apic_id_registered(void) | 24 | static inline int apic_id_registered(void) |
| 25 | { | 25 | { |
| 26 | return physid_isset(GET_APIC_ID(apic_read(APIC_ID)), phys_cpu_present_map); | 26 | return physid_isset(GET_APIC_ID(read_apic_id()), phys_cpu_present_map); |
| 27 | } | 27 | } |
| 28 | 28 | ||
| 29 | /* | 29 | /* |
diff --git a/include/asm-x86/mach-visws/smpboot_hooks.h b/include/asm-x86/mach-visws/smpboot_hooks.h index d926471fa359..c9b83e395a2e 100644 --- a/include/asm-x86/mach-visws/smpboot_hooks.h +++ b/include/asm-x86/mach-visws/smpboot_hooks.h | |||
| @@ -22,3 +22,7 @@ static inline void smpboot_restore_warm_reset_vector(void) | |||
| 22 | static inline void smpboot_setup_io_apic(void) | 22 | static inline void smpboot_setup_io_apic(void) |
| 23 | { | 23 | { |
| 24 | } | 24 | } |
| 25 | |||
| 26 | static inline void smpboot_clear_io_apic(void) | ||
| 27 | { | ||
| 28 | } | ||
diff --git a/include/asm-x86/mach_apic.h b/include/asm-x86/mach_apic.h deleted file mode 100644 index 7b7115a0c1c9..000000000000 --- a/include/asm-x86/mach_apic.h +++ /dev/null | |||
| @@ -1,29 +0,0 @@ | |||
| 1 | #ifndef __ASM_MACH_APIC_H | ||
| 2 | #define __ASM_MACH_APIC_H | ||
| 3 | |||
| 4 | /* | ||
| 5 | * Copyright 2004 James Cleverdon, IBM. | ||
| 6 | * Subject to the GNU Public License, v.2 | ||
| 7 | * | ||
| 8 | * Generic APIC sub-arch defines. | ||
| 9 | * | ||
| 10 | * Hacked for x86-64 by James Cleverdon from i386 architecture code by | ||
| 11 | * Martin Bligh, Andi Kleen, James Bottomley, John Stultz, and | ||
| 12 | * James Cleverdon. | ||
| 13 | */ | ||
| 14 | |||
| 15 | #include <asm/genapic.h> | ||
| 16 | |||
| 17 | #define INT_DELIVERY_MODE (genapic->int_delivery_mode) | ||
| 18 | #define INT_DEST_MODE (genapic->int_dest_mode) | ||
| 19 | #define TARGET_CPUS (genapic->target_cpus()) | ||
| 20 | #define vector_allocation_domain (genapic->vector_allocation_domain) | ||
| 21 | #define apic_id_registered (genapic->apic_id_registered) | ||
| 22 | #define init_apic_ldr (genapic->init_apic_ldr) | ||
| 23 | #define send_IPI_mask (genapic->send_IPI_mask) | ||
| 24 | #define send_IPI_allbutself (genapic->send_IPI_allbutself) | ||
| 25 | #define send_IPI_all (genapic->send_IPI_all) | ||
| 26 | #define cpu_mask_to_apicid (genapic->cpu_mask_to_apicid) | ||
| 27 | #define phys_pkg_id (genapic->phys_pkg_id) | ||
| 28 | |||
| 29 | #endif /* __ASM_MACH_APIC_H */ | ||
diff --git a/include/asm-x86/mc146818rtc.h b/include/asm-x86/mc146818rtc.h index cdd9f965835a..daf1ccde77af 100644 --- a/include/asm-x86/mc146818rtc.h +++ b/include/asm-x86/mc146818rtc.h | |||
| @@ -42,7 +42,7 @@ extern volatile unsigned long cmos_lock; | |||
| 42 | static inline void lock_cmos(unsigned char reg) | 42 | static inline void lock_cmos(unsigned char reg) |
| 43 | { | 43 | { |
| 44 | unsigned long new; | 44 | unsigned long new; |
| 45 | new = ((smp_processor_id()+1) << 8) | reg; | 45 | new = ((smp_processor_id() + 1) << 8) | reg; |
| 46 | for (;;) { | 46 | for (;;) { |
| 47 | if (cmos_lock) { | 47 | if (cmos_lock) { |
| 48 | cpu_relax(); | 48 | cpu_relax(); |
| @@ -57,22 +57,26 @@ static inline void unlock_cmos(void) | |||
| 57 | { | 57 | { |
| 58 | cmos_lock = 0; | 58 | cmos_lock = 0; |
| 59 | } | 59 | } |
| 60 | |||
| 60 | static inline int do_i_have_lock_cmos(void) | 61 | static inline int do_i_have_lock_cmos(void) |
| 61 | { | 62 | { |
| 62 | return (cmos_lock >> 8) == (smp_processor_id()+1); | 63 | return (cmos_lock >> 8) == (smp_processor_id() + 1); |
| 63 | } | 64 | } |
| 65 | |||
| 64 | static inline unsigned char current_lock_cmos_reg(void) | 66 | static inline unsigned char current_lock_cmos_reg(void) |
| 65 | { | 67 | { |
| 66 | return cmos_lock & 0xff; | 68 | return cmos_lock & 0xff; |
| 67 | } | 69 | } |
| 68 | #define lock_cmos_prefix(reg) \ | 70 | |
| 71 | #define lock_cmos_prefix(reg) \ | ||
| 69 | do { \ | 72 | do { \ |
| 70 | unsigned long cmos_flags; \ | 73 | unsigned long cmos_flags; \ |
| 71 | local_irq_save(cmos_flags); \ | 74 | local_irq_save(cmos_flags); \ |
| 72 | lock_cmos(reg) | 75 | lock_cmos(reg) |
| 73 | #define lock_cmos_suffix(reg) \ | 76 | |
| 74 | unlock_cmos(); \ | 77 | #define lock_cmos_suffix(reg) \ |
| 75 | local_irq_restore(cmos_flags); \ | 78 | unlock_cmos(); \ |
| 79 | local_irq_restore(cmos_flags); \ | ||
| 76 | } while (0) | 80 | } while (0) |
| 77 | #else | 81 | #else |
| 78 | #define lock_cmos_prefix(reg) do {} while (0) | 82 | #define lock_cmos_prefix(reg) do {} while (0) |
diff --git a/include/asm-x86/mca_dma.h b/include/asm-x86/mca_dma.h index fbb1f3b71279..c3dca6edc6b1 100644 --- a/include/asm-x86/mca_dma.h +++ b/include/asm-x86/mca_dma.h | |||
| @@ -12,18 +12,18 @@ | |||
| 12 | * count by 2 when using 16-bit dma; that is not handled by these functions. | 12 | * count by 2 when using 16-bit dma; that is not handled by these functions. |
| 13 | * | 13 | * |
| 14 | * Ramen Noodles are yummy. | 14 | * Ramen Noodles are yummy. |
| 15 | * | 15 | * |
| 16 | * 1998 Tymm Twillman <tymm@computer.org> | 16 | * 1998 Tymm Twillman <tymm@computer.org> |
| 17 | */ | 17 | */ |
| 18 | 18 | ||
| 19 | /* | 19 | /* |
| 20 | * Registers that are used by the DMA controller; FN is the function register | 20 | * Registers that are used by the DMA controller; FN is the function register |
| 21 | * (tell the controller what to do) and EXE is the execution register (how | 21 | * (tell the controller what to do) and EXE is the execution register (how |
| 22 | * to do it) | 22 | * to do it) |
| 23 | */ | 23 | */ |
| 24 | 24 | ||
| 25 | #define MCA_DMA_REG_FN 0x18 | 25 | #define MCA_DMA_REG_FN 0x18 |
| 26 | #define MCA_DMA_REG_EXE 0x1A | 26 | #define MCA_DMA_REG_EXE 0x1A |
| 27 | 27 | ||
| 28 | /* | 28 | /* |
| 29 | * Functions that the DMA controller can do | 29 | * Functions that the DMA controller can do |
| @@ -43,9 +43,9 @@ | |||
| 43 | 43 | ||
| 44 | /* | 44 | /* |
| 45 | * Modes (used by setting MCA_DMA_FN_MODE in the function register) | 45 | * Modes (used by setting MCA_DMA_FN_MODE in the function register) |
| 46 | * | 46 | * |
| 47 | * Note that the MODE_READ is read from memory (write to device), and | 47 | * Note that the MODE_READ is read from memory (write to device), and |
| 48 | * MODE_WRITE is vice-versa. | 48 | * MODE_WRITE is vice-versa. |
| 49 | */ | 49 | */ |
| 50 | 50 | ||
| 51 | #define MCA_DMA_MODE_XFER 0x04 /* read by default */ | 51 | #define MCA_DMA_MODE_XFER 0x04 /* read by default */ |
| @@ -63,7 +63,7 @@ | |||
| 63 | * IRQ context. | 63 | * IRQ context. |
| 64 | */ | 64 | */ |
| 65 | 65 | ||
| 66 | static __inline__ void mca_enable_dma(unsigned int dmanr) | 66 | static inline void mca_enable_dma(unsigned int dmanr) |
| 67 | { | 67 | { |
| 68 | outb(MCA_DMA_FN_RESET_MASK | dmanr, MCA_DMA_REG_FN); | 68 | outb(MCA_DMA_FN_RESET_MASK | dmanr, MCA_DMA_REG_FN); |
| 69 | } | 69 | } |
| @@ -76,7 +76,7 @@ static __inline__ void mca_enable_dma(unsigned int dmanr) | |||
| 76 | * IRQ context. | 76 | * IRQ context. |
| 77 | */ | 77 | */ |
| 78 | 78 | ||
| 79 | static __inline__ void mca_disable_dma(unsigned int dmanr) | 79 | static inline void mca_disable_dma(unsigned int dmanr) |
| 80 | { | 80 | { |
| 81 | outb(MCA_DMA_FN_MASK | dmanr, MCA_DMA_REG_FN); | 81 | outb(MCA_DMA_FN_MASK | dmanr, MCA_DMA_REG_FN); |
| 82 | } | 82 | } |
| @@ -87,10 +87,10 @@ static __inline__ void mca_disable_dma(unsigned int dmanr) | |||
| 87 | * @a: 24bit bus address | 87 | * @a: 24bit bus address |
| 88 | * | 88 | * |
| 89 | * Load the address register in the DMA controller. This has a 24bit | 89 | * Load the address register in the DMA controller. This has a 24bit |
| 90 | * limitation (16Mb). | 90 | * limitation (16Mb). |
| 91 | */ | 91 | */ |
| 92 | 92 | ||
| 93 | static __inline__ void mca_set_dma_addr(unsigned int dmanr, unsigned int a) | 93 | static inline void mca_set_dma_addr(unsigned int dmanr, unsigned int a) |
| 94 | { | 94 | { |
| 95 | outb(MCA_DMA_FN_SET_ADDR | dmanr, MCA_DMA_REG_FN); | 95 | outb(MCA_DMA_FN_SET_ADDR | dmanr, MCA_DMA_REG_FN); |
| 96 | outb(a & 0xff, MCA_DMA_REG_EXE); | 96 | outb(a & 0xff, MCA_DMA_REG_EXE); |
| @@ -106,14 +106,14 @@ static __inline__ void mca_set_dma_addr(unsigned int dmanr, unsigned int a) | |||
| 106 | * limitation (16Mb). The return is a bus address. | 106 | * limitation (16Mb). The return is a bus address. |
| 107 | */ | 107 | */ |
| 108 | 108 | ||
| 109 | static __inline__ unsigned int mca_get_dma_addr(unsigned int dmanr) | 109 | static inline unsigned int mca_get_dma_addr(unsigned int dmanr) |
| 110 | { | 110 | { |
| 111 | unsigned int addr; | 111 | unsigned int addr; |
| 112 | 112 | ||
| 113 | outb(MCA_DMA_FN_GET_ADDR | dmanr, MCA_DMA_REG_FN); | 113 | outb(MCA_DMA_FN_GET_ADDR | dmanr, MCA_DMA_REG_FN); |
| 114 | addr = inb(MCA_DMA_REG_EXE); | 114 | addr = inb(MCA_DMA_REG_EXE); |
| 115 | addr |= inb(MCA_DMA_REG_EXE) << 8; | 115 | addr |= inb(MCA_DMA_REG_EXE) << 8; |
| 116 | addr |= inb(MCA_DMA_REG_EXE) << 16; | 116 | addr |= inb(MCA_DMA_REG_EXE) << 16; |
| 117 | 117 | ||
| 118 | return addr; | 118 | return addr; |
| 119 | } | 119 | } |
| @@ -127,7 +127,7 @@ static __inline__ unsigned int mca_get_dma_addr(unsigned int dmanr) | |||
| 127 | * Setting a count of zero will not do what you expect. | 127 | * Setting a count of zero will not do what you expect. |
| 128 | */ | 128 | */ |
| 129 | 129 | ||
| 130 | static __inline__ void mca_set_dma_count(unsigned int dmanr, unsigned int count) | 130 | static inline void mca_set_dma_count(unsigned int dmanr, unsigned int count) |
| 131 | { | 131 | { |
| 132 | count--; /* transfers one more than count -- correct for this */ | 132 | count--; /* transfers one more than count -- correct for this */ |
| 133 | 133 | ||
| @@ -144,7 +144,7 @@ static __inline__ void mca_set_dma_count(unsigned int dmanr, unsigned int count) | |||
| 144 | * on this DMA channel. | 144 | * on this DMA channel. |
| 145 | */ | 145 | */ |
| 146 | 146 | ||
| 147 | static __inline__ unsigned int mca_get_dma_residue(unsigned int dmanr) | 147 | static inline unsigned int mca_get_dma_residue(unsigned int dmanr) |
| 148 | { | 148 | { |
| 149 | unsigned short count; | 149 | unsigned short count; |
| 150 | 150 | ||
| @@ -164,12 +164,12 @@ static __inline__ unsigned int mca_get_dma_residue(unsigned int dmanr) | |||
| 164 | * with an I/O port target. | 164 | * with an I/O port target. |
| 165 | */ | 165 | */ |
| 166 | 166 | ||
| 167 | static __inline__ void mca_set_dma_io(unsigned int dmanr, unsigned int io_addr) | 167 | static inline void mca_set_dma_io(unsigned int dmanr, unsigned int io_addr) |
| 168 | { | 168 | { |
| 169 | /* | 169 | /* |
| 170 | * DMA from a port address -- set the io address | 170 | * DMA from a port address -- set the io address |
| 171 | */ | 171 | */ |
| 172 | 172 | ||
| 173 | outb(MCA_DMA_FN_SET_IO | dmanr, MCA_DMA_REG_FN); | 173 | outb(MCA_DMA_FN_SET_IO | dmanr, MCA_DMA_REG_FN); |
| 174 | outb(io_addr & 0xff, MCA_DMA_REG_EXE); | 174 | outb(io_addr & 0xff, MCA_DMA_REG_EXE); |
| 175 | outb((io_addr >> 8) & 0xff, MCA_DMA_REG_EXE); | 175 | outb((io_addr >> 8) & 0xff, MCA_DMA_REG_EXE); |
| @@ -192,7 +192,7 @@ static __inline__ void mca_set_dma_io(unsigned int dmanr, unsigned int io_addr) | |||
| 192 | * %MCA_DMA_MODE_16 to do 16bit transfers. | 192 | * %MCA_DMA_MODE_16 to do 16bit transfers. |
| 193 | */ | 193 | */ |
| 194 | 194 | ||
| 195 | static __inline__ void mca_set_dma_mode(unsigned int dmanr, unsigned int mode) | 195 | static inline void mca_set_dma_mode(unsigned int dmanr, unsigned int mode) |
| 196 | { | 196 | { |
| 197 | outb(MCA_DMA_FN_SET_MODE | dmanr, MCA_DMA_REG_FN); | 197 | outb(MCA_DMA_FN_SET_MODE | dmanr, MCA_DMA_REG_FN); |
| 198 | outb(mode, MCA_DMA_REG_EXE); | 198 | outb(mode, MCA_DMA_REG_EXE); |
diff --git a/include/asm-x86/mmu.h b/include/asm-x86/mmu.h index efa962c38897..00e88679e11f 100644 --- a/include/asm-x86/mmu.h +++ b/include/asm-x86/mmu.h | |||
| @@ -10,10 +10,10 @@ | |||
| 10 | * | 10 | * |
| 11 | * cpu_vm_mask is used to optimize ldt flushing. | 11 | * cpu_vm_mask is used to optimize ldt flushing. |
| 12 | */ | 12 | */ |
| 13 | typedef struct { | 13 | typedef struct { |
| 14 | void *ldt; | 14 | void *ldt; |
| 15 | #ifdef CONFIG_X86_64 | 15 | #ifdef CONFIG_X86_64 |
| 16 | rwlock_t ldtlock; | 16 | rwlock_t ldtlock; |
| 17 | #endif | 17 | #endif |
| 18 | int size; | 18 | int size; |
| 19 | struct mutex lock; | 19 | struct mutex lock; |
diff --git a/include/asm-x86/mmu_context_32.h b/include/asm-x86/mmu_context_32.h index 8198d1cca1f3..9756ae0f1dd3 100644 --- a/include/asm-x86/mmu_context_32.h +++ b/include/asm-x86/mmu_context_32.h | |||
| @@ -62,7 +62,7 @@ static inline void switch_mm(struct mm_struct *prev, | |||
| 62 | BUG_ON(per_cpu(cpu_tlbstate, cpu).active_mm != next); | 62 | BUG_ON(per_cpu(cpu_tlbstate, cpu).active_mm != next); |
| 63 | 63 | ||
| 64 | if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) { | 64 | if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) { |
| 65 | /* We were in lazy tlb mode and leave_mm disabled | 65 | /* We were in lazy tlb mode and leave_mm disabled |
| 66 | * tlb flush IPI delivery. We must reload %cr3. | 66 | * tlb flush IPI delivery. We must reload %cr3. |
| 67 | */ | 67 | */ |
| 68 | load_cr3(next->pgd); | 68 | load_cr3(next->pgd); |
| @@ -75,10 +75,10 @@ static inline void switch_mm(struct mm_struct *prev, | |||
| 75 | #define deactivate_mm(tsk, mm) \ | 75 | #define deactivate_mm(tsk, mm) \ |
| 76 | asm("movl %0,%%gs": :"r" (0)); | 76 | asm("movl %0,%%gs": :"r" (0)); |
| 77 | 77 | ||
| 78 | #define activate_mm(prev, next) \ | 78 | #define activate_mm(prev, next) \ |
| 79 | do { \ | 79 | do { \ |
| 80 | paravirt_activate_mm(prev, next); \ | 80 | paravirt_activate_mm((prev), (next)); \ |
| 81 | switch_mm((prev),(next),NULL); \ | 81 | switch_mm((prev), (next), NULL); \ |
| 82 | } while(0); | 82 | } while (0); |
| 83 | 83 | ||
| 84 | #endif | 84 | #endif |
diff --git a/include/asm-x86/mmu_context_64.h b/include/asm-x86/mmu_context_64.h index ad6dc821ef9e..ca44c71e7fb3 100644 --- a/include/asm-x86/mmu_context_64.h +++ b/include/asm-x86/mmu_context_64.h | |||
| @@ -20,12 +20,12 @@ void destroy_context(struct mm_struct *mm); | |||
| 20 | static inline void enter_lazy_tlb(struct mm_struct *mm, struct task_struct *tsk) | 20 | static inline void enter_lazy_tlb(struct mm_struct *mm, struct task_struct *tsk) |
| 21 | { | 21 | { |
| 22 | #ifdef CONFIG_SMP | 22 | #ifdef CONFIG_SMP |
| 23 | if (read_pda(mmu_state) == TLBSTATE_OK) | 23 | if (read_pda(mmu_state) == TLBSTATE_OK) |
| 24 | write_pda(mmu_state, TLBSTATE_LAZY); | 24 | write_pda(mmu_state, TLBSTATE_LAZY); |
| 25 | #endif | 25 | #endif |
| 26 | } | 26 | } |
| 27 | 27 | ||
| 28 | static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, | 28 | static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, |
| 29 | struct task_struct *tsk) | 29 | struct task_struct *tsk) |
| 30 | { | 30 | { |
| 31 | unsigned cpu = smp_processor_id(); | 31 | unsigned cpu = smp_processor_id(); |
| @@ -39,7 +39,7 @@ static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, | |||
| 39 | cpu_set(cpu, next->cpu_vm_mask); | 39 | cpu_set(cpu, next->cpu_vm_mask); |
| 40 | load_cr3(next->pgd); | 40 | load_cr3(next->pgd); |
| 41 | 41 | ||
| 42 | if (unlikely(next->context.ldt != prev->context.ldt)) | 42 | if (unlikely(next->context.ldt != prev->context.ldt)) |
| 43 | load_LDT_nolock(&next->context); | 43 | load_LDT_nolock(&next->context); |
| 44 | } | 44 | } |
| 45 | #ifdef CONFIG_SMP | 45 | #ifdef CONFIG_SMP |
| @@ -48,7 +48,7 @@ static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, | |||
| 48 | if (read_pda(active_mm) != next) | 48 | if (read_pda(active_mm) != next) |
| 49 | BUG(); | 49 | BUG(); |
| 50 | if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) { | 50 | if (!cpu_test_and_set(cpu, next->cpu_vm_mask)) { |
| 51 | /* We were in lazy tlb mode and leave_mm disabled | 51 | /* We were in lazy tlb mode and leave_mm disabled |
| 52 | * tlb flush IPI delivery. We must reload CR3 | 52 | * tlb flush IPI delivery. We must reload CR3 |
| 53 | * to make sure to use no freed page tables. | 53 | * to make sure to use no freed page tables. |
| 54 | */ | 54 | */ |
| @@ -59,13 +59,14 @@ static inline void switch_mm(struct mm_struct *prev, struct mm_struct *next, | |||
| 59 | #endif | 59 | #endif |
| 60 | } | 60 | } |
| 61 | 61 | ||
| 62 | #define deactivate_mm(tsk,mm) do { \ | 62 | #define deactivate_mm(tsk, mm) \ |
| 63 | load_gs_index(0); \ | 63 | do { \ |
| 64 | asm volatile("movl %0,%%fs"::"r"(0)); \ | 64 | load_gs_index(0); \ |
| 65 | } while(0) | 65 | asm volatile("movl %0,%%fs"::"r"(0)); \ |
| 66 | } while (0) | ||
| 66 | 67 | ||
| 67 | #define activate_mm(prev, next) \ | 68 | #define activate_mm(prev, next) \ |
| 68 | switch_mm((prev),(next),NULL) | 69 | switch_mm((prev), (next), NULL) |
| 69 | 70 | ||
| 70 | 71 | ||
| 71 | #endif | 72 | #endif |
diff --git a/include/asm-x86/mmx.h b/include/asm-x86/mmx.h index 46b71da99869..940881218ff8 100644 --- a/include/asm-x86/mmx.h +++ b/include/asm-x86/mmx.h | |||
| @@ -6,7 +6,7 @@ | |||
| 6 | */ | 6 | */ |
| 7 | 7 | ||
| 8 | #include <linux/types.h> | 8 | #include <linux/types.h> |
| 9 | 9 | ||
| 10 | extern void *_mmx_memcpy(void *to, const void *from, size_t size); | 10 | extern void *_mmx_memcpy(void *to, const void *from, size_t size); |
| 11 | extern void mmx_clear_page(void *page); | 11 | extern void mmx_clear_page(void *page); |
| 12 | extern void mmx_copy_page(void *to, void *from); | 12 | extern void mmx_copy_page(void *to, void *from); |
diff --git a/include/asm-x86/mmzone_32.h b/include/asm-x86/mmzone_32.h index 274a59566c45..cb2cad0b65a7 100644 --- a/include/asm-x86/mmzone_32.h +++ b/include/asm-x86/mmzone_32.h | |||
| @@ -18,7 +18,7 @@ extern struct pglist_data *node_data[]; | |||
| 18 | #include <asm/srat.h> | 18 | #include <asm/srat.h> |
| 19 | #endif | 19 | #endif |
| 20 | 20 | ||
| 21 | extern int get_memcfg_numa_flat(void ); | 21 | extern int get_memcfg_numa_flat(void); |
| 22 | /* | 22 | /* |
| 23 | * This allows any one NUMA architecture to be compiled | 23 | * This allows any one NUMA architecture to be compiled |
| 24 | * for, and still fall back to the flat function if it | 24 | * for, and still fall back to the flat function if it |
| @@ -129,7 +129,7 @@ static inline int pfn_valid(int pfn) | |||
| 129 | struct pglist_data __maybe_unused \ | 129 | struct pglist_data __maybe_unused \ |
| 130 | *__alloc_bootmem_node__pgdat = (pgdat); \ | 130 | *__alloc_bootmem_node__pgdat = (pgdat); \ |
| 131 | __alloc_bootmem_node(NODE_DATA(0), (x), PAGE_SIZE, \ | 131 | __alloc_bootmem_node(NODE_DATA(0), (x), PAGE_SIZE, \ |
| 132 | __pa(MAX_DMA_ADDRESS)) \ | 132 | __pa(MAX_DMA_ADDRESS)); \ |
| 133 | }) | 133 | }) |
| 134 | #define alloc_bootmem_low_pages_node(pgdat, x) \ | 134 | #define alloc_bootmem_low_pages_node(pgdat, x) \ |
| 135 | ({ \ | 135 | ({ \ |
diff --git a/include/asm-x86/mmzone_64.h b/include/asm-x86/mmzone_64.h index ebaf9663aa8a..594bd0dc1d08 100644 --- a/include/asm-x86/mmzone_64.h +++ b/include/asm-x86/mmzone_64.h | |||
| @@ -7,7 +7,7 @@ | |||
| 7 | 7 | ||
| 8 | #ifdef CONFIG_NUMA | 8 | #ifdef CONFIG_NUMA |
| 9 | 9 | ||
| 10 | #define VIRTUAL_BUG_ON(x) | 10 | #define VIRTUAL_BUG_ON(x) |
| 11 | 11 | ||
| 12 | #include <asm/smp.h> | 12 | #include <asm/smp.h> |
| 13 | 13 | ||
| @@ -16,7 +16,7 @@ struct memnode { | |||
| 16 | int shift; | 16 | int shift; |
| 17 | unsigned int mapsize; | 17 | unsigned int mapsize; |
| 18 | s16 *map; | 18 | s16 *map; |
| 19 | s16 embedded_map[64-8]; | 19 | s16 embedded_map[64 - 8]; |
| 20 | } ____cacheline_aligned; /* total size = 128 bytes */ | 20 | } ____cacheline_aligned; /* total size = 128 bytes */ |
| 21 | extern struct memnode memnode; | 21 | extern struct memnode memnode; |
| 22 | #define memnode_shift memnode.shift | 22 | #define memnode_shift memnode.shift |
| @@ -25,27 +25,27 @@ extern struct memnode memnode; | |||
| 25 | 25 | ||
| 26 | extern struct pglist_data *node_data[]; | 26 | extern struct pglist_data *node_data[]; |
| 27 | 27 | ||
| 28 | static inline __attribute__((pure)) int phys_to_nid(unsigned long addr) | 28 | static inline __attribute__((pure)) int phys_to_nid(unsigned long addr) |
| 29 | { | 29 | { |
| 30 | unsigned nid; | 30 | unsigned nid; |
| 31 | VIRTUAL_BUG_ON(!memnodemap); | 31 | VIRTUAL_BUG_ON(!memnodemap); |
| 32 | VIRTUAL_BUG_ON((addr >> memnode_shift) >= memnodemapsize); | 32 | VIRTUAL_BUG_ON((addr >> memnode_shift) >= memnodemapsize); |
| 33 | nid = memnodemap[addr >> memnode_shift]; | 33 | nid = memnodemap[addr >> memnode_shift]; |
| 34 | VIRTUAL_BUG_ON(nid >= MAX_NUMNODES || !node_data[nid]); | 34 | VIRTUAL_BUG_ON(nid >= MAX_NUMNODES || !node_data[nid]); |
| 35 | return nid; | 35 | return nid; |
| 36 | } | 36 | } |
| 37 | 37 | ||
| 38 | #define NODE_DATA(nid) (node_data[nid]) | 38 | #define NODE_DATA(nid) (node_data[nid]) |
| 39 | 39 | ||
| 40 | #define node_start_pfn(nid) (NODE_DATA(nid)->node_start_pfn) | 40 | #define node_start_pfn(nid) (NODE_DATA(nid)->node_start_pfn) |
| 41 | #define node_end_pfn(nid) (NODE_DATA(nid)->node_start_pfn + \ | 41 | #define node_end_pfn(nid) (NODE_DATA(nid)->node_start_pfn + \ |
| 42 | NODE_DATA(nid)->node_spanned_pages) | 42 | NODE_DATA(nid)->node_spanned_pages) |
| 43 | 43 | ||
| 44 | extern int early_pfn_to_nid(unsigned long pfn); | 44 | extern int early_pfn_to_nid(unsigned long pfn); |
| 45 | 45 | ||
| 46 | #ifdef CONFIG_NUMA_EMU | 46 | #ifdef CONFIG_NUMA_EMU |
| 47 | #define FAKE_NODE_MIN_SIZE (64*1024*1024) | 47 | #define FAKE_NODE_MIN_SIZE (64 * 1024 * 1024) |
| 48 | #define FAKE_NODE_MIN_HASH_MASK (~(FAKE_NODE_MIN_SIZE - 1uL)) | 48 | #define FAKE_NODE_MIN_HASH_MASK (~(FAKE_NODE_MIN_SIZE - 1UL)) |
| 49 | #endif | 49 | #endif |
| 50 | 50 | ||
| 51 | #endif | 51 | #endif |
diff --git a/include/asm-x86/mpspec.h b/include/asm-x86/mpspec.h index 781ad74ab9e9..57a991b9c053 100644 --- a/include/asm-x86/mpspec.h +++ b/include/asm-x86/mpspec.h | |||
| @@ -1,16 +1,13 @@ | |||
| 1 | #ifndef _AM_X86_MPSPEC_H | 1 | #ifndef _AM_X86_MPSPEC_H |
| 2 | #define _AM_X86_MPSPEC_H | 2 | #define _AM_X86_MPSPEC_H |
| 3 | 3 | ||
| 4 | #include <linux/init.h> | ||
| 5 | |||
| 4 | #include <asm/mpspec_def.h> | 6 | #include <asm/mpspec_def.h> |
| 5 | 7 | ||
| 6 | #ifdef CONFIG_X86_32 | 8 | #ifdef CONFIG_X86_32 |
| 7 | #include <mach_mpspec.h> | 9 | #include <mach_mpspec.h> |
| 8 | 10 | ||
| 9 | extern int mp_bus_id_to_type[MAX_MP_BUSSES]; | ||
| 10 | extern int mp_bus_id_to_node[MAX_MP_BUSSES]; | ||
| 11 | extern int mp_bus_id_to_local[MAX_MP_BUSSES]; | ||
| 12 | extern int quad_local_to_mp_bus_id[NR_CPUS/4][4]; | ||
| 13 | |||
| 14 | extern unsigned int def_to_bigsmp; | 11 | extern unsigned int def_to_bigsmp; |
| 15 | extern int apic_version[MAX_APICS]; | 12 | extern int apic_version[MAX_APICS]; |
| 16 | extern u8 apicid_2_node[]; | 13 | extern u8 apicid_2_node[]; |
| @@ -24,27 +21,30 @@ extern int pic_mode; | |||
| 24 | /* Each PCI slot may be a combo card with its own bus. 4 IRQ pins per slot. */ | 21 | /* Each PCI slot may be a combo card with its own bus. 4 IRQ pins per slot. */ |
| 25 | #define MAX_IRQ_SOURCES (MAX_MP_BUSSES * 4) | 22 | #define MAX_IRQ_SOURCES (MAX_MP_BUSSES * 4) |
| 26 | 23 | ||
| 27 | extern DECLARE_BITMAP(mp_bus_not_pci, MAX_MP_BUSSES); | 24 | extern void early_find_smp_config(void); |
| 25 | extern void early_get_smp_config(void); | ||
| 28 | 26 | ||
| 29 | #endif | 27 | #endif |
| 30 | 28 | ||
| 29 | #if defined(CONFIG_MCA) || defined(CONFIG_EISA) | ||
| 30 | extern int mp_bus_id_to_type[MAX_MP_BUSSES]; | ||
| 31 | #endif | ||
| 32 | |||
| 33 | extern DECLARE_BITMAP(mp_bus_not_pci, MAX_MP_BUSSES); | ||
| 34 | |||
| 31 | extern int mp_bus_id_to_pci_bus[MAX_MP_BUSSES]; | 35 | extern int mp_bus_id_to_pci_bus[MAX_MP_BUSSES]; |
| 32 | 36 | ||
| 33 | extern unsigned int boot_cpu_physical_apicid; | 37 | extern unsigned int boot_cpu_physical_apicid; |
| 34 | extern int smp_found_config; | 38 | extern int smp_found_config; |
| 35 | extern int nr_ioapics; | ||
| 36 | extern int mp_irq_entries; | ||
| 37 | extern struct mpc_config_intsrc mp_irqs[MAX_IRQ_SOURCES]; | ||
| 38 | extern int mpc_default_type; | 39 | extern int mpc_default_type; |
| 39 | extern unsigned long mp_lapic_addr; | 40 | extern unsigned long mp_lapic_addr; |
| 40 | 41 | ||
| 41 | extern void find_smp_config(void); | 42 | extern void find_smp_config(void); |
| 42 | extern void get_smp_config(void); | 43 | extern void get_smp_config(void); |
| 43 | 44 | ||
| 45 | void __cpuinit generic_processor_info(int apicid, int version); | ||
| 44 | #ifdef CONFIG_ACPI | 46 | #ifdef CONFIG_ACPI |
| 45 | extern void mp_register_lapic(u8 id, u8 enabled); | 47 | extern void mp_register_ioapic(int id, u32 address, u32 gsi_base); |
| 46 | extern void mp_register_lapic_address(u64 address); | ||
| 47 | extern void mp_register_ioapic(u8 id, u32 address, u32 gsi_base); | ||
| 48 | extern void mp_override_legacy_irq(u8 bus_irq, u8 polarity, u8 trigger, | 48 | extern void mp_override_legacy_irq(u8 bus_irq, u8 polarity, u8 trigger, |
| 49 | u32 gsi); | 49 | u32 gsi); |
| 50 | extern void mp_config_acpi_legacy_irqs(void); | 50 | extern void mp_config_acpi_legacy_irqs(void); |
| @@ -53,8 +53,7 @@ extern int mp_register_gsi(u32 gsi, int edge_level, int active_high_low); | |||
| 53 | 53 | ||
| 54 | #define PHYSID_ARRAY_SIZE BITS_TO_LONGS(MAX_APICS) | 54 | #define PHYSID_ARRAY_SIZE BITS_TO_LONGS(MAX_APICS) |
| 55 | 55 | ||
| 56 | struct physid_mask | 56 | struct physid_mask { |
| 57 | { | ||
| 58 | unsigned long mask[PHYSID_ARRAY_SIZE]; | 57 | unsigned long mask[PHYSID_ARRAY_SIZE]; |
| 59 | }; | 58 | }; |
| 60 | 59 | ||
| @@ -63,34 +62,34 @@ typedef struct physid_mask physid_mask_t; | |||
| 63 | #define physid_set(physid, map) set_bit(physid, (map).mask) | 62 | #define physid_set(physid, map) set_bit(physid, (map).mask) |
| 64 | #define physid_clear(physid, map) clear_bit(physid, (map).mask) | 63 | #define physid_clear(physid, map) clear_bit(physid, (map).mask) |
| 65 | #define physid_isset(physid, map) test_bit(physid, (map).mask) | 64 | #define physid_isset(physid, map) test_bit(physid, (map).mask) |
| 66 | #define physid_test_and_set(physid, map) \ | 65 | #define physid_test_and_set(physid, map) \ |
| 67 | test_and_set_bit(physid, (map).mask) | 66 | test_and_set_bit(physid, (map).mask) |
| 68 | 67 | ||
| 69 | #define physids_and(dst, src1, src2) \ | 68 | #define physids_and(dst, src1, src2) \ |
| 70 | bitmap_and((dst).mask, (src1).mask, (src2).mask, MAX_APICS) | 69 | bitmap_and((dst).mask, (src1).mask, (src2).mask, MAX_APICS) |
| 71 | 70 | ||
| 72 | #define physids_or(dst, src1, src2) \ | 71 | #define physids_or(dst, src1, src2) \ |
| 73 | bitmap_or((dst).mask, (src1).mask, (src2).mask, MAX_APICS) | 72 | bitmap_or((dst).mask, (src1).mask, (src2).mask, MAX_APICS) |
| 74 | 73 | ||
| 75 | #define physids_clear(map) \ | 74 | #define physids_clear(map) \ |
| 76 | bitmap_zero((map).mask, MAX_APICS) | 75 | bitmap_zero((map).mask, MAX_APICS) |
| 77 | 76 | ||
| 78 | #define physids_complement(dst, src) \ | 77 | #define physids_complement(dst, src) \ |
| 79 | bitmap_complement((dst).mask, (src).mask, MAX_APICS) | 78 | bitmap_complement((dst).mask, (src).mask, MAX_APICS) |
| 80 | 79 | ||
| 81 | #define physids_empty(map) \ | 80 | #define physids_empty(map) \ |
| 82 | bitmap_empty((map).mask, MAX_APICS) | 81 | bitmap_empty((map).mask, MAX_APICS) |
| 83 | 82 | ||
| 84 | #define physids_equal(map1, map2) \ | 83 | #define physids_equal(map1, map2) \ |
| 85 | bitmap_equal((map1).mask, (map2).mask, MAX_APICS) | 84 | bitmap_equal((map1).mask, (map2).mask, MAX_APICS) |
| 86 | 85 | ||
| 87 | #define physids_weight(map) \ | 86 | #define physids_weight(map) \ |
| 88 | bitmap_weight((map).mask, MAX_APICS) | 87 | bitmap_weight((map).mask, MAX_APICS) |
| 89 | 88 | ||
| 90 | #define physids_shift_right(d, s, n) \ | 89 | #define physids_shift_right(d, s, n) \ |
| 91 | bitmap_shift_right((d).mask, (s).mask, n, MAX_APICS) | 90 | bitmap_shift_right((d).mask, (s).mask, n, MAX_APICS) |
| 92 | 91 | ||
| 93 | #define physids_shift_left(d, s, n) \ | 92 | #define physids_shift_left(d, s, n) \ |
| 94 | bitmap_shift_left((d).mask, (s).mask, n, MAX_APICS) | 93 | bitmap_shift_left((d).mask, (s).mask, n, MAX_APICS) |
| 95 | 94 | ||
| 96 | #define physids_coerce(map) ((map).mask[0]) | 95 | #define physids_coerce(map) ((map).mask[0]) |
diff --git a/include/asm-x86/mpspec_def.h b/include/asm-x86/mpspec_def.h index 3504617fe648..dc6ef85e3624 100644 --- a/include/asm-x86/mpspec_def.h +++ b/include/asm-x86/mpspec_def.h | |||
| @@ -11,7 +11,7 @@ | |||
| 11 | * information is. | 11 | * information is. |
| 12 | */ | 12 | */ |
| 13 | 13 | ||
| 14 | #define SMP_MAGIC_IDENT (('_'<<24)|('P'<<16)|('M'<<8)|'_') | 14 | #define SMP_MAGIC_IDENT (('_'<<24) | ('P'<<16) | ('M'<<8) | '_') |
| 15 | 15 | ||
| 16 | #ifdef CONFIG_X86_32 | 16 | #ifdef CONFIG_X86_32 |
| 17 | # define MAX_MPC_ENTRY 1024 | 17 | # define MAX_MPC_ENTRY 1024 |
| @@ -23,8 +23,7 @@ | |||
| 23 | # define MAX_APICS 255 | 23 | # define MAX_APICS 255 |
| 24 | #endif | 24 | #endif |
| 25 | 25 | ||
| 26 | struct intel_mp_floating | 26 | struct intel_mp_floating { |
| 27 | { | ||
| 28 | char mpf_signature[4]; /* "_MP_" */ | 27 | char mpf_signature[4]; /* "_MP_" */ |
| 29 | unsigned int mpf_physptr; /* Configuration table address */ | 28 | unsigned int mpf_physptr; /* Configuration table address */ |
| 30 | unsigned char mpf_length; /* Our length (paragraphs) */ | 29 | unsigned char mpf_length; /* Our length (paragraphs) */ |
| @@ -39,14 +38,13 @@ struct intel_mp_floating | |||
| 39 | 38 | ||
| 40 | #define MPC_SIGNATURE "PCMP" | 39 | #define MPC_SIGNATURE "PCMP" |
| 41 | 40 | ||
| 42 | struct mp_config_table | 41 | struct mp_config_table { |
| 43 | { | ||
| 44 | char mpc_signature[4]; | 42 | char mpc_signature[4]; |
| 45 | unsigned short mpc_length; /* Size of table */ | 43 | unsigned short mpc_length; /* Size of table */ |
| 46 | char mpc_spec; /* 0x01 */ | 44 | char mpc_spec; /* 0x01 */ |
| 47 | char mpc_checksum; | 45 | char mpc_checksum; |
| 48 | char mpc_oem[8]; | 46 | char mpc_oem[8]; |
| 49 | char mpc_productid[12]; | 47 | char mpc_productid[12]; |
| 50 | unsigned int mpc_oemptr; /* 0 if not present */ | 48 | unsigned int mpc_oemptr; /* 0 if not present */ |
| 51 | unsigned short mpc_oemsize; /* 0 if not present */ | 49 | unsigned short mpc_oemsize; /* 0 if not present */ |
| 52 | unsigned short mpc_oemcount; | 50 | unsigned short mpc_oemcount; |
| @@ -71,8 +69,7 @@ struct mp_config_table | |||
| 71 | #define CPU_MODEL_MASK 0x00F0 | 69 | #define CPU_MODEL_MASK 0x00F0 |
| 72 | #define CPU_FAMILY_MASK 0x0F00 | 70 | #define CPU_FAMILY_MASK 0x0F00 |
| 73 | 71 | ||
| 74 | struct mpc_config_processor | 72 | struct mpc_config_processor { |
| 75 | { | ||
| 76 | unsigned char mpc_type; | 73 | unsigned char mpc_type; |
| 77 | unsigned char mpc_apicid; /* Local APIC number */ | 74 | unsigned char mpc_apicid; /* Local APIC number */ |
| 78 | unsigned char mpc_apicver; /* Its versions */ | 75 | unsigned char mpc_apicver; /* Its versions */ |
| @@ -82,8 +79,7 @@ struct mpc_config_processor | |||
| 82 | unsigned int mpc_reserved[2]; | 79 | unsigned int mpc_reserved[2]; |
| 83 | }; | 80 | }; |
| 84 | 81 | ||
| 85 | struct mpc_config_bus | 82 | struct mpc_config_bus { |
| 86 | { | ||
| 87 | unsigned char mpc_type; | 83 | unsigned char mpc_type; |
| 88 | unsigned char mpc_busid; | 84 | unsigned char mpc_busid; |
| 89 | unsigned char mpc_bustype[6]; | 85 | unsigned char mpc_bustype[6]; |
| @@ -111,8 +107,7 @@ struct mpc_config_bus | |||
| 111 | 107 | ||
| 112 | #define MPC_APIC_USABLE 0x01 | 108 | #define MPC_APIC_USABLE 0x01 |
| 113 | 109 | ||
| 114 | struct mpc_config_ioapic | 110 | struct mpc_config_ioapic { |
| 115 | { | ||
| 116 | unsigned char mpc_type; | 111 | unsigned char mpc_type; |
| 117 | unsigned char mpc_apicid; | 112 | unsigned char mpc_apicid; |
| 118 | unsigned char mpc_apicver; | 113 | unsigned char mpc_apicver; |
| @@ -120,8 +115,7 @@ struct mpc_config_ioapic | |||
| 120 | unsigned int mpc_apicaddr; | 115 | unsigned int mpc_apicaddr; |
| 121 | }; | 116 | }; |
| 122 | 117 | ||
| 123 | struct mpc_config_intsrc | 118 | struct mpc_config_intsrc { |
| 124 | { | ||
| 125 | unsigned char mpc_type; | 119 | unsigned char mpc_type; |
| 126 | unsigned char mpc_irqtype; | 120 | unsigned char mpc_irqtype; |
| 127 | unsigned short mpc_irqflag; | 121 | unsigned short mpc_irqflag; |
| @@ -144,8 +138,7 @@ enum mp_irq_source_types { | |||
| 144 | 138 | ||
| 145 | #define MP_APIC_ALL 0xFF | 139 | #define MP_APIC_ALL 0xFF |
| 146 | 140 | ||
| 147 | struct mpc_config_lintsrc | 141 | struct mpc_config_lintsrc { |
| 148 | { | ||
| 149 | unsigned char mpc_type; | 142 | unsigned char mpc_type; |
| 150 | unsigned char mpc_irqtype; | 143 | unsigned char mpc_irqtype; |
| 151 | unsigned short mpc_irqflag; | 144 | unsigned short mpc_irqflag; |
| @@ -157,8 +150,7 @@ struct mpc_config_lintsrc | |||
| 157 | 150 | ||
| 158 | #define MPC_OEM_SIGNATURE "_OEM" | 151 | #define MPC_OEM_SIGNATURE "_OEM" |
| 159 | 152 | ||
| 160 | struct mp_config_oemtable | 153 | struct mp_config_oemtable { |
| 161 | { | ||
| 162 | char oem_signature[4]; | 154 | char oem_signature[4]; |
| 163 | unsigned short oem_length; /* Size of table */ | 155 | unsigned short oem_length; /* Size of table */ |
| 164 | char oem_rev; /* 0x01 */ | 156 | char oem_rev; /* 0x01 */ |
| @@ -166,17 +158,6 @@ struct mp_config_oemtable | |||
| 166 | char mpc_oem[8]; | 158 | char mpc_oem[8]; |
| 167 | }; | 159 | }; |
| 168 | 160 | ||
| 169 | struct mpc_config_translation | ||
| 170 | { | ||
| 171 | unsigned char mpc_type; | ||
| 172 | unsigned char trans_len; | ||
| 173 | unsigned char trans_type; | ||
| 174 | unsigned char trans_quad; | ||
| 175 | unsigned char trans_global; | ||
| 176 | unsigned char trans_local; | ||
| 177 | unsigned short trans_reserved; | ||
| 178 | }; | ||
| 179 | |||
| 180 | /* | 161 | /* |
| 181 | * Default configurations | 162 | * Default configurations |
| 182 | * | 163 | * |
| @@ -196,4 +177,3 @@ enum mp_bustype { | |||
| 196 | MP_BUS_MCA, | 177 | MP_BUS_MCA, |
| 197 | }; | 178 | }; |
| 198 | #endif | 179 | #endif |
| 199 | |||
diff --git a/include/asm-x86/msidef.h b/include/asm-x86/msidef.h index 5b8acddb70fb..296f29ce426d 100644 --- a/include/asm-x86/msidef.h +++ b/include/asm-x86/msidef.h | |||
| @@ -11,7 +11,8 @@ | |||
| 11 | 11 | ||
| 12 | #define MSI_DATA_VECTOR_SHIFT 0 | 12 | #define MSI_DATA_VECTOR_SHIFT 0 |
| 13 | #define MSI_DATA_VECTOR_MASK 0x000000ff | 13 | #define MSI_DATA_VECTOR_MASK 0x000000ff |
| 14 | #define MSI_DATA_VECTOR(v) (((v) << MSI_DATA_VECTOR_SHIFT) & MSI_DATA_VECTOR_MASK) | 14 | #define MSI_DATA_VECTOR(v) (((v) << MSI_DATA_VECTOR_SHIFT) & \ |
| 15 | MSI_DATA_VECTOR_MASK) | ||
| 15 | 16 | ||
| 16 | #define MSI_DATA_DELIVERY_MODE_SHIFT 8 | 17 | #define MSI_DATA_DELIVERY_MODE_SHIFT 8 |
| 17 | #define MSI_DATA_DELIVERY_FIXED (0 << MSI_DATA_DELIVERY_MODE_SHIFT) | 18 | #define MSI_DATA_DELIVERY_FIXED (0 << MSI_DATA_DELIVERY_MODE_SHIFT) |
| @@ -37,11 +38,14 @@ | |||
| 37 | #define MSI_ADDR_DEST_MODE_LOGICAL (1 << MSI_ADDR_DEST_MODE_SHIFT) | 38 | #define MSI_ADDR_DEST_MODE_LOGICAL (1 << MSI_ADDR_DEST_MODE_SHIFT) |
| 38 | 39 | ||
| 39 | #define MSI_ADDR_REDIRECTION_SHIFT 3 | 40 | #define MSI_ADDR_REDIRECTION_SHIFT 3 |
| 40 | #define MSI_ADDR_REDIRECTION_CPU (0 << MSI_ADDR_REDIRECTION_SHIFT) /* dedicated cpu */ | 41 | #define MSI_ADDR_REDIRECTION_CPU (0 << MSI_ADDR_REDIRECTION_SHIFT) |
| 41 | #define MSI_ADDR_REDIRECTION_LOWPRI (1 << MSI_ADDR_REDIRECTION_SHIFT) /* lowest priority */ | 42 | /* dedicated cpu */ |
| 43 | #define MSI_ADDR_REDIRECTION_LOWPRI (1 << MSI_ADDR_REDIRECTION_SHIFT) | ||
| 44 | /* lowest priority */ | ||
| 42 | 45 | ||
| 43 | #define MSI_ADDR_DEST_ID_SHIFT 12 | 46 | #define MSI_ADDR_DEST_ID_SHIFT 12 |
| 44 | #define MSI_ADDR_DEST_ID_MASK 0x00ffff0 | 47 | #define MSI_ADDR_DEST_ID_MASK 0x00ffff0 |
| 45 | #define MSI_ADDR_DEST_ID(dest) (((dest) << MSI_ADDR_DEST_ID_SHIFT) & MSI_ADDR_DEST_ID_MASK) | 48 | #define MSI_ADDR_DEST_ID(dest) (((dest) << MSI_ADDR_DEST_ID_SHIFT) & \ |
| 49 | MSI_ADDR_DEST_ID_MASK) | ||
| 46 | 50 | ||
| 47 | #endif /* ASM_MSIDEF_H */ | 51 | #endif /* ASM_MSIDEF_H */ |
diff --git a/include/asm-x86/msr-index.h b/include/asm-x86/msr-index.h index fae118a25278..09413ad39d3c 100644 --- a/include/asm-x86/msr-index.h +++ b/include/asm-x86/msr-index.h | |||
| @@ -57,6 +57,8 @@ | |||
| 57 | #define MSR_MTRRfix4K_F8000 0x0000026f | 57 | #define MSR_MTRRfix4K_F8000 0x0000026f |
| 58 | #define MSR_MTRRdefType 0x000002ff | 58 | #define MSR_MTRRdefType 0x000002ff |
| 59 | 59 | ||
| 60 | #define MSR_IA32_CR_PAT 0x00000277 | ||
| 61 | |||
| 60 | #define MSR_IA32_DEBUGCTLMSR 0x000001d9 | 62 | #define MSR_IA32_DEBUGCTLMSR 0x000001d9 |
| 61 | #define MSR_IA32_LASTBRANCHFROMIP 0x000001db | 63 | #define MSR_IA32_LASTBRANCHFROMIP 0x000001db |
| 62 | #define MSR_IA32_LASTBRANCHTOIP 0x000001dc | 64 | #define MSR_IA32_LASTBRANCHTOIP 0x000001dc |
| @@ -83,6 +85,7 @@ | |||
| 83 | /* AMD64 MSRs. Not complete. See the architecture manual for a more | 85 | /* AMD64 MSRs. Not complete. See the architecture manual for a more |
| 84 | complete list. */ | 86 | complete list. */ |
| 85 | 87 | ||
| 88 | #define MSR_AMD64_NB_CFG 0xc001001f | ||
| 86 | #define MSR_AMD64_IBSFETCHCTL 0xc0011030 | 89 | #define MSR_AMD64_IBSFETCHCTL 0xc0011030 |
| 87 | #define MSR_AMD64_IBSFETCHLINAD 0xc0011031 | 90 | #define MSR_AMD64_IBSFETCHLINAD 0xc0011031 |
| 88 | #define MSR_AMD64_IBSFETCHPHYSAD 0xc0011032 | 91 | #define MSR_AMD64_IBSFETCHPHYSAD 0xc0011032 |
| @@ -109,6 +112,7 @@ | |||
| 109 | #define MSR_K8_SYSCFG 0xc0010010 | 112 | #define MSR_K8_SYSCFG 0xc0010010 |
| 110 | #define MSR_K8_HWCR 0xc0010015 | 113 | #define MSR_K8_HWCR 0xc0010015 |
| 111 | #define MSR_K8_ENABLE_C1E 0xc0010055 | 114 | #define MSR_K8_ENABLE_C1E 0xc0010055 |
| 115 | #define MSR_K8_TSEG_ADDR 0xc0010112 | ||
| 112 | #define K8_MTRRFIXRANGE_DRAM_ENABLE 0x00040000 /* MtrrFixDramEn bit */ | 116 | #define K8_MTRRFIXRANGE_DRAM_ENABLE 0x00040000 /* MtrrFixDramEn bit */ |
| 113 | #define K8_MTRRFIXRANGE_DRAM_MODIFY 0x00080000 /* MtrrFixDramModEn bit */ | 117 | #define K8_MTRRFIXRANGE_DRAM_MODIFY 0x00080000 /* MtrrFixDramModEn bit */ |
| 114 | #define K8_MTRR_RDMEM_WRMEM_MASK 0x18181818 /* Mask: RdMem|WrMem */ | 118 | #define K8_MTRR_RDMEM_WRMEM_MASK 0x18181818 /* Mask: RdMem|WrMem */ |
diff --git a/include/asm-x86/msr.h b/include/asm-x86/msr.h index 3ca29ebebbb1..3707650a169b 100644 --- a/include/asm-x86/msr.h +++ b/include/asm-x86/msr.h | |||
| @@ -16,8 +16,8 @@ | |||
| 16 | static inline unsigned long long native_read_tscp(unsigned int *aux) | 16 | static inline unsigned long long native_read_tscp(unsigned int *aux) |
| 17 | { | 17 | { |
| 18 | unsigned long low, high; | 18 | unsigned long low, high; |
| 19 | asm volatile (".byte 0x0f,0x01,0xf9" | 19 | asm volatile(".byte 0x0f,0x01,0xf9" |
| 20 | : "=a" (low), "=d" (high), "=c" (*aux)); | 20 | : "=a" (low), "=d" (high), "=c" (*aux)); |
| 21 | return low | ((u64)high >> 32); | 21 | return low | ((u64)high >> 32); |
| 22 | } | 22 | } |
| 23 | 23 | ||
| @@ -29,7 +29,7 @@ static inline unsigned long long native_read_tscp(unsigned int *aux) | |||
| 29 | */ | 29 | */ |
| 30 | #ifdef CONFIG_X86_64 | 30 | #ifdef CONFIG_X86_64 |
| 31 | #define DECLARE_ARGS(val, low, high) unsigned low, high | 31 | #define DECLARE_ARGS(val, low, high) unsigned low, high |
| 32 | #define EAX_EDX_VAL(val, low, high) (low | ((u64)(high) << 32)) | 32 | #define EAX_EDX_VAL(val, low, high) ((low) | ((u64)(high) << 32)) |
| 33 | #define EAX_EDX_ARGS(val, low, high) "a" (low), "d" (high) | 33 | #define EAX_EDX_ARGS(val, low, high) "a" (low), "d" (high) |
| 34 | #define EAX_EDX_RET(val, low, high) "=a" (low), "=d" (high) | 34 | #define EAX_EDX_RET(val, low, high) "=a" (low), "=d" (high) |
| 35 | #else | 35 | #else |
| @@ -57,7 +57,7 @@ static inline unsigned long long native_read_msr_safe(unsigned int msr, | |||
| 57 | ".section .fixup,\"ax\"\n\t" | 57 | ".section .fixup,\"ax\"\n\t" |
| 58 | "3: mov %3,%0 ; jmp 1b\n\t" | 58 | "3: mov %3,%0 ; jmp 1b\n\t" |
| 59 | ".previous\n\t" | 59 | ".previous\n\t" |
| 60 | _ASM_EXTABLE(2b,3b) | 60 | _ASM_EXTABLE(2b, 3b) |
| 61 | : "=r" (*err), EAX_EDX_RET(val, low, high) | 61 | : "=r" (*err), EAX_EDX_RET(val, low, high) |
| 62 | : "c" (msr), "i" (-EFAULT)); | 62 | : "c" (msr), "i" (-EFAULT)); |
| 63 | return EAX_EDX_VAL(val, low, high); | 63 | return EAX_EDX_VAL(val, low, high); |
| @@ -78,10 +78,10 @@ static inline int native_write_msr_safe(unsigned int msr, | |||
| 78 | ".section .fixup,\"ax\"\n\t" | 78 | ".section .fixup,\"ax\"\n\t" |
| 79 | "3: mov %4,%0 ; jmp 1b\n\t" | 79 | "3: mov %4,%0 ; jmp 1b\n\t" |
| 80 | ".previous\n\t" | 80 | ".previous\n\t" |
| 81 | _ASM_EXTABLE(2b,3b) | 81 | _ASM_EXTABLE(2b, 3b) |
| 82 | : "=a" (err) | 82 | : "=a" (err) |
| 83 | : "c" (msr), "0" (low), "d" (high), | 83 | : "c" (msr), "0" (low), "d" (high), |
| 84 | "i" (-EFAULT)); | 84 | "i" (-EFAULT)); |
| 85 | return err; | 85 | return err; |
| 86 | } | 86 | } |
| 87 | 87 | ||
| @@ -116,23 +116,23 @@ static inline unsigned long long native_read_pmc(int counter) | |||
| 116 | * pointer indirection), this allows gcc to optimize better | 116 | * pointer indirection), this allows gcc to optimize better |
| 117 | */ | 117 | */ |
| 118 | 118 | ||
| 119 | #define rdmsr(msr,val1,val2) \ | 119 | #define rdmsr(msr, val1, val2) \ |
| 120 | do { \ | 120 | do { \ |
| 121 | u64 __val = native_read_msr(msr); \ | 121 | u64 __val = native_read_msr((msr)); \ |
| 122 | (val1) = (u32)__val; \ | 122 | (val1) = (u32)__val; \ |
| 123 | (val2) = (u32)(__val >> 32); \ | 123 | (val2) = (u32)(__val >> 32); \ |
| 124 | } while(0) | 124 | } while (0) |
| 125 | 125 | ||
| 126 | static inline void wrmsr(unsigned msr, unsigned low, unsigned high) | 126 | static inline void wrmsr(unsigned msr, unsigned low, unsigned high) |
| 127 | { | 127 | { |
| 128 | native_write_msr(msr, low, high); | 128 | native_write_msr(msr, low, high); |
| 129 | } | 129 | } |
| 130 | 130 | ||
| 131 | #define rdmsrl(msr,val) \ | 131 | #define rdmsrl(msr, val) \ |
| 132 | ((val) = native_read_msr(msr)) | 132 | ((val) = native_read_msr((msr))) |
| 133 | 133 | ||
| 134 | #define wrmsrl(msr, val) \ | 134 | #define wrmsrl(msr, val) \ |
| 135 | native_write_msr(msr, (u32)((u64)(val)), (u32)((u64)(val) >> 32)) | 135 | native_write_msr((msr), (u32)((u64)(val)), (u32)((u64)(val) >> 32)) |
| 136 | 136 | ||
| 137 | /* wrmsr with exception handling */ | 137 | /* wrmsr with exception handling */ |
| 138 | static inline int wrmsr_safe(unsigned msr, unsigned low, unsigned high) | 138 | static inline int wrmsr_safe(unsigned msr, unsigned low, unsigned high) |
| @@ -141,14 +141,22 @@ static inline int wrmsr_safe(unsigned msr, unsigned low, unsigned high) | |||
| 141 | } | 141 | } |
| 142 | 142 | ||
| 143 | /* rdmsr with exception handling */ | 143 | /* rdmsr with exception handling */ |
| 144 | #define rdmsr_safe(msr,p1,p2) \ | 144 | #define rdmsr_safe(msr, p1, p2) \ |
| 145 | ({ \ | 145 | ({ \ |
| 146 | int __err; \ | 146 | int __err; \ |
| 147 | u64 __val = native_read_msr_safe(msr, &__err); \ | 147 | u64 __val = native_read_msr_safe((msr), &__err); \ |
| 148 | (*p1) = (u32)__val; \ | 148 | (*p1) = (u32)__val; \ |
| 149 | (*p2) = (u32)(__val >> 32); \ | 149 | (*p2) = (u32)(__val >> 32); \ |
| 150 | __err; \ | 150 | __err; \ |
| 151 | }) | 151 | }) |
| 152 | |||
| 153 | static inline int rdmsrl_safe(unsigned msr, unsigned long long *p) | ||
| 154 | { | ||
| 155 | int err; | ||
| 156 | |||
| 157 | *p = native_read_msr_safe(msr, &err); | ||
| 158 | return err; | ||
| 159 | } | ||
| 152 | 160 | ||
| 153 | #define rdtscl(low) \ | 161 | #define rdtscl(low) \ |
| 154 | ((low) = (u32)native_read_tsc()) | 162 | ((low) = (u32)native_read_tsc()) |
| @@ -156,35 +164,37 @@ static inline int wrmsr_safe(unsigned msr, unsigned low, unsigned high) | |||
| 156 | #define rdtscll(val) \ | 164 | #define rdtscll(val) \ |
| 157 | ((val) = native_read_tsc()) | 165 | ((val) = native_read_tsc()) |
| 158 | 166 | ||
| 159 | #define rdpmc(counter,low,high) \ | 167 | #define rdpmc(counter, low, high) \ |
| 160 | do { \ | 168 | do { \ |
| 161 | u64 _l = native_read_pmc(counter); \ | 169 | u64 _l = native_read_pmc((counter)); \ |
| 162 | (low) = (u32)_l; \ | 170 | (low) = (u32)_l; \ |
| 163 | (high) = (u32)(_l >> 32); \ | 171 | (high) = (u32)(_l >> 32); \ |
| 164 | } while(0) | 172 | } while (0) |
| 165 | 173 | ||
| 166 | #define rdtscp(low, high, aux) \ | 174 | #define rdtscp(low, high, aux) \ |
| 167 | do { \ | 175 | do { \ |
| 168 | unsigned long long _val = native_read_tscp(&(aux)); \ | 176 | unsigned long long _val = native_read_tscp(&(aux)); \ |
| 169 | (low) = (u32)_val; \ | 177 | (low) = (u32)_val; \ |
| 170 | (high) = (u32)(_val >> 32); \ | 178 | (high) = (u32)(_val >> 32); \ |
| 171 | } while (0) | 179 | } while (0) |
| 172 | 180 | ||
| 173 | #define rdtscpll(val, aux) (val) = native_read_tscp(&(aux)) | 181 | #define rdtscpll(val, aux) (val) = native_read_tscp(&(aux)) |
| 174 | 182 | ||
| 175 | #endif /* !CONFIG_PARAVIRT */ | 183 | #endif /* !CONFIG_PARAVIRT */ |
| 176 | 184 | ||
| 177 | 185 | ||
| 178 | #define checking_wrmsrl(msr,val) wrmsr_safe(msr,(u32)(val),(u32)((val)>>32)) | 186 | #define checking_wrmsrl(msr, val) wrmsr_safe((msr), (u32)(val), \ |
| 187 | (u32)((val) >> 32)) | ||
| 179 | 188 | ||
| 180 | #define write_tsc(val1,val2) wrmsr(0x10, val1, val2) | 189 | #define write_tsc(val1, val2) wrmsr(0x10, (val1), (val2)) |
| 181 | 190 | ||
| 182 | #define write_rdtscp_aux(val) wrmsr(0xc0000103, val, 0) | 191 | #define write_rdtscp_aux(val) wrmsr(0xc0000103, (val), 0) |
| 183 | 192 | ||
| 184 | #ifdef CONFIG_SMP | 193 | #ifdef CONFIG_SMP |
| 185 | void rdmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h); | 194 | void rdmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h); |
| 186 | void wrmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h); | 195 | void wrmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h); |
| 187 | int rdmsr_safe_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h); | 196 | int rdmsr_safe_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h); |
| 197 | |||
| 188 | int wrmsr_safe_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h); | 198 | int wrmsr_safe_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h); |
| 189 | #else /* CONFIG_SMP */ | 199 | #else /* CONFIG_SMP */ |
| 190 | static inline void rdmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h) | 200 | static inline void rdmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h) |
| @@ -195,7 +205,8 @@ static inline void wrmsr_on_cpu(unsigned int cpu, u32 msr_no, u32 l, u32 h) | |||
| 195 | { | 205 | { |
| 196 | wrmsr(msr_no, l, h); | 206 | wrmsr(msr_no, l, h); |
| 197 | } | 207 | } |
| 198 | static inline int rdmsr_safe_on_cpu(unsigned int cpu, u32 msr_no, u32 *l, u32 *h) | 208 | static inline int rdmsr_safe_on_cpu(unsigned int cpu, u32 msr_no, |
| 209 | u32 *l, u32 *h) | ||
| 199 | { | 210 | { |
| 200 | return rdmsr_safe(msr_no, l, h); | 211 | return rdmsr_safe(msr_no, l, h); |
| 201 | } | 212 | } |
diff --git a/include/asm-x86/mtrr.h b/include/asm-x86/mtrr.h index 319d065800be..a69a01a51729 100644 --- a/include/asm-x86/mtrr.h +++ b/include/asm-x86/mtrr.h | |||
| @@ -28,8 +28,7 @@ | |||
| 28 | 28 | ||
| 29 | #define MTRR_IOCTL_BASE 'M' | 29 | #define MTRR_IOCTL_BASE 'M' |
| 30 | 30 | ||
| 31 | struct mtrr_sentry | 31 | struct mtrr_sentry { |
| 32 | { | ||
| 33 | unsigned long base; /* Base address */ | 32 | unsigned long base; /* Base address */ |
| 34 | unsigned int size; /* Size of region */ | 33 | unsigned int size; /* Size of region */ |
| 35 | unsigned int type; /* Type of region */ | 34 | unsigned int type; /* Type of region */ |
| @@ -41,8 +40,7 @@ struct mtrr_sentry | |||
| 41 | will break. */ | 40 | will break. */ |
| 42 | 41 | ||
| 43 | #ifdef __i386__ | 42 | #ifdef __i386__ |
| 44 | struct mtrr_gentry | 43 | struct mtrr_gentry { |
| 45 | { | ||
| 46 | unsigned int regnum; /* Register number */ | 44 | unsigned int regnum; /* Register number */ |
| 47 | unsigned long base; /* Base address */ | 45 | unsigned long base; /* Base address */ |
| 48 | unsigned int size; /* Size of region */ | 46 | unsigned int size; /* Size of region */ |
| @@ -51,8 +49,7 @@ struct mtrr_gentry | |||
| 51 | 49 | ||
| 52 | #else /* __i386__ */ | 50 | #else /* __i386__ */ |
| 53 | 51 | ||
| 54 | struct mtrr_gentry | 52 | struct mtrr_gentry { |
| 55 | { | ||
| 56 | unsigned long base; /* Base address */ | 53 | unsigned long base; /* Base address */ |
| 57 | unsigned int size; /* Size of region */ | 54 | unsigned int size; /* Size of region */ |
| 58 | unsigned int regnum; /* Register number */ | 55 | unsigned int regnum; /* Register number */ |
| @@ -86,38 +83,45 @@ struct mtrr_gentry | |||
| 86 | 83 | ||
| 87 | /* The following functions are for use by other drivers */ | 84 | /* The following functions are for use by other drivers */ |
| 88 | # ifdef CONFIG_MTRR | 85 | # ifdef CONFIG_MTRR |
| 86 | extern u8 mtrr_type_lookup(u64 addr, u64 end); | ||
| 89 | extern void mtrr_save_fixed_ranges(void *); | 87 | extern void mtrr_save_fixed_ranges(void *); |
| 90 | extern void mtrr_save_state(void); | 88 | extern void mtrr_save_state(void); |
| 91 | extern int mtrr_add (unsigned long base, unsigned long size, | 89 | extern int mtrr_add(unsigned long base, unsigned long size, |
| 92 | unsigned int type, bool increment); | 90 | unsigned int type, bool increment); |
| 93 | extern int mtrr_add_page (unsigned long base, unsigned long size, | 91 | extern int mtrr_add_page(unsigned long base, unsigned long size, |
| 94 | unsigned int type, bool increment); | 92 | unsigned int type, bool increment); |
| 95 | extern int mtrr_del (int reg, unsigned long base, unsigned long size); | 93 | extern int mtrr_del(int reg, unsigned long base, unsigned long size); |
| 96 | extern int mtrr_del_page (int reg, unsigned long base, unsigned long size); | 94 | extern int mtrr_del_page(int reg, unsigned long base, unsigned long size); |
| 97 | extern void mtrr_centaur_report_mcr(int mcr, u32 lo, u32 hi); | 95 | extern void mtrr_centaur_report_mcr(int mcr, u32 lo, u32 hi); |
| 98 | extern void mtrr_ap_init(void); | 96 | extern void mtrr_ap_init(void); |
| 99 | extern void mtrr_bp_init(void); | 97 | extern void mtrr_bp_init(void); |
| 100 | extern int mtrr_trim_uncached_memory(unsigned long end_pfn); | 98 | extern int mtrr_trim_uncached_memory(unsigned long end_pfn); |
| 99 | extern int amd_special_default_mtrr(void); | ||
| 101 | # else | 100 | # else |
| 101 | static inline u8 mtrr_type_lookup(u64 addr, u64 end) | ||
| 102 | { | ||
| 103 | /* | ||
| 104 | * Return no-MTRRs: | ||
| 105 | */ | ||
| 106 | return 0xff; | ||
| 107 | } | ||
| 102 | #define mtrr_save_fixed_ranges(arg) do {} while (0) | 108 | #define mtrr_save_fixed_ranges(arg) do {} while (0) |
| 103 | #define mtrr_save_state() do {} while (0) | 109 | #define mtrr_save_state() do {} while (0) |
| 104 | static __inline__ int mtrr_add (unsigned long base, unsigned long size, | 110 | static inline int mtrr_add(unsigned long base, unsigned long size, |
| 105 | unsigned int type, bool increment) | 111 | unsigned int type, bool increment) |
| 106 | { | 112 | { |
| 107 | return -ENODEV; | 113 | return -ENODEV; |
| 108 | } | 114 | } |
| 109 | static __inline__ int mtrr_add_page (unsigned long base, unsigned long size, | 115 | static inline int mtrr_add_page(unsigned long base, unsigned long size, |
| 110 | unsigned int type, bool increment) | 116 | unsigned int type, bool increment) |
| 111 | { | 117 | { |
| 112 | return -ENODEV; | 118 | return -ENODEV; |
| 113 | } | 119 | } |
| 114 | static __inline__ int mtrr_del (int reg, unsigned long base, | 120 | static inline int mtrr_del(int reg, unsigned long base, unsigned long size) |
| 115 | unsigned long size) | ||
| 116 | { | 121 | { |
| 117 | return -ENODEV; | 122 | return -ENODEV; |
| 118 | } | 123 | } |
| 119 | static __inline__ int mtrr_del_page (int reg, unsigned long base, | 124 | static inline int mtrr_del_page(int reg, unsigned long base, unsigned long size) |
| 120 | unsigned long size) | ||
| 121 | { | 125 | { |
| 122 | return -ENODEV; | 126 | return -ENODEV; |
| 123 | } | 127 | } |
| @@ -125,7 +129,9 @@ static inline int mtrr_trim_uncached_memory(unsigned long end_pfn) | |||
| 125 | { | 129 | { |
| 126 | return 0; | 130 | return 0; |
| 127 | } | 131 | } |
| 128 | static __inline__ void mtrr_centaur_report_mcr(int mcr, u32 lo, u32 hi) {;} | 132 | static inline void mtrr_centaur_report_mcr(int mcr, u32 lo, u32 hi) |
| 133 | { | ||
| 134 | } | ||
| 129 | 135 | ||
| 130 | #define mtrr_ap_init() do {} while (0) | 136 | #define mtrr_ap_init() do {} while (0) |
| 131 | #define mtrr_bp_init() do {} while (0) | 137 | #define mtrr_bp_init() do {} while (0) |
| @@ -134,15 +140,13 @@ static __inline__ void mtrr_centaur_report_mcr(int mcr, u32 lo, u32 hi) {;} | |||
| 134 | #ifdef CONFIG_COMPAT | 140 | #ifdef CONFIG_COMPAT |
| 135 | #include <linux/compat.h> | 141 | #include <linux/compat.h> |
| 136 | 142 | ||
| 137 | struct mtrr_sentry32 | 143 | struct mtrr_sentry32 { |
| 138 | { | ||
| 139 | compat_ulong_t base; /* Base address */ | 144 | compat_ulong_t base; /* Base address */ |
| 140 | compat_uint_t size; /* Size of region */ | 145 | compat_uint_t size; /* Size of region */ |
| 141 | compat_uint_t type; /* Type of region */ | 146 | compat_uint_t type; /* Type of region */ |
| 142 | }; | 147 | }; |
| 143 | 148 | ||
| 144 | struct mtrr_gentry32 | 149 | struct mtrr_gentry32 { |
| 145 | { | ||
| 146 | compat_ulong_t regnum; /* Register number */ | 150 | compat_ulong_t regnum; /* Register number */ |
| 147 | compat_uint_t base; /* Base address */ | 151 | compat_uint_t base; /* Base address */ |
| 148 | compat_uint_t size; /* Size of region */ | 152 | compat_uint_t size; /* Size of region */ |
| @@ -151,16 +155,17 @@ struct mtrr_gentry32 | |||
| 151 | 155 | ||
| 152 | #define MTRR_IOCTL_BASE 'M' | 156 | #define MTRR_IOCTL_BASE 'M' |
| 153 | 157 | ||
| 154 | #define MTRRIOC32_ADD_ENTRY _IOW(MTRR_IOCTL_BASE, 0, struct mtrr_sentry32) | 158 | #define MTRRIOC32_ADD_ENTRY _IOW(MTRR_IOCTL_BASE, 0, struct mtrr_sentry32) |
| 155 | #define MTRRIOC32_SET_ENTRY _IOW(MTRR_IOCTL_BASE, 1, struct mtrr_sentry32) | 159 | #define MTRRIOC32_SET_ENTRY _IOW(MTRR_IOCTL_BASE, 1, struct mtrr_sentry32) |
| 156 | #define MTRRIOC32_DEL_ENTRY _IOW(MTRR_IOCTL_BASE, 2, struct mtrr_sentry32) | 160 | #define MTRRIOC32_DEL_ENTRY _IOW(MTRR_IOCTL_BASE, 2, struct mtrr_sentry32) |
| 157 | #define MTRRIOC32_GET_ENTRY _IOWR(MTRR_IOCTL_BASE, 3, struct mtrr_gentry32) | 161 | #define MTRRIOC32_GET_ENTRY _IOWR(MTRR_IOCTL_BASE, 3, struct mtrr_gentry32) |
| 158 | #define MTRRIOC32_KILL_ENTRY _IOW(MTRR_IOCTL_BASE, 4, struct mtrr_sentry32) | 162 | #define MTRRIOC32_KILL_ENTRY _IOW(MTRR_IOCTL_BASE, 4, struct mtrr_sentry32) |
| 159 | #define MTRRIOC32_ADD_PAGE_ENTRY _IOW(MTRR_IOCTL_BASE, 5, struct mtrr_sentry32) | 163 | #define MTRRIOC32_ADD_PAGE_ENTRY _IOW(MTRR_IOCTL_BASE, 5, struct mtrr_sentry32) |
| 160 | #define MTRRIOC32_SET_PAGE_ENTRY _IOW(MTRR_IOCTL_BASE, 6, struct mtrr_sentry32) | 164 | #define MTRRIOC32_SET_PAGE_ENTRY _IOW(MTRR_IOCTL_BASE, 6, struct mtrr_sentry32) |
| 161 | #define MTRRIOC32_DEL_PAGE_ENTRY _IOW(MTRR_IOCTL_BASE, 7, struct mtrr_sentry32) | 165 | #define MTRRIOC32_DEL_PAGE_ENTRY _IOW(MTRR_IOCTL_BASE, 7, struct mtrr_sentry32) |
| 162 | #define MTRRIOC32_GET_PAGE_ENTRY _IOWR(MTRR_IOCTL_BASE, 8, struct mtrr_gentry32) | 166 | #define MTRRIOC32_GET_PAGE_ENTRY _IOWR(MTRR_IOCTL_BASE, 8, struct mtrr_gentry32) |
| 163 | #define MTRRIOC32_KILL_PAGE_ENTRY _IOW(MTRR_IOCTL_BASE, 9, struct mtrr_sentry32) | 167 | #define MTRRIOC32_KILL_PAGE_ENTRY \ |
| 168 | _IOW(MTRR_IOCTL_BASE, 9, struct mtrr_sentry32) | ||
| 164 | #endif /* CONFIG_COMPAT */ | 169 | #endif /* CONFIG_COMPAT */ |
| 165 | 170 | ||
| 166 | #endif /* __KERNEL__ */ | 171 | #endif /* __KERNEL__ */ |
diff --git a/include/asm-x86/mutex_32.h b/include/asm-x86/mutex_32.h index bbeefb96ddfd..73e928ef5f03 100644 --- a/include/asm-x86/mutex_32.h +++ b/include/asm-x86/mutex_32.h | |||
| @@ -9,7 +9,7 @@ | |||
| 9 | #ifndef _ASM_MUTEX_H | 9 | #ifndef _ASM_MUTEX_H |
| 10 | #define _ASM_MUTEX_H | 10 | #define _ASM_MUTEX_H |
| 11 | 11 | ||
| 12 | #include "asm/alternative.h" | 12 | #include <asm/alternative.h> |
| 13 | 13 | ||
| 14 | /** | 14 | /** |
| 15 | * __mutex_fastpath_lock - try to take the lock by moving the count | 15 | * __mutex_fastpath_lock - try to take the lock by moving the count |
| @@ -21,22 +21,20 @@ | |||
| 21 | * wasn't 1 originally. This function MUST leave the value lower than 1 | 21 | * wasn't 1 originally. This function MUST leave the value lower than 1 |
| 22 | * even when the "1" assertion wasn't true. | 22 | * even when the "1" assertion wasn't true. |
| 23 | */ | 23 | */ |
| 24 | #define __mutex_fastpath_lock(count, fail_fn) \ | 24 | #define __mutex_fastpath_lock(count, fail_fn) \ |
| 25 | do { \ | 25 | do { \ |
| 26 | unsigned int dummy; \ | 26 | unsigned int dummy; \ |
| 27 | \ | 27 | \ |
| 28 | typecheck(atomic_t *, count); \ | 28 | typecheck(atomic_t *, count); \ |
| 29 | typecheck_fn(void (*)(atomic_t *), fail_fn); \ | 29 | typecheck_fn(void (*)(atomic_t *), fail_fn); \ |
| 30 | \ | 30 | \ |
| 31 | __asm__ __volatile__( \ | 31 | asm volatile(LOCK_PREFIX " decl (%%eax)\n" \ |
| 32 | LOCK_PREFIX " decl (%%eax) \n" \ | 32 | " jns 1f \n" \ |
| 33 | " jns 1f \n" \ | 33 | " call " #fail_fn "\n" \ |
| 34 | " call "#fail_fn" \n" \ | 34 | "1:\n" \ |
| 35 | "1: \n" \ | 35 | : "=a" (dummy) \ |
| 36 | \ | 36 | : "a" (count) \ |
| 37 | :"=a" (dummy) \ | 37 | : "memory", "ecx", "edx"); \ |
| 38 | : "a" (count) \ | ||
| 39 | : "memory", "ecx", "edx"); \ | ||
| 40 | } while (0) | 38 | } while (0) |
| 41 | 39 | ||
| 42 | 40 | ||
| @@ -50,8 +48,8 @@ do { \ | |||
| 50 | * wasn't 1 originally. This function returns 0 if the fastpath succeeds, | 48 | * wasn't 1 originally. This function returns 0 if the fastpath succeeds, |
| 51 | * or anything the slow path function returns | 49 | * or anything the slow path function returns |
| 52 | */ | 50 | */ |
| 53 | static inline int | 51 | static inline int __mutex_fastpath_lock_retval(atomic_t *count, |
| 54 | __mutex_fastpath_lock_retval(atomic_t *count, int (*fail_fn)(atomic_t *)) | 52 | int (*fail_fn)(atomic_t *)) |
| 55 | { | 53 | { |
| 56 | if (unlikely(atomic_dec_return(count) < 0)) | 54 | if (unlikely(atomic_dec_return(count) < 0)) |
| 57 | return fail_fn(count); | 55 | return fail_fn(count); |
| @@ -72,22 +70,20 @@ __mutex_fastpath_lock_retval(atomic_t *count, int (*fail_fn)(atomic_t *)) | |||
| 72 | * __mutex_slowpath_needs_to_unlock() macro needs to return 1, it needs | 70 | * __mutex_slowpath_needs_to_unlock() macro needs to return 1, it needs |
| 73 | * to return 0 otherwise. | 71 | * to return 0 otherwise. |
| 74 | */ | 72 | */ |
| 75 | #define __mutex_fastpath_unlock(count, fail_fn) \ | 73 | #define __mutex_fastpath_unlock(count, fail_fn) \ |
| 76 | do { \ | 74 | do { \ |
| 77 | unsigned int dummy; \ | 75 | unsigned int dummy; \ |
| 78 | \ | 76 | \ |
| 79 | typecheck(atomic_t *, count); \ | 77 | typecheck(atomic_t *, count); \ |
| 80 | typecheck_fn(void (*)(atomic_t *), fail_fn); \ | 78 | typecheck_fn(void (*)(atomic_t *), fail_fn); \ |
| 81 | \ | 79 | \ |
| 82 | __asm__ __volatile__( \ | 80 | asm volatile(LOCK_PREFIX " incl (%%eax)\n" \ |
| 83 | LOCK_PREFIX " incl (%%eax) \n" \ | 81 | " jg 1f\n" \ |
| 84 | " jg 1f \n" \ | 82 | " call " #fail_fn "\n" \ |
| 85 | " call "#fail_fn" \n" \ | 83 | "1:\n" \ |
| 86 | "1: \n" \ | 84 | : "=a" (dummy) \ |
| 87 | \ | 85 | : "a" (count) \ |
| 88 | :"=a" (dummy) \ | 86 | : "memory", "ecx", "edx"); \ |
| 89 | : "a" (count) \ | ||
| 90 | : "memory", "ecx", "edx"); \ | ||
| 91 | } while (0) | 87 | } while (0) |
| 92 | 88 | ||
| 93 | #define __mutex_slowpath_needs_to_unlock() 1 | 89 | #define __mutex_slowpath_needs_to_unlock() 1 |
| @@ -104,8 +100,8 @@ do { \ | |||
| 104 | * Additionally, if the value was < 0 originally, this function must not leave | 100 | * Additionally, if the value was < 0 originally, this function must not leave |
| 105 | * it to 0 on failure. | 101 | * it to 0 on failure. |
| 106 | */ | 102 | */ |
| 107 | static inline int | 103 | static inline int __mutex_fastpath_trylock(atomic_t *count, |
| 108 | __mutex_fastpath_trylock(atomic_t *count, int (*fail_fn)(atomic_t *)) | 104 | int (*fail_fn)(atomic_t *)) |
| 109 | { | 105 | { |
| 110 | /* | 106 | /* |
| 111 | * We have two variants here. The cmpxchg based one is the best one | 107 | * We have two variants here. The cmpxchg based one is the best one |
diff --git a/include/asm-x86/mutex_64.h b/include/asm-x86/mutex_64.h index 6c2949a3c677..f3fae9becb38 100644 --- a/include/asm-x86/mutex_64.h +++ b/include/asm-x86/mutex_64.h | |||
| @@ -16,23 +16,21 @@ | |||
| 16 | * | 16 | * |
| 17 | * Atomically decrements @v and calls <fail_fn> if the result is negative. | 17 | * Atomically decrements @v and calls <fail_fn> if the result is negative. |
| 18 | */ | 18 | */ |
| 19 | #define __mutex_fastpath_lock(v, fail_fn) \ | 19 | #define __mutex_fastpath_lock(v, fail_fn) \ |
| 20 | do { \ | 20 | do { \ |
| 21 | unsigned long dummy; \ | 21 | unsigned long dummy; \ |
| 22 | \ | 22 | \ |
| 23 | typecheck(atomic_t *, v); \ | 23 | typecheck(atomic_t *, v); \ |
| 24 | typecheck_fn(void (*)(atomic_t *), fail_fn); \ | 24 | typecheck_fn(void (*)(atomic_t *), fail_fn); \ |
| 25 | \ | 25 | \ |
| 26 | __asm__ __volatile__( \ | 26 | asm volatile(LOCK_PREFIX " decl (%%rdi)\n" \ |
| 27 | LOCK_PREFIX " decl (%%rdi) \n" \ | 27 | " jns 1f \n" \ |
| 28 | " jns 1f \n" \ | 28 | " call " #fail_fn "\n" \ |
| 29 | " call "#fail_fn" \n" \ | 29 | "1:" \ |
| 30 | "1:" \ | 30 | : "=D" (dummy) \ |
| 31 | \ | 31 | : "D" (v) \ |
| 32 | :"=D" (dummy) \ | 32 | : "rax", "rsi", "rdx", "rcx", \ |
| 33 | : "D" (v) \ | 33 | "r8", "r9", "r10", "r11", "memory"); \ |
| 34 | : "rax", "rsi", "rdx", "rcx", \ | ||
| 35 | "r8", "r9", "r10", "r11", "memory"); \ | ||
| 36 | } while (0) | 34 | } while (0) |
| 37 | 35 | ||
| 38 | /** | 36 | /** |
| @@ -45,9 +43,8 @@ do { \ | |||
| 45 | * it wasn't 1 originally. This function returns 0 if the fastpath succeeds, | 43 | * it wasn't 1 originally. This function returns 0 if the fastpath succeeds, |
| 46 | * or anything the slow path function returns | 44 | * or anything the slow path function returns |
| 47 | */ | 45 | */ |
| 48 | static inline int | 46 | static inline int __mutex_fastpath_lock_retval(atomic_t *count, |
| 49 | __mutex_fastpath_lock_retval(atomic_t *count, | 47 | int (*fail_fn)(atomic_t *)) |
| 50 | int (*fail_fn)(atomic_t *)) | ||
| 51 | { | 48 | { |
| 52 | if (unlikely(atomic_dec_return(count) < 0)) | 49 | if (unlikely(atomic_dec_return(count) < 0)) |
| 53 | return fail_fn(count); | 50 | return fail_fn(count); |
| @@ -62,23 +59,21 @@ __mutex_fastpath_lock_retval(atomic_t *count, | |||
| 62 | * | 59 | * |
| 63 | * Atomically increments @v and calls <fail_fn> if the result is nonpositive. | 60 | * Atomically increments @v and calls <fail_fn> if the result is nonpositive. |
| 64 | */ | 61 | */ |
| 65 | #define __mutex_fastpath_unlock(v, fail_fn) \ | 62 | #define __mutex_fastpath_unlock(v, fail_fn) \ |
| 66 | do { \ | 63 | do { \ |
| 67 | unsigned long dummy; \ | 64 | unsigned long dummy; \ |
| 68 | \ | 65 | \ |
| 69 | typecheck(atomic_t *, v); \ | 66 | typecheck(atomic_t *, v); \ |
| 70 | typecheck_fn(void (*)(atomic_t *), fail_fn); \ | 67 | typecheck_fn(void (*)(atomic_t *), fail_fn); \ |
| 71 | \ | 68 | \ |
| 72 | __asm__ __volatile__( \ | 69 | asm volatile(LOCK_PREFIX " incl (%%rdi)\n" \ |
| 73 | LOCK_PREFIX " incl (%%rdi) \n" \ | 70 | " jg 1f\n" \ |
| 74 | " jg 1f \n" \ | 71 | " call " #fail_fn "\n" \ |
| 75 | " call "#fail_fn" \n" \ | 72 | "1:" \ |
| 76 | "1: " \ | 73 | : "=D" (dummy) \ |
| 77 | \ | 74 | : "D" (v) \ |
| 78 | :"=D" (dummy) \ | 75 | : "rax", "rsi", "rdx", "rcx", \ |
| 79 | : "D" (v) \ | 76 | "r8", "r9", "r10", "r11", "memory"); \ |
| 80 | : "rax", "rsi", "rdx", "rcx", \ | ||
| 81 | "r8", "r9", "r10", "r11", "memory"); \ | ||
| 82 | } while (0) | 77 | } while (0) |
| 83 | 78 | ||
| 84 | #define __mutex_slowpath_needs_to_unlock() 1 | 79 | #define __mutex_slowpath_needs_to_unlock() 1 |
| @@ -93,8 +88,8 @@ do { \ | |||
| 93 | * if it wasn't 1 originally. [the fallback function is never used on | 88 | * if it wasn't 1 originally. [the fallback function is never used on |
| 94 | * x86_64, because all x86_64 CPUs have a CMPXCHG instruction.] | 89 | * x86_64, because all x86_64 CPUs have a CMPXCHG instruction.] |
| 95 | */ | 90 | */ |
| 96 | static inline int | 91 | static inline int __mutex_fastpath_trylock(atomic_t *count, |
| 97 | __mutex_fastpath_trylock(atomic_t *count, int (*fail_fn)(atomic_t *)) | 92 | int (*fail_fn)(atomic_t *)) |
| 98 | { | 93 | { |
| 99 | if (likely(atomic_cmpxchg(count, 1, 0) == 1)) | 94 | if (likely(atomic_cmpxchg(count, 1, 0) == 1)) |
| 100 | return 1; | 95 | return 1; |
diff --git a/include/asm-x86/nmi.h b/include/asm-x86/nmi.h index 53ccac14cead..1e363021e72f 100644 --- a/include/asm-x86/nmi.h +++ b/include/asm-x86/nmi.h | |||
| @@ -1,5 +1,93 @@ | |||
| 1 | #ifdef CONFIG_X86_32 | 1 | #ifndef _ASM_X86_NMI_H_ |
| 2 | # include "nmi_32.h" | 2 | #define _ASM_X86_NMI_H_ |
| 3 | |||
| 4 | #include <linux/pm.h> | ||
| 5 | #include <asm/irq.h> | ||
| 6 | #include <asm/io.h> | ||
| 7 | |||
| 8 | #ifdef ARCH_HAS_NMI_WATCHDOG | ||
| 9 | |||
| 10 | /** | ||
| 11 | * do_nmi_callback | ||
| 12 | * | ||
| 13 | * Check to see if a callback exists and execute it. Return 1 | ||
| 14 | * if the handler exists and was handled successfully. | ||
| 15 | */ | ||
| 16 | int do_nmi_callback(struct pt_regs *regs, int cpu); | ||
| 17 | |||
| 18 | #ifdef CONFIG_PM | ||
| 19 | |||
| 20 | /** Replace the PM callback routine for NMI. */ | ||
| 21 | struct pm_dev *set_nmi_pm_callback(pm_callback callback); | ||
| 22 | |||
| 23 | /** Unset the PM callback routine back to the default. */ | ||
| 24 | void unset_nmi_pm_callback(struct pm_dev *dev); | ||
| 25 | |||
| 3 | #else | 26 | #else |
| 4 | # include "nmi_64.h" | 27 | |
| 28 | static inline struct pm_dev *set_nmi_pm_callback(pm_callback callback) | ||
| 29 | { | ||
| 30 | return 0; | ||
| 31 | } | ||
| 32 | |||
| 33 | static inline void unset_nmi_pm_callback(struct pm_dev *dev) | ||
| 34 | { | ||
| 35 | } | ||
| 36 | |||
| 37 | #endif /* CONFIG_PM */ | ||
| 38 | |||
| 39 | #ifdef CONFIG_X86_64 | ||
| 40 | extern void default_do_nmi(struct pt_regs *); | ||
| 41 | extern void die_nmi(char *str, struct pt_regs *regs, int do_panic); | ||
| 42 | extern void nmi_watchdog_default(void); | ||
| 43 | #else | ||
| 44 | #define nmi_watchdog_default() do {} while (0) | ||
| 45 | #endif | ||
| 46 | |||
| 47 | extern int check_nmi_watchdog(void); | ||
| 48 | extern int nmi_watchdog_enabled; | ||
| 49 | extern int unknown_nmi_panic; | ||
| 50 | extern int avail_to_resrv_perfctr_nmi_bit(unsigned int); | ||
| 51 | extern int avail_to_resrv_perfctr_nmi(unsigned int); | ||
| 52 | extern int reserve_perfctr_nmi(unsigned int); | ||
| 53 | extern void release_perfctr_nmi(unsigned int); | ||
| 54 | extern int reserve_evntsel_nmi(unsigned int); | ||
| 55 | extern void release_evntsel_nmi(unsigned int); | ||
| 56 | |||
| 57 | extern void setup_apic_nmi_watchdog(void *); | ||
| 58 | extern void stop_apic_nmi_watchdog(void *); | ||
| 59 | extern void disable_timer_nmi_watchdog(void); | ||
| 60 | extern void enable_timer_nmi_watchdog(void); | ||
| 61 | extern int nmi_watchdog_tick(struct pt_regs *regs, unsigned reason); | ||
| 62 | |||
| 63 | extern atomic_t nmi_active; | ||
| 64 | extern unsigned int nmi_watchdog; | ||
| 65 | #define NMI_DISABLED -1 | ||
| 66 | #define NMI_NONE 0 | ||
| 67 | #define NMI_IO_APIC 1 | ||
| 68 | #define NMI_LOCAL_APIC 2 | ||
| 69 | #define NMI_INVALID 3 | ||
| 70 | #define NMI_DEFAULT NMI_DISABLED | ||
| 71 | |||
| 72 | struct ctl_table; | ||
| 73 | struct file; | ||
| 74 | extern int proc_nmi_enabled(struct ctl_table *, int , struct file *, | ||
| 75 | void __user *, size_t *, loff_t *); | ||
| 76 | extern int unknown_nmi_panic; | ||
| 77 | |||
| 78 | void __trigger_all_cpu_backtrace(void); | ||
| 79 | #define trigger_all_cpu_backtrace() __trigger_all_cpu_backtrace() | ||
| 80 | |||
| 81 | #endif | ||
| 82 | |||
| 83 | void lapic_watchdog_stop(void); | ||
| 84 | int lapic_watchdog_init(unsigned nmi_hz); | ||
| 85 | int lapic_wd_event(unsigned nmi_hz); | ||
| 86 | unsigned lapic_adjust_nmi_hz(unsigned hz); | ||
| 87 | int lapic_watchdog_ok(void); | ||
| 88 | void disable_lapic_nmi_watchdog(void); | ||
| 89 | void enable_lapic_nmi_watchdog(void); | ||
| 90 | void stop_nmi(void); | ||
| 91 | void restart_nmi(void); | ||
| 92 | |||
| 5 | #endif | 93 | #endif |
diff --git a/include/asm-x86/nmi_32.h b/include/asm-x86/nmi_32.h deleted file mode 100644 index 7206c7e8a388..000000000000 --- a/include/asm-x86/nmi_32.h +++ /dev/null | |||
| @@ -1,61 +0,0 @@ | |||
| 1 | #ifndef ASM_NMI_H | ||
| 2 | #define ASM_NMI_H | ||
| 3 | |||
| 4 | #include <linux/pm.h> | ||
| 5 | #include <asm/irq.h> | ||
| 6 | |||
| 7 | #ifdef ARCH_HAS_NMI_WATCHDOG | ||
| 8 | |||
| 9 | /** | ||
| 10 | * do_nmi_callback | ||
| 11 | * | ||
| 12 | * Check to see if a callback exists and execute it. Return 1 | ||
| 13 | * if the handler exists and was handled successfully. | ||
| 14 | */ | ||
| 15 | int do_nmi_callback(struct pt_regs *regs, int cpu); | ||
| 16 | |||
| 17 | extern int nmi_watchdog_enabled; | ||
| 18 | extern int avail_to_resrv_perfctr_nmi_bit(unsigned int); | ||
| 19 | extern int avail_to_resrv_perfctr_nmi(unsigned int); | ||
| 20 | extern int reserve_perfctr_nmi(unsigned int); | ||
| 21 | extern void release_perfctr_nmi(unsigned int); | ||
| 22 | extern int reserve_evntsel_nmi(unsigned int); | ||
| 23 | extern void release_evntsel_nmi(unsigned int); | ||
| 24 | |||
| 25 | extern void setup_apic_nmi_watchdog (void *); | ||
| 26 | extern void stop_apic_nmi_watchdog (void *); | ||
| 27 | extern void disable_timer_nmi_watchdog(void); | ||
| 28 | extern void enable_timer_nmi_watchdog(void); | ||
| 29 | extern int nmi_watchdog_tick (struct pt_regs * regs, unsigned reason); | ||
| 30 | |||
| 31 | extern atomic_t nmi_active; | ||
| 32 | extern unsigned int nmi_watchdog; | ||
| 33 | #define NMI_DISABLED -1 | ||
| 34 | #define NMI_NONE 0 | ||
| 35 | #define NMI_IO_APIC 1 | ||
| 36 | #define NMI_LOCAL_APIC 2 | ||
| 37 | #define NMI_INVALID 3 | ||
| 38 | #define NMI_DEFAULT NMI_DISABLED | ||
| 39 | |||
| 40 | struct ctl_table; | ||
| 41 | struct file; | ||
| 42 | extern int proc_nmi_enabled(struct ctl_table *, int , struct file *, | ||
| 43 | void __user *, size_t *, loff_t *); | ||
| 44 | extern int unknown_nmi_panic; | ||
| 45 | |||
| 46 | void __trigger_all_cpu_backtrace(void); | ||
| 47 | #define trigger_all_cpu_backtrace() __trigger_all_cpu_backtrace() | ||
| 48 | |||
| 49 | #endif | ||
| 50 | |||
| 51 | void lapic_watchdog_stop(void); | ||
| 52 | int lapic_watchdog_init(unsigned nmi_hz); | ||
| 53 | int lapic_wd_event(unsigned nmi_hz); | ||
| 54 | unsigned lapic_adjust_nmi_hz(unsigned hz); | ||
| 55 | int lapic_watchdog_ok(void); | ||
| 56 | void disable_lapic_nmi_watchdog(void); | ||
| 57 | void enable_lapic_nmi_watchdog(void); | ||
| 58 | void stop_nmi(void); | ||
| 59 | void restart_nmi(void); | ||
| 60 | |||
| 61 | #endif /* ASM_NMI_H */ | ||
diff --git a/include/asm-x86/nmi_64.h b/include/asm-x86/nmi_64.h deleted file mode 100644 index 2eeb74e5f3ff..000000000000 --- a/include/asm-x86/nmi_64.h +++ /dev/null | |||
| @@ -1,90 +0,0 @@ | |||
| 1 | #ifndef ASM_NMI_H | ||
| 2 | #define ASM_NMI_H | ||
| 3 | |||
| 4 | #include <linux/pm.h> | ||
| 5 | #include <asm/io.h> | ||
| 6 | |||
| 7 | /** | ||
| 8 | * do_nmi_callback | ||
| 9 | * | ||
| 10 | * Check to see if a callback exists and execute it. Return 1 | ||
| 11 | * if the handler exists and was handled successfully. | ||
| 12 | */ | ||
| 13 | int do_nmi_callback(struct pt_regs *regs, int cpu); | ||
| 14 | |||
| 15 | #ifdef CONFIG_PM | ||
| 16 | |||
| 17 | /** Replace the PM callback routine for NMI. */ | ||
| 18 | struct pm_dev * set_nmi_pm_callback(pm_callback callback); | ||
| 19 | |||
| 20 | /** Unset the PM callback routine back to the default. */ | ||
| 21 | void unset_nmi_pm_callback(struct pm_dev * dev); | ||
| 22 | |||
| 23 | #else | ||
| 24 | |||
| 25 | static inline struct pm_dev * set_nmi_pm_callback(pm_callback callback) | ||
| 26 | { | ||
| 27 | return 0; | ||
| 28 | } | ||
| 29 | |||
| 30 | static inline void unset_nmi_pm_callback(struct pm_dev * dev) | ||
| 31 | { | ||
| 32 | } | ||
| 33 | |||
| 34 | #endif /* CONFIG_PM */ | ||
| 35 | |||
| 36 | extern void default_do_nmi(struct pt_regs *); | ||
| 37 | extern void die_nmi(char *str, struct pt_regs *regs, int do_panic); | ||
| 38 | |||
| 39 | #define get_nmi_reason() inb(0x61) | ||
| 40 | |||
| 41 | extern int unknown_nmi_panic; | ||
| 42 | extern int nmi_watchdog_enabled; | ||
| 43 | |||
| 44 | extern int check_nmi_watchdog(void); | ||
| 45 | extern int avail_to_resrv_perfctr_nmi_bit(unsigned int); | ||
| 46 | extern int avail_to_resrv_perfctr_nmi(unsigned int); | ||
| 47 | extern int reserve_perfctr_nmi(unsigned int); | ||
| 48 | extern void release_perfctr_nmi(unsigned int); | ||
| 49 | extern int reserve_evntsel_nmi(unsigned int); | ||
| 50 | extern void release_evntsel_nmi(unsigned int); | ||
| 51 | |||
| 52 | extern void setup_apic_nmi_watchdog (void *); | ||
| 53 | extern void stop_apic_nmi_watchdog (void *); | ||
| 54 | extern void disable_timer_nmi_watchdog(void); | ||
| 55 | extern void enable_timer_nmi_watchdog(void); | ||
| 56 | extern int nmi_watchdog_tick (struct pt_regs * regs, unsigned reason); | ||
| 57 | |||
| 58 | extern void nmi_watchdog_default(void); | ||
| 59 | |||
| 60 | extern atomic_t nmi_active; | ||
| 61 | extern unsigned int nmi_watchdog; | ||
| 62 | #define NMI_DISABLED -1 | ||
| 63 | #define NMI_NONE 0 | ||
| 64 | #define NMI_IO_APIC 1 | ||
| 65 | #define NMI_LOCAL_APIC 2 | ||
| 66 | #define NMI_INVALID 3 | ||
| 67 | #define NMI_DEFAULT NMI_DISABLED | ||
| 68 | |||
| 69 | struct ctl_table; | ||
| 70 | struct file; | ||
| 71 | extern int proc_nmi_enabled(struct ctl_table *, int , struct file *, | ||
| 72 | void __user *, size_t *, loff_t *); | ||
| 73 | |||
| 74 | extern int unknown_nmi_panic; | ||
| 75 | |||
| 76 | void __trigger_all_cpu_backtrace(void); | ||
| 77 | #define trigger_all_cpu_backtrace() __trigger_all_cpu_backtrace() | ||
| 78 | |||
| 79 | |||
| 80 | void lapic_watchdog_stop(void); | ||
| 81 | int lapic_watchdog_init(unsigned nmi_hz); | ||
| 82 | int lapic_wd_event(unsigned nmi_hz); | ||
| 83 | unsigned lapic_adjust_nmi_hz(unsigned hz); | ||
| 84 | int lapic_watchdog_ok(void); | ||
| 85 | void disable_lapic_nmi_watchdog(void); | ||
| 86 | void enable_lapic_nmi_watchdog(void); | ||
| 87 | void stop_nmi(void); | ||
| 88 | void restart_nmi(void); | ||
| 89 | |||
| 90 | #endif /* ASM_NMI_H */ | ||
diff --git a/include/asm-x86/nops.h b/include/asm-x86/nops.h index b3930ae539b3..ad0bedd10b89 100644 --- a/include/asm-x86/nops.h +++ b/include/asm-x86/nops.h | |||
| @@ -5,6 +5,8 @@ | |||
| 5 | 5 | ||
| 6 | /* generic versions from gas | 6 | /* generic versions from gas |
| 7 | 1: nop | 7 | 1: nop |
| 8 | the following instructions are NOT nops in 64-bit mode, | ||
| 9 | for 64-bit mode use K8 or P6 nops instead | ||
| 8 | 2: movl %esi,%esi | 10 | 2: movl %esi,%esi |
| 9 | 3: leal 0x00(%esi),%esi | 11 | 3: leal 0x00(%esi),%esi |
| 10 | 4: leal 0x00(,%esi,1),%esi | 12 | 4: leal 0x00(,%esi,1),%esi |
diff --git a/include/asm-x86/numa_64.h b/include/asm-x86/numa_64.h index 15fe07cde586..32c22ae0709f 100644 --- a/include/asm-x86/numa_64.h +++ b/include/asm-x86/numa_64.h | |||
| @@ -1,11 +1,12 @@ | |||
| 1 | #ifndef _ASM_X8664_NUMA_H | 1 | #ifndef _ASM_X8664_NUMA_H |
| 2 | #define _ASM_X8664_NUMA_H 1 | 2 | #define _ASM_X8664_NUMA_H 1 |
| 3 | 3 | ||
| 4 | #include <linux/nodemask.h> | 4 | #include <linux/nodemask.h> |
| 5 | #include <asm/apicdef.h> | 5 | #include <asm/apicdef.h> |
| 6 | 6 | ||
| 7 | struct bootnode { | 7 | struct bootnode { |
| 8 | u64 start,end; | 8 | u64 start; |
| 9 | u64 end; | ||
| 9 | }; | 10 | }; |
| 10 | 11 | ||
| 11 | extern int compute_hash_shift(struct bootnode *nodes, int numnodes); | 12 | extern int compute_hash_shift(struct bootnode *nodes, int numnodes); |
diff --git a/include/asm-x86/numaq.h b/include/asm-x86/numaq.h index 38f710dc37f2..94b86c31239a 100644 --- a/include/asm-x86/numaq.h +++ b/include/asm-x86/numaq.h | |||
| @@ -3,7 +3,7 @@ | |||
| 3 | * | 3 | * |
| 4 | * Copyright (C) 2002, IBM Corp. | 4 | * Copyright (C) 2002, IBM Corp. |
| 5 | * | 5 | * |
| 6 | * All rights reserved. | 6 | * All rights reserved. |
| 7 | * | 7 | * |
| 8 | * This program is free software; you can redistribute it and/or modify | 8 | * This program is free software; you can redistribute it and/or modify |
| 9 | * it under the terms of the GNU General Public License as published by | 9 | * it under the terms of the GNU General Public License as published by |
| @@ -33,7 +33,8 @@ extern int get_memcfg_numaq(void); | |||
| 33 | /* | 33 | /* |
| 34 | * SYS_CFG_DATA_PRIV_ADDR, struct eachquadmem, and struct sys_cfg_data are the | 34 | * SYS_CFG_DATA_PRIV_ADDR, struct eachquadmem, and struct sys_cfg_data are the |
| 35 | */ | 35 | */ |
| 36 | #define SYS_CFG_DATA_PRIV_ADDR 0x0009d000 /* place for scd in private quad space */ | 36 | #define SYS_CFG_DATA_PRIV_ADDR 0x0009d000 /* place for scd in private |
| 37 | quad space */ | ||
| 37 | 38 | ||
| 38 | /* | 39 | /* |
| 39 | * Communication area for each processor on lynxer-processor tests. | 40 | * Communication area for each processor on lynxer-processor tests. |
| @@ -139,7 +140,7 @@ struct sys_cfg_data { | |||
| 139 | unsigned int low_shrd_mem_base; /* 0 or 512MB or 1GB */ | 140 | unsigned int low_shrd_mem_base; /* 0 or 512MB or 1GB */ |
| 140 | unsigned int low_shrd_mem_quad_offset; /* 0,128M,256M,512M,1G */ | 141 | unsigned int low_shrd_mem_quad_offset; /* 0,128M,256M,512M,1G */ |
| 141 | /* may not be totally populated */ | 142 | /* may not be totally populated */ |
| 142 | unsigned int split_mem_enbl; /* 0 for no low shared memory */ | 143 | unsigned int split_mem_enbl; /* 0 for no low shared memory */ |
| 143 | unsigned int mmio_sz; /* Size of total system memory mapped I/O */ | 144 | unsigned int mmio_sz; /* Size of total system memory mapped I/O */ |
| 144 | /* (in MB). */ | 145 | /* (in MB). */ |
| 145 | unsigned int quad_spin_lock; /* Spare location used for quad */ | 146 | unsigned int quad_spin_lock; /* Spare location used for quad */ |
| @@ -152,7 +153,7 @@ struct sys_cfg_data { | |||
| 152 | /* | 153 | /* |
| 153 | * memory configuration area for each quad | 154 | * memory configuration area for each quad |
| 154 | */ | 155 | */ |
| 155 | struct eachquadmem eq[MAX_NUMNODES]; /* indexed by quad id */ | 156 | struct eachquadmem eq[MAX_NUMNODES]; /* indexed by quad id */ |
| 156 | }; | 157 | }; |
| 157 | 158 | ||
| 158 | static inline unsigned long *get_zholes_size(int nid) | 159 | static inline unsigned long *get_zholes_size(int nid) |
diff --git a/include/asm-x86/page.h b/include/asm-x86/page.h index a05b2896492f..6724a4bc6b7a 100644 --- a/include/asm-x86/page.h +++ b/include/asm-x86/page.h | |||
| @@ -33,10 +33,8 @@ | |||
| 33 | 33 | ||
| 34 | #ifdef CONFIG_X86_64 | 34 | #ifdef CONFIG_X86_64 |
| 35 | #include <asm/page_64.h> | 35 | #include <asm/page_64.h> |
| 36 | #define max_pfn_mapped end_pfn_map | ||
| 37 | #else | 36 | #else |
| 38 | #include <asm/page_32.h> | 37 | #include <asm/page_32.h> |
| 39 | #define max_pfn_mapped max_low_pfn | ||
| 40 | #endif /* CONFIG_X86_64 */ | 38 | #endif /* CONFIG_X86_64 */ |
| 41 | 39 | ||
| 42 | #define PAGE_OFFSET ((unsigned long)__PAGE_OFFSET) | 40 | #define PAGE_OFFSET ((unsigned long)__PAGE_OFFSET) |
| @@ -50,6 +48,8 @@ | |||
| 50 | 48 | ||
| 51 | extern int page_is_ram(unsigned long pagenr); | 49 | extern int page_is_ram(unsigned long pagenr); |
| 52 | 50 | ||
| 51 | extern unsigned long max_pfn_mapped; | ||
| 52 | |||
| 53 | struct page; | 53 | struct page; |
| 54 | 54 | ||
| 55 | static inline void clear_user_page(void *page, unsigned long vaddr, | 55 | static inline void clear_user_page(void *page, unsigned long vaddr, |
diff --git a/include/asm-x86/page_32.h b/include/asm-x86/page_32.h index 5f7257fd589b..424e82f8ae27 100644 --- a/include/asm-x86/page_32.h +++ b/include/asm-x86/page_32.h | |||
| @@ -47,7 +47,10 @@ typedef unsigned long pgdval_t; | |||
| 47 | typedef unsigned long pgprotval_t; | 47 | typedef unsigned long pgprotval_t; |
| 48 | typedef unsigned long phys_addr_t; | 48 | typedef unsigned long phys_addr_t; |
| 49 | 49 | ||
| 50 | typedef union { pteval_t pte, pte_low; } pte_t; | 50 | typedef union { |
| 51 | pteval_t pte; | ||
| 52 | pteval_t pte_low; | ||
| 53 | } pte_t; | ||
| 51 | 54 | ||
| 52 | #endif /* __ASSEMBLY__ */ | 55 | #endif /* __ASSEMBLY__ */ |
| 53 | #endif /* CONFIG_X86_PAE */ | 56 | #endif /* CONFIG_X86_PAE */ |
| @@ -61,7 +64,7 @@ typedef struct page *pgtable_t; | |||
| 61 | #endif | 64 | #endif |
| 62 | 65 | ||
| 63 | #ifndef __ASSEMBLY__ | 66 | #ifndef __ASSEMBLY__ |
| 64 | #define __phys_addr(x) ((x)-PAGE_OFFSET) | 67 | #define __phys_addr(x) ((x) - PAGE_OFFSET) |
| 65 | #define __phys_reloc_hide(x) RELOC_HIDE((x), 0) | 68 | #define __phys_reloc_hide(x) RELOC_HIDE((x), 0) |
| 66 | 69 | ||
| 67 | #ifdef CONFIG_FLATMEM | 70 | #ifdef CONFIG_FLATMEM |
| @@ -78,7 +81,7 @@ extern unsigned int __VMALLOC_RESERVE; | |||
| 78 | extern int sysctl_legacy_va_layout; | 81 | extern int sysctl_legacy_va_layout; |
| 79 | 82 | ||
| 80 | #define VMALLOC_RESERVE ((unsigned long)__VMALLOC_RESERVE) | 83 | #define VMALLOC_RESERVE ((unsigned long)__VMALLOC_RESERVE) |
| 81 | #define MAXMEM (-__PAGE_OFFSET-__VMALLOC_RESERVE) | 84 | #define MAXMEM (-__PAGE_OFFSET - __VMALLOC_RESERVE) |
| 82 | 85 | ||
| 83 | #ifdef CONFIG_X86_USE_3DNOW | 86 | #ifdef CONFIG_X86_USE_3DNOW |
| 84 | #include <asm/mmx.h> | 87 | #include <asm/mmx.h> |
diff --git a/include/asm-x86/page_64.h b/include/asm-x86/page_64.h index 143546073b95..6ea72859c491 100644 --- a/include/asm-x86/page_64.h +++ b/include/asm-x86/page_64.h | |||
| @@ -5,7 +5,7 @@ | |||
| 5 | 5 | ||
| 6 | #define THREAD_ORDER 1 | 6 | #define THREAD_ORDER 1 |
| 7 | #define THREAD_SIZE (PAGE_SIZE << THREAD_ORDER) | 7 | #define THREAD_SIZE (PAGE_SIZE << THREAD_ORDER) |
| 8 | #define CURRENT_MASK (~(THREAD_SIZE-1)) | 8 | #define CURRENT_MASK (~(THREAD_SIZE - 1)) |
| 9 | 9 | ||
| 10 | #define EXCEPTION_STACK_ORDER 0 | 10 | #define EXCEPTION_STACK_ORDER 0 |
| 11 | #define EXCEPTION_STKSZ (PAGE_SIZE << EXCEPTION_STACK_ORDER) | 11 | #define EXCEPTION_STKSZ (PAGE_SIZE << EXCEPTION_STACK_ORDER) |
| @@ -48,10 +48,10 @@ | |||
| 48 | #define __VIRTUAL_MASK_SHIFT 48 | 48 | #define __VIRTUAL_MASK_SHIFT 48 |
| 49 | 49 | ||
| 50 | /* | 50 | /* |
| 51 | * Kernel image size is limited to 128 MB (see level2_kernel_pgt in | 51 | * Kernel image size is limited to 512 MB (see level2_kernel_pgt in |
| 52 | * arch/x86/kernel/head_64.S), and it is mapped here: | 52 | * arch/x86/kernel/head_64.S), and it is mapped here: |
| 53 | */ | 53 | */ |
| 54 | #define KERNEL_IMAGE_SIZE (128*1024*1024) | 54 | #define KERNEL_IMAGE_SIZE (512 * 1024 * 1024) |
| 55 | #define KERNEL_IMAGE_START _AC(0xffffffff80000000, UL) | 55 | #define KERNEL_IMAGE_START _AC(0xffffffff80000000, UL) |
| 56 | 56 | ||
| 57 | #ifndef __ASSEMBLY__ | 57 | #ifndef __ASSEMBLY__ |
| @@ -59,7 +59,6 @@ void clear_page(void *page); | |||
| 59 | void copy_page(void *to, void *from); | 59 | void copy_page(void *to, void *from); |
| 60 | 60 | ||
| 61 | extern unsigned long end_pfn; | 61 | extern unsigned long end_pfn; |
| 62 | extern unsigned long end_pfn_map; | ||
| 63 | extern unsigned long phys_base; | 62 | extern unsigned long phys_base; |
| 64 | 63 | ||
| 65 | extern unsigned long __phys_addr(unsigned long); | 64 | extern unsigned long __phys_addr(unsigned long); |
| @@ -81,6 +80,9 @@ typedef struct { pteval_t pte; } pte_t; | |||
| 81 | 80 | ||
| 82 | #define vmemmap ((struct page *)VMEMMAP_START) | 81 | #define vmemmap ((struct page *)VMEMMAP_START) |
| 83 | 82 | ||
| 83 | extern unsigned long init_memory_mapping(unsigned long start, | ||
| 84 | unsigned long end); | ||
| 85 | |||
| 84 | #endif /* !__ASSEMBLY__ */ | 86 | #endif /* !__ASSEMBLY__ */ |
| 85 | 87 | ||
| 86 | #ifdef CONFIG_FLATMEM | 88 | #ifdef CONFIG_FLATMEM |
diff --git a/include/asm-x86/param.h b/include/asm-x86/param.h index c996ec4da0c8..6f0d0422f4ca 100644 --- a/include/asm-x86/param.h +++ b/include/asm-x86/param.h | |||
| @@ -3,8 +3,8 @@ | |||
| 3 | 3 | ||
| 4 | #ifdef __KERNEL__ | 4 | #ifdef __KERNEL__ |
| 5 | # define HZ CONFIG_HZ /* Internal kernel timer frequency */ | 5 | # define HZ CONFIG_HZ /* Internal kernel timer frequency */ |
| 6 | # define USER_HZ 100 /* .. some user interfaces are in "ticks" */ | 6 | # define USER_HZ 100 /* some user interfaces are */ |
| 7 | # define CLOCKS_PER_SEC (USER_HZ) /* like times() */ | 7 | # define CLOCKS_PER_SEC (USER_HZ) /* in "ticks" like times() */ |
| 8 | #endif | 8 | #endif |
| 9 | 9 | ||
| 10 | #ifndef HZ | 10 | #ifndef HZ |
diff --git a/include/asm-x86/paravirt.h b/include/asm-x86/paravirt.h index d6236eb46466..3d419398499b 100644 --- a/include/asm-x86/paravirt.h +++ b/include/asm-x86/paravirt.h | |||
| @@ -231,7 +231,8 @@ struct pv_mmu_ops { | |||
| 231 | void (*set_pte_at)(struct mm_struct *mm, unsigned long addr, | 231 | void (*set_pte_at)(struct mm_struct *mm, unsigned long addr, |
| 232 | pte_t *ptep, pte_t pteval); | 232 | pte_t *ptep, pte_t pteval); |
| 233 | void (*set_pmd)(pmd_t *pmdp, pmd_t pmdval); | 233 | void (*set_pmd)(pmd_t *pmdp, pmd_t pmdval); |
| 234 | void (*pte_update)(struct mm_struct *mm, unsigned long addr, pte_t *ptep); | 234 | void (*pte_update)(struct mm_struct *mm, unsigned long addr, |
| 235 | pte_t *ptep); | ||
| 235 | void (*pte_update_defer)(struct mm_struct *mm, | 236 | void (*pte_update_defer)(struct mm_struct *mm, |
| 236 | unsigned long addr, pte_t *ptep); | 237 | unsigned long addr, pte_t *ptep); |
| 237 | 238 | ||
| @@ -246,7 +247,8 @@ struct pv_mmu_ops { | |||
| 246 | void (*set_pte_atomic)(pte_t *ptep, pte_t pteval); | 247 | void (*set_pte_atomic)(pte_t *ptep, pte_t pteval); |
| 247 | void (*set_pte_present)(struct mm_struct *mm, unsigned long addr, | 248 | void (*set_pte_present)(struct mm_struct *mm, unsigned long addr, |
| 248 | pte_t *ptep, pte_t pte); | 249 | pte_t *ptep, pte_t pte); |
| 249 | void (*pte_clear)(struct mm_struct *mm, unsigned long addr, pte_t *ptep); | 250 | void (*pte_clear)(struct mm_struct *mm, unsigned long addr, |
| 251 | pte_t *ptep); | ||
| 250 | void (*pmd_clear)(pmd_t *pmdp); | 252 | void (*pmd_clear)(pmd_t *pmdp); |
| 251 | 253 | ||
| 252 | #endif /* CONFIG_X86_PAE */ | 254 | #endif /* CONFIG_X86_PAE */ |
| @@ -274,8 +276,7 @@ struct pv_mmu_ops { | |||
| 274 | /* This contains all the paravirt structures: we get a convenient | 276 | /* This contains all the paravirt structures: we get a convenient |
| 275 | * number for each function using the offset which we use to indicate | 277 | * number for each function using the offset which we use to indicate |
| 276 | * what to patch. */ | 278 | * what to patch. */ |
| 277 | struct paravirt_patch_template | 279 | struct paravirt_patch_template { |
| 278 | { | ||
| 279 | struct pv_init_ops pv_init_ops; | 280 | struct pv_init_ops pv_init_ops; |
| 280 | struct pv_time_ops pv_time_ops; | 281 | struct pv_time_ops pv_time_ops; |
| 281 | struct pv_cpu_ops pv_cpu_ops; | 282 | struct pv_cpu_ops pv_cpu_ops; |
| @@ -660,43 +661,56 @@ static inline int paravirt_write_msr(unsigned msr, unsigned low, unsigned high) | |||
| 660 | } | 661 | } |
| 661 | 662 | ||
| 662 | /* These should all do BUG_ON(_err), but our headers are too tangled. */ | 663 | /* These should all do BUG_ON(_err), but our headers are too tangled. */ |
| 663 | #define rdmsr(msr,val1,val2) do { \ | 664 | #define rdmsr(msr, val1, val2) \ |
| 665 | do { \ | ||
| 664 | int _err; \ | 666 | int _err; \ |
| 665 | u64 _l = paravirt_read_msr(msr, &_err); \ | 667 | u64 _l = paravirt_read_msr(msr, &_err); \ |
| 666 | val1 = (u32)_l; \ | 668 | val1 = (u32)_l; \ |
| 667 | val2 = _l >> 32; \ | 669 | val2 = _l >> 32; \ |
| 668 | } while(0) | 670 | } while (0) |
| 669 | 671 | ||
| 670 | #define wrmsr(msr,val1,val2) do { \ | 672 | #define wrmsr(msr, val1, val2) \ |
| 673 | do { \ | ||
| 671 | paravirt_write_msr(msr, val1, val2); \ | 674 | paravirt_write_msr(msr, val1, val2); \ |
| 672 | } while(0) | 675 | } while (0) |
| 673 | 676 | ||
| 674 | #define rdmsrl(msr,val) do { \ | 677 | #define rdmsrl(msr, val) \ |
| 678 | do { \ | ||
| 675 | int _err; \ | 679 | int _err; \ |
| 676 | val = paravirt_read_msr(msr, &_err); \ | 680 | val = paravirt_read_msr(msr, &_err); \ |
| 677 | } while(0) | 681 | } while (0) |
| 678 | 682 | ||
| 679 | #define wrmsrl(msr,val) wrmsr(msr, (u32)((u64)(val)), ((u64)(val))>>32) | 683 | #define wrmsrl(msr, val) wrmsr(msr, (u32)((u64)(val)), ((u64)(val))>>32) |
| 680 | #define wrmsr_safe(msr,a,b) paravirt_write_msr(msr, a, b) | 684 | #define wrmsr_safe(msr, a, b) paravirt_write_msr(msr, a, b) |
| 681 | 685 | ||
| 682 | /* rdmsr with exception handling */ | 686 | /* rdmsr with exception handling */ |
| 683 | #define rdmsr_safe(msr,a,b) ({ \ | 687 | #define rdmsr_safe(msr, a, b) \ |
| 688 | ({ \ | ||
| 684 | int _err; \ | 689 | int _err; \ |
| 685 | u64 _l = paravirt_read_msr(msr, &_err); \ | 690 | u64 _l = paravirt_read_msr(msr, &_err); \ |
| 686 | (*a) = (u32)_l; \ | 691 | (*a) = (u32)_l; \ |
| 687 | (*b) = _l >> 32; \ | 692 | (*b) = _l >> 32; \ |
| 688 | _err; }) | 693 | _err; \ |
| 694 | }) | ||
| 695 | |||
| 696 | static inline int rdmsrl_safe(unsigned msr, unsigned long long *p) | ||
| 697 | { | ||
| 698 | int err; | ||
| 689 | 699 | ||
| 700 | *p = paravirt_read_msr(msr, &err); | ||
| 701 | return err; | ||
| 702 | } | ||
| 690 | 703 | ||
| 691 | static inline u64 paravirt_read_tsc(void) | 704 | static inline u64 paravirt_read_tsc(void) |
| 692 | { | 705 | { |
| 693 | return PVOP_CALL0(u64, pv_cpu_ops.read_tsc); | 706 | return PVOP_CALL0(u64, pv_cpu_ops.read_tsc); |
| 694 | } | 707 | } |
| 695 | 708 | ||
| 696 | #define rdtscl(low) do { \ | 709 | #define rdtscl(low) \ |
| 710 | do { \ | ||
| 697 | u64 _l = paravirt_read_tsc(); \ | 711 | u64 _l = paravirt_read_tsc(); \ |
| 698 | low = (int)_l; \ | 712 | low = (int)_l; \ |
| 699 | } while(0) | 713 | } while (0) |
| 700 | 714 | ||
| 701 | #define rdtscll(val) (val = paravirt_read_tsc()) | 715 | #define rdtscll(val) (val = paravirt_read_tsc()) |
| 702 | 716 | ||
| @@ -711,11 +725,12 @@ static inline unsigned long long paravirt_read_pmc(int counter) | |||
| 711 | return PVOP_CALL1(u64, pv_cpu_ops.read_pmc, counter); | 725 | return PVOP_CALL1(u64, pv_cpu_ops.read_pmc, counter); |
| 712 | } | 726 | } |
| 713 | 727 | ||
| 714 | #define rdpmc(counter,low,high) do { \ | 728 | #define rdpmc(counter, low, high) \ |
| 729 | do { \ | ||
| 715 | u64 _l = paravirt_read_pmc(counter); \ | 730 | u64 _l = paravirt_read_pmc(counter); \ |
| 716 | low = (u32)_l; \ | 731 | low = (u32)_l; \ |
| 717 | high = _l >> 32; \ | 732 | high = _l >> 32; \ |
| 718 | } while(0) | 733 | } while (0) |
| 719 | 734 | ||
| 720 | static inline unsigned long long paravirt_rdtscp(unsigned int *aux) | 735 | static inline unsigned long long paravirt_rdtscp(unsigned int *aux) |
| 721 | { | 736 | { |
| @@ -794,7 +809,8 @@ static inline void set_iopl_mask(unsigned mask) | |||
| 794 | } | 809 | } |
| 795 | 810 | ||
| 796 | /* The paravirtualized I/O functions */ | 811 | /* The paravirtualized I/O functions */ |
| 797 | static inline void slow_down_io(void) { | 812 | static inline void slow_down_io(void) |
| 813 | { | ||
| 798 | pv_cpu_ops.io_delay(); | 814 | pv_cpu_ops.io_delay(); |
| 799 | #ifdef REALLY_SLOW_IO | 815 | #ifdef REALLY_SLOW_IO |
| 800 | pv_cpu_ops.io_delay(); | 816 | pv_cpu_ops.io_delay(); |
diff --git a/include/asm-x86/parport.h b/include/asm-x86/parport.h index 019cbca24a38..3c4ffeb467e9 100644 --- a/include/asm-x86/parport.h +++ b/include/asm-x86/parport.h | |||
| @@ -1,10 +1,10 @@ | |||
| 1 | #ifndef _ASM_X86_PARPORT_H | 1 | #ifndef _ASM_X86_PARPORT_H |
| 2 | #define _ASM_X86_PARPORT_H | 2 | #define _ASM_X86_PARPORT_H |
| 3 | 3 | ||
| 4 | static int __devinit parport_pc_find_isa_ports (int autoirq, int autodma); | 4 | static int __devinit parport_pc_find_isa_ports(int autoirq, int autodma); |
| 5 | static int __devinit parport_pc_find_nonpci_ports (int autoirq, int autodma) | 5 | static int __devinit parport_pc_find_nonpci_ports(int autoirq, int autodma) |
| 6 | { | 6 | { |
| 7 | return parport_pc_find_isa_ports (autoirq, autodma); | 7 | return parport_pc_find_isa_ports(autoirq, autodma); |
| 8 | } | 8 | } |
| 9 | 9 | ||
| 10 | #endif /* _ASM_X86_PARPORT_H */ | 10 | #endif /* _ASM_X86_PARPORT_H */ |
diff --git a/include/asm-x86/pat.h b/include/asm-x86/pat.h new file mode 100644 index 000000000000..8b822b5a1786 --- /dev/null +++ b/include/asm-x86/pat.h | |||
| @@ -0,0 +1,16 @@ | |||
| 1 | |||
| 2 | #ifndef _ASM_PAT_H | ||
| 3 | #define _ASM_PAT_H 1 | ||
| 4 | |||
| 5 | #include <linux/types.h> | ||
| 6 | |||
| 7 | extern int pat_wc_enabled; | ||
| 8 | |||
| 9 | extern void pat_init(void); | ||
| 10 | |||
| 11 | extern int reserve_memtype(u64 start, u64 end, | ||
| 12 | unsigned long req_type, unsigned long *ret_type); | ||
| 13 | extern int free_memtype(u64 start, u64 end); | ||
| 14 | |||
| 15 | #endif | ||
| 16 | |||
diff --git a/include/asm-x86/pci-direct.h b/include/asm-x86/pci-direct.h index 6823fa4f1afa..5b21485be573 100644 --- a/include/asm-x86/pci-direct.h +++ b/include/asm-x86/pci-direct.h | |||
| @@ -4,7 +4,7 @@ | |||
| 4 | #include <linux/types.h> | 4 | #include <linux/types.h> |
| 5 | 5 | ||
| 6 | /* Direct PCI access. This is used for PCI accesses in early boot before | 6 | /* Direct PCI access. This is used for PCI accesses in early boot before |
| 7 | the PCI subsystem works. */ | 7 | the PCI subsystem works. */ |
| 8 | 8 | ||
| 9 | extern u32 read_pci_config(u8 bus, u8 slot, u8 func, u8 offset); | 9 | extern u32 read_pci_config(u8 bus, u8 slot, u8 func, u8 offset); |
| 10 | extern u8 read_pci_config_byte(u8 bus, u8 slot, u8 func, u8 offset); | 10 | extern u8 read_pci_config_byte(u8 bus, u8 slot, u8 func, u8 offset); |
diff --git a/include/asm-x86/pci.h b/include/asm-x86/pci.h index c61190cb9e12..ddd8e248fc0a 100644 --- a/include/asm-x86/pci.h +++ b/include/asm-x86/pci.h | |||
| @@ -8,14 +8,13 @@ | |||
| 8 | #include <asm/scatterlist.h> | 8 | #include <asm/scatterlist.h> |
| 9 | #include <asm/io.h> | 9 | #include <asm/io.h> |
| 10 | 10 | ||
| 11 | |||
| 12 | #ifdef __KERNEL__ | 11 | #ifdef __KERNEL__ |
| 13 | 12 | ||
| 14 | struct pci_sysdata { | 13 | struct pci_sysdata { |
| 15 | int domain; /* PCI domain */ | 14 | int domain; /* PCI domain */ |
| 16 | int node; /* NUMA node */ | 15 | int node; /* NUMA node */ |
| 17 | #ifdef CONFIG_X86_64 | 16 | #ifdef CONFIG_X86_64 |
| 18 | void* iommu; /* IOMMU private data */ | 17 | void *iommu; /* IOMMU private data */ |
| 19 | #endif | 18 | #endif |
| 20 | }; | 19 | }; |
| 21 | 20 | ||
| @@ -52,7 +51,7 @@ extern unsigned long pci_mem_start; | |||
| 52 | #define PCIBIOS_MIN_CARDBUS_IO 0x4000 | 51 | #define PCIBIOS_MIN_CARDBUS_IO 0x4000 |
| 53 | 52 | ||
| 54 | void pcibios_config_init(void); | 53 | void pcibios_config_init(void); |
| 55 | struct pci_bus * pcibios_scan_root(int bus); | 54 | struct pci_bus *pcibios_scan_root(int bus); |
| 56 | 55 | ||
| 57 | void pcibios_set_master(struct pci_dev *dev); | 56 | void pcibios_set_master(struct pci_dev *dev); |
| 58 | void pcibios_penalize_isa_irq(int irq, int active); | 57 | void pcibios_penalize_isa_irq(int irq, int active); |
| @@ -62,7 +61,8 @@ int pcibios_set_irq_routing(struct pci_dev *dev, int pin, int irq); | |||
| 62 | 61 | ||
| 63 | #define HAVE_PCI_MMAP | 62 | #define HAVE_PCI_MMAP |
| 64 | extern int pci_mmap_page_range(struct pci_dev *dev, struct vm_area_struct *vma, | 63 | extern int pci_mmap_page_range(struct pci_dev *dev, struct vm_area_struct *vma, |
| 65 | enum pci_mmap_state mmap_state, int write_combine); | 64 | enum pci_mmap_state mmap_state, |
| 65 | int write_combine); | ||
| 66 | 66 | ||
| 67 | 67 | ||
| 68 | #ifdef CONFIG_PCI | 68 | #ifdef CONFIG_PCI |
diff --git a/include/asm-x86/pci_64.h b/include/asm-x86/pci_64.h index 374690314539..df867e5d80b1 100644 --- a/include/asm-x86/pci_64.h +++ b/include/asm-x86/pci_64.h | |||
| @@ -1,12 +1,10 @@ | |||
| 1 | #ifndef __x8664_PCI_H | 1 | #ifndef __x8664_PCI_H |
| 2 | #define __x8664_PCI_H | 2 | #define __x8664_PCI_H |
| 3 | 3 | ||
| 4 | |||
| 5 | #ifdef __KERNEL__ | 4 | #ifdef __KERNEL__ |
| 6 | 5 | ||
| 7 | |||
| 8 | #ifdef CONFIG_CALGARY_IOMMU | 6 | #ifdef CONFIG_CALGARY_IOMMU |
| 9 | static inline void* pci_iommu(struct pci_bus *bus) | 7 | static inline void *pci_iommu(struct pci_bus *bus) |
| 10 | { | 8 | { |
| 11 | struct pci_sysdata *sd = bus->sysdata; | 9 | struct pci_sysdata *sd = bus->sysdata; |
| 12 | return sd->iommu; | 10 | return sd->iommu; |
| @@ -19,11 +17,10 @@ static inline void set_pci_iommu(struct pci_bus *bus, void *val) | |||
| 19 | } | 17 | } |
| 20 | #endif /* CONFIG_CALGARY_IOMMU */ | 18 | #endif /* CONFIG_CALGARY_IOMMU */ |
| 21 | 19 | ||
| 22 | 20 | extern int (*pci_config_read)(int seg, int bus, int dev, int fn, | |
| 23 | extern int (*pci_config_read)(int seg, int bus, int dev, int fn, int reg, int len, u32 *value); | 21 | int reg, int len, u32 *value); |
| 24 | extern int (*pci_config_write)(int seg, int bus, int dev, int fn, int reg, int len, u32 value); | 22 | extern int (*pci_config_write)(int seg, int bus, int dev, int fn, |
| 25 | 23 | int reg, int len, u32 value); | |
| 26 | |||
| 27 | 24 | ||
| 28 | extern void pci_iommu_alloc(void); | 25 | extern void pci_iommu_alloc(void); |
| 29 | 26 | ||
| @@ -65,5 +62,4 @@ extern void pci_iommu_alloc(void); | |||
| 65 | 62 | ||
| 66 | #endif /* __KERNEL__ */ | 63 | #endif /* __KERNEL__ */ |
| 67 | 64 | ||
| 68 | |||
| 69 | #endif /* __x8664_PCI_H */ | 65 | #endif /* __x8664_PCI_H */ |
diff --git a/include/asm-x86/pda.h b/include/asm-x86/pda.h index c0305bff0f19..101fb9e11954 100644 --- a/include/asm-x86/pda.h +++ b/include/asm-x86/pda.h | |||
| @@ -22,7 +22,6 @@ struct x8664_pda { | |||
| 22 | offset 40!!! */ | 22 | offset 40!!! */ |
| 23 | #endif | 23 | #endif |
| 24 | char *irqstackptr; | 24 | char *irqstackptr; |
| 25 | unsigned int nodenumber; /* number of current node */ | ||
| 26 | unsigned int __softirq_pending; | 25 | unsigned int __softirq_pending; |
| 27 | unsigned int __nmi_count; /* number of NMI on this CPUs */ | 26 | unsigned int __nmi_count; /* number of NMI on this CPUs */ |
| 28 | short mmu_state; | 27 | short mmu_state; |
| @@ -58,34 +57,36 @@ extern struct x8664_pda _proxy_pda; | |||
| 58 | 57 | ||
| 59 | #define pda_offset(field) offsetof(struct x8664_pda, field) | 58 | #define pda_offset(field) offsetof(struct x8664_pda, field) |
| 60 | 59 | ||
| 61 | #define pda_to_op(op, field, val) do { \ | 60 | #define pda_to_op(op, field, val) \ |
| 62 | typedef typeof(_proxy_pda.field) T__; \ | 61 | do { \ |
| 63 | if (0) { T__ tmp__; tmp__ = (val); } /* type checking */ \ | 62 | typedef typeof(_proxy_pda.field) T__; \ |
| 64 | switch (sizeof(_proxy_pda.field)) { \ | 63 | if (0) { T__ tmp__; tmp__ = (val); } /* type checking */ \ |
| 65 | case 2: \ | 64 | switch (sizeof(_proxy_pda.field)) { \ |
| 66 | asm(op "w %1,%%gs:%c2" : \ | 65 | case 2: \ |
| 67 | "+m" (_proxy_pda.field) : \ | 66 | asm(op "w %1,%%gs:%c2" : \ |
| 68 | "ri" ((T__)val), \ | 67 | "+m" (_proxy_pda.field) : \ |
| 69 | "i"(pda_offset(field))); \ | 68 | "ri" ((T__)val), \ |
| 70 | break; \ | 69 | "i"(pda_offset(field))); \ |
| 71 | case 4: \ | 70 | break; \ |
| 72 | asm(op "l %1,%%gs:%c2" : \ | 71 | case 4: \ |
| 73 | "+m" (_proxy_pda.field) : \ | 72 | asm(op "l %1,%%gs:%c2" : \ |
| 74 | "ri" ((T__)val), \ | 73 | "+m" (_proxy_pda.field) : \ |
| 75 | "i" (pda_offset(field))); \ | 74 | "ri" ((T__)val), \ |
| 76 | break; \ | 75 | "i" (pda_offset(field))); \ |
| 77 | case 8: \ | 76 | break; \ |
| 78 | asm(op "q %1,%%gs:%c2": \ | 77 | case 8: \ |
| 79 | "+m" (_proxy_pda.field) : \ | 78 | asm(op "q %1,%%gs:%c2": \ |
| 80 | "ri" ((T__)val), \ | 79 | "+m" (_proxy_pda.field) : \ |
| 81 | "i"(pda_offset(field))); \ | 80 | "ri" ((T__)val), \ |
| 82 | break; \ | 81 | "i"(pda_offset(field))); \ |
| 83 | default: \ | 82 | break; \ |
| 84 | __bad_pda_field(); \ | 83 | default: \ |
| 85 | } \ | 84 | __bad_pda_field(); \ |
| 86 | } while (0) | 85 | } \ |
| 86 | } while (0) | ||
| 87 | 87 | ||
| 88 | #define pda_from_op(op,field) ({ \ | 88 | #define pda_from_op(op, field) \ |
| 89 | ({ \ | ||
| 89 | typeof(_proxy_pda.field) ret__; \ | 90 | typeof(_proxy_pda.field) ret__; \ |
| 90 | switch (sizeof(_proxy_pda.field)) { \ | 91 | switch (sizeof(_proxy_pda.field)) { \ |
| 91 | case 2: \ | 92 | case 2: \ |
| @@ -93,23 +94,24 @@ extern struct x8664_pda _proxy_pda; | |||
| 93 | "=r" (ret__) : \ | 94 | "=r" (ret__) : \ |
| 94 | "i" (pda_offset(field)), \ | 95 | "i" (pda_offset(field)), \ |
| 95 | "m" (_proxy_pda.field)); \ | 96 | "m" (_proxy_pda.field)); \ |
| 96 | break; \ | 97 | break; \ |
| 97 | case 4: \ | 98 | case 4: \ |
| 98 | asm(op "l %%gs:%c1,%0": \ | 99 | asm(op "l %%gs:%c1,%0": \ |
| 99 | "=r" (ret__): \ | 100 | "=r" (ret__): \ |
| 100 | "i" (pda_offset(field)), \ | 101 | "i" (pda_offset(field)), \ |
| 101 | "m" (_proxy_pda.field)); \ | 102 | "m" (_proxy_pda.field)); \ |
| 102 | break; \ | 103 | break; \ |
| 103 | case 8: \ | 104 | case 8: \ |
| 104 | asm(op "q %%gs:%c1,%0": \ | 105 | asm(op "q %%gs:%c1,%0": \ |
| 105 | "=r" (ret__) : \ | 106 | "=r" (ret__) : \ |
| 106 | "i" (pda_offset(field)), \ | 107 | "i" (pda_offset(field)), \ |
| 107 | "m" (_proxy_pda.field)); \ | 108 | "m" (_proxy_pda.field)); \ |
| 108 | break; \ | 109 | break; \ |
| 109 | default: \ | 110 | default: \ |
| 110 | __bad_pda_field(); \ | 111 | __bad_pda_field(); \ |
| 111 | } \ | 112 | } \ |
| 112 | ret__; }) | 113 | ret__; \ |
| 114 | }) | ||
| 113 | 115 | ||
| 114 | #define read_pda(field) pda_from_op("mov", field) | 116 | #define read_pda(field) pda_from_op("mov", field) |
| 115 | #define write_pda(field, val) pda_to_op("mov", field, val) | 117 | #define write_pda(field, val) pda_to_op("mov", field, val) |
| @@ -118,12 +120,13 @@ extern struct x8664_pda _proxy_pda; | |||
| 118 | #define or_pda(field, val) pda_to_op("or", field, val) | 120 | #define or_pda(field, val) pda_to_op("or", field, val) |
| 119 | 121 | ||
| 120 | /* This is not atomic against other CPUs -- CPU preemption needs to be off */ | 122 | /* This is not atomic against other CPUs -- CPU preemption needs to be off */ |
| 121 | #define test_and_clear_bit_pda(bit, field) ({ \ | 123 | #define test_and_clear_bit_pda(bit, field) \ |
| 122 | int old__; \ | 124 | ({ \ |
| 123 | asm volatile("btr %2,%%gs:%c3\n\tsbbl %0,%0" \ | 125 | int old__; \ |
| 124 | : "=r" (old__), "+m" (_proxy_pda.field) \ | 126 | asm volatile("btr %2,%%gs:%c3\n\tsbbl %0,%0" \ |
| 125 | : "dIr" (bit), "i" (pda_offset(field)) : "memory"); \ | 127 | : "=r" (old__), "+m" (_proxy_pda.field) \ |
| 126 | old__; \ | 128 | : "dIr" (bit), "i" (pda_offset(field)) : "memory");\ |
| 129 | old__; \ | ||
| 127 | }) | 130 | }) |
| 128 | 131 | ||
| 129 | #endif | 132 | #endif |
diff --git a/include/asm-x86/percpu.h b/include/asm-x86/percpu.h index 0dec00f27eb4..736fc3bb8e1e 100644 --- a/include/asm-x86/percpu.h +++ b/include/asm-x86/percpu.h | |||
| @@ -85,58 +85,62 @@ DECLARE_PER_CPU(unsigned long, this_cpu_off); | |||
| 85 | * don't give an lvalue though). */ | 85 | * don't give an lvalue though). */ |
| 86 | extern void __bad_percpu_size(void); | 86 | extern void __bad_percpu_size(void); |
| 87 | 87 | ||
| 88 | #define percpu_to_op(op,var,val) \ | 88 | #define percpu_to_op(op, var, val) \ |
| 89 | do { \ | 89 | do { \ |
| 90 | typedef typeof(var) T__; \ | 90 | typedef typeof(var) T__; \ |
| 91 | if (0) { T__ tmp__; tmp__ = (val); } \ | 91 | if (0) { \ |
| 92 | switch (sizeof(var)) { \ | 92 | T__ tmp__; \ |
| 93 | case 1: \ | 93 | tmp__ = (val); \ |
| 94 | asm(op "b %1,"__percpu_seg"%0" \ | 94 | } \ |
| 95 | : "+m" (var) \ | 95 | switch (sizeof(var)) { \ |
| 96 | :"ri" ((T__)val)); \ | 96 | case 1: \ |
| 97 | break; \ | 97 | asm(op "b %1,"__percpu_seg"%0" \ |
| 98 | case 2: \ | 98 | : "+m" (var) \ |
| 99 | asm(op "w %1,"__percpu_seg"%0" \ | 99 | : "ri" ((T__)val)); \ |
| 100 | : "+m" (var) \ | 100 | break; \ |
| 101 | :"ri" ((T__)val)); \ | 101 | case 2: \ |
| 102 | break; \ | 102 | asm(op "w %1,"__percpu_seg"%0" \ |
| 103 | case 4: \ | 103 | : "+m" (var) \ |
| 104 | asm(op "l %1,"__percpu_seg"%0" \ | 104 | : "ri" ((T__)val)); \ |
| 105 | : "+m" (var) \ | 105 | break; \ |
| 106 | :"ri" ((T__)val)); \ | 106 | case 4: \ |
| 107 | break; \ | 107 | asm(op "l %1,"__percpu_seg"%0" \ |
| 108 | default: __bad_percpu_size(); \ | 108 | : "+m" (var) \ |
| 109 | } \ | 109 | : "ri" ((T__)val)); \ |
| 110 | } while (0) | 110 | break; \ |
| 111 | 111 | default: __bad_percpu_size(); \ | |
| 112 | #define percpu_from_op(op,var) \ | 112 | } \ |
| 113 | ({ \ | 113 | } while (0) |
| 114 | typeof(var) ret__; \ | 114 | |
| 115 | switch (sizeof(var)) { \ | 115 | #define percpu_from_op(op, var) \ |
| 116 | case 1: \ | 116 | ({ \ |
| 117 | asm(op "b "__percpu_seg"%1,%0" \ | 117 | typeof(var) ret__; \ |
| 118 | : "=r" (ret__) \ | 118 | switch (sizeof(var)) { \ |
| 119 | : "m" (var)); \ | 119 | case 1: \ |
| 120 | break; \ | 120 | asm(op "b "__percpu_seg"%1,%0" \ |
| 121 | case 2: \ | 121 | : "=r" (ret__) \ |
| 122 | asm(op "w "__percpu_seg"%1,%0" \ | 122 | : "m" (var)); \ |
| 123 | : "=r" (ret__) \ | 123 | break; \ |
| 124 | : "m" (var)); \ | 124 | case 2: \ |
| 125 | break; \ | 125 | asm(op "w "__percpu_seg"%1,%0" \ |
| 126 | case 4: \ | 126 | : "=r" (ret__) \ |
| 127 | asm(op "l "__percpu_seg"%1,%0" \ | 127 | : "m" (var)); \ |
| 128 | : "=r" (ret__) \ | 128 | break; \ |
| 129 | : "m" (var)); \ | 129 | case 4: \ |
| 130 | break; \ | 130 | asm(op "l "__percpu_seg"%1,%0" \ |
| 131 | default: __bad_percpu_size(); \ | 131 | : "=r" (ret__) \ |
| 132 | } \ | 132 | : "m" (var)); \ |
| 133 | ret__; }) | 133 | break; \ |
| 134 | default: __bad_percpu_size(); \ | ||
| 135 | } \ | ||
| 136 | ret__; \ | ||
| 137 | }) | ||
| 134 | 138 | ||
| 135 | #define x86_read_percpu(var) percpu_from_op("mov", per_cpu__##var) | 139 | #define x86_read_percpu(var) percpu_from_op("mov", per_cpu__##var) |
| 136 | #define x86_write_percpu(var,val) percpu_to_op("mov", per_cpu__##var, val) | 140 | #define x86_write_percpu(var, val) percpu_to_op("mov", per_cpu__##var, val) |
| 137 | #define x86_add_percpu(var,val) percpu_to_op("add", per_cpu__##var, val) | 141 | #define x86_add_percpu(var, val) percpu_to_op("add", per_cpu__##var, val) |
| 138 | #define x86_sub_percpu(var,val) percpu_to_op("sub", per_cpu__##var, val) | 142 | #define x86_sub_percpu(var, val) percpu_to_op("sub", per_cpu__##var, val) |
| 139 | #define x86_or_percpu(var,val) percpu_to_op("or", per_cpu__##var, val) | 143 | #define x86_or_percpu(var, val) percpu_to_op("or", per_cpu__##var, val) |
| 140 | #endif /* !__ASSEMBLY__ */ | 144 | #endif /* !__ASSEMBLY__ */ |
| 141 | #endif /* !CONFIG_X86_64 */ | 145 | #endif /* !CONFIG_X86_64 */ |
| 142 | #endif /* _ASM_X86_PERCPU_H_ */ | 146 | #endif /* _ASM_X86_PERCPU_H_ */ |
diff --git a/include/asm-x86/pgtable-2level.h b/include/asm-x86/pgtable-2level.h index 701404fab308..46bc52c0eae1 100644 --- a/include/asm-x86/pgtable-2level.h +++ b/include/asm-x86/pgtable-2level.h | |||
| @@ -26,7 +26,8 @@ static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte) | |||
| 26 | native_set_pte(ptep, pte); | 26 | native_set_pte(ptep, pte); |
| 27 | } | 27 | } |
| 28 | 28 | ||
| 29 | static inline void native_set_pte_present(struct mm_struct *mm, unsigned long addr, | 29 | static inline void native_set_pte_present(struct mm_struct *mm, |
| 30 | unsigned long addr, | ||
| 30 | pte_t *ptep, pte_t pte) | 31 | pte_t *ptep, pte_t pte) |
| 31 | { | 32 | { |
| 32 | native_set_pte(ptep, pte); | 33 | native_set_pte(ptep, pte); |
| @@ -37,7 +38,8 @@ static inline void native_pmd_clear(pmd_t *pmdp) | |||
| 37 | native_set_pmd(pmdp, __pmd(0)); | 38 | native_set_pmd(pmdp, __pmd(0)); |
| 38 | } | 39 | } |
| 39 | 40 | ||
| 40 | static inline void native_pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *xp) | 41 | static inline void native_pte_clear(struct mm_struct *mm, |
| 42 | unsigned long addr, pte_t *xp) | ||
| 41 | { | 43 | { |
| 42 | *xp = native_make_pte(0); | 44 | *xp = native_make_pte(0); |
| 43 | } | 45 | } |
| @@ -61,16 +63,18 @@ static inline pte_t native_ptep_get_and_clear(pte_t *xp) | |||
| 61 | */ | 63 | */ |
| 62 | #define PTE_FILE_MAX_BITS 29 | 64 | #define PTE_FILE_MAX_BITS 29 |
| 63 | 65 | ||
| 64 | #define pte_to_pgoff(pte) \ | 66 | #define pte_to_pgoff(pte) \ |
| 65 | ((((pte).pte_low >> 1) & 0x1f ) + (((pte).pte_low >> 8) << 5 )) | 67 | ((((pte).pte_low >> 1) & 0x1f) + (((pte).pte_low >> 8) << 5)) |
| 66 | 68 | ||
| 67 | #define pgoff_to_pte(off) \ | 69 | #define pgoff_to_pte(off) \ |
| 68 | ((pte_t) { .pte_low = (((off) & 0x1f) << 1) + (((off) >> 5) << 8) + _PAGE_FILE }) | 70 | ((pte_t) { .pte_low = (((off) & 0x1f) << 1) + \ |
| 71 | (((off) >> 5) << 8) + _PAGE_FILE }) | ||
| 69 | 72 | ||
| 70 | /* Encode and de-code a swap entry */ | 73 | /* Encode and de-code a swap entry */ |
| 71 | #define __swp_type(x) (((x).val >> 1) & 0x1f) | 74 | #define __swp_type(x) (((x).val >> 1) & 0x1f) |
| 72 | #define __swp_offset(x) ((x).val >> 8) | 75 | #define __swp_offset(x) ((x).val >> 8) |
| 73 | #define __swp_entry(type, offset) ((swp_entry_t) { ((type) << 1) | ((offset) << 8) }) | 76 | #define __swp_entry(type, offset) \ |
| 77 | ((swp_entry_t) { ((type) << 1) | ((offset) << 8) }) | ||
| 74 | #define __pte_to_swp_entry(pte) ((swp_entry_t) { (pte).pte_low }) | 78 | #define __pte_to_swp_entry(pte) ((swp_entry_t) { (pte).pte_low }) |
| 75 | #define __swp_entry_to_pte(x) ((pte_t) { .pte = (x).val }) | 79 | #define __swp_entry_to_pte(x) ((pte_t) { .pte = (x).val }) |
| 76 | 80 | ||
diff --git a/include/asm-x86/pgtable-3level.h b/include/asm-x86/pgtable-3level.h index 1d763eec740f..8b4a9d44b7f4 100644 --- a/include/asm-x86/pgtable-3level.h +++ b/include/asm-x86/pgtable-3level.h | |||
| @@ -8,22 +8,26 @@ | |||
| 8 | * Copyright (C) 1999 Ingo Molnar <mingo@redhat.com> | 8 | * Copyright (C) 1999 Ingo Molnar <mingo@redhat.com> |
| 9 | */ | 9 | */ |
| 10 | 10 | ||
| 11 | #define pte_ERROR(e) \ | 11 | #define pte_ERROR(e) \ |
| 12 | printk("%s:%d: bad pte %p(%08lx%08lx).\n", __FILE__, __LINE__, &(e), (e).pte_high, (e).pte_low) | 12 | printk("%s:%d: bad pte %p(%08lx%08lx).\n", \ |
| 13 | #define pmd_ERROR(e) \ | 13 | __FILE__, __LINE__, &(e), (e).pte_high, (e).pte_low) |
| 14 | printk("%s:%d: bad pmd %p(%016Lx).\n", __FILE__, __LINE__, &(e), pmd_val(e)) | 14 | #define pmd_ERROR(e) \ |
| 15 | #define pgd_ERROR(e) \ | 15 | printk("%s:%d: bad pmd %p(%016Lx).\n", \ |
| 16 | printk("%s:%d: bad pgd %p(%016Lx).\n", __FILE__, __LINE__, &(e), pgd_val(e)) | 16 | __FILE__, __LINE__, &(e), pmd_val(e)) |
| 17 | 17 | #define pgd_ERROR(e) \ | |
| 18 | printk("%s:%d: bad pgd %p(%016Lx).\n", \ | ||
| 19 | __FILE__, __LINE__, &(e), pgd_val(e)) | ||
| 18 | 20 | ||
| 19 | static inline int pud_none(pud_t pud) | 21 | static inline int pud_none(pud_t pud) |
| 20 | { | 22 | { |
| 21 | return pud_val(pud) == 0; | 23 | return pud_val(pud) == 0; |
| 22 | } | 24 | } |
| 25 | |||
| 23 | static inline int pud_bad(pud_t pud) | 26 | static inline int pud_bad(pud_t pud) |
| 24 | { | 27 | { |
| 25 | return (pud_val(pud) & ~(PTE_MASK | _KERNPG_TABLE | _PAGE_USER)) != 0; | 28 | return (pud_val(pud) & ~(PTE_MASK | _KERNPG_TABLE | _PAGE_USER)) != 0; |
| 26 | } | 29 | } |
| 30 | |||
| 27 | static inline int pud_present(pud_t pud) | 31 | static inline int pud_present(pud_t pud) |
| 28 | { | 32 | { |
| 29 | return pud_val(pud) & _PAGE_PRESENT; | 33 | return pud_val(pud) & _PAGE_PRESENT; |
| @@ -48,7 +52,8 @@ static inline void native_set_pte(pte_t *ptep, pte_t pte) | |||
| 48 | * we are justified in merely clearing the PTE present bit, followed | 52 | * we are justified in merely clearing the PTE present bit, followed |
| 49 | * by a set. The ordering here is important. | 53 | * by a set. The ordering here is important. |
| 50 | */ | 54 | */ |
| 51 | static inline void native_set_pte_present(struct mm_struct *mm, unsigned long addr, | 55 | static inline void native_set_pte_present(struct mm_struct *mm, |
| 56 | unsigned long addr, | ||
| 52 | pte_t *ptep, pte_t pte) | 57 | pte_t *ptep, pte_t pte) |
| 53 | { | 58 | { |
| 54 | ptep->pte_low = 0; | 59 | ptep->pte_low = 0; |
| @@ -60,15 +65,17 @@ static inline void native_set_pte_present(struct mm_struct *mm, unsigned long ad | |||
| 60 | 65 | ||
| 61 | static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte) | 66 | static inline void native_set_pte_atomic(pte_t *ptep, pte_t pte) |
| 62 | { | 67 | { |
| 63 | set_64bit((unsigned long long *)(ptep),native_pte_val(pte)); | 68 | set_64bit((unsigned long long *)(ptep), native_pte_val(pte)); |
| 64 | } | 69 | } |
| 70 | |||
| 65 | static inline void native_set_pmd(pmd_t *pmdp, pmd_t pmd) | 71 | static inline void native_set_pmd(pmd_t *pmdp, pmd_t pmd) |
| 66 | { | 72 | { |
| 67 | set_64bit((unsigned long long *)(pmdp),native_pmd_val(pmd)); | 73 | set_64bit((unsigned long long *)(pmdp), native_pmd_val(pmd)); |
| 68 | } | 74 | } |
| 75 | |||
| 69 | static inline void native_set_pud(pud_t *pudp, pud_t pud) | 76 | static inline void native_set_pud(pud_t *pudp, pud_t pud) |
| 70 | { | 77 | { |
| 71 | set_64bit((unsigned long long *)(pudp),native_pud_val(pud)); | 78 | set_64bit((unsigned long long *)(pudp), native_pud_val(pud)); |
| 72 | } | 79 | } |
| 73 | 80 | ||
| 74 | /* | 81 | /* |
| @@ -76,7 +83,8 @@ static inline void native_set_pud(pud_t *pudp, pud_t pud) | |||
| 76 | * entry, so clear the bottom half first and enforce ordering with a compiler | 83 | * entry, so clear the bottom half first and enforce ordering with a compiler |
| 77 | * barrier. | 84 | * barrier. |
| 78 | */ | 85 | */ |
| 79 | static inline void native_pte_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) | 86 | static inline void native_pte_clear(struct mm_struct *mm, unsigned long addr, |
| 87 | pte_t *ptep) | ||
| 80 | { | 88 | { |
| 81 | ptep->pte_low = 0; | 89 | ptep->pte_low = 0; |
| 82 | smp_wmb(); | 90 | smp_wmb(); |
| @@ -107,20 +115,19 @@ static inline void pud_clear(pud_t *pudp) | |||
| 107 | * current pgd to avoid unnecessary TLB flushes. | 115 | * current pgd to avoid unnecessary TLB flushes. |
| 108 | */ | 116 | */ |
| 109 | pgd = read_cr3(); | 117 | pgd = read_cr3(); |
| 110 | if (__pa(pudp) >= pgd && __pa(pudp) < (pgd + sizeof(pgd_t)*PTRS_PER_PGD)) | 118 | if (__pa(pudp) >= pgd && __pa(pudp) < |
| 119 | (pgd + sizeof(pgd_t)*PTRS_PER_PGD)) | ||
| 111 | write_cr3(pgd); | 120 | write_cr3(pgd); |
| 112 | } | 121 | } |
| 113 | 122 | ||
| 114 | #define pud_page(pud) \ | 123 | #define pud_page(pud) ((struct page *) __va(pud_val(pud) & PAGE_MASK)) |
| 115 | ((struct page *) __va(pud_val(pud) & PAGE_MASK)) | ||
| 116 | 124 | ||
| 117 | #define pud_page_vaddr(pud) \ | 125 | #define pud_page_vaddr(pud) ((unsigned long) __va(pud_val(pud) & PAGE_MASK)) |
| 118 | ((unsigned long) __va(pud_val(pud) & PAGE_MASK)) | ||
| 119 | 126 | ||
| 120 | 127 | ||
| 121 | /* Find an entry in the second-level page table.. */ | 128 | /* Find an entry in the second-level page table.. */ |
| 122 | #define pmd_offset(pud, address) ((pmd_t *) pud_page(*(pud)) + \ | 129 | #define pmd_offset(pud, address) ((pmd_t *)pud_page(*(pud)) + \ |
| 123 | pmd_index(address)) | 130 | pmd_index(address)) |
| 124 | 131 | ||
| 125 | #ifdef CONFIG_SMP | 132 | #ifdef CONFIG_SMP |
| 126 | static inline pte_t native_ptep_get_and_clear(pte_t *ptep) | 133 | static inline pte_t native_ptep_get_and_clear(pte_t *ptep) |
| @@ -161,7 +168,8 @@ static inline unsigned long pte_pfn(pte_t pte) | |||
| 161 | * put the 32 bits of offset into the high part. | 168 | * put the 32 bits of offset into the high part. |
| 162 | */ | 169 | */ |
| 163 | #define pte_to_pgoff(pte) ((pte).pte_high) | 170 | #define pte_to_pgoff(pte) ((pte).pte_high) |
| 164 | #define pgoff_to_pte(off) ((pte_t) { { .pte_low = _PAGE_FILE, .pte_high = (off) } }) | 171 | #define pgoff_to_pte(off) \ |
| 172 | ((pte_t) { { .pte_low = _PAGE_FILE, .pte_high = (off) } }) | ||
| 165 | #define PTE_FILE_MAX_BITS 32 | 173 | #define PTE_FILE_MAX_BITS 32 |
| 166 | 174 | ||
| 167 | /* Encode and de-code a swap entry */ | 175 | /* Encode and de-code a swap entry */ |
diff --git a/include/asm-x86/pgtable.h b/include/asm-x86/pgtable.h index 9cf472aeb9ce..f1d9f4a03f6f 100644 --- a/include/asm-x86/pgtable.h +++ b/include/asm-x86/pgtable.h | |||
| @@ -4,13 +4,13 @@ | |||
| 4 | #define USER_PTRS_PER_PGD ((TASK_SIZE-1)/PGDIR_SIZE+1) | 4 | #define USER_PTRS_PER_PGD ((TASK_SIZE-1)/PGDIR_SIZE+1) |
| 5 | #define FIRST_USER_ADDRESS 0 | 5 | #define FIRST_USER_ADDRESS 0 |
| 6 | 6 | ||
| 7 | #define _PAGE_BIT_PRESENT 0 | 7 | #define _PAGE_BIT_PRESENT 0 /* is present */ |
| 8 | #define _PAGE_BIT_RW 1 | 8 | #define _PAGE_BIT_RW 1 /* writeable */ |
| 9 | #define _PAGE_BIT_USER 2 | 9 | #define _PAGE_BIT_USER 2 /* userspace addressable */ |
| 10 | #define _PAGE_BIT_PWT 3 | 10 | #define _PAGE_BIT_PWT 3 /* page write through */ |
| 11 | #define _PAGE_BIT_PCD 4 | 11 | #define _PAGE_BIT_PCD 4 /* page cache disabled */ |
| 12 | #define _PAGE_BIT_ACCESSED 5 | 12 | #define _PAGE_BIT_ACCESSED 5 /* was accessed (raised by CPU) */ |
| 13 | #define _PAGE_BIT_DIRTY 6 | 13 | #define _PAGE_BIT_DIRTY 6 /* was written to (raised by CPU) */ |
| 14 | #define _PAGE_BIT_FILE 6 | 14 | #define _PAGE_BIT_FILE 6 |
| 15 | #define _PAGE_BIT_PSE 7 /* 4 MB (or 2MB) page */ | 15 | #define _PAGE_BIT_PSE 7 /* 4 MB (or 2MB) page */ |
| 16 | #define _PAGE_BIT_PAT 7 /* on 4KB pages */ | 16 | #define _PAGE_BIT_PAT 7 /* on 4KB pages */ |
| @@ -48,24 +48,39 @@ | |||
| 48 | #endif | 48 | #endif |
| 49 | 49 | ||
| 50 | /* If _PAGE_PRESENT is clear, we use these: */ | 50 | /* If _PAGE_PRESENT is clear, we use these: */ |
| 51 | #define _PAGE_FILE _PAGE_DIRTY /* nonlinear file mapping, saved PTE; unset:swap */ | 51 | #define _PAGE_FILE _PAGE_DIRTY /* nonlinear file mapping, |
| 52 | * saved PTE; unset:swap */ | ||
| 52 | #define _PAGE_PROTNONE _PAGE_PSE /* if the user mapped it with PROT_NONE; | 53 | #define _PAGE_PROTNONE _PAGE_PSE /* if the user mapped it with PROT_NONE; |
| 53 | pte_present gives true */ | 54 | pte_present gives true */ |
| 54 | 55 | ||
| 55 | #define _PAGE_TABLE (_PAGE_PRESENT | _PAGE_RW | _PAGE_USER | _PAGE_ACCESSED | _PAGE_DIRTY) | 56 | #define _PAGE_TABLE (_PAGE_PRESENT | _PAGE_RW | _PAGE_USER | \ |
| 56 | #define _KERNPG_TABLE (_PAGE_PRESENT | _PAGE_RW | _PAGE_ACCESSED | _PAGE_DIRTY) | 57 | _PAGE_ACCESSED | _PAGE_DIRTY) |
| 58 | #define _KERNPG_TABLE (_PAGE_PRESENT | _PAGE_RW | _PAGE_ACCESSED | \ | ||
| 59 | _PAGE_DIRTY) | ||
| 57 | 60 | ||
| 58 | #define _PAGE_CHG_MASK (PTE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY) | 61 | #define _PAGE_CHG_MASK (PTE_MASK | _PAGE_ACCESSED | _PAGE_DIRTY) |
| 59 | 62 | ||
| 60 | #define PAGE_NONE __pgprot(_PAGE_PROTNONE | _PAGE_ACCESSED) | 63 | #define _PAGE_CACHE_MASK (_PAGE_PCD | _PAGE_PWT) |
| 61 | #define PAGE_SHARED __pgprot(_PAGE_PRESENT | _PAGE_RW | _PAGE_USER | _PAGE_ACCESSED | _PAGE_NX) | 64 | #define _PAGE_CACHE_WB (0) |
| 65 | #define _PAGE_CACHE_WC (_PAGE_PWT) | ||
| 66 | #define _PAGE_CACHE_UC_MINUS (_PAGE_PCD) | ||
| 67 | #define _PAGE_CACHE_UC (_PAGE_PCD | _PAGE_PWT) | ||
| 62 | 68 | ||
| 63 | #define PAGE_SHARED_EXEC __pgprot(_PAGE_PRESENT | _PAGE_RW | _PAGE_USER | _PAGE_ACCESSED) | 69 | #define PAGE_NONE __pgprot(_PAGE_PROTNONE | _PAGE_ACCESSED) |
| 64 | #define PAGE_COPY_NOEXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_ACCESSED | _PAGE_NX) | 70 | #define PAGE_SHARED __pgprot(_PAGE_PRESENT | _PAGE_RW | _PAGE_USER | \ |
| 65 | #define PAGE_COPY_EXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_ACCESSED) | 71 | _PAGE_ACCESSED | _PAGE_NX) |
| 72 | |||
| 73 | #define PAGE_SHARED_EXEC __pgprot(_PAGE_PRESENT | _PAGE_RW | \ | ||
| 74 | _PAGE_USER | _PAGE_ACCESSED) | ||
| 75 | #define PAGE_COPY_NOEXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | \ | ||
| 76 | _PAGE_ACCESSED | _PAGE_NX) | ||
| 77 | #define PAGE_COPY_EXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | \ | ||
| 78 | _PAGE_ACCESSED) | ||
| 66 | #define PAGE_COPY PAGE_COPY_NOEXEC | 79 | #define PAGE_COPY PAGE_COPY_NOEXEC |
| 67 | #define PAGE_READONLY __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_ACCESSED | _PAGE_NX) | 80 | #define PAGE_READONLY __pgprot(_PAGE_PRESENT | _PAGE_USER | \ |
| 68 | #define PAGE_READONLY_EXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | _PAGE_ACCESSED) | 81 | _PAGE_ACCESSED | _PAGE_NX) |
| 82 | #define PAGE_READONLY_EXEC __pgprot(_PAGE_PRESENT | _PAGE_USER | \ | ||
| 83 | _PAGE_ACCESSED) | ||
| 69 | 84 | ||
| 70 | #ifdef CONFIG_X86_32 | 85 | #ifdef CONFIG_X86_32 |
| 71 | #define _PAGE_KERNEL_EXEC \ | 86 | #define _PAGE_KERNEL_EXEC \ |
| @@ -84,6 +99,7 @@ extern pteval_t __PAGE_KERNEL, __PAGE_KERNEL_EXEC; | |||
| 84 | #define __PAGE_KERNEL_RO (__PAGE_KERNEL & ~_PAGE_RW) | 99 | #define __PAGE_KERNEL_RO (__PAGE_KERNEL & ~_PAGE_RW) |
| 85 | #define __PAGE_KERNEL_RX (__PAGE_KERNEL_EXEC & ~_PAGE_RW) | 100 | #define __PAGE_KERNEL_RX (__PAGE_KERNEL_EXEC & ~_PAGE_RW) |
| 86 | #define __PAGE_KERNEL_EXEC_NOCACHE (__PAGE_KERNEL_EXEC | _PAGE_PCD | _PAGE_PWT) | 101 | #define __PAGE_KERNEL_EXEC_NOCACHE (__PAGE_KERNEL_EXEC | _PAGE_PCD | _PAGE_PWT) |
| 102 | #define __PAGE_KERNEL_WC (__PAGE_KERNEL | _PAGE_CACHE_WC) | ||
| 87 | #define __PAGE_KERNEL_NOCACHE (__PAGE_KERNEL | _PAGE_PCD | _PAGE_PWT) | 103 | #define __PAGE_KERNEL_NOCACHE (__PAGE_KERNEL | _PAGE_PCD | _PAGE_PWT) |
| 88 | #define __PAGE_KERNEL_UC_MINUS (__PAGE_KERNEL | _PAGE_PCD) | 104 | #define __PAGE_KERNEL_UC_MINUS (__PAGE_KERNEL | _PAGE_PCD) |
| 89 | #define __PAGE_KERNEL_VSYSCALL (__PAGE_KERNEL_RX | _PAGE_USER) | 105 | #define __PAGE_KERNEL_VSYSCALL (__PAGE_KERNEL_RX | _PAGE_USER) |
| @@ -101,6 +117,7 @@ extern pteval_t __PAGE_KERNEL, __PAGE_KERNEL_EXEC; | |||
| 101 | #define PAGE_KERNEL_RO MAKE_GLOBAL(__PAGE_KERNEL_RO) | 117 | #define PAGE_KERNEL_RO MAKE_GLOBAL(__PAGE_KERNEL_RO) |
| 102 | #define PAGE_KERNEL_EXEC MAKE_GLOBAL(__PAGE_KERNEL_EXEC) | 118 | #define PAGE_KERNEL_EXEC MAKE_GLOBAL(__PAGE_KERNEL_EXEC) |
| 103 | #define PAGE_KERNEL_RX MAKE_GLOBAL(__PAGE_KERNEL_RX) | 119 | #define PAGE_KERNEL_RX MAKE_GLOBAL(__PAGE_KERNEL_RX) |
| 120 | #define PAGE_KERNEL_WC MAKE_GLOBAL(__PAGE_KERNEL_WC) | ||
| 104 | #define PAGE_KERNEL_NOCACHE MAKE_GLOBAL(__PAGE_KERNEL_NOCACHE) | 121 | #define PAGE_KERNEL_NOCACHE MAKE_GLOBAL(__PAGE_KERNEL_NOCACHE) |
| 105 | #define PAGE_KERNEL_UC_MINUS MAKE_GLOBAL(__PAGE_KERNEL_UC_MINUS) | 122 | #define PAGE_KERNEL_UC_MINUS MAKE_GLOBAL(__PAGE_KERNEL_UC_MINUS) |
| 106 | #define PAGE_KERNEL_EXEC_NOCACHE MAKE_GLOBAL(__PAGE_KERNEL_EXEC_NOCACHE) | 123 | #define PAGE_KERNEL_EXEC_NOCACHE MAKE_GLOBAL(__PAGE_KERNEL_EXEC_NOCACHE) |
| @@ -134,7 +151,7 @@ extern pteval_t __PAGE_KERNEL, __PAGE_KERNEL_EXEC; | |||
| 134 | * ZERO_PAGE is a global shared page that is always zero: used | 151 | * ZERO_PAGE is a global shared page that is always zero: used |
| 135 | * for zero-mapped memory areas etc.. | 152 | * for zero-mapped memory areas etc.. |
| 136 | */ | 153 | */ |
| 137 | extern unsigned long empty_zero_page[PAGE_SIZE/sizeof(unsigned long)]; | 154 | extern unsigned long empty_zero_page[PAGE_SIZE / sizeof(unsigned long)]; |
| 138 | #define ZERO_PAGE(vaddr) (virt_to_page(empty_zero_page)) | 155 | #define ZERO_PAGE(vaddr) (virt_to_page(empty_zero_page)) |
| 139 | 156 | ||
| 140 | extern spinlock_t pgd_lock; | 157 | extern spinlock_t pgd_lock; |
| @@ -144,30 +161,101 @@ extern struct list_head pgd_list; | |||
| 144 | * The following only work if pte_present() is true. | 161 | * The following only work if pte_present() is true. |
| 145 | * Undefined behaviour if not.. | 162 | * Undefined behaviour if not.. |
| 146 | */ | 163 | */ |
| 147 | static inline int pte_dirty(pte_t pte) { return pte_val(pte) & _PAGE_DIRTY; } | 164 | static inline int pte_dirty(pte_t pte) |
| 148 | static inline int pte_young(pte_t pte) { return pte_val(pte) & _PAGE_ACCESSED; } | 165 | { |
| 149 | static inline int pte_write(pte_t pte) { return pte_val(pte) & _PAGE_RW; } | 166 | return pte_val(pte) & _PAGE_DIRTY; |
| 150 | static inline int pte_file(pte_t pte) { return pte_val(pte) & _PAGE_FILE; } | 167 | } |
| 151 | static inline int pte_huge(pte_t pte) { return pte_val(pte) & _PAGE_PSE; } | 168 | |
| 152 | static inline int pte_global(pte_t pte) { return pte_val(pte) & _PAGE_GLOBAL; } | 169 | static inline int pte_young(pte_t pte) |
| 153 | static inline int pte_exec(pte_t pte) { return !(pte_val(pte) & _PAGE_NX); } | 170 | { |
| 154 | 171 | return pte_val(pte) & _PAGE_ACCESSED; | |
| 155 | static inline int pmd_large(pmd_t pte) { | 172 | } |
| 156 | return (pmd_val(pte) & (_PAGE_PSE|_PAGE_PRESENT)) == | 173 | |
| 157 | (_PAGE_PSE|_PAGE_PRESENT); | 174 | static inline int pte_write(pte_t pte) |
| 175 | { | ||
| 176 | return pte_val(pte) & _PAGE_RW; | ||
| 177 | } | ||
| 178 | |||
| 179 | static inline int pte_file(pte_t pte) | ||
| 180 | { | ||
| 181 | return pte_val(pte) & _PAGE_FILE; | ||
| 182 | } | ||
| 183 | |||
| 184 | static inline int pte_huge(pte_t pte) | ||
| 185 | { | ||
| 186 | return pte_val(pte) & _PAGE_PSE; | ||
| 187 | } | ||
| 188 | |||
| 189 | static inline int pte_global(pte_t pte) | ||
| 190 | { | ||
| 191 | return pte_val(pte) & _PAGE_GLOBAL; | ||
| 192 | } | ||
| 193 | |||
| 194 | static inline int pte_exec(pte_t pte) | ||
| 195 | { | ||
| 196 | return !(pte_val(pte) & _PAGE_NX); | ||
| 197 | } | ||
| 198 | |||
| 199 | static inline int pmd_large(pmd_t pte) | ||
| 200 | { | ||
| 201 | return (pmd_val(pte) & (_PAGE_PSE | _PAGE_PRESENT)) == | ||
| 202 | (_PAGE_PSE | _PAGE_PRESENT); | ||
| 203 | } | ||
| 204 | |||
| 205 | static inline pte_t pte_mkclean(pte_t pte) | ||
| 206 | { | ||
| 207 | return __pte(pte_val(pte) & ~(pteval_t)_PAGE_DIRTY); | ||
| 208 | } | ||
| 209 | |||
| 210 | static inline pte_t pte_mkold(pte_t pte) | ||
| 211 | { | ||
| 212 | return __pte(pte_val(pte) & ~(pteval_t)_PAGE_ACCESSED); | ||
| 213 | } | ||
| 214 | |||
| 215 | static inline pte_t pte_wrprotect(pte_t pte) | ||
| 216 | { | ||
| 217 | return __pte(pte_val(pte) & ~(pteval_t)_PAGE_RW); | ||
| 218 | } | ||
| 219 | |||
| 220 | static inline pte_t pte_mkexec(pte_t pte) | ||
| 221 | { | ||
| 222 | return __pte(pte_val(pte) & ~(pteval_t)_PAGE_NX); | ||
| 223 | } | ||
| 224 | |||
| 225 | static inline pte_t pte_mkdirty(pte_t pte) | ||
| 226 | { | ||
| 227 | return __pte(pte_val(pte) | _PAGE_DIRTY); | ||
| 228 | } | ||
| 229 | |||
| 230 | static inline pte_t pte_mkyoung(pte_t pte) | ||
| 231 | { | ||
| 232 | return __pte(pte_val(pte) | _PAGE_ACCESSED); | ||
| 158 | } | 233 | } |
| 159 | 234 | ||
| 160 | static inline pte_t pte_mkclean(pte_t pte) { return __pte(pte_val(pte) & ~(pteval_t)_PAGE_DIRTY); } | 235 | static inline pte_t pte_mkwrite(pte_t pte) |
| 161 | static inline pte_t pte_mkold(pte_t pte) { return __pte(pte_val(pte) & ~(pteval_t)_PAGE_ACCESSED); } | 236 | { |
| 162 | static inline pte_t pte_wrprotect(pte_t pte) { return __pte(pte_val(pte) & ~(pteval_t)_PAGE_RW); } | 237 | return __pte(pte_val(pte) | _PAGE_RW); |
| 163 | static inline pte_t pte_mkexec(pte_t pte) { return __pte(pte_val(pte) & ~(pteval_t)_PAGE_NX); } | 238 | } |
| 164 | static inline pte_t pte_mkdirty(pte_t pte) { return __pte(pte_val(pte) | _PAGE_DIRTY); } | 239 | |
| 165 | static inline pte_t pte_mkyoung(pte_t pte) { return __pte(pte_val(pte) | _PAGE_ACCESSED); } | 240 | static inline pte_t pte_mkhuge(pte_t pte) |
| 166 | static inline pte_t pte_mkwrite(pte_t pte) { return __pte(pte_val(pte) | _PAGE_RW); } | 241 | { |
| 167 | static inline pte_t pte_mkhuge(pte_t pte) { return __pte(pte_val(pte) | _PAGE_PSE); } | 242 | return __pte(pte_val(pte) | _PAGE_PSE); |
| 168 | static inline pte_t pte_clrhuge(pte_t pte) { return __pte(pte_val(pte) & ~(pteval_t)_PAGE_PSE); } | 243 | } |
| 169 | static inline pte_t pte_mkglobal(pte_t pte) { return __pte(pte_val(pte) | _PAGE_GLOBAL); } | 244 | |
| 170 | static inline pte_t pte_clrglobal(pte_t pte) { return __pte(pte_val(pte) & ~(pteval_t)_PAGE_GLOBAL); } | 245 | static inline pte_t pte_clrhuge(pte_t pte) |
| 246 | { | ||
| 247 | return __pte(pte_val(pte) & ~(pteval_t)_PAGE_PSE); | ||
| 248 | } | ||
| 249 | |||
| 250 | static inline pte_t pte_mkglobal(pte_t pte) | ||
| 251 | { | ||
| 252 | return __pte(pte_val(pte) | _PAGE_GLOBAL); | ||
| 253 | } | ||
| 254 | |||
| 255 | static inline pte_t pte_clrglobal(pte_t pte) | ||
| 256 | { | ||
| 257 | return __pte(pte_val(pte) & ~(pteval_t)_PAGE_GLOBAL); | ||
| 258 | } | ||
| 171 | 259 | ||
| 172 | extern pteval_t __supported_pte_mask; | 260 | extern pteval_t __supported_pte_mask; |
| 173 | 261 | ||
| @@ -334,7 +422,8 @@ static inline void native_set_pte_at(struct mm_struct *mm, unsigned long addr, | |||
| 334 | }) | 422 | }) |
| 335 | 423 | ||
| 336 | #define __HAVE_ARCH_PTEP_GET_AND_CLEAR | 424 | #define __HAVE_ARCH_PTEP_GET_AND_CLEAR |
| 337 | static inline pte_t ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, pte_t *ptep) | 425 | static inline pte_t ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, |
| 426 | pte_t *ptep) | ||
| 338 | { | 427 | { |
| 339 | pte_t pte = native_ptep_get_and_clear(ptep); | 428 | pte_t pte = native_ptep_get_and_clear(ptep); |
| 340 | pte_update(mm, addr, ptep); | 429 | pte_update(mm, addr, ptep); |
| @@ -342,7 +431,9 @@ static inline pte_t ptep_get_and_clear(struct mm_struct *mm, unsigned long addr, | |||
| 342 | } | 431 | } |
| 343 | 432 | ||
| 344 | #define __HAVE_ARCH_PTEP_GET_AND_CLEAR_FULL | 433 | #define __HAVE_ARCH_PTEP_GET_AND_CLEAR_FULL |
| 345 | static inline pte_t ptep_get_and_clear_full(struct mm_struct *mm, unsigned long addr, pte_t *ptep, int full) | 434 | static inline pte_t ptep_get_and_clear_full(struct mm_struct *mm, |
| 435 | unsigned long addr, pte_t *ptep, | ||
| 436 | int full) | ||
| 346 | { | 437 | { |
| 347 | pte_t pte; | 438 | pte_t pte; |
| 348 | if (full) { | 439 | if (full) { |
| @@ -358,7 +449,8 @@ static inline pte_t ptep_get_and_clear_full(struct mm_struct *mm, unsigned long | |||
| 358 | } | 449 | } |
| 359 | 450 | ||
| 360 | #define __HAVE_ARCH_PTEP_SET_WRPROTECT | 451 | #define __HAVE_ARCH_PTEP_SET_WRPROTECT |
| 361 | static inline void ptep_set_wrprotect(struct mm_struct *mm, unsigned long addr, pte_t *ptep) | 452 | static inline void ptep_set_wrprotect(struct mm_struct *mm, |
| 453 | unsigned long addr, pte_t *ptep) | ||
| 362 | { | 454 | { |
| 363 | clear_bit(_PAGE_BIT_RW, (unsigned long *)&ptep->pte); | 455 | clear_bit(_PAGE_BIT_RW, (unsigned long *)&ptep->pte); |
| 364 | pte_update(mm, addr, ptep); | 456 | pte_update(mm, addr, ptep); |
diff --git a/include/asm-x86/pgtable_32.h b/include/asm-x86/pgtable_32.h index 4e6a0fca0b47..c4a643674458 100644 --- a/include/asm-x86/pgtable_32.h +++ b/include/asm-x86/pgtable_32.h | |||
| @@ -40,13 +40,13 @@ void paging_init(void); | |||
| 40 | #ifdef CONFIG_X86_PAE | 40 | #ifdef CONFIG_X86_PAE |
| 41 | # include <asm/pgtable-3level-defs.h> | 41 | # include <asm/pgtable-3level-defs.h> |
| 42 | # define PMD_SIZE (1UL << PMD_SHIFT) | 42 | # define PMD_SIZE (1UL << PMD_SHIFT) |
| 43 | # define PMD_MASK (~(PMD_SIZE-1)) | 43 | # define PMD_MASK (~(PMD_SIZE - 1)) |
| 44 | #else | 44 | #else |
| 45 | # include <asm/pgtable-2level-defs.h> | 45 | # include <asm/pgtable-2level-defs.h> |
| 46 | #endif | 46 | #endif |
| 47 | 47 | ||
| 48 | #define PGDIR_SIZE (1UL << PGDIR_SHIFT) | 48 | #define PGDIR_SIZE (1UL << PGDIR_SHIFT) |
| 49 | #define PGDIR_MASK (~(PGDIR_SIZE-1)) | 49 | #define PGDIR_MASK (~(PGDIR_SIZE - 1)) |
| 50 | 50 | ||
| 51 | #define USER_PGD_PTRS (PAGE_OFFSET >> PGDIR_SHIFT) | 51 | #define USER_PGD_PTRS (PAGE_OFFSET >> PGDIR_SHIFT) |
| 52 | #define KERNEL_PGD_PTRS (PTRS_PER_PGD-USER_PGD_PTRS) | 52 | #define KERNEL_PGD_PTRS (PTRS_PER_PGD-USER_PGD_PTRS) |
| @@ -58,21 +58,22 @@ void paging_init(void); | |||
| 58 | * The vmalloc() routines leaves a hole of 4kB between each vmalloced | 58 | * The vmalloc() routines leaves a hole of 4kB between each vmalloced |
| 59 | * area for the same reason. ;) | 59 | * area for the same reason. ;) |
| 60 | */ | 60 | */ |
| 61 | #define VMALLOC_OFFSET (8*1024*1024) | 61 | #define VMALLOC_OFFSET (8 * 1024 * 1024) |
| 62 | #define VMALLOC_START (((unsigned long) high_memory + \ | 62 | #define VMALLOC_START (((unsigned long)high_memory + 2 * VMALLOC_OFFSET - 1) \ |
| 63 | 2*VMALLOC_OFFSET-1) & ~(VMALLOC_OFFSET-1)) | 63 | & ~(VMALLOC_OFFSET - 1)) |
| 64 | #ifdef CONFIG_X86_PAE | 64 | #ifdef CONFIG_X86_PAE |
| 65 | #define LAST_PKMAP 512 | 65 | #define LAST_PKMAP 512 |
| 66 | #else | 66 | #else |
| 67 | #define LAST_PKMAP 1024 | 67 | #define LAST_PKMAP 1024 |
| 68 | #endif | 68 | #endif |
| 69 | 69 | ||
| 70 | #define PKMAP_BASE ((FIXADDR_BOOT_START - PAGE_SIZE*(LAST_PKMAP + 1)) & PMD_MASK) | 70 | #define PKMAP_BASE ((FIXADDR_BOOT_START - PAGE_SIZE * (LAST_PKMAP + 1)) \ |
| 71 | & PMD_MASK) | ||
| 71 | 72 | ||
| 72 | #ifdef CONFIG_HIGHMEM | 73 | #ifdef CONFIG_HIGHMEM |
| 73 | # define VMALLOC_END (PKMAP_BASE-2*PAGE_SIZE) | 74 | # define VMALLOC_END (PKMAP_BASE - 2 * PAGE_SIZE) |
| 74 | #else | 75 | #else |
| 75 | # define VMALLOC_END (FIXADDR_START-2*PAGE_SIZE) | 76 | # define VMALLOC_END (FIXADDR_START - 2 * PAGE_SIZE) |
| 76 | #endif | 77 | #endif |
| 77 | 78 | ||
| 78 | /* | 79 | /* |
| @@ -88,10 +89,16 @@ extern unsigned long pg0[]; | |||
| 88 | #define pte_present(x) ((x).pte_low & (_PAGE_PRESENT | _PAGE_PROTNONE)) | 89 | #define pte_present(x) ((x).pte_low & (_PAGE_PRESENT | _PAGE_PROTNONE)) |
| 89 | 90 | ||
| 90 | /* To avoid harmful races, pmd_none(x) should check only the lower when PAE */ | 91 | /* To avoid harmful races, pmd_none(x) should check only the lower when PAE */ |
| 91 | #define pmd_none(x) (!(unsigned long)pmd_val(x)) | 92 | #define pmd_none(x) (!(unsigned long)pmd_val((x))) |
| 92 | #define pmd_present(x) (pmd_val(x) & _PAGE_PRESENT) | 93 | #define pmd_present(x) (pmd_val((x)) & _PAGE_PRESENT) |
| 93 | #define pmd_bad(x) ((pmd_val(x) & (~PAGE_MASK & ~_PAGE_USER)) != _KERNPG_TABLE) | ||
| 94 | 94 | ||
| 95 | extern int pmd_bad(pmd_t pmd); | ||
| 96 | |||
| 97 | #define pmd_bad_v1(x) \ | ||
| 98 | (_KERNPG_TABLE != (pmd_val((x)) & ~(PAGE_MASK | _PAGE_USER))) | ||
| 99 | #define pmd_bad_v2(x) \ | ||
| 100 | (_KERNPG_TABLE != (pmd_val((x)) & ~(PAGE_MASK | _PAGE_USER | \ | ||
| 101 | _PAGE_PSE | _PAGE_NX))) | ||
| 95 | 102 | ||
| 96 | #define pages_to_mb(x) ((x) >> (20-PAGE_SHIFT)) | 103 | #define pages_to_mb(x) ((x) >> (20-PAGE_SHIFT)) |
| 97 | 104 | ||
| @@ -117,17 +124,18 @@ static inline void clone_pgd_range(pgd_t *dst, pgd_t *src, int count) | |||
| 117 | } | 124 | } |
| 118 | 125 | ||
| 119 | /* | 126 | /* |
| 120 | * Macro to mark a page protection value as "uncacheable". On processors which do not support | 127 | * Macro to mark a page protection value as "uncacheable". |
| 121 | * it, this is a no-op. | 128 | * On processors which do not support it, this is a no-op. |
| 122 | */ | 129 | */ |
| 123 | #define pgprot_noncached(prot) ((boot_cpu_data.x86 > 3) \ | 130 | #define pgprot_noncached(prot) \ |
| 124 | ? (__pgprot(pgprot_val(prot) | _PAGE_PCD | _PAGE_PWT)) : (prot)) | 131 | ((boot_cpu_data.x86 > 3) \ |
| 132 | ? (__pgprot(pgprot_val(prot) | _PAGE_PCD | _PAGE_PWT)) \ | ||
| 133 | : (prot)) | ||
| 125 | 134 | ||
| 126 | /* | 135 | /* |
| 127 | * Conversion functions: convert a page and protection to a page entry, | 136 | * Conversion functions: convert a page and protection to a page entry, |
| 128 | * and a page entry and page directory to the page they refer to. | 137 | * and a page entry and page directory to the page they refer to. |
| 129 | */ | 138 | */ |
| 130 | |||
| 131 | #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot)) | 139 | #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot)) |
| 132 | 140 | ||
| 133 | /* | 141 | /* |
| @@ -136,20 +144,20 @@ static inline void clone_pgd_range(pgd_t *dst, pgd_t *src, int count) | |||
| 136 | * this macro returns the index of the entry in the pgd page which would | 144 | * this macro returns the index of the entry in the pgd page which would |
| 137 | * control the given virtual address | 145 | * control the given virtual address |
| 138 | */ | 146 | */ |
| 139 | #define pgd_index(address) (((address) >> PGDIR_SHIFT) & (PTRS_PER_PGD-1)) | 147 | #define pgd_index(address) (((address) >> PGDIR_SHIFT) & (PTRS_PER_PGD - 1)) |
| 140 | #define pgd_index_k(addr) pgd_index(addr) | 148 | #define pgd_index_k(addr) pgd_index((addr)) |
| 141 | 149 | ||
| 142 | /* | 150 | /* |
| 143 | * pgd_offset() returns a (pgd_t *) | 151 | * pgd_offset() returns a (pgd_t *) |
| 144 | * pgd_index() is used get the offset into the pgd page's array of pgd_t's; | 152 | * pgd_index() is used get the offset into the pgd page's array of pgd_t's; |
| 145 | */ | 153 | */ |
| 146 | #define pgd_offset(mm, address) ((mm)->pgd+pgd_index(address)) | 154 | #define pgd_offset(mm, address) ((mm)->pgd + pgd_index((address))) |
| 147 | 155 | ||
| 148 | /* | 156 | /* |
| 149 | * a shortcut which implies the use of the kernel's pgd, instead | 157 | * a shortcut which implies the use of the kernel's pgd, instead |
| 150 | * of a process's | 158 | * of a process's |
| 151 | */ | 159 | */ |
| 152 | #define pgd_offset_k(address) pgd_offset(&init_mm, address) | 160 | #define pgd_offset_k(address) pgd_offset(&init_mm, (address)) |
| 153 | 161 | ||
| 154 | static inline int pud_large(pud_t pud) { return 0; } | 162 | static inline int pud_large(pud_t pud) { return 0; } |
| 155 | 163 | ||
| @@ -159,8 +167,8 @@ static inline int pud_large(pud_t pud) { return 0; } | |||
| 159 | * this macro returns the index of the entry in the pmd page which would | 167 | * this macro returns the index of the entry in the pmd page which would |
| 160 | * control the given virtual address | 168 | * control the given virtual address |
| 161 | */ | 169 | */ |
| 162 | #define pmd_index(address) \ | 170 | #define pmd_index(address) \ |
| 163 | (((address) >> PMD_SHIFT) & (PTRS_PER_PMD-1)) | 171 | (((address) >> PMD_SHIFT) & (PTRS_PER_PMD - 1)) |
| 164 | 172 | ||
| 165 | /* | 173 | /* |
| 166 | * the pte page can be thought of an array like this: pte_t[PTRS_PER_PTE] | 174 | * the pte page can be thought of an array like this: pte_t[PTRS_PER_PTE] |
| @@ -168,43 +176,45 @@ static inline int pud_large(pud_t pud) { return 0; } | |||
| 168 | * this macro returns the index of the entry in the pte page which would | 176 | * this macro returns the index of the entry in the pte page which would |
| 169 | * control the given virtual address | 177 | * control the given virtual address |
| 170 | */ | 178 | */ |
| 171 | #define pte_index(address) \ | 179 | #define pte_index(address) \ |
| 172 | (((address) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)) | 180 | (((address) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)) |
| 173 | #define pte_offset_kernel(dir, address) \ | 181 | #define pte_offset_kernel(dir, address) \ |
| 174 | ((pte_t *) pmd_page_vaddr(*(dir)) + pte_index(address)) | 182 | ((pte_t *)pmd_page_vaddr(*(dir)) + pte_index((address))) |
| 175 | 183 | ||
| 176 | #define pmd_page(pmd) (pfn_to_page(pmd_val(pmd) >> PAGE_SHIFT)) | 184 | #define pmd_page(pmd) (pfn_to_page(pmd_val((pmd)) >> PAGE_SHIFT)) |
| 177 | 185 | ||
| 178 | #define pmd_page_vaddr(pmd) \ | 186 | #define pmd_page_vaddr(pmd) \ |
| 179 | ((unsigned long) __va(pmd_val(pmd) & PAGE_MASK)) | 187 | ((unsigned long)__va(pmd_val((pmd)) & PAGE_MASK)) |
| 180 | 188 | ||
| 181 | #if defined(CONFIG_HIGHPTE) | 189 | #if defined(CONFIG_HIGHPTE) |
| 182 | #define pte_offset_map(dir, address) \ | 190 | #define pte_offset_map(dir, address) \ |
| 183 | ((pte_t *)kmap_atomic_pte(pmd_page(*(dir)),KM_PTE0) + pte_index(address)) | 191 | ((pte_t *)kmap_atomic_pte(pmd_page(*(dir)), KM_PTE0) + \ |
| 184 | #define pte_offset_map_nested(dir, address) \ | 192 | pte_index((address))) |
| 185 | ((pte_t *)kmap_atomic_pte(pmd_page(*(dir)),KM_PTE1) + pte_index(address)) | 193 | #define pte_offset_map_nested(dir, address) \ |
| 186 | #define pte_unmap(pte) kunmap_atomic(pte, KM_PTE0) | 194 | ((pte_t *)kmap_atomic_pte(pmd_page(*(dir)), KM_PTE1) + \ |
| 187 | #define pte_unmap_nested(pte) kunmap_atomic(pte, KM_PTE1) | 195 | pte_index((address))) |
| 196 | #define pte_unmap(pte) kunmap_atomic((pte), KM_PTE0) | ||
| 197 | #define pte_unmap_nested(pte) kunmap_atomic((pte), KM_PTE1) | ||
| 188 | #else | 198 | #else |
| 189 | #define pte_offset_map(dir, address) \ | 199 | #define pte_offset_map(dir, address) \ |
| 190 | ((pte_t *)page_address(pmd_page(*(dir))) + pte_index(address)) | 200 | ((pte_t *)page_address(pmd_page(*(dir))) + pte_index((address))) |
| 191 | #define pte_offset_map_nested(dir, address) pte_offset_map(dir, address) | 201 | #define pte_offset_map_nested(dir, address) pte_offset_map((dir), (address)) |
| 192 | #define pte_unmap(pte) do { } while (0) | 202 | #define pte_unmap(pte) do { } while (0) |
| 193 | #define pte_unmap_nested(pte) do { } while (0) | 203 | #define pte_unmap_nested(pte) do { } while (0) |
| 194 | #endif | 204 | #endif |
| 195 | 205 | ||
| 196 | /* Clear a kernel PTE and flush it from the TLB */ | 206 | /* Clear a kernel PTE and flush it from the TLB */ |
| 197 | #define kpte_clear_flush(ptep, vaddr) \ | 207 | #define kpte_clear_flush(ptep, vaddr) \ |
| 198 | do { \ | 208 | do { \ |
| 199 | pte_clear(&init_mm, vaddr, ptep); \ | 209 | pte_clear(&init_mm, (vaddr), (ptep)); \ |
| 200 | __flush_tlb_one(vaddr); \ | 210 | __flush_tlb_one((vaddr)); \ |
| 201 | } while (0) | 211 | } while (0) |
| 202 | 212 | ||
| 203 | /* | 213 | /* |
| 204 | * The i386 doesn't have any external MMU info: the kernel page | 214 | * The i386 doesn't have any external MMU info: the kernel page |
| 205 | * tables contain all the necessary information. | 215 | * tables contain all the necessary information. |
| 206 | */ | 216 | */ |
| 207 | #define update_mmu_cache(vma,address,pte) do { } while (0) | 217 | #define update_mmu_cache(vma, address, pte) do { } while (0) |
| 208 | 218 | ||
| 209 | void native_pagetable_setup_start(pgd_t *base); | 219 | void native_pagetable_setup_start(pgd_t *base); |
| 210 | void native_pagetable_setup_done(pgd_t *base); | 220 | void native_pagetable_setup_done(pgd_t *base); |
| @@ -233,7 +243,7 @@ static inline void paravirt_pagetable_setup_done(pgd_t *base) | |||
| 233 | #define kern_addr_valid(kaddr) (0) | 243 | #define kern_addr_valid(kaddr) (0) |
| 234 | #endif | 244 | #endif |
| 235 | 245 | ||
| 236 | #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ | 246 | #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ |
| 237 | remap_pfn_range(vma, vaddr, pfn, size, prot) | 247 | remap_pfn_range(vma, vaddr, pfn, size, prot) |
| 238 | 248 | ||
| 239 | #endif /* _I386_PGTABLE_H */ | 249 | #endif /* _I386_PGTABLE_H */ |
diff --git a/include/asm-x86/pgtable_64.h b/include/asm-x86/pgtable_64.h index 0a0b77bc736a..9fd87d0b6477 100644 --- a/include/asm-x86/pgtable_64.h +++ b/include/asm-x86/pgtable_64.h | |||
| @@ -52,14 +52,18 @@ extern void paging_init(void); | |||
| 52 | 52 | ||
| 53 | #ifndef __ASSEMBLY__ | 53 | #ifndef __ASSEMBLY__ |
| 54 | 54 | ||
| 55 | #define pte_ERROR(e) \ | 55 | #define pte_ERROR(e) \ |
| 56 | printk("%s:%d: bad pte %p(%016lx).\n", __FILE__, __LINE__, &(e), pte_val(e)) | 56 | printk("%s:%d: bad pte %p(%016lx).\n", \ |
| 57 | #define pmd_ERROR(e) \ | 57 | __FILE__, __LINE__, &(e), pte_val(e)) |
| 58 | printk("%s:%d: bad pmd %p(%016lx).\n", __FILE__, __LINE__, &(e), pmd_val(e)) | 58 | #define pmd_ERROR(e) \ |
| 59 | #define pud_ERROR(e) \ | 59 | printk("%s:%d: bad pmd %p(%016lx).\n", \ |
| 60 | printk("%s:%d: bad pud %p(%016lx).\n", __FILE__, __LINE__, &(e), pud_val(e)) | 60 | __FILE__, __LINE__, &(e), pmd_val(e)) |
| 61 | #define pgd_ERROR(e) \ | 61 | #define pud_ERROR(e) \ |
| 62 | printk("%s:%d: bad pgd %p(%016lx).\n", __FILE__, __LINE__, &(e), pgd_val(e)) | 62 | printk("%s:%d: bad pud %p(%016lx).\n", \ |
| 63 | __FILE__, __LINE__, &(e), pud_val(e)) | ||
| 64 | #define pgd_ERROR(e) \ | ||
| 65 | printk("%s:%d: bad pgd %p(%016lx).\n", \ | ||
| 66 | __FILE__, __LINE__, &(e), pgd_val(e)) | ||
| 63 | 67 | ||
| 64 | #define pgd_none(x) (!pgd_val(x)) | 68 | #define pgd_none(x) (!pgd_val(x)) |
| 65 | #define pud_none(x) (!pud_val(x)) | 69 | #define pud_none(x) (!pud_val(x)) |
| @@ -87,7 +91,8 @@ static inline pte_t native_ptep_get_and_clear(pte_t *xp) | |||
| 87 | #ifdef CONFIG_SMP | 91 | #ifdef CONFIG_SMP |
| 88 | return native_make_pte(xchg(&xp->pte, 0)); | 92 | return native_make_pte(xchg(&xp->pte, 0)); |
| 89 | #else | 93 | #else |
| 90 | /* native_local_ptep_get_and_clear, but duplicated because of cyclic dependency */ | 94 | /* native_local_ptep_get_and_clear, |
| 95 | but duplicated because of cyclic dependency */ | ||
| 91 | pte_t ret = *xp; | 96 | pte_t ret = *xp; |
| 92 | native_pte_clear(NULL, 0, xp); | 97 | native_pte_clear(NULL, 0, xp); |
| 93 | return ret; | 98 | return ret; |
| @@ -119,7 +124,7 @@ static inline void native_set_pgd(pgd_t *pgdp, pgd_t pgd) | |||
| 119 | *pgdp = pgd; | 124 | *pgdp = pgd; |
| 120 | } | 125 | } |
| 121 | 126 | ||
| 122 | static inline void native_pgd_clear(pgd_t * pgd) | 127 | static inline void native_pgd_clear(pgd_t *pgd) |
| 123 | { | 128 | { |
| 124 | native_set_pgd(pgd, native_make_pgd(0)); | 129 | native_set_pgd(pgd, native_make_pgd(0)); |
| 125 | } | 130 | } |
| @@ -128,19 +133,19 @@ static inline void native_pgd_clear(pgd_t * pgd) | |||
| 128 | 133 | ||
| 129 | #endif /* !__ASSEMBLY__ */ | 134 | #endif /* !__ASSEMBLY__ */ |
| 130 | 135 | ||
| 131 | #define PMD_SIZE (_AC(1,UL) << PMD_SHIFT) | 136 | #define PMD_SIZE (_AC(1, UL) << PMD_SHIFT) |
| 132 | #define PMD_MASK (~(PMD_SIZE-1)) | 137 | #define PMD_MASK (~(PMD_SIZE - 1)) |
| 133 | #define PUD_SIZE (_AC(1,UL) << PUD_SHIFT) | 138 | #define PUD_SIZE (_AC(1, UL) << PUD_SHIFT) |
| 134 | #define PUD_MASK (~(PUD_SIZE-1)) | 139 | #define PUD_MASK (~(PUD_SIZE - 1)) |
| 135 | #define PGDIR_SIZE (_AC(1,UL) << PGDIR_SHIFT) | 140 | #define PGDIR_SIZE (_AC(1, UL) << PGDIR_SHIFT) |
| 136 | #define PGDIR_MASK (~(PGDIR_SIZE-1)) | 141 | #define PGDIR_MASK (~(PGDIR_SIZE - 1)) |
| 137 | 142 | ||
| 138 | 143 | ||
| 139 | #define MAXMEM _AC(0x3fffffffffff, UL) | 144 | #define MAXMEM _AC(0x00003fffffffffff, UL) |
| 140 | #define VMALLOC_START _AC(0xffffc20000000000, UL) | 145 | #define VMALLOC_START _AC(0xffffc20000000000, UL) |
| 141 | #define VMALLOC_END _AC(0xffffe1ffffffffff, UL) | 146 | #define VMALLOC_END _AC(0xffffe1ffffffffff, UL) |
| 142 | #define VMEMMAP_START _AC(0xffffe20000000000, UL) | 147 | #define VMEMMAP_START _AC(0xffffe20000000000, UL) |
| 143 | #define MODULES_VADDR _AC(0xffffffff88000000, UL) | 148 | #define MODULES_VADDR _AC(0xffffffffa0000000, UL) |
| 144 | #define MODULES_END _AC(0xfffffffffff00000, UL) | 149 | #define MODULES_END _AC(0xfffffffffff00000, UL) |
| 145 | #define MODULES_LEN (MODULES_END - MODULES_VADDR) | 150 | #define MODULES_LEN (MODULES_END - MODULES_VADDR) |
| 146 | 151 | ||
| @@ -153,26 +158,28 @@ static inline unsigned long pgd_bad(pgd_t pgd) | |||
| 153 | 158 | ||
| 154 | static inline unsigned long pud_bad(pud_t pud) | 159 | static inline unsigned long pud_bad(pud_t pud) |
| 155 | { | 160 | { |
| 156 | return pud_val(pud) & ~(PTE_MASK | _KERNPG_TABLE | _PAGE_USER); | 161 | return pud_val(pud) & |
| 162 | ~(PTE_MASK | _KERNPG_TABLE | _PAGE_USER | _PAGE_PSE | _PAGE_NX); | ||
| 157 | } | 163 | } |
| 158 | 164 | ||
| 159 | static inline unsigned long pmd_bad(pmd_t pmd) | 165 | static inline unsigned long pmd_bad(pmd_t pmd) |
| 160 | { | 166 | { |
| 161 | return pmd_val(pmd) & ~(PTE_MASK | _KERNPG_TABLE | _PAGE_USER); | 167 | return pmd_val(pmd) & |
| 168 | ~(PTE_MASK | _KERNPG_TABLE | _PAGE_USER | _PAGE_PSE | _PAGE_NX); | ||
| 162 | } | 169 | } |
| 163 | 170 | ||
| 164 | #define pte_none(x) (!pte_val(x)) | 171 | #define pte_none(x) (!pte_val((x))) |
| 165 | #define pte_present(x) (pte_val(x) & (_PAGE_PRESENT | _PAGE_PROTNONE)) | 172 | #define pte_present(x) (pte_val((x)) & (_PAGE_PRESENT | _PAGE_PROTNONE)) |
| 166 | 173 | ||
| 167 | #define pages_to_mb(x) ((x) >> (20-PAGE_SHIFT)) /* FIXME: is this right? */ | 174 | #define pages_to_mb(x) ((x) >> (20 - PAGE_SHIFT)) /* FIXME: is this right? */ |
| 168 | #define pte_page(x) pfn_to_page(pte_pfn(x)) | 175 | #define pte_page(x) pfn_to_page(pte_pfn((x))) |
| 169 | #define pte_pfn(x) ((pte_val(x) & __PHYSICAL_MASK) >> PAGE_SHIFT) | 176 | #define pte_pfn(x) ((pte_val((x)) & __PHYSICAL_MASK) >> PAGE_SHIFT) |
| 170 | 177 | ||
| 171 | /* | 178 | /* |
| 172 | * Macro to mark a page protection value as "uncacheable". | 179 | * Macro to mark a page protection value as "uncacheable". |
| 173 | */ | 180 | */ |
| 174 | #define pgprot_noncached(prot) (__pgprot(pgprot_val(prot) | _PAGE_PCD | _PAGE_PWT)) | 181 | #define pgprot_noncached(prot) \ |
| 175 | 182 | (__pgprot(pgprot_val((prot)) | _PAGE_PCD | _PAGE_PWT)) | |
| 176 | 183 | ||
| 177 | /* | 184 | /* |
| 178 | * Conversion functions: convert a page and protection to a page entry, | 185 | * Conversion functions: convert a page and protection to a page entry, |
| @@ -182,75 +189,81 @@ static inline unsigned long pmd_bad(pmd_t pmd) | |||
| 182 | /* | 189 | /* |
| 183 | * Level 4 access. | 190 | * Level 4 access. |
| 184 | */ | 191 | */ |
| 185 | #define pgd_page_vaddr(pgd) ((unsigned long) __va((unsigned long)pgd_val(pgd) & PTE_MASK)) | 192 | #define pgd_page_vaddr(pgd) \ |
| 186 | #define pgd_page(pgd) (pfn_to_page(pgd_val(pgd) >> PAGE_SHIFT)) | 193 | ((unsigned long)__va((unsigned long)pgd_val((pgd)) & PTE_MASK)) |
| 187 | #define pgd_index(address) (((address) >> PGDIR_SHIFT) & (PTRS_PER_PGD-1)) | 194 | #define pgd_page(pgd) (pfn_to_page(pgd_val((pgd)) >> PAGE_SHIFT)) |
| 188 | #define pgd_offset(mm, addr) ((mm)->pgd + pgd_index(addr)) | 195 | #define pgd_index(address) (((address) >> PGDIR_SHIFT) & (PTRS_PER_PGD - 1)) |
| 189 | #define pgd_offset_k(address) (init_level4_pgt + pgd_index(address)) | 196 | #define pgd_offset(mm, address) ((mm)->pgd + pgd_index((address))) |
| 197 | #define pgd_offset_k(address) (init_level4_pgt + pgd_index((address))) | ||
| 190 | #define pgd_present(pgd) (pgd_val(pgd) & _PAGE_PRESENT) | 198 | #define pgd_present(pgd) (pgd_val(pgd) & _PAGE_PRESENT) |
| 191 | static inline int pgd_large(pgd_t pgd) { return 0; } | 199 | static inline int pgd_large(pgd_t pgd) { return 0; } |
| 192 | #define mk_kernel_pgd(address) ((pgd_t){ (address) | _KERNPG_TABLE }) | 200 | #define mk_kernel_pgd(address) ((pgd_t){ (address) | _KERNPG_TABLE }) |
| 193 | 201 | ||
| 194 | /* PUD - Level3 access */ | 202 | /* PUD - Level3 access */ |
| 195 | /* to find an entry in a page-table-directory. */ | 203 | /* to find an entry in a page-table-directory. */ |
| 196 | #define pud_page_vaddr(pud) ((unsigned long) __va(pud_val(pud) & PHYSICAL_PAGE_MASK)) | 204 | #define pud_page_vaddr(pud) \ |
| 197 | #define pud_page(pud) (pfn_to_page(pud_val(pud) >> PAGE_SHIFT)) | 205 | ((unsigned long)__va(pud_val((pud)) & PHYSICAL_PAGE_MASK)) |
| 198 | #define pud_index(address) (((address) >> PUD_SHIFT) & (PTRS_PER_PUD-1)) | 206 | #define pud_page(pud) (pfn_to_page(pud_val((pud)) >> PAGE_SHIFT)) |
| 199 | #define pud_offset(pgd, address) ((pud_t *) pgd_page_vaddr(*(pgd)) + pud_index(address)) | 207 | #define pud_index(address) (((address) >> PUD_SHIFT) & (PTRS_PER_PUD - 1)) |
| 200 | #define pud_present(pud) (pud_val(pud) & _PAGE_PRESENT) | 208 | #define pud_offset(pgd, address) \ |
| 209 | ((pud_t *)pgd_page_vaddr(*(pgd)) + pud_index((address))) | ||
| 210 | #define pud_present(pud) (pud_val((pud)) & _PAGE_PRESENT) | ||
| 201 | 211 | ||
| 202 | static inline int pud_large(pud_t pte) | 212 | static inline int pud_large(pud_t pte) |
| 203 | { | 213 | { |
| 204 | return (pud_val(pte) & (_PAGE_PSE|_PAGE_PRESENT)) == | 214 | return (pud_val(pte) & (_PAGE_PSE | _PAGE_PRESENT)) == |
| 205 | (_PAGE_PSE|_PAGE_PRESENT); | 215 | (_PAGE_PSE | _PAGE_PRESENT); |
| 206 | } | 216 | } |
| 207 | 217 | ||
| 208 | /* PMD - Level 2 access */ | 218 | /* PMD - Level 2 access */ |
| 209 | #define pmd_page_vaddr(pmd) ((unsigned long) __va(pmd_val(pmd) & PTE_MASK)) | 219 | #define pmd_page_vaddr(pmd) ((unsigned long) __va(pmd_val((pmd)) & PTE_MASK)) |
| 210 | #define pmd_page(pmd) (pfn_to_page(pmd_val(pmd) >> PAGE_SHIFT)) | 220 | #define pmd_page(pmd) (pfn_to_page(pmd_val((pmd)) >> PAGE_SHIFT)) |
| 211 | 221 | ||
| 212 | #define pmd_index(address) (((address) >> PMD_SHIFT) & (PTRS_PER_PMD-1)) | 222 | #define pmd_index(address) (((address) >> PMD_SHIFT) & (PTRS_PER_PMD - 1)) |
| 213 | #define pmd_offset(dir, address) ((pmd_t *) pud_page_vaddr(*(dir)) + \ | 223 | #define pmd_offset(dir, address) ((pmd_t *)pud_page_vaddr(*(dir)) + \ |
| 214 | pmd_index(address)) | 224 | pmd_index(address)) |
| 215 | #define pmd_none(x) (!pmd_val(x)) | 225 | #define pmd_none(x) (!pmd_val((x))) |
| 216 | #define pmd_present(x) (pmd_val(x) & _PAGE_PRESENT) | 226 | #define pmd_present(x) (pmd_val((x)) & _PAGE_PRESENT) |
| 217 | #define pfn_pmd(nr,prot) (__pmd(((nr) << PAGE_SHIFT) | pgprot_val(prot))) | 227 | #define pfn_pmd(nr, prot) (__pmd(((nr) << PAGE_SHIFT) | pgprot_val((prot)))) |
| 218 | #define pmd_pfn(x) ((pmd_val(x) & __PHYSICAL_MASK) >> PAGE_SHIFT) | 228 | #define pmd_pfn(x) ((pmd_val((x)) & __PHYSICAL_MASK) >> PAGE_SHIFT) |
| 219 | 229 | ||
| 220 | #define pte_to_pgoff(pte) ((pte_val(pte) & PHYSICAL_PAGE_MASK) >> PAGE_SHIFT) | 230 | #define pte_to_pgoff(pte) ((pte_val((pte)) & PHYSICAL_PAGE_MASK) >> PAGE_SHIFT) |
| 221 | #define pgoff_to_pte(off) ((pte_t) { .pte = ((off) << PAGE_SHIFT) | _PAGE_FILE }) | 231 | #define pgoff_to_pte(off) ((pte_t) { .pte = ((off) << PAGE_SHIFT) | \ |
| 232 | _PAGE_FILE }) | ||
| 222 | #define PTE_FILE_MAX_BITS __PHYSICAL_MASK_SHIFT | 233 | #define PTE_FILE_MAX_BITS __PHYSICAL_MASK_SHIFT |
| 223 | 234 | ||
| 224 | /* PTE - Level 1 access. */ | 235 | /* PTE - Level 1 access. */ |
| 225 | 236 | ||
| 226 | /* page, protection -> pte */ | 237 | /* page, protection -> pte */ |
| 227 | #define mk_pte(page, pgprot) pfn_pte(page_to_pfn(page), (pgprot)) | 238 | #define mk_pte(page, pgprot) pfn_pte(page_to_pfn((page)), (pgprot)) |
| 228 | 239 | ||
| 229 | #define pte_index(address) \ | 240 | #define pte_index(address) (((address) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)) |
| 230 | (((address) >> PAGE_SHIFT) & (PTRS_PER_PTE - 1)) | ||
| 231 | #define pte_offset_kernel(dir, address) ((pte_t *) pmd_page_vaddr(*(dir)) + \ | 241 | #define pte_offset_kernel(dir, address) ((pte_t *) pmd_page_vaddr(*(dir)) + \ |
| 232 | pte_index(address)) | 242 | pte_index((address))) |
| 233 | 243 | ||
| 234 | /* x86-64 always has all page tables mapped. */ | 244 | /* x86-64 always has all page tables mapped. */ |
| 235 | #define pte_offset_map(dir,address) pte_offset_kernel(dir,address) | 245 | #define pte_offset_map(dir, address) pte_offset_kernel((dir), (address)) |
| 236 | #define pte_offset_map_nested(dir,address) pte_offset_kernel(dir,address) | 246 | #define pte_offset_map_nested(dir, address) pte_offset_kernel((dir), (address)) |
| 237 | #define pte_unmap(pte) /* NOP */ | 247 | #define pte_unmap(pte) /* NOP */ |
| 238 | #define pte_unmap_nested(pte) /* NOP */ | 248 | #define pte_unmap_nested(pte) /* NOP */ |
| 249 | |||
| 250 | #define update_mmu_cache(vma, address, pte) do { } while (0) | ||
| 239 | 251 | ||
| 240 | #define update_mmu_cache(vma,address,pte) do { } while (0) | 252 | extern int direct_gbpages; |
| 241 | 253 | ||
| 242 | /* Encode and de-code a swap entry */ | 254 | /* Encode and de-code a swap entry */ |
| 243 | #define __swp_type(x) (((x).val >> 1) & 0x3f) | 255 | #define __swp_type(x) (((x).val >> 1) & 0x3f) |
| 244 | #define __swp_offset(x) ((x).val >> 8) | 256 | #define __swp_offset(x) ((x).val >> 8) |
| 245 | #define __swp_entry(type, offset) ((swp_entry_t) { ((type) << 1) | ((offset) << 8) }) | 257 | #define __swp_entry(type, offset) ((swp_entry_t) { ((type) << 1) | \ |
| 246 | #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val(pte) }) | 258 | ((offset) << 8) }) |
| 259 | #define __pte_to_swp_entry(pte) ((swp_entry_t) { pte_val((pte)) }) | ||
| 247 | #define __swp_entry_to_pte(x) ((pte_t) { .pte = (x).val }) | 260 | #define __swp_entry_to_pte(x) ((pte_t) { .pte = (x).val }) |
| 248 | 261 | ||
| 249 | extern int kern_addr_valid(unsigned long addr); | 262 | extern int kern_addr_valid(unsigned long addr); |
| 250 | extern void cleanup_highmap(void); | 263 | extern void cleanup_highmap(void); |
| 251 | 264 | ||
| 252 | #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ | 265 | #define io_remap_pfn_range(vma, vaddr, pfn, size, prot) \ |
| 253 | remap_pfn_range(vma, vaddr, pfn, size, prot) | 266 | remap_pfn_range(vma, vaddr, pfn, size, prot) |
| 254 | 267 | ||
| 255 | #define HAVE_ARCH_UNMAPPED_AREA | 268 | #define HAVE_ARCH_UNMAPPED_AREA |
| 256 | #define HAVE_ARCH_UNMAPPED_AREA_TOPDOWN | 269 | #define HAVE_ARCH_UNMAPPED_AREA_TOPDOWN |
| @@ -263,8 +276,10 @@ extern void cleanup_highmap(void); | |||
| 263 | 276 | ||
| 264 | /* fs/proc/kcore.c */ | 277 | /* fs/proc/kcore.c */ |
| 265 | #define kc_vaddr_to_offset(v) ((v) & __VIRTUAL_MASK) | 278 | #define kc_vaddr_to_offset(v) ((v) & __VIRTUAL_MASK) |
| 266 | #define kc_offset_to_vaddr(o) \ | 279 | #define kc_offset_to_vaddr(o) \ |
| 267 | (((o) & (1UL << (__VIRTUAL_MASK_SHIFT-1))) ? ((o) | (~__VIRTUAL_MASK)) : (o)) | 280 | (((o) & (1UL << (__VIRTUAL_MASK_SHIFT - 1))) \ |
| 281 | ? ((o) | ~__VIRTUAL_MASK) \ | ||
| 282 | : (o)) | ||
| 268 | 283 | ||
| 269 | #define __HAVE_ARCH_PTE_SAME | 284 | #define __HAVE_ARCH_PTE_SAME |
| 270 | #endif /* !__ASSEMBLY__ */ | 285 | #endif /* !__ASSEMBLY__ */ |
diff --git a/include/asm-x86/posix_types.h b/include/asm-x86/posix_types.h index bb7133dc155d..fe312a5ba204 100644 --- a/include/asm-x86/posix_types.h +++ b/include/asm-x86/posix_types.h | |||
| @@ -1,11 +1,5 @@ | |||
| 1 | #ifdef __KERNEL__ | 1 | #ifdef __KERNEL__ |
| 2 | # ifdef CONFIG_X86_32 | 2 | # if defined(CONFIG_X86_32) || defined(__i386__) |
| 3 | # include "posix_types_32.h" | ||
| 4 | # else | ||
| 5 | # include "posix_types_64.h" | ||
| 6 | # endif | ||
| 7 | #else | ||
| 8 | # ifdef __i386__ | ||
| 9 | # include "posix_types_32.h" | 3 | # include "posix_types_32.h" |
| 10 | # else | 4 | # else |
| 11 | # include "posix_types_64.h" | 5 | # include "posix_types_64.h" |
diff --git a/include/asm-x86/posix_types_32.h b/include/asm-x86/posix_types_32.h index 015e539cdef5..b031efda37ec 100644 --- a/include/asm-x86/posix_types_32.h +++ b/include/asm-x86/posix_types_32.h | |||
| @@ -45,32 +45,39 @@ typedef struct { | |||
| 45 | #if defined(__KERNEL__) | 45 | #if defined(__KERNEL__) |
| 46 | 46 | ||
| 47 | #undef __FD_SET | 47 | #undef __FD_SET |
| 48 | #define __FD_SET(fd,fdsetp) \ | 48 | #define __FD_SET(fd,fdsetp) \ |
| 49 | __asm__ __volatile__("btsl %1,%0": \ | 49 | asm volatile("btsl %1,%0": \ |
| 50 | "+m" (*(__kernel_fd_set *) (fdsetp)):"r" ((int) (fd))) | 50 | "+m" (*(__kernel_fd_set *)(fdsetp)) \ |
| 51 | : "r" ((int)(fd))) | ||
| 51 | 52 | ||
| 52 | #undef __FD_CLR | 53 | #undef __FD_CLR |
| 53 | #define __FD_CLR(fd,fdsetp) \ | 54 | #define __FD_CLR(fd,fdsetp) \ |
| 54 | __asm__ __volatile__("btrl %1,%0": \ | 55 | asm volatile("btrl %1,%0": \ |
| 55 | "+m" (*(__kernel_fd_set *) (fdsetp)):"r" ((int) (fd))) | 56 | "+m" (*(__kernel_fd_set *)(fdsetp)) \ |
| 57 | : "r" ((int) (fd))) | ||
| 56 | 58 | ||
| 57 | #undef __FD_ISSET | 59 | #undef __FD_ISSET |
| 58 | #define __FD_ISSET(fd,fdsetp) (__extension__ ({ \ | 60 | #define __FD_ISSET(fd,fdsetp) \ |
| 59 | unsigned char __result; \ | 61 | (__extension__ \ |
| 60 | __asm__ __volatile__("btl %1,%2 ; setb %0" \ | 62 | ({ \ |
| 61 | :"=q" (__result) :"r" ((int) (fd)), \ | 63 | unsigned char __result; \ |
| 62 | "m" (*(__kernel_fd_set *) (fdsetp))); \ | 64 | asm volatile("btl %1,%2 ; setb %0" \ |
| 63 | __result; })) | 65 | : "=q" (__result) \ |
| 66 | : "r" ((int)(fd)), \ | ||
| 67 | "m" (*(__kernel_fd_set *)(fdsetp))); \ | ||
| 68 | __result; \ | ||
| 69 | })) | ||
| 64 | 70 | ||
| 65 | #undef __FD_ZERO | 71 | #undef __FD_ZERO |
| 66 | #define __FD_ZERO(fdsetp) \ | 72 | #define __FD_ZERO(fdsetp) \ |
| 67 | do { \ | 73 | do { \ |
| 68 | int __d0, __d1; \ | 74 | int __d0, __d1; \ |
| 69 | __asm__ __volatile__("cld ; rep ; stosl" \ | 75 | asm volatile("cld ; rep ; stosl" \ |
| 70 | :"=m" (*(__kernel_fd_set *) (fdsetp)), \ | 76 | : "=m" (*(__kernel_fd_set *)(fdsetp)), \ |
| 71 | "=&c" (__d0), "=&D" (__d1) \ | 77 | "=&c" (__d0), "=&D" (__d1) \ |
| 72 | :"a" (0), "1" (__FDSET_LONGS), \ | 78 | : "a" (0), "1" (__FDSET_LONGS), \ |
| 73 | "2" ((__kernel_fd_set *) (fdsetp)) : "memory"); \ | 79 | "2" ((__kernel_fd_set *)(fdsetp)) \ |
| 80 | : "memory"); \ | ||
| 74 | } while (0) | 81 | } while (0) |
| 75 | 82 | ||
| 76 | #endif /* defined(__KERNEL__) */ | 83 | #endif /* defined(__KERNEL__) */ |
diff --git a/include/asm-x86/posix_types_64.h b/include/asm-x86/posix_types_64.h index 9926aa43775b..d6624c95854a 100644 --- a/include/asm-x86/posix_types_64.h +++ b/include/asm-x86/posix_types_64.h | |||
| @@ -46,7 +46,7 @@ typedef unsigned long __kernel_old_dev_t; | |||
| 46 | #ifdef __KERNEL__ | 46 | #ifdef __KERNEL__ |
| 47 | 47 | ||
| 48 | #undef __FD_SET | 48 | #undef __FD_SET |
| 49 | static __inline__ void __FD_SET(unsigned long fd, __kernel_fd_set *fdsetp) | 49 | static inline void __FD_SET(unsigned long fd, __kernel_fd_set *fdsetp) |
| 50 | { | 50 | { |
| 51 | unsigned long _tmp = fd / __NFDBITS; | 51 | unsigned long _tmp = fd / __NFDBITS; |
| 52 | unsigned long _rem = fd % __NFDBITS; | 52 | unsigned long _rem = fd % __NFDBITS; |
| @@ -54,7 +54,7 @@ static __inline__ void __FD_SET(unsigned long fd, __kernel_fd_set *fdsetp) | |||
| 54 | } | 54 | } |
| 55 | 55 | ||
| 56 | #undef __FD_CLR | 56 | #undef __FD_CLR |
| 57 | static __inline__ void __FD_CLR(unsigned long fd, __kernel_fd_set *fdsetp) | 57 | static inline void __FD_CLR(unsigned long fd, __kernel_fd_set *fdsetp) |
| 58 | { | 58 | { |
| 59 | unsigned long _tmp = fd / __NFDBITS; | 59 | unsigned long _tmp = fd / __NFDBITS; |
| 60 | unsigned long _rem = fd % __NFDBITS; | 60 | unsigned long _rem = fd % __NFDBITS; |
| @@ -62,7 +62,7 @@ static __inline__ void __FD_CLR(unsigned long fd, __kernel_fd_set *fdsetp) | |||
| 62 | } | 62 | } |
| 63 | 63 | ||
| 64 | #undef __FD_ISSET | 64 | #undef __FD_ISSET |
| 65 | static __inline__ int __FD_ISSET(unsigned long fd, __const__ __kernel_fd_set *p) | 65 | static inline int __FD_ISSET(unsigned long fd, __const__ __kernel_fd_set *p) |
| 66 | { | 66 | { |
| 67 | unsigned long _tmp = fd / __NFDBITS; | 67 | unsigned long _tmp = fd / __NFDBITS; |
| 68 | unsigned long _rem = fd % __NFDBITS; | 68 | unsigned long _rem = fd % __NFDBITS; |
| @@ -74,36 +74,36 @@ static __inline__ int __FD_ISSET(unsigned long fd, __const__ __kernel_fd_set *p) | |||
| 74 | * for 256 and 1024-bit fd_sets respectively) | 74 | * for 256 and 1024-bit fd_sets respectively) |
| 75 | */ | 75 | */ |
| 76 | #undef __FD_ZERO | 76 | #undef __FD_ZERO |
| 77 | static __inline__ void __FD_ZERO(__kernel_fd_set *p) | 77 | static inline void __FD_ZERO(__kernel_fd_set *p) |
| 78 | { | 78 | { |
| 79 | unsigned long *tmp = p->fds_bits; | 79 | unsigned long *tmp = p->fds_bits; |
| 80 | int i; | 80 | int i; |
| 81 | 81 | ||
| 82 | if (__builtin_constant_p(__FDSET_LONGS)) { | 82 | if (__builtin_constant_p(__FDSET_LONGS)) { |
| 83 | switch (__FDSET_LONGS) { | 83 | switch (__FDSET_LONGS) { |
| 84 | case 32: | 84 | case 32: |
| 85 | tmp[ 0] = 0; tmp[ 1] = 0; tmp[ 2] = 0; tmp[ 3] = 0; | 85 | tmp[ 0] = 0; tmp[ 1] = 0; tmp[ 2] = 0; tmp[ 3] = 0; |
| 86 | tmp[ 4] = 0; tmp[ 5] = 0; tmp[ 6] = 0; tmp[ 7] = 0; | 86 | tmp[ 4] = 0; tmp[ 5] = 0; tmp[ 6] = 0; tmp[ 7] = 0; |
| 87 | tmp[ 8] = 0; tmp[ 9] = 0; tmp[10] = 0; tmp[11] = 0; | 87 | tmp[ 8] = 0; tmp[ 9] = 0; tmp[10] = 0; tmp[11] = 0; |
| 88 | tmp[12] = 0; tmp[13] = 0; tmp[14] = 0; tmp[15] = 0; | 88 | tmp[12] = 0; tmp[13] = 0; tmp[14] = 0; tmp[15] = 0; |
| 89 | tmp[16] = 0; tmp[17] = 0; tmp[18] = 0; tmp[19] = 0; | 89 | tmp[16] = 0; tmp[17] = 0; tmp[18] = 0; tmp[19] = 0; |
| 90 | tmp[20] = 0; tmp[21] = 0; tmp[22] = 0; tmp[23] = 0; | 90 | tmp[20] = 0; tmp[21] = 0; tmp[22] = 0; tmp[23] = 0; |
| 91 | tmp[24] = 0; tmp[25] = 0; tmp[26] = 0; tmp[27] = 0; | 91 | tmp[24] = 0; tmp[25] = 0; tmp[26] = 0; tmp[27] = 0; |
| 92 | tmp[28] = 0; tmp[29] = 0; tmp[30] = 0; tmp[31] = 0; | 92 | tmp[28] = 0; tmp[29] = 0; tmp[30] = 0; tmp[31] = 0; |
| 93 | return; | 93 | return; |
| 94 | case 16: | 94 | case 16: |
| 95 | tmp[ 0] = 0; tmp[ 1] = 0; tmp[ 2] = 0; tmp[ 3] = 0; | 95 | tmp[ 0] = 0; tmp[ 1] = 0; tmp[ 2] = 0; tmp[ 3] = 0; |
| 96 | tmp[ 4] = 0; tmp[ 5] = 0; tmp[ 6] = 0; tmp[ 7] = 0; | 96 | tmp[ 4] = 0; tmp[ 5] = 0; tmp[ 6] = 0; tmp[ 7] = 0; |
| 97 | tmp[ 8] = 0; tmp[ 9] = 0; tmp[10] = 0; tmp[11] = 0; | 97 | tmp[ 8] = 0; tmp[ 9] = 0; tmp[10] = 0; tmp[11] = 0; |
| 98 | tmp[12] = 0; tmp[13] = 0; tmp[14] = 0; tmp[15] = 0; | 98 | tmp[12] = 0; tmp[13] = 0; tmp[14] = 0; tmp[15] = 0; |
| 99 | return; | 99 | return; |
| 100 | case 8: | 100 | case 8: |
| 101 | tmp[ 0] = 0; tmp[ 1] = 0; tmp[ 2] = 0; tmp[ 3] = 0; | 101 | tmp[ 0] = 0; tmp[ 1] = 0; tmp[ 2] = 0; tmp[ 3] = 0; |
| 102 | tmp[ 4] = 0; tmp[ 5] = 0; tmp[ 6] = 0; tmp[ 7] = 0; | 102 | tmp[ 4] = 0; tmp[ 5] = 0; tmp[ 6] = 0; tmp[ 7] = 0; |
| 103 | return; | 103 | return; |
| 104 | case 4: | 104 | case 4: |
| 105 | tmp[ 0] = 0; tmp[ 1] = 0; tmp[ 2] = 0; tmp[ 3] = 0; | 105 | tmp[ 0] = 0; tmp[ 1] = 0; tmp[ 2] = 0; tmp[ 3] = 0; |
| 106 | return; | 106 | return; |
| 107 | } | 107 | } |
| 108 | } | 108 | } |
| 109 | i = __FDSET_LONGS; | 109 | i = __FDSET_LONGS; |
diff --git a/include/asm-x86/processor.h b/include/asm-x86/processor.h index 45a2f0ab33d0..6e26c7c717a2 100644 --- a/include/asm-x86/processor.h +++ b/include/asm-x86/processor.h | |||
| @@ -3,8 +3,7 @@ | |||
| 3 | 3 | ||
| 4 | #include <asm/processor-flags.h> | 4 | #include <asm/processor-flags.h> |
| 5 | 5 | ||
| 6 | /* migration helpers, for KVM - will be removed in 2.6.25: */ | 6 | /* migration helper, for KVM - will be removed in 2.6.25: */ |
| 7 | #include <asm/vm86.h> | ||
| 8 | #define Xgt_desc_struct desc_ptr | 7 | #define Xgt_desc_struct desc_ptr |
| 9 | 8 | ||
| 10 | /* Forward declaration, a strange C thing */ | 9 | /* Forward declaration, a strange C thing */ |
| @@ -24,6 +23,7 @@ struct mm_struct; | |||
| 24 | #include <asm/msr.h> | 23 | #include <asm/msr.h> |
| 25 | #include <asm/desc_defs.h> | 24 | #include <asm/desc_defs.h> |
| 26 | #include <asm/nops.h> | 25 | #include <asm/nops.h> |
| 26 | |||
| 27 | #include <linux/personality.h> | 27 | #include <linux/personality.h> |
| 28 | #include <linux/cpumask.h> | 28 | #include <linux/cpumask.h> |
| 29 | #include <linux/cache.h> | 29 | #include <linux/cache.h> |
| @@ -37,16 +37,18 @@ struct mm_struct; | |||
| 37 | static inline void *current_text_addr(void) | 37 | static inline void *current_text_addr(void) |
| 38 | { | 38 | { |
| 39 | void *pc; | 39 | void *pc; |
| 40 | asm volatile("mov $1f,%0\n1:":"=r" (pc)); | 40 | |
| 41 | asm volatile("mov $1f, %0; 1:":"=r" (pc)); | ||
| 42 | |||
| 41 | return pc; | 43 | return pc; |
| 42 | } | 44 | } |
| 43 | 45 | ||
| 44 | #ifdef CONFIG_X86_VSMP | 46 | #ifdef CONFIG_X86_VSMP |
| 45 | #define ARCH_MIN_TASKALIGN (1 << INTERNODE_CACHE_SHIFT) | 47 | # define ARCH_MIN_TASKALIGN (1 << INTERNODE_CACHE_SHIFT) |
| 46 | #define ARCH_MIN_MMSTRUCT_ALIGN (1 << INTERNODE_CACHE_SHIFT) | 48 | # define ARCH_MIN_MMSTRUCT_ALIGN (1 << INTERNODE_CACHE_SHIFT) |
| 47 | #else | 49 | #else |
| 48 | #define ARCH_MIN_TASKALIGN 16 | 50 | # define ARCH_MIN_TASKALIGN 16 |
| 49 | #define ARCH_MIN_MMSTRUCT_ALIGN 0 | 51 | # define ARCH_MIN_MMSTRUCT_ALIGN 0 |
| 50 | #endif | 52 | #endif |
| 51 | 53 | ||
| 52 | /* | 54 | /* |
| @@ -56,69 +58,82 @@ static inline void *current_text_addr(void) | |||
| 56 | */ | 58 | */ |
| 57 | 59 | ||
| 58 | struct cpuinfo_x86 { | 60 | struct cpuinfo_x86 { |
| 59 | __u8 x86; /* CPU family */ | 61 | __u8 x86; /* CPU family */ |
| 60 | __u8 x86_vendor; /* CPU vendor */ | 62 | __u8 x86_vendor; /* CPU vendor */ |
| 61 | __u8 x86_model; | 63 | __u8 x86_model; |
| 62 | __u8 x86_mask; | 64 | __u8 x86_mask; |
| 63 | #ifdef CONFIG_X86_32 | 65 | #ifdef CONFIG_X86_32 |
| 64 | char wp_works_ok; /* It doesn't on 386's */ | 66 | char wp_works_ok; /* It doesn't on 386's */ |
| 65 | char hlt_works_ok; /* Problems on some 486Dx4's and old 386's */ | 67 | |
| 66 | char hard_math; | 68 | /* Problems on some 486Dx4's and old 386's: */ |
| 67 | char rfu; | 69 | char hlt_works_ok; |
| 68 | char fdiv_bug; | 70 | char hard_math; |
| 69 | char f00f_bug; | 71 | char rfu; |
| 70 | char coma_bug; | 72 | char fdiv_bug; |
| 71 | char pad0; | 73 | char f00f_bug; |
| 74 | char coma_bug; | ||
| 75 | char pad0; | ||
| 72 | #else | 76 | #else |
| 73 | /* number of 4K pages in DTLB/ITLB combined(in pages)*/ | 77 | /* Number of 4K pages in DTLB/ITLB combined(in pages): */ |
| 74 | int x86_tlbsize; | 78 | int x86_tlbsize; |
| 75 | __u8 x86_virt_bits, x86_phys_bits; | 79 | __u8 x86_virt_bits; |
| 76 | /* cpuid returned core id bits */ | 80 | __u8 x86_phys_bits; |
| 77 | __u8 x86_coreid_bits; | 81 | /* CPUID returned core id bits: */ |
| 78 | /* Max extended CPUID function supported */ | 82 | __u8 x86_coreid_bits; |
| 79 | __u32 extended_cpuid_level; | 83 | /* Max extended CPUID function supported: */ |
| 84 | __u32 extended_cpuid_level; | ||
| 80 | #endif | 85 | #endif |
| 81 | int cpuid_level; /* Maximum supported CPUID level, -1=no CPUID */ | 86 | /* Maximum supported CPUID level, -1=no CPUID: */ |
| 82 | __u32 x86_capability[NCAPINTS]; | 87 | int cpuid_level; |
| 83 | char x86_vendor_id[16]; | 88 | __u32 x86_capability[NCAPINTS]; |
| 84 | char x86_model_id[64]; | 89 | char x86_vendor_id[16]; |
| 85 | int x86_cache_size; /* in KB - valid for CPUS which support this | 90 | char x86_model_id[64]; |
| 86 | call */ | 91 | /* in KB - valid for CPUS which support this call: */ |
| 87 | int x86_cache_alignment; /* In bytes */ | 92 | int x86_cache_size; |
| 88 | int x86_power; | 93 | int x86_cache_alignment; /* In bytes */ |
| 89 | unsigned long loops_per_jiffy; | 94 | int x86_power; |
| 95 | unsigned long loops_per_jiffy; | ||
| 90 | #ifdef CONFIG_SMP | 96 | #ifdef CONFIG_SMP |
| 91 | cpumask_t llc_shared_map; /* cpus sharing the last level cache */ | 97 | /* cpus sharing the last level cache: */ |
| 98 | cpumask_t llc_shared_map; | ||
| 92 | #endif | 99 | #endif |
| 93 | u16 x86_max_cores; /* cpuid returned max cores value */ | 100 | /* cpuid returned max cores value: */ |
| 94 | u16 apicid; | 101 | u16 x86_max_cores; |
| 95 | u16 x86_clflush_size; | 102 | u16 apicid; |
| 103 | u16 initial_apicid; | ||
| 104 | u16 x86_clflush_size; | ||
| 96 | #ifdef CONFIG_SMP | 105 | #ifdef CONFIG_SMP |
| 97 | u16 booted_cores; /* number of cores as seen by OS */ | 106 | /* number of cores as seen by the OS: */ |
| 98 | u16 phys_proc_id; /* Physical processor id. */ | 107 | u16 booted_cores; |
| 99 | u16 cpu_core_id; /* Core id */ | 108 | /* Physical processor id: */ |
| 100 | u16 cpu_index; /* index into per_cpu list */ | 109 | u16 phys_proc_id; |
| 110 | /* Core id: */ | ||
| 111 | u16 cpu_core_id; | ||
| 112 | /* Index into per_cpu list: */ | ||
| 113 | u16 cpu_index; | ||
| 101 | #endif | 114 | #endif |
| 102 | } __attribute__((__aligned__(SMP_CACHE_BYTES))); | 115 | } __attribute__((__aligned__(SMP_CACHE_BYTES))); |
| 103 | 116 | ||
| 104 | #define X86_VENDOR_INTEL 0 | 117 | #define X86_VENDOR_INTEL 0 |
| 105 | #define X86_VENDOR_CYRIX 1 | 118 | #define X86_VENDOR_CYRIX 1 |
| 106 | #define X86_VENDOR_AMD 2 | 119 | #define X86_VENDOR_AMD 2 |
| 107 | #define X86_VENDOR_UMC 3 | 120 | #define X86_VENDOR_UMC 3 |
| 108 | #define X86_VENDOR_NEXGEN 4 | 121 | #define X86_VENDOR_NEXGEN 4 |
| 109 | #define X86_VENDOR_CENTAUR 5 | 122 | #define X86_VENDOR_CENTAUR 5 |
| 110 | #define X86_VENDOR_TRANSMETA 7 | 123 | #define X86_VENDOR_TRANSMETA 7 |
| 111 | #define X86_VENDOR_NSC 8 | 124 | #define X86_VENDOR_NSC 8 |
| 112 | #define X86_VENDOR_NUM 9 | 125 | #define X86_VENDOR_NUM 9 |
| 113 | #define X86_VENDOR_UNKNOWN 0xff | 126 | |
| 127 | #define X86_VENDOR_UNKNOWN 0xff | ||
| 114 | 128 | ||
| 115 | /* | 129 | /* |
| 116 | * capabilities of CPUs | 130 | * capabilities of CPUs |
| 117 | */ | 131 | */ |
| 118 | extern struct cpuinfo_x86 boot_cpu_data; | 132 | extern struct cpuinfo_x86 boot_cpu_data; |
| 119 | extern struct cpuinfo_x86 new_cpu_data; | 133 | extern struct cpuinfo_x86 new_cpu_data; |
| 120 | extern struct tss_struct doublefault_tss; | 134 | |
| 121 | extern __u32 cleared_cpu_caps[NCAPINTS]; | 135 | extern struct tss_struct doublefault_tss; |
| 136 | extern __u32 cleared_cpu_caps[NCAPINTS]; | ||
| 122 | 137 | ||
| 123 | #ifdef CONFIG_SMP | 138 | #ifdef CONFIG_SMP |
| 124 | DECLARE_PER_CPU(struct cpuinfo_x86, cpu_info); | 139 | DECLARE_PER_CPU(struct cpuinfo_x86, cpu_info); |
| @@ -129,7 +144,18 @@ DECLARE_PER_CPU(struct cpuinfo_x86, cpu_info); | |||
| 129 | #define current_cpu_data boot_cpu_data | 144 | #define current_cpu_data boot_cpu_data |
| 130 | #endif | 145 | #endif |
| 131 | 146 | ||
| 132 | void cpu_detect(struct cpuinfo_x86 *c); | 147 | static inline int hlt_works(int cpu) |
| 148 | { | ||
| 149 | #ifdef CONFIG_X86_32 | ||
| 150 | return cpu_data(cpu).hlt_works_ok; | ||
| 151 | #else | ||
| 152 | return 1; | ||
| 153 | #endif | ||
| 154 | } | ||
| 155 | |||
| 156 | #define cache_line_size() (boot_cpu_data.x86_cache_alignment) | ||
| 157 | |||
| 158 | extern void cpu_detect(struct cpuinfo_x86 *c); | ||
| 133 | 159 | ||
| 134 | extern void identify_cpu(struct cpuinfo_x86 *); | 160 | extern void identify_cpu(struct cpuinfo_x86 *); |
| 135 | extern void identify_boot_cpu(void); | 161 | extern void identify_boot_cpu(void); |
| @@ -146,15 +172,15 @@ static inline void detect_ht(struct cpuinfo_x86 *c) {} | |||
| 146 | #endif | 172 | #endif |
| 147 | 173 | ||
| 148 | static inline void native_cpuid(unsigned int *eax, unsigned int *ebx, | 174 | static inline void native_cpuid(unsigned int *eax, unsigned int *ebx, |
| 149 | unsigned int *ecx, unsigned int *edx) | 175 | unsigned int *ecx, unsigned int *edx) |
| 150 | { | 176 | { |
| 151 | /* ecx is often an input as well as an output. */ | 177 | /* ecx is often an input as well as an output. */ |
| 152 | __asm__("cpuid" | 178 | asm("cpuid" |
| 153 | : "=a" (*eax), | 179 | : "=a" (*eax), |
| 154 | "=b" (*ebx), | 180 | "=b" (*ebx), |
| 155 | "=c" (*ecx), | 181 | "=c" (*ecx), |
| 156 | "=d" (*edx) | 182 | "=d" (*edx) |
| 157 | : "0" (*eax), "2" (*ecx)); | 183 | : "0" (*eax), "2" (*ecx)); |
| 158 | } | 184 | } |
| 159 | 185 | ||
| 160 | static inline void load_cr3(pgd_t *pgdir) | 186 | static inline void load_cr3(pgd_t *pgdir) |
| @@ -165,54 +191,67 @@ static inline void load_cr3(pgd_t *pgdir) | |||
| 165 | #ifdef CONFIG_X86_32 | 191 | #ifdef CONFIG_X86_32 |
| 166 | /* This is the TSS defined by the hardware. */ | 192 | /* This is the TSS defined by the hardware. */ |
| 167 | struct x86_hw_tss { | 193 | struct x86_hw_tss { |
| 168 | unsigned short back_link, __blh; | 194 | unsigned short back_link, __blh; |
| 169 | unsigned long sp0; | 195 | unsigned long sp0; |
| 170 | unsigned short ss0, __ss0h; | 196 | unsigned short ss0, __ss0h; |
| 171 | unsigned long sp1; | 197 | unsigned long sp1; |
| 172 | unsigned short ss1, __ss1h; /* ss1 caches MSR_IA32_SYSENTER_CS */ | 198 | /* ss1 caches MSR_IA32_SYSENTER_CS: */ |
| 173 | unsigned long sp2; | 199 | unsigned short ss1, __ss1h; |
| 174 | unsigned short ss2, __ss2h; | 200 | unsigned long sp2; |
| 175 | unsigned long __cr3; | 201 | unsigned short ss2, __ss2h; |
| 176 | unsigned long ip; | 202 | unsigned long __cr3; |
| 177 | unsigned long flags; | 203 | unsigned long ip; |
| 178 | unsigned long ax, cx, dx, bx; | 204 | unsigned long flags; |
| 179 | unsigned long sp, bp, si, di; | 205 | unsigned long ax; |
| 180 | unsigned short es, __esh; | 206 | unsigned long cx; |
| 181 | unsigned short cs, __csh; | 207 | unsigned long dx; |
| 182 | unsigned short ss, __ssh; | 208 | unsigned long bx; |
| 183 | unsigned short ds, __dsh; | 209 | unsigned long sp; |
| 184 | unsigned short fs, __fsh; | 210 | unsigned long bp; |
| 185 | unsigned short gs, __gsh; | 211 | unsigned long si; |
| 186 | unsigned short ldt, __ldth; | 212 | unsigned long di; |
| 187 | unsigned short trace, io_bitmap_base; | 213 | unsigned short es, __esh; |
| 214 | unsigned short cs, __csh; | ||
| 215 | unsigned short ss, __ssh; | ||
| 216 | unsigned short ds, __dsh; | ||
| 217 | unsigned short fs, __fsh; | ||
| 218 | unsigned short gs, __gsh; | ||
| 219 | unsigned short ldt, __ldth; | ||
| 220 | unsigned short trace; | ||
| 221 | unsigned short io_bitmap_base; | ||
| 222 | |||
| 188 | } __attribute__((packed)); | 223 | } __attribute__((packed)); |
| 189 | #else | 224 | #else |
| 190 | struct x86_hw_tss { | 225 | struct x86_hw_tss { |
| 191 | u32 reserved1; | 226 | u32 reserved1; |
| 192 | u64 sp0; | 227 | u64 sp0; |
| 193 | u64 sp1; | 228 | u64 sp1; |
| 194 | u64 sp2; | 229 | u64 sp2; |
| 195 | u64 reserved2; | 230 | u64 reserved2; |
| 196 | u64 ist[7]; | 231 | u64 ist[7]; |
| 197 | u32 reserved3; | 232 | u32 reserved3; |
| 198 | u32 reserved4; | 233 | u32 reserved4; |
| 199 | u16 reserved5; | 234 | u16 reserved5; |
| 200 | u16 io_bitmap_base; | 235 | u16 io_bitmap_base; |
| 236 | |||
| 201 | } __attribute__((packed)) ____cacheline_aligned; | 237 | } __attribute__((packed)) ____cacheline_aligned; |
| 202 | #endif | 238 | #endif |
| 203 | 239 | ||
| 204 | /* | 240 | /* |
| 205 | * Size of io_bitmap. | 241 | * IO-bitmap sizes: |
| 206 | */ | 242 | */ |
| 207 | #define IO_BITMAP_BITS 65536 | 243 | #define IO_BITMAP_BITS 65536 |
| 208 | #define IO_BITMAP_BYTES (IO_BITMAP_BITS/8) | 244 | #define IO_BITMAP_BYTES (IO_BITMAP_BITS/8) |
| 209 | #define IO_BITMAP_LONGS (IO_BITMAP_BYTES/sizeof(long)) | 245 | #define IO_BITMAP_LONGS (IO_BITMAP_BYTES/sizeof(long)) |
| 210 | #define IO_BITMAP_OFFSET offsetof(struct tss_struct, io_bitmap) | 246 | #define IO_BITMAP_OFFSET offsetof(struct tss_struct, io_bitmap) |
| 211 | #define INVALID_IO_BITMAP_OFFSET 0x8000 | 247 | #define INVALID_IO_BITMAP_OFFSET 0x8000 |
| 212 | #define INVALID_IO_BITMAP_OFFSET_LAZY 0x9000 | 248 | #define INVALID_IO_BITMAP_OFFSET_LAZY 0x9000 |
| 213 | 249 | ||
| 214 | struct tss_struct { | 250 | struct tss_struct { |
| 215 | struct x86_hw_tss x86_tss; | 251 | /* |
| 252 | * The hardware state: | ||
| 253 | */ | ||
| 254 | struct x86_hw_tss x86_tss; | ||
| 216 | 255 | ||
| 217 | /* | 256 | /* |
| 218 | * The extra 1 is there because the CPU will access an | 257 | * The extra 1 is there because the CPU will access an |
| @@ -220,90 +259,108 @@ struct tss_struct { | |||
| 220 | * bitmap. The extra byte must be all 1 bits, and must | 259 | * bitmap. The extra byte must be all 1 bits, and must |
| 221 | * be within the limit. | 260 | * be within the limit. |
| 222 | */ | 261 | */ |
| 223 | unsigned long io_bitmap[IO_BITMAP_LONGS + 1]; | 262 | unsigned long io_bitmap[IO_BITMAP_LONGS + 1]; |
| 224 | /* | 263 | /* |
| 225 | * Cache the current maximum and the last task that used the bitmap: | 264 | * Cache the current maximum and the last task that used the bitmap: |
| 226 | */ | 265 | */ |
| 227 | unsigned long io_bitmap_max; | 266 | unsigned long io_bitmap_max; |
| 228 | struct thread_struct *io_bitmap_owner; | 267 | struct thread_struct *io_bitmap_owner; |
| 268 | |||
| 229 | /* | 269 | /* |
| 230 | * pads the TSS to be cacheline-aligned (size is 0x100) | 270 | * Pad the TSS to be cacheline-aligned (size is 0x100): |
| 231 | */ | 271 | */ |
| 232 | unsigned long __cacheline_filler[35]; | 272 | unsigned long __cacheline_filler[35]; |
| 233 | /* | 273 | /* |
| 234 | * .. and then another 0x100 bytes for emergency kernel stack | 274 | * .. and then another 0x100 bytes for the emergency kernel stack: |
| 235 | */ | 275 | */ |
| 236 | unsigned long stack[64]; | 276 | unsigned long stack[64]; |
| 277 | |||
| 237 | } __attribute__((packed)); | 278 | } __attribute__((packed)); |
| 238 | 279 | ||
| 239 | DECLARE_PER_CPU(struct tss_struct, init_tss); | 280 | DECLARE_PER_CPU(struct tss_struct, init_tss); |
| 240 | 281 | ||
| 241 | /* Save the original ist values for checking stack pointers during debugging */ | 282 | /* |
| 283 | * Save the original ist values for checking stack pointers during debugging | ||
| 284 | */ | ||
| 242 | struct orig_ist { | 285 | struct orig_ist { |
| 243 | unsigned long ist[7]; | 286 | unsigned long ist[7]; |
| 244 | }; | 287 | }; |
| 245 | 288 | ||
| 246 | #define MXCSR_DEFAULT 0x1f80 | 289 | #define MXCSR_DEFAULT 0x1f80 |
| 247 | 290 | ||
| 248 | struct i387_fsave_struct { | 291 | struct i387_fsave_struct { |
| 249 | u32 cwd; | 292 | u32 cwd; /* FPU Control Word */ |
| 250 | u32 swd; | 293 | u32 swd; /* FPU Status Word */ |
| 251 | u32 twd; | 294 | u32 twd; /* FPU Tag Word */ |
| 252 | u32 fip; | 295 | u32 fip; /* FPU IP Offset */ |
| 253 | u32 fcs; | 296 | u32 fcs; /* FPU IP Selector */ |
| 254 | u32 foo; | 297 | u32 foo; /* FPU Operand Pointer Offset */ |
| 255 | u32 fos; | 298 | u32 fos; /* FPU Operand Pointer Selector */ |
| 256 | u32 st_space[20]; /* 8*10 bytes for each FP-reg = 80 bytes */ | 299 | |
| 257 | u32 status; /* software status information */ | 300 | /* 8*10 bytes for each FP-reg = 80 bytes: */ |
| 301 | u32 st_space[20]; | ||
| 302 | |||
| 303 | /* Software status information [not touched by FSAVE ]: */ | ||
| 304 | u32 status; | ||
| 258 | }; | 305 | }; |
| 259 | 306 | ||
| 260 | struct i387_fxsave_struct { | 307 | struct i387_fxsave_struct { |
| 261 | u16 cwd; | 308 | u16 cwd; /* Control Word */ |
| 262 | u16 swd; | 309 | u16 swd; /* Status Word */ |
| 263 | u16 twd; | 310 | u16 twd; /* Tag Word */ |
| 264 | u16 fop; | 311 | u16 fop; /* Last Instruction Opcode */ |
| 265 | union { | 312 | union { |
| 266 | struct { | 313 | struct { |
| 267 | u64 rip; | 314 | u64 rip; /* Instruction Pointer */ |
| 268 | u64 rdp; | 315 | u64 rdp; /* Data Pointer */ |
| 269 | }; | 316 | }; |
| 270 | struct { | 317 | struct { |
| 271 | u32 fip; | 318 | u32 fip; /* FPU IP Offset */ |
| 272 | u32 fcs; | 319 | u32 fcs; /* FPU IP Selector */ |
| 273 | u32 foo; | 320 | u32 foo; /* FPU Operand Offset */ |
| 274 | u32 fos; | 321 | u32 fos; /* FPU Operand Selector */ |
| 275 | }; | 322 | }; |
| 276 | }; | 323 | }; |
| 277 | u32 mxcsr; | 324 | u32 mxcsr; /* MXCSR Register State */ |
| 278 | u32 mxcsr_mask; | 325 | u32 mxcsr_mask; /* MXCSR Mask */ |
| 279 | u32 st_space[32]; /* 8*16 bytes for each FP-reg = 128 bytes */ | 326 | |
| 280 | u32 xmm_space[64]; /* 16*16 bytes for each XMM-reg = 256 bytes */ | 327 | /* 8*16 bytes for each FP-reg = 128 bytes: */ |
| 281 | u32 padding[24]; | 328 | u32 st_space[32]; |
| 329 | |||
| 330 | /* 16*16 bytes for each XMM-reg = 256 bytes: */ | ||
| 331 | u32 xmm_space[64]; | ||
| 332 | |||
| 333 | u32 padding[24]; | ||
| 334 | |||
| 282 | } __attribute__((aligned(16))); | 335 | } __attribute__((aligned(16))); |
| 283 | 336 | ||
| 284 | struct i387_soft_struct { | 337 | struct i387_soft_struct { |
| 285 | u32 cwd; | 338 | u32 cwd; |
| 286 | u32 swd; | 339 | u32 swd; |
| 287 | u32 twd; | 340 | u32 twd; |
| 288 | u32 fip; | 341 | u32 fip; |
| 289 | u32 fcs; | 342 | u32 fcs; |
| 290 | u32 foo; | 343 | u32 foo; |
| 291 | u32 fos; | 344 | u32 fos; |
| 292 | u32 st_space[20]; /* 8*10 bytes for each FP-reg = 80 bytes */ | 345 | /* 8*10 bytes for each FP-reg = 80 bytes: */ |
| 293 | u8 ftop, changed, lookahead, no_update, rm, alimit; | 346 | u32 st_space[20]; |
| 294 | struct info *info; | 347 | u8 ftop; |
| 295 | u32 entry_eip; | 348 | u8 changed; |
| 349 | u8 lookahead; | ||
| 350 | u8 no_update; | ||
| 351 | u8 rm; | ||
| 352 | u8 alimit; | ||
| 353 | struct info *info; | ||
| 354 | u32 entry_eip; | ||
| 296 | }; | 355 | }; |
| 297 | 356 | ||
| 298 | union i387_union { | 357 | union i387_union { |
| 299 | struct i387_fsave_struct fsave; | 358 | struct i387_fsave_struct fsave; |
| 300 | struct i387_fxsave_struct fxsave; | 359 | struct i387_fxsave_struct fxsave; |
| 301 | struct i387_soft_struct soft; | 360 | struct i387_soft_struct soft; |
| 302 | }; | 361 | }; |
| 303 | 362 | ||
| 304 | #ifdef CONFIG_X86_32 | 363 | #ifdef CONFIG_X86_64 |
| 305 | DECLARE_PER_CPU(u8, cpu_llc_id); | ||
| 306 | #else | ||
| 307 | DECLARE_PER_CPU(struct orig_ist, orig_ist); | 364 | DECLARE_PER_CPU(struct orig_ist, orig_ist); |
| 308 | #endif | 365 | #endif |
| 309 | 366 | ||
| @@ -313,42 +370,50 @@ extern unsigned int init_intel_cacheinfo(struct cpuinfo_x86 *c); | |||
| 313 | extern unsigned short num_cache_leaves; | 370 | extern unsigned short num_cache_leaves; |
| 314 | 371 | ||
| 315 | struct thread_struct { | 372 | struct thread_struct { |
| 316 | /* cached TLS descriptors. */ | 373 | /* Cached TLS descriptors: */ |
| 317 | struct desc_struct tls_array[GDT_ENTRY_TLS_ENTRIES]; | 374 | struct desc_struct tls_array[GDT_ENTRY_TLS_ENTRIES]; |
| 318 | unsigned long sp0; | 375 | unsigned long sp0; |
| 319 | unsigned long sp; | 376 | unsigned long sp; |
| 320 | #ifdef CONFIG_X86_32 | 377 | #ifdef CONFIG_X86_32 |
| 321 | unsigned long sysenter_cs; | 378 | unsigned long sysenter_cs; |
| 322 | #else | 379 | #else |
| 323 | unsigned long usersp; /* Copy from PDA */ | 380 | unsigned long usersp; /* Copy from PDA */ |
| 324 | unsigned short es, ds, fsindex, gsindex; | 381 | unsigned short es; |
| 382 | unsigned short ds; | ||
| 383 | unsigned short fsindex; | ||
| 384 | unsigned short gsindex; | ||
| 325 | #endif | 385 | #endif |
| 326 | unsigned long ip; | 386 | unsigned long ip; |
| 327 | unsigned long fs; | 387 | unsigned long fs; |
| 328 | unsigned long gs; | 388 | unsigned long gs; |
| 329 | /* Hardware debugging registers */ | 389 | /* Hardware debugging registers: */ |
| 330 | unsigned long debugreg0; | 390 | unsigned long debugreg0; |
| 331 | unsigned long debugreg1; | 391 | unsigned long debugreg1; |
| 332 | unsigned long debugreg2; | 392 | unsigned long debugreg2; |
| 333 | unsigned long debugreg3; | 393 | unsigned long debugreg3; |
| 334 | unsigned long debugreg6; | 394 | unsigned long debugreg6; |
| 335 | unsigned long debugreg7; | 395 | unsigned long debugreg7; |
| 336 | /* fault info */ | 396 | /* Fault info: */ |
| 337 | unsigned long cr2, trap_no, error_code; | 397 | unsigned long cr2; |
| 338 | /* floating point info */ | 398 | unsigned long trap_no; |
| 399 | unsigned long error_code; | ||
| 400 | /* Floating point info: */ | ||
| 339 | union i387_union i387 __attribute__((aligned(16)));; | 401 | union i387_union i387 __attribute__((aligned(16)));; |
| 340 | #ifdef CONFIG_X86_32 | 402 | #ifdef CONFIG_X86_32 |
| 341 | /* virtual 86 mode info */ | 403 | /* Virtual 86 mode info */ |
| 342 | struct vm86_struct __user *vm86_info; | 404 | struct vm86_struct __user *vm86_info; |
| 343 | unsigned long screen_bitmap; | 405 | unsigned long screen_bitmap; |
| 344 | unsigned long v86flags, v86mask, saved_sp0; | 406 | unsigned long v86flags; |
| 345 | unsigned int saved_fs, saved_gs; | 407 | unsigned long v86mask; |
| 408 | unsigned long saved_sp0; | ||
| 409 | unsigned int saved_fs; | ||
| 410 | unsigned int saved_gs; | ||
| 346 | #endif | 411 | #endif |
| 347 | /* IO permissions */ | 412 | /* IO permissions: */ |
| 348 | unsigned long *io_bitmap_ptr; | 413 | unsigned long *io_bitmap_ptr; |
| 349 | unsigned long iopl; | 414 | unsigned long iopl; |
| 350 | /* max allowed port in the bitmap, in bytes: */ | 415 | /* Max allowed port in the bitmap, in bytes: */ |
| 351 | unsigned io_bitmap_max; | 416 | unsigned io_bitmap_max; |
| 352 | /* MSR_IA32_DEBUGCTLMSR value to switch in if TIF_DEBUGCTLMSR is set. */ | 417 | /* MSR_IA32_DEBUGCTLMSR value to switch in if TIF_DEBUGCTLMSR is set. */ |
| 353 | unsigned long debugctlmsr; | 418 | unsigned long debugctlmsr; |
| 354 | /* Debug Store - if not 0 points to a DS Save Area configuration; | 419 | /* Debug Store - if not 0 points to a DS Save Area configuration; |
| @@ -358,21 +423,27 @@ struct thread_struct { | |||
| 358 | 423 | ||
| 359 | static inline unsigned long native_get_debugreg(int regno) | 424 | static inline unsigned long native_get_debugreg(int regno) |
| 360 | { | 425 | { |
| 361 | unsigned long val = 0; /* Damn you, gcc! */ | 426 | unsigned long val = 0; /* Damn you, gcc! */ |
| 362 | 427 | ||
| 363 | switch (regno) { | 428 | switch (regno) { |
| 364 | case 0: | 429 | case 0: |
| 365 | asm("mov %%db0, %0" :"=r" (val)); break; | 430 | asm("mov %%db0, %0" :"=r" (val)); |
| 431 | break; | ||
| 366 | case 1: | 432 | case 1: |
| 367 | asm("mov %%db1, %0" :"=r" (val)); break; | 433 | asm("mov %%db1, %0" :"=r" (val)); |
| 434 | break; | ||
| 368 | case 2: | 435 | case 2: |
| 369 | asm("mov %%db2, %0" :"=r" (val)); break; | 436 | asm("mov %%db2, %0" :"=r" (val)); |
| 437 | break; | ||
| 370 | case 3: | 438 | case 3: |
| 371 | asm("mov %%db3, %0" :"=r" (val)); break; | 439 | asm("mov %%db3, %0" :"=r" (val)); |
| 440 | break; | ||
| 372 | case 6: | 441 | case 6: |
| 373 | asm("mov %%db6, %0" :"=r" (val)); break; | 442 | asm("mov %%db6, %0" :"=r" (val)); |
| 443 | break; | ||
| 374 | case 7: | 444 | case 7: |
| 375 | asm("mov %%db7, %0" :"=r" (val)); break; | 445 | asm("mov %%db7, %0" :"=r" (val)); |
| 446 | break; | ||
| 376 | default: | 447 | default: |
| 377 | BUG(); | 448 | BUG(); |
| 378 | } | 449 | } |
| @@ -383,22 +454,22 @@ static inline void native_set_debugreg(int regno, unsigned long value) | |||
| 383 | { | 454 | { |
| 384 | switch (regno) { | 455 | switch (regno) { |
| 385 | case 0: | 456 | case 0: |
| 386 | asm("mov %0,%%db0" : /* no output */ :"r" (value)); | 457 | asm("mov %0, %%db0" ::"r" (value)); |
| 387 | break; | 458 | break; |
| 388 | case 1: | 459 | case 1: |
| 389 | asm("mov %0,%%db1" : /* no output */ :"r" (value)); | 460 | asm("mov %0, %%db1" ::"r" (value)); |
| 390 | break; | 461 | break; |
| 391 | case 2: | 462 | case 2: |
| 392 | asm("mov %0,%%db2" : /* no output */ :"r" (value)); | 463 | asm("mov %0, %%db2" ::"r" (value)); |
| 393 | break; | 464 | break; |
| 394 | case 3: | 465 | case 3: |
| 395 | asm("mov %0,%%db3" : /* no output */ :"r" (value)); | 466 | asm("mov %0, %%db3" ::"r" (value)); |
| 396 | break; | 467 | break; |
| 397 | case 6: | 468 | case 6: |
| 398 | asm("mov %0,%%db6" : /* no output */ :"r" (value)); | 469 | asm("mov %0, %%db6" ::"r" (value)); |
| 399 | break; | 470 | break; |
| 400 | case 7: | 471 | case 7: |
| 401 | asm("mov %0,%%db7" : /* no output */ :"r" (value)); | 472 | asm("mov %0, %%db7" ::"r" (value)); |
| 402 | break; | 473 | break; |
| 403 | default: | 474 | default: |
| 404 | BUG(); | 475 | BUG(); |
| @@ -412,23 +483,24 @@ static inline void native_set_iopl_mask(unsigned mask) | |||
| 412 | { | 483 | { |
| 413 | #ifdef CONFIG_X86_32 | 484 | #ifdef CONFIG_X86_32 |
| 414 | unsigned int reg; | 485 | unsigned int reg; |
| 415 | __asm__ __volatile__ ("pushfl;" | 486 | |
| 416 | "popl %0;" | 487 | asm volatile ("pushfl;" |
| 417 | "andl %1, %0;" | 488 | "popl %0;" |
| 418 | "orl %2, %0;" | 489 | "andl %1, %0;" |
| 419 | "pushl %0;" | 490 | "orl %2, %0;" |
| 420 | "popfl" | 491 | "pushl %0;" |
| 421 | : "=&r" (reg) | 492 | "popfl" |
| 422 | : "i" (~X86_EFLAGS_IOPL), "r" (mask)); | 493 | : "=&r" (reg) |
| 494 | : "i" (~X86_EFLAGS_IOPL), "r" (mask)); | ||
| 423 | #endif | 495 | #endif |
| 424 | } | 496 | } |
| 425 | 497 | ||
| 426 | static inline void native_load_sp0(struct tss_struct *tss, | 498 | static inline void |
| 427 | struct thread_struct *thread) | 499 | native_load_sp0(struct tss_struct *tss, struct thread_struct *thread) |
| 428 | { | 500 | { |
| 429 | tss->x86_tss.sp0 = thread->sp0; | 501 | tss->x86_tss.sp0 = thread->sp0; |
| 430 | #ifdef CONFIG_X86_32 | 502 | #ifdef CONFIG_X86_32 |
| 431 | /* Only happens when SEP is enabled, no need to test "SEP"arately */ | 503 | /* Only happens when SEP is enabled, no need to test "SEP"arately: */ |
| 432 | if (unlikely(tss->x86_tss.ss1 != thread->sysenter_cs)) { | 504 | if (unlikely(tss->x86_tss.ss1 != thread->sysenter_cs)) { |
| 433 | tss->x86_tss.ss1 = thread->sysenter_cs; | 505 | tss->x86_tss.ss1 = thread->sysenter_cs; |
| 434 | wrmsr(MSR_IA32_SYSENTER_CS, thread->sysenter_cs, 0); | 506 | wrmsr(MSR_IA32_SYSENTER_CS, thread->sysenter_cs, 0); |
| @@ -446,8 +518,8 @@ static inline void native_swapgs(void) | |||
| 446 | #ifdef CONFIG_PARAVIRT | 518 | #ifdef CONFIG_PARAVIRT |
| 447 | #include <asm/paravirt.h> | 519 | #include <asm/paravirt.h> |
| 448 | #else | 520 | #else |
| 449 | #define __cpuid native_cpuid | 521 | #define __cpuid native_cpuid |
| 450 | #define paravirt_enabled() 0 | 522 | #define paravirt_enabled() 0 |
| 451 | 523 | ||
| 452 | /* | 524 | /* |
| 453 | * These special macros can be used to get or set a debugging register | 525 | * These special macros can be used to get or set a debugging register |
| @@ -473,11 +545,12 @@ static inline void load_sp0(struct tss_struct *tss, | |||
| 473 | * enable), so that any CPU's that boot up | 545 | * enable), so that any CPU's that boot up |
| 474 | * after us can get the correct flags. | 546 | * after us can get the correct flags. |
| 475 | */ | 547 | */ |
| 476 | extern unsigned long mmu_cr4_features; | 548 | extern unsigned long mmu_cr4_features; |
| 477 | 549 | ||
| 478 | static inline void set_in_cr4(unsigned long mask) | 550 | static inline void set_in_cr4(unsigned long mask) |
| 479 | { | 551 | { |
| 480 | unsigned cr4; | 552 | unsigned cr4; |
| 553 | |||
| 481 | mmu_cr4_features |= mask; | 554 | mmu_cr4_features |= mask; |
| 482 | cr4 = read_cr4(); | 555 | cr4 = read_cr4(); |
| 483 | cr4 |= mask; | 556 | cr4 |= mask; |
| @@ -487,6 +560,7 @@ static inline void set_in_cr4(unsigned long mask) | |||
| 487 | static inline void clear_in_cr4(unsigned long mask) | 560 | static inline void clear_in_cr4(unsigned long mask) |
| 488 | { | 561 | { |
| 489 | unsigned cr4; | 562 | unsigned cr4; |
| 563 | |||
| 490 | mmu_cr4_features &= ~mask; | 564 | mmu_cr4_features &= ~mask; |
| 491 | cr4 = read_cr4(); | 565 | cr4 = read_cr4(); |
| 492 | cr4 &= ~mask; | 566 | cr4 &= ~mask; |
| @@ -494,42 +568,42 @@ static inline void clear_in_cr4(unsigned long mask) | |||
| 494 | } | 568 | } |
| 495 | 569 | ||
| 496 | struct microcode_header { | 570 | struct microcode_header { |
| 497 | unsigned int hdrver; | 571 | unsigned int hdrver; |
| 498 | unsigned int rev; | 572 | unsigned int rev; |
| 499 | unsigned int date; | 573 | unsigned int date; |
| 500 | unsigned int sig; | 574 | unsigned int sig; |
| 501 | unsigned int cksum; | 575 | unsigned int cksum; |
| 502 | unsigned int ldrver; | 576 | unsigned int ldrver; |
| 503 | unsigned int pf; | 577 | unsigned int pf; |
| 504 | unsigned int datasize; | 578 | unsigned int datasize; |
| 505 | unsigned int totalsize; | 579 | unsigned int totalsize; |
| 506 | unsigned int reserved[3]; | 580 | unsigned int reserved[3]; |
| 507 | }; | 581 | }; |
| 508 | 582 | ||
| 509 | struct microcode { | 583 | struct microcode { |
| 510 | struct microcode_header hdr; | 584 | struct microcode_header hdr; |
| 511 | unsigned int bits[0]; | 585 | unsigned int bits[0]; |
| 512 | }; | 586 | }; |
| 513 | 587 | ||
| 514 | typedef struct microcode microcode_t; | 588 | typedef struct microcode microcode_t; |
| 515 | typedef struct microcode_header microcode_header_t; | 589 | typedef struct microcode_header microcode_header_t; |
| 516 | 590 | ||
| 517 | /* microcode format is extended from prescott processors */ | 591 | /* microcode format is extended from prescott processors */ |
| 518 | struct extended_signature { | 592 | struct extended_signature { |
| 519 | unsigned int sig; | 593 | unsigned int sig; |
| 520 | unsigned int pf; | 594 | unsigned int pf; |
| 521 | unsigned int cksum; | 595 | unsigned int cksum; |
| 522 | }; | 596 | }; |
| 523 | 597 | ||
| 524 | struct extended_sigtable { | 598 | struct extended_sigtable { |
| 525 | unsigned int count; | 599 | unsigned int count; |
| 526 | unsigned int cksum; | 600 | unsigned int cksum; |
| 527 | unsigned int reserved[3]; | 601 | unsigned int reserved[3]; |
| 528 | struct extended_signature sigs[0]; | 602 | struct extended_signature sigs[0]; |
| 529 | }; | 603 | }; |
| 530 | 604 | ||
| 531 | typedef struct { | 605 | typedef struct { |
| 532 | unsigned long seg; | 606 | unsigned long seg; |
| 533 | } mm_segment_t; | 607 | } mm_segment_t; |
| 534 | 608 | ||
| 535 | 609 | ||
| @@ -541,7 +615,7 @@ extern int kernel_thread(int (*fn)(void *), void *arg, unsigned long flags); | |||
| 541 | /* Free all resources held by a thread. */ | 615 | /* Free all resources held by a thread. */ |
| 542 | extern void release_thread(struct task_struct *); | 616 | extern void release_thread(struct task_struct *); |
| 543 | 617 | ||
| 544 | /* Prepare to copy thread state - unlazy all lazy status */ | 618 | /* Prepare to copy thread state - unlazy all lazy state */ |
| 545 | extern void prepare_to_copy(struct task_struct *tsk); | 619 | extern void prepare_to_copy(struct task_struct *tsk); |
| 546 | 620 | ||
| 547 | unsigned long get_wchan(struct task_struct *p); | 621 | unsigned long get_wchan(struct task_struct *p); |
| @@ -578,118 +652,137 @@ static inline unsigned int cpuid_eax(unsigned int op) | |||
| 578 | unsigned int eax, ebx, ecx, edx; | 652 | unsigned int eax, ebx, ecx, edx; |
| 579 | 653 | ||
| 580 | cpuid(op, &eax, &ebx, &ecx, &edx); | 654 | cpuid(op, &eax, &ebx, &ecx, &edx); |
| 655 | |||
| 581 | return eax; | 656 | return eax; |
| 582 | } | 657 | } |
| 658 | |||
| 583 | static inline unsigned int cpuid_ebx(unsigned int op) | 659 | static inline unsigned int cpuid_ebx(unsigned int op) |
| 584 | { | 660 | { |
| 585 | unsigned int eax, ebx, ecx, edx; | 661 | unsigned int eax, ebx, ecx, edx; |
| 586 | 662 | ||
| 587 | cpuid(op, &eax, &ebx, &ecx, &edx); | 663 | cpuid(op, &eax, &ebx, &ecx, &edx); |
| 664 | |||
| 588 | return ebx; | 665 | return ebx; |
| 589 | } | 666 | } |
| 667 | |||
| 590 | static inline unsigned int cpuid_ecx(unsigned int op) | 668 | static inline unsigned int cpuid_ecx(unsigned int op) |
| 591 | { | 669 | { |
| 592 | unsigned int eax, ebx, ecx, edx; | 670 | unsigned int eax, ebx, ecx, edx; |
| 593 | 671 | ||
| 594 | cpuid(op, &eax, &ebx, &ecx, &edx); | 672 | cpuid(op, &eax, &ebx, &ecx, &edx); |
| 673 | |||
| 595 | return ecx; | 674 | return ecx; |
| 596 | } | 675 | } |
| 676 | |||
| 597 | static inline unsigned int cpuid_edx(unsigned int op) | 677 | static inline unsigned int cpuid_edx(unsigned int op) |
| 598 | { | 678 | { |
| 599 | unsigned int eax, ebx, ecx, edx; | 679 | unsigned int eax, ebx, ecx, edx; |
| 600 | 680 | ||
| 601 | cpuid(op, &eax, &ebx, &ecx, &edx); | 681 | cpuid(op, &eax, &ebx, &ecx, &edx); |
| 682 | |||
| 602 | return edx; | 683 | return edx; |
| 603 | } | 684 | } |
| 604 | 685 | ||
| 605 | /* REP NOP (PAUSE) is a good thing to insert into busy-wait loops. */ | 686 | /* REP NOP (PAUSE) is a good thing to insert into busy-wait loops. */ |
| 606 | static inline void rep_nop(void) | 687 | static inline void rep_nop(void) |
| 607 | { | 688 | { |
| 608 | __asm__ __volatile__("rep;nop": : :"memory"); | 689 | asm volatile("rep; nop" ::: "memory"); |
| 609 | } | 690 | } |
| 610 | 691 | ||
| 611 | /* Stop speculative execution */ | 692 | static inline void cpu_relax(void) |
| 693 | { | ||
| 694 | rep_nop(); | ||
| 695 | } | ||
| 696 | |||
| 697 | /* Stop speculative execution: */ | ||
| 612 | static inline void sync_core(void) | 698 | static inline void sync_core(void) |
| 613 | { | 699 | { |
| 614 | int tmp; | 700 | int tmp; |
| 701 | |||
| 615 | asm volatile("cpuid" : "=a" (tmp) : "0" (1) | 702 | asm volatile("cpuid" : "=a" (tmp) : "0" (1) |
| 616 | : "ebx", "ecx", "edx", "memory"); | 703 | : "ebx", "ecx", "edx", "memory"); |
| 617 | } | 704 | } |
| 618 | 705 | ||
| 619 | #define cpu_relax() rep_nop() | ||
| 620 | |||
| 621 | static inline void __monitor(const void *eax, unsigned long ecx, | 706 | static inline void __monitor(const void *eax, unsigned long ecx, |
| 622 | unsigned long edx) | 707 | unsigned long edx) |
| 623 | { | 708 | { |
| 624 | /* "monitor %eax,%ecx,%edx;" */ | 709 | /* "monitor %eax, %ecx, %edx;" */ |
| 625 | asm volatile( | 710 | asm volatile(".byte 0x0f, 0x01, 0xc8;" |
| 626 | ".byte 0x0f,0x01,0xc8;" | 711 | :: "a" (eax), "c" (ecx), "d"(edx)); |
| 627 | : :"a" (eax), "c" (ecx), "d"(edx)); | ||
| 628 | } | 712 | } |
| 629 | 713 | ||
| 630 | static inline void __mwait(unsigned long eax, unsigned long ecx) | 714 | static inline void __mwait(unsigned long eax, unsigned long ecx) |
| 631 | { | 715 | { |
| 632 | /* "mwait %eax,%ecx;" */ | 716 | /* "mwait %eax, %ecx;" */ |
| 633 | asm volatile( | 717 | asm volatile(".byte 0x0f, 0x01, 0xc9;" |
| 634 | ".byte 0x0f,0x01,0xc9;" | 718 | :: "a" (eax), "c" (ecx)); |
| 635 | : :"a" (eax), "c" (ecx)); | ||
| 636 | } | 719 | } |
| 637 | 720 | ||
| 638 | static inline void __sti_mwait(unsigned long eax, unsigned long ecx) | 721 | static inline void __sti_mwait(unsigned long eax, unsigned long ecx) |
| 639 | { | 722 | { |
| 640 | /* "mwait %eax,%ecx;" */ | 723 | /* "mwait %eax, %ecx;" */ |
| 641 | asm volatile( | 724 | asm volatile("sti; .byte 0x0f, 0x01, 0xc9;" |
| 642 | "sti; .byte 0x0f,0x01,0xc9;" | 725 | :: "a" (eax), "c" (ecx)); |
| 643 | : :"a" (eax), "c" (ecx)); | ||
| 644 | } | 726 | } |
| 645 | 727 | ||
| 646 | extern void mwait_idle_with_hints(unsigned long eax, unsigned long ecx); | 728 | extern void mwait_idle_with_hints(unsigned long eax, unsigned long ecx); |
| 647 | 729 | ||
| 648 | extern int force_mwait; | 730 | extern int force_mwait; |
| 649 | 731 | ||
| 650 | extern void select_idle_routine(const struct cpuinfo_x86 *c); | 732 | extern void select_idle_routine(const struct cpuinfo_x86 *c); |
| 651 | 733 | ||
| 652 | extern unsigned long boot_option_idle_override; | 734 | extern unsigned long boot_option_idle_override; |
| 653 | 735 | ||
| 654 | extern void enable_sep_cpu(void); | 736 | extern void enable_sep_cpu(void); |
| 655 | extern int sysenter_setup(void); | 737 | extern int sysenter_setup(void); |
| 656 | 738 | ||
| 657 | /* Defined in head.S */ | 739 | /* Defined in head.S */ |
| 658 | extern struct desc_ptr early_gdt_descr; | 740 | extern struct desc_ptr early_gdt_descr; |
| 659 | 741 | ||
| 660 | extern void cpu_set_gdt(int); | 742 | extern void cpu_set_gdt(int); |
| 661 | extern void switch_to_new_gdt(void); | 743 | extern void switch_to_new_gdt(void); |
| 662 | extern void cpu_init(void); | 744 | extern void cpu_init(void); |
| 663 | extern void init_gdt(int cpu); | 745 | extern void init_gdt(int cpu); |
| 664 | 746 | ||
| 665 | /* from system description table in BIOS. Mostly for MCA use, but | 747 | static inline void update_debugctlmsr(unsigned long debugctlmsr) |
| 666 | * others may find it useful. */ | 748 | { |
| 667 | extern unsigned int machine_id; | 749 | #ifndef CONFIG_X86_DEBUGCTLMSR |
| 668 | extern unsigned int machine_submodel_id; | 750 | if (boot_cpu_data.x86 < 6) |
| 669 | extern unsigned int BIOS_revision; | 751 | return; |
| 752 | #endif | ||
| 753 | wrmsrl(MSR_IA32_DEBUGCTLMSR, debugctlmsr); | ||
| 754 | } | ||
| 670 | 755 | ||
| 671 | /* Boot loader type from the setup header */ | 756 | /* |
| 672 | extern int bootloader_type; | 757 | * from system description table in BIOS. Mostly for MCA use, but |
| 758 | * others may find it useful: | ||
| 759 | */ | ||
| 760 | extern unsigned int machine_id; | ||
| 761 | extern unsigned int machine_submodel_id; | ||
| 762 | extern unsigned int BIOS_revision; | ||
| 763 | |||
| 764 | /* Boot loader type from the setup header: */ | ||
| 765 | extern int bootloader_type; | ||
| 673 | 766 | ||
| 674 | extern char ignore_fpu_irq; | 767 | extern char ignore_fpu_irq; |
| 675 | #define cache_line_size() (boot_cpu_data.x86_cache_alignment) | ||
| 676 | 768 | ||
| 677 | #define HAVE_ARCH_PICK_MMAP_LAYOUT 1 | 769 | #define HAVE_ARCH_PICK_MMAP_LAYOUT 1 |
| 678 | #define ARCH_HAS_PREFETCHW | 770 | #define ARCH_HAS_PREFETCHW |
| 679 | #define ARCH_HAS_SPINLOCK_PREFETCH | 771 | #define ARCH_HAS_SPINLOCK_PREFETCH |
| 680 | 772 | ||
| 681 | #ifdef CONFIG_X86_32 | 773 | #ifdef CONFIG_X86_32 |
| 682 | #define BASE_PREFETCH ASM_NOP4 | 774 | # define BASE_PREFETCH ASM_NOP4 |
| 683 | #define ARCH_HAS_PREFETCH | 775 | # define ARCH_HAS_PREFETCH |
| 684 | #else | 776 | #else |
| 685 | #define BASE_PREFETCH "prefetcht0 (%1)" | 777 | # define BASE_PREFETCH "prefetcht0 (%1)" |
| 686 | #endif | 778 | #endif |
| 687 | 779 | ||
| 688 | /* Prefetch instructions for Pentium III and AMD Athlon */ | 780 | /* |
| 689 | /* It's not worth to care about 3dnow! prefetches for the K6 | 781 | * Prefetch instructions for Pentium III (+) and AMD Athlon (+) |
| 690 | because they are microcoded there and very slow. | 782 | * |
| 691 | However we don't do prefetches for pre XP Athlons currently | 783 | * It's not worth to care about 3dnow prefetches for the K6 |
| 692 | That should be fixed. */ | 784 | * because they are microcoded there and very slow. |
| 785 | */ | ||
| 693 | static inline void prefetch(const void *x) | 786 | static inline void prefetch(const void *x) |
| 694 | { | 787 | { |
| 695 | alternative_input(BASE_PREFETCH, | 788 | alternative_input(BASE_PREFETCH, |
| @@ -698,8 +791,11 @@ static inline void prefetch(const void *x) | |||
| 698 | "r" (x)); | 791 | "r" (x)); |
| 699 | } | 792 | } |
| 700 | 793 | ||
| 701 | /* 3dnow! prefetch to get an exclusive cache line. Useful for | 794 | /* |
| 702 | spinlocks to avoid one state transition in the cache coherency protocol. */ | 795 | * 3dnow prefetch to get an exclusive cache line. |
| 796 | * Useful for spinlocks to avoid one state transition in the | ||
| 797 | * cache coherency protocol: | ||
| 798 | */ | ||
| 703 | static inline void prefetchw(const void *x) | 799 | static inline void prefetchw(const void *x) |
| 704 | { | 800 | { |
| 705 | alternative_input(BASE_PREFETCH, | 801 | alternative_input(BASE_PREFETCH, |
| @@ -708,21 +804,25 @@ static inline void prefetchw(const void *x) | |||
| 708 | "r" (x)); | 804 | "r" (x)); |
| 709 | } | 805 | } |
| 710 | 806 | ||
| 711 | #define spin_lock_prefetch(x) prefetchw(x) | 807 | static inline void spin_lock_prefetch(const void *x) |
| 808 | { | ||
| 809 | prefetchw(x); | ||
| 810 | } | ||
| 811 | |||
| 712 | #ifdef CONFIG_X86_32 | 812 | #ifdef CONFIG_X86_32 |
| 713 | /* | 813 | /* |
| 714 | * User space process size: 3GB (default). | 814 | * User space process size: 3GB (default). |
| 715 | */ | 815 | */ |
| 716 | #define TASK_SIZE (PAGE_OFFSET) | 816 | #define TASK_SIZE PAGE_OFFSET |
| 717 | #define STACK_TOP TASK_SIZE | 817 | #define STACK_TOP TASK_SIZE |
| 718 | #define STACK_TOP_MAX STACK_TOP | 818 | #define STACK_TOP_MAX STACK_TOP |
| 719 | 819 | ||
| 720 | #define INIT_THREAD { \ | 820 | #define INIT_THREAD { \ |
| 721 | .sp0 = sizeof(init_stack) + (long)&init_stack, \ | 821 | .sp0 = sizeof(init_stack) + (long)&init_stack, \ |
| 722 | .vm86_info = NULL, \ | 822 | .vm86_info = NULL, \ |
| 723 | .sysenter_cs = __KERNEL_CS, \ | 823 | .sysenter_cs = __KERNEL_CS, \ |
| 724 | .io_bitmap_ptr = NULL, \ | 824 | .io_bitmap_ptr = NULL, \ |
| 725 | .fs = __KERNEL_PERCPU, \ | 825 | .fs = __KERNEL_PERCPU, \ |
| 726 | } | 826 | } |
| 727 | 827 | ||
| 728 | /* | 828 | /* |
| @@ -731,28 +831,15 @@ static inline void prefetchw(const void *x) | |||
| 731 | * permission bitmap. The extra byte must be all 1 bits, and must | 831 | * permission bitmap. The extra byte must be all 1 bits, and must |
| 732 | * be within the limit. | 832 | * be within the limit. |
| 733 | */ | 833 | */ |
| 734 | #define INIT_TSS { \ | 834 | #define INIT_TSS { \ |
| 735 | .x86_tss = { \ | 835 | .x86_tss = { \ |
| 736 | .sp0 = sizeof(init_stack) + (long)&init_stack, \ | 836 | .sp0 = sizeof(init_stack) + (long)&init_stack, \ |
| 737 | .ss0 = __KERNEL_DS, \ | 837 | .ss0 = __KERNEL_DS, \ |
| 738 | .ss1 = __KERNEL_CS, \ | 838 | .ss1 = __KERNEL_CS, \ |
| 739 | .io_bitmap_base = INVALID_IO_BITMAP_OFFSET, \ | 839 | .io_bitmap_base = INVALID_IO_BITMAP_OFFSET, \ |
| 740 | }, \ | 840 | }, \ |
| 741 | .io_bitmap = { [0 ... IO_BITMAP_LONGS] = ~0 }, \ | 841 | .io_bitmap = { [0 ... IO_BITMAP_LONGS] = ~0 }, \ |
| 742 | } | 842 | } |
| 743 | |||
| 744 | #define start_thread(regs, new_eip, new_esp) do { \ | ||
| 745 | __asm__("movl %0,%%gs": :"r" (0)); \ | ||
| 746 | regs->fs = 0; \ | ||
| 747 | set_fs(USER_DS); \ | ||
| 748 | regs->ds = __USER_DS; \ | ||
| 749 | regs->es = __USER_DS; \ | ||
| 750 | regs->ss = __USER_DS; \ | ||
| 751 | regs->cs = __USER_CS; \ | ||
| 752 | regs->ip = new_eip; \ | ||
| 753 | regs->sp = new_esp; \ | ||
| 754 | } while (0) | ||
| 755 | |||
| 756 | 843 | ||
| 757 | extern unsigned long thread_saved_pc(struct task_struct *tsk); | 844 | extern unsigned long thread_saved_pc(struct task_struct *tsk); |
| 758 | 845 | ||
| @@ -780,24 +867,24 @@ extern unsigned long thread_saved_pc(struct task_struct *tsk); | |||
| 780 | __regs__ - 1; \ | 867 | __regs__ - 1; \ |
| 781 | }) | 868 | }) |
| 782 | 869 | ||
| 783 | #define KSTK_ESP(task) (task_pt_regs(task)->sp) | 870 | #define KSTK_ESP(task) (task_pt_regs(task)->sp) |
| 784 | 871 | ||
| 785 | #else | 872 | #else |
| 786 | /* | 873 | /* |
| 787 | * User space process size. 47bits minus one guard page. | 874 | * User space process size. 47bits minus one guard page. |
| 788 | */ | 875 | */ |
| 789 | #define TASK_SIZE64 (0x800000000000UL - 4096) | 876 | #define TASK_SIZE64 ((1UL << 47) - PAGE_SIZE) |
| 790 | 877 | ||
| 791 | /* This decides where the kernel will search for a free chunk of vm | 878 | /* This decides where the kernel will search for a free chunk of vm |
| 792 | * space during mmap's. | 879 | * space during mmap's. |
| 793 | */ | 880 | */ |
| 794 | #define IA32_PAGE_OFFSET ((current->personality & ADDR_LIMIT_3GB) ? \ | 881 | #define IA32_PAGE_OFFSET ((current->personality & ADDR_LIMIT_3GB) ? \ |
| 795 | 0xc0000000 : 0xFFFFe000) | 882 | 0xc0000000 : 0xFFFFe000) |
| 796 | 883 | ||
| 797 | #define TASK_SIZE (test_thread_flag(TIF_IA32) ? \ | 884 | #define TASK_SIZE (test_thread_flag(TIF_IA32) ? \ |
| 798 | IA32_PAGE_OFFSET : TASK_SIZE64) | 885 | IA32_PAGE_OFFSET : TASK_SIZE64) |
| 799 | #define TASK_SIZE_OF(child) ((test_tsk_thread_flag(child, TIF_IA32)) ? \ | 886 | #define TASK_SIZE_OF(child) ((test_tsk_thread_flag(child, TIF_IA32)) ? \ |
| 800 | IA32_PAGE_OFFSET : TASK_SIZE64) | 887 | IA32_PAGE_OFFSET : TASK_SIZE64) |
| 801 | 888 | ||
| 802 | #define STACK_TOP TASK_SIZE | 889 | #define STACK_TOP TASK_SIZE |
| 803 | #define STACK_TOP_MAX TASK_SIZE64 | 890 | #define STACK_TOP_MAX TASK_SIZE64 |
| @@ -810,33 +897,25 @@ extern unsigned long thread_saved_pc(struct task_struct *tsk); | |||
| 810 | .x86_tss.sp0 = (unsigned long)&init_stack + sizeof(init_stack) \ | 897 | .x86_tss.sp0 = (unsigned long)&init_stack + sizeof(init_stack) \ |
| 811 | } | 898 | } |
| 812 | 899 | ||
| 813 | #define start_thread(regs, new_rip, new_rsp) do { \ | ||
| 814 | asm volatile("movl %0,%%fs; movl %0,%%es; movl %0,%%ds": :"r" (0)); \ | ||
| 815 | load_gs_index(0); \ | ||
| 816 | (regs)->ip = (new_rip); \ | ||
| 817 | (regs)->sp = (new_rsp); \ | ||
| 818 | write_pda(oldrsp, (new_rsp)); \ | ||
| 819 | (regs)->cs = __USER_CS; \ | ||
| 820 | (regs)->ss = __USER_DS; \ | ||
| 821 | (regs)->flags = 0x200; \ | ||
| 822 | set_fs(USER_DS); \ | ||
| 823 | } while (0) | ||
| 824 | |||
| 825 | /* | 900 | /* |
| 826 | * Return saved PC of a blocked thread. | 901 | * Return saved PC of a blocked thread. |
| 827 | * What is this good for? it will be always the scheduler or ret_from_fork. | 902 | * What is this good for? it will be always the scheduler or ret_from_fork. |
| 828 | */ | 903 | */ |
| 829 | #define thread_saved_pc(t) (*(unsigned long *)((t)->thread.sp - 8)) | 904 | #define thread_saved_pc(t) (*(unsigned long *)((t)->thread.sp - 8)) |
| 830 | 905 | ||
| 831 | #define task_pt_regs(tsk) ((struct pt_regs *)(tsk)->thread.sp0 - 1) | 906 | #define task_pt_regs(tsk) ((struct pt_regs *)(tsk)->thread.sp0 - 1) |
| 832 | #define KSTK_ESP(tsk) -1 /* sorry. doesn't work for syscall. */ | 907 | #define KSTK_ESP(tsk) -1 /* sorry. doesn't work for syscall. */ |
| 833 | #endif /* CONFIG_X86_64 */ | 908 | #endif /* CONFIG_X86_64 */ |
| 834 | 909 | ||
| 835 | /* This decides where the kernel will search for a free chunk of vm | 910 | extern void start_thread(struct pt_regs *regs, unsigned long new_ip, |
| 911 | unsigned long new_sp); | ||
| 912 | |||
| 913 | /* | ||
| 914 | * This decides where the kernel will search for a free chunk of vm | ||
| 836 | * space during mmap's. | 915 | * space during mmap's. |
| 837 | */ | 916 | */ |
| 838 | #define TASK_UNMAPPED_BASE (PAGE_ALIGN(TASK_SIZE / 3)) | 917 | #define TASK_UNMAPPED_BASE (PAGE_ALIGN(TASK_SIZE / 3)) |
| 839 | 918 | ||
| 840 | #define KSTK_EIP(task) (task_pt_regs(task)->ip) | 919 | #define KSTK_EIP(task) (task_pt_regs(task)->ip) |
| 841 | 920 | ||
| 842 | #endif | 921 | #endif |
diff --git a/include/asm-x86/proto.h b/include/asm-x86/proto.h index 68563c0709ac..1e17bcce450e 100644 --- a/include/asm-x86/proto.h +++ b/include/asm-x86/proto.h | |||
| @@ -7,8 +7,6 @@ | |||
| 7 | 7 | ||
| 8 | extern void early_idt_handler(void); | 8 | extern void early_idt_handler(void); |
| 9 | 9 | ||
| 10 | extern void init_memory_mapping(unsigned long start, unsigned long end); | ||
| 11 | |||
| 12 | extern void system_call(void); | 10 | extern void system_call(void); |
| 13 | extern void syscall_init(void); | 11 | extern void syscall_init(void); |
| 14 | 12 | ||
| @@ -26,7 +24,7 @@ extern int reboot_force; | |||
| 26 | 24 | ||
| 27 | long do_arch_prctl(struct task_struct *task, int code, unsigned long addr); | 25 | long do_arch_prctl(struct task_struct *task, int code, unsigned long addr); |
| 28 | 26 | ||
| 29 | #define round_up(x,y) (((x) + (y) - 1) & ~((y)-1)) | 27 | #define round_up(x, y) (((x) + (y) - 1) & ~((y) - 1)) |
| 30 | #define round_down(x,y) ((x) & ~((y)-1)) | 28 | #define round_down(x, y) ((x) & ~((y) - 1)) |
| 31 | 29 | ||
| 32 | #endif | 30 | #endif |
diff --git a/include/asm-x86/ptrace.h b/include/asm-x86/ptrace.h index d9e04b46a440..24ec061566c5 100644 --- a/include/asm-x86/ptrace.h +++ b/include/asm-x86/ptrace.h | |||
| @@ -36,23 +36,23 @@ struct pt_regs { | |||
| 36 | #else /* __KERNEL__ */ | 36 | #else /* __KERNEL__ */ |
| 37 | 37 | ||
| 38 | struct pt_regs { | 38 | struct pt_regs { |
| 39 | long bx; | 39 | unsigned long bx; |
| 40 | long cx; | 40 | unsigned long cx; |
| 41 | long dx; | 41 | unsigned long dx; |
| 42 | long si; | 42 | unsigned long si; |
| 43 | long di; | 43 | unsigned long di; |
| 44 | long bp; | 44 | unsigned long bp; |
| 45 | long ax; | 45 | unsigned long ax; |
| 46 | int ds; | 46 | unsigned long ds; |
| 47 | int es; | 47 | unsigned long es; |
| 48 | int fs; | 48 | unsigned long fs; |
| 49 | /* int gs; */ | 49 | /* int gs; */ |
| 50 | long orig_ax; | 50 | unsigned long orig_ax; |
| 51 | long ip; | 51 | unsigned long ip; |
| 52 | int cs; | 52 | unsigned long cs; |
| 53 | long flags; | 53 | unsigned long flags; |
| 54 | long sp; | 54 | unsigned long sp; |
| 55 | int ss; | 55 | unsigned long ss; |
| 56 | }; | 56 | }; |
| 57 | 57 | ||
| 58 | #include <asm/vm86.h> | 58 | #include <asm/vm86.h> |
| @@ -140,12 +140,16 @@ extern unsigned long | |||
| 140 | convert_ip_to_linear(struct task_struct *child, struct pt_regs *regs); | 140 | convert_ip_to_linear(struct task_struct *child, struct pt_regs *regs); |
| 141 | 141 | ||
| 142 | #ifdef CONFIG_X86_32 | 142 | #ifdef CONFIG_X86_32 |
| 143 | extern void send_sigtrap(struct task_struct *tsk, struct pt_regs *regs, int error_code); | 143 | extern void send_sigtrap(struct task_struct *tsk, struct pt_regs *regs, |
| 144 | int error_code); | ||
| 144 | #else | 145 | #else |
| 145 | void signal_fault(struct pt_regs *regs, void __user *frame, char *where); | 146 | void signal_fault(struct pt_regs *regs, void __user *frame, char *where); |
| 146 | #endif | 147 | #endif |
| 147 | 148 | ||
| 148 | #define regs_return_value(regs) ((regs)->ax) | 149 | static inline unsigned long regs_return_value(struct pt_regs *regs) |
| 150 | { | ||
| 151 | return regs->ax; | ||
| 152 | } | ||
| 149 | 153 | ||
| 150 | /* | 154 | /* |
| 151 | * user_mode_vm(regs) determines whether a register set came from user mode. | 155 | * user_mode_vm(regs) determines whether a register set came from user mode. |
| @@ -166,8 +170,8 @@ static inline int user_mode(struct pt_regs *regs) | |||
| 166 | static inline int user_mode_vm(struct pt_regs *regs) | 170 | static inline int user_mode_vm(struct pt_regs *regs) |
| 167 | { | 171 | { |
| 168 | #ifdef CONFIG_X86_32 | 172 | #ifdef CONFIG_X86_32 |
| 169 | return ((regs->cs & SEGMENT_RPL_MASK) | | 173 | return ((regs->cs & SEGMENT_RPL_MASK) | (regs->flags & X86_VM_MASK)) >= |
| 170 | (regs->flags & VM_MASK)) >= USER_RPL; | 174 | USER_RPL; |
| 171 | #else | 175 | #else |
| 172 | return user_mode(regs); | 176 | return user_mode(regs); |
| 173 | #endif | 177 | #endif |
| @@ -176,7 +180,7 @@ static inline int user_mode_vm(struct pt_regs *regs) | |||
| 176 | static inline int v8086_mode(struct pt_regs *regs) | 180 | static inline int v8086_mode(struct pt_regs *regs) |
| 177 | { | 181 | { |
| 178 | #ifdef CONFIG_X86_32 | 182 | #ifdef CONFIG_X86_32 |
| 179 | return (regs->flags & VM_MASK); | 183 | return (regs->flags & X86_VM_MASK); |
| 180 | #else | 184 | #else |
| 181 | return 0; /* No V86 mode support in long mode */ | 185 | return 0; /* No V86 mode support in long mode */ |
| 182 | #endif | 186 | #endif |
diff --git a/include/asm-x86/reboot.h b/include/asm-x86/reboot.h index e9e3ffc22c07..6b5233b4f84b 100644 --- a/include/asm-x86/reboot.h +++ b/include/asm-x86/reboot.h | |||
| @@ -3,8 +3,7 @@ | |||
| 3 | 3 | ||
| 4 | struct pt_regs; | 4 | struct pt_regs; |
| 5 | 5 | ||
| 6 | struct machine_ops | 6 | struct machine_ops { |
| 7 | { | ||
| 8 | void (*restart)(char *cmd); | 7 | void (*restart)(char *cmd); |
| 9 | void (*halt)(void); | 8 | void (*halt)(void); |
| 10 | void (*power_off)(void); | 9 | void (*power_off)(void); |
diff --git a/include/asm-x86/resume-trace.h b/include/asm-x86/resume-trace.h index 46f725b0bc82..2557514d7ef6 100644 --- a/include/asm-x86/resume-trace.h +++ b/include/asm-x86/resume-trace.h | |||
| @@ -3,16 +3,17 @@ | |||
| 3 | 3 | ||
| 4 | #include <asm/asm.h> | 4 | #include <asm/asm.h> |
| 5 | 5 | ||
| 6 | #define TRACE_RESUME(user) do { \ | 6 | #define TRACE_RESUME(user) \ |
| 7 | do { \ | ||
| 7 | if (pm_trace_enabled) { \ | 8 | if (pm_trace_enabled) { \ |
| 8 | void *tracedata; \ | 9 | void *tracedata; \ |
| 9 | asm volatile(_ASM_MOV_UL " $1f,%0\n" \ | 10 | asm volatile(_ASM_MOV_UL " $1f,%0\n" \ |
| 10 | ".section .tracedata,\"a\"\n" \ | 11 | ".section .tracedata,\"a\"\n" \ |
| 11 | "1:\t.word %c1\n\t" \ | 12 | "1:\t.word %c1\n\t" \ |
| 12 | _ASM_PTR " %c2\n" \ | 13 | _ASM_PTR " %c2\n" \ |
| 13 | ".previous" \ | 14 | ".previous" \ |
| 14 | :"=r" (tracedata) \ | 15 | :"=r" (tracedata) \ |
| 15 | : "i" (__LINE__), "i" (__FILE__)); \ | 16 | : "i" (__LINE__), "i" (__FILE__)); \ |
| 16 | generate_resume_trace(tracedata, user); \ | 17 | generate_resume_trace(tracedata, user); \ |
| 17 | } \ | 18 | } \ |
| 18 | } while (0) | 19 | } while (0) |
diff --git a/include/asm-x86/rio.h b/include/asm-x86/rio.h index 97cdcc9887ba..3451c576e6af 100644 --- a/include/asm-x86/rio.h +++ b/include/asm-x86/rio.h | |||
| @@ -11,53 +11,53 @@ | |||
| 11 | #define RIO_TABLE_VERSION 3 | 11 | #define RIO_TABLE_VERSION 3 |
| 12 | 12 | ||
| 13 | struct rio_table_hdr { | 13 | struct rio_table_hdr { |
| 14 | u8 version; /* Version number of this data structure */ | 14 | u8 version; /* Version number of this data structure */ |
| 15 | u8 num_scal_dev; /* # of Scalability devices */ | 15 | u8 num_scal_dev; /* # of Scalability devices */ |
| 16 | u8 num_rio_dev; /* # of RIO I/O devices */ | 16 | u8 num_rio_dev; /* # of RIO I/O devices */ |
| 17 | } __attribute__((packed)); | 17 | } __attribute__((packed)); |
| 18 | 18 | ||
| 19 | struct scal_detail { | 19 | struct scal_detail { |
| 20 | u8 node_id; /* Scalability Node ID */ | 20 | u8 node_id; /* Scalability Node ID */ |
| 21 | u32 CBAR; /* Address of 1MB register space */ | 21 | u32 CBAR; /* Address of 1MB register space */ |
| 22 | u8 port0node; /* Node ID port connected to: 0xFF=None */ | 22 | u8 port0node; /* Node ID port connected to: 0xFF=None */ |
| 23 | u8 port0port; /* Port num port connected to: 0,1,2, or */ | 23 | u8 port0port; /* Port num port connected to: 0,1,2, or */ |
| 24 | /* 0xFF=None */ | 24 | /* 0xFF=None */ |
| 25 | u8 port1node; /* Node ID port connected to: 0xFF = None */ | 25 | u8 port1node; /* Node ID port connected to: 0xFF = None */ |
| 26 | u8 port1port; /* Port num port connected to: 0,1,2, or */ | 26 | u8 port1port; /* Port num port connected to: 0,1,2, or */ |
| 27 | /* 0xFF=None */ | 27 | /* 0xFF=None */ |
| 28 | u8 port2node; /* Node ID port connected to: 0xFF = None */ | 28 | u8 port2node; /* Node ID port connected to: 0xFF = None */ |
| 29 | u8 port2port; /* Port num port connected to: 0,1,2, or */ | 29 | u8 port2port; /* Port num port connected to: 0,1,2, or */ |
| 30 | /* 0xFF=None */ | 30 | /* 0xFF=None */ |
| 31 | u8 chassis_num; /* 1 based Chassis number (1 = boot node) */ | 31 | u8 chassis_num; /* 1 based Chassis number (1 = boot node) */ |
| 32 | } __attribute__((packed)); | 32 | } __attribute__((packed)); |
| 33 | 33 | ||
| 34 | struct rio_detail { | 34 | struct rio_detail { |
| 35 | u8 node_id; /* RIO Node ID */ | 35 | u8 node_id; /* RIO Node ID */ |
| 36 | u32 BBAR; /* Address of 1MB register space */ | 36 | u32 BBAR; /* Address of 1MB register space */ |
| 37 | u8 type; /* Type of device */ | 37 | u8 type; /* Type of device */ |
| 38 | u8 owner_id; /* Node ID of Hurricane that owns this */ | 38 | u8 owner_id; /* Node ID of Hurricane that owns this */ |
| 39 | /* node */ | 39 | /* node */ |
| 40 | u8 port0node; /* Node ID port connected to: 0xFF=None */ | 40 | u8 port0node; /* Node ID port connected to: 0xFF=None */ |
| 41 | u8 port0port; /* Port num port connected to: 0,1,2, or */ | 41 | u8 port0port; /* Port num port connected to: 0,1,2, or */ |
| 42 | /* 0xFF=None */ | 42 | /* 0xFF=None */ |
| 43 | u8 port1node; /* Node ID port connected to: 0xFF=None */ | 43 | u8 port1node; /* Node ID port connected to: 0xFF=None */ |
| 44 | u8 port1port; /* Port num port connected to: 0,1,2, or */ | 44 | u8 port1port; /* Port num port connected to: 0,1,2, or */ |
| 45 | /* 0xFF=None */ | 45 | /* 0xFF=None */ |
| 46 | u8 first_slot; /* Lowest slot number below this Calgary */ | 46 | u8 first_slot; /* Lowest slot number below this Calgary */ |
| 47 | u8 status; /* Bit 0 = 1 : the XAPIC is used */ | 47 | u8 status; /* Bit 0 = 1 : the XAPIC is used */ |
| 48 | /* = 0 : the XAPIC is not used, ie: */ | 48 | /* = 0 : the XAPIC is not used, ie: */ |
| 49 | /* ints fwded to another XAPIC */ | 49 | /* ints fwded to another XAPIC */ |
| 50 | /* Bits1:7 Reserved */ | 50 | /* Bits1:7 Reserved */ |
| 51 | u8 WP_index; /* instance index - lower ones have */ | 51 | u8 WP_index; /* instance index - lower ones have */ |
| 52 | /* lower slot numbers/PCI bus numbers */ | 52 | /* lower slot numbers/PCI bus numbers */ |
| 53 | u8 chassis_num; /* 1 based Chassis number */ | 53 | u8 chassis_num; /* 1 based Chassis number */ |
| 54 | } __attribute__((packed)); | 54 | } __attribute__((packed)); |
| 55 | 55 | ||
| 56 | enum { | 56 | enum { |
| 57 | HURR_SCALABILTY = 0, /* Hurricane Scalability info */ | 57 | HURR_SCALABILTY = 0, /* Hurricane Scalability info */ |
| 58 | HURR_RIOIB = 2, /* Hurricane RIOIB info */ | 58 | HURR_RIOIB = 2, /* Hurricane RIOIB info */ |
| 59 | COMPAT_CALGARY = 4, /* Compatibility Calgary */ | 59 | COMPAT_CALGARY = 4, /* Compatibility Calgary */ |
| 60 | ALT_CALGARY = 5, /* Second Planar Calgary */ | 60 | ALT_CALGARY = 5, /* Second Planar Calgary */ |
| 61 | }; | 61 | }; |
| 62 | 62 | ||
| 63 | /* | 63 | /* |
diff --git a/include/asm-x86/rwsem.h b/include/asm-x86/rwsem.h index 520a379f4b80..750f2a3542b3 100644 --- a/include/asm-x86/rwsem.h +++ b/include/asm-x86/rwsem.h | |||
| @@ -56,14 +56,16 @@ extern asmregparm struct rw_semaphore * | |||
| 56 | /* | 56 | /* |
| 57 | * the semaphore definition | 57 | * the semaphore definition |
| 58 | */ | 58 | */ |
| 59 | struct rw_semaphore { | 59 | |
| 60 | signed long count; | ||
| 61 | #define RWSEM_UNLOCKED_VALUE 0x00000000 | 60 | #define RWSEM_UNLOCKED_VALUE 0x00000000 |
| 62 | #define RWSEM_ACTIVE_BIAS 0x00000001 | 61 | #define RWSEM_ACTIVE_BIAS 0x00000001 |
| 63 | #define RWSEM_ACTIVE_MASK 0x0000ffff | 62 | #define RWSEM_ACTIVE_MASK 0x0000ffff |
| 64 | #define RWSEM_WAITING_BIAS (-0x00010000) | 63 | #define RWSEM_WAITING_BIAS (-0x00010000) |
| 65 | #define RWSEM_ACTIVE_READ_BIAS RWSEM_ACTIVE_BIAS | 64 | #define RWSEM_ACTIVE_READ_BIAS RWSEM_ACTIVE_BIAS |
| 66 | #define RWSEM_ACTIVE_WRITE_BIAS (RWSEM_WAITING_BIAS + RWSEM_ACTIVE_BIAS) | 65 | #define RWSEM_ACTIVE_WRITE_BIAS (RWSEM_WAITING_BIAS + RWSEM_ACTIVE_BIAS) |
| 66 | |||
| 67 | struct rw_semaphore { | ||
| 68 | signed long count; | ||
| 67 | spinlock_t wait_lock; | 69 | spinlock_t wait_lock; |
| 68 | struct list_head wait_list; | 70 | struct list_head wait_list; |
| 69 | #ifdef CONFIG_DEBUG_LOCK_ALLOC | 71 | #ifdef CONFIG_DEBUG_LOCK_ALLOC |
| @@ -78,11 +80,13 @@ struct rw_semaphore { | |||
| 78 | #endif | 80 | #endif |
| 79 | 81 | ||
| 80 | 82 | ||
| 81 | #define __RWSEM_INITIALIZER(name) \ | 83 | #define __RWSEM_INITIALIZER(name) \ |
| 82 | { RWSEM_UNLOCKED_VALUE, __SPIN_LOCK_UNLOCKED((name).wait_lock), \ | 84 | { \ |
| 83 | LIST_HEAD_INIT((name).wait_list) __RWSEM_DEP_MAP_INIT(name) } | 85 | RWSEM_UNLOCKED_VALUE, __SPIN_LOCK_UNLOCKED((name).wait_lock), \ |
| 86 | LIST_HEAD_INIT((name).wait_list) __RWSEM_DEP_MAP_INIT(name) \ | ||
| 87 | } | ||
| 84 | 88 | ||
| 85 | #define DECLARE_RWSEM(name) \ | 89 | #define DECLARE_RWSEM(name) \ |
| 86 | struct rw_semaphore name = __RWSEM_INITIALIZER(name) | 90 | struct rw_semaphore name = __RWSEM_INITIALIZER(name) |
| 87 | 91 | ||
| 88 | extern void __init_rwsem(struct rw_semaphore *sem, const char *name, | 92 | extern void __init_rwsem(struct rw_semaphore *sem, const char *name, |
| @@ -100,16 +104,16 @@ do { \ | |||
| 100 | */ | 104 | */ |
| 101 | static inline void __down_read(struct rw_semaphore *sem) | 105 | static inline void __down_read(struct rw_semaphore *sem) |
| 102 | { | 106 | { |
| 103 | __asm__ __volatile__( | 107 | asm volatile("# beginning down_read\n\t" |
| 104 | "# beginning down_read\n\t" | 108 | LOCK_PREFIX " incl (%%eax)\n\t" |
| 105 | LOCK_PREFIX " incl (%%eax)\n\t" /* adds 0x00000001, returns the old value */ | 109 | /* adds 0x00000001, returns the old value */ |
| 106 | " jns 1f\n" | 110 | " jns 1f\n" |
| 107 | " call call_rwsem_down_read_failed\n" | 111 | " call call_rwsem_down_read_failed\n" |
| 108 | "1:\n\t" | 112 | "1:\n\t" |
| 109 | "# ending down_read\n\t" | 113 | "# ending down_read\n\t" |
| 110 | : "+m" (sem->count) | 114 | : "+m" (sem->count) |
| 111 | : "a" (sem) | 115 | : "a" (sem) |
| 112 | : "memory", "cc"); | 116 | : "memory", "cc"); |
| 113 | } | 117 | } |
| 114 | 118 | ||
| 115 | /* | 119 | /* |
| @@ -118,21 +122,20 @@ LOCK_PREFIX " incl (%%eax)\n\t" /* adds 0x00000001, returns the old value | |||
| 118 | static inline int __down_read_trylock(struct rw_semaphore *sem) | 122 | static inline int __down_read_trylock(struct rw_semaphore *sem) |
| 119 | { | 123 | { |
| 120 | __s32 result, tmp; | 124 | __s32 result, tmp; |
| 121 | __asm__ __volatile__( | 125 | asm volatile("# beginning __down_read_trylock\n\t" |
| 122 | "# beginning __down_read_trylock\n\t" | 126 | " movl %0,%1\n\t" |
| 123 | " movl %0,%1\n\t" | 127 | "1:\n\t" |
| 124 | "1:\n\t" | 128 | " movl %1,%2\n\t" |
| 125 | " movl %1,%2\n\t" | 129 | " addl %3,%2\n\t" |
| 126 | " addl %3,%2\n\t" | 130 | " jle 2f\n\t" |
| 127 | " jle 2f\n\t" | 131 | LOCK_PREFIX " cmpxchgl %2,%0\n\t" |
| 128 | LOCK_PREFIX " cmpxchgl %2,%0\n\t" | 132 | " jnz 1b\n\t" |
| 129 | " jnz 1b\n\t" | 133 | "2:\n\t" |
| 130 | "2:\n\t" | 134 | "# ending __down_read_trylock\n\t" |
| 131 | "# ending __down_read_trylock\n\t" | 135 | : "+m" (sem->count), "=&a" (result), "=&r" (tmp) |
| 132 | : "+m" (sem->count), "=&a" (result), "=&r" (tmp) | 136 | : "i" (RWSEM_ACTIVE_READ_BIAS) |
| 133 | : "i" (RWSEM_ACTIVE_READ_BIAS) | 137 | : "memory", "cc"); |
| 134 | : "memory", "cc"); | 138 | return result >= 0 ? 1 : 0; |
| 135 | return result>=0 ? 1 : 0; | ||
| 136 | } | 139 | } |
| 137 | 140 | ||
| 138 | /* | 141 | /* |
| @@ -143,17 +146,18 @@ static inline void __down_write_nested(struct rw_semaphore *sem, int subclass) | |||
| 143 | int tmp; | 146 | int tmp; |
| 144 | 147 | ||
| 145 | tmp = RWSEM_ACTIVE_WRITE_BIAS; | 148 | tmp = RWSEM_ACTIVE_WRITE_BIAS; |
| 146 | __asm__ __volatile__( | 149 | asm volatile("# beginning down_write\n\t" |
| 147 | "# beginning down_write\n\t" | 150 | LOCK_PREFIX " xadd %%edx,(%%eax)\n\t" |
| 148 | LOCK_PREFIX " xadd %%edx,(%%eax)\n\t" /* subtract 0x0000ffff, returns the old value */ | 151 | /* subtract 0x0000ffff, returns the old value */ |
| 149 | " testl %%edx,%%edx\n\t" /* was the count 0 before? */ | 152 | " testl %%edx,%%edx\n\t" |
| 150 | " jz 1f\n" | 153 | /* was the count 0 before? */ |
| 151 | " call call_rwsem_down_write_failed\n" | 154 | " jz 1f\n" |
| 152 | "1:\n" | 155 | " call call_rwsem_down_write_failed\n" |
| 153 | "# ending down_write" | 156 | "1:\n" |
| 154 | : "+m" (sem->count), "=d" (tmp) | 157 | "# ending down_write" |
| 155 | : "a" (sem), "1" (tmp) | 158 | : "+m" (sem->count), "=d" (tmp) |
| 156 | : "memory", "cc"); | 159 | : "a" (sem), "1" (tmp) |
| 160 | : "memory", "cc"); | ||
| 157 | } | 161 | } |
| 158 | 162 | ||
| 159 | static inline void __down_write(struct rw_semaphore *sem) | 163 | static inline void __down_write(struct rw_semaphore *sem) |
| @@ -167,7 +171,7 @@ static inline void __down_write(struct rw_semaphore *sem) | |||
| 167 | static inline int __down_write_trylock(struct rw_semaphore *sem) | 171 | static inline int __down_write_trylock(struct rw_semaphore *sem) |
| 168 | { | 172 | { |
| 169 | signed long ret = cmpxchg(&sem->count, | 173 | signed long ret = cmpxchg(&sem->count, |
| 170 | RWSEM_UNLOCKED_VALUE, | 174 | RWSEM_UNLOCKED_VALUE, |
| 171 | RWSEM_ACTIVE_WRITE_BIAS); | 175 | RWSEM_ACTIVE_WRITE_BIAS); |
| 172 | if (ret == RWSEM_UNLOCKED_VALUE) | 176 | if (ret == RWSEM_UNLOCKED_VALUE) |
| 173 | return 1; | 177 | return 1; |
| @@ -180,16 +184,16 @@ static inline int __down_write_trylock(struct rw_semaphore *sem) | |||
| 180 | static inline void __up_read(struct rw_semaphore *sem) | 184 | static inline void __up_read(struct rw_semaphore *sem) |
| 181 | { | 185 | { |
| 182 | __s32 tmp = -RWSEM_ACTIVE_READ_BIAS; | 186 | __s32 tmp = -RWSEM_ACTIVE_READ_BIAS; |
| 183 | __asm__ __volatile__( | 187 | asm volatile("# beginning __up_read\n\t" |
| 184 | "# beginning __up_read\n\t" | 188 | LOCK_PREFIX " xadd %%edx,(%%eax)\n\t" |
| 185 | LOCK_PREFIX " xadd %%edx,(%%eax)\n\t" /* subtracts 1, returns the old value */ | 189 | /* subtracts 1, returns the old value */ |
| 186 | " jns 1f\n\t" | 190 | " jns 1f\n\t" |
| 187 | " call call_rwsem_wake\n" | 191 | " call call_rwsem_wake\n" |
| 188 | "1:\n" | 192 | "1:\n" |
| 189 | "# ending __up_read\n" | 193 | "# ending __up_read\n" |
| 190 | : "+m" (sem->count), "=d" (tmp) | 194 | : "+m" (sem->count), "=d" (tmp) |
| 191 | : "a" (sem), "1" (tmp) | 195 | : "a" (sem), "1" (tmp) |
| 192 | : "memory", "cc"); | 196 | : "memory", "cc"); |
| 193 | } | 197 | } |
| 194 | 198 | ||
| 195 | /* | 199 | /* |
| @@ -197,17 +201,18 @@ LOCK_PREFIX " xadd %%edx,(%%eax)\n\t" /* subtracts 1, returns the old valu | |||
| 197 | */ | 201 | */ |
| 198 | static inline void __up_write(struct rw_semaphore *sem) | 202 | static inline void __up_write(struct rw_semaphore *sem) |
| 199 | { | 203 | { |
| 200 | __asm__ __volatile__( | 204 | asm volatile("# beginning __up_write\n\t" |
| 201 | "# beginning __up_write\n\t" | 205 | " movl %2,%%edx\n\t" |
| 202 | " movl %2,%%edx\n\t" | 206 | LOCK_PREFIX " xaddl %%edx,(%%eax)\n\t" |
| 203 | LOCK_PREFIX " xaddl %%edx,(%%eax)\n\t" /* tries to transition 0xffff0001 -> 0x00000000 */ | 207 | /* tries to transition |
| 204 | " jz 1f\n" | 208 | 0xffff0001 -> 0x00000000 */ |
| 205 | " call call_rwsem_wake\n" | 209 | " jz 1f\n" |
| 206 | "1:\n\t" | 210 | " call call_rwsem_wake\n" |
| 207 | "# ending __up_write\n" | 211 | "1:\n\t" |
| 208 | : "+m" (sem->count) | 212 | "# ending __up_write\n" |
| 209 | : "a" (sem), "i" (-RWSEM_ACTIVE_WRITE_BIAS) | 213 | : "+m" (sem->count) |
| 210 | : "memory", "cc", "edx"); | 214 | : "a" (sem), "i" (-RWSEM_ACTIVE_WRITE_BIAS) |
| 215 | : "memory", "cc", "edx"); | ||
| 211 | } | 216 | } |
| 212 | 217 | ||
| 213 | /* | 218 | /* |
| @@ -215,16 +220,16 @@ LOCK_PREFIX " xaddl %%edx,(%%eax)\n\t" /* tries to transition 0xffff0001 -> | |||
| 215 | */ | 220 | */ |
| 216 | static inline void __downgrade_write(struct rw_semaphore *sem) | 221 | static inline void __downgrade_write(struct rw_semaphore *sem) |
| 217 | { | 222 | { |
| 218 | __asm__ __volatile__( | 223 | asm volatile("# beginning __downgrade_write\n\t" |
| 219 | "# beginning __downgrade_write\n\t" | 224 | LOCK_PREFIX " addl %2,(%%eax)\n\t" |
| 220 | LOCK_PREFIX " addl %2,(%%eax)\n\t" /* transitions 0xZZZZ0001 -> 0xYYYY0001 */ | 225 | /* transitions 0xZZZZ0001 -> 0xYYYY0001 */ |
| 221 | " jns 1f\n\t" | 226 | " jns 1f\n\t" |
| 222 | " call call_rwsem_downgrade_wake\n" | 227 | " call call_rwsem_downgrade_wake\n" |
| 223 | "1:\n\t" | 228 | "1:\n\t" |
| 224 | "# ending __downgrade_write\n" | 229 | "# ending __downgrade_write\n" |
| 225 | : "+m" (sem->count) | 230 | : "+m" (sem->count) |
| 226 | : "a" (sem), "i" (-RWSEM_WAITING_BIAS) | 231 | : "a" (sem), "i" (-RWSEM_WAITING_BIAS) |
| 227 | : "memory", "cc"); | 232 | : "memory", "cc"); |
| 228 | } | 233 | } |
| 229 | 234 | ||
| 230 | /* | 235 | /* |
| @@ -232,10 +237,9 @@ LOCK_PREFIX " addl %2,(%%eax)\n\t" /* transitions 0xZZZZ0001 -> 0xYYYY0001 | |||
| 232 | */ | 237 | */ |
| 233 | static inline void rwsem_atomic_add(int delta, struct rw_semaphore *sem) | 238 | static inline void rwsem_atomic_add(int delta, struct rw_semaphore *sem) |
| 234 | { | 239 | { |
| 235 | __asm__ __volatile__( | 240 | asm volatile(LOCK_PREFIX "addl %1,%0" |
| 236 | LOCK_PREFIX "addl %1,%0" | 241 | : "+m" (sem->count) |
| 237 | : "+m" (sem->count) | 242 | : "ir" (delta)); |
| 238 | : "ir" (delta)); | ||
| 239 | } | 243 | } |
| 240 | 244 | ||
| 241 | /* | 245 | /* |
| @@ -245,12 +249,11 @@ static inline int rwsem_atomic_update(int delta, struct rw_semaphore *sem) | |||
| 245 | { | 249 | { |
| 246 | int tmp = delta; | 250 | int tmp = delta; |
| 247 | 251 | ||
| 248 | __asm__ __volatile__( | 252 | asm volatile(LOCK_PREFIX "xadd %0,%1" |
| 249 | LOCK_PREFIX "xadd %0,%1" | 253 | : "+r" (tmp), "+m" (sem->count) |
| 250 | : "+r" (tmp), "+m" (sem->count) | 254 | : : "memory"); |
| 251 | : : "memory"); | ||
| 252 | 255 | ||
| 253 | return tmp+delta; | 256 | return tmp + delta; |
| 254 | } | 257 | } |
| 255 | 258 | ||
| 256 | static inline int rwsem_is_locked(struct rw_semaphore *sem) | 259 | static inline int rwsem_is_locked(struct rw_semaphore *sem) |
diff --git a/include/asm-x86/segment.h b/include/asm-x86/segment.h index 23f0535fec61..ed5131dd7d92 100644 --- a/include/asm-x86/segment.h +++ b/include/asm-x86/segment.h | |||
| @@ -191,13 +191,14 @@ | |||
| 191 | #define SEGMENT_TI_MASK 0x4 | 191 | #define SEGMENT_TI_MASK 0x4 |
| 192 | 192 | ||
| 193 | #define IDT_ENTRIES 256 | 193 | #define IDT_ENTRIES 256 |
| 194 | #define NUM_EXCEPTION_VECTORS 32 | ||
| 194 | #define GDT_SIZE (GDT_ENTRIES * 8) | 195 | #define GDT_SIZE (GDT_ENTRIES * 8) |
| 195 | #define GDT_ENTRY_TLS_ENTRIES 3 | 196 | #define GDT_ENTRY_TLS_ENTRIES 3 |
| 196 | #define TLS_SIZE (GDT_ENTRY_TLS_ENTRIES * 8) | 197 | #define TLS_SIZE (GDT_ENTRY_TLS_ENTRIES * 8) |
| 197 | 198 | ||
| 198 | #ifdef __KERNEL__ | 199 | #ifdef __KERNEL__ |
| 199 | #ifndef __ASSEMBLY__ | 200 | #ifndef __ASSEMBLY__ |
| 200 | extern const char early_idt_handlers[IDT_ENTRIES][10]; | 201 | extern const char early_idt_handlers[NUM_EXCEPTION_VECTORS][10]; |
| 201 | #endif | 202 | #endif |
| 202 | #endif | 203 | #endif |
| 203 | 204 | ||
diff --git a/include/asm-x86/setup.h b/include/asm-x86/setup.h index 071e054abd82..fa6763af8d26 100644 --- a/include/asm-x86/setup.h +++ b/include/asm-x86/setup.h | |||
| @@ -4,6 +4,10 @@ | |||
| 4 | #define COMMAND_LINE_SIZE 2048 | 4 | #define COMMAND_LINE_SIZE 2048 |
| 5 | 5 | ||
| 6 | #ifndef __ASSEMBLY__ | 6 | #ifndef __ASSEMBLY__ |
| 7 | |||
| 8 | /* Interrupt control for vSMPowered x86_64 systems */ | ||
| 9 | void vsmp_init(void); | ||
| 10 | |||
| 7 | char *machine_specific_memory_setup(void); | 11 | char *machine_specific_memory_setup(void); |
| 8 | #ifndef CONFIG_PARAVIRT | 12 | #ifndef CONFIG_PARAVIRT |
| 9 | #define paravirt_post_allocator_init() do {} while (0) | 13 | #define paravirt_post_allocator_init() do {} while (0) |
| @@ -51,8 +55,8 @@ struct e820entry; | |||
| 51 | char * __init machine_specific_memory_setup(void); | 55 | char * __init machine_specific_memory_setup(void); |
| 52 | char *memory_setup(void); | 56 | char *memory_setup(void); |
| 53 | 57 | ||
| 54 | int __init copy_e820_map(struct e820entry * biosmap, int nr_map); | 58 | int __init copy_e820_map(struct e820entry *biosmap, int nr_map); |
| 55 | int __init sanitize_e820_map(struct e820entry * biosmap, char * pnr_map); | 59 | int __init sanitize_e820_map(struct e820entry *biosmap, char *pnr_map); |
| 56 | void __init add_memory_region(unsigned long long start, | 60 | void __init add_memory_region(unsigned long long start, |
| 57 | unsigned long long size, int type); | 61 | unsigned long long size, int type); |
| 58 | 62 | ||
diff --git a/include/asm-x86/sigcontext.h b/include/asm-x86/sigcontext.h index d743947f4c77..2f9c884d2c0f 100644 --- a/include/asm-x86/sigcontext.h +++ b/include/asm-x86/sigcontext.h | |||
| @@ -79,7 +79,7 @@ struct sigcontext { | |||
| 79 | unsigned long flags; | 79 | unsigned long flags; |
| 80 | unsigned long sp_at_signal; | 80 | unsigned long sp_at_signal; |
| 81 | unsigned short ss, __ssh; | 81 | unsigned short ss, __ssh; |
| 82 | struct _fpstate __user * fpstate; | 82 | struct _fpstate __user *fpstate; |
| 83 | unsigned long oldmask; | 83 | unsigned long oldmask; |
| 84 | unsigned long cr2; | 84 | unsigned long cr2; |
| 85 | }; | 85 | }; |
| @@ -107,7 +107,7 @@ struct sigcontext { | |||
| 107 | unsigned long eflags; | 107 | unsigned long eflags; |
| 108 | unsigned long esp_at_signal; | 108 | unsigned long esp_at_signal; |
| 109 | unsigned short ss, __ssh; | 109 | unsigned short ss, __ssh; |
| 110 | struct _fpstate __user * fpstate; | 110 | struct _fpstate __user *fpstate; |
| 111 | unsigned long oldmask; | 111 | unsigned long oldmask; |
| 112 | unsigned long cr2; | 112 | unsigned long cr2; |
| 113 | }; | 113 | }; |
| @@ -121,7 +121,8 @@ struct sigcontext { | |||
| 121 | struct _fpstate { | 121 | struct _fpstate { |
| 122 | __u16 cwd; | 122 | __u16 cwd; |
| 123 | __u16 swd; | 123 | __u16 swd; |
| 124 | __u16 twd; /* Note this is not the same as the 32bit/x87/FSAVE twd */ | 124 | __u16 twd; /* Note this is not the same as the |
| 125 | 32bit/x87/FSAVE twd */ | ||
| 125 | __u16 fop; | 126 | __u16 fop; |
| 126 | __u64 rip; | 127 | __u64 rip; |
| 127 | __u64 rdp; | 128 | __u64 rdp; |
diff --git a/include/asm-x86/sigcontext32.h b/include/asm-x86/sigcontext32.h index 6ffab4fd593a..57a9686fb491 100644 --- a/include/asm-x86/sigcontext32.h +++ b/include/asm-x86/sigcontext32.h | |||
| @@ -26,7 +26,7 @@ struct _fpstate_ia32 { | |||
| 26 | __u32 cw; | 26 | __u32 cw; |
| 27 | __u32 sw; | 27 | __u32 sw; |
| 28 | __u32 tag; /* not compatible to 64bit twd */ | 28 | __u32 tag; /* not compatible to 64bit twd */ |
| 29 | __u32 ipoff; | 29 | __u32 ipoff; |
| 30 | __u32 cssel; | 30 | __u32 cssel; |
| 31 | __u32 dataoff; | 31 | __u32 dataoff; |
| 32 | __u32 datasel; | 32 | __u32 datasel; |
| @@ -39,7 +39,7 @@ struct _fpstate_ia32 { | |||
| 39 | __u32 mxcsr; | 39 | __u32 mxcsr; |
| 40 | __u32 reserved; | 40 | __u32 reserved; |
| 41 | struct _fpxreg _fxsr_st[8]; | 41 | struct _fpxreg _fxsr_st[8]; |
| 42 | struct _xmmreg _xmm[8]; /* It's actually 16 */ | 42 | struct _xmmreg _xmm[8]; /* It's actually 16 */ |
| 43 | __u32 padding[56]; | 43 | __u32 padding[56]; |
| 44 | }; | 44 | }; |
| 45 | 45 | ||
diff --git a/include/asm-x86/signal.h b/include/asm-x86/signal.h index aee7eca585ab..f15186d39c69 100644 --- a/include/asm-x86/signal.h +++ b/include/asm-x86/signal.h | |||
| @@ -185,61 +185,61 @@ typedef struct sigaltstack { | |||
| 185 | 185 | ||
| 186 | #define __HAVE_ARCH_SIG_BITOPS | 186 | #define __HAVE_ARCH_SIG_BITOPS |
| 187 | 187 | ||
| 188 | #define sigaddset(set,sig) \ | 188 | #define sigaddset(set,sig) \ |
| 189 | (__builtin_constantp(sig) ? \ | 189 | (__builtin_constantp(sig) \ |
| 190 | __const_sigaddset((set),(sig)) : \ | 190 | ? __const_sigaddset((set), (sig)) \ |
| 191 | __gen_sigaddset((set),(sig))) | 191 | : __gen_sigaddset((set), (sig))) |
| 192 | 192 | ||
| 193 | static __inline__ void __gen_sigaddset(sigset_t *set, int _sig) | 193 | static inline void __gen_sigaddset(sigset_t *set, int _sig) |
| 194 | { | 194 | { |
| 195 | __asm__("btsl %1,%0" : "+m"(*set) : "Ir"(_sig - 1) : "cc"); | 195 | asm("btsl %1,%0" : "+m"(*set) : "Ir"(_sig - 1) : "cc"); |
| 196 | } | 196 | } |
| 197 | 197 | ||
| 198 | static __inline__ void __const_sigaddset(sigset_t *set, int _sig) | 198 | static inline void __const_sigaddset(sigset_t *set, int _sig) |
| 199 | { | 199 | { |
| 200 | unsigned long sig = _sig - 1; | 200 | unsigned long sig = _sig - 1; |
| 201 | set->sig[sig / _NSIG_BPW] |= 1 << (sig % _NSIG_BPW); | 201 | set->sig[sig / _NSIG_BPW] |= 1 << (sig % _NSIG_BPW); |
| 202 | } | 202 | } |
| 203 | 203 | ||
| 204 | #define sigdelset(set,sig) \ | 204 | #define sigdelset(set, sig) \ |
| 205 | (__builtin_constant_p(sig) ? \ | 205 | (__builtin_constant_p(sig) \ |
| 206 | __const_sigdelset((set),(sig)) : \ | 206 | ? __const_sigdelset((set), (sig)) \ |
| 207 | __gen_sigdelset((set),(sig))) | 207 | : __gen_sigdelset((set), (sig))) |
| 208 | 208 | ||
| 209 | 209 | ||
| 210 | static __inline__ void __gen_sigdelset(sigset_t *set, int _sig) | 210 | static inline void __gen_sigdelset(sigset_t *set, int _sig) |
| 211 | { | 211 | { |
| 212 | __asm__("btrl %1,%0" : "+m"(*set) : "Ir"(_sig - 1) : "cc"); | 212 | asm("btrl %1,%0" : "+m"(*set) : "Ir"(_sig - 1) : "cc"); |
| 213 | } | 213 | } |
| 214 | 214 | ||
| 215 | static __inline__ void __const_sigdelset(sigset_t *set, int _sig) | 215 | static inline void __const_sigdelset(sigset_t *set, int _sig) |
| 216 | { | 216 | { |
| 217 | unsigned long sig = _sig - 1; | 217 | unsigned long sig = _sig - 1; |
| 218 | set->sig[sig / _NSIG_BPW] &= ~(1 << (sig % _NSIG_BPW)); | 218 | set->sig[sig / _NSIG_BPW] &= ~(1 << (sig % _NSIG_BPW)); |
| 219 | } | 219 | } |
| 220 | 220 | ||
| 221 | static __inline__ int __const_sigismember(sigset_t *set, int _sig) | 221 | static inline int __const_sigismember(sigset_t *set, int _sig) |
| 222 | { | 222 | { |
| 223 | unsigned long sig = _sig - 1; | 223 | unsigned long sig = _sig - 1; |
| 224 | return 1 & (set->sig[sig / _NSIG_BPW] >> (sig % _NSIG_BPW)); | 224 | return 1 & (set->sig[sig / _NSIG_BPW] >> (sig % _NSIG_BPW)); |
| 225 | } | 225 | } |
| 226 | 226 | ||
| 227 | static __inline__ int __gen_sigismember(sigset_t *set, int _sig) | 227 | static inline int __gen_sigismember(sigset_t *set, int _sig) |
| 228 | { | 228 | { |
| 229 | int ret; | 229 | int ret; |
| 230 | __asm__("btl %2,%1\n\tsbbl %0,%0" | 230 | asm("btl %2,%1\n\tsbbl %0,%0" |
| 231 | : "=r"(ret) : "m"(*set), "Ir"(_sig-1) : "cc"); | 231 | : "=r"(ret) : "m"(*set), "Ir"(_sig-1) : "cc"); |
| 232 | return ret; | 232 | return ret; |
| 233 | } | 233 | } |
| 234 | 234 | ||
| 235 | #define sigismember(set,sig) \ | 235 | #define sigismember(set, sig) \ |
| 236 | (__builtin_constant_p(sig) ? \ | 236 | (__builtin_constant_p(sig) \ |
| 237 | __const_sigismember((set),(sig)) : \ | 237 | ? __const_sigismember((set), (sig)) \ |
| 238 | __gen_sigismember((set),(sig))) | 238 | : __gen_sigismember((set), (sig))) |
| 239 | 239 | ||
| 240 | static __inline__ int sigfindinword(unsigned long word) | 240 | static inline int sigfindinword(unsigned long word) |
| 241 | { | 241 | { |
| 242 | __asm__("bsfl %1,%0" : "=r"(word) : "rm"(word) : "cc"); | 242 | asm("bsfl %1,%0" : "=r"(word) : "rm"(word) : "cc"); |
| 243 | return word; | 243 | return word; |
| 244 | } | 244 | } |
| 245 | 245 | ||
diff --git a/include/asm-x86/smp.h b/include/asm-x86/smp.h index f2e8319a6b0b..62ebdec394b9 100644 --- a/include/asm-x86/smp.h +++ b/include/asm-x86/smp.h | |||
| @@ -1,5 +1,209 @@ | |||
| 1 | #ifdef CONFIG_X86_32 | 1 | #ifndef _ASM_X86_SMP_H_ |
| 2 | # include "smp_32.h" | 2 | #define _ASM_X86_SMP_H_ |
| 3 | #ifndef __ASSEMBLY__ | ||
| 4 | #include <linux/cpumask.h> | ||
| 5 | #include <linux/init.h> | ||
| 6 | #include <asm/percpu.h> | ||
| 7 | |||
| 8 | /* | ||
| 9 | * We need the APIC definitions automatically as part of 'smp.h' | ||
| 10 | */ | ||
| 11 | #ifdef CONFIG_X86_LOCAL_APIC | ||
| 12 | # include <asm/mpspec.h> | ||
| 13 | # include <asm/apic.h> | ||
| 14 | # ifdef CONFIG_X86_IO_APIC | ||
| 15 | # include <asm/io_apic.h> | ||
| 16 | # endif | ||
| 17 | #endif | ||
| 18 | #include <asm/pda.h> | ||
| 19 | #include <asm/thread_info.h> | ||
| 20 | |||
| 21 | extern cpumask_t cpu_callout_map; | ||
| 22 | extern cpumask_t cpu_initialized; | ||
| 23 | extern cpumask_t cpu_callin_map; | ||
| 24 | |||
| 25 | extern void (*mtrr_hook)(void); | ||
| 26 | extern void zap_low_mappings(void); | ||
| 27 | |||
| 28 | extern int smp_num_siblings; | ||
| 29 | extern unsigned int num_processors; | ||
| 30 | extern cpumask_t cpu_initialized; | ||
| 31 | |||
| 32 | #ifdef CONFIG_SMP | ||
| 33 | extern u16 x86_cpu_to_apicid_init[]; | ||
| 34 | extern u16 x86_bios_cpu_apicid_init[]; | ||
| 35 | extern void *x86_cpu_to_apicid_early_ptr; | ||
| 36 | extern void *x86_bios_cpu_apicid_early_ptr; | ||
| 3 | #else | 37 | #else |
| 4 | # include "smp_64.h" | 38 | #define x86_cpu_to_apicid_early_ptr NULL |
| 39 | #define x86_bios_cpu_apicid_early_ptr NULL | ||
| 40 | #endif | ||
| 41 | |||
| 42 | DECLARE_PER_CPU(cpumask_t, cpu_sibling_map); | ||
| 43 | DECLARE_PER_CPU(cpumask_t, cpu_core_map); | ||
| 44 | DECLARE_PER_CPU(u16, cpu_llc_id); | ||
| 45 | DECLARE_PER_CPU(u16, x86_cpu_to_apicid); | ||
| 46 | DECLARE_PER_CPU(u16, x86_bios_cpu_apicid); | ||
| 47 | |||
| 48 | /* Static state in head.S used to set up a CPU */ | ||
| 49 | extern struct { | ||
| 50 | void *sp; | ||
| 51 | unsigned short ss; | ||
| 52 | } stack_start; | ||
| 53 | |||
| 54 | struct smp_ops { | ||
| 55 | void (*smp_prepare_boot_cpu)(void); | ||
| 56 | void (*smp_prepare_cpus)(unsigned max_cpus); | ||
| 57 | int (*cpu_up)(unsigned cpu); | ||
| 58 | void (*smp_cpus_done)(unsigned max_cpus); | ||
| 59 | |||
| 60 | void (*smp_send_stop)(void); | ||
| 61 | void (*smp_send_reschedule)(int cpu); | ||
| 62 | int (*smp_call_function_mask)(cpumask_t mask, | ||
| 63 | void (*func)(void *info), void *info, | ||
| 64 | int wait); | ||
| 65 | }; | ||
| 66 | |||
| 67 | /* Globals due to paravirt */ | ||
| 68 | extern void set_cpu_sibling_map(int cpu); | ||
| 69 | |||
| 70 | #ifdef CONFIG_SMP | ||
| 71 | #ifndef CONFIG_PARAVIRT | ||
| 72 | #define startup_ipi_hook(phys_apicid, start_eip, start_esp) do { } while (0) | ||
| 73 | #endif | ||
| 74 | extern struct smp_ops smp_ops; | ||
| 75 | |||
| 76 | static inline void smp_send_stop(void) | ||
| 77 | { | ||
| 78 | smp_ops.smp_send_stop(); | ||
| 79 | } | ||
| 80 | |||
| 81 | static inline void smp_prepare_boot_cpu(void) | ||
| 82 | { | ||
| 83 | smp_ops.smp_prepare_boot_cpu(); | ||
| 84 | } | ||
| 85 | |||
| 86 | static inline void smp_prepare_cpus(unsigned int max_cpus) | ||
| 87 | { | ||
| 88 | smp_ops.smp_prepare_cpus(max_cpus); | ||
| 89 | } | ||
| 90 | |||
| 91 | static inline void smp_cpus_done(unsigned int max_cpus) | ||
| 92 | { | ||
| 93 | smp_ops.smp_cpus_done(max_cpus); | ||
| 94 | } | ||
| 95 | |||
| 96 | static inline int __cpu_up(unsigned int cpu) | ||
| 97 | { | ||
| 98 | return smp_ops.cpu_up(cpu); | ||
| 99 | } | ||
| 100 | |||
| 101 | static inline void smp_send_reschedule(int cpu) | ||
| 102 | { | ||
| 103 | smp_ops.smp_send_reschedule(cpu); | ||
| 104 | } | ||
| 105 | |||
| 106 | static inline int smp_call_function_mask(cpumask_t mask, | ||
| 107 | void (*func) (void *info), void *info, | ||
| 108 | int wait) | ||
| 109 | { | ||
| 110 | return smp_ops.smp_call_function_mask(mask, func, info, wait); | ||
| 111 | } | ||
| 112 | |||
| 113 | void native_smp_prepare_boot_cpu(void); | ||
| 114 | void native_smp_prepare_cpus(unsigned int max_cpus); | ||
| 115 | void native_smp_cpus_done(unsigned int max_cpus); | ||
| 116 | int native_cpu_up(unsigned int cpunum); | ||
| 117 | |||
| 118 | extern int __cpu_disable(void); | ||
| 119 | extern void __cpu_die(unsigned int cpu); | ||
| 120 | |||
| 121 | extern void prefill_possible_map(void); | ||
| 122 | |||
| 123 | void smp_store_cpu_info(int id); | ||
| 124 | #define cpu_physical_id(cpu) per_cpu(x86_cpu_to_apicid, cpu) | ||
| 125 | |||
| 126 | /* We don't mark CPUs online until __cpu_up(), so we need another measure */ | ||
| 127 | static inline int num_booting_cpus(void) | ||
| 128 | { | ||
| 129 | return cpus_weight(cpu_callout_map); | ||
| 130 | } | ||
| 131 | #endif /* CONFIG_SMP */ | ||
| 132 | |||
| 133 | extern unsigned disabled_cpus __cpuinitdata; | ||
| 134 | |||
| 135 | #ifdef CONFIG_X86_32_SMP | ||
| 136 | /* | ||
| 137 | * This function is needed by all SMP systems. It must _always_ be valid | ||
| 138 | * from the initial startup. We map APIC_BASE very early in page_setup(), | ||
| 139 | * so this is correct in the x86 case. | ||
| 140 | */ | ||
| 141 | DECLARE_PER_CPU(int, cpu_number); | ||
| 142 | #define raw_smp_processor_id() (x86_read_percpu(cpu_number)) | ||
| 143 | extern int safe_smp_processor_id(void); | ||
| 144 | |||
| 145 | #elif defined(CONFIG_X86_64_SMP) | ||
| 146 | #define raw_smp_processor_id() read_pda(cpunumber) | ||
| 147 | |||
| 148 | #define stack_smp_processor_id() \ | ||
| 149 | ({ \ | ||
| 150 | struct thread_info *ti; \ | ||
| 151 | __asm__("andq %%rsp,%0; ":"=r" (ti) : "0" (CURRENT_MASK)); \ | ||
| 152 | ti->cpu; \ | ||
| 153 | }) | ||
| 154 | #define safe_smp_processor_id() smp_processor_id() | ||
| 155 | |||
| 156 | #else /* !CONFIG_X86_32_SMP && !CONFIG_X86_64_SMP */ | ||
| 157 | #define cpu_physical_id(cpu) boot_cpu_physical_apicid | ||
| 158 | #define safe_smp_processor_id() 0 | ||
| 159 | #define stack_smp_processor_id() 0 | ||
| 160 | #endif | ||
| 161 | |||
| 162 | #ifdef CONFIG_X86_LOCAL_APIC | ||
| 163 | |||
| 164 | static inline int logical_smp_processor_id(void) | ||
| 165 | { | ||
| 166 | /* we don't want to mark this access volatile - bad code generation */ | ||
| 167 | return GET_APIC_LOGICAL_ID(*(u32 *)(APIC_BASE + APIC_LDR)); | ||
| 168 | } | ||
| 169 | |||
| 170 | #ifndef CONFIG_X86_64 | ||
| 171 | static inline unsigned int read_apic_id(void) | ||
| 172 | { | ||
| 173 | return *(u32 *)(APIC_BASE + APIC_ID); | ||
| 174 | } | ||
| 175 | #else | ||
| 176 | extern unsigned int read_apic_id(void); | ||
| 177 | #endif | ||
| 178 | |||
| 179 | |||
| 180 | # ifdef APIC_DEFINITION | ||
| 181 | extern int hard_smp_processor_id(void); | ||
| 182 | # else | ||
| 183 | # include <mach_apicdef.h> | ||
| 184 | static inline int hard_smp_processor_id(void) | ||
| 185 | { | ||
| 186 | /* we don't want to mark this access volatile - bad code generation */ | ||
| 187 | return GET_APIC_ID(read_apic_id()); | ||
| 188 | } | ||
| 189 | # endif /* APIC_DEFINITION */ | ||
| 190 | |||
| 191 | #else /* CONFIG_X86_LOCAL_APIC */ | ||
| 192 | |||
| 193 | # ifndef CONFIG_SMP | ||
| 194 | # define hard_smp_processor_id() 0 | ||
| 195 | # endif | ||
| 196 | |||
| 197 | #endif /* CONFIG_X86_LOCAL_APIC */ | ||
| 198 | |||
| 199 | #ifdef CONFIG_HOTPLUG_CPU | ||
| 200 | extern void cpu_exit_clear(void); | ||
| 201 | extern void cpu_uninit(void); | ||
| 202 | extern void remove_siblinginfo(int cpu); | ||
| 203 | #endif | ||
| 204 | |||
| 205 | extern void smp_alloc_memory(void); | ||
| 206 | extern void lock_ipi_call_lock(void); | ||
| 207 | extern void unlock_ipi_call_lock(void); | ||
| 208 | #endif /* __ASSEMBLY__ */ | ||
| 5 | #endif | 209 | #endif |
diff --git a/include/asm-x86/smp_32.h b/include/asm-x86/smp_32.h deleted file mode 100644 index 56152e312287..000000000000 --- a/include/asm-x86/smp_32.h +++ /dev/null | |||
| @@ -1,165 +0,0 @@ | |||
| 1 | #ifndef __ASM_SMP_H | ||
| 2 | #define __ASM_SMP_H | ||
| 3 | |||
| 4 | #ifndef __ASSEMBLY__ | ||
| 5 | #include <linux/cpumask.h> | ||
| 6 | #include <linux/init.h> | ||
| 7 | |||
| 8 | /* | ||
| 9 | * We need the APIC definitions automatically as part of 'smp.h' | ||
| 10 | */ | ||
| 11 | #ifdef CONFIG_X86_LOCAL_APIC | ||
| 12 | # include <asm/mpspec.h> | ||
| 13 | # include <asm/apic.h> | ||
| 14 | # ifdef CONFIG_X86_IO_APIC | ||
| 15 | # include <asm/io_apic.h> | ||
| 16 | # endif | ||
| 17 | #endif | ||
| 18 | |||
| 19 | extern cpumask_t cpu_callout_map; | ||
| 20 | extern cpumask_t cpu_callin_map; | ||
| 21 | |||
| 22 | extern int smp_num_siblings; | ||
| 23 | extern unsigned int num_processors; | ||
| 24 | |||
| 25 | extern void smp_alloc_memory(void); | ||
| 26 | extern void lock_ipi_call_lock(void); | ||
| 27 | extern void unlock_ipi_call_lock(void); | ||
| 28 | |||
| 29 | extern void (*mtrr_hook) (void); | ||
| 30 | extern void zap_low_mappings (void); | ||
| 31 | |||
| 32 | extern u8 __initdata x86_cpu_to_apicid_init[]; | ||
| 33 | extern void *x86_cpu_to_apicid_early_ptr; | ||
| 34 | |||
| 35 | DECLARE_PER_CPU(cpumask_t, cpu_sibling_map); | ||
| 36 | DECLARE_PER_CPU(cpumask_t, cpu_core_map); | ||
| 37 | DECLARE_PER_CPU(u8, cpu_llc_id); | ||
| 38 | DECLARE_PER_CPU(u8, x86_cpu_to_apicid); | ||
| 39 | |||
| 40 | #ifdef CONFIG_HOTPLUG_CPU | ||
| 41 | extern void cpu_exit_clear(void); | ||
| 42 | extern void cpu_uninit(void); | ||
| 43 | extern void remove_siblinginfo(int cpu); | ||
| 44 | #endif | ||
| 45 | |||
| 46 | /* Globals due to paravirt */ | ||
| 47 | extern void set_cpu_sibling_map(int cpu); | ||
| 48 | |||
| 49 | struct smp_ops | ||
| 50 | { | ||
| 51 | void (*smp_prepare_boot_cpu)(void); | ||
| 52 | void (*smp_prepare_cpus)(unsigned max_cpus); | ||
| 53 | int (*cpu_up)(unsigned cpu); | ||
| 54 | void (*smp_cpus_done)(unsigned max_cpus); | ||
| 55 | |||
| 56 | void (*smp_send_stop)(void); | ||
| 57 | void (*smp_send_reschedule)(int cpu); | ||
| 58 | int (*smp_call_function_mask)(cpumask_t mask, | ||
| 59 | void (*func)(void *info), void *info, | ||
| 60 | int wait); | ||
| 61 | }; | ||
| 62 | |||
| 63 | #ifdef CONFIG_SMP | ||
| 64 | extern struct smp_ops smp_ops; | ||
| 65 | |||
| 66 | static inline void smp_prepare_boot_cpu(void) | ||
| 67 | { | ||
| 68 | smp_ops.smp_prepare_boot_cpu(); | ||
| 69 | } | ||
| 70 | static inline void smp_prepare_cpus(unsigned int max_cpus) | ||
| 71 | { | ||
| 72 | smp_ops.smp_prepare_cpus(max_cpus); | ||
| 73 | } | ||
| 74 | static inline int __cpu_up(unsigned int cpu) | ||
| 75 | { | ||
| 76 | return smp_ops.cpu_up(cpu); | ||
| 77 | } | ||
| 78 | static inline void smp_cpus_done(unsigned int max_cpus) | ||
| 79 | { | ||
| 80 | smp_ops.smp_cpus_done(max_cpus); | ||
| 81 | } | ||
| 82 | |||
| 83 | static inline void smp_send_stop(void) | ||
| 84 | { | ||
| 85 | smp_ops.smp_send_stop(); | ||
| 86 | } | ||
| 87 | static inline void smp_send_reschedule(int cpu) | ||
| 88 | { | ||
| 89 | smp_ops.smp_send_reschedule(cpu); | ||
| 90 | } | ||
| 91 | static inline int smp_call_function_mask(cpumask_t mask, | ||
| 92 | void (*func) (void *info), void *info, | ||
| 93 | int wait) | ||
| 94 | { | ||
| 95 | return smp_ops.smp_call_function_mask(mask, func, info, wait); | ||
| 96 | } | ||
| 97 | |||
| 98 | void native_smp_prepare_boot_cpu(void); | ||
| 99 | void native_smp_prepare_cpus(unsigned int max_cpus); | ||
| 100 | int native_cpu_up(unsigned int cpunum); | ||
| 101 | void native_smp_cpus_done(unsigned int max_cpus); | ||
| 102 | |||
| 103 | #ifndef CONFIG_PARAVIRT | ||
| 104 | #define startup_ipi_hook(phys_apicid, start_eip, start_esp) do { } while (0) | ||
| 105 | #endif | ||
| 106 | |||
| 107 | extern int __cpu_disable(void); | ||
| 108 | extern void __cpu_die(unsigned int cpu); | ||
| 109 | |||
| 110 | /* | ||
| 111 | * This function is needed by all SMP systems. It must _always_ be valid | ||
| 112 | * from the initial startup. We map APIC_BASE very early in page_setup(), | ||
| 113 | * so this is correct in the x86 case. | ||
| 114 | */ | ||
| 115 | DECLARE_PER_CPU(int, cpu_number); | ||
| 116 | #define raw_smp_processor_id() (x86_read_percpu(cpu_number)) | ||
| 117 | |||
| 118 | #define cpu_physical_id(cpu) per_cpu(x86_cpu_to_apicid, cpu) | ||
| 119 | |||
| 120 | extern int safe_smp_processor_id(void); | ||
| 121 | |||
| 122 | void __cpuinit smp_store_cpu_info(int id); | ||
| 123 | |||
| 124 | /* We don't mark CPUs online until __cpu_up(), so we need another measure */ | ||
| 125 | static inline int num_booting_cpus(void) | ||
| 126 | { | ||
| 127 | return cpus_weight(cpu_callout_map); | ||
| 128 | } | ||
| 129 | |||
| 130 | #else /* CONFIG_SMP */ | ||
| 131 | |||
| 132 | #define safe_smp_processor_id() 0 | ||
| 133 | #define cpu_physical_id(cpu) boot_cpu_physical_apicid | ||
| 134 | |||
| 135 | #endif /* !CONFIG_SMP */ | ||
| 136 | |||
| 137 | #ifdef CONFIG_X86_LOCAL_APIC | ||
| 138 | |||
| 139 | static __inline int logical_smp_processor_id(void) | ||
| 140 | { | ||
| 141 | /* we don't want to mark this access volatile - bad code generation */ | ||
| 142 | return GET_APIC_LOGICAL_ID(*(u32 *)(APIC_BASE + APIC_LDR)); | ||
| 143 | } | ||
| 144 | |||
| 145 | # ifdef APIC_DEFINITION | ||
| 146 | extern int hard_smp_processor_id(void); | ||
| 147 | # else | ||
| 148 | # include <mach_apicdef.h> | ||
| 149 | static inline int hard_smp_processor_id(void) | ||
| 150 | { | ||
| 151 | /* we don't want to mark this access volatile - bad code generation */ | ||
| 152 | return GET_APIC_ID(*(u32 *)(APIC_BASE + APIC_ID)); | ||
| 153 | } | ||
| 154 | # endif /* APIC_DEFINITION */ | ||
| 155 | |||
| 156 | #else /* CONFIG_X86_LOCAL_APIC */ | ||
| 157 | |||
| 158 | # ifndef CONFIG_SMP | ||
| 159 | # define hard_smp_processor_id() 0 | ||
| 160 | # endif | ||
| 161 | |||
| 162 | #endif /* CONFIG_X86_LOCAL_APIC */ | ||
| 163 | |||
| 164 | #endif /* !ASSEMBLY */ | ||
| 165 | #endif | ||
diff --git a/include/asm-x86/smp_64.h b/include/asm-x86/smp_64.h deleted file mode 100644 index e0a75519ad21..000000000000 --- a/include/asm-x86/smp_64.h +++ /dev/null | |||
| @@ -1,101 +0,0 @@ | |||
| 1 | #ifndef __ASM_SMP_H | ||
| 2 | #define __ASM_SMP_H | ||
| 3 | |||
| 4 | #include <linux/cpumask.h> | ||
| 5 | #include <linux/init.h> | ||
| 6 | |||
| 7 | /* | ||
| 8 | * We need the APIC definitions automatically as part of 'smp.h' | ||
| 9 | */ | ||
| 10 | #include <asm/apic.h> | ||
| 11 | #include <asm/io_apic.h> | ||
| 12 | #include <asm/mpspec.h> | ||
| 13 | #include <asm/pda.h> | ||
| 14 | #include <asm/thread_info.h> | ||
| 15 | |||
| 16 | extern cpumask_t cpu_callout_map; | ||
| 17 | extern cpumask_t cpu_initialized; | ||
| 18 | |||
| 19 | extern int smp_num_siblings; | ||
| 20 | extern unsigned int num_processors; | ||
| 21 | |||
| 22 | extern void smp_alloc_memory(void); | ||
| 23 | extern void lock_ipi_call_lock(void); | ||
| 24 | extern void unlock_ipi_call_lock(void); | ||
| 25 | |||
| 26 | extern int smp_call_function_mask(cpumask_t mask, void (*func)(void *), | ||
| 27 | void *info, int wait); | ||
| 28 | |||
| 29 | extern u16 __initdata x86_cpu_to_apicid_init[]; | ||
| 30 | extern u16 __initdata x86_bios_cpu_apicid_init[]; | ||
| 31 | extern void *x86_cpu_to_apicid_early_ptr; | ||
| 32 | extern void *x86_bios_cpu_apicid_early_ptr; | ||
| 33 | |||
| 34 | DECLARE_PER_CPU(cpumask_t, cpu_sibling_map); | ||
| 35 | DECLARE_PER_CPU(cpumask_t, cpu_core_map); | ||
| 36 | DECLARE_PER_CPU(u16, cpu_llc_id); | ||
| 37 | DECLARE_PER_CPU(u16, x86_cpu_to_apicid); | ||
| 38 | DECLARE_PER_CPU(u16, x86_bios_cpu_apicid); | ||
| 39 | |||
| 40 | static inline int cpu_present_to_apicid(int mps_cpu) | ||
| 41 | { | ||
| 42 | if (cpu_present(mps_cpu)) | ||
| 43 | return (int)per_cpu(x86_bios_cpu_apicid, mps_cpu); | ||
| 44 | else | ||
| 45 | return BAD_APICID; | ||
| 46 | } | ||
| 47 | |||
| 48 | #ifdef CONFIG_SMP | ||
| 49 | |||
| 50 | #define SMP_TRAMPOLINE_BASE 0x6000 | ||
| 51 | |||
| 52 | extern int __cpu_disable(void); | ||
| 53 | extern void __cpu_die(unsigned int cpu); | ||
| 54 | extern void prefill_possible_map(void); | ||
| 55 | extern unsigned __cpuinitdata disabled_cpus; | ||
| 56 | |||
| 57 | #define raw_smp_processor_id() read_pda(cpunumber) | ||
| 58 | #define cpu_physical_id(cpu) per_cpu(x86_cpu_to_apicid, cpu) | ||
| 59 | |||
| 60 | #define stack_smp_processor_id() \ | ||
| 61 | ({ \ | ||
| 62 | struct thread_info *ti; \ | ||
| 63 | __asm__("andq %%rsp,%0; ":"=r" (ti) : "0" (CURRENT_MASK)); \ | ||
| 64 | ti->cpu; \ | ||
| 65 | }) | ||
| 66 | |||
| 67 | /* | ||
| 68 | * On x86 all CPUs are mapped 1:1 to the APIC space. This simplifies | ||
| 69 | * scheduling and IPI sending and compresses data structures. | ||
| 70 | */ | ||
| 71 | static inline int num_booting_cpus(void) | ||
| 72 | { | ||
| 73 | return cpus_weight(cpu_callout_map); | ||
| 74 | } | ||
| 75 | |||
| 76 | extern void smp_send_reschedule(int cpu); | ||
| 77 | |||
| 78 | #else /* CONFIG_SMP */ | ||
| 79 | |||
| 80 | extern unsigned int boot_cpu_id; | ||
| 81 | #define cpu_physical_id(cpu) boot_cpu_id | ||
| 82 | #define stack_smp_processor_id() 0 | ||
| 83 | |||
| 84 | #endif /* !CONFIG_SMP */ | ||
| 85 | |||
| 86 | #define safe_smp_processor_id() smp_processor_id() | ||
| 87 | |||
| 88 | static __inline int logical_smp_processor_id(void) | ||
| 89 | { | ||
| 90 | /* we don't want to mark this access volatile - bad code generation */ | ||
| 91 | return GET_APIC_LOGICAL_ID(*(u32 *)(APIC_BASE + APIC_LDR)); | ||
| 92 | } | ||
| 93 | |||
| 94 | static inline int hard_smp_processor_id(void) | ||
| 95 | { | ||
| 96 | /* we don't want to mark this access volatile - bad code generation */ | ||
| 97 | return GET_APIC_ID(*(u32 *)(APIC_BASE + APIC_ID)); | ||
| 98 | } | ||
| 99 | |||
| 100 | #endif | ||
| 101 | |||
diff --git a/include/asm-x86/sparsemem.h b/include/asm-x86/sparsemem.h index fa58cd55411a..9bd48b0a534b 100644 --- a/include/asm-x86/sparsemem.h +++ b/include/asm-x86/sparsemem.h | |||
| @@ -16,7 +16,7 @@ | |||
| 16 | 16 | ||
| 17 | #ifdef CONFIG_X86_32 | 17 | #ifdef CONFIG_X86_32 |
| 18 | # ifdef CONFIG_X86_PAE | 18 | # ifdef CONFIG_X86_PAE |
| 19 | # define SECTION_SIZE_BITS 30 | 19 | # define SECTION_SIZE_BITS 29 |
| 20 | # define MAX_PHYSADDR_BITS 36 | 20 | # define MAX_PHYSADDR_BITS 36 |
| 21 | # define MAX_PHYSMEM_BITS 36 | 21 | # define MAX_PHYSMEM_BITS 36 |
| 22 | # else | 22 | # else |
| @@ -26,8 +26,8 @@ | |||
| 26 | # endif | 26 | # endif |
| 27 | #else /* CONFIG_X86_32 */ | 27 | #else /* CONFIG_X86_32 */ |
| 28 | # define SECTION_SIZE_BITS 27 /* matt - 128 is convenient right now */ | 28 | # define SECTION_SIZE_BITS 27 /* matt - 128 is convenient right now */ |
| 29 | # define MAX_PHYSADDR_BITS 40 | 29 | # define MAX_PHYSADDR_BITS 44 |
| 30 | # define MAX_PHYSMEM_BITS 40 | 30 | # define MAX_PHYSMEM_BITS 44 |
| 31 | #endif | 31 | #endif |
| 32 | 32 | ||
| 33 | #endif /* CONFIG_SPARSEMEM */ | 33 | #endif /* CONFIG_SPARSEMEM */ |
diff --git a/include/asm-x86/spinlock.h b/include/asm-x86/spinlock.h index 23804c1890ff..bc6376f1bc5a 100644 --- a/include/asm-x86/spinlock.h +++ b/include/asm-x86/spinlock.h | |||
| @@ -78,11 +78,11 @@ static inline int __raw_spin_is_contended(raw_spinlock_t *lock) | |||
| 78 | return (((tmp >> 8) & 0xff) - (tmp & 0xff)) > 1; | 78 | return (((tmp >> 8) & 0xff) - (tmp & 0xff)) > 1; |
| 79 | } | 79 | } |
| 80 | 80 | ||
| 81 | static inline void __raw_spin_lock(raw_spinlock_t *lock) | 81 | static __always_inline void __raw_spin_lock(raw_spinlock_t *lock) |
| 82 | { | 82 | { |
| 83 | short inc = 0x0100; | 83 | short inc = 0x0100; |
| 84 | 84 | ||
| 85 | __asm__ __volatile__ ( | 85 | asm volatile ( |
| 86 | LOCK_PREFIX "xaddw %w0, %1\n" | 86 | LOCK_PREFIX "xaddw %w0, %1\n" |
| 87 | "1:\t" | 87 | "1:\t" |
| 88 | "cmpb %h0, %b0\n\t" | 88 | "cmpb %h0, %b0\n\t" |
| @@ -92,42 +92,40 @@ static inline void __raw_spin_lock(raw_spinlock_t *lock) | |||
| 92 | /* don't need lfence here, because loads are in-order */ | 92 | /* don't need lfence here, because loads are in-order */ |
| 93 | "jmp 1b\n" | 93 | "jmp 1b\n" |
| 94 | "2:" | 94 | "2:" |
| 95 | :"+Q" (inc), "+m" (lock->slock) | 95 | : "+Q" (inc), "+m" (lock->slock) |
| 96 | : | 96 | : |
| 97 | :"memory", "cc"); | 97 | : "memory", "cc"); |
| 98 | } | 98 | } |
| 99 | 99 | ||
| 100 | #define __raw_spin_lock_flags(lock, flags) __raw_spin_lock(lock) | 100 | #define __raw_spin_lock_flags(lock, flags) __raw_spin_lock(lock) |
| 101 | 101 | ||
| 102 | static inline int __raw_spin_trylock(raw_spinlock_t *lock) | 102 | static __always_inline int __raw_spin_trylock(raw_spinlock_t *lock) |
| 103 | { | 103 | { |
| 104 | int tmp; | 104 | int tmp; |
| 105 | short new; | 105 | short new; |
| 106 | 106 | ||
| 107 | asm volatile( | 107 | asm volatile("movw %2,%w0\n\t" |
| 108 | "movw %2,%w0\n\t" | 108 | "cmpb %h0,%b0\n\t" |
| 109 | "cmpb %h0,%b0\n\t" | 109 | "jne 1f\n\t" |
| 110 | "jne 1f\n\t" | 110 | "movw %w0,%w1\n\t" |
| 111 | "movw %w0,%w1\n\t" | 111 | "incb %h1\n\t" |
| 112 | "incb %h1\n\t" | 112 | "lock ; cmpxchgw %w1,%2\n\t" |
| 113 | "lock ; cmpxchgw %w1,%2\n\t" | 113 | "1:" |
| 114 | "1:" | 114 | "sete %b1\n\t" |
| 115 | "sete %b1\n\t" | 115 | "movzbl %b1,%0\n\t" |
| 116 | "movzbl %b1,%0\n\t" | 116 | : "=&a" (tmp), "=Q" (new), "+m" (lock->slock) |
| 117 | :"=&a" (tmp), "=Q" (new), "+m" (lock->slock) | 117 | : |
| 118 | : | 118 | : "memory", "cc"); |
| 119 | : "memory", "cc"); | ||
| 120 | 119 | ||
| 121 | return tmp; | 120 | return tmp; |
| 122 | } | 121 | } |
| 123 | 122 | ||
| 124 | static inline void __raw_spin_unlock(raw_spinlock_t *lock) | 123 | static __always_inline void __raw_spin_unlock(raw_spinlock_t *lock) |
| 125 | { | 124 | { |
| 126 | __asm__ __volatile__( | 125 | asm volatile(UNLOCK_LOCK_PREFIX "incb %0" |
| 127 | UNLOCK_LOCK_PREFIX "incb %0" | 126 | : "+m" (lock->slock) |
| 128 | :"+m" (lock->slock) | 127 | : |
| 129 | : | 128 | : "memory", "cc"); |
| 130 | :"memory", "cc"); | ||
| 131 | } | 129 | } |
| 132 | #else | 130 | #else |
| 133 | static inline int __raw_spin_is_locked(raw_spinlock_t *lock) | 131 | static inline int __raw_spin_is_locked(raw_spinlock_t *lock) |
| @@ -144,60 +142,57 @@ static inline int __raw_spin_is_contended(raw_spinlock_t *lock) | |||
| 144 | return (((tmp >> 16) & 0xffff) - (tmp & 0xffff)) > 1; | 142 | return (((tmp >> 16) & 0xffff) - (tmp & 0xffff)) > 1; |
| 145 | } | 143 | } |
| 146 | 144 | ||
| 147 | static inline void __raw_spin_lock(raw_spinlock_t *lock) | 145 | static __always_inline void __raw_spin_lock(raw_spinlock_t *lock) |
| 148 | { | 146 | { |
| 149 | int inc = 0x00010000; | 147 | int inc = 0x00010000; |
| 150 | int tmp; | 148 | int tmp; |
| 151 | 149 | ||
| 152 | __asm__ __volatile__ ( | 150 | asm volatile("lock ; xaddl %0, %1\n" |
| 153 | "lock ; xaddl %0, %1\n" | 151 | "movzwl %w0, %2\n\t" |
| 154 | "movzwl %w0, %2\n\t" | 152 | "shrl $16, %0\n\t" |
| 155 | "shrl $16, %0\n\t" | 153 | "1:\t" |
| 156 | "1:\t" | 154 | "cmpl %0, %2\n\t" |
| 157 | "cmpl %0, %2\n\t" | 155 | "je 2f\n\t" |
| 158 | "je 2f\n\t" | 156 | "rep ; nop\n\t" |
| 159 | "rep ; nop\n\t" | 157 | "movzwl %1, %2\n\t" |
| 160 | "movzwl %1, %2\n\t" | 158 | /* don't need lfence here, because loads are in-order */ |
| 161 | /* don't need lfence here, because loads are in-order */ | 159 | "jmp 1b\n" |
| 162 | "jmp 1b\n" | 160 | "2:" |
| 163 | "2:" | 161 | : "+Q" (inc), "+m" (lock->slock), "=r" (tmp) |
| 164 | :"+Q" (inc), "+m" (lock->slock), "=r" (tmp) | 162 | : |
| 165 | : | 163 | : "memory", "cc"); |
| 166 | :"memory", "cc"); | ||
| 167 | } | 164 | } |
| 168 | 165 | ||
| 169 | #define __raw_spin_lock_flags(lock, flags) __raw_spin_lock(lock) | 166 | #define __raw_spin_lock_flags(lock, flags) __raw_spin_lock(lock) |
| 170 | 167 | ||
| 171 | static inline int __raw_spin_trylock(raw_spinlock_t *lock) | 168 | static __always_inline int __raw_spin_trylock(raw_spinlock_t *lock) |
| 172 | { | 169 | { |
| 173 | int tmp; | 170 | int tmp; |
| 174 | int new; | 171 | int new; |
| 175 | 172 | ||
| 176 | asm volatile( | 173 | asm volatile("movl %2,%0\n\t" |
| 177 | "movl %2,%0\n\t" | 174 | "movl %0,%1\n\t" |
| 178 | "movl %0,%1\n\t" | 175 | "roll $16, %0\n\t" |
| 179 | "roll $16, %0\n\t" | 176 | "cmpl %0,%1\n\t" |
| 180 | "cmpl %0,%1\n\t" | 177 | "jne 1f\n\t" |
| 181 | "jne 1f\n\t" | 178 | "addl $0x00010000, %1\n\t" |
| 182 | "addl $0x00010000, %1\n\t" | 179 | "lock ; cmpxchgl %1,%2\n\t" |
| 183 | "lock ; cmpxchgl %1,%2\n\t" | 180 | "1:" |
| 184 | "1:" | 181 | "sete %b1\n\t" |
| 185 | "sete %b1\n\t" | 182 | "movzbl %b1,%0\n\t" |
| 186 | "movzbl %b1,%0\n\t" | 183 | : "=&a" (tmp), "=r" (new), "+m" (lock->slock) |
| 187 | :"=&a" (tmp), "=r" (new), "+m" (lock->slock) | 184 | : |
| 188 | : | 185 | : "memory", "cc"); |
| 189 | : "memory", "cc"); | ||
| 190 | 186 | ||
| 191 | return tmp; | 187 | return tmp; |
| 192 | } | 188 | } |
| 193 | 189 | ||
| 194 | static inline void __raw_spin_unlock(raw_spinlock_t *lock) | 190 | static __always_inline void __raw_spin_unlock(raw_spinlock_t *lock) |
| 195 | { | 191 | { |
| 196 | __asm__ __volatile__( | 192 | asm volatile(UNLOCK_LOCK_PREFIX "incw %0" |
| 197 | UNLOCK_LOCK_PREFIX "incw %0" | 193 | : "+m" (lock->slock) |
| 198 | :"+m" (lock->slock) | 194 | : |
| 199 | : | 195 | : "memory", "cc"); |
| 200 | :"memory", "cc"); | ||
| 201 | } | 196 | } |
| 202 | #endif | 197 | #endif |
| 203 | 198 | ||
diff --git a/include/asm-x86/srat.h b/include/asm-x86/srat.h index 165ab4bdc02b..f4bba131d068 100644 --- a/include/asm-x86/srat.h +++ b/include/asm-x86/srat.h | |||
| @@ -1,5 +1,5 @@ | |||
| 1 | /* | 1 | /* |
| 2 | * Some of the code in this file has been gleaned from the 64 bit | 2 | * Some of the code in this file has been gleaned from the 64 bit |
| 3 | * discontigmem support code base. | 3 | * discontigmem support code base. |
| 4 | * | 4 | * |
| 5 | * Copyright (C) 2002, IBM Corp. | 5 | * Copyright (C) 2002, IBM Corp. |
diff --git a/include/asm-x86/string_32.h b/include/asm-x86/string_32.h index c5d13a86dea7..b49369ad9a61 100644 --- a/include/asm-x86/string_32.h +++ b/include/asm-x86/string_32.h | |||
| @@ -3,7 +3,7 @@ | |||
| 3 | 3 | ||
| 4 | #ifdef __KERNEL__ | 4 | #ifdef __KERNEL__ |
| 5 | 5 | ||
| 6 | /* Let gcc decide wether to inline or use the out of line functions */ | 6 | /* Let gcc decide whether to inline or use the out of line functions */ |
| 7 | 7 | ||
| 8 | #define __HAVE_ARCH_STRCPY | 8 | #define __HAVE_ARCH_STRCPY |
| 9 | extern char *strcpy(char *dest, const char *src); | 9 | extern char *strcpy(char *dest, const char *src); |
diff --git a/include/asm-x86/string_64.h b/include/asm-x86/string_64.h index e583da7918fb..52b5ab383395 100644 --- a/include/asm-x86/string_64.h +++ b/include/asm-x86/string_64.h | |||
| @@ -3,26 +3,24 @@ | |||
| 3 | 3 | ||
| 4 | #ifdef __KERNEL__ | 4 | #ifdef __KERNEL__ |
| 5 | 5 | ||
| 6 | /* Written 2002 by Andi Kleen */ | 6 | /* Written 2002 by Andi Kleen */ |
| 7 | 7 | ||
| 8 | /* Only used for special circumstances. Stolen from i386/string.h */ | 8 | /* Only used for special circumstances. Stolen from i386/string.h */ |
| 9 | static __always_inline void * | 9 | static __always_inline void *__inline_memcpy(void *to, const void *from, size_t n) |
| 10 | __inline_memcpy(void * to, const void * from, size_t n) | ||
| 11 | { | 10 | { |
| 12 | unsigned long d0, d1, d2; | 11 | unsigned long d0, d1, d2; |
| 13 | __asm__ __volatile__( | 12 | asm volatile("rep ; movsl\n\t" |
| 14 | "rep ; movsl\n\t" | 13 | "testb $2,%b4\n\t" |
| 15 | "testb $2,%b4\n\t" | 14 | "je 1f\n\t" |
| 16 | "je 1f\n\t" | 15 | "movsw\n" |
| 17 | "movsw\n" | 16 | "1:\ttestb $1,%b4\n\t" |
| 18 | "1:\ttestb $1,%b4\n\t" | 17 | "je 2f\n\t" |
| 19 | "je 2f\n\t" | 18 | "movsb\n" |
| 20 | "movsb\n" | 19 | "2:" |
| 21 | "2:" | 20 | : "=&c" (d0), "=&D" (d1), "=&S" (d2) |
| 22 | : "=&c" (d0), "=&D" (d1), "=&S" (d2) | 21 | : "0" (n / 4), "q" (n), "1" ((long)to), "2" ((long)from) |
| 23 | :"0" (n/4), "q" (n),"1" ((long) to),"2" ((long) from) | 22 | : "memory"); |
| 24 | : "memory"); | 23 | return to; |
| 25 | return (to); | ||
| 26 | } | 24 | } |
| 27 | 25 | ||
| 28 | /* Even with __builtin_ the compiler may decide to use the out of line | 26 | /* Even with __builtin_ the compiler may decide to use the out of line |
| @@ -32,28 +30,30 @@ return (to); | |||
| 32 | #if (__GNUC__ == 4 && __GNUC_MINOR__ >= 3) || __GNUC__ > 4 | 30 | #if (__GNUC__ == 4 && __GNUC_MINOR__ >= 3) || __GNUC__ > 4 |
| 33 | extern void *memcpy(void *to, const void *from, size_t len); | 31 | extern void *memcpy(void *to, const void *from, size_t len); |
| 34 | #else | 32 | #else |
| 35 | extern void *__memcpy(void *to, const void *from, size_t len); | 33 | extern void *__memcpy(void *to, const void *from, size_t len); |
| 36 | #define memcpy(dst,src,len) \ | 34 | #define memcpy(dst, src, len) \ |
| 37 | ({ size_t __len = (len); \ | 35 | ({ \ |
| 38 | void *__ret; \ | 36 | size_t __len = (len); \ |
| 39 | if (__builtin_constant_p(len) && __len >= 64) \ | 37 | void *__ret; \ |
| 40 | __ret = __memcpy((dst),(src),__len); \ | 38 | if (__builtin_constant_p(len) && __len >= 64) \ |
| 41 | else \ | 39 | __ret = __memcpy((dst), (src), __len); \ |
| 42 | __ret = __builtin_memcpy((dst),(src),__len); \ | 40 | else \ |
| 43 | __ret; }) | 41 | __ret = __builtin_memcpy((dst), (src), __len); \ |
| 42 | __ret; \ | ||
| 43 | }) | ||
| 44 | #endif | 44 | #endif |
| 45 | 45 | ||
| 46 | #define __HAVE_ARCH_MEMSET | 46 | #define __HAVE_ARCH_MEMSET |
| 47 | void *memset(void *s, int c, size_t n); | 47 | void *memset(void *s, int c, size_t n); |
| 48 | 48 | ||
| 49 | #define __HAVE_ARCH_MEMMOVE | 49 | #define __HAVE_ARCH_MEMMOVE |
| 50 | void * memmove(void * dest,const void *src,size_t count); | 50 | void *memmove(void *dest, const void *src, size_t count); |
| 51 | 51 | ||
| 52 | int memcmp(const void * cs,const void * ct,size_t count); | 52 | int memcmp(const void *cs, const void *ct, size_t count); |
| 53 | size_t strlen(const char * s); | 53 | size_t strlen(const char *s); |
| 54 | char *strcpy(char * dest,const char *src); | 54 | char *strcpy(char *dest, const char *src); |
| 55 | char *strcat(char * dest, const char * src); | 55 | char *strcat(char *dest, const char *src); |
| 56 | int strcmp(const char * cs,const char * ct); | 56 | int strcmp(const char *cs, const char *ct); |
| 57 | 57 | ||
| 58 | #endif /* __KERNEL__ */ | 58 | #endif /* __KERNEL__ */ |
| 59 | 59 | ||
diff --git a/include/asm-x86/suspend_32.h b/include/asm-x86/suspend_32.h index 1bbda3ad7796..24e1c080aa8a 100644 --- a/include/asm-x86/suspend_32.h +++ b/include/asm-x86/suspend_32.h | |||
| @@ -10,7 +10,7 @@ static inline int arch_prepare_suspend(void) { return 0; } | |||
| 10 | 10 | ||
| 11 | /* image of the saved processor state */ | 11 | /* image of the saved processor state */ |
| 12 | struct saved_context { | 12 | struct saved_context { |
| 13 | u16 es, fs, gs, ss; | 13 | u16 es, fs, gs, ss; |
| 14 | unsigned long cr0, cr2, cr3, cr4; | 14 | unsigned long cr0, cr2, cr3, cr4; |
| 15 | struct desc_ptr gdt; | 15 | struct desc_ptr gdt; |
| 16 | struct desc_ptr idt; | 16 | struct desc_ptr idt; |
| @@ -32,11 +32,11 @@ extern unsigned long saved_edi; | |||
| 32 | static inline void acpi_save_register_state(unsigned long return_point) | 32 | static inline void acpi_save_register_state(unsigned long return_point) |
| 33 | { | 33 | { |
| 34 | saved_eip = return_point; | 34 | saved_eip = return_point; |
| 35 | asm volatile ("movl %%esp,%0" : "=m" (saved_esp)); | 35 | asm volatile("movl %%esp,%0" : "=m" (saved_esp)); |
| 36 | asm volatile ("movl %%ebp,%0" : "=m" (saved_ebp)); | 36 | asm volatile("movl %%ebp,%0" : "=m" (saved_ebp)); |
| 37 | asm volatile ("movl %%ebx,%0" : "=m" (saved_ebx)); | 37 | asm volatile("movl %%ebx,%0" : "=m" (saved_ebx)); |
| 38 | asm volatile ("movl %%edi,%0" : "=m" (saved_edi)); | 38 | asm volatile("movl %%edi,%0" : "=m" (saved_edi)); |
| 39 | asm volatile ("movl %%esi,%0" : "=m" (saved_esi)); | 39 | asm volatile("movl %%esi,%0" : "=m" (saved_esi)); |
| 40 | } | 40 | } |
| 41 | 41 | ||
| 42 | #define acpi_restore_register_state() do {} while (0) | 42 | #define acpi_restore_register_state() do {} while (0) |
diff --git a/include/asm-x86/suspend_64.h b/include/asm-x86/suspend_64.h index 2eb92cb81a0d..dc3262b43072 100644 --- a/include/asm-x86/suspend_64.h +++ b/include/asm-x86/suspend_64.h | |||
| @@ -9,8 +9,7 @@ | |||
| 9 | #include <asm/desc.h> | 9 | #include <asm/desc.h> |
| 10 | #include <asm/i387.h> | 10 | #include <asm/i387.h> |
| 11 | 11 | ||
| 12 | static inline int | 12 | static inline int arch_prepare_suspend(void) |
| 13 | arch_prepare_suspend(void) | ||
| 14 | { | 13 | { |
| 15 | return 0; | 14 | return 0; |
| 16 | } | 15 | } |
| @@ -25,7 +24,7 @@ arch_prepare_suspend(void) | |||
| 25 | */ | 24 | */ |
| 26 | struct saved_context { | 25 | struct saved_context { |
| 27 | struct pt_regs regs; | 26 | struct pt_regs regs; |
| 28 | u16 ds, es, fs, gs, ss; | 27 | u16 ds, es, fs, gs, ss; |
| 29 | unsigned long gs_base, gs_kernel_base, fs_base; | 28 | unsigned long gs_base, gs_kernel_base, fs_base; |
| 30 | unsigned long cr0, cr2, cr3, cr4, cr8; | 29 | unsigned long cr0, cr2, cr3, cr4, cr8; |
| 31 | unsigned long efer; | 30 | unsigned long efer; |
diff --git a/include/asm-x86/swiotlb.h b/include/asm-x86/swiotlb.h index f9c589539a82..f5d9e74b1e4a 100644 --- a/include/asm-x86/swiotlb.h +++ b/include/asm-x86/swiotlb.h | |||
| @@ -8,15 +8,15 @@ | |||
| 8 | extern dma_addr_t swiotlb_map_single(struct device *hwdev, void *ptr, | 8 | extern dma_addr_t swiotlb_map_single(struct device *hwdev, void *ptr, |
| 9 | size_t size, int dir); | 9 | size_t size, int dir); |
| 10 | extern void *swiotlb_alloc_coherent(struct device *hwdev, size_t size, | 10 | extern void *swiotlb_alloc_coherent(struct device *hwdev, size_t size, |
| 11 | dma_addr_t *dma_handle, gfp_t flags); | 11 | dma_addr_t *dma_handle, gfp_t flags); |
| 12 | extern void swiotlb_unmap_single(struct device *hwdev, dma_addr_t dev_addr, | 12 | extern void swiotlb_unmap_single(struct device *hwdev, dma_addr_t dev_addr, |
| 13 | size_t size, int dir); | 13 | size_t size, int dir); |
| 14 | extern void swiotlb_sync_single_for_cpu(struct device *hwdev, | 14 | extern void swiotlb_sync_single_for_cpu(struct device *hwdev, |
| 15 | dma_addr_t dev_addr, | 15 | dma_addr_t dev_addr, |
| 16 | size_t size, int dir); | 16 | size_t size, int dir); |
| 17 | extern void swiotlb_sync_single_for_device(struct device *hwdev, | 17 | extern void swiotlb_sync_single_for_device(struct device *hwdev, |
| 18 | dma_addr_t dev_addr, | 18 | dma_addr_t dev_addr, |
| 19 | size_t size, int dir); | 19 | size_t size, int dir); |
| 20 | extern void swiotlb_sync_single_range_for_cpu(struct device *hwdev, | 20 | extern void swiotlb_sync_single_range_for_cpu(struct device *hwdev, |
| 21 | dma_addr_t dev_addr, | 21 | dma_addr_t dev_addr, |
| 22 | unsigned long offset, | 22 | unsigned long offset, |
| @@ -26,18 +26,18 @@ extern void swiotlb_sync_single_range_for_device(struct device *hwdev, | |||
| 26 | unsigned long offset, | 26 | unsigned long offset, |
| 27 | size_t size, int dir); | 27 | size_t size, int dir); |
| 28 | extern void swiotlb_sync_sg_for_cpu(struct device *hwdev, | 28 | extern void swiotlb_sync_sg_for_cpu(struct device *hwdev, |
| 29 | struct scatterlist *sg, int nelems, | 29 | struct scatterlist *sg, int nelems, |
| 30 | int dir); | 30 | int dir); |
| 31 | extern void swiotlb_sync_sg_for_device(struct device *hwdev, | 31 | extern void swiotlb_sync_sg_for_device(struct device *hwdev, |
| 32 | struct scatterlist *sg, int nelems, | 32 | struct scatterlist *sg, int nelems, |
| 33 | int dir); | 33 | int dir); |
| 34 | extern int swiotlb_map_sg(struct device *hwdev, struct scatterlist *sg, | 34 | extern int swiotlb_map_sg(struct device *hwdev, struct scatterlist *sg, |
| 35 | int nents, int direction); | 35 | int nents, int direction); |
| 36 | extern void swiotlb_unmap_sg(struct device *hwdev, struct scatterlist *sg, | 36 | extern void swiotlb_unmap_sg(struct device *hwdev, struct scatterlist *sg, |
| 37 | int nents, int direction); | 37 | int nents, int direction); |
| 38 | extern int swiotlb_dma_mapping_error(dma_addr_t dma_addr); | 38 | extern int swiotlb_dma_mapping_error(dma_addr_t dma_addr); |
| 39 | extern void swiotlb_free_coherent (struct device *hwdev, size_t size, | 39 | extern void swiotlb_free_coherent(struct device *hwdev, size_t size, |
| 40 | void *vaddr, dma_addr_t dma_handle); | 40 | void *vaddr, dma_addr_t dma_handle); |
| 41 | extern int swiotlb_dma_supported(struct device *hwdev, u64 mask); | 41 | extern int swiotlb_dma_supported(struct device *hwdev, u64 mask); |
| 42 | extern void swiotlb_init(void); | 42 | extern void swiotlb_init(void); |
| 43 | 43 | ||
diff --git a/include/asm-x86/sync_bitops.h b/include/asm-x86/sync_bitops.h index 6b775c905666..b47a1d0b8a83 100644 --- a/include/asm-x86/sync_bitops.h +++ b/include/asm-x86/sync_bitops.h | |||
| @@ -13,7 +13,7 @@ | |||
| 13 | * bit 0 is the LSB of addr; bit 32 is the LSB of (addr+1). | 13 | * bit 0 is the LSB of addr; bit 32 is the LSB of (addr+1). |
| 14 | */ | 14 | */ |
| 15 | 15 | ||
| 16 | #define ADDR (*(volatile long *) addr) | 16 | #define ADDR (*(volatile long *)addr) |
| 17 | 17 | ||
| 18 | /** | 18 | /** |
| 19 | * sync_set_bit - Atomically set a bit in memory | 19 | * sync_set_bit - Atomically set a bit in memory |
| @@ -26,12 +26,12 @@ | |||
| 26 | * Note that @nr may be almost arbitrarily large; this function is not | 26 | * Note that @nr may be almost arbitrarily large; this function is not |
| 27 | * restricted to acting on a single-word quantity. | 27 | * restricted to acting on a single-word quantity. |
| 28 | */ | 28 | */ |
| 29 | static inline void sync_set_bit(int nr, volatile unsigned long * addr) | 29 | static inline void sync_set_bit(int nr, volatile unsigned long *addr) |
| 30 | { | 30 | { |
| 31 | __asm__ __volatile__("lock; btsl %1,%0" | 31 | asm volatile("lock; btsl %1,%0" |
| 32 | :"+m" (ADDR) | 32 | : "+m" (ADDR) |
| 33 | :"Ir" (nr) | 33 | : "Ir" (nr) |
| 34 | : "memory"); | 34 | : "memory"); |
| 35 | } | 35 | } |
| 36 | 36 | ||
| 37 | /** | 37 | /** |
| @@ -44,12 +44,12 @@ static inline void sync_set_bit(int nr, volatile unsigned long * addr) | |||
| 44 | * you should call smp_mb__before_clear_bit() and/or smp_mb__after_clear_bit() | 44 | * you should call smp_mb__before_clear_bit() and/or smp_mb__after_clear_bit() |
| 45 | * in order to ensure changes are visible on other processors. | 45 | * in order to ensure changes are visible on other processors. |
| 46 | */ | 46 | */ |
| 47 | static inline void sync_clear_bit(int nr, volatile unsigned long * addr) | 47 | static inline void sync_clear_bit(int nr, volatile unsigned long *addr) |
| 48 | { | 48 | { |
| 49 | __asm__ __volatile__("lock; btrl %1,%0" | 49 | asm volatile("lock; btrl %1,%0" |
| 50 | :"+m" (ADDR) | 50 | : "+m" (ADDR) |
| 51 | :"Ir" (nr) | 51 | : "Ir" (nr) |
| 52 | : "memory"); | 52 | : "memory"); |
| 53 | } | 53 | } |
| 54 | 54 | ||
| 55 | /** | 55 | /** |
| @@ -61,12 +61,12 @@ static inline void sync_clear_bit(int nr, volatile unsigned long * addr) | |||
| 61 | * Note that @nr may be almost arbitrarily large; this function is not | 61 | * Note that @nr may be almost arbitrarily large; this function is not |
| 62 | * restricted to acting on a single-word quantity. | 62 | * restricted to acting on a single-word quantity. |
| 63 | */ | 63 | */ |
| 64 | static inline void sync_change_bit(int nr, volatile unsigned long * addr) | 64 | static inline void sync_change_bit(int nr, volatile unsigned long *addr) |
| 65 | { | 65 | { |
| 66 | __asm__ __volatile__("lock; btcl %1,%0" | 66 | asm volatile("lock; btcl %1,%0" |
| 67 | :"+m" (ADDR) | 67 | : "+m" (ADDR) |
| 68 | :"Ir" (nr) | 68 | : "Ir" (nr) |
| 69 | : "memory"); | 69 | : "memory"); |
| 70 | } | 70 | } |
| 71 | 71 | ||
| 72 | /** | 72 | /** |
| @@ -77,13 +77,13 @@ static inline void sync_change_bit(int nr, volatile unsigned long * addr) | |||
| 77 | * This operation is atomic and cannot be reordered. | 77 | * This operation is atomic and cannot be reordered. |
| 78 | * It also implies a memory barrier. | 78 | * It also implies a memory barrier. |
| 79 | */ | 79 | */ |
| 80 | static inline int sync_test_and_set_bit(int nr, volatile unsigned long * addr) | 80 | static inline int sync_test_and_set_bit(int nr, volatile unsigned long *addr) |
| 81 | { | 81 | { |
| 82 | int oldbit; | 82 | int oldbit; |
| 83 | 83 | ||
| 84 | __asm__ __volatile__("lock; btsl %2,%1\n\tsbbl %0,%0" | 84 | asm volatile("lock; btsl %2,%1\n\tsbbl %0,%0" |
| 85 | :"=r" (oldbit),"+m" (ADDR) | 85 | : "=r" (oldbit), "+m" (ADDR) |
| 86 | :"Ir" (nr) : "memory"); | 86 | : "Ir" (nr) : "memory"); |
| 87 | return oldbit; | 87 | return oldbit; |
| 88 | } | 88 | } |
| 89 | 89 | ||
| @@ -95,13 +95,13 @@ static inline int sync_test_and_set_bit(int nr, volatile unsigned long * addr) | |||
| 95 | * This operation is atomic and cannot be reordered. | 95 | * This operation is atomic and cannot be reordered. |
| 96 | * It also implies a memory barrier. | 96 | * It also implies a memory barrier. |
| 97 | */ | 97 | */ |
| 98 | static inline int sync_test_and_clear_bit(int nr, volatile unsigned long * addr) | 98 | static inline int sync_test_and_clear_bit(int nr, volatile unsigned long *addr) |
| 99 | { | 99 | { |
| 100 | int oldbit; | 100 | int oldbit; |
| 101 | 101 | ||
| 102 | __asm__ __volatile__("lock; btrl %2,%1\n\tsbbl %0,%0" | 102 | asm volatile("lock; btrl %2,%1\n\tsbbl %0,%0" |
| 103 | :"=r" (oldbit),"+m" (ADDR) | 103 | : "=r" (oldbit), "+m" (ADDR) |
| 104 | :"Ir" (nr) : "memory"); | 104 | : "Ir" (nr) : "memory"); |
| 105 | return oldbit; | 105 | return oldbit; |
| 106 | } | 106 | } |
| 107 | 107 | ||
| @@ -113,36 +113,17 @@ static inline int sync_test_and_clear_bit(int nr, volatile unsigned long * addr) | |||
| 113 | * This operation is atomic and cannot be reordered. | 113 | * This operation is atomic and cannot be reordered. |
| 114 | * It also implies a memory barrier. | 114 | * It also implies a memory barrier. |
| 115 | */ | 115 | */ |
| 116 | static inline int sync_test_and_change_bit(int nr, volatile unsigned long* addr) | 116 | static inline int sync_test_and_change_bit(int nr, volatile unsigned long *addr) |
| 117 | { | 117 | { |
| 118 | int oldbit; | 118 | int oldbit; |
| 119 | 119 | ||
| 120 | __asm__ __volatile__("lock; btcl %2,%1\n\tsbbl %0,%0" | 120 | asm volatile("lock; btcl %2,%1\n\tsbbl %0,%0" |
| 121 | :"=r" (oldbit),"+m" (ADDR) | 121 | : "=r" (oldbit), "+m" (ADDR) |
| 122 | :"Ir" (nr) : "memory"); | 122 | : "Ir" (nr) : "memory"); |
| 123 | return oldbit; | 123 | return oldbit; |
| 124 | } | 124 | } |
| 125 | 125 | ||
| 126 | static __always_inline int sync_constant_test_bit(int nr, const volatile unsigned long *addr) | 126 | #define sync_test_bit(nr, addr) test_bit(nr, addr) |
| 127 | { | ||
| 128 | return ((1UL << (nr & 31)) & | ||
| 129 | (((const volatile unsigned int *)addr)[nr >> 5])) != 0; | ||
| 130 | } | ||
| 131 | |||
| 132 | static inline int sync_var_test_bit(int nr, const volatile unsigned long * addr) | ||
| 133 | { | ||
| 134 | int oldbit; | ||
| 135 | |||
| 136 | __asm__ __volatile__("btl %2,%1\n\tsbbl %0,%0" | ||
| 137 | :"=r" (oldbit) | ||
| 138 | :"m" (ADDR),"Ir" (nr)); | ||
| 139 | return oldbit; | ||
| 140 | } | ||
| 141 | |||
| 142 | #define sync_test_bit(nr,addr) \ | ||
| 143 | (__builtin_constant_p(nr) ? \ | ||
| 144 | sync_constant_test_bit((nr),(addr)) : \ | ||
| 145 | sync_var_test_bit((nr),(addr))) | ||
| 146 | 127 | ||
| 147 | #undef ADDR | 128 | #undef ADDR |
| 148 | 129 | ||
diff --git a/include/asm-x86/system.h b/include/asm-x86/system.h index 9cff02ffe6c2..a2f04cd79b29 100644 --- a/include/asm-x86/system.h +++ b/include/asm-x86/system.h | |||
| @@ -27,22 +27,44 @@ struct task_struct *__switch_to(struct task_struct *prev, | |||
| 27 | * Saving eflags is important. It switches not only IOPL between tasks, | 27 | * Saving eflags is important. It switches not only IOPL between tasks, |
| 28 | * it also protects other tasks from NT leaking through sysenter etc. | 28 | * it also protects other tasks from NT leaking through sysenter etc. |
| 29 | */ | 29 | */ |
| 30 | #define switch_to(prev, next, last) do { \ | 30 | #define switch_to(prev, next, last) \ |
| 31 | unsigned long esi, edi; \ | 31 | do { \ |
| 32 | asm volatile("pushfl\n\t" /* Save flags */ \ | 32 | /* \ |
| 33 | "pushl %%ebp\n\t" \ | 33 | * Context-switching clobbers all registers, so we clobber \ |
| 34 | "movl %%esp,%0\n\t" /* save ESP */ \ | 34 | * them explicitly, via unused output variables. \ |
| 35 | "movl %5,%%esp\n\t" /* restore ESP */ \ | 35 | * (EAX and EBP is not listed because EBP is saved/restored \ |
| 36 | "movl $1f,%1\n\t" /* save EIP */ \ | 36 | * explicitly for wchan access and EAX is the return value of \ |
| 37 | "pushl %6\n\t" /* restore EIP */ \ | 37 | * __switch_to()) \ |
| 38 | "jmp __switch_to\n" \ | 38 | */ \ |
| 39 | unsigned long ebx, ecx, edx, esi, edi; \ | ||
| 40 | \ | ||
| 41 | asm volatile("pushfl\n\t" /* save flags */ \ | ||
| 42 | "pushl %%ebp\n\t" /* save EBP */ \ | ||
| 43 | "movl %%esp,%[prev_sp]\n\t" /* save ESP */ \ | ||
| 44 | "movl %[next_sp],%%esp\n\t" /* restore ESP */ \ | ||
| 45 | "movl $1f,%[prev_ip]\n\t" /* save EIP */ \ | ||
| 46 | "pushl %[next_ip]\n\t" /* restore EIP */ \ | ||
| 47 | "jmp __switch_to\n" /* regparm call */ \ | ||
| 39 | "1:\t" \ | 48 | "1:\t" \ |
| 40 | "popl %%ebp\n\t" \ | 49 | "popl %%ebp\n\t" /* restore EBP */ \ |
| 41 | "popfl" \ | 50 | "popfl\n" /* restore flags */ \ |
| 42 | :"=m" (prev->thread.sp), "=m" (prev->thread.ip), \ | 51 | \ |
| 43 | "=a" (last), "=S" (esi), "=D" (edi) \ | 52 | /* output parameters */ \ |
| 44 | :"m" (next->thread.sp), "m" (next->thread.ip), \ | 53 | : [prev_sp] "=m" (prev->thread.sp), \ |
| 45 | "2" (prev), "d" (next)); \ | 54 | [prev_ip] "=m" (prev->thread.ip), \ |
| 55 | "=a" (last), \ | ||
| 56 | \ | ||
| 57 | /* clobbered output registers: */ \ | ||
| 58 | "=b" (ebx), "=c" (ecx), "=d" (edx), \ | ||
| 59 | "=S" (esi), "=D" (edi) \ | ||
| 60 | \ | ||
| 61 | /* input parameters: */ \ | ||
| 62 | : [next_sp] "m" (next->thread.sp), \ | ||
| 63 | [next_ip] "m" (next->thread.ip), \ | ||
| 64 | \ | ||
| 65 | /* regparm parameters for __switch_to(): */ \ | ||
| 66 | [prev] "a" (prev), \ | ||
| 67 | [next] "d" (next)); \ | ||
| 46 | } while (0) | 68 | } while (0) |
| 47 | 69 | ||
| 48 | /* | 70 | /* |
| @@ -122,35 +144,34 @@ extern void load_gs_index(unsigned); | |||
| 122 | */ | 144 | */ |
| 123 | #define loadsegment(seg, value) \ | 145 | #define loadsegment(seg, value) \ |
| 124 | asm volatile("\n" \ | 146 | asm volatile("\n" \ |
| 125 | "1:\t" \ | 147 | "1:\t" \ |
| 126 | "movl %k0,%%" #seg "\n" \ | 148 | "movl %k0,%%" #seg "\n" \ |
| 127 | "2:\n" \ | 149 | "2:\n" \ |
| 128 | ".section .fixup,\"ax\"\n" \ | 150 | ".section .fixup,\"ax\"\n" \ |
| 129 | "3:\t" \ | 151 | "3:\t" \ |
| 130 | "movl %k1, %%" #seg "\n\t" \ | 152 | "movl %k1, %%" #seg "\n\t" \ |
| 131 | "jmp 2b\n" \ | 153 | "jmp 2b\n" \ |
| 132 | ".previous\n" \ | 154 | ".previous\n" \ |
| 133 | _ASM_EXTABLE(1b,3b) \ | 155 | _ASM_EXTABLE(1b,3b) \ |
| 134 | : :"r" (value), "r" (0)) | 156 | : :"r" (value), "r" (0)) |
| 135 | 157 | ||
| 136 | 158 | ||
| 137 | /* | 159 | /* |
| 138 | * Save a segment register away | 160 | * Save a segment register away |
| 139 | */ | 161 | */ |
| 140 | #define savesegment(seg, value) \ | 162 | #define savesegment(seg, value) \ |
| 141 | asm volatile("mov %%" #seg ",%0":"=rm" (value)) | 163 | asm volatile("mov %%" #seg ",%0":"=rm" (value)) |
| 142 | 164 | ||
| 143 | static inline unsigned long get_limit(unsigned long segment) | 165 | static inline unsigned long get_limit(unsigned long segment) |
| 144 | { | 166 | { |
| 145 | unsigned long __limit; | 167 | unsigned long __limit; |
| 146 | __asm__("lsll %1,%0" | 168 | asm("lsll %1,%0" : "=r" (__limit) : "r" (segment)); |
| 147 | :"=r" (__limit):"r" (segment)); | 169 | return __limit + 1; |
| 148 | return __limit+1; | ||
| 149 | } | 170 | } |
| 150 | 171 | ||
| 151 | static inline void native_clts(void) | 172 | static inline void native_clts(void) |
| 152 | { | 173 | { |
| 153 | asm volatile ("clts"); | 174 | asm volatile("clts"); |
| 154 | } | 175 | } |
| 155 | 176 | ||
| 156 | /* | 177 | /* |
| @@ -165,43 +186,43 @@ static unsigned long __force_order; | |||
| 165 | static inline unsigned long native_read_cr0(void) | 186 | static inline unsigned long native_read_cr0(void) |
| 166 | { | 187 | { |
| 167 | unsigned long val; | 188 | unsigned long val; |
| 168 | asm volatile("mov %%cr0,%0\n\t" :"=r" (val), "=m" (__force_order)); | 189 | asm volatile("mov %%cr0,%0\n\t" : "=r" (val), "=m" (__force_order)); |
| 169 | return val; | 190 | return val; |
| 170 | } | 191 | } |
| 171 | 192 | ||
| 172 | static inline void native_write_cr0(unsigned long val) | 193 | static inline void native_write_cr0(unsigned long val) |
| 173 | { | 194 | { |
| 174 | asm volatile("mov %0,%%cr0": :"r" (val), "m" (__force_order)); | 195 | asm volatile("mov %0,%%cr0": : "r" (val), "m" (__force_order)); |
| 175 | } | 196 | } |
| 176 | 197 | ||
| 177 | static inline unsigned long native_read_cr2(void) | 198 | static inline unsigned long native_read_cr2(void) |
| 178 | { | 199 | { |
| 179 | unsigned long val; | 200 | unsigned long val; |
| 180 | asm volatile("mov %%cr2,%0\n\t" :"=r" (val), "=m" (__force_order)); | 201 | asm volatile("mov %%cr2,%0\n\t" : "=r" (val), "=m" (__force_order)); |
| 181 | return val; | 202 | return val; |
| 182 | } | 203 | } |
| 183 | 204 | ||
| 184 | static inline void native_write_cr2(unsigned long val) | 205 | static inline void native_write_cr2(unsigned long val) |
| 185 | { | 206 | { |
| 186 | asm volatile("mov %0,%%cr2": :"r" (val), "m" (__force_order)); | 207 | asm volatile("mov %0,%%cr2": : "r" (val), "m" (__force_order)); |
| 187 | } | 208 | } |
| 188 | 209 | ||
| 189 | static inline unsigned long native_read_cr3(void) | 210 | static inline unsigned long native_read_cr3(void) |
| 190 | { | 211 | { |
| 191 | unsigned long val; | 212 | unsigned long val; |
| 192 | asm volatile("mov %%cr3,%0\n\t" :"=r" (val), "=m" (__force_order)); | 213 | asm volatile("mov %%cr3,%0\n\t" : "=r" (val), "=m" (__force_order)); |
| 193 | return val; | 214 | return val; |
| 194 | } | 215 | } |
| 195 | 216 | ||
| 196 | static inline void native_write_cr3(unsigned long val) | 217 | static inline void native_write_cr3(unsigned long val) |
| 197 | { | 218 | { |
| 198 | asm volatile("mov %0,%%cr3": :"r" (val), "m" (__force_order)); | 219 | asm volatile("mov %0,%%cr3": : "r" (val), "m" (__force_order)); |
| 199 | } | 220 | } |
| 200 | 221 | ||
| 201 | static inline unsigned long native_read_cr4(void) | 222 | static inline unsigned long native_read_cr4(void) |
| 202 | { | 223 | { |
| 203 | unsigned long val; | 224 | unsigned long val; |
| 204 | asm volatile("mov %%cr4,%0\n\t" :"=r" (val), "=m" (__force_order)); | 225 | asm volatile("mov %%cr4,%0\n\t" : "=r" (val), "=m" (__force_order)); |
| 205 | return val; | 226 | return val; |
| 206 | } | 227 | } |
| 207 | 228 | ||
| @@ -213,7 +234,7 @@ static inline unsigned long native_read_cr4_safe(void) | |||
| 213 | #ifdef CONFIG_X86_32 | 234 | #ifdef CONFIG_X86_32 |
| 214 | asm volatile("1: mov %%cr4, %0\n" | 235 | asm volatile("1: mov %%cr4, %0\n" |
| 215 | "2:\n" | 236 | "2:\n" |
| 216 | _ASM_EXTABLE(1b,2b) | 237 | _ASM_EXTABLE(1b, 2b) |
| 217 | : "=r" (val), "=m" (__force_order) : "0" (0)); | 238 | : "=r" (val), "=m" (__force_order) : "0" (0)); |
| 218 | #else | 239 | #else |
| 219 | val = native_read_cr4(); | 240 | val = native_read_cr4(); |
| @@ -223,7 +244,7 @@ static inline unsigned long native_read_cr4_safe(void) | |||
| 223 | 244 | ||
| 224 | static inline void native_write_cr4(unsigned long val) | 245 | static inline void native_write_cr4(unsigned long val) |
| 225 | { | 246 | { |
| 226 | asm volatile("mov %0,%%cr4": :"r" (val), "m" (__force_order)); | 247 | asm volatile("mov %0,%%cr4": : "r" (val), "m" (__force_order)); |
| 227 | } | 248 | } |
| 228 | 249 | ||
| 229 | #ifdef CONFIG_X86_64 | 250 | #ifdef CONFIG_X86_64 |
| @@ -244,6 +265,7 @@ static inline void native_wbinvd(void) | |||
| 244 | { | 265 | { |
| 245 | asm volatile("wbinvd": : :"memory"); | 266 | asm volatile("wbinvd": : :"memory"); |
| 246 | } | 267 | } |
| 268 | |||
| 247 | #ifdef CONFIG_PARAVIRT | 269 | #ifdef CONFIG_PARAVIRT |
| 248 | #include <asm/paravirt.h> | 270 | #include <asm/paravirt.h> |
| 249 | #else | 271 | #else |
| @@ -276,7 +298,7 @@ static inline void clflush(volatile void *__p) | |||
| 276 | asm volatile("clflush %0" : "+m" (*(volatile char __force *)__p)); | 298 | asm volatile("clflush %0" : "+m" (*(volatile char __force *)__p)); |
| 277 | } | 299 | } |
| 278 | 300 | ||
| 279 | #define nop() __asm__ __volatile__ ("nop") | 301 | #define nop() asm volatile ("nop") |
| 280 | 302 | ||
| 281 | void disable_hlt(void); | 303 | void disable_hlt(void); |
| 282 | void enable_hlt(void); | 304 | void enable_hlt(void); |
| @@ -296,16 +318,7 @@ void default_idle(void); | |||
| 296 | */ | 318 | */ |
| 297 | #ifdef CONFIG_X86_32 | 319 | #ifdef CONFIG_X86_32 |
| 298 | /* | 320 | /* |
| 299 | * For now, "wmb()" doesn't actually do anything, as all | 321 | * Some non-Intel clones support out of order store. wmb() ceases to be a |
| 300 | * Intel CPU's follow what Intel calls a *Processor Order*, | ||
| 301 | * in which all writes are seen in the program order even | ||
| 302 | * outside the CPU. | ||
| 303 | * | ||
| 304 | * I expect future Intel CPU's to have a weaker ordering, | ||
| 305 | * but I'd also expect them to finally get their act together | ||
| 306 | * and add some real memory barriers if so. | ||
| 307 | * | ||
| 308 | * Some non intel clones support out of order store. wmb() ceases to be a | ||
| 309 | * nop for these. | 322 | * nop for these. |
| 310 | */ | 323 | */ |
| 311 | #define mb() alternative("lock; addl $0,0(%%esp)", "mfence", X86_FEATURE_XMM2) | 324 | #define mb() alternative("lock; addl $0,0(%%esp)", "mfence", X86_FEATURE_XMM2) |
| @@ -384,7 +397,7 @@ void default_idle(void); | |||
| 384 | # define smp_wmb() barrier() | 397 | # define smp_wmb() barrier() |
| 385 | #endif | 398 | #endif |
| 386 | #define smp_read_barrier_depends() read_barrier_depends() | 399 | #define smp_read_barrier_depends() read_barrier_depends() |
| 387 | #define set_mb(var, value) do { (void) xchg(&var, value); } while (0) | 400 | #define set_mb(var, value) do { (void)xchg(&var, value); } while (0) |
| 388 | #else | 401 | #else |
| 389 | #define smp_mb() barrier() | 402 | #define smp_mb() barrier() |
| 390 | #define smp_rmb() barrier() | 403 | #define smp_rmb() barrier() |
diff --git a/include/asm-x86/tce.h b/include/asm-x86/tce.h index cd955d3d112f..b1a4ea00df78 100644 --- a/include/asm-x86/tce.h +++ b/include/asm-x86/tce.h | |||
| @@ -39,7 +39,7 @@ struct iommu_table; | |||
| 39 | #define TCE_RPN_MASK 0x0000fffffffff000ULL | 39 | #define TCE_RPN_MASK 0x0000fffffffff000ULL |
| 40 | 40 | ||
| 41 | extern void tce_build(struct iommu_table *tbl, unsigned long index, | 41 | extern void tce_build(struct iommu_table *tbl, unsigned long index, |
| 42 | unsigned int npages, unsigned long uaddr, int direction); | 42 | unsigned int npages, unsigned long uaddr, int direction); |
| 43 | extern void tce_free(struct iommu_table *tbl, long index, unsigned int npages); | 43 | extern void tce_free(struct iommu_table *tbl, long index, unsigned int npages); |
| 44 | extern void * __init alloc_tce_table(void); | 44 | extern void * __init alloc_tce_table(void); |
| 45 | extern void __init free_tce_table(void *tbl); | 45 | extern void __init free_tce_table(void *tbl); |
diff --git a/include/asm-x86/thread_info_32.h b/include/asm-x86/thread_info_32.h index 5bd508260ffb..4e053fa561a9 100644 --- a/include/asm-x86/thread_info_32.h +++ b/include/asm-x86/thread_info_32.h | |||
| @@ -20,7 +20,8 @@ | |||
| 20 | * low level task data that entry.S needs immediate access to | 20 | * low level task data that entry.S needs immediate access to |
| 21 | * - this struct should fit entirely inside of one cache line | 21 | * - this struct should fit entirely inside of one cache line |
| 22 | * - this struct shares the supervisor stack pages | 22 | * - this struct shares the supervisor stack pages |
| 23 | * - if the contents of this structure are changed, the assembly constants must also be changed | 23 | * - if the contents of this structure are changed, |
| 24 | * the assembly constants must also be changed | ||
| 24 | */ | 25 | */ |
| 25 | #ifndef __ASSEMBLY__ | 26 | #ifndef __ASSEMBLY__ |
| 26 | 27 | ||
| @@ -30,18 +31,16 @@ struct thread_info { | |||
| 30 | unsigned long flags; /* low level flags */ | 31 | unsigned long flags; /* low level flags */ |
| 31 | unsigned long status; /* thread-synchronous flags */ | 32 | unsigned long status; /* thread-synchronous flags */ |
| 32 | __u32 cpu; /* current CPU */ | 33 | __u32 cpu; /* current CPU */ |
| 33 | int preempt_count; /* 0 => preemptable, <0 => BUG */ | 34 | int preempt_count; /* 0 => preemptable, |
| 34 | 35 | <0 => BUG */ | |
| 35 | |||
| 36 | mm_segment_t addr_limit; /* thread address space: | 36 | mm_segment_t addr_limit; /* thread address space: |
| 37 | 0-0xBFFFFFFF for user-thead | 37 | 0-0xBFFFFFFF user-thread |
| 38 | 0-0xFFFFFFFF for kernel-thread | 38 | 0-0xFFFFFFFF kernel-thread |
| 39 | */ | 39 | */ |
| 40 | void *sysenter_return; | 40 | void *sysenter_return; |
| 41 | struct restart_block restart_block; | 41 | struct restart_block restart_block; |
| 42 | 42 | unsigned long previous_esp; /* ESP of the previous stack in | |
| 43 | unsigned long previous_esp; /* ESP of the previous stack in case | 43 | case of nested (IRQ) stacks |
| 44 | of nested (IRQ) stacks | ||
| 45 | */ | 44 | */ |
| 46 | __u8 supervisor_stack[0]; | 45 | __u8 supervisor_stack[0]; |
| 47 | }; | 46 | }; |
| @@ -90,15 +89,16 @@ register unsigned long current_stack_pointer asm("esp") __used; | |||
| 90 | /* how to get the thread information struct from C */ | 89 | /* how to get the thread information struct from C */ |
| 91 | static inline struct thread_info *current_thread_info(void) | 90 | static inline struct thread_info *current_thread_info(void) |
| 92 | { | 91 | { |
| 93 | return (struct thread_info *)(current_stack_pointer & ~(THREAD_SIZE - 1)); | 92 | return (struct thread_info *) |
| 93 | (current_stack_pointer & ~(THREAD_SIZE - 1)); | ||
| 94 | } | 94 | } |
| 95 | 95 | ||
| 96 | /* thread information allocation */ | 96 | /* thread information allocation */ |
| 97 | #ifdef CONFIG_DEBUG_STACK_USAGE | 97 | #ifdef CONFIG_DEBUG_STACK_USAGE |
| 98 | #define alloc_thread_info(tsk) ((struct thread_info *) \ | 98 | #define alloc_thread_info(tsk) ((struct thread_info *) \ |
| 99 | __get_free_pages(GFP_KERNEL| __GFP_ZERO, get_order(THREAD_SIZE))) | 99 | __get_free_pages(GFP_KERNEL | __GFP_ZERO, get_order(THREAD_SIZE))) |
| 100 | #else | 100 | #else |
| 101 | #define alloc_thread_info(tsk) ((struct thread_info *) \ | 101 | #define alloc_thread_info(tsk) ((struct thread_info *) \ |
| 102 | __get_free_pages(GFP_KERNEL, get_order(THREAD_SIZE))) | 102 | __get_free_pages(GFP_KERNEL, get_order(THREAD_SIZE))) |
| 103 | #endif | 103 | #endif |
| 104 | 104 | ||
| @@ -107,7 +107,7 @@ static inline struct thread_info *current_thread_info(void) | |||
| 107 | #else /* !__ASSEMBLY__ */ | 107 | #else /* !__ASSEMBLY__ */ |
| 108 | 108 | ||
| 109 | /* how to get the thread information struct from ASM */ | 109 | /* how to get the thread information struct from ASM */ |
| 110 | #define GET_THREAD_INFO(reg) \ | 110 | #define GET_THREAD_INFO(reg) \ |
| 111 | movl $-THREAD_SIZE, reg; \ | 111 | movl $-THREAD_SIZE, reg; \ |
| 112 | andl %esp, reg | 112 | andl %esp, reg |
| 113 | 113 | ||
| @@ -119,14 +119,16 @@ static inline struct thread_info *current_thread_info(void) | |||
| 119 | 119 | ||
| 120 | /* | 120 | /* |
| 121 | * thread information flags | 121 | * thread information flags |
| 122 | * - these are process state flags that various assembly files may need to access | 122 | * - these are process state flags that various |
| 123 | * assembly files may need to access | ||
| 123 | * - pending work-to-be-done flags are in LSW | 124 | * - pending work-to-be-done flags are in LSW |
| 124 | * - other flags in MSW | 125 | * - other flags in MSW |
| 125 | */ | 126 | */ |
| 126 | #define TIF_SYSCALL_TRACE 0 /* syscall trace active */ | 127 | #define TIF_SYSCALL_TRACE 0 /* syscall trace active */ |
| 127 | #define TIF_SIGPENDING 1 /* signal pending */ | 128 | #define TIF_SIGPENDING 1 /* signal pending */ |
| 128 | #define TIF_NEED_RESCHED 2 /* rescheduling necessary */ | 129 | #define TIF_NEED_RESCHED 2 /* rescheduling necessary */ |
| 129 | #define TIF_SINGLESTEP 3 /* restore singlestep on return to user mode */ | 130 | #define TIF_SINGLESTEP 3 /* restore singlestep on return to |
| 131 | user mode */ | ||
| 130 | #define TIF_IRET 4 /* return with iret */ | 132 | #define TIF_IRET 4 /* return with iret */ |
| 131 | #define TIF_SYSCALL_EMU 5 /* syscall emulation active */ | 133 | #define TIF_SYSCALL_EMU 5 /* syscall emulation active */ |
| 132 | #define TIF_SYSCALL_AUDIT 6 /* syscall auditing active */ | 134 | #define TIF_SYSCALL_AUDIT 6 /* syscall auditing active */ |
| @@ -143,36 +145,36 @@ static inline struct thread_info *current_thread_info(void) | |||
| 143 | #define TIF_DS_AREA_MSR 23 /* uses thread_struct.ds_area_msr */ | 145 | #define TIF_DS_AREA_MSR 23 /* uses thread_struct.ds_area_msr */ |
| 144 | #define TIF_BTS_TRACE_TS 24 /* record scheduling event timestamps */ | 146 | #define TIF_BTS_TRACE_TS 24 /* record scheduling event timestamps */ |
| 145 | 147 | ||
| 146 | #define _TIF_SYSCALL_TRACE (1<<TIF_SYSCALL_TRACE) | 148 | #define _TIF_SYSCALL_TRACE (1 << TIF_SYSCALL_TRACE) |
| 147 | #define _TIF_SIGPENDING (1<<TIF_SIGPENDING) | 149 | #define _TIF_SIGPENDING (1 << TIF_SIGPENDING) |
| 148 | #define _TIF_NEED_RESCHED (1<<TIF_NEED_RESCHED) | 150 | #define _TIF_NEED_RESCHED (1 << TIF_NEED_RESCHED) |
| 149 | #define _TIF_SINGLESTEP (1<<TIF_SINGLESTEP) | 151 | #define _TIF_SINGLESTEP (1 << TIF_SINGLESTEP) |
| 150 | #define _TIF_IRET (1<<TIF_IRET) | 152 | #define _TIF_IRET (1 << TIF_IRET) |
| 151 | #define _TIF_SYSCALL_EMU (1<<TIF_SYSCALL_EMU) | 153 | #define _TIF_SYSCALL_EMU (1 << TIF_SYSCALL_EMU) |
| 152 | #define _TIF_SYSCALL_AUDIT (1<<TIF_SYSCALL_AUDIT) | 154 | #define _TIF_SYSCALL_AUDIT (1 << TIF_SYSCALL_AUDIT) |
| 153 | #define _TIF_SECCOMP (1<<TIF_SECCOMP) | 155 | #define _TIF_SECCOMP (1 << TIF_SECCOMP) |
| 154 | #define _TIF_RESTORE_SIGMASK (1<<TIF_RESTORE_SIGMASK) | 156 | #define _TIF_RESTORE_SIGMASK (1 << TIF_RESTORE_SIGMASK) |
| 155 | #define _TIF_HRTICK_RESCHED (1<<TIF_HRTICK_RESCHED) | 157 | #define _TIF_HRTICK_RESCHED (1 << TIF_HRTICK_RESCHED) |
| 156 | #define _TIF_DEBUG (1<<TIF_DEBUG) | 158 | #define _TIF_DEBUG (1 << TIF_DEBUG) |
| 157 | #define _TIF_IO_BITMAP (1<<TIF_IO_BITMAP) | 159 | #define _TIF_IO_BITMAP (1 << TIF_IO_BITMAP) |
| 158 | #define _TIF_FREEZE (1<<TIF_FREEZE) | 160 | #define _TIF_FREEZE (1 << TIF_FREEZE) |
| 159 | #define _TIF_NOTSC (1<<TIF_NOTSC) | 161 | #define _TIF_NOTSC (1 << TIF_NOTSC) |
| 160 | #define _TIF_FORCED_TF (1<<TIF_FORCED_TF) | 162 | #define _TIF_FORCED_TF (1 << TIF_FORCED_TF) |
| 161 | #define _TIF_DEBUGCTLMSR (1<<TIF_DEBUGCTLMSR) | 163 | #define _TIF_DEBUGCTLMSR (1 << TIF_DEBUGCTLMSR) |
| 162 | #define _TIF_DS_AREA_MSR (1<<TIF_DS_AREA_MSR) | 164 | #define _TIF_DS_AREA_MSR (1 << TIF_DS_AREA_MSR) |
| 163 | #define _TIF_BTS_TRACE_TS (1<<TIF_BTS_TRACE_TS) | 165 | #define _TIF_BTS_TRACE_TS (1 << TIF_BTS_TRACE_TS) |
| 164 | 166 | ||
| 165 | /* work to do on interrupt/exception return */ | 167 | /* work to do on interrupt/exception return */ |
| 166 | #define _TIF_WORK_MASK \ | 168 | #define _TIF_WORK_MASK \ |
| 167 | (0x0000FFFF & ~(_TIF_SYSCALL_TRACE | _TIF_SYSCALL_AUDIT | \ | 169 | (0x0000FFFF & ~(_TIF_SYSCALL_TRACE | _TIF_SYSCALL_AUDIT | \ |
| 168 | _TIF_SECCOMP | _TIF_SYSCALL_EMU)) | 170 | _TIF_SECCOMP | _TIF_SYSCALL_EMU)) |
| 169 | /* work to do on any return to u-space */ | 171 | /* work to do on any return to u-space */ |
| 170 | #define _TIF_ALLWORK_MASK (0x0000FFFF & ~_TIF_SECCOMP) | 172 | #define _TIF_ALLWORK_MASK (0x0000FFFF & ~_TIF_SECCOMP) |
| 171 | 173 | ||
| 172 | /* flags to check in __switch_to() */ | 174 | /* flags to check in __switch_to() */ |
| 173 | #define _TIF_WORK_CTXSW \ | 175 | #define _TIF_WORK_CTXSW \ |
| 174 | (_TIF_IO_BITMAP | _TIF_NOTSC | _TIF_DEBUGCTLMSR | \ | 176 | (_TIF_IO_BITMAP | _TIF_NOTSC | _TIF_DEBUGCTLMSR | \ |
| 175 | _TIF_DS_AREA_MSR | _TIF_BTS_TRACE_TS) | 177 | _TIF_DS_AREA_MSR | _TIF_BTS_TRACE_TS) |
| 176 | #define _TIF_WORK_CTXSW_PREV _TIF_WORK_CTXSW | 178 | #define _TIF_WORK_CTXSW_PREV _TIF_WORK_CTXSW |
| 177 | #define _TIF_WORK_CTXSW_NEXT (_TIF_WORK_CTXSW | _TIF_DEBUG) | 179 | #define _TIF_WORK_CTXSW_NEXT (_TIF_WORK_CTXSW | _TIF_DEBUG) |
| 178 | 180 | ||
| @@ -184,8 +186,10 @@ static inline struct thread_info *current_thread_info(void) | |||
| 184 | * ever touches our thread-synchronous status, so we don't | 186 | * ever touches our thread-synchronous status, so we don't |
| 185 | * have to worry about atomic accesses. | 187 | * have to worry about atomic accesses. |
| 186 | */ | 188 | */ |
| 187 | #define TS_USEDFPU 0x0001 /* FPU was used by this task this quantum (SMP) */ | 189 | #define TS_USEDFPU 0x0001 /* FPU was used by this task |
| 188 | #define TS_POLLING 0x0002 /* True if in idle loop and not sleeping */ | 190 | this quantum (SMP) */ |
| 191 | #define TS_POLLING 0x0002 /* True if in idle loop | ||
| 192 | and not sleeping */ | ||
| 189 | 193 | ||
| 190 | #define tsk_is_polling(t) (task_thread_info(t)->status & TS_POLLING) | 194 | #define tsk_is_polling(t) (task_thread_info(t)->status & TS_POLLING) |
| 191 | 195 | ||
diff --git a/include/asm-x86/thread_info_64.h b/include/asm-x86/thread_info_64.h index 6c9b214b8fc3..1e5c6f6152cd 100644 --- a/include/asm-x86/thread_info_64.h +++ b/include/asm-x86/thread_info_64.h | |||
| @@ -29,9 +29,9 @@ struct thread_info { | |||
| 29 | __u32 flags; /* low level flags */ | 29 | __u32 flags; /* low level flags */ |
| 30 | __u32 status; /* thread synchronous flags */ | 30 | __u32 status; /* thread synchronous flags */ |
| 31 | __u32 cpu; /* current CPU */ | 31 | __u32 cpu; /* current CPU */ |
| 32 | int preempt_count; /* 0 => preemptable, <0 => BUG */ | 32 | int preempt_count; /* 0 => preemptable, |
| 33 | 33 | <0 => BUG */ | |
| 34 | mm_segment_t addr_limit; | 34 | mm_segment_t addr_limit; |
| 35 | struct restart_block restart_block; | 35 | struct restart_block restart_block; |
| 36 | #ifdef CONFIG_IA32_EMULATION | 36 | #ifdef CONFIG_IA32_EMULATION |
| 37 | void __user *sysenter_return; | 37 | void __user *sysenter_return; |
| @@ -61,17 +61,17 @@ struct thread_info { | |||
| 61 | #define init_stack (init_thread_union.stack) | 61 | #define init_stack (init_thread_union.stack) |
| 62 | 62 | ||
| 63 | static inline struct thread_info *current_thread_info(void) | 63 | static inline struct thread_info *current_thread_info(void) |
| 64 | { | 64 | { |
| 65 | struct thread_info *ti; | 65 | struct thread_info *ti; |
| 66 | ti = (void *)(read_pda(kernelstack) + PDA_STACKOFFSET - THREAD_SIZE); | 66 | ti = (void *)(read_pda(kernelstack) + PDA_STACKOFFSET - THREAD_SIZE); |
| 67 | return ti; | 67 | return ti; |
| 68 | } | 68 | } |
| 69 | 69 | ||
| 70 | /* do not use in interrupt context */ | 70 | /* do not use in interrupt context */ |
| 71 | static inline struct thread_info *stack_thread_info(void) | 71 | static inline struct thread_info *stack_thread_info(void) |
| 72 | { | 72 | { |
| 73 | struct thread_info *ti; | 73 | struct thread_info *ti; |
| 74 | __asm__("andq %%rsp,%0; ":"=r" (ti) : "0" (~(THREAD_SIZE - 1))); | 74 | asm("andq %%rsp,%0; " : "=r" (ti) : "0" (~(THREAD_SIZE - 1))); |
| 75 | return ti; | 75 | return ti; |
| 76 | } | 76 | } |
| 77 | 77 | ||
| @@ -82,8 +82,8 @@ static inline struct thread_info *stack_thread_info(void) | |||
| 82 | #define THREAD_FLAGS GFP_KERNEL | 82 | #define THREAD_FLAGS GFP_KERNEL |
| 83 | #endif | 83 | #endif |
| 84 | 84 | ||
| 85 | #define alloc_thread_info(tsk) \ | 85 | #define alloc_thread_info(tsk) \ |
| 86 | ((struct thread_info *) __get_free_pages(THREAD_FLAGS, THREAD_ORDER)) | 86 | ((struct thread_info *)__get_free_pages(THREAD_FLAGS, THREAD_ORDER)) |
| 87 | 87 | ||
| 88 | #define free_thread_info(ti) free_pages((unsigned long) (ti), THREAD_ORDER) | 88 | #define free_thread_info(ti) free_pages((unsigned long) (ti), THREAD_ORDER) |
| 89 | 89 | ||
| @@ -98,7 +98,8 @@ static inline struct thread_info *stack_thread_info(void) | |||
| 98 | 98 | ||
| 99 | /* | 99 | /* |
| 100 | * thread information flags | 100 | * thread information flags |
| 101 | * - these are process state flags that various assembly files may need to access | 101 | * - these are process state flags that various assembly files |
| 102 | * may need to access | ||
| 102 | * - pending work-to-be-done flags are in LSW | 103 | * - pending work-to-be-done flags are in LSW |
| 103 | * - other flags in MSW | 104 | * - other flags in MSW |
| 104 | * Warning: layout of LSW is hardcoded in entry.S | 105 | * Warning: layout of LSW is hardcoded in entry.S |
| @@ -114,7 +115,7 @@ static inline struct thread_info *stack_thread_info(void) | |||
| 114 | #define TIF_MCE_NOTIFY 10 /* notify userspace of an MCE */ | 115 | #define TIF_MCE_NOTIFY 10 /* notify userspace of an MCE */ |
| 115 | #define TIF_HRTICK_RESCHED 11 /* reprogram hrtick timer */ | 116 | #define TIF_HRTICK_RESCHED 11 /* reprogram hrtick timer */ |
| 116 | /* 16 free */ | 117 | /* 16 free */ |
| 117 | #define TIF_IA32 17 /* 32bit process */ | 118 | #define TIF_IA32 17 /* 32bit process */ |
| 118 | #define TIF_FORK 18 /* ret_from_fork */ | 119 | #define TIF_FORK 18 /* ret_from_fork */ |
| 119 | #define TIF_ABI_PENDING 19 | 120 | #define TIF_ABI_PENDING 19 |
| 120 | #define TIF_MEMDIE 20 | 121 | #define TIF_MEMDIE 20 |
| @@ -126,39 +127,40 @@ static inline struct thread_info *stack_thread_info(void) | |||
| 126 | #define TIF_DS_AREA_MSR 26 /* uses thread_struct.ds_area_msr */ | 127 | #define TIF_DS_AREA_MSR 26 /* uses thread_struct.ds_area_msr */ |
| 127 | #define TIF_BTS_TRACE_TS 27 /* record scheduling event timestamps */ | 128 | #define TIF_BTS_TRACE_TS 27 /* record scheduling event timestamps */ |
| 128 | 129 | ||
| 129 | #define _TIF_SYSCALL_TRACE (1<<TIF_SYSCALL_TRACE) | 130 | #define _TIF_SYSCALL_TRACE (1 << TIF_SYSCALL_TRACE) |
| 130 | #define _TIF_SIGPENDING (1<<TIF_SIGPENDING) | 131 | #define _TIF_SIGPENDING (1 << TIF_SIGPENDING) |
| 131 | #define _TIF_SINGLESTEP (1<<TIF_SINGLESTEP) | 132 | #define _TIF_SINGLESTEP (1 << TIF_SINGLESTEP) |
| 132 | #define _TIF_NEED_RESCHED (1<<TIF_NEED_RESCHED) | 133 | #define _TIF_NEED_RESCHED (1 << TIF_NEED_RESCHED) |
| 133 | #define _TIF_IRET (1<<TIF_IRET) | 134 | #define _TIF_IRET (1 << TIF_IRET) |
| 134 | #define _TIF_SYSCALL_AUDIT (1<<TIF_SYSCALL_AUDIT) | 135 | #define _TIF_SYSCALL_AUDIT (1 << TIF_SYSCALL_AUDIT) |
| 135 | #define _TIF_SECCOMP (1<<TIF_SECCOMP) | 136 | #define _TIF_SECCOMP (1 << TIF_SECCOMP) |
| 136 | #define _TIF_RESTORE_SIGMASK (1<<TIF_RESTORE_SIGMASK) | 137 | #define _TIF_RESTORE_SIGMASK (1 << TIF_RESTORE_SIGMASK) |
| 137 | #define _TIF_MCE_NOTIFY (1<<TIF_MCE_NOTIFY) | 138 | #define _TIF_MCE_NOTIFY (1 << TIF_MCE_NOTIFY) |
| 138 | #define _TIF_HRTICK_RESCHED (1<<TIF_HRTICK_RESCHED) | 139 | #define _TIF_HRTICK_RESCHED (1 << TIF_HRTICK_RESCHED) |
| 139 | #define _TIF_IA32 (1<<TIF_IA32) | 140 | #define _TIF_IA32 (1 << TIF_IA32) |
| 140 | #define _TIF_FORK (1<<TIF_FORK) | 141 | #define _TIF_FORK (1 << TIF_FORK) |
| 141 | #define _TIF_ABI_PENDING (1<<TIF_ABI_PENDING) | 142 | #define _TIF_ABI_PENDING (1 << TIF_ABI_PENDING) |
| 142 | #define _TIF_DEBUG (1<<TIF_DEBUG) | 143 | #define _TIF_DEBUG (1 << TIF_DEBUG) |
| 143 | #define _TIF_IO_BITMAP (1<<TIF_IO_BITMAP) | 144 | #define _TIF_IO_BITMAP (1 << TIF_IO_BITMAP) |
| 144 | #define _TIF_FREEZE (1<<TIF_FREEZE) | 145 | #define _TIF_FREEZE (1 << TIF_FREEZE) |
| 145 | #define _TIF_FORCED_TF (1<<TIF_FORCED_TF) | 146 | #define _TIF_FORCED_TF (1 << TIF_FORCED_TF) |
| 146 | #define _TIF_DEBUGCTLMSR (1<<TIF_DEBUGCTLMSR) | 147 | #define _TIF_DEBUGCTLMSR (1 << TIF_DEBUGCTLMSR) |
| 147 | #define _TIF_DS_AREA_MSR (1<<TIF_DS_AREA_MSR) | 148 | #define _TIF_DS_AREA_MSR (1 << TIF_DS_AREA_MSR) |
| 148 | #define _TIF_BTS_TRACE_TS (1<<TIF_BTS_TRACE_TS) | 149 | #define _TIF_BTS_TRACE_TS (1 << TIF_BTS_TRACE_TS) |
| 149 | 150 | ||
| 150 | /* work to do on interrupt/exception return */ | 151 | /* work to do on interrupt/exception return */ |
| 151 | #define _TIF_WORK_MASK \ | 152 | #define _TIF_WORK_MASK \ |
| 152 | (0x0000FFFF & ~(_TIF_SYSCALL_TRACE|_TIF_SYSCALL_AUDIT|_TIF_SINGLESTEP|_TIF_SECCOMP)) | 153 | (0x0000FFFF & \ |
| 154 | ~(_TIF_SYSCALL_TRACE|_TIF_SYSCALL_AUDIT|_TIF_SINGLESTEP|_TIF_SECCOMP)) | ||
| 153 | /* work to do on any return to user space */ | 155 | /* work to do on any return to user space */ |
| 154 | #define _TIF_ALLWORK_MASK (0x0000FFFF & ~_TIF_SECCOMP) | 156 | #define _TIF_ALLWORK_MASK (0x0000FFFF & ~_TIF_SECCOMP) |
| 155 | 157 | ||
| 156 | #define _TIF_DO_NOTIFY_MASK \ | 158 | #define _TIF_DO_NOTIFY_MASK \ |
| 157 | (_TIF_SIGPENDING|_TIF_SINGLESTEP|_TIF_MCE_NOTIFY|_TIF_HRTICK_RESCHED) | 159 | (_TIF_SIGPENDING|_TIF_SINGLESTEP|_TIF_MCE_NOTIFY|_TIF_HRTICK_RESCHED) |
| 158 | 160 | ||
| 159 | /* flags to check in __switch_to() */ | 161 | /* flags to check in __switch_to() */ |
| 160 | #define _TIF_WORK_CTXSW \ | 162 | #define _TIF_WORK_CTXSW \ |
| 161 | (_TIF_IO_BITMAP|_TIF_DEBUGCTLMSR|_TIF_DS_AREA_MSR|_TIF_BTS_TRACE_TS) | 163 | (_TIF_IO_BITMAP|_TIF_DEBUGCTLMSR|_TIF_DS_AREA_MSR|_TIF_BTS_TRACE_TS) |
| 162 | #define _TIF_WORK_CTXSW_PREV _TIF_WORK_CTXSW | 164 | #define _TIF_WORK_CTXSW_PREV _TIF_WORK_CTXSW |
| 163 | #define _TIF_WORK_CTXSW_NEXT (_TIF_WORK_CTXSW|_TIF_DEBUG) | 165 | #define _TIF_WORK_CTXSW_NEXT (_TIF_WORK_CTXSW|_TIF_DEBUG) |
| 164 | 166 | ||
| @@ -171,9 +173,11 @@ static inline struct thread_info *stack_thread_info(void) | |||
| 171 | * ever touches our thread-synchronous status, so we don't | 173 | * ever touches our thread-synchronous status, so we don't |
| 172 | * have to worry about atomic accesses. | 174 | * have to worry about atomic accesses. |
| 173 | */ | 175 | */ |
| 174 | #define TS_USEDFPU 0x0001 /* FPU was used by this task this quantum (SMP) */ | 176 | #define TS_USEDFPU 0x0001 /* FPU was used by this task |
| 177 | this quantum (SMP) */ | ||
| 175 | #define TS_COMPAT 0x0002 /* 32bit syscall active */ | 178 | #define TS_COMPAT 0x0002 /* 32bit syscall active */ |
| 176 | #define TS_POLLING 0x0004 /* true if in idle loop and not sleeping */ | 179 | #define TS_POLLING 0x0004 /* true if in idle loop |
| 180 | and not sleeping */ | ||
| 177 | 181 | ||
| 178 | #define tsk_is_polling(t) (task_thread_info(t)->status & TS_POLLING) | 182 | #define tsk_is_polling(t) (task_thread_info(t)->status & TS_POLLING) |
| 179 | 183 | ||
diff --git a/include/asm-x86/tlbflush.h b/include/asm-x86/tlbflush.h index 3998709ed637..0c0674d94255 100644 --- a/include/asm-x86/tlbflush.h +++ b/include/asm-x86/tlbflush.h | |||
| @@ -32,7 +32,7 @@ static inline void __native_flush_tlb_global(void) | |||
| 32 | 32 | ||
| 33 | static inline void __native_flush_tlb_single(unsigned long addr) | 33 | static inline void __native_flush_tlb_single(unsigned long addr) |
| 34 | { | 34 | { |
| 35 | __asm__ __volatile__("invlpg (%0)" ::"r" (addr) : "memory"); | 35 | asm volatile("invlpg (%0)" ::"r" (addr) : "memory"); |
| 36 | } | 36 | } |
| 37 | 37 | ||
| 38 | static inline void __flush_tlb_all(void) | 38 | static inline void __flush_tlb_all(void) |
| @@ -134,8 +134,7 @@ void native_flush_tlb_others(const cpumask_t *cpumask, struct mm_struct *mm, | |||
| 134 | #define TLBSTATE_LAZY 2 | 134 | #define TLBSTATE_LAZY 2 |
| 135 | 135 | ||
| 136 | #ifdef CONFIG_X86_32 | 136 | #ifdef CONFIG_X86_32 |
| 137 | struct tlb_state | 137 | struct tlb_state { |
| 138 | { | ||
| 139 | struct mm_struct *active_mm; | 138 | struct mm_struct *active_mm; |
| 140 | int state; | 139 | int state; |
| 141 | char __cacheline_padding[L1_CACHE_BYTES-8]; | 140 | char __cacheline_padding[L1_CACHE_BYTES-8]; |
diff --git a/include/asm-x86/topology.h b/include/asm-x86/topology.h index 8af05a93f097..81a29eb08ac4 100644 --- a/include/asm-x86/topology.h +++ b/include/asm-x86/topology.h | |||
| @@ -32,13 +32,18 @@ | |||
| 32 | /* Mappings between logical cpu number and node number */ | 32 | /* Mappings between logical cpu number and node number */ |
| 33 | #ifdef CONFIG_X86_32 | 33 | #ifdef CONFIG_X86_32 |
| 34 | extern int cpu_to_node_map[]; | 34 | extern int cpu_to_node_map[]; |
| 35 | |||
| 36 | #else | 35 | #else |
| 36 | /* Returns the number of the current Node. */ | ||
| 37 | #define numa_node_id() (early_cpu_to_node(raw_smp_processor_id())) | ||
| 38 | #endif | ||
| 39 | |||
| 37 | DECLARE_PER_CPU(int, x86_cpu_to_node_map); | 40 | DECLARE_PER_CPU(int, x86_cpu_to_node_map); |
| 41 | |||
| 42 | #ifdef CONFIG_SMP | ||
| 38 | extern int x86_cpu_to_node_map_init[]; | 43 | extern int x86_cpu_to_node_map_init[]; |
| 39 | extern void *x86_cpu_to_node_map_early_ptr; | 44 | extern void *x86_cpu_to_node_map_early_ptr; |
| 40 | /* Returns the number of the current Node. */ | 45 | #else |
| 41 | #define numa_node_id() (early_cpu_to_node(raw_smp_processor_id())) | 46 | #define x86_cpu_to_node_map_early_ptr NULL |
| 42 | #endif | 47 | #endif |
| 43 | 48 | ||
| 44 | extern cpumask_t node_to_cpumask_map[]; | 49 | extern cpumask_t node_to_cpumask_map[]; |
| @@ -54,6 +59,8 @@ static inline int cpu_to_node(int cpu) | |||
| 54 | } | 59 | } |
| 55 | 60 | ||
| 56 | #else /* CONFIG_X86_64 */ | 61 | #else /* CONFIG_X86_64 */ |
| 62 | |||
| 63 | #ifdef CONFIG_SMP | ||
| 57 | static inline int early_cpu_to_node(int cpu) | 64 | static inline int early_cpu_to_node(int cpu) |
| 58 | { | 65 | { |
| 59 | int *cpu_to_node_map = x86_cpu_to_node_map_early_ptr; | 66 | int *cpu_to_node_map = x86_cpu_to_node_map_early_ptr; |
| @@ -65,21 +72,21 @@ static inline int early_cpu_to_node(int cpu) | |||
| 65 | else | 72 | else |
| 66 | return NUMA_NO_NODE; | 73 | return NUMA_NO_NODE; |
| 67 | } | 74 | } |
| 75 | #else | ||
| 76 | #define early_cpu_to_node(cpu) cpu_to_node(cpu) | ||
| 77 | #endif | ||
| 68 | 78 | ||
| 69 | static inline int cpu_to_node(int cpu) | 79 | static inline int cpu_to_node(int cpu) |
| 70 | { | 80 | { |
| 71 | #ifdef CONFIG_DEBUG_PER_CPU_MAPS | 81 | #ifdef CONFIG_DEBUG_PER_CPU_MAPS |
| 72 | if (x86_cpu_to_node_map_early_ptr) { | 82 | if (x86_cpu_to_node_map_early_ptr) { |
| 73 | printk("KERN_NOTICE cpu_to_node(%d): usage too early!\n", | 83 | printk("KERN_NOTICE cpu_to_node(%d): usage too early!\n", |
| 74 | (int)cpu); | 84 | (int)cpu); |
| 75 | dump_stack(); | 85 | dump_stack(); |
| 76 | return ((int *)x86_cpu_to_node_map_early_ptr)[cpu]; | 86 | return ((int *)x86_cpu_to_node_map_early_ptr)[cpu]; |
| 77 | } | 87 | } |
| 78 | #endif | 88 | #endif |
| 79 | if (per_cpu_offset(cpu)) | 89 | return per_cpu(x86_cpu_to_node_map, cpu); |
| 80 | return per_cpu(x86_cpu_to_node_map, cpu); | ||
| 81 | else | ||
| 82 | return NUMA_NO_NODE; | ||
| 83 | } | 90 | } |
| 84 | #endif /* CONFIG_X86_64 */ | 91 | #endif /* CONFIG_X86_64 */ |
| 85 | 92 | ||
diff --git a/include/asm-x86/trampoline.h b/include/asm-x86/trampoline.h new file mode 100644 index 000000000000..b156b08d0131 --- /dev/null +++ b/include/asm-x86/trampoline.h | |||
| @@ -0,0 +1,21 @@ | |||
| 1 | #ifndef __TRAMPOLINE_HEADER | ||
| 2 | #define __TRAMPOLINE_HEADER | ||
| 3 | |||
| 4 | #ifndef __ASSEMBLY__ | ||
| 5 | |||
| 6 | /* | ||
| 7 | * Trampoline 80x86 program as an array. | ||
| 8 | */ | ||
| 9 | extern const unsigned char trampoline_data []; | ||
| 10 | extern const unsigned char trampoline_end []; | ||
| 11 | extern unsigned char *trampoline_base; | ||
| 12 | |||
| 13 | extern unsigned long init_rsp; | ||
| 14 | extern unsigned long initial_code; | ||
| 15 | |||
| 16 | #define TRAMPOLINE_BASE 0x6000 | ||
| 17 | extern unsigned long setup_trampoline(void); | ||
| 18 | |||
| 19 | #endif /* __ASSEMBLY__ */ | ||
| 20 | |||
| 21 | #endif /* __TRAMPOLINE_HEADER */ | ||
diff --git a/include/asm-x86/tsc.h b/include/asm-x86/tsc.h index 7d3e27f7d484..d2d8eb5b55f5 100644 --- a/include/asm-x86/tsc.h +++ b/include/asm-x86/tsc.h | |||
| @@ -42,7 +42,7 @@ static inline cycles_t vget_cycles(void) | |||
| 42 | if (!cpu_has_tsc) | 42 | if (!cpu_has_tsc) |
| 43 | return 0; | 43 | return 0; |
| 44 | #endif | 44 | #endif |
| 45 | return (cycles_t) __native_read_tsc(); | 45 | return (cycles_t)__native_read_tsc(); |
| 46 | } | 46 | } |
| 47 | 47 | ||
| 48 | extern void tsc_init(void); | 48 | extern void tsc_init(void); |
diff --git a/include/asm-x86/uaccess_32.h b/include/asm-x86/uaccess_32.h index fcc570ec4fee..8e7595c1f34e 100644 --- a/include/asm-x86/uaccess_32.h +++ b/include/asm-x86/uaccess_32.h | |||
| @@ -32,7 +32,7 @@ | |||
| 32 | #define get_fs() (current_thread_info()->addr_limit) | 32 | #define get_fs() (current_thread_info()->addr_limit) |
| 33 | #define set_fs(x) (current_thread_info()->addr_limit = (x)) | 33 | #define set_fs(x) (current_thread_info()->addr_limit = (x)) |
| 34 | 34 | ||
| 35 | #define segment_eq(a,b) ((a).seg == (b).seg) | 35 | #define segment_eq(a, b) ((a).seg == (b).seg) |
| 36 | 36 | ||
| 37 | /* | 37 | /* |
| 38 | * movsl can be slow when source and dest are not both 8-byte aligned | 38 | * movsl can be slow when source and dest are not both 8-byte aligned |
| @@ -43,7 +43,9 @@ extern struct movsl_mask { | |||
| 43 | } ____cacheline_aligned_in_smp movsl_mask; | 43 | } ____cacheline_aligned_in_smp movsl_mask; |
| 44 | #endif | 44 | #endif |
| 45 | 45 | ||
| 46 | #define __addr_ok(addr) ((unsigned long __force)(addr) < (current_thread_info()->addr_limit.seg)) | 46 | #define __addr_ok(addr) \ |
| 47 | ((unsigned long __force)(addr) < \ | ||
| 48 | (current_thread_info()->addr_limit.seg)) | ||
| 47 | 49 | ||
| 48 | /* | 50 | /* |
| 49 | * Test whether a block of memory is a valid user space address. | 51 | * Test whether a block of memory is a valid user space address. |
| @@ -54,13 +56,16 @@ extern struct movsl_mask { | |||
| 54 | * | 56 | * |
| 55 | * This needs 33-bit arithmetic. We have a carry... | 57 | * This needs 33-bit arithmetic. We have a carry... |
| 56 | */ | 58 | */ |
| 57 | #define __range_ok(addr,size) ({ \ | 59 | #define __range_ok(addr, size) \ |
| 58 | unsigned long flag,roksum; \ | 60 | ({ \ |
| 59 | __chk_user_ptr(addr); \ | 61 | unsigned long flag, roksum; \ |
| 60 | asm("addl %3,%1 ; sbbl %0,%0; cmpl %1,%4; sbbl $0,%0" \ | 62 | __chk_user_ptr(addr); \ |
| 61 | :"=&r" (flag), "=r" (roksum) \ | 63 | asm("addl %3,%1 ; sbbl %0,%0; cmpl %1,%4; sbbl $0,%0" \ |
| 62 | :"1" (addr),"g" ((int)(size)),"rm" (current_thread_info()->addr_limit.seg)); \ | 64 | :"=&r" (flag), "=r" (roksum) \ |
| 63 | flag; }) | 65 | :"1" (addr), "g" ((int)(size)), \ |
| 66 | "rm" (current_thread_info()->addr_limit.seg)); \ | ||
| 67 | flag; \ | ||
| 68 | }) | ||
| 64 | 69 | ||
| 65 | /** | 70 | /** |
| 66 | * access_ok: - Checks if a user space pointer is valid | 71 | * access_ok: - Checks if a user space pointer is valid |
| @@ -81,7 +86,7 @@ extern struct movsl_mask { | |||
| 81 | * checks that the pointer is in the user space range - after calling | 86 | * checks that the pointer is in the user space range - after calling |
| 82 | * this function, memory access functions may still return -EFAULT. | 87 | * this function, memory access functions may still return -EFAULT. |
| 83 | */ | 88 | */ |
| 84 | #define access_ok(type,addr,size) (likely(__range_ok(addr,size) == 0)) | 89 | #define access_ok(type, addr, size) (likely(__range_ok(addr, size) == 0)) |
| 85 | 90 | ||
| 86 | /* | 91 | /* |
| 87 | * The exception table consists of pairs of addresses: the first is the | 92 | * The exception table consists of pairs of addresses: the first is the |
| @@ -96,8 +101,7 @@ extern struct movsl_mask { | |||
| 96 | * on our cache or tlb entries. | 101 | * on our cache or tlb entries. |
| 97 | */ | 102 | */ |
| 98 | 103 | ||
| 99 | struct exception_table_entry | 104 | struct exception_table_entry { |
| 100 | { | ||
| 101 | unsigned long insn, fixup; | 105 | unsigned long insn, fixup; |
| 102 | }; | 106 | }; |
| 103 | 107 | ||
| @@ -122,13 +126,15 @@ extern void __get_user_1(void); | |||
| 122 | extern void __get_user_2(void); | 126 | extern void __get_user_2(void); |
| 123 | extern void __get_user_4(void); | 127 | extern void __get_user_4(void); |
| 124 | 128 | ||
| 125 | #define __get_user_x(size,ret,x,ptr) \ | 129 | #define __get_user_x(size, ret, x, ptr) \ |
| 126 | __asm__ __volatile__("call __get_user_" #size \ | 130 | asm volatile("call __get_user_" #size \ |
| 127 | :"=a" (ret),"=d" (x) \ | 131 | :"=a" (ret),"=d" (x) \ |
| 128 | :"0" (ptr)) | 132 | :"0" (ptr)) |
| 133 | |||
| 129 | 134 | ||
| 135 | /* Careful: we have to cast the result to the type of the pointer | ||
| 136 | * for sign reasons */ | ||
| 130 | 137 | ||
| 131 | /* Careful: we have to cast the result to the type of the pointer for sign reasons */ | ||
| 132 | /** | 138 | /** |
| 133 | * get_user: - Get a simple variable from user space. | 139 | * get_user: - Get a simple variable from user space. |
| 134 | * @x: Variable to store result. | 140 | * @x: Variable to store result. |
| @@ -146,15 +152,24 @@ extern void __get_user_4(void); | |||
| 146 | * Returns zero on success, or -EFAULT on error. | 152 | * Returns zero on success, or -EFAULT on error. |
| 147 | * On error, the variable @x is set to zero. | 153 | * On error, the variable @x is set to zero. |
| 148 | */ | 154 | */ |
| 149 | #define get_user(x,ptr) \ | 155 | #define get_user(x, ptr) \ |
| 150 | ({ int __ret_gu; \ | 156 | ({ \ |
| 157 | int __ret_gu; \ | ||
| 151 | unsigned long __val_gu; \ | 158 | unsigned long __val_gu; \ |
| 152 | __chk_user_ptr(ptr); \ | 159 | __chk_user_ptr(ptr); \ |
| 153 | switch(sizeof (*(ptr))) { \ | 160 | switch (sizeof(*(ptr))) { \ |
| 154 | case 1: __get_user_x(1,__ret_gu,__val_gu,ptr); break; \ | 161 | case 1: \ |
| 155 | case 2: __get_user_x(2,__ret_gu,__val_gu,ptr); break; \ | 162 | __get_user_x(1, __ret_gu, __val_gu, ptr); \ |
| 156 | case 4: __get_user_x(4,__ret_gu,__val_gu,ptr); break; \ | 163 | break; \ |
| 157 | default: __get_user_x(X,__ret_gu,__val_gu,ptr); break; \ | 164 | case 2: \ |
| 165 | __get_user_x(2, __ret_gu, __val_gu, ptr); \ | ||
| 166 | break; \ | ||
| 167 | case 4: \ | ||
| 168 | __get_user_x(4, __ret_gu, __val_gu, ptr); \ | ||
| 169 | break; \ | ||
| 170 | default: \ | ||
| 171 | __get_user_x(X, __ret_gu, __val_gu, ptr); \ | ||
| 172 | break; \ | ||
| 158 | } \ | 173 | } \ |
| 159 | (x) = (__typeof__(*(ptr)))__val_gu; \ | 174 | (x) = (__typeof__(*(ptr)))__val_gu; \ |
| 160 | __ret_gu; \ | 175 | __ret_gu; \ |
| @@ -171,11 +186,25 @@ extern void __put_user_2(void); | |||
| 171 | extern void __put_user_4(void); | 186 | extern void __put_user_4(void); |
| 172 | extern void __put_user_8(void); | 187 | extern void __put_user_8(void); |
| 173 | 188 | ||
| 174 | #define __put_user_1(x, ptr) __asm__ __volatile__("call __put_user_1":"=a" (__ret_pu):"0" ((typeof(*(ptr)))(x)), "c" (ptr)) | 189 | #define __put_user_1(x, ptr) \ |
| 175 | #define __put_user_2(x, ptr) __asm__ __volatile__("call __put_user_2":"=a" (__ret_pu):"0" ((typeof(*(ptr)))(x)), "c" (ptr)) | 190 | asm volatile("call __put_user_1" : "=a" (__ret_pu) \ |
| 176 | #define __put_user_4(x, ptr) __asm__ __volatile__("call __put_user_4":"=a" (__ret_pu):"0" ((typeof(*(ptr)))(x)), "c" (ptr)) | 191 | : "0" ((typeof(*(ptr)))(x)), "c" (ptr)) |
| 177 | #define __put_user_8(x, ptr) __asm__ __volatile__("call __put_user_8":"=a" (__ret_pu):"A" ((typeof(*(ptr)))(x)), "c" (ptr)) | 192 | |
| 178 | #define __put_user_X(x, ptr) __asm__ __volatile__("call __put_user_X":"=a" (__ret_pu):"c" (ptr)) | 193 | #define __put_user_2(x, ptr) \ |
| 194 | asm volatile("call __put_user_2" : "=a" (__ret_pu) \ | ||
| 195 | : "0" ((typeof(*(ptr)))(x)), "c" (ptr)) | ||
| 196 | |||
| 197 | #define __put_user_4(x, ptr) \ | ||
| 198 | asm volatile("call __put_user_4" : "=a" (__ret_pu) \ | ||
| 199 | : "0" ((typeof(*(ptr)))(x)), "c" (ptr)) | ||
| 200 | |||
| 201 | #define __put_user_8(x, ptr) \ | ||
| 202 | asm volatile("call __put_user_8" : "=a" (__ret_pu) \ | ||
| 203 | : "A" ((typeof(*(ptr)))(x)), "c" (ptr)) | ||
| 204 | |||
| 205 | #define __put_user_X(x, ptr) \ | ||
| 206 | asm volatile("call __put_user_X" : "=a" (__ret_pu) \ | ||
| 207 | : "c" (ptr)) | ||
| 179 | 208 | ||
| 180 | /** | 209 | /** |
| 181 | * put_user: - Write a simple value into user space. | 210 | * put_user: - Write a simple value into user space. |
| @@ -195,32 +224,43 @@ extern void __put_user_8(void); | |||
| 195 | */ | 224 | */ |
| 196 | #ifdef CONFIG_X86_WP_WORKS_OK | 225 | #ifdef CONFIG_X86_WP_WORKS_OK |
| 197 | 226 | ||
| 198 | #define put_user(x,ptr) \ | 227 | #define put_user(x, ptr) \ |
| 199 | ({ int __ret_pu; \ | 228 | ({ \ |
| 229 | int __ret_pu; \ | ||
| 200 | __typeof__(*(ptr)) __pu_val; \ | 230 | __typeof__(*(ptr)) __pu_val; \ |
| 201 | __chk_user_ptr(ptr); \ | 231 | __chk_user_ptr(ptr); \ |
| 202 | __pu_val = x; \ | 232 | __pu_val = x; \ |
| 203 | switch(sizeof(*(ptr))) { \ | 233 | switch (sizeof(*(ptr))) { \ |
| 204 | case 1: __put_user_1(__pu_val, ptr); break; \ | 234 | case 1: \ |
| 205 | case 2: __put_user_2(__pu_val, ptr); break; \ | 235 | __put_user_1(__pu_val, ptr); \ |
| 206 | case 4: __put_user_4(__pu_val, ptr); break; \ | 236 | break; \ |
| 207 | case 8: __put_user_8(__pu_val, ptr); break; \ | 237 | case 2: \ |
| 208 | default:__put_user_X(__pu_val, ptr); break; \ | 238 | __put_user_2(__pu_val, ptr); \ |
| 239 | break; \ | ||
| 240 | case 4: \ | ||
| 241 | __put_user_4(__pu_val, ptr); \ | ||
| 242 | break; \ | ||
| 243 | case 8: \ | ||
| 244 | __put_user_8(__pu_val, ptr); \ | ||
| 245 | break; \ | ||
| 246 | default: \ | ||
| 247 | __put_user_X(__pu_val, ptr); \ | ||
| 248 | break; \ | ||
| 209 | } \ | 249 | } \ |
| 210 | __ret_pu; \ | 250 | __ret_pu; \ |
| 211 | }) | 251 | }) |
| 212 | 252 | ||
| 213 | #else | 253 | #else |
| 214 | #define put_user(x,ptr) \ | 254 | #define put_user(x, ptr) \ |
| 215 | ({ \ | 255 | ({ \ |
| 216 | int __ret_pu; \ | 256 | int __ret_pu; \ |
| 217 | __typeof__(*(ptr)) __pus_tmp = x; \ | 257 | __typeof__(*(ptr))__pus_tmp = x; \ |
| 218 | __ret_pu=0; \ | 258 | __ret_pu = 0; \ |
| 219 | if(unlikely(__copy_to_user_ll(ptr, &__pus_tmp, \ | 259 | if (unlikely(__copy_to_user_ll(ptr, &__pus_tmp, \ |
| 220 | sizeof(*(ptr))) != 0)) \ | 260 | sizeof(*(ptr))) != 0)) \ |
| 221 | __ret_pu=-EFAULT; \ | 261 | __ret_pu = -EFAULT; \ |
| 222 | __ret_pu; \ | 262 | __ret_pu; \ |
| 223 | }) | 263 | }) |
| 224 | 264 | ||
| 225 | 265 | ||
| 226 | #endif | 266 | #endif |
| @@ -245,8 +285,8 @@ extern void __put_user_8(void); | |||
| 245 | * Returns zero on success, or -EFAULT on error. | 285 | * Returns zero on success, or -EFAULT on error. |
| 246 | * On error, the variable @x is set to zero. | 286 | * On error, the variable @x is set to zero. |
| 247 | */ | 287 | */ |
| 248 | #define __get_user(x,ptr) \ | 288 | #define __get_user(x, ptr) \ |
| 249 | __get_user_nocheck((x),(ptr),sizeof(*(ptr))) | 289 | __get_user_nocheck((x), (ptr), sizeof(*(ptr))) |
| 250 | 290 | ||
| 251 | 291 | ||
| 252 | /** | 292 | /** |
| @@ -268,54 +308,62 @@ extern void __put_user_8(void); | |||
| 268 | * | 308 | * |
| 269 | * Returns zero on success, or -EFAULT on error. | 309 | * Returns zero on success, or -EFAULT on error. |
| 270 | */ | 310 | */ |
| 271 | #define __put_user(x,ptr) \ | 311 | #define __put_user(x, ptr) \ |
| 272 | __put_user_nocheck((__typeof__(*(ptr)))(x),(ptr),sizeof(*(ptr))) | 312 | __put_user_nocheck((__typeof__(*(ptr)))(x), (ptr), sizeof(*(ptr))) |
| 273 | 313 | ||
| 274 | #define __put_user_nocheck(x,ptr,size) \ | 314 | #define __put_user_nocheck(x, ptr, size) \ |
| 275 | ({ \ | 315 | ({ \ |
| 276 | long __pu_err; \ | 316 | long __pu_err; \ |
| 277 | __put_user_size((x),(ptr),(size),__pu_err,-EFAULT); \ | 317 | __put_user_size((x), (ptr), (size), __pu_err, -EFAULT); \ |
| 278 | __pu_err; \ | 318 | __pu_err; \ |
| 279 | }) | 319 | }) |
| 280 | 320 | ||
| 281 | 321 | ||
| 282 | #define __put_user_u64(x, addr, err) \ | 322 | #define __put_user_u64(x, addr, err) \ |
| 283 | __asm__ __volatile__( \ | 323 | asm volatile("1: movl %%eax,0(%2)\n" \ |
| 284 | "1: movl %%eax,0(%2)\n" \ | 324 | "2: movl %%edx,4(%2)\n" \ |
| 285 | "2: movl %%edx,4(%2)\n" \ | 325 | "3:\n" \ |
| 286 | "3:\n" \ | 326 | ".section .fixup,\"ax\"\n" \ |
| 287 | ".section .fixup,\"ax\"\n" \ | 327 | "4: movl %3,%0\n" \ |
| 288 | "4: movl %3,%0\n" \ | 328 | " jmp 3b\n" \ |
| 289 | " jmp 3b\n" \ | 329 | ".previous\n" \ |
| 290 | ".previous\n" \ | 330 | _ASM_EXTABLE(1b, 4b) \ |
| 291 | _ASM_EXTABLE(1b,4b) \ | 331 | _ASM_EXTABLE(2b, 4b) \ |
| 292 | _ASM_EXTABLE(2b,4b) \ | 332 | : "=r" (err) \ |
| 293 | : "=r"(err) \ | 333 | : "A" (x), "r" (addr), "i" (-EFAULT), "0" (err)) |
| 294 | : "A" (x), "r" (addr), "i"(-EFAULT), "0"(err)) | ||
| 295 | 334 | ||
| 296 | #ifdef CONFIG_X86_WP_WORKS_OK | 335 | #ifdef CONFIG_X86_WP_WORKS_OK |
| 297 | 336 | ||
| 298 | #define __put_user_size(x,ptr,size,retval,errret) \ | 337 | #define __put_user_size(x, ptr, size, retval, errret) \ |
| 299 | do { \ | 338 | do { \ |
| 300 | retval = 0; \ | 339 | retval = 0; \ |
| 301 | __chk_user_ptr(ptr); \ | 340 | __chk_user_ptr(ptr); \ |
| 302 | switch (size) { \ | 341 | switch (size) { \ |
| 303 | case 1: __put_user_asm(x,ptr,retval,"b","b","iq",errret);break; \ | 342 | case 1: \ |
| 304 | case 2: __put_user_asm(x,ptr,retval,"w","w","ir",errret);break; \ | 343 | __put_user_asm(x, ptr, retval, "b", "b", "iq", errret); \ |
| 305 | case 4: __put_user_asm(x,ptr,retval,"l","","ir",errret); break; \ | 344 | break; \ |
| 306 | case 8: __put_user_u64((__typeof__(*ptr))(x),ptr,retval); break;\ | 345 | case 2: \ |
| 307 | default: __put_user_bad(); \ | 346 | __put_user_asm(x, ptr, retval, "w", "w", "ir", errret); \ |
| 347 | break; \ | ||
| 348 | case 4: \ | ||
| 349 | __put_user_asm(x, ptr, retval, "l", "", "ir", errret); \ | ||
| 350 | break; \ | ||
| 351 | case 8: \ | ||
| 352 | __put_user_u64((__typeof__(*ptr))(x), ptr, retval); \ | ||
| 353 | break; \ | ||
| 354 | default: \ | ||
| 355 | __put_user_bad(); \ | ||
| 308 | } \ | 356 | } \ |
| 309 | } while (0) | 357 | } while (0) |
| 310 | 358 | ||
| 311 | #else | 359 | #else |
| 312 | 360 | ||
| 313 | #define __put_user_size(x,ptr,size,retval,errret) \ | 361 | #define __put_user_size(x, ptr, size, retval, errret) \ |
| 314 | do { \ | 362 | do { \ |
| 315 | __typeof__(*(ptr)) __pus_tmp = x; \ | 363 | __typeof__(*(ptr))__pus_tmp = x; \ |
| 316 | retval = 0; \ | 364 | retval = 0; \ |
| 317 | \ | 365 | \ |
| 318 | if(unlikely(__copy_to_user_ll(ptr, &__pus_tmp, size) != 0)) \ | 366 | if (unlikely(__copy_to_user_ll(ptr, &__pus_tmp, size) != 0)) \ |
| 319 | retval = errret; \ | 367 | retval = errret; \ |
| 320 | } while (0) | 368 | } while (0) |
| 321 | 369 | ||
| @@ -329,65 +377,70 @@ struct __large_struct { unsigned long buf[100]; }; | |||
| 329 | * aliasing issues. | 377 | * aliasing issues. |
| 330 | */ | 378 | */ |
| 331 | #define __put_user_asm(x, addr, err, itype, rtype, ltype, errret) \ | 379 | #define __put_user_asm(x, addr, err, itype, rtype, ltype, errret) \ |
| 332 | __asm__ __volatile__( \ | 380 | asm volatile("1: mov"itype" %"rtype"1,%2\n" \ |
| 333 | "1: mov"itype" %"rtype"1,%2\n" \ | 381 | "2:\n" \ |
| 334 | "2:\n" \ | 382 | ".section .fixup,\"ax\"\n" \ |
| 335 | ".section .fixup,\"ax\"\n" \ | 383 | "3: movl %3,%0\n" \ |
| 336 | "3: movl %3,%0\n" \ | 384 | " jmp 2b\n" \ |
| 337 | " jmp 2b\n" \ | 385 | ".previous\n" \ |
| 338 | ".previous\n" \ | 386 | _ASM_EXTABLE(1b, 3b) \ |
| 339 | _ASM_EXTABLE(1b,3b) \ | 387 | : "=r"(err) \ |
| 340 | : "=r"(err) \ | 388 | : ltype (x), "m" (__m(addr)), "i" (errret), "0" (err)) |
| 341 | : ltype (x), "m"(__m(addr)), "i"(errret), "0"(err)) | 389 | |
| 342 | 390 | ||
| 343 | 391 | #define __get_user_nocheck(x, ptr, size) \ | |
| 344 | #define __get_user_nocheck(x,ptr,size) \ | 392 | ({ \ |
| 345 | ({ \ | 393 | long __gu_err; \ |
| 346 | long __gu_err; \ | 394 | unsigned long __gu_val; \ |
| 347 | unsigned long __gu_val; \ | 395 | __get_user_size(__gu_val, (ptr), (size), __gu_err, -EFAULT); \ |
| 348 | __get_user_size(__gu_val,(ptr),(size),__gu_err,-EFAULT);\ | 396 | (x) = (__typeof__(*(ptr)))__gu_val; \ |
| 349 | (x) = (__typeof__(*(ptr)))__gu_val; \ | 397 | __gu_err; \ |
| 350 | __gu_err; \ | ||
| 351 | }) | 398 | }) |
| 352 | 399 | ||
| 353 | extern long __get_user_bad(void); | 400 | extern long __get_user_bad(void); |
| 354 | 401 | ||
| 355 | #define __get_user_size(x,ptr,size,retval,errret) \ | 402 | #define __get_user_size(x, ptr, size, retval, errret) \ |
| 356 | do { \ | 403 | do { \ |
| 357 | retval = 0; \ | 404 | retval = 0; \ |
| 358 | __chk_user_ptr(ptr); \ | 405 | __chk_user_ptr(ptr); \ |
| 359 | switch (size) { \ | 406 | switch (size) { \ |
| 360 | case 1: __get_user_asm(x,ptr,retval,"b","b","=q",errret);break; \ | 407 | case 1: \ |
| 361 | case 2: __get_user_asm(x,ptr,retval,"w","w","=r",errret);break; \ | 408 | __get_user_asm(x, ptr, retval, "b", "b", "=q", errret); \ |
| 362 | case 4: __get_user_asm(x,ptr,retval,"l","","=r",errret);break; \ | 409 | break; \ |
| 363 | default: (x) = __get_user_bad(); \ | 410 | case 2: \ |
| 411 | __get_user_asm(x, ptr, retval, "w", "w", "=r", errret); \ | ||
| 412 | break; \ | ||
| 413 | case 4: \ | ||
| 414 | __get_user_asm(x, ptr, retval, "l", "", "=r", errret); \ | ||
| 415 | break; \ | ||
| 416 | default: \ | ||
| 417 | (x) = __get_user_bad(); \ | ||
| 364 | } \ | 418 | } \ |
| 365 | } while (0) | 419 | } while (0) |
| 366 | 420 | ||
| 367 | #define __get_user_asm(x, addr, err, itype, rtype, ltype, errret) \ | 421 | #define __get_user_asm(x, addr, err, itype, rtype, ltype, errret) \ |
| 368 | __asm__ __volatile__( \ | 422 | asm volatile("1: mov"itype" %2,%"rtype"1\n" \ |
| 369 | "1: mov"itype" %2,%"rtype"1\n" \ | 423 | "2:\n" \ |
| 370 | "2:\n" \ | 424 | ".section .fixup,\"ax\"\n" \ |
| 371 | ".section .fixup,\"ax\"\n" \ | 425 | "3: movl %3,%0\n" \ |
| 372 | "3: movl %3,%0\n" \ | 426 | " xor"itype" %"rtype"1,%"rtype"1\n" \ |
| 373 | " xor"itype" %"rtype"1,%"rtype"1\n" \ | 427 | " jmp 2b\n" \ |
| 374 | " jmp 2b\n" \ | 428 | ".previous\n" \ |
| 375 | ".previous\n" \ | 429 | _ASM_EXTABLE(1b, 3b) \ |
| 376 | _ASM_EXTABLE(1b,3b) \ | 430 | : "=r" (err), ltype (x) \ |
| 377 | : "=r"(err), ltype (x) \ | 431 | : "m" (__m(addr)), "i" (errret), "0" (err)) |
| 378 | : "m"(__m(addr)), "i"(errret), "0"(err)) | 432 | |
| 379 | 433 | ||
| 380 | 434 | unsigned long __must_check __copy_to_user_ll | |
| 381 | unsigned long __must_check __copy_to_user_ll(void __user *to, | 435 | (void __user *to, const void *from, unsigned long n); |
| 382 | const void *from, unsigned long n); | 436 | unsigned long __must_check __copy_from_user_ll |
| 383 | unsigned long __must_check __copy_from_user_ll(void *to, | 437 | (void *to, const void __user *from, unsigned long n); |
| 384 | const void __user *from, unsigned long n); | 438 | unsigned long __must_check __copy_from_user_ll_nozero |
| 385 | unsigned long __must_check __copy_from_user_ll_nozero(void *to, | 439 | (void *to, const void __user *from, unsigned long n); |
| 386 | const void __user *from, unsigned long n); | 440 | unsigned long __must_check __copy_from_user_ll_nocache |
| 387 | unsigned long __must_check __copy_from_user_ll_nocache(void *to, | 441 | (void *to, const void __user *from, unsigned long n); |
| 388 | const void __user *from, unsigned long n); | 442 | unsigned long __must_check __copy_from_user_ll_nocache_nozero |
| 389 | unsigned long __must_check __copy_from_user_ll_nocache_nozero(void *to, | 443 | (void *to, const void __user *from, unsigned long n); |
| 390 | const void __user *from, unsigned long n); | ||
| 391 | 444 | ||
| 392 | /** | 445 | /** |
| 393 | * __copy_to_user_inatomic: - Copy a block of data into user space, with less checking. | 446 | * __copy_to_user_inatomic: - Copy a block of data into user space, with less checking. |
| @@ -416,13 +469,16 @@ __copy_to_user_inatomic(void __user *to, const void *from, unsigned long n) | |||
| 416 | 469 | ||
| 417 | switch (n) { | 470 | switch (n) { |
| 418 | case 1: | 471 | case 1: |
| 419 | __put_user_size(*(u8 *)from, (u8 __user *)to, 1, ret, 1); | 472 | __put_user_size(*(u8 *)from, (u8 __user *)to, |
| 473 | 1, ret, 1); | ||
| 420 | return ret; | 474 | return ret; |
| 421 | case 2: | 475 | case 2: |
| 422 | __put_user_size(*(u16 *)from, (u16 __user *)to, 2, ret, 2); | 476 | __put_user_size(*(u16 *)from, (u16 __user *)to, |
| 477 | 2, ret, 2); | ||
| 423 | return ret; | 478 | return ret; |
| 424 | case 4: | 479 | case 4: |
| 425 | __put_user_size(*(u32 *)from, (u32 __user *)to, 4, ret, 4); | 480 | __put_user_size(*(u32 *)from, (u32 __user *)to, |
| 481 | 4, ret, 4); | ||
| 426 | return ret; | 482 | return ret; |
| 427 | } | 483 | } |
| 428 | } | 484 | } |
| @@ -545,19 +601,21 @@ static __always_inline unsigned long __copy_from_user_nocache(void *to, | |||
| 545 | } | 601 | } |
| 546 | 602 | ||
| 547 | static __always_inline unsigned long | 603 | static __always_inline unsigned long |
| 548 | __copy_from_user_inatomic_nocache(void *to, const void __user *from, unsigned long n) | 604 | __copy_from_user_inatomic_nocache(void *to, const void __user *from, |
| 605 | unsigned long n) | ||
| 549 | { | 606 | { |
| 550 | return __copy_from_user_ll_nocache_nozero(to, from, n); | 607 | return __copy_from_user_ll_nocache_nozero(to, from, n); |
| 551 | } | 608 | } |
| 552 | 609 | ||
| 553 | unsigned long __must_check copy_to_user(void __user *to, | 610 | unsigned long __must_check copy_to_user(void __user *to, |
| 554 | const void *from, unsigned long n); | 611 | const void *from, unsigned long n); |
| 555 | unsigned long __must_check copy_from_user(void *to, | 612 | unsigned long __must_check copy_from_user(void *to, |
| 556 | const void __user *from, unsigned long n); | 613 | const void __user *from, |
| 614 | unsigned long n); | ||
| 557 | long __must_check strncpy_from_user(char *dst, const char __user *src, | 615 | long __must_check strncpy_from_user(char *dst, const char __user *src, |
| 558 | long count); | 616 | long count); |
| 559 | long __must_check __strncpy_from_user(char *dst, | 617 | long __must_check __strncpy_from_user(char *dst, |
| 560 | const char __user *src, long count); | 618 | const char __user *src, long count); |
| 561 | 619 | ||
| 562 | /** | 620 | /** |
| 563 | * strlen_user: - Get the size of a string in user space. | 621 | * strlen_user: - Get the size of a string in user space. |
diff --git a/include/asm-x86/uaccess_64.h b/include/asm-x86/uaccess_64.h index b87eb4ba8f9d..b8a2f4339903 100644 --- a/include/asm-x86/uaccess_64.h +++ b/include/asm-x86/uaccess_64.h | |||
| @@ -29,23 +29,27 @@ | |||
| 29 | #define get_fs() (current_thread_info()->addr_limit) | 29 | #define get_fs() (current_thread_info()->addr_limit) |
| 30 | #define set_fs(x) (current_thread_info()->addr_limit = (x)) | 30 | #define set_fs(x) (current_thread_info()->addr_limit = (x)) |
| 31 | 31 | ||
| 32 | #define segment_eq(a,b) ((a).seg == (b).seg) | 32 | #define segment_eq(a, b) ((a).seg == (b).seg) |
| 33 | 33 | ||
| 34 | #define __addr_ok(addr) (!((unsigned long)(addr) & (current_thread_info()->addr_limit.seg))) | 34 | #define __addr_ok(addr) (!((unsigned long)(addr) & \ |
| 35 | (current_thread_info()->addr_limit.seg))) | ||
| 35 | 36 | ||
| 36 | /* | 37 | /* |
| 37 | * Uhhuh, this needs 65-bit arithmetic. We have a carry.. | 38 | * Uhhuh, this needs 65-bit arithmetic. We have a carry.. |
| 38 | */ | 39 | */ |
| 39 | #define __range_not_ok(addr,size) ({ \ | 40 | #define __range_not_ok(addr, size) \ |
| 40 | unsigned long flag,roksum; \ | 41 | ({ \ |
| 41 | __chk_user_ptr(addr); \ | 42 | unsigned long flag, roksum; \ |
| 42 | asm("# range_ok\n\r" \ | 43 | __chk_user_ptr(addr); \ |
| 43 | "addq %3,%1 ; sbbq %0,%0 ; cmpq %1,%4 ; sbbq $0,%0" \ | 44 | asm("# range_ok\n\r" \ |
| 44 | :"=&r" (flag), "=r" (roksum) \ | 45 | "addq %3,%1 ; sbbq %0,%0 ; cmpq %1,%4 ; sbbq $0,%0" \ |
| 45 | :"1" (addr),"g" ((long)(size)),"g" (current_thread_info()->addr_limit.seg)); \ | 46 | : "=&r" (flag), "=r" (roksum) \ |
| 46 | flag; }) | 47 | : "1" (addr), "g" ((long)(size)), \ |
| 48 | "g" (current_thread_info()->addr_limit.seg)); \ | ||
| 49 | flag; \ | ||
| 50 | }) | ||
| 47 | 51 | ||
| 48 | #define access_ok(type, addr, size) (__range_not_ok(addr,size) == 0) | 52 | #define access_ok(type, addr, size) (__range_not_ok(addr, size) == 0) |
| 49 | 53 | ||
| 50 | /* | 54 | /* |
| 51 | * The exception table consists of pairs of addresses: the first is the | 55 | * The exception table consists of pairs of addresses: the first is the |
| @@ -60,8 +64,7 @@ | |||
| 60 | * on our cache or tlb entries. | 64 | * on our cache or tlb entries. |
| 61 | */ | 65 | */ |
| 62 | 66 | ||
| 63 | struct exception_table_entry | 67 | struct exception_table_entry { |
| 64 | { | ||
| 65 | unsigned long insn, fixup; | 68 | unsigned long insn, fixup; |
| 66 | }; | 69 | }; |
| 67 | 70 | ||
| @@ -84,23 +87,36 @@ extern int fixup_exception(struct pt_regs *regs); | |||
| 84 | * accesses to the same area of user memory). | 87 | * accesses to the same area of user memory). |
| 85 | */ | 88 | */ |
| 86 | 89 | ||
| 87 | #define __get_user_x(size,ret,x,ptr) \ | 90 | #define __get_user_x(size, ret, x, ptr) \ |
| 88 | asm volatile("call __get_user_" #size \ | 91 | asm volatile("call __get_user_" #size \ |
| 89 | :"=a" (ret),"=d" (x) \ | 92 | : "=a" (ret),"=d" (x) \ |
| 90 | :"c" (ptr) \ | 93 | : "c" (ptr) \ |
| 91 | :"r8") | 94 | : "r8") |
| 95 | |||
| 96 | /* Careful: we have to cast the result to the type of the pointer | ||
| 97 | * for sign reasons */ | ||
| 92 | 98 | ||
| 93 | /* Careful: we have to cast the result to the type of the pointer for sign reasons */ | 99 | #define get_user(x, ptr) \ |
| 94 | #define get_user(x,ptr) \ | 100 | ({ \ |
| 95 | ({ unsigned long __val_gu; \ | 101 | unsigned long __val_gu; \ |
| 96 | int __ret_gu; \ | 102 | int __ret_gu; \ |
| 97 | __chk_user_ptr(ptr); \ | 103 | __chk_user_ptr(ptr); \ |
| 98 | switch(sizeof (*(ptr))) { \ | 104 | switch (sizeof(*(ptr))) { \ |
| 99 | case 1: __get_user_x(1,__ret_gu,__val_gu,ptr); break; \ | 105 | case 1: \ |
| 100 | case 2: __get_user_x(2,__ret_gu,__val_gu,ptr); break; \ | 106 | __get_user_x(1, __ret_gu, __val_gu, ptr); \ |
| 101 | case 4: __get_user_x(4,__ret_gu,__val_gu,ptr); break; \ | 107 | break; \ |
| 102 | case 8: __get_user_x(8,__ret_gu,__val_gu,ptr); break; \ | 108 | case 2: \ |
| 103 | default: __get_user_bad(); break; \ | 109 | __get_user_x(2, __ret_gu, __val_gu, ptr); \ |
| 110 | break; \ | ||
| 111 | case 4: \ | ||
| 112 | __get_user_x(4, __ret_gu, __val_gu, ptr); \ | ||
| 113 | break; \ | ||
| 114 | case 8: \ | ||
| 115 | __get_user_x(8, __ret_gu, __val_gu, ptr); \ | ||
| 116 | break; \ | ||
| 117 | default: \ | ||
| 118 | __get_user_bad(); \ | ||
| 119 | break; \ | ||
| 104 | } \ | 120 | } \ |
| 105 | (x) = (__force typeof(*(ptr)))__val_gu; \ | 121 | (x) = (__force typeof(*(ptr)))__val_gu; \ |
| 106 | __ret_gu; \ | 122 | __ret_gu; \ |
| @@ -112,55 +128,73 @@ extern void __put_user_4(void); | |||
| 112 | extern void __put_user_8(void); | 128 | extern void __put_user_8(void); |
| 113 | extern void __put_user_bad(void); | 129 | extern void __put_user_bad(void); |
| 114 | 130 | ||
| 115 | #define __put_user_x(size,ret,x,ptr) \ | 131 | #define __put_user_x(size, ret, x, ptr) \ |
| 116 | asm volatile("call __put_user_" #size \ | 132 | asm volatile("call __put_user_" #size \ |
| 117 | :"=a" (ret) \ | 133 | :"=a" (ret) \ |
| 118 | :"c" (ptr),"d" (x) \ | 134 | :"c" (ptr),"d" (x) \ |
| 119 | :"r8") | 135 | :"r8") |
| 120 | 136 | ||
| 121 | #define put_user(x,ptr) \ | 137 | #define put_user(x, ptr) \ |
| 122 | __put_user_check((__typeof__(*(ptr)))(x),(ptr),sizeof(*(ptr))) | 138 | __put_user_check((__typeof__(*(ptr)))(x), (ptr), sizeof(*(ptr))) |
| 123 | 139 | ||
| 124 | #define __get_user(x,ptr) \ | 140 | #define __get_user(x, ptr) \ |
| 125 | __get_user_nocheck((x),(ptr),sizeof(*(ptr))) | 141 | __get_user_nocheck((x), (ptr), sizeof(*(ptr))) |
| 126 | #define __put_user(x,ptr) \ | 142 | #define __put_user(x, ptr) \ |
| 127 | __put_user_nocheck((__typeof__(*(ptr)))(x),(ptr),sizeof(*(ptr))) | 143 | __put_user_nocheck((__typeof__(*(ptr)))(x), (ptr), sizeof(*(ptr))) |
| 128 | 144 | ||
| 129 | #define __get_user_unaligned __get_user | 145 | #define __get_user_unaligned __get_user |
| 130 | #define __put_user_unaligned __put_user | 146 | #define __put_user_unaligned __put_user |
| 131 | 147 | ||
| 132 | #define __put_user_nocheck(x,ptr,size) \ | 148 | #define __put_user_nocheck(x, ptr, size) \ |
| 133 | ({ \ | 149 | ({ \ |
| 134 | int __pu_err; \ | 150 | int __pu_err; \ |
| 135 | __put_user_size((x),(ptr),(size),__pu_err); \ | 151 | __put_user_size((x), (ptr), (size), __pu_err); \ |
| 136 | __pu_err; \ | 152 | __pu_err; \ |
| 137 | }) | 153 | }) |
| 138 | 154 | ||
| 139 | 155 | ||
| 140 | #define __put_user_check(x,ptr,size) \ | 156 | #define __put_user_check(x, ptr, size) \ |
| 141 | ({ \ | 157 | ({ \ |
| 142 | int __pu_err; \ | 158 | int __pu_err; \ |
| 143 | typeof(*(ptr)) __user *__pu_addr = (ptr); \ | 159 | typeof(*(ptr)) __user *__pu_addr = (ptr); \ |
| 144 | switch (size) { \ | 160 | switch (size) { \ |
| 145 | case 1: __put_user_x(1,__pu_err,x,__pu_addr); break; \ | 161 | case 1: \ |
| 146 | case 2: __put_user_x(2,__pu_err,x,__pu_addr); break; \ | 162 | __put_user_x(1, __pu_err, x, __pu_addr); \ |
| 147 | case 4: __put_user_x(4,__pu_err,x,__pu_addr); break; \ | 163 | break; \ |
| 148 | case 8: __put_user_x(8,__pu_err,x,__pu_addr); break; \ | 164 | case 2: \ |
| 149 | default: __put_user_bad(); \ | 165 | __put_user_x(2, __pu_err, x, __pu_addr); \ |
| 150 | } \ | 166 | break; \ |
| 151 | __pu_err; \ | 167 | case 4: \ |
| 168 | __put_user_x(4, __pu_err, x, __pu_addr); \ | ||
| 169 | break; \ | ||
| 170 | case 8: \ | ||
| 171 | __put_user_x(8, __pu_err, x, __pu_addr); \ | ||
| 172 | break; \ | ||
| 173 | default: \ | ||
| 174 | __put_user_bad(); \ | ||
| 175 | } \ | ||
| 176 | __pu_err; \ | ||
| 152 | }) | 177 | }) |
| 153 | 178 | ||
| 154 | #define __put_user_size(x,ptr,size,retval) \ | 179 | #define __put_user_size(x, ptr, size, retval) \ |
| 155 | do { \ | 180 | do { \ |
| 156 | retval = 0; \ | 181 | retval = 0; \ |
| 157 | __chk_user_ptr(ptr); \ | 182 | __chk_user_ptr(ptr); \ |
| 158 | switch (size) { \ | 183 | switch (size) { \ |
| 159 | case 1: __put_user_asm(x,ptr,retval,"b","b","iq",-EFAULT); break;\ | 184 | case 1: \ |
| 160 | case 2: __put_user_asm(x,ptr,retval,"w","w","ir",-EFAULT); break;\ | 185 | __put_user_asm(x, ptr, retval, "b", "b", "iq", -EFAULT);\ |
| 161 | case 4: __put_user_asm(x,ptr,retval,"l","k","ir",-EFAULT); break;\ | 186 | break; \ |
| 162 | case 8: __put_user_asm(x,ptr,retval,"q","","Zr",-EFAULT); break;\ | 187 | case 2: \ |
| 163 | default: __put_user_bad(); \ | 188 | __put_user_asm(x, ptr, retval, "w", "w", "ir", -EFAULT);\ |
| 189 | break; \ | ||
| 190 | case 4: \ | ||
| 191 | __put_user_asm(x, ptr, retval, "l", "k", "ir", -EFAULT);\ | ||
| 192 | break; \ | ||
| 193 | case 8: \ | ||
| 194 | __put_user_asm(x, ptr, retval, "q", "", "Zr", -EFAULT); \ | ||
| 195 | break; \ | ||
| 196 | default: \ | ||
| 197 | __put_user_bad(); \ | ||
| 164 | } \ | 198 | } \ |
| 165 | } while (0) | 199 | } while (0) |
| 166 | 200 | ||
| @@ -174,23 +208,22 @@ struct __large_struct { unsigned long buf[100]; }; | |||
| 174 | * aliasing issues. | 208 | * aliasing issues. |
| 175 | */ | 209 | */ |
| 176 | #define __put_user_asm(x, addr, err, itype, rtype, ltype, errno) \ | 210 | #define __put_user_asm(x, addr, err, itype, rtype, ltype, errno) \ |
| 177 | asm volatile( \ | 211 | asm volatile("1: mov"itype" %"rtype"1,%2\n" \ |
| 178 | "1: mov"itype" %"rtype"1,%2\n" \ | 212 | "2:\n" \ |
| 179 | "2:\n" \ | 213 | ".section .fixup, \"ax\"\n" \ |
| 180 | ".section .fixup,\"ax\"\n" \ | 214 | "3: mov %3,%0\n" \ |
| 181 | "3: mov %3,%0\n" \ | 215 | " jmp 2b\n" \ |
| 182 | " jmp 2b\n" \ | 216 | ".previous\n" \ |
| 183 | ".previous\n" \ | 217 | _ASM_EXTABLE(1b, 3b) \ |
| 184 | _ASM_EXTABLE(1b,3b) \ | 218 | : "=r"(err) \ |
| 185 | : "=r"(err) \ | 219 | : ltype (x), "m" (__m(addr)), "i" (errno), "0" (err)) |
| 186 | : ltype (x), "m"(__m(addr)), "i"(errno), "0"(err)) | 220 | |
| 187 | 221 | ||
| 188 | 222 | #define __get_user_nocheck(x, ptr, size) \ | |
| 189 | #define __get_user_nocheck(x,ptr,size) \ | ||
| 190 | ({ \ | 223 | ({ \ |
| 191 | int __gu_err; \ | 224 | int __gu_err; \ |
| 192 | unsigned long __gu_val; \ | 225 | unsigned long __gu_val; \ |
| 193 | __get_user_size(__gu_val,(ptr),(size),__gu_err); \ | 226 | __get_user_size(__gu_val, (ptr), (size), __gu_err); \ |
| 194 | (x) = (__force typeof(*(ptr)))__gu_val; \ | 227 | (x) = (__force typeof(*(ptr)))__gu_val; \ |
| 195 | __gu_err; \ | 228 | __gu_err; \ |
| 196 | }) | 229 | }) |
| @@ -201,31 +234,39 @@ extern int __get_user_4(void); | |||
| 201 | extern int __get_user_8(void); | 234 | extern int __get_user_8(void); |
| 202 | extern int __get_user_bad(void); | 235 | extern int __get_user_bad(void); |
| 203 | 236 | ||
| 204 | #define __get_user_size(x,ptr,size,retval) \ | 237 | #define __get_user_size(x, ptr, size, retval) \ |
| 205 | do { \ | 238 | do { \ |
| 206 | retval = 0; \ | 239 | retval = 0; \ |
| 207 | __chk_user_ptr(ptr); \ | 240 | __chk_user_ptr(ptr); \ |
| 208 | switch (size) { \ | 241 | switch (size) { \ |
| 209 | case 1: __get_user_asm(x,ptr,retval,"b","b","=q",-EFAULT); break;\ | 242 | case 1: \ |
| 210 | case 2: __get_user_asm(x,ptr,retval,"w","w","=r",-EFAULT); break;\ | 243 | __get_user_asm(x, ptr, retval, "b", "b", "=q", -EFAULT);\ |
| 211 | case 4: __get_user_asm(x,ptr,retval,"l","k","=r",-EFAULT); break;\ | 244 | break; \ |
| 212 | case 8: __get_user_asm(x,ptr,retval,"q","","=r",-EFAULT); break;\ | 245 | case 2: \ |
| 213 | default: (x) = __get_user_bad(); \ | 246 | __get_user_asm(x, ptr, retval, "w", "w", "=r", -EFAULT);\ |
| 247 | break; \ | ||
| 248 | case 4: \ | ||
| 249 | __get_user_asm(x, ptr, retval, "l", "k", "=r", -EFAULT);\ | ||
| 250 | break; \ | ||
| 251 | case 8: \ | ||
| 252 | __get_user_asm(x, ptr, retval, "q", "", "=r", -EFAULT); \ | ||
| 253 | break; \ | ||
| 254 | default: \ | ||
| 255 | (x) = __get_user_bad(); \ | ||
| 214 | } \ | 256 | } \ |
| 215 | } while (0) | 257 | } while (0) |
| 216 | 258 | ||
| 217 | #define __get_user_asm(x, addr, err, itype, rtype, ltype, errno) \ | 259 | #define __get_user_asm(x, addr, err, itype, rtype, ltype, errno) \ |
| 218 | asm volatile( \ | 260 | asm volatile("1: mov"itype" %2,%"rtype"1\n" \ |
| 219 | "1: mov"itype" %2,%"rtype"1\n" \ | 261 | "2:\n" \ |
| 220 | "2:\n" \ | 262 | ".section .fixup, \"ax\"\n" \ |
| 221 | ".section .fixup,\"ax\"\n" \ | 263 | "3: mov %3,%0\n" \ |
| 222 | "3: mov %3,%0\n" \ | 264 | " xor"itype" %"rtype"1,%"rtype"1\n" \ |
| 223 | " xor"itype" %"rtype"1,%"rtype"1\n" \ | 265 | " jmp 2b\n" \ |
| 224 | " jmp 2b\n" \ | 266 | ".previous\n" \ |
| 225 | ".previous\n" \ | 267 | _ASM_EXTABLE(1b, 3b) \ |
| 226 | _ASM_EXTABLE(1b,3b) \ | 268 | : "=r" (err), ltype (x) \ |
| 227 | : "=r"(err), ltype (x) \ | 269 | : "m" (__m(addr)), "i"(errno), "0"(err)) |
| 228 | : "m"(__m(addr)), "i"(errno), "0"(err)) | ||
| 229 | 270 | ||
| 230 | /* | 271 | /* |
| 231 | * Copy To/From Userspace | 272 | * Copy To/From Userspace |
| @@ -244,110 +285,142 @@ copy_in_user(void __user *to, const void __user *from, unsigned len); | |||
| 244 | 285 | ||
| 245 | static __always_inline __must_check | 286 | static __always_inline __must_check |
| 246 | int __copy_from_user(void *dst, const void __user *src, unsigned size) | 287 | int __copy_from_user(void *dst, const void __user *src, unsigned size) |
| 247 | { | 288 | { |
| 248 | int ret = 0; | 289 | int ret = 0; |
| 249 | if (!__builtin_constant_p(size)) | 290 | if (!__builtin_constant_p(size)) |
| 250 | return copy_user_generic(dst,(__force void *)src,size); | 291 | return copy_user_generic(dst, (__force void *)src, size); |
| 251 | switch (size) { | 292 | switch (size) { |
| 252 | case 1:__get_user_asm(*(u8*)dst,(u8 __user *)src,ret,"b","b","=q",1); | 293 | case 1:__get_user_asm(*(u8 *)dst, (u8 __user *)src, |
| 294 | ret, "b", "b", "=q", 1); | ||
| 253 | return ret; | 295 | return ret; |
| 254 | case 2:__get_user_asm(*(u16*)dst,(u16 __user *)src,ret,"w","w","=r",2); | 296 | case 2:__get_user_asm(*(u16 *)dst, (u16 __user *)src, |
| 297 | ret, "w", "w", "=r", 2); | ||
| 255 | return ret; | 298 | return ret; |
| 256 | case 4:__get_user_asm(*(u32*)dst,(u32 __user *)src,ret,"l","k","=r",4); | 299 | case 4:__get_user_asm(*(u32 *)dst, (u32 __user *)src, |
| 300 | ret, "l", "k", "=r", 4); | ||
| 301 | return ret; | ||
| 302 | case 8:__get_user_asm(*(u64 *)dst, (u64 __user *)src, | ||
| 303 | ret, "q", "", "=r", 8); | ||
| 257 | return ret; | 304 | return ret; |
| 258 | case 8:__get_user_asm(*(u64*)dst,(u64 __user *)src,ret,"q","","=r",8); | ||
| 259 | return ret; | ||
| 260 | case 10: | 305 | case 10: |
| 261 | __get_user_asm(*(u64*)dst,(u64 __user *)src,ret,"q","","=r",16); | 306 | __get_user_asm(*(u64 *)dst, (u64 __user *)src, |
| 262 | if (unlikely(ret)) return ret; | 307 | ret, "q", "", "=r", 16); |
| 263 | __get_user_asm(*(u16*)(8+(char*)dst),(u16 __user *)(8+(char __user *)src),ret,"w","w","=r",2); | 308 | if (unlikely(ret)) |
| 264 | return ret; | 309 | return ret; |
| 310 | __get_user_asm(*(u16 *)(8 + (char *)dst), | ||
| 311 | (u16 __user *)(8 + (char __user *)src), | ||
| 312 | ret, "w", "w", "=r", 2); | ||
| 313 | return ret; | ||
| 265 | case 16: | 314 | case 16: |
| 266 | __get_user_asm(*(u64*)dst,(u64 __user *)src,ret,"q","","=r",16); | 315 | __get_user_asm(*(u64 *)dst, (u64 __user *)src, |
| 267 | if (unlikely(ret)) return ret; | 316 | ret, "q", "", "=r", 16); |
| 268 | __get_user_asm(*(u64*)(8+(char*)dst),(u64 __user *)(8+(char __user *)src),ret,"q","","=r",8); | 317 | if (unlikely(ret)) |
| 269 | return ret; | 318 | return ret; |
| 319 | __get_user_asm(*(u64 *)(8 + (char *)dst), | ||
| 320 | (u64 __user *)(8 + (char __user *)src), | ||
| 321 | ret, "q", "", "=r", 8); | ||
| 322 | return ret; | ||
| 270 | default: | 323 | default: |
| 271 | return copy_user_generic(dst,(__force void *)src,size); | 324 | return copy_user_generic(dst, (__force void *)src, size); |
| 272 | } | 325 | } |
| 273 | } | 326 | } |
| 274 | 327 | ||
| 275 | static __always_inline __must_check | 328 | static __always_inline __must_check |
| 276 | int __copy_to_user(void __user *dst, const void *src, unsigned size) | 329 | int __copy_to_user(void __user *dst, const void *src, unsigned size) |
| 277 | { | 330 | { |
| 278 | int ret = 0; | 331 | int ret = 0; |
| 279 | if (!__builtin_constant_p(size)) | 332 | if (!__builtin_constant_p(size)) |
| 280 | return copy_user_generic((__force void *)dst,src,size); | 333 | return copy_user_generic((__force void *)dst, src, size); |
| 281 | switch (size) { | 334 | switch (size) { |
| 282 | case 1:__put_user_asm(*(u8*)src,(u8 __user *)dst,ret,"b","b","iq",1); | 335 | case 1:__put_user_asm(*(u8 *)src, (u8 __user *)dst, |
| 336 | ret, "b", "b", "iq", 1); | ||
| 283 | return ret; | 337 | return ret; |
| 284 | case 2:__put_user_asm(*(u16*)src,(u16 __user *)dst,ret,"w","w","ir",2); | 338 | case 2:__put_user_asm(*(u16 *)src, (u16 __user *)dst, |
| 339 | ret, "w", "w", "ir", 2); | ||
| 285 | return ret; | 340 | return ret; |
| 286 | case 4:__put_user_asm(*(u32*)src,(u32 __user *)dst,ret,"l","k","ir",4); | 341 | case 4:__put_user_asm(*(u32 *)src, (u32 __user *)dst, |
| 342 | ret, "l", "k", "ir", 4); | ||
| 343 | return ret; | ||
| 344 | case 8:__put_user_asm(*(u64 *)src, (u64 __user *)dst, | ||
| 345 | ret, "q", "", "ir", 8); | ||
| 287 | return ret; | 346 | return ret; |
| 288 | case 8:__put_user_asm(*(u64*)src,(u64 __user *)dst,ret,"q","","ir",8); | ||
| 289 | return ret; | ||
| 290 | case 10: | 347 | case 10: |
| 291 | __put_user_asm(*(u64*)src,(u64 __user *)dst,ret,"q","","ir",10); | 348 | __put_user_asm(*(u64 *)src, (u64 __user *)dst, |
| 292 | if (unlikely(ret)) return ret; | 349 | ret, "q", "", "ir", 10); |
| 350 | if (unlikely(ret)) | ||
| 351 | return ret; | ||
| 293 | asm("":::"memory"); | 352 | asm("":::"memory"); |
| 294 | __put_user_asm(4[(u16*)src],4+(u16 __user *)dst,ret,"w","w","ir",2); | 353 | __put_user_asm(4[(u16 *)src], 4 + (u16 __user *)dst, |
| 295 | return ret; | 354 | ret, "w", "w", "ir", 2); |
| 355 | return ret; | ||
| 296 | case 16: | 356 | case 16: |
| 297 | __put_user_asm(*(u64*)src,(u64 __user *)dst,ret,"q","","ir",16); | 357 | __put_user_asm(*(u64 *)src, (u64 __user *)dst, |
| 298 | if (unlikely(ret)) return ret; | 358 | ret, "q", "", "ir", 16); |
| 359 | if (unlikely(ret)) | ||
| 360 | return ret; | ||
| 299 | asm("":::"memory"); | 361 | asm("":::"memory"); |
| 300 | __put_user_asm(1[(u64*)src],1+(u64 __user *)dst,ret,"q","","ir",8); | 362 | __put_user_asm(1[(u64 *)src], 1 + (u64 __user *)dst, |
| 301 | return ret; | 363 | ret, "q", "", "ir", 8); |
| 364 | return ret; | ||
| 302 | default: | 365 | default: |
| 303 | return copy_user_generic((__force void *)dst,src,size); | 366 | return copy_user_generic((__force void *)dst, src, size); |
| 304 | } | 367 | } |
| 305 | } | 368 | } |
| 306 | 369 | ||
| 307 | static __always_inline __must_check | 370 | static __always_inline __must_check |
| 308 | int __copy_in_user(void __user *dst, const void __user *src, unsigned size) | 371 | int __copy_in_user(void __user *dst, const void __user *src, unsigned size) |
| 309 | { | 372 | { |
| 310 | int ret = 0; | 373 | int ret = 0; |
| 311 | if (!__builtin_constant_p(size)) | 374 | if (!__builtin_constant_p(size)) |
| 312 | return copy_user_generic((__force void *)dst,(__force void *)src,size); | 375 | return copy_user_generic((__force void *)dst, |
| 313 | switch (size) { | 376 | (__force void *)src, size); |
| 314 | case 1: { | 377 | switch (size) { |
| 378 | case 1: { | ||
| 315 | u8 tmp; | 379 | u8 tmp; |
| 316 | __get_user_asm(tmp,(u8 __user *)src,ret,"b","b","=q",1); | 380 | __get_user_asm(tmp, (u8 __user *)src, |
| 381 | ret, "b", "b", "=q", 1); | ||
| 317 | if (likely(!ret)) | 382 | if (likely(!ret)) |
| 318 | __put_user_asm(tmp,(u8 __user *)dst,ret,"b","b","iq",1); | 383 | __put_user_asm(tmp, (u8 __user *)dst, |
| 384 | ret, "b", "b", "iq", 1); | ||
| 319 | return ret; | 385 | return ret; |
| 320 | } | 386 | } |
| 321 | case 2: { | 387 | case 2: { |
| 322 | u16 tmp; | 388 | u16 tmp; |
| 323 | __get_user_asm(tmp,(u16 __user *)src,ret,"w","w","=r",2); | 389 | __get_user_asm(tmp, (u16 __user *)src, |
| 390 | ret, "w", "w", "=r", 2); | ||
| 324 | if (likely(!ret)) | 391 | if (likely(!ret)) |
| 325 | __put_user_asm(tmp,(u16 __user *)dst,ret,"w","w","ir",2); | 392 | __put_user_asm(tmp, (u16 __user *)dst, |
| 393 | ret, "w", "w", "ir", 2); | ||
| 326 | return ret; | 394 | return ret; |
| 327 | } | 395 | } |
| 328 | 396 | ||
| 329 | case 4: { | 397 | case 4: { |
| 330 | u32 tmp; | 398 | u32 tmp; |
| 331 | __get_user_asm(tmp,(u32 __user *)src,ret,"l","k","=r",4); | 399 | __get_user_asm(tmp, (u32 __user *)src, |
| 400 | ret, "l", "k", "=r", 4); | ||
| 332 | if (likely(!ret)) | 401 | if (likely(!ret)) |
| 333 | __put_user_asm(tmp,(u32 __user *)dst,ret,"l","k","ir",4); | 402 | __put_user_asm(tmp, (u32 __user *)dst, |
| 403 | ret, "l", "k", "ir", 4); | ||
| 334 | return ret; | 404 | return ret; |
| 335 | } | 405 | } |
| 336 | case 8: { | 406 | case 8: { |
| 337 | u64 tmp; | 407 | u64 tmp; |
| 338 | __get_user_asm(tmp,(u64 __user *)src,ret,"q","","=r",8); | 408 | __get_user_asm(tmp, (u64 __user *)src, |
| 409 | ret, "q", "", "=r", 8); | ||
| 339 | if (likely(!ret)) | 410 | if (likely(!ret)) |
| 340 | __put_user_asm(tmp,(u64 __user *)dst,ret,"q","","ir",8); | 411 | __put_user_asm(tmp, (u64 __user *)dst, |
| 412 | ret, "q", "", "ir", 8); | ||
| 341 | return ret; | 413 | return ret; |
| 342 | } | 414 | } |
| 343 | default: | 415 | default: |
| 344 | return copy_user_generic((__force void *)dst,(__force void *)src,size); | 416 | return copy_user_generic((__force void *)dst, |
| 417 | (__force void *)src, size); | ||
| 345 | } | 418 | } |
| 346 | } | 419 | } |
| 347 | 420 | ||
| 348 | __must_check long | 421 | __must_check long |
| 349 | strncpy_from_user(char *dst, const char __user *src, long count); | 422 | strncpy_from_user(char *dst, const char __user *src, long count); |
| 350 | __must_check long | 423 | __must_check long |
| 351 | __strncpy_from_user(char *dst, const char __user *src, long count); | 424 | __strncpy_from_user(char *dst, const char __user *src, long count); |
| 352 | __must_check long strnlen_user(const char __user *str, long n); | 425 | __must_check long strnlen_user(const char __user *str, long n); |
| 353 | __must_check long __strnlen_user(const char __user *str, long n); | 426 | __must_check long __strnlen_user(const char __user *str, long n); |
| @@ -355,7 +428,8 @@ __must_check long strlen_user(const char __user *str); | |||
| 355 | __must_check unsigned long clear_user(void __user *mem, unsigned long len); | 428 | __must_check unsigned long clear_user(void __user *mem, unsigned long len); |
| 356 | __must_check unsigned long __clear_user(void __user *mem, unsigned long len); | 429 | __must_check unsigned long __clear_user(void __user *mem, unsigned long len); |
| 357 | 430 | ||
| 358 | __must_check long __copy_from_user_inatomic(void *dst, const void __user *src, unsigned size); | 431 | __must_check long __copy_from_user_inatomic(void *dst, const void __user *src, |
| 432 | unsigned size); | ||
| 359 | 433 | ||
| 360 | static __must_check __always_inline int | 434 | static __must_check __always_inline int |
| 361 | __copy_to_user_inatomic(void __user *dst, const void *src, unsigned size) | 435 | __copy_to_user_inatomic(void __user *dst, const void *src, unsigned size) |
| @@ -364,15 +438,19 @@ __copy_to_user_inatomic(void __user *dst, const void *src, unsigned size) | |||
| 364 | } | 438 | } |
| 365 | 439 | ||
| 366 | #define ARCH_HAS_NOCACHE_UACCESS 1 | 440 | #define ARCH_HAS_NOCACHE_UACCESS 1 |
| 367 | extern long __copy_user_nocache(void *dst, const void __user *src, unsigned size, int zerorest); | 441 | extern long __copy_user_nocache(void *dst, const void __user *src, |
| 442 | unsigned size, int zerorest); | ||
| 368 | 443 | ||
| 369 | static inline int __copy_from_user_nocache(void *dst, const void __user *src, unsigned size) | 444 | static inline int __copy_from_user_nocache(void *dst, const void __user *src, |
| 445 | unsigned size) | ||
| 370 | { | 446 | { |
| 371 | might_sleep(); | 447 | might_sleep(); |
| 372 | return __copy_user_nocache(dst, src, size, 1); | 448 | return __copy_user_nocache(dst, src, size, 1); |
| 373 | } | 449 | } |
| 374 | 450 | ||
| 375 | static inline int __copy_from_user_inatomic_nocache(void *dst, const void __user *src, unsigned size) | 451 | static inline int __copy_from_user_inatomic_nocache(void *dst, |
| 452 | const void __user *src, | ||
| 453 | unsigned size) | ||
| 376 | { | 454 | { |
| 377 | return __copy_user_nocache(dst, src, size, 0); | 455 | return __copy_user_nocache(dst, src, size, 0); |
| 378 | } | 456 | } |
diff --git a/include/asm-x86/unaligned.h b/include/asm-x86/unaligned.h index 913598d4f761..d270ffe72759 100644 --- a/include/asm-x86/unaligned.h +++ b/include/asm-x86/unaligned.h | |||
| @@ -32,6 +32,6 @@ | |||
| 32 | * | 32 | * |
| 33 | * Note that unaligned accesses can be very expensive on some architectures. | 33 | * Note that unaligned accesses can be very expensive on some architectures. |
| 34 | */ | 34 | */ |
| 35 | #define put_unaligned(val, ptr) ((void)( *(ptr) = (val) )) | 35 | #define put_unaligned(val, ptr) ((void)(*(ptr) = (val))) |
| 36 | 36 | ||
| 37 | #endif /* _ASM_X86_UNALIGNED_H */ | 37 | #endif /* _ASM_X86_UNALIGNED_H */ |
diff --git a/include/asm-x86/unistd.h b/include/asm-x86/unistd.h index 2a58ed3e51d8..effc7ad8e12f 100644 --- a/include/asm-x86/unistd.h +++ b/include/asm-x86/unistd.h | |||
| @@ -1,11 +1,5 @@ | |||
| 1 | #ifdef __KERNEL__ | 1 | #ifdef __KERNEL__ |
| 2 | # ifdef CONFIG_X86_32 | 2 | # if defined(CONFIG_X86_32) || defined(__i386__) |
| 3 | # include "unistd_32.h" | ||
| 4 | # else | ||
| 5 | # include "unistd_64.h" | ||
| 6 | # endif | ||
| 7 | #else | ||
| 8 | # ifdef __i386__ | ||
| 9 | # include "unistd_32.h" | 3 | # include "unistd_32.h" |
| 10 | # else | 4 | # else |
| 11 | # include "unistd_64.h" | 5 | # include "unistd_64.h" |
diff --git a/include/asm-x86/unistd_32.h b/include/asm-x86/unistd_32.h index 984123a68f7c..8317d94771d3 100644 --- a/include/asm-x86/unistd_32.h +++ b/include/asm-x86/unistd_32.h | |||
| @@ -81,7 +81,7 @@ | |||
| 81 | #define __NR_sigpending 73 | 81 | #define __NR_sigpending 73 |
| 82 | #define __NR_sethostname 74 | 82 | #define __NR_sethostname 74 |
| 83 | #define __NR_setrlimit 75 | 83 | #define __NR_setrlimit 75 |
| 84 | #define __NR_getrlimit 76 /* Back compatible 2Gig limited rlimit */ | 84 | #define __NR_getrlimit 76 /* Back compatible 2Gig limited rlimit */ |
| 85 | #define __NR_getrusage 77 | 85 | #define __NR_getrusage 77 |
| 86 | #define __NR_gettimeofday 78 | 86 | #define __NR_gettimeofday 78 |
| 87 | #define __NR_settimeofday 79 | 87 | #define __NR_settimeofday 79 |
diff --git a/include/asm-x86/unistd_64.h b/include/asm-x86/unistd_64.h index 3883ceb54ef5..fe26e36d0f51 100644 --- a/include/asm-x86/unistd_64.h +++ b/include/asm-x86/unistd_64.h | |||
| @@ -2,7 +2,7 @@ | |||
| 2 | #define _ASM_X86_64_UNISTD_H_ | 2 | #define _ASM_X86_64_UNISTD_H_ |
| 3 | 3 | ||
| 4 | #ifndef __SYSCALL | 4 | #ifndef __SYSCALL |
| 5 | #define __SYSCALL(a,b) | 5 | #define __SYSCALL(a, b) |
| 6 | #endif | 6 | #endif |
| 7 | 7 | ||
| 8 | /* | 8 | /* |
diff --git a/include/asm-x86/user32.h b/include/asm-x86/user32.h index f769872debea..a3d910047879 100644 --- a/include/asm-x86/user32.h +++ b/include/asm-x86/user32.h | |||
| @@ -1,7 +1,8 @@ | |||
| 1 | #ifndef USER32_H | 1 | #ifndef USER32_H |
| 2 | #define USER32_H 1 | 2 | #define USER32_H 1 |
| 3 | 3 | ||
| 4 | /* IA32 compatible user structures for ptrace. These should be used for 32bit coredumps too. */ | 4 | /* IA32 compatible user structures for ptrace. |
| 5 | * These should be used for 32bit coredumps too. */ | ||
| 5 | 6 | ||
| 6 | struct user_i387_ia32_struct { | 7 | struct user_i387_ia32_struct { |
| 7 | u32 cwd; | 8 | u32 cwd; |
| @@ -42,9 +43,9 @@ struct user_regs_struct32 { | |||
| 42 | }; | 43 | }; |
| 43 | 44 | ||
| 44 | struct user32 { | 45 | struct user32 { |
| 45 | struct user_regs_struct32 regs; /* Where the registers are actually stored */ | 46 | struct user_regs_struct32 regs; /* Where the registers are actually stored */ |
| 46 | int u_fpvalid; /* True if math co-processor being used. */ | 47 | int u_fpvalid; /* True if math co-processor being used. */ |
| 47 | /* for this mess. Not yet used. */ | 48 | /* for this mess. Not yet used. */ |
| 48 | struct user_i387_ia32_struct i387; /* Math Co-processor registers. */ | 49 | struct user_i387_ia32_struct i387; /* Math Co-processor registers. */ |
| 49 | /* The rest of this junk is to help gdb figure out what goes where */ | 50 | /* The rest of this junk is to help gdb figure out what goes where */ |
| 50 | __u32 u_tsize; /* Text segment size (pages). */ | 51 | __u32 u_tsize; /* Text segment size (pages). */ |
diff --git a/include/asm-x86/user_32.h b/include/asm-x86/user_32.h index 6157da6f882c..d6e51edc259d 100644 --- a/include/asm-x86/user_32.h +++ b/include/asm-x86/user_32.h | |||
| @@ -100,10 +100,10 @@ struct user_regs_struct { | |||
| 100 | struct user{ | 100 | struct user{ |
| 101 | /* We start with the registers, to mimic the way that "memory" is returned | 101 | /* We start with the registers, to mimic the way that "memory" is returned |
| 102 | from the ptrace(3,...) function. */ | 102 | from the ptrace(3,...) function. */ |
| 103 | struct user_regs_struct regs; /* Where the registers are actually stored */ | 103 | struct user_regs_struct regs; /* Where the registers are actually stored */ |
| 104 | /* ptrace does not yet supply these. Someday.... */ | 104 | /* ptrace does not yet supply these. Someday.... */ |
| 105 | int u_fpvalid; /* True if math co-processor being used. */ | 105 | int u_fpvalid; /* True if math co-processor being used. */ |
| 106 | /* for this mess. Not yet used. */ | 106 | /* for this mess. Not yet used. */ |
| 107 | struct user_i387_struct i387; /* Math Co-processor registers. */ | 107 | struct user_i387_struct i387; /* Math Co-processor registers. */ |
| 108 | /* The rest of this junk is to help gdb figure out what goes where */ | 108 | /* The rest of this junk is to help gdb figure out what goes where */ |
| 109 | unsigned long int u_tsize; /* Text segment size (pages). */ | 109 | unsigned long int u_tsize; /* Text segment size (pages). */ |
| @@ -118,7 +118,7 @@ struct user{ | |||
| 118 | int reserved; /* No longer used */ | 118 | int reserved; /* No longer used */ |
| 119 | unsigned long u_ar0; /* Used by gdb to help find the values for */ | 119 | unsigned long u_ar0; /* Used by gdb to help find the values for */ |
| 120 | /* the registers. */ | 120 | /* the registers. */ |
| 121 | struct user_i387_struct* u_fpstate; /* Math Co-processor pointer. */ | 121 | struct user_i387_struct *u_fpstate; /* Math Co-processor pointer. */ |
| 122 | unsigned long magic; /* To uniquely identify a core file */ | 122 | unsigned long magic; /* To uniquely identify a core file */ |
| 123 | char u_comm[32]; /* User command that was responsible */ | 123 | char u_comm[32]; /* User command that was responsible */ |
| 124 | int u_debugreg[8]; | 124 | int u_debugreg[8]; |
diff --git a/include/asm-x86/user_64.h b/include/asm-x86/user_64.h index 963616455609..6037b634c77f 100644 --- a/include/asm-x86/user_64.h +++ b/include/asm-x86/user_64.h | |||
| @@ -45,12 +45,13 @@ | |||
| 45 | */ | 45 | */ |
| 46 | 46 | ||
| 47 | /* This matches the 64bit FXSAVE format as defined by AMD. It is the same | 47 | /* This matches the 64bit FXSAVE format as defined by AMD. It is the same |
| 48 | as the 32bit format defined by Intel, except that the selector:offset pairs for | 48 | as the 32bit format defined by Intel, except that the selector:offset pairs |
| 49 | data and eip are replaced with flat 64bit pointers. */ | 49 | for data and eip are replaced with flat 64bit pointers. */ |
| 50 | struct user_i387_struct { | 50 | struct user_i387_struct { |
| 51 | unsigned short cwd; | 51 | unsigned short cwd; |
| 52 | unsigned short swd; | 52 | unsigned short swd; |
| 53 | unsigned short twd; /* Note this is not the same as the 32bit/x87/FSAVE twd */ | 53 | unsigned short twd; /* Note this is not the same as |
| 54 | the 32bit/x87/FSAVE twd */ | ||
| 54 | unsigned short fop; | 55 | unsigned short fop; |
| 55 | __u64 rip; | 56 | __u64 rip; |
| 56 | __u64 rdp; | 57 | __u64 rdp; |
| @@ -97,13 +98,14 @@ struct user_regs_struct { | |||
| 97 | /* When the kernel dumps core, it starts by dumping the user struct - | 98 | /* When the kernel dumps core, it starts by dumping the user struct - |
| 98 | this will be used by gdb to figure out where the data and stack segments | 99 | this will be used by gdb to figure out where the data and stack segments |
| 99 | are within the file, and what virtual addresses to use. */ | 100 | are within the file, and what virtual addresses to use. */ |
| 100 | struct user{ | 101 | |
| 102 | struct user { | ||
| 101 | /* We start with the registers, to mimic the way that "memory" is returned | 103 | /* We start with the registers, to mimic the way that "memory" is returned |
| 102 | from the ptrace(3,...) function. */ | 104 | from the ptrace(3,...) function. */ |
| 103 | struct user_regs_struct regs; /* Where the registers are actually stored */ | 105 | struct user_regs_struct regs; /* Where the registers are actually stored */ |
| 104 | /* ptrace does not yet supply these. Someday.... */ | 106 | /* ptrace does not yet supply these. Someday.... */ |
| 105 | int u_fpvalid; /* True if math co-processor being used. */ | 107 | int u_fpvalid; /* True if math co-processor being used. */ |
| 106 | /* for this mess. Not yet used. */ | 108 | /* for this mess. Not yet used. */ |
| 107 | int pad0; | 109 | int pad0; |
| 108 | struct user_i387_struct i387; /* Math Co-processor registers. */ | 110 | struct user_i387_struct i387; /* Math Co-processor registers. */ |
| 109 | /* The rest of this junk is to help gdb figure out what goes where */ | 111 | /* The rest of this junk is to help gdb figure out what goes where */ |
| @@ -120,7 +122,7 @@ struct user{ | |||
| 120 | int pad1; | 122 | int pad1; |
| 121 | unsigned long u_ar0; /* Used by gdb to help find the values for */ | 123 | unsigned long u_ar0; /* Used by gdb to help find the values for */ |
| 122 | /* the registers. */ | 124 | /* the registers. */ |
| 123 | struct user_i387_struct* u_fpstate; /* Math Co-processor pointer. */ | 125 | struct user_i387_struct *u_fpstate; /* Math Co-processor pointer. */ |
| 124 | unsigned long magic; /* To uniquely identify a core file */ | 126 | unsigned long magic; /* To uniquely identify a core file */ |
| 125 | char u_comm[32]; /* User command that was responsible */ | 127 | char u_comm[32]; /* User command that was responsible */ |
| 126 | unsigned long u_debugreg[8]; | 128 | unsigned long u_debugreg[8]; |
diff --git a/include/asm-x86/uv/uv_hub.h b/include/asm-x86/uv/uv_hub.h new file mode 100644 index 000000000000..26b9240d1e23 --- /dev/null +++ b/include/asm-x86/uv/uv_hub.h | |||
| @@ -0,0 +1,284 @@ | |||
| 1 | /* | ||
| 2 | * This file is subject to the terms and conditions of the GNU General Public | ||
| 3 | * License. See the file "COPYING" in the main directory of this archive | ||
| 4 | * for more details. | ||
| 5 | * | ||
| 6 | * SGI UV architectural definitions | ||
| 7 | * | ||
| 8 | * Copyright (C) 2007 Silicon Graphics, Inc. All rights reserved. | ||
| 9 | */ | ||
| 10 | |||
| 11 | #ifndef __ASM_X86_UV_HUB_H__ | ||
| 12 | #define __ASM_X86_UV_HUB_H__ | ||
| 13 | |||
| 14 | #include <linux/numa.h> | ||
| 15 | #include <linux/percpu.h> | ||
| 16 | #include <asm/types.h> | ||
| 17 | #include <asm/percpu.h> | ||
| 18 | |||
| 19 | |||
| 20 | /* | ||
| 21 | * Addressing Terminology | ||
| 22 | * | ||
| 23 | * NASID - network ID of a router, Mbrick or Cbrick. Nasid values of | ||
| 24 | * routers always have low bit of 1, C/MBricks have low bit | ||
| 25 | * equal to 0. Most addressing macros that target UV hub chips | ||
| 26 | * right shift the NASID by 1 to exclude the always-zero bit. | ||
| 27 | * | ||
| 28 | * SNASID - NASID right shifted by 1 bit. | ||
| 29 | * | ||
| 30 | * | ||
| 31 | * Memory/UV-HUB Processor Socket Address Format: | ||
| 32 | * +--------+---------------+---------------------+ | ||
| 33 | * |00..0000| SNASID | NodeOffset | | ||
| 34 | * +--------+---------------+---------------------+ | ||
| 35 | * <--- N bits --->|<--------M bits -----> | ||
| 36 | * | ||
| 37 | * M number of node offset bits (35 .. 40) | ||
| 38 | * N number of SNASID bits (0 .. 10) | ||
| 39 | * | ||
| 40 | * Note: M + N cannot currently exceed 44 (x86_64) or 46 (IA64). | ||
| 41 | * The actual values are configuration dependent and are set at | ||
| 42 | * boot time | ||
| 43 | * | ||
| 44 | * APICID format | ||
| 45 | * NOTE!!!!!! This is the current format of the APICID. However, code | ||
| 46 | * should assume that this will change in the future. Use functions | ||
| 47 | * in this file for all APICID bit manipulations and conversion. | ||
| 48 | * | ||
| 49 | * 1111110000000000 | ||
| 50 | * 5432109876543210 | ||
| 51 | * nnnnnnnnnnlc0cch | ||
| 52 | * sssssssssss | ||
| 53 | * | ||
| 54 | * n = snasid bits | ||
| 55 | * l = socket number on board | ||
| 56 | * c = core | ||
| 57 | * h = hyperthread | ||
| 58 | * s = bits that are in the socket CSR | ||
| 59 | * | ||
| 60 | * Note: Processor only supports 12 bits in the APICID register. The ACPI | ||
| 61 | * tables hold all 16 bits. Software needs to be aware of this. | ||
| 62 | * | ||
| 63 | * Unless otherwise specified, all references to APICID refer to | ||
| 64 | * the FULL value contained in ACPI tables, not the subset in the | ||
| 65 | * processor APICID register. | ||
| 66 | */ | ||
| 67 | |||
| 68 | |||
| 69 | /* | ||
| 70 | * Maximum number of bricks in all partitions and in all coherency domains. | ||
| 71 | * This is the total number of bricks accessible in the numalink fabric. It | ||
| 72 | * includes all C & M bricks. Routers are NOT included. | ||
| 73 | * | ||
| 74 | * This value is also the value of the maximum number of non-router NASIDs | ||
| 75 | * in the numalink fabric. | ||
| 76 | * | ||
| 77 | * NOTE: a brick may be 1 or 2 OS nodes. Don't get these confused. | ||
| 78 | */ | ||
| 79 | #define UV_MAX_NUMALINK_BLADES 16384 | ||
| 80 | |||
| 81 | /* | ||
| 82 | * Maximum number of C/Mbricks within a software SSI (hardware may support | ||
| 83 | * more). | ||
| 84 | */ | ||
| 85 | #define UV_MAX_SSI_BLADES 256 | ||
| 86 | |||
| 87 | /* | ||
| 88 | * The largest possible NASID of a C or M brick (+ 2) | ||
| 89 | */ | ||
| 90 | #define UV_MAX_NASID_VALUE (UV_MAX_NUMALINK_NODES * 2) | ||
| 91 | |||
| 92 | /* | ||
| 93 | * The following defines attributes of the HUB chip. These attributes are | ||
| 94 | * frequently referenced and are kept in the per-cpu data areas of each cpu. | ||
| 95 | * They are kept together in a struct to minimize cache misses. | ||
| 96 | */ | ||
| 97 | struct uv_hub_info_s { | ||
| 98 | unsigned long global_mmr_base; | ||
| 99 | unsigned short local_nasid; | ||
| 100 | unsigned short gnode_upper; | ||
| 101 | unsigned short coherency_domain_number; | ||
| 102 | unsigned short numa_blade_id; | ||
| 103 | unsigned char blade_processor_id; | ||
| 104 | unsigned char m_val; | ||
| 105 | unsigned char n_val; | ||
| 106 | }; | ||
| 107 | DECLARE_PER_CPU(struct uv_hub_info_s, __uv_hub_info); | ||
| 108 | #define uv_hub_info (&__get_cpu_var(__uv_hub_info)) | ||
| 109 | #define uv_cpu_hub_info(cpu) (&per_cpu(__uv_hub_info, cpu)) | ||
| 110 | |||
| 111 | /* | ||
| 112 | * Local & Global MMR space macros. | ||
| 113 | * Note: macros are intended to be used ONLY by inline functions | ||
| 114 | * in this file - not by other kernel code. | ||
| 115 | */ | ||
| 116 | #define UV_SNASID(n) ((n) >> 1) | ||
| 117 | #define UV_NASID(n) ((n) << 1) | ||
| 118 | |||
| 119 | #define UV_LOCAL_MMR_BASE 0xf4000000UL | ||
| 120 | #define UV_GLOBAL_MMR32_BASE 0xf8000000UL | ||
| 121 | #define UV_GLOBAL_MMR64_BASE (uv_hub_info->global_mmr_base) | ||
| 122 | |||
| 123 | #define UV_GLOBAL_MMR32_SNASID_MASK 0x3ff | ||
| 124 | #define UV_GLOBAL_MMR32_SNASID_SHIFT 15 | ||
| 125 | #define UV_GLOBAL_MMR64_SNASID_SHIFT 26 | ||
| 126 | |||
| 127 | #define UV_GLOBAL_MMR32_NASID_BITS(n) \ | ||
| 128 | (((UV_SNASID(n) & UV_GLOBAL_MMR32_SNASID_MASK)) << \ | ||
| 129 | (UV_GLOBAL_MMR32_SNASID_SHIFT)) | ||
| 130 | |||
| 131 | #define UV_GLOBAL_MMR64_NASID_BITS(n) \ | ||
| 132 | ((unsigned long)UV_SNASID(n) << UV_GLOBAL_MMR64_SNASID_SHIFT) | ||
| 133 | |||
| 134 | #define UV_APIC_NASID_SHIFT 6 | ||
| 135 | |||
| 136 | /* | ||
| 137 | * Extract a NASID from an APICID (full apicid, not processor subset) | ||
| 138 | */ | ||
| 139 | static inline int uv_apicid_to_nasid(int apicid) | ||
| 140 | { | ||
| 141 | return (UV_NASID(apicid >> UV_APIC_NASID_SHIFT)); | ||
| 142 | } | ||
| 143 | |||
| 144 | /* | ||
| 145 | * Access global MMRs using the low memory MMR32 space. This region supports | ||
| 146 | * faster MMR access but not all MMRs are accessible in this space. | ||
| 147 | */ | ||
| 148 | static inline unsigned long *uv_global_mmr32_address(int nasid, | ||
| 149 | unsigned long offset) | ||
| 150 | { | ||
| 151 | return __va(UV_GLOBAL_MMR32_BASE | | ||
| 152 | UV_GLOBAL_MMR32_NASID_BITS(nasid) | offset); | ||
| 153 | } | ||
| 154 | |||
| 155 | static inline void uv_write_global_mmr32(int nasid, unsigned long offset, | ||
| 156 | unsigned long val) | ||
| 157 | { | ||
| 158 | *uv_global_mmr32_address(nasid, offset) = val; | ||
| 159 | } | ||
| 160 | |||
| 161 | static inline unsigned long uv_read_global_mmr32(int nasid, | ||
| 162 | unsigned long offset) | ||
| 163 | { | ||
| 164 | return *uv_global_mmr32_address(nasid, offset); | ||
| 165 | } | ||
| 166 | |||
| 167 | /* | ||
| 168 | * Access Global MMR space using the MMR space located at the top of physical | ||
| 169 | * memory. | ||
| 170 | */ | ||
| 171 | static inline unsigned long *uv_global_mmr64_address(int nasid, | ||
| 172 | unsigned long offset) | ||
| 173 | { | ||
| 174 | return __va(UV_GLOBAL_MMR64_BASE | | ||
| 175 | UV_GLOBAL_MMR64_NASID_BITS(nasid) | offset); | ||
| 176 | } | ||
| 177 | |||
| 178 | static inline void uv_write_global_mmr64(int nasid, unsigned long offset, | ||
| 179 | unsigned long val) | ||
| 180 | { | ||
| 181 | *uv_global_mmr64_address(nasid, offset) = val; | ||
| 182 | } | ||
| 183 | |||
| 184 | static inline unsigned long uv_read_global_mmr64(int nasid, | ||
| 185 | unsigned long offset) | ||
| 186 | { | ||
| 187 | return *uv_global_mmr64_address(nasid, offset); | ||
| 188 | } | ||
| 189 | |||
| 190 | /* | ||
| 191 | * Access node local MMRs. Faster than using global space but only local MMRs | ||
| 192 | * are accessible. | ||
| 193 | */ | ||
| 194 | static inline unsigned long *uv_local_mmr_address(unsigned long offset) | ||
| 195 | { | ||
| 196 | return __va(UV_LOCAL_MMR_BASE | offset); | ||
| 197 | } | ||
| 198 | |||
| 199 | static inline unsigned long uv_read_local_mmr(unsigned long offset) | ||
| 200 | { | ||
| 201 | return *uv_local_mmr_address(offset); | ||
| 202 | } | ||
| 203 | |||
| 204 | static inline void uv_write_local_mmr(unsigned long offset, unsigned long val) | ||
| 205 | { | ||
| 206 | *uv_local_mmr_address(offset) = val; | ||
| 207 | } | ||
| 208 | |||
| 209 | /* | ||
| 210 | * Structures and definitions for converting between cpu, node, and blade | ||
| 211 | * numbers. | ||
| 212 | */ | ||
| 213 | struct uv_blade_info { | ||
| 214 | unsigned short nr_posible_cpus; | ||
| 215 | unsigned short nr_online_cpus; | ||
| 216 | unsigned short nasid; | ||
| 217 | }; | ||
| 218 | struct uv_blade_info *uv_blade_info; | ||
| 219 | extern short *uv_node_to_blade; | ||
| 220 | extern short *uv_cpu_to_blade; | ||
| 221 | extern short uv_possible_blades; | ||
| 222 | |||
| 223 | /* Blade-local cpu number of current cpu. Numbered 0 .. <# cpus on the blade> */ | ||
| 224 | static inline int uv_blade_processor_id(void) | ||
| 225 | { | ||
| 226 | return uv_hub_info->blade_processor_id; | ||
| 227 | } | ||
| 228 | |||
| 229 | /* Blade number of current cpu. Numnbered 0 .. <#blades -1> */ | ||
| 230 | static inline int uv_numa_blade_id(void) | ||
| 231 | { | ||
| 232 | return uv_hub_info->numa_blade_id; | ||
| 233 | } | ||
| 234 | |||
| 235 | /* Convert a cpu number to the the UV blade number */ | ||
| 236 | static inline int uv_cpu_to_blade_id(int cpu) | ||
| 237 | { | ||
| 238 | return uv_cpu_to_blade[cpu]; | ||
| 239 | } | ||
| 240 | |||
| 241 | /* Convert linux node number to the UV blade number */ | ||
| 242 | static inline int uv_node_to_blade_id(int nid) | ||
| 243 | { | ||
| 244 | return uv_node_to_blade[nid]; | ||
| 245 | } | ||
| 246 | |||
| 247 | /* Convert a blade id to the NASID of the blade */ | ||
| 248 | static inline int uv_blade_to_nasid(int bid) | ||
| 249 | { | ||
| 250 | return uv_blade_info[bid].nasid; | ||
| 251 | } | ||
| 252 | |||
| 253 | /* Determine the number of possible cpus on a blade */ | ||
| 254 | static inline int uv_blade_nr_possible_cpus(int bid) | ||
| 255 | { | ||
| 256 | return uv_blade_info[bid].nr_posible_cpus; | ||
| 257 | } | ||
| 258 | |||
| 259 | /* Determine the number of online cpus on a blade */ | ||
| 260 | static inline int uv_blade_nr_online_cpus(int bid) | ||
| 261 | { | ||
| 262 | return uv_blade_info[bid].nr_online_cpus; | ||
| 263 | } | ||
| 264 | |||
| 265 | /* Convert a cpu id to the NASID of the blade containing the cpu */ | ||
| 266 | static inline int uv_cpu_to_nasid(int cpu) | ||
| 267 | { | ||
| 268 | return uv_blade_info[uv_cpu_to_blade_id(cpu)].nasid; | ||
| 269 | } | ||
| 270 | |||
| 271 | /* Convert a node number to the NASID of the blade */ | ||
| 272 | static inline int uv_node_to_nasid(int nid) | ||
| 273 | { | ||
| 274 | return uv_blade_info[uv_node_to_blade_id(nid)].nasid; | ||
| 275 | } | ||
| 276 | |||
| 277 | /* Maximum possible number of blades */ | ||
| 278 | static inline int uv_num_possible_blades(void) | ||
| 279 | { | ||
| 280 | return uv_possible_blades; | ||
| 281 | } | ||
| 282 | |||
| 283 | #endif /* __ASM_X86_UV_HUB__ */ | ||
| 284 | |||
diff --git a/include/asm-x86/uv/uv_mmrs.h b/include/asm-x86/uv/uv_mmrs.h new file mode 100644 index 000000000000..3b69fe6b6376 --- /dev/null +++ b/include/asm-x86/uv/uv_mmrs.h | |||
| @@ -0,0 +1,373 @@ | |||
| 1 | /* | ||
| 2 | * This file is subject to the terms and conditions of the GNU General Public | ||
| 3 | * License. See the file "COPYING" in the main directory of this archive | ||
| 4 | * for more details. | ||
| 5 | * | ||
| 6 | * SGI UV MMR definitions | ||
| 7 | * | ||
| 8 | * Copyright (C) 2007-2008 Silicon Graphics, Inc. All rights reserved. | ||
| 9 | */ | ||
| 10 | |||
| 11 | #ifndef __ASM_X86_UV_MMRS__ | ||
| 12 | #define __ASM_X86_UV_MMRS__ | ||
| 13 | |||
| 14 | /* | ||
| 15 | * AUTO GENERATED - Do not edit | ||
| 16 | */ | ||
| 17 | |||
| 18 | #define UV_MMR_ENABLE (1UL << 63) | ||
| 19 | |||
| 20 | /* ========================================================================= */ | ||
| 21 | /* UVH_IPI_INT */ | ||
| 22 | /* ========================================================================= */ | ||
| 23 | #define UVH_IPI_INT 0x60500UL | ||
| 24 | #define UVH_IPI_INT_32 0x0360 | ||
| 25 | |||
| 26 | #define UVH_IPI_INT_VECTOR_SHFT 0 | ||
| 27 | #define UVH_IPI_INT_VECTOR_MASK 0x00000000000000ffUL | ||
| 28 | #define UVH_IPI_INT_DELIVERY_MODE_SHFT 8 | ||
| 29 | #define UVH_IPI_INT_DELIVERY_MODE_MASK 0x0000000000000700UL | ||
| 30 | #define UVH_IPI_INT_DESTMODE_SHFT 11 | ||
| 31 | #define UVH_IPI_INT_DESTMODE_MASK 0x0000000000000800UL | ||
| 32 | #define UVH_IPI_INT_APIC_ID_SHFT 16 | ||
| 33 | #define UVH_IPI_INT_APIC_ID_MASK 0x0000ffffffff0000UL | ||
| 34 | #define UVH_IPI_INT_SEND_SHFT 63 | ||
| 35 | #define UVH_IPI_INT_SEND_MASK 0x8000000000000000UL | ||
| 36 | |||
| 37 | union uvh_ipi_int_u { | ||
| 38 | unsigned long v; | ||
| 39 | struct uvh_ipi_int_s { | ||
| 40 | unsigned long vector_ : 8; /* RW */ | ||
| 41 | unsigned long delivery_mode : 3; /* RW */ | ||
| 42 | unsigned long destmode : 1; /* RW */ | ||
| 43 | unsigned long rsvd_12_15 : 4; /* */ | ||
| 44 | unsigned long apic_id : 32; /* RW */ | ||
| 45 | unsigned long rsvd_48_62 : 15; /* */ | ||
| 46 | unsigned long send : 1; /* WP */ | ||
| 47 | } s; | ||
| 48 | }; | ||
| 49 | |||
| 50 | /* ========================================================================= */ | ||
| 51 | /* UVH_LB_BAU_INTD_PAYLOAD_QUEUE_FIRST */ | ||
| 52 | /* ========================================================================= */ | ||
| 53 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_FIRST 0x320050UL | ||
| 54 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_FIRST_32 0x009f0 | ||
| 55 | |||
| 56 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_FIRST_ADDRESS_SHFT 4 | ||
| 57 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_FIRST_ADDRESS_MASK 0x000007fffffffff0UL | ||
| 58 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_FIRST_NODE_ID_SHFT 49 | ||
| 59 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_FIRST_NODE_ID_MASK 0x7ffe000000000000UL | ||
| 60 | |||
| 61 | union uvh_lb_bau_intd_payload_queue_first_u { | ||
| 62 | unsigned long v; | ||
| 63 | struct uvh_lb_bau_intd_payload_queue_first_s { | ||
| 64 | unsigned long rsvd_0_3: 4; /* */ | ||
| 65 | unsigned long address : 39; /* RW */ | ||
| 66 | unsigned long rsvd_43_48: 6; /* */ | ||
| 67 | unsigned long node_id : 14; /* RW */ | ||
| 68 | unsigned long rsvd_63 : 1; /* */ | ||
| 69 | } s; | ||
| 70 | }; | ||
| 71 | |||
| 72 | /* ========================================================================= */ | ||
| 73 | /* UVH_LB_BAU_INTD_PAYLOAD_QUEUE_LAST */ | ||
| 74 | /* ========================================================================= */ | ||
| 75 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_LAST 0x320060UL | ||
| 76 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_LAST_32 0x009f8 | ||
| 77 | |||
| 78 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_LAST_ADDRESS_SHFT 4 | ||
| 79 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_LAST_ADDRESS_MASK 0x000007fffffffff0UL | ||
| 80 | |||
| 81 | union uvh_lb_bau_intd_payload_queue_last_u { | ||
| 82 | unsigned long v; | ||
| 83 | struct uvh_lb_bau_intd_payload_queue_last_s { | ||
| 84 | unsigned long rsvd_0_3: 4; /* */ | ||
| 85 | unsigned long address : 39; /* RW */ | ||
| 86 | unsigned long rsvd_43_63: 21; /* */ | ||
| 87 | } s; | ||
| 88 | }; | ||
| 89 | |||
| 90 | /* ========================================================================= */ | ||
| 91 | /* UVH_LB_BAU_INTD_PAYLOAD_QUEUE_TAIL */ | ||
| 92 | /* ========================================================================= */ | ||
| 93 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_TAIL 0x320070UL | ||
| 94 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_TAIL_32 0x00a00 | ||
| 95 | |||
| 96 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_TAIL_ADDRESS_SHFT 4 | ||
| 97 | #define UVH_LB_BAU_INTD_PAYLOAD_QUEUE_TAIL_ADDRESS_MASK 0x000007fffffffff0UL | ||
| 98 | |||
| 99 | union uvh_lb_bau_intd_payload_queue_tail_u { | ||
| 100 | unsigned long v; | ||
| 101 | struct uvh_lb_bau_intd_payload_queue_tail_s { | ||
| 102 | unsigned long rsvd_0_3: 4; /* */ | ||
| 103 | unsigned long address : 39; /* RW */ | ||
| 104 | unsigned long rsvd_43_63: 21; /* */ | ||
| 105 | } s; | ||
| 106 | }; | ||
| 107 | |||
| 108 | /* ========================================================================= */ | ||
| 109 | /* UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE */ | ||
| 110 | /* ========================================================================= */ | ||
| 111 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE 0x320080UL | ||
| 112 | |||
| 113 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_0_SHFT 0 | ||
| 114 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_0_MASK 0x0000000000000001UL | ||
| 115 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_1_SHFT 1 | ||
| 116 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_1_MASK 0x0000000000000002UL | ||
| 117 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_2_SHFT 2 | ||
| 118 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_2_MASK 0x0000000000000004UL | ||
| 119 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_3_SHFT 3 | ||
| 120 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_3_MASK 0x0000000000000008UL | ||
| 121 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_4_SHFT 4 | ||
| 122 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_4_MASK 0x0000000000000010UL | ||
| 123 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_5_SHFT 5 | ||
| 124 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_5_MASK 0x0000000000000020UL | ||
| 125 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_6_SHFT 6 | ||
| 126 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_6_MASK 0x0000000000000040UL | ||
| 127 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_7_SHFT 7 | ||
| 128 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_PENDING_7_MASK 0x0000000000000080UL | ||
| 129 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_0_SHFT 8 | ||
| 130 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_0_MASK 0x0000000000000100UL | ||
| 131 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_1_SHFT 9 | ||
| 132 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_1_MASK 0x0000000000000200UL | ||
| 133 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_2_SHFT 10 | ||
| 134 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_2_MASK 0x0000000000000400UL | ||
| 135 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_3_SHFT 11 | ||
| 136 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_3_MASK 0x0000000000000800UL | ||
| 137 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_4_SHFT 12 | ||
| 138 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_4_MASK 0x0000000000001000UL | ||
| 139 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_5_SHFT 13 | ||
| 140 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_5_MASK 0x0000000000002000UL | ||
| 141 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_6_SHFT 14 | ||
| 142 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_6_MASK 0x0000000000004000UL | ||
| 143 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_7_SHFT 15 | ||
| 144 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_TIMEOUT_7_MASK 0x0000000000008000UL | ||
| 145 | union uvh_lb_bau_intd_software_acknowledge_u { | ||
| 146 | unsigned long v; | ||
| 147 | struct uvh_lb_bau_intd_software_acknowledge_s { | ||
| 148 | unsigned long pending_0 : 1; /* RW, W1C */ | ||
| 149 | unsigned long pending_1 : 1; /* RW, W1C */ | ||
| 150 | unsigned long pending_2 : 1; /* RW, W1C */ | ||
| 151 | unsigned long pending_3 : 1; /* RW, W1C */ | ||
| 152 | unsigned long pending_4 : 1; /* RW, W1C */ | ||
| 153 | unsigned long pending_5 : 1; /* RW, W1C */ | ||
| 154 | unsigned long pending_6 : 1; /* RW, W1C */ | ||
| 155 | unsigned long pending_7 : 1; /* RW, W1C */ | ||
| 156 | unsigned long timeout_0 : 1; /* RW, W1C */ | ||
| 157 | unsigned long timeout_1 : 1; /* RW, W1C */ | ||
| 158 | unsigned long timeout_2 : 1; /* RW, W1C */ | ||
| 159 | unsigned long timeout_3 : 1; /* RW, W1C */ | ||
| 160 | unsigned long timeout_4 : 1; /* RW, W1C */ | ||
| 161 | unsigned long timeout_5 : 1; /* RW, W1C */ | ||
| 162 | unsigned long timeout_6 : 1; /* RW, W1C */ | ||
| 163 | unsigned long timeout_7 : 1; /* RW, W1C */ | ||
| 164 | unsigned long rsvd_16_63: 48; /* */ | ||
| 165 | } s; | ||
| 166 | }; | ||
| 167 | |||
| 168 | /* ========================================================================= */ | ||
| 169 | /* UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_ALIAS */ | ||
| 170 | /* ========================================================================= */ | ||
| 171 | #define UVH_LB_BAU_INTD_SOFTWARE_ACKNOWLEDGE_ALIAS 0x0000000000320088UL | ||
| 172 | |||
| 173 | /* ========================================================================= */ | ||
| 174 | /* UVH_LB_BAU_SB_ACTIVATION_CONTROL */ | ||
| 175 | /* ========================================================================= */ | ||
| 176 | #define UVH_LB_BAU_SB_ACTIVATION_CONTROL 0x320020UL | ||
| 177 | #define UVH_LB_BAU_SB_ACTIVATION_CONTROL_32 0x009d8 | ||
| 178 | |||
| 179 | #define UVH_LB_BAU_SB_ACTIVATION_CONTROL_INDEX_SHFT 0 | ||
| 180 | #define UVH_LB_BAU_SB_ACTIVATION_CONTROL_INDEX_MASK 0x000000000000003fUL | ||
| 181 | #define UVH_LB_BAU_SB_ACTIVATION_CONTROL_PUSH_SHFT 62 | ||
| 182 | #define UVH_LB_BAU_SB_ACTIVATION_CONTROL_PUSH_MASK 0x4000000000000000UL | ||
| 183 | #define UVH_LB_BAU_SB_ACTIVATION_CONTROL_INIT_SHFT 63 | ||
| 184 | #define UVH_LB_BAU_SB_ACTIVATION_CONTROL_INIT_MASK 0x8000000000000000UL | ||
| 185 | |||
| 186 | union uvh_lb_bau_sb_activation_control_u { | ||
| 187 | unsigned long v; | ||
| 188 | struct uvh_lb_bau_sb_activation_control_s { | ||
| 189 | unsigned long index : 6; /* RW */ | ||
| 190 | unsigned long rsvd_6_61: 56; /* */ | ||
| 191 | unsigned long push : 1; /* WP */ | ||
| 192 | unsigned long init : 1; /* WP */ | ||
| 193 | } s; | ||
| 194 | }; | ||
| 195 | |||
| 196 | /* ========================================================================= */ | ||
| 197 | /* UVH_LB_BAU_SB_ACTIVATION_STATUS_0 */ | ||
| 198 | /* ========================================================================= */ | ||
| 199 | #define UVH_LB_BAU_SB_ACTIVATION_STATUS_0 0x320030UL | ||
| 200 | #define UVH_LB_BAU_SB_ACTIVATION_STATUS_0_32 0x009e0 | ||
| 201 | |||
| 202 | #define UVH_LB_BAU_SB_ACTIVATION_STATUS_0_STATUS_SHFT 0 | ||
| 203 | #define UVH_LB_BAU_SB_ACTIVATION_STATUS_0_STATUS_MASK 0xffffffffffffffffUL | ||
| 204 | |||
| 205 | union uvh_lb_bau_sb_activation_status_0_u { | ||
| 206 | unsigned long v; | ||
| 207 | struct uvh_lb_bau_sb_activation_status_0_s { | ||
| 208 | unsigned long status : 64; /* RW */ | ||
| 209 | } s; | ||
| 210 | }; | ||
| 211 | |||
| 212 | /* ========================================================================= */ | ||
| 213 | /* UVH_LB_BAU_SB_ACTIVATION_STATUS_1 */ | ||
| 214 | /* ========================================================================= */ | ||
| 215 | #define UVH_LB_BAU_SB_ACTIVATION_STATUS_1 0x320040UL | ||
| 216 | #define UVH_LB_BAU_SB_ACTIVATION_STATUS_1_32 0x009e8 | ||
| 217 | |||
| 218 | #define UVH_LB_BAU_SB_ACTIVATION_STATUS_1_STATUS_SHFT 0 | ||
| 219 | #define UVH_LB_BAU_SB_ACTIVATION_STATUS_1_STATUS_MASK 0xffffffffffffffffUL | ||
| 220 | |||
| 221 | union uvh_lb_bau_sb_activation_status_1_u { | ||
| 222 | unsigned long v; | ||
| 223 | struct uvh_lb_bau_sb_activation_status_1_s { | ||
| 224 | unsigned long status : 64; /* RW */ | ||
| 225 | } s; | ||
| 226 | }; | ||
| 227 | |||
| 228 | /* ========================================================================= */ | ||
| 229 | /* UVH_LB_BAU_SB_DESCRIPTOR_BASE */ | ||
| 230 | /* ========================================================================= */ | ||
| 231 | #define UVH_LB_BAU_SB_DESCRIPTOR_BASE 0x320010UL | ||
| 232 | #define UVH_LB_BAU_SB_DESCRIPTOR_BASE_32 0x009d0 | ||
| 233 | |||
| 234 | #define UVH_LB_BAU_SB_DESCRIPTOR_BASE_PAGE_ADDRESS_SHFT 12 | ||
| 235 | #define UVH_LB_BAU_SB_DESCRIPTOR_BASE_PAGE_ADDRESS_MASK 0x000007fffffff000UL | ||
| 236 | #define UVH_LB_BAU_SB_DESCRIPTOR_BASE_NODE_ID_SHFT 49 | ||
| 237 | #define UVH_LB_BAU_SB_DESCRIPTOR_BASE_NODE_ID_MASK 0x7ffe000000000000UL | ||
| 238 | |||
| 239 | union uvh_lb_bau_sb_descriptor_base_u { | ||
| 240 | unsigned long v; | ||
| 241 | struct uvh_lb_bau_sb_descriptor_base_s { | ||
| 242 | unsigned long rsvd_0_11 : 12; /* */ | ||
| 243 | unsigned long page_address : 31; /* RW */ | ||
| 244 | unsigned long rsvd_43_48 : 6; /* */ | ||
| 245 | unsigned long node_id : 14; /* RW */ | ||
| 246 | unsigned long rsvd_63 : 1; /* */ | ||
| 247 | } s; | ||
| 248 | }; | ||
| 249 | |||
| 250 | /* ========================================================================= */ | ||
| 251 | /* UVH_NODE_ID */ | ||
| 252 | /* ========================================================================= */ | ||
| 253 | #define UVH_NODE_ID 0x0UL | ||
| 254 | |||
| 255 | #define UVH_NODE_ID_FORCE1_SHFT 0 | ||
| 256 | #define UVH_NODE_ID_FORCE1_MASK 0x0000000000000001UL | ||
| 257 | #define UVH_NODE_ID_MANUFACTURER_SHFT 1 | ||
| 258 | #define UVH_NODE_ID_MANUFACTURER_MASK 0x0000000000000ffeUL | ||
| 259 | #define UVH_NODE_ID_PART_NUMBER_SHFT 12 | ||
| 260 | #define UVH_NODE_ID_PART_NUMBER_MASK 0x000000000ffff000UL | ||
| 261 | #define UVH_NODE_ID_REVISION_SHFT 28 | ||
| 262 | #define UVH_NODE_ID_REVISION_MASK 0x00000000f0000000UL | ||
| 263 | #define UVH_NODE_ID_NODE_ID_SHFT 32 | ||
| 264 | #define UVH_NODE_ID_NODE_ID_MASK 0x00007fff00000000UL | ||
| 265 | #define UVH_NODE_ID_NODES_PER_BIT_SHFT 48 | ||
| 266 | #define UVH_NODE_ID_NODES_PER_BIT_MASK 0x007f000000000000UL | ||
| 267 | #define UVH_NODE_ID_NI_PORT_SHFT 56 | ||
| 268 | #define UVH_NODE_ID_NI_PORT_MASK 0x0f00000000000000UL | ||
| 269 | |||
| 270 | union uvh_node_id_u { | ||
| 271 | unsigned long v; | ||
| 272 | struct uvh_node_id_s { | ||
| 273 | unsigned long force1 : 1; /* RO */ | ||
| 274 | unsigned long manufacturer : 11; /* RO */ | ||
| 275 | unsigned long part_number : 16; /* RO */ | ||
| 276 | unsigned long revision : 4; /* RO */ | ||
| 277 | unsigned long node_id : 15; /* RW */ | ||
| 278 | unsigned long rsvd_47 : 1; /* */ | ||
| 279 | unsigned long nodes_per_bit : 7; /* RW */ | ||
| 280 | unsigned long rsvd_55 : 1; /* */ | ||
| 281 | unsigned long ni_port : 4; /* RO */ | ||
| 282 | unsigned long rsvd_60_63 : 4; /* */ | ||
| 283 | } s; | ||
| 284 | }; | ||
| 285 | |||
| 286 | /* ========================================================================= */ | ||
| 287 | /* UVH_RH_GAM_GRU_OVERLAY_CONFIG_MMR */ | ||
| 288 | /* ========================================================================= */ | ||
| 289 | #define UVH_RH_GAM_GRU_OVERLAY_CONFIG_MMR 0x1600010UL | ||
| 290 | |||
| 291 | #define UVH_RH_GAM_GRU_OVERLAY_CONFIG_MMR_BASE_SHFT 28 | ||
| 292 | #define UVH_RH_GAM_GRU_OVERLAY_CONFIG_MMR_BASE_MASK 0x00003ffff0000000UL | ||
| 293 | #define UVH_RH_GAM_GRU_OVERLAY_CONFIG_MMR_GR4_SHFT 46 | ||
| 294 | #define UVH_RH_GAM_GRU_OVERLAY_CONFIG_MMR_GR4_MASK 0x0000400000000000UL | ||
| 295 | #define UVH_RH_GAM_GRU_OVERLAY_CONFIG_MMR_N_GRU_SHFT 52 | ||
| 296 | #define UVH_RH_GAM_GRU_OVERLAY_CONFIG_MMR_N_GRU_MASK 0x00f0000000000000UL | ||
| 297 | #define UVH_RH_GAM_GRU_OVERLAY_CONFIG_MMR_ENABLE_SHFT 63 | ||
| 298 | #define UVH_RH_GAM_GRU_OVERLAY_CONFIG_MMR_ENABLE_MASK 0x8000000000000000UL | ||
| 299 | |||
| 300 | union uvh_rh_gam_gru_overlay_config_mmr_u { | ||
| 301 | unsigned long v; | ||
| 302 | struct uvh_rh_gam_gru_overlay_config_mmr_s { | ||
| 303 | unsigned long rsvd_0_27: 28; /* */ | ||
| 304 | unsigned long base : 18; /* RW */ | ||
| 305 | unsigned long gr4 : 1; /* RW */ | ||
| 306 | unsigned long rsvd_47_51: 5; /* */ | ||
| 307 | unsigned long n_gru : 4; /* RW */ | ||
| 308 | unsigned long rsvd_56_62: 7; /* */ | ||
| 309 | unsigned long enable : 1; /* RW */ | ||
| 310 | } s; | ||
| 311 | }; | ||
| 312 | |||
| 313 | /* ========================================================================= */ | ||
| 314 | /* UVH_RH_GAM_MMR_OVERLAY_CONFIG_MMR */ | ||
| 315 | /* ========================================================================= */ | ||
| 316 | #define UVH_RH_GAM_MMR_OVERLAY_CONFIG_MMR 0x1600028UL | ||
| 317 | |||
| 318 | #define UVH_RH_GAM_MMR_OVERLAY_CONFIG_MMR_BASE_SHFT 26 | ||
| 319 | #define UVH_RH_GAM_MMR_OVERLAY_CONFIG_MMR_BASE_MASK 0x00003ffffc000000UL | ||
| 320 | #define UVH_RH_GAM_MMR_OVERLAY_CONFIG_MMR_DUAL_HUB_SHFT 46 | ||
| 321 | #define UVH_RH_GAM_MMR_OVERLAY_CONFIG_MMR_DUAL_HUB_MASK 0x0000400000000000UL | ||
| 322 | #define UVH_RH_GAM_MMR_OVERLAY_CONFIG_MMR_ENABLE_SHFT 63 | ||
| 323 | #define UVH_RH_GAM_MMR_OVERLAY_CONFIG_MMR_ENABLE_MASK 0x8000000000000000UL | ||
| 324 | |||
| 325 | union uvh_rh_gam_mmr_overlay_config_mmr_u { | ||
| 326 | unsigned long v; | ||
| 327 | struct uvh_rh_gam_mmr_overlay_config_mmr_s { | ||
| 328 | unsigned long rsvd_0_25: 26; /* */ | ||
| 329 | unsigned long base : 20; /* RW */ | ||
| 330 | unsigned long dual_hub : 1; /* RW */ | ||
| 331 | unsigned long rsvd_47_62: 16; /* */ | ||
| 332 | unsigned long enable : 1; /* RW */ | ||
| 333 | } s; | ||
| 334 | }; | ||
| 335 | |||
| 336 | /* ========================================================================= */ | ||
| 337 | /* UVH_RTC */ | ||
| 338 | /* ========================================================================= */ | ||
| 339 | #define UVH_RTC 0x28000UL | ||
| 340 | |||
| 341 | #define UVH_RTC_REAL_TIME_CLOCK_SHFT 0 | ||
| 342 | #define UVH_RTC_REAL_TIME_CLOCK_MASK 0x00ffffffffffffffUL | ||
| 343 | |||
| 344 | union uvh_rtc_u { | ||
| 345 | unsigned long v; | ||
| 346 | struct uvh_rtc_s { | ||
| 347 | unsigned long real_time_clock : 56; /* RW */ | ||
| 348 | unsigned long rsvd_56_63 : 8; /* */ | ||
| 349 | } s; | ||
| 350 | }; | ||
| 351 | |||
| 352 | /* ========================================================================= */ | ||
| 353 | /* UVH_SI_ADDR_MAP_CONFIG */ | ||
| 354 | /* ========================================================================= */ | ||
| 355 | #define UVH_SI_ADDR_MAP_CONFIG 0xc80000UL | ||
| 356 | |||
| 357 | #define UVH_SI_ADDR_MAP_CONFIG_M_SKT_SHFT 0 | ||
| 358 | #define UVH_SI_ADDR_MAP_CONFIG_M_SKT_MASK 0x000000000000003fUL | ||
| 359 | #define UVH_SI_ADDR_MAP_CONFIG_N_SKT_SHFT 8 | ||
| 360 | #define UVH_SI_ADDR_MAP_CONFIG_N_SKT_MASK 0x0000000000000f00UL | ||
| 361 | |||
| 362 | union uvh_si_addr_map_config_u { | ||
| 363 | unsigned long v; | ||
| 364 | struct uvh_si_addr_map_config_s { | ||
| 365 | unsigned long m_skt : 6; /* RW */ | ||
| 366 | unsigned long rsvd_6_7: 2; /* */ | ||
| 367 | unsigned long n_skt : 4; /* RW */ | ||
| 368 | unsigned long rsvd_12_63: 52; /* */ | ||
| 369 | } s; | ||
| 370 | }; | ||
| 371 | |||
| 372 | |||
| 373 | #endif /* __ASM_X86_UV_MMRS__ */ | ||
diff --git a/include/asm-x86/vdso.h b/include/asm-x86/vdso.h index 629bcb6e8e45..86e085e003d2 100644 --- a/include/asm-x86/vdso.h +++ b/include/asm-x86/vdso.h | |||
| @@ -8,9 +8,11 @@ extern const char VDSO64_PRELINK[]; | |||
| 8 | * Given a pointer to the vDSO image, find the pointer to VDSO64_name | 8 | * Given a pointer to the vDSO image, find the pointer to VDSO64_name |
| 9 | * as that symbol is defined in the vDSO sources or linker script. | 9 | * as that symbol is defined in the vDSO sources or linker script. |
| 10 | */ | 10 | */ |
| 11 | #define VDSO64_SYMBOL(base, name) ({ \ | 11 | #define VDSO64_SYMBOL(base, name) \ |
| 12 | extern const char VDSO64_##name[]; \ | 12 | ({ \ |
| 13 | (void *) (VDSO64_##name - VDSO64_PRELINK + (unsigned long) (base)); }) | 13 | extern const char VDSO64_##name[]; \ |
| 14 | (void *)(VDSO64_##name - VDSO64_PRELINK + (unsigned long)(base)); \ | ||
| 15 | }) | ||
| 14 | #endif | 16 | #endif |
| 15 | 17 | ||
| 16 | #if defined CONFIG_X86_32 || defined CONFIG_COMPAT | 18 | #if defined CONFIG_X86_32 || defined CONFIG_COMPAT |
| @@ -20,9 +22,18 @@ extern const char VDSO32_PRELINK[]; | |||
| 20 | * Given a pointer to the vDSO image, find the pointer to VDSO32_name | 22 | * Given a pointer to the vDSO image, find the pointer to VDSO32_name |
| 21 | * as that symbol is defined in the vDSO sources or linker script. | 23 | * as that symbol is defined in the vDSO sources or linker script. |
| 22 | */ | 24 | */ |
| 23 | #define VDSO32_SYMBOL(base, name) ({ \ | 25 | #define VDSO32_SYMBOL(base, name) \ |
| 24 | extern const char VDSO32_##name[]; \ | 26 | ({ \ |
| 25 | (void *) (VDSO32_##name - VDSO32_PRELINK + (unsigned long) (base)); }) | 27 | extern const char VDSO32_##name[]; \ |
| 28 | (void *)(VDSO32_##name - VDSO32_PRELINK + (unsigned long)(base)); \ | ||
| 29 | }) | ||
| 26 | #endif | 30 | #endif |
| 27 | 31 | ||
| 32 | /* | ||
| 33 | * These symbols are defined with the addresses in the vsyscall page. | ||
| 34 | * See vsyscall-sigreturn.S. | ||
| 35 | */ | ||
| 36 | extern void __user __kernel_sigreturn; | ||
| 37 | extern void __user __kernel_rt_sigreturn; | ||
| 38 | |||
| 28 | #endif /* asm-x86/vdso.h */ | 39 | #endif /* asm-x86/vdso.h */ |
diff --git a/include/asm-x86/vga.h b/include/asm-x86/vga.h index 0ecf68ac03aa..0ccf804377e6 100644 --- a/include/asm-x86/vga.h +++ b/include/asm-x86/vga.h | |||
| @@ -12,9 +12,9 @@ | |||
| 12 | * access the videoram directly without any black magic. | 12 | * access the videoram directly without any black magic. |
| 13 | */ | 13 | */ |
| 14 | 14 | ||
| 15 | #define VGA_MAP_MEM(x,s) (unsigned long)phys_to_virt(x) | 15 | #define VGA_MAP_MEM(x, s) (unsigned long)phys_to_virt(x) |
| 16 | 16 | ||
| 17 | #define vga_readb(x) (*(x)) | 17 | #define vga_readb(x) (*(x)) |
| 18 | #define vga_writeb(x,y) (*(y) = (x)) | 18 | #define vga_writeb(x, y) (*(y) = (x)) |
| 19 | 19 | ||
| 20 | #endif | 20 | #endif |
diff --git a/include/asm-x86/vm86.h b/include/asm-x86/vm86.h index c92fe4af52e8..074b357146df 100644 --- a/include/asm-x86/vm86.h +++ b/include/asm-x86/vm86.h | |||
| @@ -12,19 +12,13 @@ | |||
| 12 | * Linus | 12 | * Linus |
| 13 | */ | 13 | */ |
| 14 | 14 | ||
| 15 | #define TF_MASK 0x00000100 | 15 | #include <asm/processor-flags.h> |
| 16 | #define IF_MASK 0x00000200 | 16 | |
| 17 | #define IOPL_MASK 0x00003000 | ||
| 18 | #define NT_MASK 0x00004000 | ||
| 19 | #ifdef CONFIG_VM86 | 17 | #ifdef CONFIG_VM86 |
| 20 | #define VM_MASK 0x00020000 | 18 | #define X86_VM_MASK X86_EFLAGS_VM |
| 21 | #else | 19 | #else |
| 22 | #define VM_MASK 0 /* ignored */ | 20 | #define X86_VM_MASK 0 /* No VM86 support */ |
| 23 | #endif | 21 | #endif |
| 24 | #define AC_MASK 0x00040000 | ||
| 25 | #define VIF_MASK 0x00080000 /* virtual interrupt flag */ | ||
| 26 | #define VIP_MASK 0x00100000 /* virtual interrupt pending */ | ||
| 27 | #define ID_MASK 0x00200000 | ||
| 28 | 22 | ||
| 29 | #define BIOSSEG 0x0f000 | 23 | #define BIOSSEG 0x0f000 |
| 30 | 24 | ||
| @@ -42,9 +36,11 @@ | |||
| 42 | #define VM86_ARG(retval) ((retval) >> 8) | 36 | #define VM86_ARG(retval) ((retval) >> 8) |
| 43 | 37 | ||
| 44 | #define VM86_SIGNAL 0 /* return due to signal */ | 38 | #define VM86_SIGNAL 0 /* return due to signal */ |
| 45 | #define VM86_UNKNOWN 1 /* unhandled GP fault - IO-instruction or similar */ | 39 | #define VM86_UNKNOWN 1 /* unhandled GP fault |
| 40 | - IO-instruction or similar */ | ||
| 46 | #define VM86_INTx 2 /* int3/int x instruction (ARG = x) */ | 41 | #define VM86_INTx 2 /* int3/int x instruction (ARG = x) */ |
| 47 | #define VM86_STI 3 /* sti/popf/iret instruction enabled virtual interrupts */ | 42 | #define VM86_STI 3 /* sti/popf/iret instruction enabled |
| 43 | virtual interrupts */ | ||
| 48 | 44 | ||
| 49 | /* | 45 | /* |
| 50 | * Additional return values when invoking new vm86() | 46 | * Additional return values when invoking new vm86() |
| @@ -205,7 +201,8 @@ void release_vm86_irqs(struct task_struct *); | |||
| 205 | #define handle_vm86_fault(a, b) | 201 | #define handle_vm86_fault(a, b) |
| 206 | #define release_vm86_irqs(a) | 202 | #define release_vm86_irqs(a) |
| 207 | 203 | ||
| 208 | static inline int handle_vm86_trap(struct kernel_vm86_regs *a, long b, int c) { | 204 | static inline int handle_vm86_trap(struct kernel_vm86_regs *a, long b, int c) |
| 205 | { | ||
| 209 | return 0; | 206 | return 0; |
| 210 | } | 207 | } |
| 211 | 208 | ||
diff --git a/include/asm-x86/vmi.h b/include/asm-x86/vmi.h index eb8bd892c01e..b7c0dea119fe 100644 --- a/include/asm-x86/vmi.h +++ b/include/asm-x86/vmi.h | |||
| @@ -155,9 +155,9 @@ | |||
| 155 | 155 | ||
| 156 | #ifndef __ASSEMBLY__ | 156 | #ifndef __ASSEMBLY__ |
| 157 | struct vmi_relocation_info { | 157 | struct vmi_relocation_info { |
| 158 | unsigned char *eip; | 158 | unsigned char *eip; |
| 159 | unsigned char type; | 159 | unsigned char type; |
| 160 | unsigned char reserved[3]; | 160 | unsigned char reserved[3]; |
| 161 | }; | 161 | }; |
| 162 | #endif | 162 | #endif |
| 163 | 163 | ||
| @@ -173,53 +173,53 @@ struct vmi_relocation_info { | |||
| 173 | #ifndef __ASSEMBLY__ | 173 | #ifndef __ASSEMBLY__ |
| 174 | 174 | ||
| 175 | struct vrom_header { | 175 | struct vrom_header { |
| 176 | u16 rom_signature; // option ROM signature | 176 | u16 rom_signature; /* option ROM signature */ |
| 177 | u8 rom_length; // ROM length in 512 byte chunks | 177 | u8 rom_length; /* ROM length in 512 byte chunks */ |
| 178 | u8 rom_entry[4]; // 16-bit code entry point | 178 | u8 rom_entry[4]; /* 16-bit code entry point */ |
| 179 | u8 rom_pad0; // 4-byte align pad | 179 | u8 rom_pad0; /* 4-byte align pad */ |
| 180 | u32 vrom_signature; // VROM identification signature | 180 | u32 vrom_signature; /* VROM identification signature */ |
| 181 | u8 api_version_min;// Minor version of API | 181 | u8 api_version_min;/* Minor version of API */ |
| 182 | u8 api_version_maj;// Major version of API | 182 | u8 api_version_maj;/* Major version of API */ |
| 183 | u8 jump_slots; // Number of jump slots | 183 | u8 jump_slots; /* Number of jump slots */ |
| 184 | u8 reserved1; // Reserved for expansion | 184 | u8 reserved1; /* Reserved for expansion */ |
| 185 | u32 virtual_top; // Hypervisor virtual address start | 185 | u32 virtual_top; /* Hypervisor virtual address start */ |
| 186 | u16 reserved2; // Reserved for expansion | 186 | u16 reserved2; /* Reserved for expansion */ |
| 187 | u16 license_offs; // Offset to License string | 187 | u16 license_offs; /* Offset to License string */ |
| 188 | u16 pci_header_offs;// Offset to PCI OPROM header | 188 | u16 pci_header_offs;/* Offset to PCI OPROM header */ |
| 189 | u16 pnp_header_offs;// Offset to PnP OPROM header | 189 | u16 pnp_header_offs;/* Offset to PnP OPROM header */ |
| 190 | u32 rom_pad3; // PnP reserverd / VMI reserved | 190 | u32 rom_pad3; /* PnP reserverd / VMI reserved */ |
| 191 | u8 reserved[96]; // Reserved for headers | 191 | u8 reserved[96]; /* Reserved for headers */ |
| 192 | char vmi_init[8]; // VMI_Init jump point | 192 | char vmi_init[8]; /* VMI_Init jump point */ |
| 193 | char get_reloc[8]; // VMI_GetRelocationInfo jump point | 193 | char get_reloc[8]; /* VMI_GetRelocationInfo jump point */ |
| 194 | } __attribute__((packed)); | 194 | } __attribute__((packed)); |
| 195 | 195 | ||
| 196 | struct pnp_header { | 196 | struct pnp_header { |
| 197 | char sig[4]; | 197 | char sig[4]; |
| 198 | char rev; | 198 | char rev; |
| 199 | char size; | 199 | char size; |
| 200 | short next; | 200 | short next; |
| 201 | short res; | 201 | short res; |
| 202 | long devID; | 202 | long devID; |
| 203 | unsigned short manufacturer_offset; | 203 | unsigned short manufacturer_offset; |
| 204 | unsigned short product_offset; | 204 | unsigned short product_offset; |
| 205 | } __attribute__((packed)); | 205 | } __attribute__((packed)); |
| 206 | 206 | ||
| 207 | struct pci_header { | 207 | struct pci_header { |
| 208 | char sig[4]; | 208 | char sig[4]; |
| 209 | short vendorID; | 209 | short vendorID; |
| 210 | short deviceID; | 210 | short deviceID; |
| 211 | short vpdData; | 211 | short vpdData; |
| 212 | short size; | 212 | short size; |
| 213 | char rev; | 213 | char rev; |
| 214 | char class; | 214 | char class; |
| 215 | char subclass; | 215 | char subclass; |
| 216 | char interface; | 216 | char interface; |
| 217 | short chunks; | 217 | short chunks; |
| 218 | char rom_version_min; | 218 | char rom_version_min; |
| 219 | char rom_version_maj; | 219 | char rom_version_maj; |
| 220 | char codetype; | 220 | char codetype; |
| 221 | char lastRom; | 221 | char lastRom; |
| 222 | short reserved; | 222 | short reserved; |
| 223 | } __attribute__((packed)); | 223 | } __attribute__((packed)); |
| 224 | 224 | ||
| 225 | /* Function prototypes for bootstrapping */ | 225 | /* Function prototypes for bootstrapping */ |
diff --git a/include/asm-x86/voyager.h b/include/asm-x86/voyager.h index 91a9932937ab..9c811d2e6f91 100644 --- a/include/asm-x86/voyager.h +++ b/include/asm-x86/voyager.h | |||
| @@ -91,8 +91,7 @@ | |||
| 91 | #define VOYAGER_WRITE_CONFIG 0x2 | 91 | #define VOYAGER_WRITE_CONFIG 0x2 |
| 92 | #define VOYAGER_BYPASS 0xff | 92 | #define VOYAGER_BYPASS 0xff |
| 93 | 93 | ||
| 94 | typedef struct voyager_asic | 94 | typedef struct voyager_asic { |
| 95 | { | ||
| 96 | __u8 asic_addr; /* ASIC address; Level 4 */ | 95 | __u8 asic_addr; /* ASIC address; Level 4 */ |
| 97 | __u8 asic_type; /* ASIC type */ | 96 | __u8 asic_type; /* ASIC type */ |
| 98 | __u8 asic_id; /* ASIC id */ | 97 | __u8 asic_id; /* ASIC id */ |
| @@ -113,7 +112,7 @@ typedef struct voyager_module { | |||
| 113 | __u16 largest_reg; /* Largest register in the scan path */ | 112 | __u16 largest_reg; /* Largest register in the scan path */ |
| 114 | __u16 smallest_reg; /* Smallest register in the scan path */ | 113 | __u16 smallest_reg; /* Smallest register in the scan path */ |
| 115 | voyager_asic_t *asic; /* First ASIC in scan path (CAT_I) */ | 114 | voyager_asic_t *asic; /* First ASIC in scan path (CAT_I) */ |
| 116 | struct voyager_module *submodule; /* Submodule pointer */ | 115 | struct voyager_module *submodule; /* Submodule pointer */ |
| 117 | struct voyager_module *next; /* Next module in linked list */ | 116 | struct voyager_module *next; /* Next module in linked list */ |
| 118 | } voyager_module_t; | 117 | } voyager_module_t; |
| 119 | 118 | ||
| @@ -135,7 +134,7 @@ typedef struct voyager_eeprom_hdr { | |||
| 135 | __u16 cct_offset; | 134 | __u16 cct_offset; |
| 136 | __u16 log_length; /* length of err log */ | 135 | __u16 log_length; /* length of err log */ |
| 137 | __u16 xsum_end; /* offset to end of | 136 | __u16 xsum_end; /* offset to end of |
| 138 | checksum */ | 137 | checksum */ |
| 139 | __u8 reserved[4]; | 138 | __u8 reserved[4]; |
| 140 | __u8 sflag; /* starting sentinal */ | 139 | __u8 sflag; /* starting sentinal */ |
| 141 | __u8 part_number[13]; /* prom part number */ | 140 | __u8 part_number[13]; /* prom part number */ |
| @@ -148,7 +147,8 @@ typedef struct voyager_eeprom_hdr { | |||
| 148 | 147 | ||
| 149 | 148 | ||
| 150 | 149 | ||
| 151 | #define VOYAGER_EPROM_SIZE_OFFSET ((__u16)(&(((voyager_eprom_hdr_t *)0)->ee_size))) | 150 | #define VOYAGER_EPROM_SIZE_OFFSET \ |
| 151 | ((__u16)(&(((voyager_eprom_hdr_t *)0)->ee_size))) | ||
| 152 | #define VOYAGER_XSUM_END_OFFSET 0x2a | 152 | #define VOYAGER_XSUM_END_OFFSET 0x2a |
| 153 | 153 | ||
| 154 | /* the following three definitions are for internal table layouts | 154 | /* the following three definitions are for internal table layouts |
| @@ -199,7 +199,7 @@ typedef struct voyager_asic_data_table { | |||
| 199 | #define VOYAGER_WCBIC_TOM_L 0x4 | 199 | #define VOYAGER_WCBIC_TOM_L 0x4 |
| 200 | #define VOYAGER_WCBIC_TOM_H 0x5 | 200 | #define VOYAGER_WCBIC_TOM_H 0x5 |
| 201 | 201 | ||
| 202 | /* register defines for Voyager Memory Contol (VMC) | 202 | /* register defines for Voyager Memory Contol (VMC) |
| 203 | * these are present on L4 machines only */ | 203 | * these are present on L4 machines only */ |
| 204 | #define VOYAGER_VMC1 0x81 | 204 | #define VOYAGER_VMC1 0x81 |
| 205 | #define VOYAGER_VMC2 0x91 | 205 | #define VOYAGER_VMC2 0x91 |
| @@ -334,7 +334,7 @@ typedef struct { | |||
| 334 | 334 | ||
| 335 | struct QuadDescription { | 335 | struct QuadDescription { |
| 336 | __u8 Type; /* for type 0 (DYADIC or MONADIC) all fields | 336 | __u8 Type; /* for type 0 (DYADIC or MONADIC) all fields |
| 337 | * will be zero except for slot */ | 337 | * will be zero except for slot */ |
| 338 | __u8 StructureVersion; | 338 | __u8 StructureVersion; |
| 339 | __u32 CPI_BaseAddress; | 339 | __u32 CPI_BaseAddress; |
| 340 | __u32 LARC_BankSize; | 340 | __u32 LARC_BankSize; |
| @@ -342,7 +342,7 @@ struct QuadDescription { | |||
| 342 | __u8 Slot; /* Processor slots 1 - 4 */ | 342 | __u8 Slot; /* Processor slots 1 - 4 */ |
| 343 | } __attribute__((packed)); | 343 | } __attribute__((packed)); |
| 344 | 344 | ||
| 345 | struct ProcBoardInfo { | 345 | struct ProcBoardInfo { |
| 346 | __u8 Type; | 346 | __u8 Type; |
| 347 | __u8 StructureVersion; | 347 | __u8 StructureVersion; |
| 348 | __u8 NumberOfBoards; | 348 | __u8 NumberOfBoards; |
| @@ -382,19 +382,30 @@ struct CPU_Info { | |||
| 382 | * packed in it by our friend the compiler. | 382 | * packed in it by our friend the compiler. |
| 383 | */ | 383 | */ |
| 384 | typedef struct { | 384 | typedef struct { |
| 385 | __u8 Mailbox_SUS; /* Written to by SUS to give commands/response to the OS */ | 385 | __u8 Mailbox_SUS; /* Written to by SUS to give |
| 386 | __u8 Mailbox_OS; /* Written to by the OS to give commands/response to SUS */ | 386 | commands/response to the OS */ |
| 387 | __u8 SUS_MailboxVersion; /* Tells the OS which iteration of the interface SUS supports */ | 387 | __u8 Mailbox_OS; /* Written to by the OS to give |
| 388 | __u8 OS_MailboxVersion; /* Tells SUS which iteration of the interface the OS supports */ | 388 | commands/response to SUS */ |
| 389 | __u32 OS_Flags; /* Flags set by the OS as info for SUS */ | 389 | __u8 SUS_MailboxVersion; /* Tells the OS which iteration of the |
| 390 | __u32 SUS_Flags; /* Flags set by SUS as info for the OS */ | 390 | interface SUS supports */ |
| 391 | __u32 WatchDogPeriod; /* Watchdog period (in seconds) which the DP uses to see if the OS is dead */ | 391 | __u8 OS_MailboxVersion; /* Tells SUS which iteration of the |
| 392 | interface the OS supports */ | ||
| 393 | __u32 OS_Flags; /* Flags set by the OS as info for | ||
| 394 | SUS */ | ||
| 395 | __u32 SUS_Flags; /* Flags set by SUS as info | ||
| 396 | for the OS */ | ||
| 397 | __u32 WatchDogPeriod; /* Watchdog period (in seconds) which | ||
| 398 | the DP uses to see if the OS | ||
| 399 | is dead */ | ||
| 392 | __u32 WatchDogCount; /* Updated by the OS on every tic. */ | 400 | __u32 WatchDogCount; /* Updated by the OS on every tic. */ |
| 393 | __u32 MemoryFor_SUS_ErrorLog; /* Flat 32 bit address which tells SUS where to stuff the SUS error log on a dump */ | 401 | __u32 MemoryFor_SUS_ErrorLog; /* Flat 32 bit address which tells SUS |
| 394 | MC_SlotInformation_t MC_SlotInfo[NUMBER_OF_MC_BUSSES*SLOTS_PER_MC_BUS]; /* Storage for MCA POS data */ | 402 | where to stuff the SUS error log |
| 403 | on a dump */ | ||
| 404 | MC_SlotInformation_t MC_SlotInfo[NUMBER_OF_MC_BUSSES*SLOTS_PER_MC_BUS]; | ||
| 405 | /* Storage for MCA POS data */ | ||
| 395 | /* All new SECOND_PASS_INTERFACE fields added from this point */ | 406 | /* All new SECOND_PASS_INTERFACE fields added from this point */ |
| 396 | struct ProcBoardInfo *BoardData; | 407 | struct ProcBoardInfo *BoardData; |
| 397 | struct CPU_Info *CPU_Data; | 408 | struct CPU_Info *CPU_Data; |
| 398 | /* All new fields must be added from this point */ | 409 | /* All new fields must be added from this point */ |
| 399 | } Voyager_KernelSUS_Mbox_t; | 410 | } Voyager_KernelSUS_Mbox_t; |
| 400 | 411 | ||
| @@ -478,7 +489,7 @@ struct voyager_SUS { | |||
| 478 | __u32 SUS_errorlog; | 489 | __u32 SUS_errorlog; |
| 479 | /* lots of system configuration stuff under here */ | 490 | /* lots of system configuration stuff under here */ |
| 480 | }; | 491 | }; |
| 481 | 492 | ||
| 482 | /* Variables exported by voyager_smp */ | 493 | /* Variables exported by voyager_smp */ |
| 483 | extern __u32 voyager_extended_vic_processors; | 494 | extern __u32 voyager_extended_vic_processors; |
| 484 | extern __u32 voyager_allowed_boot_processors; | 495 | extern __u32 voyager_allowed_boot_processors; |
diff --git a/include/asm-x86/xor_32.h b/include/asm-x86/xor_32.h index a41ef1bdd424..067b5c1835a3 100644 --- a/include/asm-x86/xor_32.h +++ b/include/asm-x86/xor_32.h | |||
| @@ -16,12 +16,12 @@ | |||
| 16 | * Copyright (C) 1998 Ingo Molnar. | 16 | * Copyright (C) 1998 Ingo Molnar. |
| 17 | */ | 17 | */ |
| 18 | 18 | ||
| 19 | #define LD(x,y) " movq 8*("#x")(%1), %%mm"#y" ;\n" | 19 | #define LD(x, y) " movq 8*("#x")(%1), %%mm"#y" ;\n" |
| 20 | #define ST(x,y) " movq %%mm"#y", 8*("#x")(%1) ;\n" | 20 | #define ST(x, y) " movq %%mm"#y", 8*("#x")(%1) ;\n" |
| 21 | #define XO1(x,y) " pxor 8*("#x")(%2), %%mm"#y" ;\n" | 21 | #define XO1(x, y) " pxor 8*("#x")(%2), %%mm"#y" ;\n" |
| 22 | #define XO2(x,y) " pxor 8*("#x")(%3), %%mm"#y" ;\n" | 22 | #define XO2(x, y) " pxor 8*("#x")(%3), %%mm"#y" ;\n" |
| 23 | #define XO3(x,y) " pxor 8*("#x")(%4), %%mm"#y" ;\n" | 23 | #define XO3(x, y) " pxor 8*("#x")(%4), %%mm"#y" ;\n" |
| 24 | #define XO4(x,y) " pxor 8*("#x")(%5), %%mm"#y" ;\n" | 24 | #define XO4(x, y) " pxor 8*("#x")(%5), %%mm"#y" ;\n" |
| 25 | 25 | ||
| 26 | #include <asm/i387.h> | 26 | #include <asm/i387.h> |
| 27 | 27 | ||
| @@ -32,24 +32,24 @@ xor_pII_mmx_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) | |||
| 32 | 32 | ||
| 33 | kernel_fpu_begin(); | 33 | kernel_fpu_begin(); |
| 34 | 34 | ||
| 35 | __asm__ __volatile__ ( | 35 | asm volatile( |
| 36 | #undef BLOCK | 36 | #undef BLOCK |
| 37 | #define BLOCK(i) \ | 37 | #define BLOCK(i) \ |
| 38 | LD(i,0) \ | 38 | LD(i, 0) \ |
| 39 | LD(i+1,1) \ | 39 | LD(i + 1, 1) \ |
| 40 | LD(i+2,2) \ | 40 | LD(i + 2, 2) \ |
| 41 | LD(i+3,3) \ | 41 | LD(i + 3, 3) \ |
| 42 | XO1(i,0) \ | 42 | XO1(i, 0) \ |
| 43 | ST(i,0) \ | 43 | ST(i, 0) \ |
| 44 | XO1(i+1,1) \ | 44 | XO1(i+1, 1) \ |
| 45 | ST(i+1,1) \ | 45 | ST(i+1, 1) \ |
| 46 | XO1(i+2,2) \ | 46 | XO1(i + 2, 2) \ |
| 47 | ST(i+2,2) \ | 47 | ST(i + 2, 2) \ |
| 48 | XO1(i+3,3) \ | 48 | XO1(i + 3, 3) \ |
| 49 | ST(i+3,3) | 49 | ST(i + 3, 3) |
| 50 | 50 | ||
| 51 | " .align 32 ;\n" | 51 | " .align 32 ;\n" |
| 52 | " 1: ;\n" | 52 | " 1: ;\n" |
| 53 | 53 | ||
| 54 | BLOCK(0) | 54 | BLOCK(0) |
| 55 | BLOCK(4) | 55 | BLOCK(4) |
| @@ -76,25 +76,25 @@ xor_pII_mmx_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 76 | 76 | ||
| 77 | kernel_fpu_begin(); | 77 | kernel_fpu_begin(); |
| 78 | 78 | ||
| 79 | __asm__ __volatile__ ( | 79 | asm volatile( |
| 80 | #undef BLOCK | 80 | #undef BLOCK |
| 81 | #define BLOCK(i) \ | 81 | #define BLOCK(i) \ |
| 82 | LD(i,0) \ | 82 | LD(i, 0) \ |
| 83 | LD(i+1,1) \ | 83 | LD(i + 1, 1) \ |
| 84 | LD(i+2,2) \ | 84 | LD(i + 2, 2) \ |
| 85 | LD(i+3,3) \ | 85 | LD(i + 3, 3) \ |
| 86 | XO1(i,0) \ | 86 | XO1(i, 0) \ |
| 87 | XO1(i+1,1) \ | 87 | XO1(i + 1, 1) \ |
| 88 | XO1(i+2,2) \ | 88 | XO1(i + 2, 2) \ |
| 89 | XO1(i+3,3) \ | 89 | XO1(i + 3, 3) \ |
| 90 | XO2(i,0) \ | 90 | XO2(i, 0) \ |
| 91 | ST(i,0) \ | 91 | ST(i, 0) \ |
| 92 | XO2(i+1,1) \ | 92 | XO2(i + 1, 1) \ |
| 93 | ST(i+1,1) \ | 93 | ST(i + 1, 1) \ |
| 94 | XO2(i+2,2) \ | 94 | XO2(i + 2, 2) \ |
| 95 | ST(i+2,2) \ | 95 | ST(i + 2, 2) \ |
| 96 | XO2(i+3,3) \ | 96 | XO2(i + 3, 3) \ |
| 97 | ST(i+3,3) | 97 | ST(i + 3, 3) |
| 98 | 98 | ||
| 99 | " .align 32 ;\n" | 99 | " .align 32 ;\n" |
| 100 | " 1: ;\n" | 100 | " 1: ;\n" |
| @@ -125,29 +125,29 @@ xor_pII_mmx_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 125 | 125 | ||
| 126 | kernel_fpu_begin(); | 126 | kernel_fpu_begin(); |
| 127 | 127 | ||
| 128 | __asm__ __volatile__ ( | 128 | asm volatile( |
| 129 | #undef BLOCK | 129 | #undef BLOCK |
| 130 | #define BLOCK(i) \ | 130 | #define BLOCK(i) \ |
| 131 | LD(i,0) \ | 131 | LD(i, 0) \ |
| 132 | LD(i+1,1) \ | 132 | LD(i + 1, 1) \ |
| 133 | LD(i+2,2) \ | 133 | LD(i + 2, 2) \ |
| 134 | LD(i+3,3) \ | 134 | LD(i + 3, 3) \ |
| 135 | XO1(i,0) \ | 135 | XO1(i, 0) \ |
| 136 | XO1(i+1,1) \ | 136 | XO1(i + 1, 1) \ |
| 137 | XO1(i+2,2) \ | 137 | XO1(i + 2, 2) \ |
| 138 | XO1(i+3,3) \ | 138 | XO1(i + 3, 3) \ |
| 139 | XO2(i,0) \ | 139 | XO2(i, 0) \ |
| 140 | XO2(i+1,1) \ | 140 | XO2(i + 1, 1) \ |
| 141 | XO2(i+2,2) \ | 141 | XO2(i + 2, 2) \ |
| 142 | XO2(i+3,3) \ | 142 | XO2(i + 3, 3) \ |
| 143 | XO3(i,0) \ | 143 | XO3(i, 0) \ |
| 144 | ST(i,0) \ | 144 | ST(i, 0) \ |
| 145 | XO3(i+1,1) \ | 145 | XO3(i + 1, 1) \ |
| 146 | ST(i+1,1) \ | 146 | ST(i + 1, 1) \ |
| 147 | XO3(i+2,2) \ | 147 | XO3(i + 2, 2) \ |
| 148 | ST(i+2,2) \ | 148 | ST(i + 2, 2) \ |
| 149 | XO3(i+3,3) \ | 149 | XO3(i + 3, 3) \ |
| 150 | ST(i+3,3) | 150 | ST(i + 3, 3) |
| 151 | 151 | ||
| 152 | " .align 32 ;\n" | 152 | " .align 32 ;\n" |
| 153 | " 1: ;\n" | 153 | " 1: ;\n" |
| @@ -186,35 +186,35 @@ xor_pII_mmx_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 186 | because we modify p4 and p5 there, but we can't mark them | 186 | because we modify p4 and p5 there, but we can't mark them |
| 187 | as read/write, otherwise we'd overflow the 10-asm-operands | 187 | as read/write, otherwise we'd overflow the 10-asm-operands |
| 188 | limit of GCC < 3.1. */ | 188 | limit of GCC < 3.1. */ |
| 189 | __asm__ ("" : "+r" (p4), "+r" (p5)); | 189 | asm("" : "+r" (p4), "+r" (p5)); |
| 190 | 190 | ||
| 191 | __asm__ __volatile__ ( | 191 | asm volatile( |
| 192 | #undef BLOCK | 192 | #undef BLOCK |
| 193 | #define BLOCK(i) \ | 193 | #define BLOCK(i) \ |
| 194 | LD(i,0) \ | 194 | LD(i, 0) \ |
| 195 | LD(i+1,1) \ | 195 | LD(i + 1, 1) \ |
| 196 | LD(i+2,2) \ | 196 | LD(i + 2, 2) \ |
| 197 | LD(i+3,3) \ | 197 | LD(i + 3, 3) \ |
| 198 | XO1(i,0) \ | 198 | XO1(i, 0) \ |
| 199 | XO1(i+1,1) \ | 199 | XO1(i + 1, 1) \ |
| 200 | XO1(i+2,2) \ | 200 | XO1(i + 2, 2) \ |
| 201 | XO1(i+3,3) \ | 201 | XO1(i + 3, 3) \ |
| 202 | XO2(i,0) \ | 202 | XO2(i, 0) \ |
| 203 | XO2(i+1,1) \ | 203 | XO2(i + 1, 1) \ |
| 204 | XO2(i+2,2) \ | 204 | XO2(i + 2, 2) \ |
| 205 | XO2(i+3,3) \ | 205 | XO2(i + 3, 3) \ |
| 206 | XO3(i,0) \ | 206 | XO3(i, 0) \ |
| 207 | XO3(i+1,1) \ | 207 | XO3(i + 1, 1) \ |
| 208 | XO3(i+2,2) \ | 208 | XO3(i + 2, 2) \ |
| 209 | XO3(i+3,3) \ | 209 | XO3(i + 3, 3) \ |
| 210 | XO4(i,0) \ | 210 | XO4(i, 0) \ |
| 211 | ST(i,0) \ | 211 | ST(i, 0) \ |
| 212 | XO4(i+1,1) \ | 212 | XO4(i + 1, 1) \ |
| 213 | ST(i+1,1) \ | 213 | ST(i + 1, 1) \ |
| 214 | XO4(i+2,2) \ | 214 | XO4(i + 2, 2) \ |
| 215 | ST(i+2,2) \ | 215 | ST(i + 2, 2) \ |
| 216 | XO4(i+3,3) \ | 216 | XO4(i + 3, 3) \ |
| 217 | ST(i+3,3) | 217 | ST(i + 3, 3) |
| 218 | 218 | ||
| 219 | " .align 32 ;\n" | 219 | " .align 32 ;\n" |
| 220 | " 1: ;\n" | 220 | " 1: ;\n" |
| @@ -233,13 +233,13 @@ xor_pII_mmx_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 233 | " jnz 1b ;\n" | 233 | " jnz 1b ;\n" |
| 234 | : "+r" (lines), | 234 | : "+r" (lines), |
| 235 | "+r" (p1), "+r" (p2), "+r" (p3) | 235 | "+r" (p1), "+r" (p2), "+r" (p3) |
| 236 | : "r" (p4), "r" (p5) | 236 | : "r" (p4), "r" (p5) |
| 237 | : "memory"); | 237 | : "memory"); |
| 238 | 238 | ||
| 239 | /* p4 and p5 were modified, and now the variables are dead. | 239 | /* p4 and p5 were modified, and now the variables are dead. |
| 240 | Clobber them just to be sure nobody does something stupid | 240 | Clobber them just to be sure nobody does something stupid |
| 241 | like assuming they have some legal value. */ | 241 | like assuming they have some legal value. */ |
| 242 | __asm__ ("" : "=r" (p4), "=r" (p5)); | 242 | asm("" : "=r" (p4), "=r" (p5)); |
| 243 | 243 | ||
| 244 | kernel_fpu_end(); | 244 | kernel_fpu_end(); |
| 245 | } | 245 | } |
| @@ -259,7 +259,7 @@ xor_p5_mmx_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) | |||
| 259 | 259 | ||
| 260 | kernel_fpu_begin(); | 260 | kernel_fpu_begin(); |
| 261 | 261 | ||
| 262 | __asm__ __volatile__ ( | 262 | asm volatile( |
| 263 | " .align 32 ;\n" | 263 | " .align 32 ;\n" |
| 264 | " 1: ;\n" | 264 | " 1: ;\n" |
| 265 | " movq (%1), %%mm0 ;\n" | 265 | " movq (%1), %%mm0 ;\n" |
| @@ -286,7 +286,7 @@ xor_p5_mmx_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) | |||
| 286 | " pxor 56(%2), %%mm7 ;\n" | 286 | " pxor 56(%2), %%mm7 ;\n" |
| 287 | " movq %%mm6, 48(%1) ;\n" | 287 | " movq %%mm6, 48(%1) ;\n" |
| 288 | " movq %%mm7, 56(%1) ;\n" | 288 | " movq %%mm7, 56(%1) ;\n" |
| 289 | 289 | ||
| 290 | " addl $64, %1 ;\n" | 290 | " addl $64, %1 ;\n" |
| 291 | " addl $64, %2 ;\n" | 291 | " addl $64, %2 ;\n" |
| 292 | " decl %0 ;\n" | 292 | " decl %0 ;\n" |
| @@ -307,7 +307,7 @@ xor_p5_mmx_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 307 | 307 | ||
| 308 | kernel_fpu_begin(); | 308 | kernel_fpu_begin(); |
| 309 | 309 | ||
| 310 | __asm__ __volatile__ ( | 310 | asm volatile( |
| 311 | " .align 32,0x90 ;\n" | 311 | " .align 32,0x90 ;\n" |
| 312 | " 1: ;\n" | 312 | " 1: ;\n" |
| 313 | " movq (%1), %%mm0 ;\n" | 313 | " movq (%1), %%mm0 ;\n" |
| @@ -342,7 +342,7 @@ xor_p5_mmx_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 342 | " pxor 56(%3), %%mm7 ;\n" | 342 | " pxor 56(%3), %%mm7 ;\n" |
| 343 | " movq %%mm6, 48(%1) ;\n" | 343 | " movq %%mm6, 48(%1) ;\n" |
| 344 | " movq %%mm7, 56(%1) ;\n" | 344 | " movq %%mm7, 56(%1) ;\n" |
| 345 | 345 | ||
| 346 | " addl $64, %1 ;\n" | 346 | " addl $64, %1 ;\n" |
| 347 | " addl $64, %2 ;\n" | 347 | " addl $64, %2 ;\n" |
| 348 | " addl $64, %3 ;\n" | 348 | " addl $64, %3 ;\n" |
| @@ -364,7 +364,7 @@ xor_p5_mmx_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 364 | 364 | ||
| 365 | kernel_fpu_begin(); | 365 | kernel_fpu_begin(); |
| 366 | 366 | ||
| 367 | __asm__ __volatile__ ( | 367 | asm volatile( |
| 368 | " .align 32,0x90 ;\n" | 368 | " .align 32,0x90 ;\n" |
| 369 | " 1: ;\n" | 369 | " 1: ;\n" |
| 370 | " movq (%1), %%mm0 ;\n" | 370 | " movq (%1), %%mm0 ;\n" |
| @@ -407,7 +407,7 @@ xor_p5_mmx_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 407 | " pxor 56(%4), %%mm7 ;\n" | 407 | " pxor 56(%4), %%mm7 ;\n" |
| 408 | " movq %%mm6, 48(%1) ;\n" | 408 | " movq %%mm6, 48(%1) ;\n" |
| 409 | " movq %%mm7, 56(%1) ;\n" | 409 | " movq %%mm7, 56(%1) ;\n" |
| 410 | 410 | ||
| 411 | " addl $64, %1 ;\n" | 411 | " addl $64, %1 ;\n" |
| 412 | " addl $64, %2 ;\n" | 412 | " addl $64, %2 ;\n" |
| 413 | " addl $64, %3 ;\n" | 413 | " addl $64, %3 ;\n" |
| @@ -436,9 +436,9 @@ xor_p5_mmx_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 436 | because we modify p4 and p5 there, but we can't mark them | 436 | because we modify p4 and p5 there, but we can't mark them |
| 437 | as read/write, otherwise we'd overflow the 10-asm-operands | 437 | as read/write, otherwise we'd overflow the 10-asm-operands |
| 438 | limit of GCC < 3.1. */ | 438 | limit of GCC < 3.1. */ |
| 439 | __asm__ ("" : "+r" (p4), "+r" (p5)); | 439 | asm("" : "+r" (p4), "+r" (p5)); |
| 440 | 440 | ||
| 441 | __asm__ __volatile__ ( | 441 | asm volatile( |
| 442 | " .align 32,0x90 ;\n" | 442 | " .align 32,0x90 ;\n" |
| 443 | " 1: ;\n" | 443 | " 1: ;\n" |
| 444 | " movq (%1), %%mm0 ;\n" | 444 | " movq (%1), %%mm0 ;\n" |
| @@ -489,7 +489,7 @@ xor_p5_mmx_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 489 | " pxor 56(%5), %%mm7 ;\n" | 489 | " pxor 56(%5), %%mm7 ;\n" |
| 490 | " movq %%mm6, 48(%1) ;\n" | 490 | " movq %%mm6, 48(%1) ;\n" |
| 491 | " movq %%mm7, 56(%1) ;\n" | 491 | " movq %%mm7, 56(%1) ;\n" |
| 492 | 492 | ||
| 493 | " addl $64, %1 ;\n" | 493 | " addl $64, %1 ;\n" |
| 494 | " addl $64, %2 ;\n" | 494 | " addl $64, %2 ;\n" |
| 495 | " addl $64, %3 ;\n" | 495 | " addl $64, %3 ;\n" |
| @@ -505,7 +505,7 @@ xor_p5_mmx_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 505 | /* p4 and p5 were modified, and now the variables are dead. | 505 | /* p4 and p5 were modified, and now the variables are dead. |
| 506 | Clobber them just to be sure nobody does something stupid | 506 | Clobber them just to be sure nobody does something stupid |
| 507 | like assuming they have some legal value. */ | 507 | like assuming they have some legal value. */ |
| 508 | __asm__ ("" : "=r" (p4), "=r" (p5)); | 508 | asm("" : "=r" (p4), "=r" (p5)); |
| 509 | 509 | ||
| 510 | kernel_fpu_end(); | 510 | kernel_fpu_end(); |
| 511 | } | 511 | } |
| @@ -531,11 +531,12 @@ static struct xor_block_template xor_block_p5_mmx = { | |||
| 531 | * Copyright (C) 1999 Zach Brown (with obvious credit due Ingo) | 531 | * Copyright (C) 1999 Zach Brown (with obvious credit due Ingo) |
| 532 | */ | 532 | */ |
| 533 | 533 | ||
| 534 | #define XMMS_SAVE do { \ | 534 | #define XMMS_SAVE \ |
| 535 | do { \ | ||
| 535 | preempt_disable(); \ | 536 | preempt_disable(); \ |
| 536 | cr0 = read_cr0(); \ | 537 | cr0 = read_cr0(); \ |
| 537 | clts(); \ | 538 | clts(); \ |
| 538 | __asm__ __volatile__ ( \ | 539 | asm volatile( \ |
| 539 | "movups %%xmm0,(%0) ;\n\t" \ | 540 | "movups %%xmm0,(%0) ;\n\t" \ |
| 540 | "movups %%xmm1,0x10(%0) ;\n\t" \ | 541 | "movups %%xmm1,0x10(%0) ;\n\t" \ |
| 541 | "movups %%xmm2,0x20(%0) ;\n\t" \ | 542 | "movups %%xmm2,0x20(%0) ;\n\t" \ |
| @@ -543,10 +544,11 @@ static struct xor_block_template xor_block_p5_mmx = { | |||
| 543 | : \ | 544 | : \ |
| 544 | : "r" (xmm_save) \ | 545 | : "r" (xmm_save) \ |
| 545 | : "memory"); \ | 546 | : "memory"); \ |
| 546 | } while(0) | 547 | } while (0) |
| 547 | 548 | ||
| 548 | #define XMMS_RESTORE do { \ | 549 | #define XMMS_RESTORE \ |
| 549 | __asm__ __volatile__ ( \ | 550 | do { \ |
| 551 | asm volatile( \ | ||
| 550 | "sfence ;\n\t" \ | 552 | "sfence ;\n\t" \ |
| 551 | "movups (%0),%%xmm0 ;\n\t" \ | 553 | "movups (%0),%%xmm0 ;\n\t" \ |
| 552 | "movups 0x10(%0),%%xmm1 ;\n\t" \ | 554 | "movups 0x10(%0),%%xmm1 ;\n\t" \ |
| @@ -557,76 +559,76 @@ static struct xor_block_template xor_block_p5_mmx = { | |||
| 557 | : "memory"); \ | 559 | : "memory"); \ |
| 558 | write_cr0(cr0); \ | 560 | write_cr0(cr0); \ |
| 559 | preempt_enable(); \ | 561 | preempt_enable(); \ |
| 560 | } while(0) | 562 | } while (0) |
| 561 | 563 | ||
| 562 | #define ALIGN16 __attribute__((aligned(16))) | 564 | #define ALIGN16 __attribute__((aligned(16))) |
| 563 | 565 | ||
| 564 | #define OFFS(x) "16*("#x")" | 566 | #define OFFS(x) "16*("#x")" |
| 565 | #define PF_OFFS(x) "256+16*("#x")" | 567 | #define PF_OFFS(x) "256+16*("#x")" |
| 566 | #define PF0(x) " prefetchnta "PF_OFFS(x)"(%1) ;\n" | 568 | #define PF0(x) " prefetchnta "PF_OFFS(x)"(%1) ;\n" |
| 567 | #define LD(x,y) " movaps "OFFS(x)"(%1), %%xmm"#y" ;\n" | 569 | #define LD(x, y) " movaps "OFFS(x)"(%1), %%xmm"#y" ;\n" |
| 568 | #define ST(x,y) " movaps %%xmm"#y", "OFFS(x)"(%1) ;\n" | 570 | #define ST(x, y) " movaps %%xmm"#y", "OFFS(x)"(%1) ;\n" |
| 569 | #define PF1(x) " prefetchnta "PF_OFFS(x)"(%2) ;\n" | 571 | #define PF1(x) " prefetchnta "PF_OFFS(x)"(%2) ;\n" |
| 570 | #define PF2(x) " prefetchnta "PF_OFFS(x)"(%3) ;\n" | 572 | #define PF2(x) " prefetchnta "PF_OFFS(x)"(%3) ;\n" |
| 571 | #define PF3(x) " prefetchnta "PF_OFFS(x)"(%4) ;\n" | 573 | #define PF3(x) " prefetchnta "PF_OFFS(x)"(%4) ;\n" |
| 572 | #define PF4(x) " prefetchnta "PF_OFFS(x)"(%5) ;\n" | 574 | #define PF4(x) " prefetchnta "PF_OFFS(x)"(%5) ;\n" |
| 573 | #define PF5(x) " prefetchnta "PF_OFFS(x)"(%6) ;\n" | 575 | #define PF5(x) " prefetchnta "PF_OFFS(x)"(%6) ;\n" |
| 574 | #define XO1(x,y) " xorps "OFFS(x)"(%2), %%xmm"#y" ;\n" | 576 | #define XO1(x, y) " xorps "OFFS(x)"(%2), %%xmm"#y" ;\n" |
| 575 | #define XO2(x,y) " xorps "OFFS(x)"(%3), %%xmm"#y" ;\n" | 577 | #define XO2(x, y) " xorps "OFFS(x)"(%3), %%xmm"#y" ;\n" |
| 576 | #define XO3(x,y) " xorps "OFFS(x)"(%4), %%xmm"#y" ;\n" | 578 | #define XO3(x, y) " xorps "OFFS(x)"(%4), %%xmm"#y" ;\n" |
| 577 | #define XO4(x,y) " xorps "OFFS(x)"(%5), %%xmm"#y" ;\n" | 579 | #define XO4(x, y) " xorps "OFFS(x)"(%5), %%xmm"#y" ;\n" |
| 578 | #define XO5(x,y) " xorps "OFFS(x)"(%6), %%xmm"#y" ;\n" | 580 | #define XO5(x, y) " xorps "OFFS(x)"(%6), %%xmm"#y" ;\n" |
| 579 | 581 | ||
| 580 | 582 | ||
| 581 | static void | 583 | static void |
| 582 | xor_sse_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) | 584 | xor_sse_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) |
| 583 | { | 585 | { |
| 584 | unsigned long lines = bytes >> 8; | 586 | unsigned long lines = bytes >> 8; |
| 585 | char xmm_save[16*4] ALIGN16; | 587 | char xmm_save[16*4] ALIGN16; |
| 586 | int cr0; | 588 | int cr0; |
| 587 | 589 | ||
| 588 | XMMS_SAVE; | 590 | XMMS_SAVE; |
| 589 | 591 | ||
| 590 | __asm__ __volatile__ ( | 592 | asm volatile( |
| 591 | #undef BLOCK | 593 | #undef BLOCK |
| 592 | #define BLOCK(i) \ | 594 | #define BLOCK(i) \ |
| 593 | LD(i,0) \ | 595 | LD(i, 0) \ |
| 594 | LD(i+1,1) \ | 596 | LD(i + 1, 1) \ |
| 595 | PF1(i) \ | 597 | PF1(i) \ |
| 596 | PF1(i+2) \ | 598 | PF1(i + 2) \ |
| 597 | LD(i+2,2) \ | 599 | LD(i + 2, 2) \ |
| 598 | LD(i+3,3) \ | 600 | LD(i + 3, 3) \ |
| 599 | PF0(i+4) \ | 601 | PF0(i + 4) \ |
| 600 | PF0(i+6) \ | 602 | PF0(i + 6) \ |
| 601 | XO1(i,0) \ | 603 | XO1(i, 0) \ |
| 602 | XO1(i+1,1) \ | 604 | XO1(i + 1, 1) \ |
| 603 | XO1(i+2,2) \ | 605 | XO1(i + 2, 2) \ |
| 604 | XO1(i+3,3) \ | 606 | XO1(i + 3, 3) \ |
| 605 | ST(i,0) \ | 607 | ST(i, 0) \ |
| 606 | ST(i+1,1) \ | 608 | ST(i + 1, 1) \ |
| 607 | ST(i+2,2) \ | 609 | ST(i + 2, 2) \ |
| 608 | ST(i+3,3) \ | 610 | ST(i + 3, 3) \ |
| 609 | 611 | ||
| 610 | 612 | ||
| 611 | PF0(0) | 613 | PF0(0) |
| 612 | PF0(2) | 614 | PF0(2) |
| 613 | 615 | ||
| 614 | " .align 32 ;\n" | 616 | " .align 32 ;\n" |
| 615 | " 1: ;\n" | 617 | " 1: ;\n" |
| 616 | 618 | ||
| 617 | BLOCK(0) | 619 | BLOCK(0) |
| 618 | BLOCK(4) | 620 | BLOCK(4) |
| 619 | BLOCK(8) | 621 | BLOCK(8) |
| 620 | BLOCK(12) | 622 | BLOCK(12) |
| 621 | 623 | ||
| 622 | " addl $256, %1 ;\n" | 624 | " addl $256, %1 ;\n" |
| 623 | " addl $256, %2 ;\n" | 625 | " addl $256, %2 ;\n" |
| 624 | " decl %0 ;\n" | 626 | " decl %0 ;\n" |
| 625 | " jnz 1b ;\n" | 627 | " jnz 1b ;\n" |
| 626 | : "+r" (lines), | 628 | : "+r" (lines), |
| 627 | "+r" (p1), "+r" (p2) | 629 | "+r" (p1), "+r" (p2) |
| 628 | : | 630 | : |
| 629 | : "memory"); | 631 | : "memory"); |
| 630 | 632 | ||
| 631 | XMMS_RESTORE; | 633 | XMMS_RESTORE; |
| 632 | } | 634 | } |
| @@ -635,59 +637,59 @@ static void | |||
| 635 | xor_sse_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, | 637 | xor_sse_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, |
| 636 | unsigned long *p3) | 638 | unsigned long *p3) |
| 637 | { | 639 | { |
| 638 | unsigned long lines = bytes >> 8; | 640 | unsigned long lines = bytes >> 8; |
| 639 | char xmm_save[16*4] ALIGN16; | 641 | char xmm_save[16*4] ALIGN16; |
| 640 | int cr0; | 642 | int cr0; |
| 641 | 643 | ||
| 642 | XMMS_SAVE; | 644 | XMMS_SAVE; |
| 643 | 645 | ||
| 644 | __asm__ __volatile__ ( | 646 | asm volatile( |
| 645 | #undef BLOCK | 647 | #undef BLOCK |
| 646 | #define BLOCK(i) \ | 648 | #define BLOCK(i) \ |
| 647 | PF1(i) \ | 649 | PF1(i) \ |
| 648 | PF1(i+2) \ | 650 | PF1(i + 2) \ |
| 649 | LD(i,0) \ | 651 | LD(i,0) \ |
| 650 | LD(i+1,1) \ | 652 | LD(i + 1, 1) \ |
| 651 | LD(i+2,2) \ | 653 | LD(i + 2, 2) \ |
| 652 | LD(i+3,3) \ | 654 | LD(i + 3, 3) \ |
| 653 | PF2(i) \ | 655 | PF2(i) \ |
| 654 | PF2(i+2) \ | 656 | PF2(i + 2) \ |
| 655 | PF0(i+4) \ | 657 | PF0(i + 4) \ |
| 656 | PF0(i+6) \ | 658 | PF0(i + 6) \ |
| 657 | XO1(i,0) \ | 659 | XO1(i,0) \ |
| 658 | XO1(i+1,1) \ | 660 | XO1(i + 1, 1) \ |
| 659 | XO1(i+2,2) \ | 661 | XO1(i + 2, 2) \ |
| 660 | XO1(i+3,3) \ | 662 | XO1(i + 3, 3) \ |
| 661 | XO2(i,0) \ | 663 | XO2(i,0) \ |
| 662 | XO2(i+1,1) \ | 664 | XO2(i + 1, 1) \ |
| 663 | XO2(i+2,2) \ | 665 | XO2(i + 2, 2) \ |
| 664 | XO2(i+3,3) \ | 666 | XO2(i + 3, 3) \ |
| 665 | ST(i,0) \ | 667 | ST(i,0) \ |
| 666 | ST(i+1,1) \ | 668 | ST(i + 1, 1) \ |
| 667 | ST(i+2,2) \ | 669 | ST(i + 2, 2) \ |
| 668 | ST(i+3,3) \ | 670 | ST(i + 3, 3) \ |
| 669 | 671 | ||
| 670 | 672 | ||
| 671 | PF0(0) | 673 | PF0(0) |
| 672 | PF0(2) | 674 | PF0(2) |
| 673 | 675 | ||
| 674 | " .align 32 ;\n" | 676 | " .align 32 ;\n" |
| 675 | " 1: ;\n" | 677 | " 1: ;\n" |
| 676 | 678 | ||
| 677 | BLOCK(0) | 679 | BLOCK(0) |
| 678 | BLOCK(4) | 680 | BLOCK(4) |
| 679 | BLOCK(8) | 681 | BLOCK(8) |
| 680 | BLOCK(12) | 682 | BLOCK(12) |
| 681 | 683 | ||
| 682 | " addl $256, %1 ;\n" | 684 | " addl $256, %1 ;\n" |
| 683 | " addl $256, %2 ;\n" | 685 | " addl $256, %2 ;\n" |
| 684 | " addl $256, %3 ;\n" | 686 | " addl $256, %3 ;\n" |
| 685 | " decl %0 ;\n" | 687 | " decl %0 ;\n" |
| 686 | " jnz 1b ;\n" | 688 | " jnz 1b ;\n" |
| 687 | : "+r" (lines), | 689 | : "+r" (lines), |
| 688 | "+r" (p1), "+r"(p2), "+r"(p3) | 690 | "+r" (p1), "+r"(p2), "+r"(p3) |
| 689 | : | 691 | : |
| 690 | : "memory" ); | 692 | : "memory" ); |
| 691 | 693 | ||
| 692 | XMMS_RESTORE; | 694 | XMMS_RESTORE; |
| 693 | } | 695 | } |
| @@ -696,66 +698,66 @@ static void | |||
| 696 | xor_sse_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, | 698 | xor_sse_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, |
| 697 | unsigned long *p3, unsigned long *p4) | 699 | unsigned long *p3, unsigned long *p4) |
| 698 | { | 700 | { |
| 699 | unsigned long lines = bytes >> 8; | 701 | unsigned long lines = bytes >> 8; |
| 700 | char xmm_save[16*4] ALIGN16; | 702 | char xmm_save[16*4] ALIGN16; |
| 701 | int cr0; | 703 | int cr0; |
| 702 | 704 | ||
| 703 | XMMS_SAVE; | 705 | XMMS_SAVE; |
| 704 | 706 | ||
| 705 | __asm__ __volatile__ ( | 707 | asm volatile( |
| 706 | #undef BLOCK | 708 | #undef BLOCK |
| 707 | #define BLOCK(i) \ | 709 | #define BLOCK(i) \ |
| 708 | PF1(i) \ | 710 | PF1(i) \ |
| 709 | PF1(i+2) \ | 711 | PF1(i + 2) \ |
| 710 | LD(i,0) \ | 712 | LD(i,0) \ |
| 711 | LD(i+1,1) \ | 713 | LD(i + 1, 1) \ |
| 712 | LD(i+2,2) \ | 714 | LD(i + 2, 2) \ |
| 713 | LD(i+3,3) \ | 715 | LD(i + 3, 3) \ |
| 714 | PF2(i) \ | 716 | PF2(i) \ |
| 715 | PF2(i+2) \ | 717 | PF2(i + 2) \ |
| 716 | XO1(i,0) \ | 718 | XO1(i,0) \ |
| 717 | XO1(i+1,1) \ | 719 | XO1(i + 1, 1) \ |
| 718 | XO1(i+2,2) \ | 720 | XO1(i + 2, 2) \ |
| 719 | XO1(i+3,3) \ | 721 | XO1(i + 3, 3) \ |
| 720 | PF3(i) \ | 722 | PF3(i) \ |
| 721 | PF3(i+2) \ | 723 | PF3(i + 2) \ |
| 722 | PF0(i+4) \ | 724 | PF0(i + 4) \ |
| 723 | PF0(i+6) \ | 725 | PF0(i + 6) \ |
| 724 | XO2(i,0) \ | 726 | XO2(i,0) \ |
| 725 | XO2(i+1,1) \ | 727 | XO2(i + 1, 1) \ |
| 726 | XO2(i+2,2) \ | 728 | XO2(i + 2, 2) \ |
| 727 | XO2(i+3,3) \ | 729 | XO2(i + 3, 3) \ |
| 728 | XO3(i,0) \ | 730 | XO3(i,0) \ |
| 729 | XO3(i+1,1) \ | 731 | XO3(i + 1, 1) \ |
| 730 | XO3(i+2,2) \ | 732 | XO3(i + 2, 2) \ |
| 731 | XO3(i+3,3) \ | 733 | XO3(i + 3, 3) \ |
| 732 | ST(i,0) \ | 734 | ST(i,0) \ |
| 733 | ST(i+1,1) \ | 735 | ST(i + 1, 1) \ |
| 734 | ST(i+2,2) \ | 736 | ST(i + 2, 2) \ |
| 735 | ST(i+3,3) \ | 737 | ST(i + 3, 3) \ |
| 736 | 738 | ||
| 737 | 739 | ||
| 738 | PF0(0) | 740 | PF0(0) |
| 739 | PF0(2) | 741 | PF0(2) |
| 740 | 742 | ||
| 741 | " .align 32 ;\n" | 743 | " .align 32 ;\n" |
| 742 | " 1: ;\n" | 744 | " 1: ;\n" |
| 743 | 745 | ||
| 744 | BLOCK(0) | 746 | BLOCK(0) |
| 745 | BLOCK(4) | 747 | BLOCK(4) |
| 746 | BLOCK(8) | 748 | BLOCK(8) |
| 747 | BLOCK(12) | 749 | BLOCK(12) |
| 748 | 750 | ||
| 749 | " addl $256, %1 ;\n" | 751 | " addl $256, %1 ;\n" |
| 750 | " addl $256, %2 ;\n" | 752 | " addl $256, %2 ;\n" |
| 751 | " addl $256, %3 ;\n" | 753 | " addl $256, %3 ;\n" |
| 752 | " addl $256, %4 ;\n" | 754 | " addl $256, %4 ;\n" |
| 753 | " decl %0 ;\n" | 755 | " decl %0 ;\n" |
| 754 | " jnz 1b ;\n" | 756 | " jnz 1b ;\n" |
| 755 | : "+r" (lines), | 757 | : "+r" (lines), |
| 756 | "+r" (p1), "+r" (p2), "+r" (p3), "+r" (p4) | 758 | "+r" (p1), "+r" (p2), "+r" (p3), "+r" (p4) |
| 757 | : | 759 | : |
| 758 | : "memory" ); | 760 | : "memory" ); |
| 759 | 761 | ||
| 760 | XMMS_RESTORE; | 762 | XMMS_RESTORE; |
| 761 | } | 763 | } |
| @@ -764,7 +766,7 @@ static void | |||
| 764 | xor_sse_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, | 766 | xor_sse_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, |
| 765 | unsigned long *p3, unsigned long *p4, unsigned long *p5) | 767 | unsigned long *p3, unsigned long *p4, unsigned long *p5) |
| 766 | { | 768 | { |
| 767 | unsigned long lines = bytes >> 8; | 769 | unsigned long lines = bytes >> 8; |
| 768 | char xmm_save[16*4] ALIGN16; | 770 | char xmm_save[16*4] ALIGN16; |
| 769 | int cr0; | 771 | int cr0; |
| 770 | 772 | ||
| @@ -776,65 +778,65 @@ xor_sse_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 776 | because we modify p4 and p5 there, but we can't mark them | 778 | because we modify p4 and p5 there, but we can't mark them |
| 777 | as read/write, otherwise we'd overflow the 10-asm-operands | 779 | as read/write, otherwise we'd overflow the 10-asm-operands |
| 778 | limit of GCC < 3.1. */ | 780 | limit of GCC < 3.1. */ |
| 779 | __asm__ ("" : "+r" (p4), "+r" (p5)); | 781 | asm("" : "+r" (p4), "+r" (p5)); |
| 780 | 782 | ||
| 781 | __asm__ __volatile__ ( | 783 | asm volatile( |
| 782 | #undef BLOCK | 784 | #undef BLOCK |
| 783 | #define BLOCK(i) \ | 785 | #define BLOCK(i) \ |
| 784 | PF1(i) \ | 786 | PF1(i) \ |
| 785 | PF1(i+2) \ | 787 | PF1(i + 2) \ |
| 786 | LD(i,0) \ | 788 | LD(i,0) \ |
| 787 | LD(i+1,1) \ | 789 | LD(i + 1, 1) \ |
| 788 | LD(i+2,2) \ | 790 | LD(i + 2, 2) \ |
| 789 | LD(i+3,3) \ | 791 | LD(i + 3, 3) \ |
| 790 | PF2(i) \ | 792 | PF2(i) \ |
| 791 | PF2(i+2) \ | 793 | PF2(i + 2) \ |
| 792 | XO1(i,0) \ | 794 | XO1(i,0) \ |
| 793 | XO1(i+1,1) \ | 795 | XO1(i + 1, 1) \ |
| 794 | XO1(i+2,2) \ | 796 | XO1(i + 2, 2) \ |
| 795 | XO1(i+3,3) \ | 797 | XO1(i + 3, 3) \ |
| 796 | PF3(i) \ | 798 | PF3(i) \ |
| 797 | PF3(i+2) \ | 799 | PF3(i + 2) \ |
| 798 | XO2(i,0) \ | 800 | XO2(i,0) \ |
| 799 | XO2(i+1,1) \ | 801 | XO2(i + 1, 1) \ |
| 800 | XO2(i+2,2) \ | 802 | XO2(i + 2, 2) \ |
| 801 | XO2(i+3,3) \ | 803 | XO2(i + 3, 3) \ |
| 802 | PF4(i) \ | 804 | PF4(i) \ |
| 803 | PF4(i+2) \ | 805 | PF4(i + 2) \ |
| 804 | PF0(i+4) \ | 806 | PF0(i + 4) \ |
| 805 | PF0(i+6) \ | 807 | PF0(i + 6) \ |
| 806 | XO3(i,0) \ | 808 | XO3(i,0) \ |
| 807 | XO3(i+1,1) \ | 809 | XO3(i + 1, 1) \ |
| 808 | XO3(i+2,2) \ | 810 | XO3(i + 2, 2) \ |
| 809 | XO3(i+3,3) \ | 811 | XO3(i + 3, 3) \ |
| 810 | XO4(i,0) \ | 812 | XO4(i,0) \ |
| 811 | XO4(i+1,1) \ | 813 | XO4(i + 1, 1) \ |
| 812 | XO4(i+2,2) \ | 814 | XO4(i + 2, 2) \ |
| 813 | XO4(i+3,3) \ | 815 | XO4(i + 3, 3) \ |
| 814 | ST(i,0) \ | 816 | ST(i,0) \ |
| 815 | ST(i+1,1) \ | 817 | ST(i + 1, 1) \ |
| 816 | ST(i+2,2) \ | 818 | ST(i + 2, 2) \ |
| 817 | ST(i+3,3) \ | 819 | ST(i + 3, 3) \ |
| 818 | 820 | ||
| 819 | 821 | ||
| 820 | PF0(0) | 822 | PF0(0) |
| 821 | PF0(2) | 823 | PF0(2) |
| 822 | 824 | ||
| 823 | " .align 32 ;\n" | 825 | " .align 32 ;\n" |
| 824 | " 1: ;\n" | 826 | " 1: ;\n" |
| 825 | 827 | ||
| 826 | BLOCK(0) | 828 | BLOCK(0) |
| 827 | BLOCK(4) | 829 | BLOCK(4) |
| 828 | BLOCK(8) | 830 | BLOCK(8) |
| 829 | BLOCK(12) | 831 | BLOCK(12) |
| 830 | 832 | ||
| 831 | " addl $256, %1 ;\n" | 833 | " addl $256, %1 ;\n" |
| 832 | " addl $256, %2 ;\n" | 834 | " addl $256, %2 ;\n" |
| 833 | " addl $256, %3 ;\n" | 835 | " addl $256, %3 ;\n" |
| 834 | " addl $256, %4 ;\n" | 836 | " addl $256, %4 ;\n" |
| 835 | " addl $256, %5 ;\n" | 837 | " addl $256, %5 ;\n" |
| 836 | " decl %0 ;\n" | 838 | " decl %0 ;\n" |
| 837 | " jnz 1b ;\n" | 839 | " jnz 1b ;\n" |
| 838 | : "+r" (lines), | 840 | : "+r" (lines), |
| 839 | "+r" (p1), "+r" (p2), "+r" (p3) | 841 | "+r" (p1), "+r" (p2), "+r" (p3) |
| 840 | : "r" (p4), "r" (p5) | 842 | : "r" (p4), "r" (p5) |
| @@ -843,17 +845,17 @@ xor_sse_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 843 | /* p4 and p5 were modified, and now the variables are dead. | 845 | /* p4 and p5 were modified, and now the variables are dead. |
| 844 | Clobber them just to be sure nobody does something stupid | 846 | Clobber them just to be sure nobody does something stupid |
| 845 | like assuming they have some legal value. */ | 847 | like assuming they have some legal value. */ |
| 846 | __asm__ ("" : "=r" (p4), "=r" (p5)); | 848 | asm("" : "=r" (p4), "=r" (p5)); |
| 847 | 849 | ||
| 848 | XMMS_RESTORE; | 850 | XMMS_RESTORE; |
| 849 | } | 851 | } |
| 850 | 852 | ||
| 851 | static struct xor_block_template xor_block_pIII_sse = { | 853 | static struct xor_block_template xor_block_pIII_sse = { |
| 852 | .name = "pIII_sse", | 854 | .name = "pIII_sse", |
| 853 | .do_2 = xor_sse_2, | 855 | .do_2 = xor_sse_2, |
| 854 | .do_3 = xor_sse_3, | 856 | .do_3 = xor_sse_3, |
| 855 | .do_4 = xor_sse_4, | 857 | .do_4 = xor_sse_4, |
| 856 | .do_5 = xor_sse_5, | 858 | .do_5 = xor_sse_5, |
| 857 | }; | 859 | }; |
| 858 | 860 | ||
| 859 | /* Also try the generic routines. */ | 861 | /* Also try the generic routines. */ |
| @@ -861,21 +863,21 @@ static struct xor_block_template xor_block_pIII_sse = { | |||
| 861 | 863 | ||
| 862 | #undef XOR_TRY_TEMPLATES | 864 | #undef XOR_TRY_TEMPLATES |
| 863 | #define XOR_TRY_TEMPLATES \ | 865 | #define XOR_TRY_TEMPLATES \ |
| 864 | do { \ | 866 | do { \ |
| 865 | xor_speed(&xor_block_8regs); \ | 867 | xor_speed(&xor_block_8regs); \ |
| 866 | xor_speed(&xor_block_8regs_p); \ | 868 | xor_speed(&xor_block_8regs_p); \ |
| 867 | xor_speed(&xor_block_32regs); \ | 869 | xor_speed(&xor_block_32regs); \ |
| 868 | xor_speed(&xor_block_32regs_p); \ | 870 | xor_speed(&xor_block_32regs_p); \ |
| 869 | if (cpu_has_xmm) \ | 871 | if (cpu_has_xmm) \ |
| 870 | xor_speed(&xor_block_pIII_sse); \ | 872 | xor_speed(&xor_block_pIII_sse); \ |
| 871 | if (cpu_has_mmx) { \ | 873 | if (cpu_has_mmx) { \ |
| 872 | xor_speed(&xor_block_pII_mmx); \ | 874 | xor_speed(&xor_block_pII_mmx); \ |
| 873 | xor_speed(&xor_block_p5_mmx); \ | 875 | xor_speed(&xor_block_p5_mmx); \ |
| 874 | } \ | 876 | } \ |
| 875 | } while (0) | 877 | } while (0) |
| 876 | 878 | ||
| 877 | /* We force the use of the SSE xor block because it can write around L2. | 879 | /* We force the use of the SSE xor block because it can write around L2. |
| 878 | We may also be able to load into the L1 only depending on how the cpu | 880 | We may also be able to load into the L1 only depending on how the cpu |
| 879 | deals with a load to a line that is being prefetched. */ | 881 | deals with a load to a line that is being prefetched. */ |
| 880 | #define XOR_SELECT_TEMPLATE(FASTEST) \ | 882 | #define XOR_SELECT_TEMPLATE(FASTEST) \ |
| 881 | (cpu_has_xmm ? &xor_block_pIII_sse : FASTEST) | 883 | (cpu_has_xmm ? &xor_block_pIII_sse : FASTEST) |
diff --git a/include/asm-x86/xor_64.h b/include/asm-x86/xor_64.h index 1eee7fcb2420..24957e39ac8a 100644 --- a/include/asm-x86/xor_64.h +++ b/include/asm-x86/xor_64.h | |||
| @@ -24,20 +24,23 @@ | |||
| 24 | */ | 24 | */ |
| 25 | 25 | ||
| 26 | /* | 26 | /* |
| 27 | * x86-64 changes / gcc fixes from Andi Kleen. | 27 | * x86-64 changes / gcc fixes from Andi Kleen. |
| 28 | * Copyright 2002 Andi Kleen, SuSE Labs. | 28 | * Copyright 2002 Andi Kleen, SuSE Labs. |
| 29 | * | 29 | * |
| 30 | * This hasn't been optimized for the hammer yet, but there are likely | 30 | * This hasn't been optimized for the hammer yet, but there are likely |
| 31 | * no advantages to be gotten from x86-64 here anyways. | 31 | * no advantages to be gotten from x86-64 here anyways. |
| 32 | */ | 32 | */ |
| 33 | 33 | ||
| 34 | typedef struct { unsigned long a,b; } __attribute__((aligned(16))) xmm_store_t; | 34 | typedef struct { |
| 35 | unsigned long a, b; | ||
| 36 | } __attribute__((aligned(16))) xmm_store_t; | ||
| 35 | 37 | ||
| 36 | /* Doesn't use gcc to save the XMM registers, because there is no easy way to | 38 | /* Doesn't use gcc to save the XMM registers, because there is no easy way to |
| 37 | tell it to do a clts before the register saving. */ | 39 | tell it to do a clts before the register saving. */ |
| 38 | #define XMMS_SAVE do { \ | 40 | #define XMMS_SAVE \ |
| 41 | do { \ | ||
| 39 | preempt_disable(); \ | 42 | preempt_disable(); \ |
| 40 | asm volatile ( \ | 43 | asm volatile( \ |
| 41 | "movq %%cr0,%0 ;\n\t" \ | 44 | "movq %%cr0,%0 ;\n\t" \ |
| 42 | "clts ;\n\t" \ | 45 | "clts ;\n\t" \ |
| 43 | "movups %%xmm0,(%1) ;\n\t" \ | 46 | "movups %%xmm0,(%1) ;\n\t" \ |
| @@ -47,10 +50,11 @@ typedef struct { unsigned long a,b; } __attribute__((aligned(16))) xmm_store_t; | |||
| 47 | : "=&r" (cr0) \ | 50 | : "=&r" (cr0) \ |
| 48 | : "r" (xmm_save) \ | 51 | : "r" (xmm_save) \ |
| 49 | : "memory"); \ | 52 | : "memory"); \ |
| 50 | } while(0) | 53 | } while (0) |
| 51 | 54 | ||
| 52 | #define XMMS_RESTORE do { \ | 55 | #define XMMS_RESTORE \ |
| 53 | asm volatile ( \ | 56 | do { \ |
| 57 | asm volatile( \ | ||
| 54 | "sfence ;\n\t" \ | 58 | "sfence ;\n\t" \ |
| 55 | "movups (%1),%%xmm0 ;\n\t" \ | 59 | "movups (%1),%%xmm0 ;\n\t" \ |
| 56 | "movups 0x10(%1),%%xmm1 ;\n\t" \ | 60 | "movups 0x10(%1),%%xmm1 ;\n\t" \ |
| @@ -61,72 +65,72 @@ typedef struct { unsigned long a,b; } __attribute__((aligned(16))) xmm_store_t; | |||
| 61 | : "r" (cr0), "r" (xmm_save) \ | 65 | : "r" (cr0), "r" (xmm_save) \ |
| 62 | : "memory"); \ | 66 | : "memory"); \ |
| 63 | preempt_enable(); \ | 67 | preempt_enable(); \ |
| 64 | } while(0) | 68 | } while (0) |
| 65 | 69 | ||
| 66 | #define OFFS(x) "16*("#x")" | 70 | #define OFFS(x) "16*("#x")" |
| 67 | #define PF_OFFS(x) "256+16*("#x")" | 71 | #define PF_OFFS(x) "256+16*("#x")" |
| 68 | #define PF0(x) " prefetchnta "PF_OFFS(x)"(%[p1]) ;\n" | 72 | #define PF0(x) " prefetchnta "PF_OFFS(x)"(%[p1]) ;\n" |
| 69 | #define LD(x,y) " movaps "OFFS(x)"(%[p1]), %%xmm"#y" ;\n" | 73 | #define LD(x, y) " movaps "OFFS(x)"(%[p1]), %%xmm"#y" ;\n" |
| 70 | #define ST(x,y) " movaps %%xmm"#y", "OFFS(x)"(%[p1]) ;\n" | 74 | #define ST(x, y) " movaps %%xmm"#y", "OFFS(x)"(%[p1]) ;\n" |
| 71 | #define PF1(x) " prefetchnta "PF_OFFS(x)"(%[p2]) ;\n" | 75 | #define PF1(x) " prefetchnta "PF_OFFS(x)"(%[p2]) ;\n" |
| 72 | #define PF2(x) " prefetchnta "PF_OFFS(x)"(%[p3]) ;\n" | 76 | #define PF2(x) " prefetchnta "PF_OFFS(x)"(%[p3]) ;\n" |
| 73 | #define PF3(x) " prefetchnta "PF_OFFS(x)"(%[p4]) ;\n" | 77 | #define PF3(x) " prefetchnta "PF_OFFS(x)"(%[p4]) ;\n" |
| 74 | #define PF4(x) " prefetchnta "PF_OFFS(x)"(%[p5]) ;\n" | 78 | #define PF4(x) " prefetchnta "PF_OFFS(x)"(%[p5]) ;\n" |
| 75 | #define PF5(x) " prefetchnta "PF_OFFS(x)"(%[p6]) ;\n" | 79 | #define PF5(x) " prefetchnta "PF_OFFS(x)"(%[p6]) ;\n" |
| 76 | #define XO1(x,y) " xorps "OFFS(x)"(%[p2]), %%xmm"#y" ;\n" | 80 | #define XO1(x, y) " xorps "OFFS(x)"(%[p2]), %%xmm"#y" ;\n" |
| 77 | #define XO2(x,y) " xorps "OFFS(x)"(%[p3]), %%xmm"#y" ;\n" | 81 | #define XO2(x, y) " xorps "OFFS(x)"(%[p3]), %%xmm"#y" ;\n" |
| 78 | #define XO3(x,y) " xorps "OFFS(x)"(%[p4]), %%xmm"#y" ;\n" | 82 | #define XO3(x, y) " xorps "OFFS(x)"(%[p4]), %%xmm"#y" ;\n" |
| 79 | #define XO4(x,y) " xorps "OFFS(x)"(%[p5]), %%xmm"#y" ;\n" | 83 | #define XO4(x, y) " xorps "OFFS(x)"(%[p5]), %%xmm"#y" ;\n" |
| 80 | #define XO5(x,y) " xorps "OFFS(x)"(%[p6]), %%xmm"#y" ;\n" | 84 | #define XO5(x, y) " xorps "OFFS(x)"(%[p6]), %%xmm"#y" ;\n" |
| 81 | 85 | ||
| 82 | 86 | ||
| 83 | static void | 87 | static void |
| 84 | xor_sse_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) | 88 | xor_sse_2(unsigned long bytes, unsigned long *p1, unsigned long *p2) |
| 85 | { | 89 | { |
| 86 | unsigned int lines = bytes >> 8; | 90 | unsigned int lines = bytes >> 8; |
| 87 | unsigned long cr0; | 91 | unsigned long cr0; |
| 88 | xmm_store_t xmm_save[4]; | 92 | xmm_store_t xmm_save[4]; |
| 89 | 93 | ||
| 90 | XMMS_SAVE; | 94 | XMMS_SAVE; |
| 91 | 95 | ||
| 92 | asm volatile ( | 96 | asm volatile( |
| 93 | #undef BLOCK | 97 | #undef BLOCK |
| 94 | #define BLOCK(i) \ | 98 | #define BLOCK(i) \ |
| 95 | LD(i,0) \ | 99 | LD(i, 0) \ |
| 96 | LD(i+1,1) \ | 100 | LD(i + 1, 1) \ |
| 97 | PF1(i) \ | 101 | PF1(i) \ |
| 98 | PF1(i+2) \ | 102 | PF1(i + 2) \ |
| 99 | LD(i+2,2) \ | 103 | LD(i + 2, 2) \ |
| 100 | LD(i+3,3) \ | 104 | LD(i + 3, 3) \ |
| 101 | PF0(i+4) \ | 105 | PF0(i + 4) \ |
| 102 | PF0(i+6) \ | 106 | PF0(i + 6) \ |
| 103 | XO1(i,0) \ | 107 | XO1(i, 0) \ |
| 104 | XO1(i+1,1) \ | 108 | XO1(i + 1, 1) \ |
| 105 | XO1(i+2,2) \ | 109 | XO1(i + 2, 2) \ |
| 106 | XO1(i+3,3) \ | 110 | XO1(i + 3, 3) \ |
| 107 | ST(i,0) \ | 111 | ST(i, 0) \ |
| 108 | ST(i+1,1) \ | 112 | ST(i + 1, 1) \ |
| 109 | ST(i+2,2) \ | 113 | ST(i + 2, 2) \ |
| 110 | ST(i+3,3) \ | 114 | ST(i + 3, 3) \ |
| 111 | 115 | ||
| 112 | 116 | ||
| 113 | PF0(0) | 117 | PF0(0) |
| 114 | PF0(2) | 118 | PF0(2) |
| 115 | 119 | ||
| 116 | " .align 32 ;\n" | 120 | " .align 32 ;\n" |
| 117 | " 1: ;\n" | 121 | " 1: ;\n" |
| 118 | 122 | ||
| 119 | BLOCK(0) | 123 | BLOCK(0) |
| 120 | BLOCK(4) | 124 | BLOCK(4) |
| 121 | BLOCK(8) | 125 | BLOCK(8) |
| 122 | BLOCK(12) | 126 | BLOCK(12) |
| 123 | 127 | ||
| 124 | " addq %[inc], %[p1] ;\n" | 128 | " addq %[inc], %[p1] ;\n" |
| 125 | " addq %[inc], %[p2] ;\n" | 129 | " addq %[inc], %[p2] ;\n" |
| 126 | " decl %[cnt] ; jnz 1b" | 130 | " decl %[cnt] ; jnz 1b" |
| 127 | : [p1] "+r" (p1), [p2] "+r" (p2), [cnt] "+r" (lines) | 131 | : [p1] "+r" (p1), [p2] "+r" (p2), [cnt] "+r" (lines) |
| 128 | : [inc] "r" (256UL) | 132 | : [inc] "r" (256UL) |
| 129 | : "memory"); | 133 | : "memory"); |
| 130 | 134 | ||
| 131 | XMMS_RESTORE; | 135 | XMMS_RESTORE; |
| 132 | } | 136 | } |
| @@ -141,52 +145,52 @@ xor_sse_3(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 141 | 145 | ||
| 142 | XMMS_SAVE; | 146 | XMMS_SAVE; |
| 143 | 147 | ||
| 144 | __asm__ __volatile__ ( | 148 | asm volatile( |
| 145 | #undef BLOCK | 149 | #undef BLOCK |
| 146 | #define BLOCK(i) \ | 150 | #define BLOCK(i) \ |
| 147 | PF1(i) \ | 151 | PF1(i) \ |
| 148 | PF1(i+2) \ | 152 | PF1(i + 2) \ |
| 149 | LD(i,0) \ | 153 | LD(i, 0) \ |
| 150 | LD(i+1,1) \ | 154 | LD(i + 1, 1) \ |
| 151 | LD(i+2,2) \ | 155 | LD(i + 2, 2) \ |
| 152 | LD(i+3,3) \ | 156 | LD(i + 3, 3) \ |
| 153 | PF2(i) \ | 157 | PF2(i) \ |
| 154 | PF2(i+2) \ | 158 | PF2(i + 2) \ |
| 155 | PF0(i+4) \ | 159 | PF0(i + 4) \ |
| 156 | PF0(i+6) \ | 160 | PF0(i + 6) \ |
| 157 | XO1(i,0) \ | 161 | XO1(i, 0) \ |
| 158 | XO1(i+1,1) \ | 162 | XO1(i + 1, 1) \ |
| 159 | XO1(i+2,2) \ | 163 | XO1(i + 2, 2) \ |
| 160 | XO1(i+3,3) \ | 164 | XO1(i + 3, 3) \ |
| 161 | XO2(i,0) \ | 165 | XO2(i, 0) \ |
| 162 | XO2(i+1,1) \ | 166 | XO2(i + 1, 1) \ |
| 163 | XO2(i+2,2) \ | 167 | XO2(i + 2, 2) \ |
| 164 | XO2(i+3,3) \ | 168 | XO2(i + 3, 3) \ |
| 165 | ST(i,0) \ | 169 | ST(i, 0) \ |
| 166 | ST(i+1,1) \ | 170 | ST(i + 1, 1) \ |
| 167 | ST(i+2,2) \ | 171 | ST(i + 2, 2) \ |
| 168 | ST(i+3,3) \ | 172 | ST(i + 3, 3) \ |
| 169 | 173 | ||
| 170 | 174 | ||
| 171 | PF0(0) | 175 | PF0(0) |
| 172 | PF0(2) | 176 | PF0(2) |
| 173 | 177 | ||
| 174 | " .align 32 ;\n" | 178 | " .align 32 ;\n" |
| 175 | " 1: ;\n" | 179 | " 1: ;\n" |
| 176 | 180 | ||
| 177 | BLOCK(0) | 181 | BLOCK(0) |
| 178 | BLOCK(4) | 182 | BLOCK(4) |
| 179 | BLOCK(8) | 183 | BLOCK(8) |
| 180 | BLOCK(12) | 184 | BLOCK(12) |
| 181 | 185 | ||
| 182 | " addq %[inc], %[p1] ;\n" | 186 | " addq %[inc], %[p1] ;\n" |
| 183 | " addq %[inc], %[p2] ;\n" | 187 | " addq %[inc], %[p2] ;\n" |
| 184 | " addq %[inc], %[p3] ;\n" | 188 | " addq %[inc], %[p3] ;\n" |
| 185 | " decl %[cnt] ; jnz 1b" | 189 | " decl %[cnt] ; jnz 1b" |
| 186 | : [cnt] "+r" (lines), | 190 | : [cnt] "+r" (lines), |
| 187 | [p1] "+r" (p1), [p2] "+r" (p2), [p3] "+r" (p3) | 191 | [p1] "+r" (p1), [p2] "+r" (p2), [p3] "+r" (p3) |
| 188 | : [inc] "r" (256UL) | 192 | : [inc] "r" (256UL) |
| 189 | : "memory"); | 193 | : "memory"); |
| 190 | XMMS_RESTORE; | 194 | XMMS_RESTORE; |
| 191 | } | 195 | } |
| 192 | 196 | ||
| @@ -195,64 +199,64 @@ xor_sse_4(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 195 | unsigned long *p3, unsigned long *p4) | 199 | unsigned long *p3, unsigned long *p4) |
| 196 | { | 200 | { |
| 197 | unsigned int lines = bytes >> 8; | 201 | unsigned int lines = bytes >> 8; |
| 198 | xmm_store_t xmm_save[4]; | 202 | xmm_store_t xmm_save[4]; |
| 199 | unsigned long cr0; | 203 | unsigned long cr0; |
| 200 | 204 | ||
| 201 | XMMS_SAVE; | 205 | XMMS_SAVE; |
| 202 | 206 | ||
| 203 | __asm__ __volatile__ ( | 207 | asm volatile( |
| 204 | #undef BLOCK | 208 | #undef BLOCK |
| 205 | #define BLOCK(i) \ | 209 | #define BLOCK(i) \ |
| 206 | PF1(i) \ | 210 | PF1(i) \ |
| 207 | PF1(i+2) \ | 211 | PF1(i + 2) \ |
| 208 | LD(i,0) \ | 212 | LD(i, 0) \ |
| 209 | LD(i+1,1) \ | 213 | LD(i + 1, 1) \ |
| 210 | LD(i+2,2) \ | 214 | LD(i + 2, 2) \ |
| 211 | LD(i+3,3) \ | 215 | LD(i + 3, 3) \ |
| 212 | PF2(i) \ | 216 | PF2(i) \ |
| 213 | PF2(i+2) \ | 217 | PF2(i + 2) \ |
| 214 | XO1(i,0) \ | 218 | XO1(i, 0) \ |
| 215 | XO1(i+1,1) \ | 219 | XO1(i + 1, 1) \ |
| 216 | XO1(i+2,2) \ | 220 | XO1(i + 2, 2) \ |
| 217 | XO1(i+3,3) \ | 221 | XO1(i + 3, 3) \ |
| 218 | PF3(i) \ | 222 | PF3(i) \ |
| 219 | PF3(i+2) \ | 223 | PF3(i + 2) \ |
| 220 | PF0(i+4) \ | 224 | PF0(i + 4) \ |
| 221 | PF0(i+6) \ | 225 | PF0(i + 6) \ |
| 222 | XO2(i,0) \ | 226 | XO2(i, 0) \ |
| 223 | XO2(i+1,1) \ | 227 | XO2(i + 1, 1) \ |
| 224 | XO2(i+2,2) \ | 228 | XO2(i + 2, 2) \ |
| 225 | XO2(i+3,3) \ | 229 | XO2(i + 3, 3) \ |
| 226 | XO3(i,0) \ | 230 | XO3(i, 0) \ |
| 227 | XO3(i+1,1) \ | 231 | XO3(i + 1, 1) \ |
| 228 | XO3(i+2,2) \ | 232 | XO3(i + 2, 2) \ |
| 229 | XO3(i+3,3) \ | 233 | XO3(i + 3, 3) \ |
| 230 | ST(i,0) \ | 234 | ST(i, 0) \ |
| 231 | ST(i+1,1) \ | 235 | ST(i + 1, 1) \ |
| 232 | ST(i+2,2) \ | 236 | ST(i + 2, 2) \ |
| 233 | ST(i+3,3) \ | 237 | ST(i + 3, 3) \ |
| 234 | 238 | ||
| 235 | 239 | ||
| 236 | PF0(0) | 240 | PF0(0) |
| 237 | PF0(2) | 241 | PF0(2) |
| 238 | 242 | ||
| 239 | " .align 32 ;\n" | 243 | " .align 32 ;\n" |
| 240 | " 1: ;\n" | 244 | " 1: ;\n" |
| 241 | 245 | ||
| 242 | BLOCK(0) | 246 | BLOCK(0) |
| 243 | BLOCK(4) | 247 | BLOCK(4) |
| 244 | BLOCK(8) | 248 | BLOCK(8) |
| 245 | BLOCK(12) | 249 | BLOCK(12) |
| 246 | 250 | ||
| 247 | " addq %[inc], %[p1] ;\n" | 251 | " addq %[inc], %[p1] ;\n" |
| 248 | " addq %[inc], %[p2] ;\n" | 252 | " addq %[inc], %[p2] ;\n" |
| 249 | " addq %[inc], %[p3] ;\n" | 253 | " addq %[inc], %[p3] ;\n" |
| 250 | " addq %[inc], %[p4] ;\n" | 254 | " addq %[inc], %[p4] ;\n" |
| 251 | " decl %[cnt] ; jnz 1b" | 255 | " decl %[cnt] ; jnz 1b" |
| 252 | : [cnt] "+c" (lines), | 256 | : [cnt] "+c" (lines), |
| 253 | [p1] "+r" (p1), [p2] "+r" (p2), [p3] "+r" (p3), [p4] "+r" (p4) | 257 | [p1] "+r" (p1), [p2] "+r" (p2), [p3] "+r" (p3), [p4] "+r" (p4) |
| 254 | : [inc] "r" (256UL) | 258 | : [inc] "r" (256UL) |
| 255 | : "memory" ); | 259 | : "memory" ); |
| 256 | 260 | ||
| 257 | XMMS_RESTORE; | 261 | XMMS_RESTORE; |
| 258 | } | 262 | } |
| @@ -261,70 +265,70 @@ static void | |||
| 261 | xor_sse_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, | 265 | xor_sse_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, |
| 262 | unsigned long *p3, unsigned long *p4, unsigned long *p5) | 266 | unsigned long *p3, unsigned long *p4, unsigned long *p5) |
| 263 | { | 267 | { |
| 264 | unsigned int lines = bytes >> 8; | 268 | unsigned int lines = bytes >> 8; |
| 265 | xmm_store_t xmm_save[4]; | 269 | xmm_store_t xmm_save[4]; |
| 266 | unsigned long cr0; | 270 | unsigned long cr0; |
| 267 | 271 | ||
| 268 | XMMS_SAVE; | 272 | XMMS_SAVE; |
| 269 | 273 | ||
| 270 | __asm__ __volatile__ ( | 274 | asm volatile( |
| 271 | #undef BLOCK | 275 | #undef BLOCK |
| 272 | #define BLOCK(i) \ | 276 | #define BLOCK(i) \ |
| 273 | PF1(i) \ | 277 | PF1(i) \ |
| 274 | PF1(i+2) \ | 278 | PF1(i + 2) \ |
| 275 | LD(i,0) \ | 279 | LD(i, 0) \ |
| 276 | LD(i+1,1) \ | 280 | LD(i + 1, 1) \ |
| 277 | LD(i+2,2) \ | 281 | LD(i + 2, 2) \ |
| 278 | LD(i+3,3) \ | 282 | LD(i + 3, 3) \ |
| 279 | PF2(i) \ | 283 | PF2(i) \ |
| 280 | PF2(i+2) \ | 284 | PF2(i + 2) \ |
| 281 | XO1(i,0) \ | 285 | XO1(i, 0) \ |
| 282 | XO1(i+1,1) \ | 286 | XO1(i + 1, 1) \ |
| 283 | XO1(i+2,2) \ | 287 | XO1(i + 2, 2) \ |
| 284 | XO1(i+3,3) \ | 288 | XO1(i + 3, 3) \ |
| 285 | PF3(i) \ | 289 | PF3(i) \ |
| 286 | PF3(i+2) \ | 290 | PF3(i + 2) \ |
| 287 | XO2(i,0) \ | 291 | XO2(i, 0) \ |
| 288 | XO2(i+1,1) \ | 292 | XO2(i + 1, 1) \ |
| 289 | XO2(i+2,2) \ | 293 | XO2(i + 2, 2) \ |
| 290 | XO2(i+3,3) \ | 294 | XO2(i + 3, 3) \ |
| 291 | PF4(i) \ | 295 | PF4(i) \ |
| 292 | PF4(i+2) \ | 296 | PF4(i + 2) \ |
| 293 | PF0(i+4) \ | 297 | PF0(i + 4) \ |
| 294 | PF0(i+6) \ | 298 | PF0(i + 6) \ |
| 295 | XO3(i,0) \ | 299 | XO3(i, 0) \ |
| 296 | XO3(i+1,1) \ | 300 | XO3(i + 1, 1) \ |
| 297 | XO3(i+2,2) \ | 301 | XO3(i + 2, 2) \ |
| 298 | XO3(i+3,3) \ | 302 | XO3(i + 3, 3) \ |
| 299 | XO4(i,0) \ | 303 | XO4(i, 0) \ |
| 300 | XO4(i+1,1) \ | 304 | XO4(i + 1, 1) \ |
| 301 | XO4(i+2,2) \ | 305 | XO4(i + 2, 2) \ |
| 302 | XO4(i+3,3) \ | 306 | XO4(i + 3, 3) \ |
| 303 | ST(i,0) \ | 307 | ST(i, 0) \ |
| 304 | ST(i+1,1) \ | 308 | ST(i + 1, 1) \ |
| 305 | ST(i+2,2) \ | 309 | ST(i + 2, 2) \ |
| 306 | ST(i+3,3) \ | 310 | ST(i + 3, 3) \ |
| 307 | 311 | ||
| 308 | 312 | ||
| 309 | PF0(0) | 313 | PF0(0) |
| 310 | PF0(2) | 314 | PF0(2) |
| 311 | 315 | ||
| 312 | " .align 32 ;\n" | 316 | " .align 32 ;\n" |
| 313 | " 1: ;\n" | 317 | " 1: ;\n" |
| 314 | 318 | ||
| 315 | BLOCK(0) | 319 | BLOCK(0) |
| 316 | BLOCK(4) | 320 | BLOCK(4) |
| 317 | BLOCK(8) | 321 | BLOCK(8) |
| 318 | BLOCK(12) | 322 | BLOCK(12) |
| 319 | 323 | ||
| 320 | " addq %[inc], %[p1] ;\n" | 324 | " addq %[inc], %[p1] ;\n" |
| 321 | " addq %[inc], %[p2] ;\n" | 325 | " addq %[inc], %[p2] ;\n" |
| 322 | " addq %[inc], %[p3] ;\n" | 326 | " addq %[inc], %[p3] ;\n" |
| 323 | " addq %[inc], %[p4] ;\n" | 327 | " addq %[inc], %[p4] ;\n" |
| 324 | " addq %[inc], %[p5] ;\n" | 328 | " addq %[inc], %[p5] ;\n" |
| 325 | " decl %[cnt] ; jnz 1b" | 329 | " decl %[cnt] ; jnz 1b" |
| 326 | : [cnt] "+c" (lines), | 330 | : [cnt] "+c" (lines), |
| 327 | [p1] "+r" (p1), [p2] "+r" (p2), [p3] "+r" (p3), [p4] "+r" (p4), | 331 | [p1] "+r" (p1), [p2] "+r" (p2), [p3] "+r" (p3), [p4] "+r" (p4), |
| 328 | [p5] "+r" (p5) | 332 | [p5] "+r" (p5) |
| 329 | : [inc] "r" (256UL) | 333 | : [inc] "r" (256UL) |
| 330 | : "memory"); | 334 | : "memory"); |
| @@ -333,18 +337,18 @@ xor_sse_5(unsigned long bytes, unsigned long *p1, unsigned long *p2, | |||
| 333 | } | 337 | } |
| 334 | 338 | ||
| 335 | static struct xor_block_template xor_block_sse = { | 339 | static struct xor_block_template xor_block_sse = { |
| 336 | .name = "generic_sse", | 340 | .name = "generic_sse", |
| 337 | .do_2 = xor_sse_2, | 341 | .do_2 = xor_sse_2, |
| 338 | .do_3 = xor_sse_3, | 342 | .do_3 = xor_sse_3, |
| 339 | .do_4 = xor_sse_4, | 343 | .do_4 = xor_sse_4, |
| 340 | .do_5 = xor_sse_5, | 344 | .do_5 = xor_sse_5, |
| 341 | }; | 345 | }; |
| 342 | 346 | ||
| 343 | #undef XOR_TRY_TEMPLATES | 347 | #undef XOR_TRY_TEMPLATES |
| 344 | #define XOR_TRY_TEMPLATES \ | 348 | #define XOR_TRY_TEMPLATES \ |
| 345 | do { \ | 349 | do { \ |
| 346 | xor_speed(&xor_block_sse); \ | 350 | xor_speed(&xor_block_sse); \ |
| 347 | } while (0) | 351 | } while (0) |
| 348 | 352 | ||
| 349 | /* We force the use of the SSE xor block because it can write around L2. | 353 | /* We force the use of the SSE xor block because it can write around L2. |
| 350 | We may also be able to load into the L1 only depending on how the cpu | 354 | We may also be able to load into the L1 only depending on how the cpu |
