diff options
author | Zachary Amsden <zach@vmware.com> | 2007-02-13 07:26:21 -0500 |
---|---|---|
committer | Andi Kleen <andi@basil.nowhere.org> | 2007-02-13 07:26:21 -0500 |
commit | 9226d125d94c7e4964dd41cc5e9ca2ff84091d01 (patch) | |
tree | 935d6e80ff843e1d7b54e0fd9386ef2e0d31aa3d /include/asm-i386 | |
parent | c119ecce894120790903ef535dac3e105f3d6cde (diff) |
[PATCH] i386: paravirt CPU hypercall batching mode
The VMI ROM has a mode where hypercalls can be queued and batched. This turns
out to be a significant win during context switch, but must be done at a
specific point before side effects to CPU state are visible to subsequent
instructions. This is similar to the MMU batching hooks already provided.
The same hooks could be used by the Xen backend to implement a context switch
multicall.
To explain a bit more about lazy modes in the paravirt patches, basically, the
idea is that only one of lazy CPU or MMU mode can be active at any given time.
Lazy MMU mode is similar to this lazy CPU mode, and allows for batching of
multiple PTE updates (say, inside a remap loop), but to avoid keeping some
kind of state machine about when to flush cpu or mmu updates, we just allow
one or the other to be active. Although there is no real reason a more
comprehensive scheme could not be implemented, there is also no demonstrated
need for this extra complexity.
Signed-off-by: Zachary Amsden <zach@vmware.com>
Signed-off-by: Andi Kleen <ak@suse.de>
Cc: Andi Kleen <ak@suse.de>
Cc: Jeremy Fitzhardinge <jeremy@xensource.com>
Cc: Rusty Russell <rusty@rustcorp.com.au>
Cc: Chris Wright <chrisw@sous-sol.org>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Diffstat (limited to 'include/asm-i386')
-rw-r--r-- | include/asm-i386/paravirt.h | 15 |
1 files changed, 15 insertions, 0 deletions
diff --git a/include/asm-i386/paravirt.h b/include/asm-i386/paravirt.h index 53da276a2ec2..38e5164bd0e7 100644 --- a/include/asm-i386/paravirt.h +++ b/include/asm-i386/paravirt.h | |||
@@ -146,6 +146,8 @@ struct paravirt_ops | |||
146 | void (fastcall *pmd_clear)(pmd_t *pmdp); | 146 | void (fastcall *pmd_clear)(pmd_t *pmdp); |
147 | #endif | 147 | #endif |
148 | 148 | ||
149 | void (fastcall *set_lazy_mode)(int mode); | ||
150 | |||
149 | /* These two are jmp to, not actually called. */ | 151 | /* These two are jmp to, not actually called. */ |
150 | void (fastcall *irq_enable_sysexit)(void); | 152 | void (fastcall *irq_enable_sysexit)(void); |
151 | void (fastcall *iret)(void); | 153 | void (fastcall *iret)(void); |
@@ -386,6 +388,19 @@ static inline void pmd_clear(pmd_t *pmdp) | |||
386 | } | 388 | } |
387 | #endif | 389 | #endif |
388 | 390 | ||
391 | /* Lazy mode for batching updates / context switch */ | ||
392 | #define PARAVIRT_LAZY_NONE 0 | ||
393 | #define PARAVIRT_LAZY_MMU 1 | ||
394 | #define PARAVIRT_LAZY_CPU 2 | ||
395 | |||
396 | #define __HAVE_ARCH_ENTER_LAZY_CPU_MODE | ||
397 | #define arch_enter_lazy_cpu_mode() paravirt_ops.set_lazy_mode(PARAVIRT_LAZY_CPU) | ||
398 | #define arch_leave_lazy_cpu_mode() paravirt_ops.set_lazy_mode(PARAVIRT_LAZY_NONE) | ||
399 | |||
400 | #define __HAVE_ARCH_ENTER_LAZY_MMU_MODE | ||
401 | #define arch_enter_lazy_mmu_mode() paravirt_ops.set_lazy_mode(PARAVIRT_LAZY_MMU) | ||
402 | #define arch_leave_lazy_mmu_mode() paravirt_ops.set_lazy_mode(PARAVIRT_LAZY_NONE) | ||
403 | |||
389 | /* These all sit in the .parainstructions section to tell us what to patch. */ | 404 | /* These all sit in the .parainstructions section to tell us what to patch. */ |
390 | struct paravirt_patch { | 405 | struct paravirt_patch { |
391 | u8 *instr; /* original instructions */ | 406 | u8 *instr; /* original instructions */ |