diff options
author | Marc Zyngier <marc.zyngier@arm.com> | 2013-10-08 13:38:13 -0400 |
---|---|---|
committer | Christoffer Dall <christoffer.dall@linaro.org> | 2013-10-15 21:02:05 -0400 |
commit | 1f5580986a3667e9d67b65d916bb4249fd86a400 (patch) | |
tree | b1f2090427f597bcf768ba7170e80bd4f0b50465 | |
parent | a7265fb1751ffbfad553afc7f592a6dac6be48de (diff) |
ARM: KVM: Yield CPU when vcpu executes a WFE
On an (even slightly) oversubscribed system, spinlocks are quickly
becoming a bottleneck, as some vcpus are spinning, waiting for a
lock to be released, while the vcpu holding the lock may not be
running at all.
This creates contention, and the observed slowdown is 40x for
hackbench. No, this isn't a typo.
The solution is to trap blocking WFEs and tell KVM that we're
now spinning. This ensures that other vpus will get a scheduling
boost, allowing the lock to be released more quickly. Also, using
CONFIG_HAVE_KVM_CPU_RELAX_INTERCEPT slightly improves the performance
when the VM is severely overcommited.
Quick test to estimate the performance: hackbench 1 process 1000
2xA15 host (baseline): 1.843s
2xA15 guest w/o patch: 2.083s
4xA15 guest w/o patch: 80.212s
8xA15 guest w/o patch: Could not be bothered to find out
2xA15 guest w/ patch: 2.102s
4xA15 guest w/ patch: 3.205s
8xA15 guest w/ patch: 6.887s
So we go from a 40x degradation to 1.5x in the 2x overcommit case,
which is vaguely more acceptable.
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
Signed-off-by: Christoffer Dall <christoffer.dall@linaro.org>
-rw-r--r-- | arch/arm/include/asm/kvm_arm.h | 4 | ||||
-rw-r--r-- | arch/arm/kvm/Kconfig | 1 | ||||
-rw-r--r-- | arch/arm/kvm/handle_exit.c | 6 |
3 files changed, 9 insertions, 2 deletions
diff --git a/arch/arm/include/asm/kvm_arm.h b/arch/arm/include/asm/kvm_arm.h index d556f03bca17..fe395b7b1ce2 100644 --- a/arch/arm/include/asm/kvm_arm.h +++ b/arch/arm/include/asm/kvm_arm.h | |||
@@ -67,7 +67,7 @@ | |||
67 | */ | 67 | */ |
68 | #define HCR_GUEST_MASK (HCR_TSC | HCR_TSW | HCR_TWI | HCR_VM | HCR_BSU_IS | \ | 68 | #define HCR_GUEST_MASK (HCR_TSC | HCR_TSW | HCR_TWI | HCR_VM | HCR_BSU_IS | \ |
69 | HCR_FB | HCR_TAC | HCR_AMO | HCR_IMO | HCR_FMO | \ | 69 | HCR_FB | HCR_TAC | HCR_AMO | HCR_IMO | HCR_FMO | \ |
70 | HCR_SWIO | HCR_TIDCP) | 70 | HCR_TWE | HCR_SWIO | HCR_TIDCP) |
71 | #define HCR_VIRT_EXCP_MASK (HCR_VA | HCR_VI | HCR_VF) | 71 | #define HCR_VIRT_EXCP_MASK (HCR_VA | HCR_VI | HCR_VF) |
72 | 72 | ||
73 | /* System Control Register (SCTLR) bits */ | 73 | /* System Control Register (SCTLR) bits */ |
@@ -208,6 +208,8 @@ | |||
208 | #define HSR_EC_DABT (0x24) | 208 | #define HSR_EC_DABT (0x24) |
209 | #define HSR_EC_DABT_HYP (0x25) | 209 | #define HSR_EC_DABT_HYP (0x25) |
210 | 210 | ||
211 | #define HSR_WFI_IS_WFE (1U << 0) | ||
212 | |||
211 | #define HSR_HVC_IMM_MASK ((1UL << 16) - 1) | 213 | #define HSR_HVC_IMM_MASK ((1UL << 16) - 1) |
212 | 214 | ||
213 | #define HSR_DABT_S1PTW (1U << 7) | 215 | #define HSR_DABT_S1PTW (1U << 7) |
diff --git a/arch/arm/kvm/Kconfig b/arch/arm/kvm/Kconfig index ebf5015508b5..466bd299b1a8 100644 --- a/arch/arm/kvm/Kconfig +++ b/arch/arm/kvm/Kconfig | |||
@@ -20,6 +20,7 @@ config KVM | |||
20 | bool "Kernel-based Virtual Machine (KVM) support" | 20 | bool "Kernel-based Virtual Machine (KVM) support" |
21 | select PREEMPT_NOTIFIERS | 21 | select PREEMPT_NOTIFIERS |
22 | select ANON_INODES | 22 | select ANON_INODES |
23 | select HAVE_KVM_CPU_RELAX_INTERCEPT | ||
23 | select KVM_MMIO | 24 | select KVM_MMIO |
24 | select KVM_ARM_HOST | 25 | select KVM_ARM_HOST |
25 | depends on ARM_VIRT_EXT && ARM_LPAE | 26 | depends on ARM_VIRT_EXT && ARM_LPAE |
diff --git a/arch/arm/kvm/handle_exit.c b/arch/arm/kvm/handle_exit.c index df4c82d47ad7..c4c496f7619c 100644 --- a/arch/arm/kvm/handle_exit.c +++ b/arch/arm/kvm/handle_exit.c | |||
@@ -84,7 +84,11 @@ static int handle_dabt_hyp(struct kvm_vcpu *vcpu, struct kvm_run *run) | |||
84 | static int kvm_handle_wfi(struct kvm_vcpu *vcpu, struct kvm_run *run) | 84 | static int kvm_handle_wfi(struct kvm_vcpu *vcpu, struct kvm_run *run) |
85 | { | 85 | { |
86 | trace_kvm_wfi(*vcpu_pc(vcpu)); | 86 | trace_kvm_wfi(*vcpu_pc(vcpu)); |
87 | kvm_vcpu_block(vcpu); | 87 | if (kvm_vcpu_get_hsr(vcpu) & HSR_WFI_IS_WFE) |
88 | kvm_vcpu_on_spin(vcpu); | ||
89 | else | ||
90 | kvm_vcpu_block(vcpu); | ||
91 | |||
88 | return 1; | 92 | return 1; |
89 | } | 93 | } |
90 | 94 | ||