aboutsummaryrefslogtreecommitdiffstats
path: root/virt/kvm
diff options
context:
space:
mode:
authorMarc Zyngier <marc.zyngier@arm.com>2019-01-20 15:32:31 -0500
committerMarc Zyngier <marc.zyngier@arm.com>2019-02-19 16:05:50 -0500
commitbae561c0cff7e2cde99990ac3b5a58f815d6789d (patch)
tree22c26f029b5513d1de69b42c0faa484fd07cbc04 /virt/kvm
parent793acf870ea3e492c6bb3edb5f8657d9c4f4903f (diff)
KVM: arm/arm64: arch_timer: Mark physical interrupt active when a virtual interrupt is pending
When a guest gets scheduled, KVM performs a "load" operation, which for the timer includes evaluating the virtual "active" state of the interrupt, and replicating it on the physical side. This ensures that the deactivation in the guest will also take place in the physical GIC distributor. If the interrupt is not yet active, we flag it as inactive on the physical side. This means that on restoring the timer registers, if the timer has expired, we'll immediately take an interrupt. That's absolutely fine, as the interrupt will then be flagged as active on the physical side. What this assumes though is that we'll enter the guest right after having taken the interrupt, and that the guest will quickly ACK the interrupt, making it active at on the virtual side. It turns out that quite often, this assumption doesn't really hold. The guest may be preempted on the back on this interrupt, either from kernel space or whilst running at EL1 when a host interrupt fires. When this happens, we repeat the whole sequence on the next load (interrupt marked as inactive, timer registers restored, interrupt fires). And if it takes a really long time for a guest to activate the interrupt (as it does with nested virt), we end-up with many such events in quick succession, leading to the guest only making very slow progress. This can also be seen with the number of virtual timer interrupt on the host being far greater than the same number in the guest. An easy way to fix this is to evaluate the timer state when performing the "load" operation, just like we do when the interrupt actually fires. If the timer has a pending virtual interrupt at this stage, then we can safely flag the physical interrupt as being active, which prevents spurious exits. Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
Diffstat (limited to 'virt/kvm')
-rw-r--r--virt/kvm/arm/arch_timer.c15
1 files changed, 12 insertions, 3 deletions
diff --git a/virt/kvm/arm/arch_timer.c b/virt/kvm/arm/arch_timer.c
index 17f9de73cc8a..af8f2f1d01cc 100644
--- a/virt/kvm/arm/arch_timer.c
+++ b/virt/kvm/arm/arch_timer.c
@@ -487,12 +487,21 @@ static inline void set_timer_irq_phys_active(struct arch_timer_context *ctx, boo
487static void kvm_timer_vcpu_load_gic(struct arch_timer_context *ctx) 487static void kvm_timer_vcpu_load_gic(struct arch_timer_context *ctx)
488{ 488{
489 struct kvm_vcpu *vcpu = ctx->vcpu; 489 struct kvm_vcpu *vcpu = ctx->vcpu;
490 bool phys_active; 490 bool phys_active = false;
491
492 /*
493 * Update the timer output so that it is likely to match the
494 * state we're about to restore. If the timer expires between
495 * this point and the register restoration, we'll take the
496 * interrupt anyway.
497 */
498 kvm_timer_update_irq(ctx->vcpu, kvm_timer_should_fire(ctx), ctx);
491 499
492 if (irqchip_in_kernel(vcpu->kvm)) 500 if (irqchip_in_kernel(vcpu->kvm))
493 phys_active = kvm_vgic_map_is_active(vcpu, ctx->irq.irq); 501 phys_active = kvm_vgic_map_is_active(vcpu, ctx->irq.irq);
494 else 502
495 phys_active = ctx->irq.level; 503 phys_active |= ctx->irq.level;
504
496 set_timer_irq_phys_active(ctx, phys_active); 505 set_timer_irq_phys_active(ctx, phys_active);
497} 506}
498 507