diff options
author | Avi Kivity <avi@qumranet.com> | 2007-09-10 11:10:54 -0400 |
---|---|---|
committer | Avi Kivity <avi@qumranet.com> | 2007-10-13 04:18:28 -0400 |
commit | 04d2cc7780d48a212843e38d46402d97fa1f4774 (patch) | |
tree | a209131bad59abcf574abbaae23145db3c4005e0 /drivers/kvm | |
parent | 29bd8a78082f2d7e2165a735f50b5c716ef3213b (diff) |
KVM: Move main vcpu loop into subarch independent code
This simplifies adding new code as well as reducing overall code size.
Signed-off-by: Avi Kivity <avi@qumranet.com>
Diffstat (limited to 'drivers/kvm')
-rw-r--r-- | drivers/kvm/kvm.h | 9 | ||||
-rw-r--r-- | drivers/kvm/kvm_main.c | 124 | ||||
-rw-r--r-- | drivers/kvm/svm.c | 142 | ||||
-rw-r--r-- | drivers/kvm/vmx.c | 129 |
4 files changed, 187 insertions, 217 deletions
diff --git a/drivers/kvm/kvm.h b/drivers/kvm/kvm.h index 42bb225ad6c1..d93ab48424c6 100644 --- a/drivers/kvm/kvm.h +++ b/drivers/kvm/kvm.h | |||
@@ -453,13 +453,16 @@ struct kvm_x86_ops { | |||
453 | /* Create, but do not attach this VCPU */ | 453 | /* Create, but do not attach this VCPU */ |
454 | struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned id); | 454 | struct kvm_vcpu *(*vcpu_create)(struct kvm *kvm, unsigned id); |
455 | void (*vcpu_free)(struct kvm_vcpu *vcpu); | 455 | void (*vcpu_free)(struct kvm_vcpu *vcpu); |
456 | void (*vcpu_reset)(struct kvm_vcpu *vcpu); | ||
456 | 457 | ||
458 | void (*prepare_guest_switch)(struct kvm_vcpu *vcpu); | ||
457 | void (*vcpu_load)(struct kvm_vcpu *vcpu, int cpu); | 459 | void (*vcpu_load)(struct kvm_vcpu *vcpu, int cpu); |
458 | void (*vcpu_put)(struct kvm_vcpu *vcpu); | 460 | void (*vcpu_put)(struct kvm_vcpu *vcpu); |
459 | void (*vcpu_decache)(struct kvm_vcpu *vcpu); | 461 | void (*vcpu_decache)(struct kvm_vcpu *vcpu); |
460 | 462 | ||
461 | int (*set_guest_debug)(struct kvm_vcpu *vcpu, | 463 | int (*set_guest_debug)(struct kvm_vcpu *vcpu, |
462 | struct kvm_debug_guest *dbg); | 464 | struct kvm_debug_guest *dbg); |
465 | void (*guest_debug_pre)(struct kvm_vcpu *vcpu); | ||
463 | int (*get_msr)(struct kvm_vcpu *vcpu, u32 msr_index, u64 *pdata); | 466 | int (*get_msr)(struct kvm_vcpu *vcpu, u32 msr_index, u64 *pdata); |
464 | int (*set_msr)(struct kvm_vcpu *vcpu, u32 msr_index, u64 data); | 467 | int (*set_msr)(struct kvm_vcpu *vcpu, u32 msr_index, u64 data); |
465 | u64 (*get_segment_base)(struct kvm_vcpu *vcpu, int seg); | 468 | u64 (*get_segment_base)(struct kvm_vcpu *vcpu, int seg); |
@@ -491,12 +494,16 @@ struct kvm_x86_ops { | |||
491 | 494 | ||
492 | void (*inject_gp)(struct kvm_vcpu *vcpu, unsigned err_code); | 495 | void (*inject_gp)(struct kvm_vcpu *vcpu, unsigned err_code); |
493 | 496 | ||
494 | int (*run)(struct kvm_vcpu *vcpu, struct kvm_run *run); | 497 | void (*run)(struct kvm_vcpu *vcpu, struct kvm_run *run); |
498 | int (*handle_exit)(struct kvm_run *run, struct kvm_vcpu *vcpu); | ||
495 | void (*skip_emulated_instruction)(struct kvm_vcpu *vcpu); | 499 | void (*skip_emulated_instruction)(struct kvm_vcpu *vcpu); |
496 | void (*patch_hypercall)(struct kvm_vcpu *vcpu, | 500 | void (*patch_hypercall)(struct kvm_vcpu *vcpu, |
497 | unsigned char *hypercall_addr); | 501 | unsigned char *hypercall_addr); |
498 | int (*get_irq)(struct kvm_vcpu *vcpu); | 502 | int (*get_irq)(struct kvm_vcpu *vcpu); |
499 | void (*set_irq)(struct kvm_vcpu *vcpu, int vec); | 503 | void (*set_irq)(struct kvm_vcpu *vcpu, int vec); |
504 | void (*inject_pending_irq)(struct kvm_vcpu *vcpu); | ||
505 | void (*inject_pending_vectors)(struct kvm_vcpu *vcpu, | ||
506 | struct kvm_run *run); | ||
500 | }; | 507 | }; |
501 | 508 | ||
502 | extern struct kvm_x86_ops *kvm_x86_ops; | 509 | extern struct kvm_x86_ops *kvm_x86_ops; |
diff --git a/drivers/kvm/kvm_main.c b/drivers/kvm/kvm_main.c index 9bfa1bcd26e9..e17b433152cb 100644 --- a/drivers/kvm/kvm_main.c +++ b/drivers/kvm/kvm_main.c | |||
@@ -38,6 +38,7 @@ | |||
38 | #include <linux/cpumask.h> | 38 | #include <linux/cpumask.h> |
39 | #include <linux/smp.h> | 39 | #include <linux/smp.h> |
40 | #include <linux/anon_inodes.h> | 40 | #include <linux/anon_inodes.h> |
41 | #include <linux/profile.h> | ||
41 | 42 | ||
42 | #include <asm/processor.h> | 43 | #include <asm/processor.h> |
43 | #include <asm/msr.h> | 44 | #include <asm/msr.h> |
@@ -1970,6 +1971,127 @@ int kvm_emulate_pio_string(struct kvm_vcpu *vcpu, struct kvm_run *run, int in, | |||
1970 | } | 1971 | } |
1971 | EXPORT_SYMBOL_GPL(kvm_emulate_pio_string); | 1972 | EXPORT_SYMBOL_GPL(kvm_emulate_pio_string); |
1972 | 1973 | ||
1974 | /* | ||
1975 | * Check if userspace requested an interrupt window, and that the | ||
1976 | * interrupt window is open. | ||
1977 | * | ||
1978 | * No need to exit to userspace if we already have an interrupt queued. | ||
1979 | */ | ||
1980 | static int dm_request_for_irq_injection(struct kvm_vcpu *vcpu, | ||
1981 | struct kvm_run *kvm_run) | ||
1982 | { | ||
1983 | return (!vcpu->irq_summary && | ||
1984 | kvm_run->request_interrupt_window && | ||
1985 | vcpu->interrupt_window_open && | ||
1986 | (kvm_x86_ops->get_rflags(vcpu) & X86_EFLAGS_IF)); | ||
1987 | } | ||
1988 | |||
1989 | static void post_kvm_run_save(struct kvm_vcpu *vcpu, | ||
1990 | struct kvm_run *kvm_run) | ||
1991 | { | ||
1992 | kvm_run->if_flag = (kvm_x86_ops->get_rflags(vcpu) & X86_EFLAGS_IF) != 0; | ||
1993 | kvm_run->cr8 = get_cr8(vcpu); | ||
1994 | kvm_run->apic_base = kvm_get_apic_base(vcpu); | ||
1995 | if (irqchip_in_kernel(vcpu->kvm)) | ||
1996 | kvm_run->ready_for_interrupt_injection = 1; | ||
1997 | else | ||
1998 | kvm_run->ready_for_interrupt_injection = | ||
1999 | (vcpu->interrupt_window_open && | ||
2000 | vcpu->irq_summary == 0); | ||
2001 | } | ||
2002 | |||
2003 | static int __vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) | ||
2004 | { | ||
2005 | int r; | ||
2006 | |||
2007 | if (unlikely(vcpu->mp_state == VCPU_MP_STATE_SIPI_RECEIVED)) { | ||
2008 | printk("vcpu %d received sipi with vector # %x\n", | ||
2009 | vcpu->vcpu_id, vcpu->sipi_vector); | ||
2010 | kvm_lapic_reset(vcpu); | ||
2011 | kvm_x86_ops->vcpu_reset(vcpu); | ||
2012 | vcpu->mp_state = VCPU_MP_STATE_RUNNABLE; | ||
2013 | } | ||
2014 | |||
2015 | preempted: | ||
2016 | if (vcpu->guest_debug.enabled) | ||
2017 | kvm_x86_ops->guest_debug_pre(vcpu); | ||
2018 | |||
2019 | again: | ||
2020 | r = kvm_mmu_reload(vcpu); | ||
2021 | if (unlikely(r)) | ||
2022 | goto out; | ||
2023 | |||
2024 | preempt_disable(); | ||
2025 | |||
2026 | kvm_x86_ops->prepare_guest_switch(vcpu); | ||
2027 | kvm_load_guest_fpu(vcpu); | ||
2028 | |||
2029 | local_irq_disable(); | ||
2030 | |||
2031 | if (signal_pending(current)) { | ||
2032 | local_irq_enable(); | ||
2033 | preempt_enable(); | ||
2034 | r = -EINTR; | ||
2035 | kvm_run->exit_reason = KVM_EXIT_INTR; | ||
2036 | ++vcpu->stat.signal_exits; | ||
2037 | goto out; | ||
2038 | } | ||
2039 | |||
2040 | if (irqchip_in_kernel(vcpu->kvm)) | ||
2041 | kvm_x86_ops->inject_pending_irq(vcpu); | ||
2042 | else if (!vcpu->mmio_read_completed) | ||
2043 | kvm_x86_ops->inject_pending_vectors(vcpu, kvm_run); | ||
2044 | |||
2045 | vcpu->guest_mode = 1; | ||
2046 | |||
2047 | if (vcpu->requests) | ||
2048 | if (test_and_clear_bit(KVM_TLB_FLUSH, &vcpu->requests)) | ||
2049 | kvm_x86_ops->tlb_flush(vcpu); | ||
2050 | |||
2051 | kvm_x86_ops->run(vcpu, kvm_run); | ||
2052 | |||
2053 | vcpu->guest_mode = 0; | ||
2054 | local_irq_enable(); | ||
2055 | |||
2056 | ++vcpu->stat.exits; | ||
2057 | |||
2058 | preempt_enable(); | ||
2059 | |||
2060 | /* | ||
2061 | * Profile KVM exit RIPs: | ||
2062 | */ | ||
2063 | if (unlikely(prof_on == KVM_PROFILING)) { | ||
2064 | kvm_x86_ops->cache_regs(vcpu); | ||
2065 | profile_hit(KVM_PROFILING, (void *)vcpu->rip); | ||
2066 | } | ||
2067 | |||
2068 | r = kvm_x86_ops->handle_exit(kvm_run, vcpu); | ||
2069 | |||
2070 | if (r > 0) { | ||
2071 | if (dm_request_for_irq_injection(vcpu, kvm_run)) { | ||
2072 | r = -EINTR; | ||
2073 | kvm_run->exit_reason = KVM_EXIT_INTR; | ||
2074 | ++vcpu->stat.request_irq_exits; | ||
2075 | goto out; | ||
2076 | } | ||
2077 | if (!need_resched()) { | ||
2078 | ++vcpu->stat.light_exits; | ||
2079 | goto again; | ||
2080 | } | ||
2081 | } | ||
2082 | |||
2083 | out: | ||
2084 | if (r > 0) { | ||
2085 | kvm_resched(vcpu); | ||
2086 | goto preempted; | ||
2087 | } | ||
2088 | |||
2089 | post_kvm_run_save(vcpu, kvm_run); | ||
2090 | |||
2091 | return r; | ||
2092 | } | ||
2093 | |||
2094 | |||
1973 | static int kvm_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) | 2095 | static int kvm_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) |
1974 | { | 2096 | { |
1975 | int r; | 2097 | int r; |
@@ -2017,7 +2139,7 @@ static int kvm_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) | |||
2017 | kvm_x86_ops->decache_regs(vcpu); | 2139 | kvm_x86_ops->decache_regs(vcpu); |
2018 | } | 2140 | } |
2019 | 2141 | ||
2020 | r = kvm_x86_ops->run(vcpu, kvm_run); | 2142 | r = __vcpu_run(vcpu, kvm_run); |
2021 | 2143 | ||
2022 | out: | 2144 | out: |
2023 | if (vcpu->sigset_active) | 2145 | if (vcpu->sigset_active) |
diff --git a/drivers/kvm/svm.c b/drivers/kvm/svm.c index 7b22d396c149..95681ea16382 100644 --- a/drivers/kvm/svm.c +++ b/drivers/kvm/svm.c | |||
@@ -22,7 +22,6 @@ | |||
22 | #include <linux/kernel.h> | 22 | #include <linux/kernel.h> |
23 | #include <linux/vmalloc.h> | 23 | #include <linux/vmalloc.h> |
24 | #include <linux/highmem.h> | 24 | #include <linux/highmem.h> |
25 | #include <linux/profile.h> | ||
26 | #include <linux/sched.h> | 25 | #include <linux/sched.h> |
27 | 26 | ||
28 | #include <asm/desc.h> | 27 | #include <asm/desc.h> |
@@ -50,6 +49,8 @@ MODULE_LICENSE("GPL"); | |||
50 | #define SVM_FEATURE_LBRV (1 << 1) | 49 | #define SVM_FEATURE_LBRV (1 << 1) |
51 | #define SVM_DEATURE_SVML (1 << 2) | 50 | #define SVM_DEATURE_SVML (1 << 2) |
52 | 51 | ||
52 | static void kvm_reput_irq(struct vcpu_svm *svm); | ||
53 | |||
53 | static inline struct vcpu_svm *to_svm(struct kvm_vcpu *vcpu) | 54 | static inline struct vcpu_svm *to_svm(struct kvm_vcpu *vcpu) |
54 | { | 55 | { |
55 | return container_of(vcpu, struct vcpu_svm, vcpu); | 56 | return container_of(vcpu, struct vcpu_svm, vcpu); |
@@ -555,6 +556,13 @@ static void init_vmcb(struct vmcb *vmcb) | |||
555 | /* rdx = ?? */ | 556 | /* rdx = ?? */ |
556 | } | 557 | } |
557 | 558 | ||
559 | static void svm_vcpu_reset(struct kvm_vcpu *vcpu) | ||
560 | { | ||
561 | struct vcpu_svm *svm = to_svm(vcpu); | ||
562 | |||
563 | init_vmcb(svm->vmcb); | ||
564 | } | ||
565 | |||
558 | static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) | 566 | static struct kvm_vcpu *svm_create_vcpu(struct kvm *kvm, unsigned int id) |
559 | { | 567 | { |
560 | struct vcpu_svm *svm; | 568 | struct vcpu_svm *svm; |
@@ -1252,10 +1260,20 @@ static int (*svm_exit_handlers[])(struct vcpu_svm *svm, | |||
1252 | }; | 1260 | }; |
1253 | 1261 | ||
1254 | 1262 | ||
1255 | static int handle_exit(struct vcpu_svm *svm, struct kvm_run *kvm_run) | 1263 | static int handle_exit(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu) |
1256 | { | 1264 | { |
1265 | struct vcpu_svm *svm = to_svm(vcpu); | ||
1257 | u32 exit_code = svm->vmcb->control.exit_code; | 1266 | u32 exit_code = svm->vmcb->control.exit_code; |
1258 | 1267 | ||
1268 | kvm_reput_irq(svm); | ||
1269 | |||
1270 | if (svm->vmcb->control.exit_code == SVM_EXIT_ERR) { | ||
1271 | kvm_run->exit_reason = KVM_EXIT_FAIL_ENTRY; | ||
1272 | kvm_run->fail_entry.hardware_entry_failure_reason | ||
1273 | = svm->vmcb->control.exit_code; | ||
1274 | return 0; | ||
1275 | } | ||
1276 | |||
1259 | if (is_external_interrupt(svm->vmcb->control.exit_int_info) && | 1277 | if (is_external_interrupt(svm->vmcb->control.exit_int_info) && |
1260 | exit_code != SVM_EXIT_EXCP_BASE + PF_VECTOR) | 1278 | exit_code != SVM_EXIT_EXCP_BASE + PF_VECTOR) |
1261 | printk(KERN_ERR "%s: unexpected exit_ini_info 0x%x " | 1279 | printk(KERN_ERR "%s: unexpected exit_ini_info 0x%x " |
@@ -1313,11 +1331,11 @@ static void svm_set_irq(struct kvm_vcpu *vcpu, int irq) | |||
1313 | svm_inject_irq(svm, irq); | 1331 | svm_inject_irq(svm, irq); |
1314 | } | 1332 | } |
1315 | 1333 | ||
1316 | static void svm_intr_assist(struct vcpu_svm *svm) | 1334 | static void svm_intr_assist(struct kvm_vcpu *vcpu) |
1317 | { | 1335 | { |
1336 | struct vcpu_svm *svm = to_svm(vcpu); | ||
1318 | struct vmcb *vmcb = svm->vmcb; | 1337 | struct vmcb *vmcb = svm->vmcb; |
1319 | int intr_vector = -1; | 1338 | int intr_vector = -1; |
1320 | struct kvm_vcpu *vcpu = &svm->vcpu; | ||
1321 | 1339 | ||
1322 | kvm_inject_pending_timer_irqs(vcpu); | 1340 | kvm_inject_pending_timer_irqs(vcpu); |
1323 | if ((vmcb->control.exit_int_info & SVM_EVTINJ_VALID) && | 1341 | if ((vmcb->control.exit_int_info & SVM_EVTINJ_VALID) && |
@@ -1376,9 +1394,10 @@ static void svm_do_inject_vector(struct vcpu_svm *svm) | |||
1376 | svm_inject_irq(svm, irq); | 1394 | svm_inject_irq(svm, irq); |
1377 | } | 1395 | } |
1378 | 1396 | ||
1379 | static void do_interrupt_requests(struct vcpu_svm *svm, | 1397 | static void do_interrupt_requests(struct kvm_vcpu *vcpu, |
1380 | struct kvm_run *kvm_run) | 1398 | struct kvm_run *kvm_run) |
1381 | { | 1399 | { |
1400 | struct vcpu_svm *svm = to_svm(vcpu); | ||
1382 | struct vmcb_control_area *control = &svm->vmcb->control; | 1401 | struct vmcb_control_area *control = &svm->vmcb->control; |
1383 | 1402 | ||
1384 | svm->vcpu.interrupt_window_open = | 1403 | svm->vcpu.interrupt_window_open = |
@@ -1401,35 +1420,6 @@ static void do_interrupt_requests(struct vcpu_svm *svm, | |||
1401 | control->intercept &= ~(1ULL << INTERCEPT_VINTR); | 1420 | control->intercept &= ~(1ULL << INTERCEPT_VINTR); |
1402 | } | 1421 | } |
1403 | 1422 | ||
1404 | static void post_kvm_run_save(struct vcpu_svm *svm, | ||
1405 | struct kvm_run *kvm_run) | ||
1406 | { | ||
1407 | if (irqchip_in_kernel(svm->vcpu.kvm)) | ||
1408 | kvm_run->ready_for_interrupt_injection = 1; | ||
1409 | else | ||
1410 | kvm_run->ready_for_interrupt_injection = | ||
1411 | (svm->vcpu.interrupt_window_open && | ||
1412 | svm->vcpu.irq_summary == 0); | ||
1413 | kvm_run->if_flag = (svm->vmcb->save.rflags & X86_EFLAGS_IF) != 0; | ||
1414 | kvm_run->cr8 = get_cr8(&svm->vcpu); | ||
1415 | kvm_run->apic_base = kvm_get_apic_base(&svm->vcpu); | ||
1416 | } | ||
1417 | |||
1418 | /* | ||
1419 | * Check if userspace requested an interrupt window, and that the | ||
1420 | * interrupt window is open. | ||
1421 | * | ||
1422 | * No need to exit to userspace if we already have an interrupt queued. | ||
1423 | */ | ||
1424 | static int dm_request_for_irq_injection(struct vcpu_svm *svm, | ||
1425 | struct kvm_run *kvm_run) | ||
1426 | { | ||
1427 | return (!svm->vcpu.irq_summary && | ||
1428 | kvm_run->request_interrupt_window && | ||
1429 | svm->vcpu.interrupt_window_open && | ||
1430 | (svm->vmcb->save.rflags & X86_EFLAGS_IF)); | ||
1431 | } | ||
1432 | |||
1433 | static void save_db_regs(unsigned long *db_regs) | 1423 | static void save_db_regs(unsigned long *db_regs) |
1434 | { | 1424 | { |
1435 | asm volatile ("mov %%dr0, %0" : "=r"(db_regs[0])); | 1425 | asm volatile ("mov %%dr0, %0" : "=r"(db_regs[0])); |
@@ -1451,38 +1441,16 @@ static void svm_flush_tlb(struct kvm_vcpu *vcpu) | |||
1451 | force_new_asid(vcpu); | 1441 | force_new_asid(vcpu); |
1452 | } | 1442 | } |
1453 | 1443 | ||
1454 | static int svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) | 1444 | static void svm_prepare_guest_switch(struct kvm_vcpu *vcpu) |
1445 | { | ||
1446 | } | ||
1447 | |||
1448 | static void svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) | ||
1455 | { | 1449 | { |
1456 | struct vcpu_svm *svm = to_svm(vcpu); | 1450 | struct vcpu_svm *svm = to_svm(vcpu); |
1457 | u16 fs_selector; | 1451 | u16 fs_selector; |
1458 | u16 gs_selector; | 1452 | u16 gs_selector; |
1459 | u16 ldt_selector; | 1453 | u16 ldt_selector; |
1460 | int r; | ||
1461 | |||
1462 | again: | ||
1463 | r = kvm_mmu_reload(vcpu); | ||
1464 | if (unlikely(r)) | ||
1465 | return r; | ||
1466 | |||
1467 | clgi(); | ||
1468 | |||
1469 | if (signal_pending(current)) { | ||
1470 | stgi(); | ||
1471 | ++vcpu->stat.signal_exits; | ||
1472 | post_kvm_run_save(svm, kvm_run); | ||
1473 | kvm_run->exit_reason = KVM_EXIT_INTR; | ||
1474 | return -EINTR; | ||
1475 | } | ||
1476 | |||
1477 | if (irqchip_in_kernel(vcpu->kvm)) | ||
1478 | svm_intr_assist(svm); | ||
1479 | else if (!vcpu->mmio_read_completed) | ||
1480 | do_interrupt_requests(svm, kvm_run); | ||
1481 | |||
1482 | vcpu->guest_mode = 1; | ||
1483 | if (vcpu->requests) | ||
1484 | if (test_and_clear_bit(KVM_TLB_FLUSH, &vcpu->requests)) | ||
1485 | svm_flush_tlb(vcpu); | ||
1486 | 1454 | ||
1487 | pre_svm_run(svm); | 1455 | pre_svm_run(svm); |
1488 | 1456 | ||
@@ -1501,10 +1469,9 @@ again: | |||
1501 | load_db_regs(svm->db_regs); | 1469 | load_db_regs(svm->db_regs); |
1502 | } | 1470 | } |
1503 | 1471 | ||
1504 | if (vcpu->fpu_active) { | 1472 | clgi(); |
1505 | fx_save(&vcpu->host_fx_image); | 1473 | |
1506 | fx_restore(&vcpu->guest_fx_image); | 1474 | local_irq_enable(); |
1507 | } | ||
1508 | 1475 | ||
1509 | asm volatile ( | 1476 | asm volatile ( |
1510 | #ifdef CONFIG_X86_64 | 1477 | #ifdef CONFIG_X86_64 |
@@ -1612,12 +1579,9 @@ again: | |||
1612 | #endif | 1579 | #endif |
1613 | : "cc", "memory" ); | 1580 | : "cc", "memory" ); |
1614 | 1581 | ||
1615 | vcpu->guest_mode = 0; | 1582 | local_irq_disable(); |
1616 | 1583 | ||
1617 | if (vcpu->fpu_active) { | 1584 | stgi(); |
1618 | fx_save(&vcpu->guest_fx_image); | ||
1619 | fx_restore(&vcpu->host_fx_image); | ||
1620 | } | ||
1621 | 1585 | ||
1622 | if ((svm->vmcb->save.dr7 & 0xff)) | 1586 | if ((svm->vmcb->save.dr7 & 0xff)) |
1623 | load_db_regs(svm->host_db_regs); | 1587 | load_db_regs(svm->host_db_regs); |
@@ -1635,40 +1599,7 @@ again: | |||
1635 | 1599 | ||
1636 | reload_tss(vcpu); | 1600 | reload_tss(vcpu); |
1637 | 1601 | ||
1638 | /* | ||
1639 | * Profile KVM exit RIPs: | ||
1640 | */ | ||
1641 | if (unlikely(prof_on == KVM_PROFILING)) | ||
1642 | profile_hit(KVM_PROFILING, | ||
1643 | (void *)(unsigned long)svm->vmcb->save.rip); | ||
1644 | |||
1645 | stgi(); | ||
1646 | |||
1647 | kvm_reput_irq(svm); | ||
1648 | |||
1649 | svm->next_rip = 0; | 1602 | svm->next_rip = 0; |
1650 | |||
1651 | if (svm->vmcb->control.exit_code == SVM_EXIT_ERR) { | ||
1652 | kvm_run->exit_reason = KVM_EXIT_FAIL_ENTRY; | ||
1653 | kvm_run->fail_entry.hardware_entry_failure_reason | ||
1654 | = svm->vmcb->control.exit_code; | ||
1655 | post_kvm_run_save(svm, kvm_run); | ||
1656 | return 0; | ||
1657 | } | ||
1658 | |||
1659 | r = handle_exit(svm, kvm_run); | ||
1660 | if (r > 0) { | ||
1661 | if (dm_request_for_irq_injection(svm, kvm_run)) { | ||
1662 | ++vcpu->stat.request_irq_exits; | ||
1663 | post_kvm_run_save(svm, kvm_run); | ||
1664 | kvm_run->exit_reason = KVM_EXIT_INTR; | ||
1665 | return -EINTR; | ||
1666 | } | ||
1667 | kvm_resched(vcpu); | ||
1668 | goto again; | ||
1669 | } | ||
1670 | post_kvm_run_save(svm, kvm_run); | ||
1671 | return r; | ||
1672 | } | 1603 | } |
1673 | 1604 | ||
1674 | static void svm_set_cr3(struct kvm_vcpu *vcpu, unsigned long root) | 1605 | static void svm_set_cr3(struct kvm_vcpu *vcpu, unsigned long root) |
@@ -1752,7 +1683,9 @@ static struct kvm_x86_ops svm_x86_ops = { | |||
1752 | 1683 | ||
1753 | .vcpu_create = svm_create_vcpu, | 1684 | .vcpu_create = svm_create_vcpu, |
1754 | .vcpu_free = svm_free_vcpu, | 1685 | .vcpu_free = svm_free_vcpu, |
1686 | .vcpu_reset = svm_vcpu_reset, | ||
1755 | 1687 | ||
1688 | .prepare_guest_switch = svm_prepare_guest_switch, | ||
1756 | .vcpu_load = svm_vcpu_load, | 1689 | .vcpu_load = svm_vcpu_load, |
1757 | .vcpu_put = svm_vcpu_put, | 1690 | .vcpu_put = svm_vcpu_put, |
1758 | .vcpu_decache = svm_vcpu_decache, | 1691 | .vcpu_decache = svm_vcpu_decache, |
@@ -1786,10 +1719,13 @@ static struct kvm_x86_ops svm_x86_ops = { | |||
1786 | .inject_gp = svm_inject_gp, | 1719 | .inject_gp = svm_inject_gp, |
1787 | 1720 | ||
1788 | .run = svm_vcpu_run, | 1721 | .run = svm_vcpu_run, |
1722 | .handle_exit = handle_exit, | ||
1789 | .skip_emulated_instruction = skip_emulated_instruction, | 1723 | .skip_emulated_instruction = skip_emulated_instruction, |
1790 | .patch_hypercall = svm_patch_hypercall, | 1724 | .patch_hypercall = svm_patch_hypercall, |
1791 | .get_irq = svm_get_irq, | 1725 | .get_irq = svm_get_irq, |
1792 | .set_irq = svm_set_irq, | 1726 | .set_irq = svm_set_irq, |
1727 | .inject_pending_irq = svm_intr_assist, | ||
1728 | .inject_pending_vectors = do_interrupt_requests, | ||
1793 | }; | 1729 | }; |
1794 | 1730 | ||
1795 | static int __init svm_init(void) | 1731 | static int __init svm_init(void) |
diff --git a/drivers/kvm/vmx.c b/drivers/kvm/vmx.c index 713f78a89595..fa4277d520ca 100644 --- a/drivers/kvm/vmx.c +++ b/drivers/kvm/vmx.c | |||
@@ -25,7 +25,6 @@ | |||
25 | #include <linux/kernel.h> | 25 | #include <linux/kernel.h> |
26 | #include <linux/mm.h> | 26 | #include <linux/mm.h> |
27 | #include <linux/highmem.h> | 27 | #include <linux/highmem.h> |
28 | #include <linux/profile.h> | ||
29 | #include <linux/sched.h> | 28 | #include <linux/sched.h> |
30 | 29 | ||
31 | #include <asm/io.h> | 30 | #include <asm/io.h> |
@@ -355,8 +354,10 @@ static void load_transition_efer(struct vcpu_vmx *vmx) | |||
355 | vmx->vcpu.stat.efer_reload++; | 354 | vmx->vcpu.stat.efer_reload++; |
356 | } | 355 | } |
357 | 356 | ||
358 | static void vmx_save_host_state(struct vcpu_vmx *vmx) | 357 | static void vmx_save_host_state(struct kvm_vcpu *vcpu) |
359 | { | 358 | { |
359 | struct vcpu_vmx *vmx = to_vmx(vcpu); | ||
360 | |||
360 | if (vmx->host_state.loaded) | 361 | if (vmx->host_state.loaded) |
361 | return; | 362 | return; |
362 | 363 | ||
@@ -1598,6 +1599,13 @@ out: | |||
1598 | return ret; | 1599 | return ret; |
1599 | } | 1600 | } |
1600 | 1601 | ||
1602 | static void vmx_vcpu_reset(struct kvm_vcpu *vcpu) | ||
1603 | { | ||
1604 | struct vcpu_vmx *vmx = to_vmx(vcpu); | ||
1605 | |||
1606 | vmx_vcpu_setup(vmx); | ||
1607 | } | ||
1608 | |||
1601 | static void inject_rmode_irq(struct kvm_vcpu *vcpu, int irq) | 1609 | static void inject_rmode_irq(struct kvm_vcpu *vcpu, int irq) |
1602 | { | 1610 | { |
1603 | u16 ent[2]; | 1611 | u16 ent[2]; |
@@ -2019,20 +2027,6 @@ static int handle_tpr_below_threshold(struct kvm_vcpu *vcpu, | |||
2019 | return 1; | 2027 | return 1; |
2020 | } | 2028 | } |
2021 | 2029 | ||
2022 | static void post_kvm_run_save(struct kvm_vcpu *vcpu, | ||
2023 | struct kvm_run *kvm_run) | ||
2024 | { | ||
2025 | kvm_run->if_flag = (vmcs_readl(GUEST_RFLAGS) & X86_EFLAGS_IF) != 0; | ||
2026 | kvm_run->cr8 = get_cr8(vcpu); | ||
2027 | kvm_run->apic_base = kvm_get_apic_base(vcpu); | ||
2028 | if (irqchip_in_kernel(vcpu->kvm)) | ||
2029 | kvm_run->ready_for_interrupt_injection = 1; | ||
2030 | else | ||
2031 | kvm_run->ready_for_interrupt_injection = | ||
2032 | (vcpu->interrupt_window_open && | ||
2033 | vcpu->irq_summary == 0); | ||
2034 | } | ||
2035 | |||
2036 | static int handle_interrupt_window(struct kvm_vcpu *vcpu, | 2030 | static int handle_interrupt_window(struct kvm_vcpu *vcpu, |
2037 | struct kvm_run *kvm_run) | 2031 | struct kvm_run *kvm_run) |
2038 | { | 2032 | { |
@@ -2123,21 +2117,6 @@ static int kvm_handle_exit(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu) | |||
2123 | return 0; | 2117 | return 0; |
2124 | } | 2118 | } |
2125 | 2119 | ||
2126 | /* | ||
2127 | * Check if userspace requested an interrupt window, and that the | ||
2128 | * interrupt window is open. | ||
2129 | * | ||
2130 | * No need to exit to userspace if we already have an interrupt queued. | ||
2131 | */ | ||
2132 | static int dm_request_for_irq_injection(struct kvm_vcpu *vcpu, | ||
2133 | struct kvm_run *kvm_run) | ||
2134 | { | ||
2135 | return (!vcpu->irq_summary && | ||
2136 | kvm_run->request_interrupt_window && | ||
2137 | vcpu->interrupt_window_open && | ||
2138 | (vmcs_readl(GUEST_RFLAGS) & X86_EFLAGS_IF)); | ||
2139 | } | ||
2140 | |||
2141 | static void vmx_flush_tlb(struct kvm_vcpu *vcpu) | 2120 | static void vmx_flush_tlb(struct kvm_vcpu *vcpu) |
2142 | { | 2121 | { |
2143 | } | 2122 | } |
@@ -2214,59 +2193,15 @@ static void vmx_intr_assist(struct kvm_vcpu *vcpu) | |||
2214 | enable_irq_window(vcpu); | 2193 | enable_irq_window(vcpu); |
2215 | } | 2194 | } |
2216 | 2195 | ||
2217 | static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) | 2196 | static void vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) |
2218 | { | 2197 | { |
2219 | struct vcpu_vmx *vmx = to_vmx(vcpu); | 2198 | struct vcpu_vmx *vmx = to_vmx(vcpu); |
2220 | int r; | ||
2221 | |||
2222 | if (unlikely(vcpu->mp_state == VCPU_MP_STATE_SIPI_RECEIVED)) { | ||
2223 | printk("vcpu %d received sipi with vector # %x\n", | ||
2224 | vcpu->vcpu_id, vcpu->sipi_vector); | ||
2225 | kvm_lapic_reset(vcpu); | ||
2226 | vmx_vcpu_setup(vmx); | ||
2227 | vcpu->mp_state = VCPU_MP_STATE_RUNNABLE; | ||
2228 | } | ||
2229 | |||
2230 | preempted: | ||
2231 | if (vcpu->guest_debug.enabled) | ||
2232 | kvm_guest_debug_pre(vcpu); | ||
2233 | |||
2234 | again: | ||
2235 | r = kvm_mmu_reload(vcpu); | ||
2236 | if (unlikely(r)) | ||
2237 | goto out; | ||
2238 | |||
2239 | preempt_disable(); | ||
2240 | |||
2241 | vmx_save_host_state(vmx); | ||
2242 | kvm_load_guest_fpu(vcpu); | ||
2243 | 2199 | ||
2244 | /* | 2200 | /* |
2245 | * Loading guest fpu may have cleared host cr0.ts | 2201 | * Loading guest fpu may have cleared host cr0.ts |
2246 | */ | 2202 | */ |
2247 | vmcs_writel(HOST_CR0, read_cr0()); | 2203 | vmcs_writel(HOST_CR0, read_cr0()); |
2248 | 2204 | ||
2249 | local_irq_disable(); | ||
2250 | |||
2251 | if (signal_pending(current)) { | ||
2252 | local_irq_enable(); | ||
2253 | preempt_enable(); | ||
2254 | r = -EINTR; | ||
2255 | kvm_run->exit_reason = KVM_EXIT_INTR; | ||
2256 | ++vcpu->stat.signal_exits; | ||
2257 | goto out; | ||
2258 | } | ||
2259 | |||
2260 | if (irqchip_in_kernel(vcpu->kvm)) | ||
2261 | vmx_intr_assist(vcpu); | ||
2262 | else if (!vcpu->mmio_read_completed) | ||
2263 | do_interrupt_requests(vcpu, kvm_run); | ||
2264 | |||
2265 | vcpu->guest_mode = 1; | ||
2266 | if (vcpu->requests) | ||
2267 | if (test_and_clear_bit(KVM_TLB_FLUSH, &vcpu->requests)) | ||
2268 | vmx_flush_tlb(vcpu); | ||
2269 | |||
2270 | asm ( | 2205 | asm ( |
2271 | /* Store host registers */ | 2206 | /* Store host registers */ |
2272 | #ifdef CONFIG_X86_64 | 2207 | #ifdef CONFIG_X86_64 |
@@ -2383,46 +2318,10 @@ again: | |||
2383 | [cr2]"i"(offsetof(struct kvm_vcpu, cr2)) | 2318 | [cr2]"i"(offsetof(struct kvm_vcpu, cr2)) |
2384 | : "cc", "memory" ); | 2319 | : "cc", "memory" ); |
2385 | 2320 | ||
2386 | vcpu->guest_mode = 0; | ||
2387 | local_irq_enable(); | ||
2388 | |||
2389 | ++vcpu->stat.exits; | ||
2390 | |||
2391 | vcpu->interrupt_window_open = (vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) & 3) == 0; | 2321 | vcpu->interrupt_window_open = (vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) & 3) == 0; |
2392 | 2322 | ||
2393 | asm ("mov %0, %%ds; mov %0, %%es" : : "r"(__USER_DS)); | 2323 | asm ("mov %0, %%ds; mov %0, %%es" : : "r"(__USER_DS)); |
2394 | vmx->launched = 1; | 2324 | vmx->launched = 1; |
2395 | |||
2396 | preempt_enable(); | ||
2397 | |||
2398 | /* | ||
2399 | * Profile KVM exit RIPs: | ||
2400 | */ | ||
2401 | if (unlikely(prof_on == KVM_PROFILING)) | ||
2402 | profile_hit(KVM_PROFILING, (void *)vmcs_readl(GUEST_RIP)); | ||
2403 | |||
2404 | r = kvm_handle_exit(kvm_run, vcpu); | ||
2405 | if (r > 0) { | ||
2406 | if (dm_request_for_irq_injection(vcpu, kvm_run)) { | ||
2407 | r = -EINTR; | ||
2408 | kvm_run->exit_reason = KVM_EXIT_INTR; | ||
2409 | ++vcpu->stat.request_irq_exits; | ||
2410 | goto out; | ||
2411 | } | ||
2412 | if (!need_resched()) { | ||
2413 | ++vcpu->stat.light_exits; | ||
2414 | goto again; | ||
2415 | } | ||
2416 | } | ||
2417 | |||
2418 | out: | ||
2419 | if (r > 0) { | ||
2420 | kvm_resched(vcpu); | ||
2421 | goto preempted; | ||
2422 | } | ||
2423 | |||
2424 | post_kvm_run_save(vcpu, kvm_run); | ||
2425 | return r; | ||
2426 | } | 2325 | } |
2427 | 2326 | ||
2428 | static void vmx_inject_page_fault(struct kvm_vcpu *vcpu, | 2327 | static void vmx_inject_page_fault(struct kvm_vcpu *vcpu, |
@@ -2560,12 +2459,15 @@ static struct kvm_x86_ops vmx_x86_ops = { | |||
2560 | 2459 | ||
2561 | .vcpu_create = vmx_create_vcpu, | 2460 | .vcpu_create = vmx_create_vcpu, |
2562 | .vcpu_free = vmx_free_vcpu, | 2461 | .vcpu_free = vmx_free_vcpu, |
2462 | .vcpu_reset = vmx_vcpu_reset, | ||
2563 | 2463 | ||
2464 | .prepare_guest_switch = vmx_save_host_state, | ||
2564 | .vcpu_load = vmx_vcpu_load, | 2465 | .vcpu_load = vmx_vcpu_load, |
2565 | .vcpu_put = vmx_vcpu_put, | 2466 | .vcpu_put = vmx_vcpu_put, |
2566 | .vcpu_decache = vmx_vcpu_decache, | 2467 | .vcpu_decache = vmx_vcpu_decache, |
2567 | 2468 | ||
2568 | .set_guest_debug = set_guest_debug, | 2469 | .set_guest_debug = set_guest_debug, |
2470 | .guest_debug_pre = kvm_guest_debug_pre, | ||
2569 | .get_msr = vmx_get_msr, | 2471 | .get_msr = vmx_get_msr, |
2570 | .set_msr = vmx_set_msr, | 2472 | .set_msr = vmx_set_msr, |
2571 | .get_segment_base = vmx_get_segment_base, | 2473 | .get_segment_base = vmx_get_segment_base, |
@@ -2594,10 +2496,13 @@ static struct kvm_x86_ops vmx_x86_ops = { | |||
2594 | .inject_gp = vmx_inject_gp, | 2496 | .inject_gp = vmx_inject_gp, |
2595 | 2497 | ||
2596 | .run = vmx_vcpu_run, | 2498 | .run = vmx_vcpu_run, |
2499 | .handle_exit = kvm_handle_exit, | ||
2597 | .skip_emulated_instruction = skip_emulated_instruction, | 2500 | .skip_emulated_instruction = skip_emulated_instruction, |
2598 | .patch_hypercall = vmx_patch_hypercall, | 2501 | .patch_hypercall = vmx_patch_hypercall, |
2599 | .get_irq = vmx_get_irq, | 2502 | .get_irq = vmx_get_irq, |
2600 | .set_irq = vmx_inject_irq, | 2503 | .set_irq = vmx_inject_irq, |
2504 | .inject_pending_irq = vmx_intr_assist, | ||
2505 | .inject_pending_vectors = do_interrupt_requests, | ||
2601 | }; | 2506 | }; |
2602 | 2507 | ||
2603 | static int __init vmx_init(void) | 2508 | static int __init vmx_init(void) |