diff options
| author | Avi Kivity <avi@qumranet.com> | 2007-01-05 19:36:54 -0500 |
|---|---|---|
| committer | Linus Torvalds <torvalds@woody.osdl.org> | 2007-01-06 02:55:27 -0500 |
| commit | e2dec939db126989808853d218e426daaeebc9e2 (patch) | |
| tree | 5c742e609e43090df396fc1c7a6b4c526099dbea | |
| parent | 714b93da1a6d97307dfafb9915517879d8a66c0d (diff) | |
[PATCH] KVM: MMU: Detect oom conditions and propagate error to userspace
Signed-off-by: Avi Kivity <avi@qumranet.com>
Acked-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Andrew Morton <akpm@osdl.org>
Signed-off-by: Linus Torvalds <torvalds@osdl.org>
| -rw-r--r-- | drivers/kvm/mmu.c | 32 | ||||
| -rw-r--r-- | drivers/kvm/paging_tmpl.h | 8 | ||||
| -rw-r--r-- | drivers/kvm/svm.c | 14 | ||||
| -rw-r--r-- | drivers/kvm/vmx.c | 15 |
4 files changed, 50 insertions, 19 deletions
diff --git a/drivers/kvm/mmu.c b/drivers/kvm/mmu.c index e96362aa7947..7761089ef3bc 100644 --- a/drivers/kvm/mmu.c +++ b/drivers/kvm/mmu.c | |||
| @@ -166,19 +166,20 @@ static int is_rmap_pte(u64 pte) | |||
| 166 | == (PT_WRITABLE_MASK | PT_PRESENT_MASK); | 166 | == (PT_WRITABLE_MASK | PT_PRESENT_MASK); |
| 167 | } | 167 | } |
| 168 | 168 | ||
| 169 | static void mmu_topup_memory_cache(struct kvm_mmu_memory_cache *cache, | 169 | static int mmu_topup_memory_cache(struct kvm_mmu_memory_cache *cache, |
| 170 | size_t objsize, int min) | 170 | size_t objsize, int min) |
| 171 | { | 171 | { |
| 172 | void *obj; | 172 | void *obj; |
| 173 | 173 | ||
| 174 | if (cache->nobjs >= min) | 174 | if (cache->nobjs >= min) |
| 175 | return; | 175 | return 0; |
| 176 | while (cache->nobjs < ARRAY_SIZE(cache->objects)) { | 176 | while (cache->nobjs < ARRAY_SIZE(cache->objects)) { |
| 177 | obj = kzalloc(objsize, GFP_NOWAIT); | 177 | obj = kzalloc(objsize, GFP_NOWAIT); |
| 178 | if (!obj) | 178 | if (!obj) |
| 179 | BUG(); | 179 | return -ENOMEM; |
| 180 | cache->objects[cache->nobjs++] = obj; | 180 | cache->objects[cache->nobjs++] = obj; |
| 181 | } | 181 | } |
| 182 | return 0; | ||
| 182 | } | 183 | } |
| 183 | 184 | ||
| 184 | static void mmu_free_memory_cache(struct kvm_mmu_memory_cache *mc) | 185 | static void mmu_free_memory_cache(struct kvm_mmu_memory_cache *mc) |
| @@ -187,12 +188,18 @@ static void mmu_free_memory_cache(struct kvm_mmu_memory_cache *mc) | |||
| 187 | kfree(mc->objects[--mc->nobjs]); | 188 | kfree(mc->objects[--mc->nobjs]); |
| 188 | } | 189 | } |
| 189 | 190 | ||
| 190 | static void mmu_topup_memory_caches(struct kvm_vcpu *vcpu) | 191 | static int mmu_topup_memory_caches(struct kvm_vcpu *vcpu) |
| 191 | { | 192 | { |
| 192 | mmu_topup_memory_cache(&vcpu->mmu_pte_chain_cache, | 193 | int r; |
| 193 | sizeof(struct kvm_pte_chain), 4); | 194 | |
| 194 | mmu_topup_memory_cache(&vcpu->mmu_rmap_desc_cache, | 195 | r = mmu_topup_memory_cache(&vcpu->mmu_pte_chain_cache, |
| 195 | sizeof(struct kvm_rmap_desc), 1); | 196 | sizeof(struct kvm_pte_chain), 4); |
| 197 | if (r) | ||
| 198 | goto out; | ||
| 199 | r = mmu_topup_memory_cache(&vcpu->mmu_rmap_desc_cache, | ||
| 200 | sizeof(struct kvm_rmap_desc), 1); | ||
| 201 | out: | ||
| 202 | return r; | ||
| 196 | } | 203 | } |
| 197 | 204 | ||
| 198 | static void mmu_free_memory_caches(struct kvm_vcpu *vcpu) | 205 | static void mmu_free_memory_caches(struct kvm_vcpu *vcpu) |
| @@ -824,8 +831,11 @@ static int nonpaging_page_fault(struct kvm_vcpu *vcpu, gva_t gva, | |||
| 824 | { | 831 | { |
| 825 | gpa_t addr = gva; | 832 | gpa_t addr = gva; |
| 826 | hpa_t paddr; | 833 | hpa_t paddr; |
| 834 | int r; | ||
| 827 | 835 | ||
| 828 | mmu_topup_memory_caches(vcpu); | 836 | r = mmu_topup_memory_caches(vcpu); |
| 837 | if (r) | ||
| 838 | return r; | ||
| 829 | 839 | ||
| 830 | ASSERT(vcpu); | 840 | ASSERT(vcpu); |
| 831 | ASSERT(VALID_PAGE(vcpu->mmu.root_hpa)); | 841 | ASSERT(VALID_PAGE(vcpu->mmu.root_hpa)); |
| @@ -1052,7 +1062,7 @@ int kvm_mmu_reset_context(struct kvm_vcpu *vcpu) | |||
| 1052 | r = init_kvm_mmu(vcpu); | 1062 | r = init_kvm_mmu(vcpu); |
| 1053 | if (r < 0) | 1063 | if (r < 0) |
| 1054 | goto out; | 1064 | goto out; |
| 1055 | mmu_topup_memory_caches(vcpu); | 1065 | r = mmu_topup_memory_caches(vcpu); |
| 1056 | out: | 1066 | out: |
| 1057 | return r; | 1067 | return r; |
| 1058 | } | 1068 | } |
diff --git a/drivers/kvm/paging_tmpl.h b/drivers/kvm/paging_tmpl.h index 4e6670ff1847..32b385188454 100644 --- a/drivers/kvm/paging_tmpl.h +++ b/drivers/kvm/paging_tmpl.h | |||
| @@ -339,7 +339,8 @@ static int FNAME(fix_write_pf)(struct kvm_vcpu *vcpu, | |||
| 339 | * - normal guest page fault due to the guest pte marked not present, not | 339 | * - normal guest page fault due to the guest pte marked not present, not |
| 340 | * writable, or not executable | 340 | * writable, or not executable |
| 341 | * | 341 | * |
| 342 | * Returns: 1 if we need to emulate the instruction, 0 otherwise | 342 | * Returns: 1 if we need to emulate the instruction, 0 otherwise, or |
| 343 | * a negative value on error. | ||
| 343 | */ | 344 | */ |
| 344 | static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr, | 345 | static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr, |
| 345 | u32 error_code) | 346 | u32 error_code) |
| @@ -351,10 +352,13 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr, | |||
| 351 | u64 *shadow_pte; | 352 | u64 *shadow_pte; |
| 352 | int fixed; | 353 | int fixed; |
| 353 | int write_pt = 0; | 354 | int write_pt = 0; |
| 355 | int r; | ||
| 354 | 356 | ||
| 355 | pgprintk("%s: addr %lx err %x\n", __FUNCTION__, addr, error_code); | 357 | pgprintk("%s: addr %lx err %x\n", __FUNCTION__, addr, error_code); |
| 356 | 358 | ||
| 357 | mmu_topup_memory_caches(vcpu); | 359 | r = mmu_topup_memory_caches(vcpu); |
| 360 | if (r) | ||
| 361 | return r; | ||
| 358 | 362 | ||
| 359 | /* | 363 | /* |
| 360 | * Look up the shadow pte for the faulting address. | 364 | * Look up the shadow pte for the faulting address. |
diff --git a/drivers/kvm/svm.c b/drivers/kvm/svm.c index 99250011a471..af1e7b3f9171 100644 --- a/drivers/kvm/svm.c +++ b/drivers/kvm/svm.c | |||
| @@ -852,6 +852,7 @@ static int pf_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) | |||
| 852 | u64 fault_address; | 852 | u64 fault_address; |
| 853 | u32 error_code; | 853 | u32 error_code; |
| 854 | enum emulation_result er; | 854 | enum emulation_result er; |
| 855 | int r; | ||
| 855 | 856 | ||
| 856 | if (is_external_interrupt(exit_int_info)) | 857 | if (is_external_interrupt(exit_int_info)) |
| 857 | push_irq(vcpu, exit_int_info & SVM_EVTINJ_VEC_MASK); | 858 | push_irq(vcpu, exit_int_info & SVM_EVTINJ_VEC_MASK); |
| @@ -860,7 +861,12 @@ static int pf_interception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) | |||
| 860 | 861 | ||
| 861 | fault_address = vcpu->svm->vmcb->control.exit_info_2; | 862 | fault_address = vcpu->svm->vmcb->control.exit_info_2; |
| 862 | error_code = vcpu->svm->vmcb->control.exit_info_1; | 863 | error_code = vcpu->svm->vmcb->control.exit_info_1; |
| 863 | if (!kvm_mmu_page_fault(vcpu, fault_address, error_code)) { | 864 | r = kvm_mmu_page_fault(vcpu, fault_address, error_code); |
| 865 | if (r < 0) { | ||
| 866 | spin_unlock(&vcpu->kvm->lock); | ||
| 867 | return r; | ||
| 868 | } | ||
| 869 | if (!r) { | ||
| 864 | spin_unlock(&vcpu->kvm->lock); | 870 | spin_unlock(&vcpu->kvm->lock); |
| 865 | return 1; | 871 | return 1; |
| 866 | } | 872 | } |
| @@ -1398,6 +1404,7 @@ static int svm_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) | |||
| 1398 | u16 fs_selector; | 1404 | u16 fs_selector; |
| 1399 | u16 gs_selector; | 1405 | u16 gs_selector; |
| 1400 | u16 ldt_selector; | 1406 | u16 ldt_selector; |
| 1407 | int r; | ||
| 1401 | 1408 | ||
| 1402 | again: | 1409 | again: |
| 1403 | do_interrupt_requests(vcpu, kvm_run); | 1410 | do_interrupt_requests(vcpu, kvm_run); |
| @@ -1565,7 +1572,8 @@ again: | |||
| 1565 | return 0; | 1572 | return 0; |
| 1566 | } | 1573 | } |
| 1567 | 1574 | ||
| 1568 | if (handle_exit(vcpu, kvm_run)) { | 1575 | r = handle_exit(vcpu, kvm_run); |
| 1576 | if (r > 0) { | ||
| 1569 | if (signal_pending(current)) { | 1577 | if (signal_pending(current)) { |
| 1570 | ++kvm_stat.signal_exits; | 1578 | ++kvm_stat.signal_exits; |
| 1571 | post_kvm_run_save(vcpu, kvm_run); | 1579 | post_kvm_run_save(vcpu, kvm_run); |
| @@ -1581,7 +1589,7 @@ again: | |||
| 1581 | goto again; | 1589 | goto again; |
| 1582 | } | 1590 | } |
| 1583 | post_kvm_run_save(vcpu, kvm_run); | 1591 | post_kvm_run_save(vcpu, kvm_run); |
| 1584 | return 0; | 1592 | return r; |
| 1585 | } | 1593 | } |
| 1586 | 1594 | ||
| 1587 | static void svm_flush_tlb(struct kvm_vcpu *vcpu) | 1595 | static void svm_flush_tlb(struct kvm_vcpu *vcpu) |
diff --git a/drivers/kvm/vmx.c b/drivers/kvm/vmx.c index 59178ad4d344..ed3956739771 100644 --- a/drivers/kvm/vmx.c +++ b/drivers/kvm/vmx.c | |||
| @@ -1289,6 +1289,7 @@ static int handle_exception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) | |||
| 1289 | unsigned long cr2, rip; | 1289 | unsigned long cr2, rip; |
| 1290 | u32 vect_info; | 1290 | u32 vect_info; |
| 1291 | enum emulation_result er; | 1291 | enum emulation_result er; |
| 1292 | int r; | ||
| 1292 | 1293 | ||
| 1293 | vect_info = vmcs_read32(IDT_VECTORING_INFO_FIELD); | 1294 | vect_info = vmcs_read32(IDT_VECTORING_INFO_FIELD); |
| 1294 | intr_info = vmcs_read32(VM_EXIT_INTR_INFO); | 1295 | intr_info = vmcs_read32(VM_EXIT_INTR_INFO); |
| @@ -1317,7 +1318,12 @@ static int handle_exception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) | |||
| 1317 | cr2 = vmcs_readl(EXIT_QUALIFICATION); | 1318 | cr2 = vmcs_readl(EXIT_QUALIFICATION); |
| 1318 | 1319 | ||
| 1319 | spin_lock(&vcpu->kvm->lock); | 1320 | spin_lock(&vcpu->kvm->lock); |
| 1320 | if (!kvm_mmu_page_fault(vcpu, cr2, error_code)) { | 1321 | r = kvm_mmu_page_fault(vcpu, cr2, error_code); |
| 1322 | if (r < 0) { | ||
| 1323 | spin_unlock(&vcpu->kvm->lock); | ||
| 1324 | return r; | ||
| 1325 | } | ||
| 1326 | if (!r) { | ||
| 1321 | spin_unlock(&vcpu->kvm->lock); | 1327 | spin_unlock(&vcpu->kvm->lock); |
| 1322 | return 1; | 1328 | return 1; |
| 1323 | } | 1329 | } |
| @@ -1680,6 +1686,7 @@ static int vmx_vcpu_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) | |||
| 1680 | u8 fail; | 1686 | u8 fail; |
| 1681 | u16 fs_sel, gs_sel, ldt_sel; | 1687 | u16 fs_sel, gs_sel, ldt_sel; |
| 1682 | int fs_gs_ldt_reload_needed; | 1688 | int fs_gs_ldt_reload_needed; |
| 1689 | int r; | ||
| 1683 | 1690 | ||
| 1684 | again: | 1691 | again: |
| 1685 | /* | 1692 | /* |
| @@ -1853,6 +1860,7 @@ again: | |||
| 1853 | if (fail) { | 1860 | if (fail) { |
| 1854 | kvm_run->exit_type = KVM_EXIT_TYPE_FAIL_ENTRY; | 1861 | kvm_run->exit_type = KVM_EXIT_TYPE_FAIL_ENTRY; |
| 1855 | kvm_run->exit_reason = vmcs_read32(VM_INSTRUCTION_ERROR); | 1862 | kvm_run->exit_reason = vmcs_read32(VM_INSTRUCTION_ERROR); |
| 1863 | r = 0; | ||
| 1856 | } else { | 1864 | } else { |
| 1857 | if (fs_gs_ldt_reload_needed) { | 1865 | if (fs_gs_ldt_reload_needed) { |
| 1858 | load_ldt(ldt_sel); | 1866 | load_ldt(ldt_sel); |
| @@ -1872,7 +1880,8 @@ again: | |||
| 1872 | } | 1880 | } |
| 1873 | vcpu->launched = 1; | 1881 | vcpu->launched = 1; |
| 1874 | kvm_run->exit_type = KVM_EXIT_TYPE_VM_EXIT; | 1882 | kvm_run->exit_type = KVM_EXIT_TYPE_VM_EXIT; |
| 1875 | if (kvm_handle_exit(kvm_run, vcpu)) { | 1883 | r = kvm_handle_exit(kvm_run, vcpu); |
| 1884 | if (r > 0) { | ||
| 1876 | /* Give scheduler a change to reschedule. */ | 1885 | /* Give scheduler a change to reschedule. */ |
| 1877 | if (signal_pending(current)) { | 1886 | if (signal_pending(current)) { |
| 1878 | ++kvm_stat.signal_exits; | 1887 | ++kvm_stat.signal_exits; |
| @@ -1892,7 +1901,7 @@ again: | |||
| 1892 | } | 1901 | } |
| 1893 | 1902 | ||
| 1894 | post_kvm_run_save(vcpu, kvm_run); | 1903 | post_kvm_run_save(vcpu, kvm_run); |
| 1895 | return 0; | 1904 | return r; |
| 1896 | } | 1905 | } |
| 1897 | 1906 | ||
| 1898 | static void vmx_flush_tlb(struct kvm_vcpu *vcpu) | 1907 | static void vmx_flush_tlb(struct kvm_vcpu *vcpu) |
