diff options
Diffstat (limited to 'arch/x86/kvm/mmu.c')
-rw-r--r-- | arch/x86/kvm/mmu.c | 91 |
1 files changed, 75 insertions, 16 deletions
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c index 9cafbb499813..47b2c3288b6b 100644 --- a/arch/x86/kvm/mmu.c +++ b/arch/x86/kvm/mmu.c | |||
@@ -554,14 +554,18 @@ static int host_mapping_level(struct kvm *kvm, gfn_t gfn) | |||
554 | return ret; | 554 | return ret; |
555 | } | 555 | } |
556 | 556 | ||
557 | static int mapping_level(struct kvm_vcpu *vcpu, gfn_t large_gfn) | 557 | static bool mapping_level_dirty_bitmap(struct kvm_vcpu *vcpu, gfn_t large_gfn) |
558 | { | 558 | { |
559 | struct kvm_memory_slot *slot; | 559 | struct kvm_memory_slot *slot; |
560 | int host_level, level, max_level; | ||
561 | |||
562 | slot = gfn_to_memslot(vcpu->kvm, large_gfn); | 560 | slot = gfn_to_memslot(vcpu->kvm, large_gfn); |
563 | if (slot && slot->dirty_bitmap) | 561 | if (slot && slot->dirty_bitmap) |
564 | return PT_PAGE_TABLE_LEVEL; | 562 | return true; |
563 | return false; | ||
564 | } | ||
565 | |||
566 | static int mapping_level(struct kvm_vcpu *vcpu, gfn_t large_gfn) | ||
567 | { | ||
568 | int host_level, level, max_level; | ||
565 | 569 | ||
566 | host_level = host_mapping_level(vcpu->kvm, large_gfn); | 570 | host_level = host_mapping_level(vcpu->kvm, large_gfn); |
567 | 571 | ||
@@ -2281,6 +2285,48 @@ static int kvm_handle_bad_page(struct kvm *kvm, gfn_t gfn, pfn_t pfn) | |||
2281 | return 1; | 2285 | return 1; |
2282 | } | 2286 | } |
2283 | 2287 | ||
2288 | static void transparent_hugepage_adjust(struct kvm_vcpu *vcpu, | ||
2289 | gfn_t *gfnp, pfn_t *pfnp, int *levelp) | ||
2290 | { | ||
2291 | pfn_t pfn = *pfnp; | ||
2292 | gfn_t gfn = *gfnp; | ||
2293 | int level = *levelp; | ||
2294 | |||
2295 | /* | ||
2296 | * Check if it's a transparent hugepage. If this would be an | ||
2297 | * hugetlbfs page, level wouldn't be set to | ||
2298 | * PT_PAGE_TABLE_LEVEL and there would be no adjustment done | ||
2299 | * here. | ||
2300 | */ | ||
2301 | if (!is_error_pfn(pfn) && !kvm_is_mmio_pfn(pfn) && | ||
2302 | level == PT_PAGE_TABLE_LEVEL && | ||
2303 | PageTransCompound(pfn_to_page(pfn)) && | ||
2304 | !has_wrprotected_page(vcpu->kvm, gfn, PT_DIRECTORY_LEVEL)) { | ||
2305 | unsigned long mask; | ||
2306 | /* | ||
2307 | * mmu_notifier_retry was successful and we hold the | ||
2308 | * mmu_lock here, so the pmd can't become splitting | ||
2309 | * from under us, and in turn | ||
2310 | * __split_huge_page_refcount() can't run from under | ||
2311 | * us and we can safely transfer the refcount from | ||
2312 | * PG_tail to PG_head as we switch the pfn to tail to | ||
2313 | * head. | ||
2314 | */ | ||
2315 | *levelp = level = PT_DIRECTORY_LEVEL; | ||
2316 | mask = KVM_PAGES_PER_HPAGE(level) - 1; | ||
2317 | VM_BUG_ON((gfn & mask) != (pfn & mask)); | ||
2318 | if (pfn & mask) { | ||
2319 | gfn &= ~mask; | ||
2320 | *gfnp = gfn; | ||
2321 | kvm_release_pfn_clean(pfn); | ||
2322 | pfn &= ~mask; | ||
2323 | if (!get_page_unless_zero(pfn_to_page(pfn))) | ||
2324 | BUG(); | ||
2325 | *pfnp = pfn; | ||
2326 | } | ||
2327 | } | ||
2328 | } | ||
2329 | |||
2284 | static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn, | 2330 | static bool try_async_pf(struct kvm_vcpu *vcpu, bool prefault, gfn_t gfn, |
2285 | gva_t gva, pfn_t *pfn, bool write, bool *writable); | 2331 | gva_t gva, pfn_t *pfn, bool write, bool *writable); |
2286 | 2332 | ||
@@ -2289,20 +2335,25 @@ static int nonpaging_map(struct kvm_vcpu *vcpu, gva_t v, int write, gfn_t gfn, | |||
2289 | { | 2335 | { |
2290 | int r; | 2336 | int r; |
2291 | int level; | 2337 | int level; |
2338 | int force_pt_level; | ||
2292 | pfn_t pfn; | 2339 | pfn_t pfn; |
2293 | unsigned long mmu_seq; | 2340 | unsigned long mmu_seq; |
2294 | bool map_writable; | 2341 | bool map_writable; |
2295 | 2342 | ||
2296 | level = mapping_level(vcpu, gfn); | 2343 | force_pt_level = mapping_level_dirty_bitmap(vcpu, gfn); |
2297 | 2344 | if (likely(!force_pt_level)) { | |
2298 | /* | 2345 | level = mapping_level(vcpu, gfn); |
2299 | * This path builds a PAE pagetable - so we can map 2mb pages at | 2346 | /* |
2300 | * maximum. Therefore check if the level is larger than that. | 2347 | * This path builds a PAE pagetable - so we can map |
2301 | */ | 2348 | * 2mb pages at maximum. Therefore check if the level |
2302 | if (level > PT_DIRECTORY_LEVEL) | 2349 | * is larger than that. |
2303 | level = PT_DIRECTORY_LEVEL; | 2350 | */ |
2351 | if (level > PT_DIRECTORY_LEVEL) | ||
2352 | level = PT_DIRECTORY_LEVEL; | ||
2304 | 2353 | ||
2305 | gfn &= ~(KVM_PAGES_PER_HPAGE(level) - 1); | 2354 | gfn &= ~(KVM_PAGES_PER_HPAGE(level) - 1); |
2355 | } else | ||
2356 | level = PT_PAGE_TABLE_LEVEL; | ||
2306 | 2357 | ||
2307 | mmu_seq = vcpu->kvm->mmu_notifier_seq; | 2358 | mmu_seq = vcpu->kvm->mmu_notifier_seq; |
2308 | smp_rmb(); | 2359 | smp_rmb(); |
@@ -2318,6 +2369,8 @@ static int nonpaging_map(struct kvm_vcpu *vcpu, gva_t v, int write, gfn_t gfn, | |||
2318 | if (mmu_notifier_retry(vcpu, mmu_seq)) | 2369 | if (mmu_notifier_retry(vcpu, mmu_seq)) |
2319 | goto out_unlock; | 2370 | goto out_unlock; |
2320 | kvm_mmu_free_some_pages(vcpu); | 2371 | kvm_mmu_free_some_pages(vcpu); |
2372 | if (likely(!force_pt_level)) | ||
2373 | transparent_hugepage_adjust(vcpu, &gfn, &pfn, &level); | ||
2321 | r = __direct_map(vcpu, v, write, map_writable, level, gfn, pfn, | 2374 | r = __direct_map(vcpu, v, write, map_writable, level, gfn, pfn, |
2322 | prefault); | 2375 | prefault); |
2323 | spin_unlock(&vcpu->kvm->mmu_lock); | 2376 | spin_unlock(&vcpu->kvm->mmu_lock); |
@@ -2655,6 +2708,7 @@ static int tdp_page_fault(struct kvm_vcpu *vcpu, gva_t gpa, u32 error_code, | |||
2655 | pfn_t pfn; | 2708 | pfn_t pfn; |
2656 | int r; | 2709 | int r; |
2657 | int level; | 2710 | int level; |
2711 | int force_pt_level; | ||
2658 | gfn_t gfn = gpa >> PAGE_SHIFT; | 2712 | gfn_t gfn = gpa >> PAGE_SHIFT; |
2659 | unsigned long mmu_seq; | 2713 | unsigned long mmu_seq; |
2660 | int write = error_code & PFERR_WRITE_MASK; | 2714 | int write = error_code & PFERR_WRITE_MASK; |
@@ -2667,9 +2721,12 @@ static int tdp_page_fault(struct kvm_vcpu *vcpu, gva_t gpa, u32 error_code, | |||
2667 | if (r) | 2721 | if (r) |
2668 | return r; | 2722 | return r; |
2669 | 2723 | ||
2670 | level = mapping_level(vcpu, gfn); | 2724 | force_pt_level = mapping_level_dirty_bitmap(vcpu, gfn); |
2671 | 2725 | if (likely(!force_pt_level)) { | |
2672 | gfn &= ~(KVM_PAGES_PER_HPAGE(level) - 1); | 2726 | level = mapping_level(vcpu, gfn); |
2727 | gfn &= ~(KVM_PAGES_PER_HPAGE(level) - 1); | ||
2728 | } else | ||
2729 | level = PT_PAGE_TABLE_LEVEL; | ||
2673 | 2730 | ||
2674 | mmu_seq = vcpu->kvm->mmu_notifier_seq; | 2731 | mmu_seq = vcpu->kvm->mmu_notifier_seq; |
2675 | smp_rmb(); | 2732 | smp_rmb(); |
@@ -2684,6 +2741,8 @@ static int tdp_page_fault(struct kvm_vcpu *vcpu, gva_t gpa, u32 error_code, | |||
2684 | if (mmu_notifier_retry(vcpu, mmu_seq)) | 2741 | if (mmu_notifier_retry(vcpu, mmu_seq)) |
2685 | goto out_unlock; | 2742 | goto out_unlock; |
2686 | kvm_mmu_free_some_pages(vcpu); | 2743 | kvm_mmu_free_some_pages(vcpu); |
2744 | if (likely(!force_pt_level)) | ||
2745 | transparent_hugepage_adjust(vcpu, &gfn, &pfn, &level); | ||
2687 | r = __direct_map(vcpu, gpa, write, map_writable, | 2746 | r = __direct_map(vcpu, gpa, write, map_writable, |
2688 | level, gfn, pfn, prefault); | 2747 | level, gfn, pfn, prefault); |
2689 | spin_unlock(&vcpu->kvm->mmu_lock); | 2748 | spin_unlock(&vcpu->kvm->mmu_lock); |