diff options
author | Lai Jiangshan <laijs@cn.fujitsu.com> | 2010-05-26 04:49:59 -0400 |
---|---|---|
committer | Avi Kivity <avi@redhat.com> | 2010-08-01 03:35:52 -0400 |
commit | 2032a93d66fa282ba0f2ea9152eeff9511fa9a96 (patch) | |
tree | f0bbd50c3f237fa93c87207dc29602cbdcf9ad63 /arch/x86/kvm/mmu.c | |
parent | c8174f7b35b3018c4c7b3237ed1c792e454fd5c3 (diff) |
KVM: MMU: Don't allocate gfns page for direct mmu pages
When sp->role.direct is set, sp->gfns does not contain any essential
information, leaf sptes reachable from this sp are for a continuous
guest physical memory range (a linear range).
So sp->gfns[i] (if it was set) equals to sp->gfn + i. (PT_PAGE_TABLE_LEVEL)
Obviously, it is not essential information, we can calculate it when need.
It means we don't need sp->gfns when sp->role.direct=1,
Thus we can save one page usage for every kvm_mmu_page.
Note:
Access to sp->gfns must be wrapped by kvm_mmu_page_get_gfn()
or kvm_mmu_page_set_gfn().
It is only exposed in FNAME(sync_page).
Signed-off-by: Lai Jiangshan <laijs@cn.fujitsu.com>
Signed-off-by: Avi Kivity <avi@redhat.com>
Diffstat (limited to 'arch/x86/kvm/mmu.c')
-rw-r--r-- | arch/x86/kvm/mmu.c | 38 |
1 files changed, 29 insertions, 9 deletions
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c index 07673487fd5d..f46b6c9aff27 100644 --- a/arch/x86/kvm/mmu.c +++ b/arch/x86/kvm/mmu.c | |||
@@ -397,6 +397,22 @@ static void mmu_free_rmap_desc(struct kvm_rmap_desc *rd) | |||
397 | kmem_cache_free(rmap_desc_cache, rd); | 397 | kmem_cache_free(rmap_desc_cache, rd); |
398 | } | 398 | } |
399 | 399 | ||
400 | static gfn_t kvm_mmu_page_get_gfn(struct kvm_mmu_page *sp, int index) | ||
401 | { | ||
402 | if (!sp->role.direct) | ||
403 | return sp->gfns[index]; | ||
404 | |||
405 | return sp->gfn + (index << ((sp->role.level - 1) * PT64_LEVEL_BITS)); | ||
406 | } | ||
407 | |||
408 | static void kvm_mmu_page_set_gfn(struct kvm_mmu_page *sp, int index, gfn_t gfn) | ||
409 | { | ||
410 | if (sp->role.direct) | ||
411 | BUG_ON(gfn != kvm_mmu_page_get_gfn(sp, index)); | ||
412 | else | ||
413 | sp->gfns[index] = gfn; | ||
414 | } | ||
415 | |||
400 | /* | 416 | /* |
401 | * Return the pointer to the largepage write count for a given | 417 | * Return the pointer to the largepage write count for a given |
402 | * gfn, handling slots that are not large page aligned. | 418 | * gfn, handling slots that are not large page aligned. |
@@ -547,7 +563,7 @@ static int rmap_add(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn) | |||
547 | return count; | 563 | return count; |
548 | gfn = unalias_gfn(vcpu->kvm, gfn); | 564 | gfn = unalias_gfn(vcpu->kvm, gfn); |
549 | sp = page_header(__pa(spte)); | 565 | sp = page_header(__pa(spte)); |
550 | sp->gfns[spte - sp->spt] = gfn; | 566 | kvm_mmu_page_set_gfn(sp, spte - sp->spt, gfn); |
551 | rmapp = gfn_to_rmap(vcpu->kvm, gfn, sp->role.level); | 567 | rmapp = gfn_to_rmap(vcpu->kvm, gfn, sp->role.level); |
552 | if (!*rmapp) { | 568 | if (!*rmapp) { |
553 | rmap_printk("rmap_add: %p %llx 0->1\n", spte, *spte); | 569 | rmap_printk("rmap_add: %p %llx 0->1\n", spte, *spte); |
@@ -605,6 +621,7 @@ static void rmap_remove(struct kvm *kvm, u64 *spte) | |||
605 | struct kvm_rmap_desc *prev_desc; | 621 | struct kvm_rmap_desc *prev_desc; |
606 | struct kvm_mmu_page *sp; | 622 | struct kvm_mmu_page *sp; |
607 | pfn_t pfn; | 623 | pfn_t pfn; |
624 | gfn_t gfn; | ||
608 | unsigned long *rmapp; | 625 | unsigned long *rmapp; |
609 | int i; | 626 | int i; |
610 | 627 | ||
@@ -616,7 +633,8 @@ static void rmap_remove(struct kvm *kvm, u64 *spte) | |||
616 | kvm_set_pfn_accessed(pfn); | 633 | kvm_set_pfn_accessed(pfn); |
617 | if (is_writable_pte(*spte)) | 634 | if (is_writable_pte(*spte)) |
618 | kvm_set_pfn_dirty(pfn); | 635 | kvm_set_pfn_dirty(pfn); |
619 | rmapp = gfn_to_rmap(kvm, sp->gfns[spte - sp->spt], sp->role.level); | 636 | gfn = kvm_mmu_page_get_gfn(sp, spte - sp->spt); |
637 | rmapp = gfn_to_rmap(kvm, gfn, sp->role.level); | ||
620 | if (!*rmapp) { | 638 | if (!*rmapp) { |
621 | printk(KERN_ERR "rmap_remove: %p %llx 0->BUG\n", spte, *spte); | 639 | printk(KERN_ERR "rmap_remove: %p %llx 0->BUG\n", spte, *spte); |
622 | BUG(); | 640 | BUG(); |
@@ -900,7 +918,8 @@ static void kvm_mmu_free_page(struct kvm *kvm, struct kvm_mmu_page *sp) | |||
900 | ASSERT(is_empty_shadow_page(sp->spt)); | 918 | ASSERT(is_empty_shadow_page(sp->spt)); |
901 | list_del(&sp->link); | 919 | list_del(&sp->link); |
902 | __free_page(virt_to_page(sp->spt)); | 920 | __free_page(virt_to_page(sp->spt)); |
903 | __free_page(virt_to_page(sp->gfns)); | 921 | if (!sp->role.direct) |
922 | __free_page(virt_to_page(sp->gfns)); | ||
904 | kmem_cache_free(mmu_page_header_cache, sp); | 923 | kmem_cache_free(mmu_page_header_cache, sp); |
905 | ++kvm->arch.n_free_mmu_pages; | 924 | ++kvm->arch.n_free_mmu_pages; |
906 | } | 925 | } |
@@ -911,13 +930,15 @@ static unsigned kvm_page_table_hashfn(gfn_t gfn) | |||
911 | } | 930 | } |
912 | 931 | ||
913 | static struct kvm_mmu_page *kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, | 932 | static struct kvm_mmu_page *kvm_mmu_alloc_page(struct kvm_vcpu *vcpu, |
914 | u64 *parent_pte) | 933 | u64 *parent_pte, int direct) |
915 | { | 934 | { |
916 | struct kvm_mmu_page *sp; | 935 | struct kvm_mmu_page *sp; |
917 | 936 | ||
918 | sp = mmu_memory_cache_alloc(&vcpu->arch.mmu_page_header_cache, sizeof *sp); | 937 | sp = mmu_memory_cache_alloc(&vcpu->arch.mmu_page_header_cache, sizeof *sp); |
919 | sp->spt = mmu_memory_cache_alloc(&vcpu->arch.mmu_page_cache, PAGE_SIZE); | 938 | sp->spt = mmu_memory_cache_alloc(&vcpu->arch.mmu_page_cache, PAGE_SIZE); |
920 | sp->gfns = mmu_memory_cache_alloc(&vcpu->arch.mmu_page_cache, PAGE_SIZE); | 939 | if (!direct) |
940 | sp->gfns = mmu_memory_cache_alloc(&vcpu->arch.mmu_page_cache, | ||
941 | PAGE_SIZE); | ||
921 | set_page_private(virt_to_page(sp->spt), (unsigned long)sp); | 942 | set_page_private(virt_to_page(sp->spt), (unsigned long)sp); |
922 | list_add(&sp->link, &vcpu->kvm->arch.active_mmu_pages); | 943 | list_add(&sp->link, &vcpu->kvm->arch.active_mmu_pages); |
923 | bitmap_zero(sp->slot_bitmap, KVM_MEMORY_SLOTS + KVM_PRIVATE_MEM_SLOTS); | 944 | bitmap_zero(sp->slot_bitmap, KVM_MEMORY_SLOTS + KVM_PRIVATE_MEM_SLOTS); |
@@ -1386,7 +1407,7 @@ static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu, | |||
1386 | return sp; | 1407 | return sp; |
1387 | } | 1408 | } |
1388 | ++vcpu->kvm->stat.mmu_cache_miss; | 1409 | ++vcpu->kvm->stat.mmu_cache_miss; |
1389 | sp = kvm_mmu_alloc_page(vcpu, parent_pte); | 1410 | sp = kvm_mmu_alloc_page(vcpu, parent_pte, direct); |
1390 | if (!sp) | 1411 | if (!sp) |
1391 | return sp; | 1412 | return sp; |
1392 | sp->gfn = gfn; | 1413 | sp->gfn = gfn; |
@@ -3403,7 +3424,7 @@ void inspect_spte_has_rmap(struct kvm *kvm, u64 *sptep) | |||
3403 | 3424 | ||
3404 | if (*sptep & PT_WRITABLE_MASK) { | 3425 | if (*sptep & PT_WRITABLE_MASK) { |
3405 | rev_sp = page_header(__pa(sptep)); | 3426 | rev_sp = page_header(__pa(sptep)); |
3406 | gfn = rev_sp->gfns[sptep - rev_sp->spt]; | 3427 | gfn = kvm_mmu_page_get_gfn(rev_sp, sptep - rev_sp->spt); |
3407 | 3428 | ||
3408 | if (!gfn_to_memslot(kvm, gfn)) { | 3429 | if (!gfn_to_memslot(kvm, gfn)) { |
3409 | if (!printk_ratelimit()) | 3430 | if (!printk_ratelimit()) |
@@ -3417,8 +3438,7 @@ void inspect_spte_has_rmap(struct kvm *kvm, u64 *sptep) | |||
3417 | return; | 3438 | return; |
3418 | } | 3439 | } |
3419 | 3440 | ||
3420 | rmapp = gfn_to_rmap(kvm, rev_sp->gfns[sptep - rev_sp->spt], | 3441 | rmapp = gfn_to_rmap(kvm, gfn, rev_sp->role.level); |
3421 | rev_sp->role.level); | ||
3422 | if (!*rmapp) { | 3442 | if (!*rmapp) { |
3423 | if (!printk_ratelimit()) | 3443 | if (!printk_ratelimit()) |
3424 | return; | 3444 | return; |