diff options
Diffstat (limited to 'arch/x86/kvm/mmu.c')
-rw-r--r-- | arch/x86/kvm/mmu.c | 20 |
1 files changed, 10 insertions, 10 deletions
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c index b0037a77e56b..59bf1d9553a7 100644 --- a/arch/x86/kvm/mmu.c +++ b/arch/x86/kvm/mmu.c | |||
@@ -1738,11 +1738,11 @@ static int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn) | |||
1738 | LIST_HEAD(invalid_list); | 1738 | LIST_HEAD(invalid_list); |
1739 | int r; | 1739 | int r; |
1740 | 1740 | ||
1741 | pgprintk("%s: looking for gfn %lx\n", __func__, gfn); | 1741 | pgprintk("%s: looking for gfn %llx\n", __func__, gfn); |
1742 | r = 0; | 1742 | r = 0; |
1743 | 1743 | ||
1744 | for_each_gfn_indirect_valid_sp(kvm, sp, gfn, node) { | 1744 | for_each_gfn_indirect_valid_sp(kvm, sp, gfn, node) { |
1745 | pgprintk("%s: gfn %lx role %x\n", __func__, gfn, | 1745 | pgprintk("%s: gfn %llx role %x\n", __func__, gfn, |
1746 | sp->role.word); | 1746 | sp->role.word); |
1747 | r = 1; | 1747 | r = 1; |
1748 | kvm_mmu_prepare_zap_page(kvm, sp, &invalid_list); | 1748 | kvm_mmu_prepare_zap_page(kvm, sp, &invalid_list); |
@@ -1758,7 +1758,7 @@ static void mmu_unshadow(struct kvm *kvm, gfn_t gfn) | |||
1758 | LIST_HEAD(invalid_list); | 1758 | LIST_HEAD(invalid_list); |
1759 | 1759 | ||
1760 | for_each_gfn_indirect_valid_sp(kvm, sp, gfn, node) { | 1760 | for_each_gfn_indirect_valid_sp(kvm, sp, gfn, node) { |
1761 | pgprintk("%s: zap %lx %x\n", | 1761 | pgprintk("%s: zap %llx %x\n", |
1762 | __func__, gfn, sp->role.word); | 1762 | __func__, gfn, sp->role.word); |
1763 | kvm_mmu_prepare_zap_page(kvm, sp, &invalid_list); | 1763 | kvm_mmu_prepare_zap_page(kvm, sp, &invalid_list); |
1764 | } | 1764 | } |
@@ -2002,7 +2002,7 @@ static int set_spte(struct kvm_vcpu *vcpu, u64 *sptep, | |||
2002 | goto set_pte; | 2002 | goto set_pte; |
2003 | 2003 | ||
2004 | if (mmu_need_write_protect(vcpu, gfn, can_unsync)) { | 2004 | if (mmu_need_write_protect(vcpu, gfn, can_unsync)) { |
2005 | pgprintk("%s: found shadow page for %lx, marking ro\n", | 2005 | pgprintk("%s: found shadow page for %llx, marking ro\n", |
2006 | __func__, gfn); | 2006 | __func__, gfn); |
2007 | ret = 1; | 2007 | ret = 1; |
2008 | pte_access &= ~ACC_WRITE_MASK; | 2008 | pte_access &= ~ACC_WRITE_MASK; |
@@ -2031,7 +2031,7 @@ static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep, | |||
2031 | int rmap_count; | 2031 | int rmap_count; |
2032 | 2032 | ||
2033 | pgprintk("%s: spte %llx access %x write_fault %d" | 2033 | pgprintk("%s: spte %llx access %x write_fault %d" |
2034 | " user_fault %d gfn %lx\n", | 2034 | " user_fault %d gfn %llx\n", |
2035 | __func__, *sptep, pt_access, | 2035 | __func__, *sptep, pt_access, |
2036 | write_fault, user_fault, gfn); | 2036 | write_fault, user_fault, gfn); |
2037 | 2037 | ||
@@ -2050,7 +2050,7 @@ static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep, | |||
2050 | __set_spte(sptep, shadow_trap_nonpresent_pte); | 2050 | __set_spte(sptep, shadow_trap_nonpresent_pte); |
2051 | kvm_flush_remote_tlbs(vcpu->kvm); | 2051 | kvm_flush_remote_tlbs(vcpu->kvm); |
2052 | } else if (pfn != spte_to_pfn(*sptep)) { | 2052 | } else if (pfn != spte_to_pfn(*sptep)) { |
2053 | pgprintk("hfn old %lx new %lx\n", | 2053 | pgprintk("hfn old %llx new %llx\n", |
2054 | spte_to_pfn(*sptep), pfn); | 2054 | spte_to_pfn(*sptep), pfn); |
2055 | drop_spte(vcpu->kvm, sptep, shadow_trap_nonpresent_pte); | 2055 | drop_spte(vcpu->kvm, sptep, shadow_trap_nonpresent_pte); |
2056 | kvm_flush_remote_tlbs(vcpu->kvm); | 2056 | kvm_flush_remote_tlbs(vcpu->kvm); |
@@ -2067,7 +2067,7 @@ static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep, | |||
2067 | } | 2067 | } |
2068 | 2068 | ||
2069 | pgprintk("%s: setting spte %llx\n", __func__, *sptep); | 2069 | pgprintk("%s: setting spte %llx\n", __func__, *sptep); |
2070 | pgprintk("instantiating %s PTE (%s) at %ld (%llx) addr %p\n", | 2070 | pgprintk("instantiating %s PTE (%s) at %llx (%llx) addr %p\n", |
2071 | is_large_pte(*sptep)? "2MB" : "4kB", | 2071 | is_large_pte(*sptep)? "2MB" : "4kB", |
2072 | *sptep & PT_PRESENT_MASK ?"RW":"R", gfn, | 2072 | *sptep & PT_PRESENT_MASK ?"RW":"R", gfn, |
2073 | *sptep, sptep); | 2073 | *sptep, sptep); |
@@ -3651,9 +3651,9 @@ void inspect_spte_has_rmap(struct kvm *kvm, u64 *sptep) | |||
3651 | if (!gfn_to_memslot(kvm, gfn)) { | 3651 | if (!gfn_to_memslot(kvm, gfn)) { |
3652 | if (!printk_ratelimit()) | 3652 | if (!printk_ratelimit()) |
3653 | return; | 3653 | return; |
3654 | printk(KERN_ERR "%s: no memslot for gfn %ld\n", | 3654 | printk(KERN_ERR "%s: no memslot for gfn %llx\n", |
3655 | audit_msg, gfn); | 3655 | audit_msg, gfn); |
3656 | printk(KERN_ERR "%s: index %ld of sp (gfn=%lx)\n", | 3656 | printk(KERN_ERR "%s: index %ld of sp (gfn=%llx)\n", |
3657 | audit_msg, (long int)(sptep - rev_sp->spt), | 3657 | audit_msg, (long int)(sptep - rev_sp->spt), |
3658 | rev_sp->gfn); | 3658 | rev_sp->gfn); |
3659 | dump_stack(); | 3659 | dump_stack(); |
@@ -3728,7 +3728,7 @@ static void audit_write_protection(struct kvm_vcpu *vcpu) | |||
3728 | while (spte) { | 3728 | while (spte) { |
3729 | if (is_writable_pte(*spte)) | 3729 | if (is_writable_pte(*spte)) |
3730 | printk(KERN_ERR "%s: (%s) shadow page has " | 3730 | printk(KERN_ERR "%s: (%s) shadow page has " |
3731 | "writable mappings: gfn %lx role %x\n", | 3731 | "writable mappings: gfn %llx role %x\n", |
3732 | __func__, audit_msg, sp->gfn, | 3732 | __func__, audit_msg, sp->gfn, |
3733 | sp->role.word); | 3733 | sp->role.word); |
3734 | spte = rmap_next(vcpu->kvm, rmapp, spte); | 3734 | spte = rmap_next(vcpu->kvm, rmapp, spte); |