aboutsummaryrefslogtreecommitdiffstats
path: root/arch/x86/kvm/mmu.c
diff options
context:
space:
mode:
Diffstat (limited to 'arch/x86/kvm/mmu.c')
-rw-r--r--arch/x86/kvm/mmu.c20
1 files changed, 10 insertions, 10 deletions
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
index b0037a77e56b..59bf1d9553a7 100644
--- a/arch/x86/kvm/mmu.c
+++ b/arch/x86/kvm/mmu.c
@@ -1738,11 +1738,11 @@ static int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn)
1738 LIST_HEAD(invalid_list); 1738 LIST_HEAD(invalid_list);
1739 int r; 1739 int r;
1740 1740
1741 pgprintk("%s: looking for gfn %lx\n", __func__, gfn); 1741 pgprintk("%s: looking for gfn %llx\n", __func__, gfn);
1742 r = 0; 1742 r = 0;
1743 1743
1744 for_each_gfn_indirect_valid_sp(kvm, sp, gfn, node) { 1744 for_each_gfn_indirect_valid_sp(kvm, sp, gfn, node) {
1745 pgprintk("%s: gfn %lx role %x\n", __func__, gfn, 1745 pgprintk("%s: gfn %llx role %x\n", __func__, gfn,
1746 sp->role.word); 1746 sp->role.word);
1747 r = 1; 1747 r = 1;
1748 kvm_mmu_prepare_zap_page(kvm, sp, &invalid_list); 1748 kvm_mmu_prepare_zap_page(kvm, sp, &invalid_list);
@@ -1758,7 +1758,7 @@ static void mmu_unshadow(struct kvm *kvm, gfn_t gfn)
1758 LIST_HEAD(invalid_list); 1758 LIST_HEAD(invalid_list);
1759 1759
1760 for_each_gfn_indirect_valid_sp(kvm, sp, gfn, node) { 1760 for_each_gfn_indirect_valid_sp(kvm, sp, gfn, node) {
1761 pgprintk("%s: zap %lx %x\n", 1761 pgprintk("%s: zap %llx %x\n",
1762 __func__, gfn, sp->role.word); 1762 __func__, gfn, sp->role.word);
1763 kvm_mmu_prepare_zap_page(kvm, sp, &invalid_list); 1763 kvm_mmu_prepare_zap_page(kvm, sp, &invalid_list);
1764 } 1764 }
@@ -2002,7 +2002,7 @@ static int set_spte(struct kvm_vcpu *vcpu, u64 *sptep,
2002 goto set_pte; 2002 goto set_pte;
2003 2003
2004 if (mmu_need_write_protect(vcpu, gfn, can_unsync)) { 2004 if (mmu_need_write_protect(vcpu, gfn, can_unsync)) {
2005 pgprintk("%s: found shadow page for %lx, marking ro\n", 2005 pgprintk("%s: found shadow page for %llx, marking ro\n",
2006 __func__, gfn); 2006 __func__, gfn);
2007 ret = 1; 2007 ret = 1;
2008 pte_access &= ~ACC_WRITE_MASK; 2008 pte_access &= ~ACC_WRITE_MASK;
@@ -2031,7 +2031,7 @@ static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep,
2031 int rmap_count; 2031 int rmap_count;
2032 2032
2033 pgprintk("%s: spte %llx access %x write_fault %d" 2033 pgprintk("%s: spte %llx access %x write_fault %d"
2034 " user_fault %d gfn %lx\n", 2034 " user_fault %d gfn %llx\n",
2035 __func__, *sptep, pt_access, 2035 __func__, *sptep, pt_access,
2036 write_fault, user_fault, gfn); 2036 write_fault, user_fault, gfn);
2037 2037
@@ -2050,7 +2050,7 @@ static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep,
2050 __set_spte(sptep, shadow_trap_nonpresent_pte); 2050 __set_spte(sptep, shadow_trap_nonpresent_pte);
2051 kvm_flush_remote_tlbs(vcpu->kvm); 2051 kvm_flush_remote_tlbs(vcpu->kvm);
2052 } else if (pfn != spte_to_pfn(*sptep)) { 2052 } else if (pfn != spte_to_pfn(*sptep)) {
2053 pgprintk("hfn old %lx new %lx\n", 2053 pgprintk("hfn old %llx new %llx\n",
2054 spte_to_pfn(*sptep), pfn); 2054 spte_to_pfn(*sptep), pfn);
2055 drop_spte(vcpu->kvm, sptep, shadow_trap_nonpresent_pte); 2055 drop_spte(vcpu->kvm, sptep, shadow_trap_nonpresent_pte);
2056 kvm_flush_remote_tlbs(vcpu->kvm); 2056 kvm_flush_remote_tlbs(vcpu->kvm);
@@ -2067,7 +2067,7 @@ static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *sptep,
2067 } 2067 }
2068 2068
2069 pgprintk("%s: setting spte %llx\n", __func__, *sptep); 2069 pgprintk("%s: setting spte %llx\n", __func__, *sptep);
2070 pgprintk("instantiating %s PTE (%s) at %ld (%llx) addr %p\n", 2070 pgprintk("instantiating %s PTE (%s) at %llx (%llx) addr %p\n",
2071 is_large_pte(*sptep)? "2MB" : "4kB", 2071 is_large_pte(*sptep)? "2MB" : "4kB",
2072 *sptep & PT_PRESENT_MASK ?"RW":"R", gfn, 2072 *sptep & PT_PRESENT_MASK ?"RW":"R", gfn,
2073 *sptep, sptep); 2073 *sptep, sptep);
@@ -3651,9 +3651,9 @@ void inspect_spte_has_rmap(struct kvm *kvm, u64 *sptep)
3651 if (!gfn_to_memslot(kvm, gfn)) { 3651 if (!gfn_to_memslot(kvm, gfn)) {
3652 if (!printk_ratelimit()) 3652 if (!printk_ratelimit())
3653 return; 3653 return;
3654 printk(KERN_ERR "%s: no memslot for gfn %ld\n", 3654 printk(KERN_ERR "%s: no memslot for gfn %llx\n",
3655 audit_msg, gfn); 3655 audit_msg, gfn);
3656 printk(KERN_ERR "%s: index %ld of sp (gfn=%lx)\n", 3656 printk(KERN_ERR "%s: index %ld of sp (gfn=%llx)\n",
3657 audit_msg, (long int)(sptep - rev_sp->spt), 3657 audit_msg, (long int)(sptep - rev_sp->spt),
3658 rev_sp->gfn); 3658 rev_sp->gfn);
3659 dump_stack(); 3659 dump_stack();
@@ -3728,7 +3728,7 @@ static void audit_write_protection(struct kvm_vcpu *vcpu)
3728 while (spte) { 3728 while (spte) {
3729 if (is_writable_pte(*spte)) 3729 if (is_writable_pte(*spte))
3730 printk(KERN_ERR "%s: (%s) shadow page has " 3730 printk(KERN_ERR "%s: (%s) shadow page has "
3731 "writable mappings: gfn %lx role %x\n", 3731 "writable mappings: gfn %llx role %x\n",
3732 __func__, audit_msg, sp->gfn, 3732 __func__, audit_msg, sp->gfn,
3733 sp->role.word); 3733 sp->role.word);
3734 spte = rmap_next(vcpu->kvm, rmapp, spte); 3734 spte = rmap_next(vcpu->kvm, rmapp, spte);