aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--arch/x86/kvm/lapic.c8
-rw-r--r--arch/x86/kvm/mmu.c35
-rw-r--r--arch/x86/kvm/paging_tmpl.h14
-rw-r--r--arch/x86/kvm/svm.c14
-rw-r--r--arch/x86/kvm/vmx.c6
-rw-r--r--arch/x86/kvm/x86.c12
6 files changed, 44 insertions, 45 deletions
diff --git a/arch/x86/kvm/lapic.c b/arch/x86/kvm/lapic.c
index 68a6b1511934..31280df7d2e3 100644
--- a/arch/x86/kvm/lapic.c
+++ b/arch/x86/kvm/lapic.c
@@ -658,7 +658,7 @@ static void start_apic_timer(struct kvm_lapic *apic)
658 apic_debug("%s: bus cycle is %" PRId64 "ns, now 0x%016" 658 apic_debug("%s: bus cycle is %" PRId64 "ns, now 0x%016"
659 PRIx64 ", " 659 PRIx64 ", "
660 "timer initial count 0x%x, period %lldns, " 660 "timer initial count 0x%x, period %lldns, "
661 "expire @ 0x%016" PRIx64 ".\n", __FUNCTION__, 661 "expire @ 0x%016" PRIx64 ".\n", __func__,
662 APIC_BUS_CYCLE_NS, ktime_to_ns(now), 662 APIC_BUS_CYCLE_NS, ktime_to_ns(now),
663 apic_get_reg(apic, APIC_TMICT), 663 apic_get_reg(apic, APIC_TMICT),
664 apic->timer.period, 664 apic->timer.period,
@@ -691,7 +691,7 @@ static void apic_mmio_write(struct kvm_io_device *this,
691 /* too common printing */ 691 /* too common printing */
692 if (offset != APIC_EOI) 692 if (offset != APIC_EOI)
693 apic_debug("%s: offset 0x%x with length 0x%x, and value is " 693 apic_debug("%s: offset 0x%x with length 0x%x, and value is "
694 "0x%x\n", __FUNCTION__, offset, len, val); 694 "0x%x\n", __func__, offset, len, val);
695 695
696 offset &= 0xff0; 696 offset &= 0xff0;
697 697
@@ -869,7 +869,7 @@ void kvm_lapic_reset(struct kvm_vcpu *vcpu)
869 struct kvm_lapic *apic; 869 struct kvm_lapic *apic;
870 int i; 870 int i;
871 871
872 apic_debug("%s\n", __FUNCTION__); 872 apic_debug("%s\n", __func__);
873 873
874 ASSERT(vcpu); 874 ASSERT(vcpu);
875 apic = vcpu->arch.apic; 875 apic = vcpu->arch.apic;
@@ -907,7 +907,7 @@ void kvm_lapic_reset(struct kvm_vcpu *vcpu)
907 apic_update_ppr(apic); 907 apic_update_ppr(apic);
908 908
909 apic_debug(KERN_INFO "%s: vcpu=%p, id=%d, base_msr=" 909 apic_debug(KERN_INFO "%s: vcpu=%p, id=%d, base_msr="
910 "0x%016" PRIx64 ", base_address=0x%0lx.\n", __FUNCTION__, 910 "0x%016" PRIx64 ", base_address=0x%0lx.\n", __func__,
911 vcpu, kvm_apic_id(apic), 911 vcpu, kvm_apic_id(apic),
912 vcpu->arch.apic_base, apic->base_address); 912 vcpu->arch.apic_base, apic->base_address);
913} 913}
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
index 1932a3aeda1d..414405b6ec13 100644
--- a/arch/x86/kvm/mmu.c
+++ b/arch/x86/kvm/mmu.c
@@ -649,7 +649,7 @@ static int is_empty_shadow_page(u64 *spt)
649 649
650 for (pos = spt, end = pos + PAGE_SIZE / sizeof(u64); pos != end; pos++) 650 for (pos = spt, end = pos + PAGE_SIZE / sizeof(u64); pos != end; pos++)
651 if (*pos != shadow_trap_nonpresent_pte) { 651 if (*pos != shadow_trap_nonpresent_pte) {
652 printk(KERN_ERR "%s: %p %llx\n", __FUNCTION__, 652 printk(KERN_ERR "%s: %p %llx\n", __func__,
653 pos, *pos); 653 pos, *pos);
654 return 0; 654 return 0;
655 } 655 }
@@ -772,14 +772,14 @@ static struct kvm_mmu_page *kvm_mmu_lookup_page(struct kvm *kvm, gfn_t gfn)
772 struct kvm_mmu_page *sp; 772 struct kvm_mmu_page *sp;
773 struct hlist_node *node; 773 struct hlist_node *node;
774 774
775 pgprintk("%s: looking for gfn %lx\n", __FUNCTION__, gfn); 775 pgprintk("%s: looking for gfn %lx\n", __func__, gfn);
776 index = kvm_page_table_hashfn(gfn); 776 index = kvm_page_table_hashfn(gfn);
777 bucket = &kvm->arch.mmu_page_hash[index]; 777 bucket = &kvm->arch.mmu_page_hash[index];
778 hlist_for_each_entry(sp, node, bucket, hash_link) 778 hlist_for_each_entry(sp, node, bucket, hash_link)
779 if (sp->gfn == gfn && !sp->role.metaphysical 779 if (sp->gfn == gfn && !sp->role.metaphysical
780 && !sp->role.invalid) { 780 && !sp->role.invalid) {
781 pgprintk("%s: found role %x\n", 781 pgprintk("%s: found role %x\n",
782 __FUNCTION__, sp->role.word); 782 __func__, sp->role.word);
783 return sp; 783 return sp;
784 } 784 }
785 return NULL; 785 return NULL;
@@ -810,21 +810,21 @@ static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu,
810 quadrant &= (1 << ((PT32_PT_BITS - PT64_PT_BITS) * level)) - 1; 810 quadrant &= (1 << ((PT32_PT_BITS - PT64_PT_BITS) * level)) - 1;
811 role.quadrant = quadrant; 811 role.quadrant = quadrant;
812 } 812 }
813 pgprintk("%s: looking gfn %lx role %x\n", __FUNCTION__, 813 pgprintk("%s: looking gfn %lx role %x\n", __func__,
814 gfn, role.word); 814 gfn, role.word);
815 index = kvm_page_table_hashfn(gfn); 815 index = kvm_page_table_hashfn(gfn);
816 bucket = &vcpu->kvm->arch.mmu_page_hash[index]; 816 bucket = &vcpu->kvm->arch.mmu_page_hash[index];
817 hlist_for_each_entry(sp, node, bucket, hash_link) 817 hlist_for_each_entry(sp, node, bucket, hash_link)
818 if (sp->gfn == gfn && sp->role.word == role.word) { 818 if (sp->gfn == gfn && sp->role.word == role.word) {
819 mmu_page_add_parent_pte(vcpu, sp, parent_pte); 819 mmu_page_add_parent_pte(vcpu, sp, parent_pte);
820 pgprintk("%s: found\n", __FUNCTION__); 820 pgprintk("%s: found\n", __func__);
821 return sp; 821 return sp;
822 } 822 }
823 ++vcpu->kvm->stat.mmu_cache_miss; 823 ++vcpu->kvm->stat.mmu_cache_miss;
824 sp = kvm_mmu_alloc_page(vcpu, parent_pte); 824 sp = kvm_mmu_alloc_page(vcpu, parent_pte);
825 if (!sp) 825 if (!sp)
826 return sp; 826 return sp;
827 pgprintk("%s: adding gfn %lx role %x\n", __FUNCTION__, gfn, role.word); 827 pgprintk("%s: adding gfn %lx role %x\n", __func__, gfn, role.word);
828 sp->gfn = gfn; 828 sp->gfn = gfn;
829 sp->role = role; 829 sp->role = role;
830 hlist_add_head(&sp->hash_link, bucket); 830 hlist_add_head(&sp->hash_link, bucket);
@@ -960,13 +960,13 @@ static int kvm_mmu_unprotect_page(struct kvm *kvm, gfn_t gfn)
960 struct hlist_node *node, *n; 960 struct hlist_node *node, *n;
961 int r; 961 int r;
962 962
963 pgprintk("%s: looking for gfn %lx\n", __FUNCTION__, gfn); 963 pgprintk("%s: looking for gfn %lx\n", __func__, gfn);
964 r = 0; 964 r = 0;
965 index = kvm_page_table_hashfn(gfn); 965 index = kvm_page_table_hashfn(gfn);
966 bucket = &kvm->arch.mmu_page_hash[index]; 966 bucket = &kvm->arch.mmu_page_hash[index];
967 hlist_for_each_entry_safe(sp, node, n, bucket, hash_link) 967 hlist_for_each_entry_safe(sp, node, n, bucket, hash_link)
968 if (sp->gfn == gfn && !sp->role.metaphysical) { 968 if (sp->gfn == gfn && !sp->role.metaphysical) {
969 pgprintk("%s: gfn %lx role %x\n", __FUNCTION__, gfn, 969 pgprintk("%s: gfn %lx role %x\n", __func__, gfn,
970 sp->role.word); 970 sp->role.word);
971 kvm_mmu_zap_page(kvm, sp); 971 kvm_mmu_zap_page(kvm, sp);
972 r = 1; 972 r = 1;
@@ -979,7 +979,7 @@ static void mmu_unshadow(struct kvm *kvm, gfn_t gfn)
979 struct kvm_mmu_page *sp; 979 struct kvm_mmu_page *sp;
980 980
981 while ((sp = kvm_mmu_lookup_page(kvm, gfn)) != NULL) { 981 while ((sp = kvm_mmu_lookup_page(kvm, gfn)) != NULL) {
982 pgprintk("%s: zap %lx %x\n", __FUNCTION__, gfn, sp->role.word); 982 pgprintk("%s: zap %lx %x\n", __func__, gfn, sp->role.word);
983 kvm_mmu_zap_page(kvm, sp); 983 kvm_mmu_zap_page(kvm, sp);
984 } 984 }
985} 985}
@@ -1021,7 +1021,7 @@ static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *shadow_pte,
1021 1021
1022 pgprintk("%s: spte %llx access %x write_fault %d" 1022 pgprintk("%s: spte %llx access %x write_fault %d"
1023 " user_fault %d gfn %lx\n", 1023 " user_fault %d gfn %lx\n",
1024 __FUNCTION__, *shadow_pte, pt_access, 1024 __func__, *shadow_pte, pt_access,
1025 write_fault, user_fault, gfn); 1025 write_fault, user_fault, gfn);
1026 1026
1027 if (is_rmap_pte(*shadow_pte)) { 1027 if (is_rmap_pte(*shadow_pte)) {
@@ -1047,7 +1047,6 @@ static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *shadow_pte,
1047 } 1047 }
1048 } 1048 }
1049 1049
1050
1051 /* 1050 /*
1052 * We don't set the accessed bit, since we sometimes want to see 1051 * We don't set the accessed bit, since we sometimes want to see
1053 * whether the guest actually used the pte (in order to detect 1052 * whether the guest actually used the pte (in order to detect
@@ -1081,7 +1080,7 @@ static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *shadow_pte,
1081 if (shadow || 1080 if (shadow ||
1082 (largepage && has_wrprotected_page(vcpu->kvm, gfn))) { 1081 (largepage && has_wrprotected_page(vcpu->kvm, gfn))) {
1083 pgprintk("%s: found shadow page for %lx, marking ro\n", 1082 pgprintk("%s: found shadow page for %lx, marking ro\n",
1084 __FUNCTION__, gfn); 1083 __func__, gfn);
1085 pte_access &= ~ACC_WRITE_MASK; 1084 pte_access &= ~ACC_WRITE_MASK;
1086 if (is_writeble_pte(spte)) { 1085 if (is_writeble_pte(spte)) {
1087 spte &= ~PT_WRITABLE_MASK; 1086 spte &= ~PT_WRITABLE_MASK;
@@ -1097,7 +1096,7 @@ unshadowed:
1097 if (pte_access & ACC_WRITE_MASK) 1096 if (pte_access & ACC_WRITE_MASK)
1098 mark_page_dirty(vcpu->kvm, gfn); 1097 mark_page_dirty(vcpu->kvm, gfn);
1099 1098
1100 pgprintk("%s: setting spte %llx\n", __FUNCTION__, spte); 1099 pgprintk("%s: setting spte %llx\n", __func__, spte);
1101 pgprintk("instantiating %s PTE (%s) at %d (%llx) addr %llx\n", 1100 pgprintk("instantiating %s PTE (%s) at %d (%llx) addr %llx\n",
1102 (spte&PT_PAGE_SIZE_MASK)? "2MB" : "4kB", 1101 (spte&PT_PAGE_SIZE_MASK)? "2MB" : "4kB",
1103 (spte&PT_WRITABLE_MASK)?"RW":"R", gfn, spte, shadow_pte); 1102 (spte&PT_WRITABLE_MASK)?"RW":"R", gfn, spte, shadow_pte);
@@ -1317,7 +1316,7 @@ static int nonpaging_page_fault(struct kvm_vcpu *vcpu, gva_t gva,
1317 gfn_t gfn; 1316 gfn_t gfn;
1318 int r; 1317 int r;
1319 1318
1320 pgprintk("%s: gva %lx error %x\n", __FUNCTION__, gva, error_code); 1319 pgprintk("%s: gva %lx error %x\n", __func__, gva, error_code);
1321 r = mmu_topup_memory_caches(vcpu); 1320 r = mmu_topup_memory_caches(vcpu);
1322 if (r) 1321 if (r)
1323 return r; 1322 return r;
@@ -1395,7 +1394,7 @@ void kvm_mmu_flush_tlb(struct kvm_vcpu *vcpu)
1395 1394
1396static void paging_new_cr3(struct kvm_vcpu *vcpu) 1395static void paging_new_cr3(struct kvm_vcpu *vcpu)
1397{ 1396{
1398 pgprintk("%s: cr3 %lx\n", __FUNCTION__, vcpu->arch.cr3); 1397 pgprintk("%s: cr3 %lx\n", __func__, vcpu->arch.cr3);
1399 mmu_free_roots(vcpu); 1398 mmu_free_roots(vcpu);
1400} 1399}
1401 1400
@@ -1691,7 +1690,7 @@ void kvm_mmu_pte_write(struct kvm_vcpu *vcpu, gpa_t gpa,
1691 int npte; 1690 int npte;
1692 int r; 1691 int r;
1693 1692
1694 pgprintk("%s: gpa %llx bytes %d\n", __FUNCTION__, gpa, bytes); 1693 pgprintk("%s: gpa %llx bytes %d\n", __func__, gpa, bytes);
1695 mmu_guess_page_from_pte_write(vcpu, gpa, new, bytes); 1694 mmu_guess_page_from_pte_write(vcpu, gpa, new, bytes);
1696 spin_lock(&vcpu->kvm->mmu_lock); 1695 spin_lock(&vcpu->kvm->mmu_lock);
1697 kvm_mmu_free_some_pages(vcpu); 1696 kvm_mmu_free_some_pages(vcpu);
@@ -2139,7 +2138,7 @@ static void audit_rmap(struct kvm_vcpu *vcpu)
2139 2138
2140 if (n_rmap != n_actual) 2139 if (n_rmap != n_actual)
2141 printk(KERN_ERR "%s: (%s) rmap %d actual %d\n", 2140 printk(KERN_ERR "%s: (%s) rmap %d actual %d\n",
2142 __FUNCTION__, audit_msg, n_rmap, n_actual); 2141 __func__, audit_msg, n_rmap, n_actual);
2143} 2142}
2144 2143
2145static void audit_write_protection(struct kvm_vcpu *vcpu) 2144static void audit_write_protection(struct kvm_vcpu *vcpu)
@@ -2159,7 +2158,7 @@ static void audit_write_protection(struct kvm_vcpu *vcpu)
2159 if (*rmapp) 2158 if (*rmapp)
2160 printk(KERN_ERR "%s: (%s) shadow page has writable" 2159 printk(KERN_ERR "%s: (%s) shadow page has writable"
2161 " mappings: gfn %lx role %x\n", 2160 " mappings: gfn %lx role %x\n",
2162 __FUNCTION__, audit_msg, sp->gfn, 2161 __func__, audit_msg, sp->gfn,
2163 sp->role.word); 2162 sp->role.word);
2164 } 2163 }
2165} 2164}
diff --git a/arch/x86/kvm/paging_tmpl.h b/arch/x86/kvm/paging_tmpl.h
index 17f9d160ca34..57abbd091143 100644
--- a/arch/x86/kvm/paging_tmpl.h
+++ b/arch/x86/kvm/paging_tmpl.h
@@ -130,7 +130,7 @@ static int FNAME(walk_addr)(struct guest_walker *walker,
130 unsigned index, pt_access, pte_access; 130 unsigned index, pt_access, pte_access;
131 gpa_t pte_gpa; 131 gpa_t pte_gpa;
132 132
133 pgprintk("%s: addr %lx\n", __FUNCTION__, addr); 133 pgprintk("%s: addr %lx\n", __func__, addr);
134walk: 134walk:
135 walker->level = vcpu->arch.mmu.root_level; 135 walker->level = vcpu->arch.mmu.root_level;
136 pte = vcpu->arch.cr3; 136 pte = vcpu->arch.cr3;
@@ -155,7 +155,7 @@ walk:
155 pte_gpa += index * sizeof(pt_element_t); 155 pte_gpa += index * sizeof(pt_element_t);
156 walker->table_gfn[walker->level - 1] = table_gfn; 156 walker->table_gfn[walker->level - 1] = table_gfn;
157 walker->pte_gpa[walker->level - 1] = pte_gpa; 157 walker->pte_gpa[walker->level - 1] = pte_gpa;
158 pgprintk("%s: table_gfn[%d] %lx\n", __FUNCTION__, 158 pgprintk("%s: table_gfn[%d] %lx\n", __func__,
159 walker->level - 1, table_gfn); 159 walker->level - 1, table_gfn);
160 160
161 kvm_read_guest(vcpu->kvm, pte_gpa, &pte, sizeof(pte)); 161 kvm_read_guest(vcpu->kvm, pte_gpa, &pte, sizeof(pte));
@@ -222,7 +222,7 @@ walk:
222 walker->pt_access = pt_access; 222 walker->pt_access = pt_access;
223 walker->pte_access = pte_access; 223 walker->pte_access = pte_access;
224 pgprintk("%s: pte %llx pte_access %x pt_access %x\n", 224 pgprintk("%s: pte %llx pte_access %x pt_access %x\n",
225 __FUNCTION__, (u64)pte, pt_access, pte_access); 225 __func__, (u64)pte, pt_access, pte_access);
226 return 1; 226 return 1;
227 227
228not_present: 228not_present:
@@ -256,7 +256,7 @@ static void FNAME(update_pte)(struct kvm_vcpu *vcpu, struct kvm_mmu_page *page,
256 set_shadow_pte(spte, shadow_notrap_nonpresent_pte); 256 set_shadow_pte(spte, shadow_notrap_nonpresent_pte);
257 return; 257 return;
258 } 258 }
259 pgprintk("%s: gpte %llx spte %p\n", __FUNCTION__, (u64)gpte, spte); 259 pgprintk("%s: gpte %llx spte %p\n", __func__, (u64)gpte, spte);
260 pte_access = page->role.access & FNAME(gpte_access)(vcpu, gpte); 260 pte_access = page->role.access & FNAME(gpte_access)(vcpu, gpte);
261 if (gpte_to_gfn(gpte) != vcpu->arch.update_pte.gfn) 261 if (gpte_to_gfn(gpte) != vcpu->arch.update_pte.gfn)
262 return; 262 return;
@@ -381,7 +381,7 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr,
381 struct page *page; 381 struct page *page;
382 int largepage = 0; 382 int largepage = 0;
383 383
384 pgprintk("%s: addr %lx err %x\n", __FUNCTION__, addr, error_code); 384 pgprintk("%s: addr %lx err %x\n", __func__, addr, error_code);
385 kvm_mmu_audit(vcpu, "pre page fault"); 385 kvm_mmu_audit(vcpu, "pre page fault");
386 386
387 r = mmu_topup_memory_caches(vcpu); 387 r = mmu_topup_memory_caches(vcpu);
@@ -399,7 +399,7 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr,
399 * The page is not mapped by the guest. Let the guest handle it. 399 * The page is not mapped by the guest. Let the guest handle it.
400 */ 400 */
401 if (!r) { 401 if (!r) {
402 pgprintk("%s: guest page fault\n", __FUNCTION__); 402 pgprintk("%s: guest page fault\n", __func__);
403 inject_page_fault(vcpu, addr, walker.error_code); 403 inject_page_fault(vcpu, addr, walker.error_code);
404 vcpu->arch.last_pt_write_count = 0; /* reset fork detector */ 404 vcpu->arch.last_pt_write_count = 0; /* reset fork detector */
405 up_read(&vcpu->kvm->slots_lock); 405 up_read(&vcpu->kvm->slots_lock);
@@ -431,7 +431,7 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr,
431 shadow_pte = FNAME(fetch)(vcpu, addr, &walker, user_fault, write_fault, 431 shadow_pte = FNAME(fetch)(vcpu, addr, &walker, user_fault, write_fault,
432 largepage, &write_pt, page); 432 largepage, &write_pt, page);
433 433
434 pgprintk("%s: shadow pte %p %llx ptwrite %d\n", __FUNCTION__, 434 pgprintk("%s: shadow pte %p %llx ptwrite %d\n", __func__,
435 shadow_pte, *shadow_pte, write_pt); 435 shadow_pte, *shadow_pte, write_pt);
436 436
437 if (!write_pt) 437 if (!write_pt)
diff --git a/arch/x86/kvm/svm.c b/arch/x86/kvm/svm.c
index ff6e5c8da3c6..b2c667fe6832 100644
--- a/arch/x86/kvm/svm.c
+++ b/arch/x86/kvm/svm.c
@@ -230,12 +230,12 @@ static void skip_emulated_instruction(struct kvm_vcpu *vcpu)
230 struct vcpu_svm *svm = to_svm(vcpu); 230 struct vcpu_svm *svm = to_svm(vcpu);
231 231
232 if (!svm->next_rip) { 232 if (!svm->next_rip) {
233 printk(KERN_DEBUG "%s: NOP\n", __FUNCTION__); 233 printk(KERN_DEBUG "%s: NOP\n", __func__);
234 return; 234 return;
235 } 235 }
236 if (svm->next_rip - svm->vmcb->save.rip > MAX_INST_SIZE) 236 if (svm->next_rip - svm->vmcb->save.rip > MAX_INST_SIZE)
237 printk(KERN_ERR "%s: ip 0x%llx next 0x%llx\n", 237 printk(KERN_ERR "%s: ip 0x%llx next 0x%llx\n",
238 __FUNCTION__, 238 __func__,
239 svm->vmcb->save.rip, 239 svm->vmcb->save.rip,
240 svm->next_rip); 240 svm->next_rip);
241 241
@@ -996,7 +996,7 @@ static void svm_set_dr(struct kvm_vcpu *vcpu, int dr, unsigned long value,
996 } 996 }
997 default: 997 default:
998 printk(KERN_DEBUG "%s: unexpected dr %u\n", 998 printk(KERN_DEBUG "%s: unexpected dr %u\n",
999 __FUNCTION__, dr); 999 __func__, dr);
1000 *exception = UD_VECTOR; 1000 *exception = UD_VECTOR;
1001 return; 1001 return;
1002 } 1002 }
@@ -1109,7 +1109,7 @@ static int invalid_op_interception(struct vcpu_svm *svm,
1109static int task_switch_interception(struct vcpu_svm *svm, 1109static int task_switch_interception(struct vcpu_svm *svm,
1110 struct kvm_run *kvm_run) 1110 struct kvm_run *kvm_run)
1111{ 1111{
1112 pr_unimpl(&svm->vcpu, "%s: task switch is unsupported\n", __FUNCTION__); 1112 pr_unimpl(&svm->vcpu, "%s: task switch is unsupported\n", __func__);
1113 kvm_run->exit_reason = KVM_EXIT_UNKNOWN; 1113 kvm_run->exit_reason = KVM_EXIT_UNKNOWN;
1114 return 0; 1114 return 0;
1115} 1115}
@@ -1125,7 +1125,7 @@ static int emulate_on_interception(struct vcpu_svm *svm,
1125 struct kvm_run *kvm_run) 1125 struct kvm_run *kvm_run)
1126{ 1126{
1127 if (emulate_instruction(&svm->vcpu, NULL, 0, 0, 0) != EMULATE_DONE) 1127 if (emulate_instruction(&svm->vcpu, NULL, 0, 0, 0) != EMULATE_DONE)
1128 pr_unimpl(&svm->vcpu, "%s: failed\n", __FUNCTION__); 1128 pr_unimpl(&svm->vcpu, "%s: failed\n", __func__);
1129 return 1; 1129 return 1;
1130} 1130}
1131 1131
@@ -1257,7 +1257,7 @@ static int svm_set_msr(struct kvm_vcpu *vcpu, unsigned ecx, u64 data)
1257 case MSR_IA32_DEBUGCTLMSR: 1257 case MSR_IA32_DEBUGCTLMSR:
1258 if (!svm_has(SVM_FEATURE_LBRV)) { 1258 if (!svm_has(SVM_FEATURE_LBRV)) {
1259 pr_unimpl(vcpu, "%s: MSR_IA32_DEBUGCTL 0x%llx, nop\n", 1259 pr_unimpl(vcpu, "%s: MSR_IA32_DEBUGCTL 0x%llx, nop\n",
1260 __FUNCTION__, data); 1260 __func__, data);
1261 break; 1261 break;
1262 } 1262 }
1263 if (data & DEBUGCTL_RESERVED_BITS) 1263 if (data & DEBUGCTL_RESERVED_BITS)
@@ -1419,7 +1419,7 @@ static int handle_exit(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
1419 exit_code != SVM_EXIT_NPF) 1419 exit_code != SVM_EXIT_NPF)
1420 printk(KERN_ERR "%s: unexpected exit_ini_info 0x%x " 1420 printk(KERN_ERR "%s: unexpected exit_ini_info 0x%x "
1421 "exit_code 0x%x\n", 1421 "exit_code 0x%x\n",
1422 __FUNCTION__, svm->vmcb->control.exit_int_info, 1422 __func__, svm->vmcb->control.exit_int_info,
1423 exit_code); 1423 exit_code);
1424 1424
1425 if (exit_code >= ARRAY_SIZE(svm_exit_handlers) 1425 if (exit_code >= ARRAY_SIZE(svm_exit_handlers)
diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c
index 50345032974d..7ef710afceba 100644
--- a/arch/x86/kvm/vmx.c
+++ b/arch/x86/kvm/vmx.c
@@ -1254,7 +1254,7 @@ static void enter_lmode(struct kvm_vcpu *vcpu)
1254 guest_tr_ar = vmcs_read32(GUEST_TR_AR_BYTES); 1254 guest_tr_ar = vmcs_read32(GUEST_TR_AR_BYTES);
1255 if ((guest_tr_ar & AR_TYPE_MASK) != AR_TYPE_BUSY_64_TSS) { 1255 if ((guest_tr_ar & AR_TYPE_MASK) != AR_TYPE_BUSY_64_TSS) {
1256 printk(KERN_DEBUG "%s: tss fixup for long mode. \n", 1256 printk(KERN_DEBUG "%s: tss fixup for long mode. \n",
1257 __FUNCTION__); 1257 __func__);
1258 vmcs_write32(GUEST_TR_AR_BYTES, 1258 vmcs_write32(GUEST_TR_AR_BYTES,
1259 (guest_tr_ar & ~AR_TYPE_MASK) 1259 (guest_tr_ar & ~AR_TYPE_MASK)
1260 | AR_TYPE_BUSY_64_TSS); 1260 | AR_TYPE_BUSY_64_TSS);
@@ -1909,7 +1909,7 @@ static int handle_exception(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run)
1909 if ((vect_info & VECTORING_INFO_VALID_MASK) && 1909 if ((vect_info & VECTORING_INFO_VALID_MASK) &&
1910 !is_page_fault(intr_info)) 1910 !is_page_fault(intr_info))
1911 printk(KERN_ERR "%s: unexpected, vectoring info 0x%x " 1911 printk(KERN_ERR "%s: unexpected, vectoring info 0x%x "
1912 "intr info 0x%x\n", __FUNCTION__, vect_info, intr_info); 1912 "intr info 0x%x\n", __func__, vect_info, intr_info);
1913 1913
1914 if (!irqchip_in_kernel(vcpu->kvm) && is_external_interrupt(vect_info)) { 1914 if (!irqchip_in_kernel(vcpu->kvm) && is_external_interrupt(vect_info)) {
1915 int irq = vect_info & VECTORING_INFO_VECTOR_MASK; 1915 int irq = vect_info & VECTORING_INFO_VECTOR_MASK;
@@ -2275,7 +2275,7 @@ static int kvm_handle_exit(struct kvm_run *kvm_run, struct kvm_vcpu *vcpu)
2275 if ((vectoring_info & VECTORING_INFO_VALID_MASK) && 2275 if ((vectoring_info & VECTORING_INFO_VALID_MASK) &&
2276 exit_reason != EXIT_REASON_EXCEPTION_NMI) 2276 exit_reason != EXIT_REASON_EXCEPTION_NMI)
2277 printk(KERN_WARNING "%s: unexpected, valid vectoring info and " 2277 printk(KERN_WARNING "%s: unexpected, valid vectoring info and "
2278 "exit reason is 0x%x\n", __FUNCTION__, exit_reason); 2278 "exit reason is 0x%x\n", __func__, exit_reason);
2279 if (exit_reason < kvm_vmx_max_exit_handlers 2279 if (exit_reason < kvm_vmx_max_exit_handlers
2280 && kvm_vmx_exit_handlers[exit_reason]) 2280 && kvm_vmx_exit_handlers[exit_reason])
2281 return kvm_vmx_exit_handlers[exit_reason](vcpu, kvm_run); 2281 return kvm_vmx_exit_handlers[exit_reason](vcpu, kvm_run);
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
index 491eda308289..bf78d6522d3d 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -563,15 +563,15 @@ int kvm_set_msr_common(struct kvm_vcpu *vcpu, u32 msr, u64 data)
563 break; 563 break;
564 case MSR_IA32_MC0_STATUS: 564 case MSR_IA32_MC0_STATUS:
565 pr_unimpl(vcpu, "%s: MSR_IA32_MC0_STATUS 0x%llx, nop\n", 565 pr_unimpl(vcpu, "%s: MSR_IA32_MC0_STATUS 0x%llx, nop\n",
566 __FUNCTION__, data); 566 __func__, data);
567 break; 567 break;
568 case MSR_IA32_MCG_STATUS: 568 case MSR_IA32_MCG_STATUS:
569 pr_unimpl(vcpu, "%s: MSR_IA32_MCG_STATUS 0x%llx, nop\n", 569 pr_unimpl(vcpu, "%s: MSR_IA32_MCG_STATUS 0x%llx, nop\n",
570 __FUNCTION__, data); 570 __func__, data);
571 break; 571 break;
572 case MSR_IA32_MCG_CTL: 572 case MSR_IA32_MCG_CTL:
573 pr_unimpl(vcpu, "%s: MSR_IA32_MCG_CTL 0x%llx, nop\n", 573 pr_unimpl(vcpu, "%s: MSR_IA32_MCG_CTL 0x%llx, nop\n",
574 __FUNCTION__, data); 574 __func__, data);
575 break; 575 break;
576 case MSR_IA32_UCODE_REV: 576 case MSR_IA32_UCODE_REV:
577 case MSR_IA32_UCODE_WRITE: 577 case MSR_IA32_UCODE_WRITE:
@@ -1939,7 +1939,7 @@ int emulator_get_dr(struct x86_emulate_ctxt *ctxt, int dr, unsigned long *dest)
1939 *dest = kvm_x86_ops->get_dr(vcpu, dr); 1939 *dest = kvm_x86_ops->get_dr(vcpu, dr);
1940 return X86EMUL_CONTINUE; 1940 return X86EMUL_CONTINUE;
1941 default: 1941 default:
1942 pr_unimpl(vcpu, "%s: unexpected dr %u\n", __FUNCTION__, dr); 1942 pr_unimpl(vcpu, "%s: unexpected dr %u\n", __func__, dr);
1943 return X86EMUL_UNHANDLEABLE; 1943 return X86EMUL_UNHANDLEABLE;
1944 } 1944 }
1945} 1945}
@@ -2486,7 +2486,7 @@ unsigned long realmode_get_cr(struct kvm_vcpu *vcpu, int cr)
2486 case 8: 2486 case 8:
2487 return kvm_get_cr8(vcpu); 2487 return kvm_get_cr8(vcpu);
2488 default: 2488 default:
2489 vcpu_printf(vcpu, "%s: unexpected cr %u\n", __FUNCTION__, cr); 2489 vcpu_printf(vcpu, "%s: unexpected cr %u\n", __func__, cr);
2490 return 0; 2490 return 0;
2491 } 2491 }
2492} 2492}
@@ -2512,7 +2512,7 @@ void realmode_set_cr(struct kvm_vcpu *vcpu, int cr, unsigned long val,
2512 kvm_set_cr8(vcpu, val & 0xfUL); 2512 kvm_set_cr8(vcpu, val & 0xfUL);
2513 break; 2513 break;
2514 default: 2514 default:
2515 vcpu_printf(vcpu, "%s: unexpected cr %u\n", __FUNCTION__, cr); 2515 vcpu_printf(vcpu, "%s: unexpected cr %u\n", __func__, cr);
2516 } 2516 }
2517} 2517}
2518 2518