aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--arch/powerpc/kvm/book3s_hv_rm_mmu.c23
-rw-r--r--include/linux/kvm_host.h25
-rw-r--r--virt/kvm/kvm_main.c21
3 files changed, 28 insertions, 41 deletions
diff --git a/arch/powerpc/kvm/book3s_hv_rm_mmu.c b/arch/powerpc/kvm/book3s_hv_rm_mmu.c
index 5f3c60b89faf..def880aea63a 100644
--- a/arch/powerpc/kvm/book3s_hv_rm_mmu.c
+++ b/arch/powerpc/kvm/book3s_hv_rm_mmu.c
@@ -21,25 +21,6 @@
21#include <asm/synch.h> 21#include <asm/synch.h>
22#include <asm/ppc-opcode.h> 22#include <asm/ppc-opcode.h>
23 23
24/*
25 * Since this file is built in even if KVM is a module, we need
26 * a local copy of this function for the case where kvm_main.c is
27 * modular.
28 */
29static struct kvm_memory_slot *builtin_gfn_to_memslot(struct kvm *kvm,
30 gfn_t gfn)
31{
32 struct kvm_memslots *slots;
33 struct kvm_memory_slot *memslot;
34
35 slots = kvm_memslots(kvm);
36 kvm_for_each_memslot(memslot, slots)
37 if (gfn >= memslot->base_gfn &&
38 gfn < memslot->base_gfn + memslot->npages)
39 return memslot;
40 return NULL;
41}
42
43/* Translate address of a vmalloc'd thing to a linear map address */ 24/* Translate address of a vmalloc'd thing to a linear map address */
44static void *real_vmalloc_addr(void *x) 25static void *real_vmalloc_addr(void *x)
45{ 26{
@@ -99,7 +80,7 @@ static void remove_revmap_chain(struct kvm *kvm, long pte_index,
99 rcbits = hpte_r & (HPTE_R_R | HPTE_R_C); 80 rcbits = hpte_r & (HPTE_R_R | HPTE_R_C);
100 ptel = rev->guest_rpte |= rcbits; 81 ptel = rev->guest_rpte |= rcbits;
101 gfn = hpte_rpn(ptel, hpte_page_size(hpte_v, ptel)); 82 gfn = hpte_rpn(ptel, hpte_page_size(hpte_v, ptel));
102 memslot = builtin_gfn_to_memslot(kvm, gfn); 83 memslot = __gfn_to_memslot(kvm_memslots(kvm), gfn);
103 if (!memslot || (memslot->flags & KVM_MEMSLOT_INVALID)) 84 if (!memslot || (memslot->flags & KVM_MEMSLOT_INVALID))
104 return; 85 return;
105 86
@@ -181,7 +162,7 @@ long kvmppc_h_enter(struct kvm_vcpu *vcpu, unsigned long flags,
181 /* Find the memslot (if any) for this address */ 162 /* Find the memslot (if any) for this address */
182 gpa = (ptel & HPTE_R_RPN) & ~(psize - 1); 163 gpa = (ptel & HPTE_R_RPN) & ~(psize - 1);
183 gfn = gpa >> PAGE_SHIFT; 164 gfn = gpa >> PAGE_SHIFT;
184 memslot = builtin_gfn_to_memslot(kvm, gfn); 165 memslot = __gfn_to_memslot(kvm_memslots(kvm), gfn);
185 pa = 0; 166 pa = 0;
186 is_io = ~0ul; 167 is_io = ~0ul;
187 rmap = NULL; 168 rmap = NULL;
diff --git a/include/linux/kvm_host.h b/include/linux/kvm_host.h
index eada8e69fe58..9698080c902b 100644
--- a/include/linux/kvm_host.h
+++ b/include/linux/kvm_host.h
@@ -651,6 +651,31 @@ static inline void kvm_guest_exit(void)
651 current->flags &= ~PF_VCPU; 651 current->flags &= ~PF_VCPU;
652} 652}
653 653
654/*
655 * search_memslots() and __gfn_to_memslot() are here because they are
656 * used in non-modular code in arch/powerpc/kvm/book3s_hv_rm_mmu.c.
657 * gfn_to_memslot() itself isn't here as an inline because that would
658 * bloat other code too much.
659 */
660static inline struct kvm_memory_slot *
661search_memslots(struct kvm_memslots *slots, gfn_t gfn)
662{
663 struct kvm_memory_slot *memslot;
664
665 kvm_for_each_memslot(memslot, slots)
666 if (gfn >= memslot->base_gfn &&
667 gfn < memslot->base_gfn + memslot->npages)
668 return memslot;
669
670 return NULL;
671}
672
673static inline struct kvm_memory_slot *
674__gfn_to_memslot(struct kvm_memslots *slots, gfn_t gfn)
675{
676 return search_memslots(slots, gfn);
677}
678
654static inline int memslot_id(struct kvm *kvm, gfn_t gfn) 679static inline int memslot_id(struct kvm *kvm, gfn_t gfn)
655{ 680{
656 return gfn_to_memslot(kvm, gfn)->id; 681 return gfn_to_memslot(kvm, gfn)->id;
diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
index 9f32bffd37c0..470e30520fe8 100644
--- a/virt/kvm/kvm_main.c
+++ b/virt/kvm/kvm_main.c
@@ -640,19 +640,6 @@ static int kvm_create_dirty_bitmap(struct kvm_memory_slot *memslot)
640} 640}
641#endif /* !CONFIG_S390 */ 641#endif /* !CONFIG_S390 */
642 642
643static struct kvm_memory_slot *
644search_memslots(struct kvm_memslots *slots, gfn_t gfn)
645{
646 struct kvm_memory_slot *memslot;
647
648 kvm_for_each_memslot(memslot, slots)
649 if (gfn >= memslot->base_gfn &&
650 gfn < memslot->base_gfn + memslot->npages)
651 return memslot;
652
653 return NULL;
654}
655
656static int cmp_memslot(const void *slot1, const void *slot2) 643static int cmp_memslot(const void *slot1, const void *slot2)
657{ 644{
658 struct kvm_memory_slot *s1, *s2; 645 struct kvm_memory_slot *s1, *s2;
@@ -1031,12 +1018,6 @@ int kvm_is_error_hva(unsigned long addr)
1031} 1018}
1032EXPORT_SYMBOL_GPL(kvm_is_error_hva); 1019EXPORT_SYMBOL_GPL(kvm_is_error_hva);
1033 1020
1034static struct kvm_memory_slot *__gfn_to_memslot(struct kvm_memslots *slots,
1035 gfn_t gfn)
1036{
1037 return search_memslots(slots, gfn);
1038}
1039
1040struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn) 1021struct kvm_memory_slot *gfn_to_memslot(struct kvm *kvm, gfn_t gfn)
1041{ 1022{
1042 return __gfn_to_memslot(kvm_memslots(kvm), gfn); 1023 return __gfn_to_memslot(kvm_memslots(kvm), gfn);
@@ -1459,7 +1440,7 @@ int kvm_gfn_to_hva_cache_init(struct kvm *kvm, struct gfn_to_hva_cache *ghc,
1459 1440
1460 ghc->gpa = gpa; 1441 ghc->gpa = gpa;
1461 ghc->generation = slots->generation; 1442 ghc->generation = slots->generation;
1462 ghc->memslot = __gfn_to_memslot(slots, gfn); 1443 ghc->memslot = gfn_to_memslot(kvm, gfn);
1463 ghc->hva = gfn_to_hva_many(ghc->memslot, gfn, NULL); 1444 ghc->hva = gfn_to_hva_many(ghc->memslot, gfn, NULL);
1464 if (!kvm_is_error_hva(ghc->hva)) 1445 if (!kvm_is_error_hva(ghc->hva))
1465 ghc->hva += offset; 1446 ghc->hva += offset;