diff options
Diffstat (limited to 'arch/powerpc/kvm')
-rw-r--r-- | arch/powerpc/kvm/book3s_mmu_hpte.c | 40 |
1 files changed, 35 insertions, 5 deletions
diff --git a/arch/powerpc/kvm/book3s_mmu_hpte.c b/arch/powerpc/kvm/book3s_mmu_hpte.c index b64389362446..02c64ab99c97 100644 --- a/arch/powerpc/kvm/book3s_mmu_hpte.c +++ b/arch/powerpc/kvm/book3s_mmu_hpte.c | |||
@@ -45,6 +45,12 @@ static inline u64 kvmppc_mmu_hash_pte(u64 eaddr) | |||
45 | return hash_64(eaddr >> PTE_SIZE, HPTEG_HASH_BITS_PTE); | 45 | return hash_64(eaddr >> PTE_SIZE, HPTEG_HASH_BITS_PTE); |
46 | } | 46 | } |
47 | 47 | ||
48 | static inline u64 kvmppc_mmu_hash_pte_long(u64 eaddr) | ||
49 | { | ||
50 | return hash_64((eaddr & 0x0ffff000) >> PTE_SIZE, | ||
51 | HPTEG_HASH_BITS_PTE_LONG); | ||
52 | } | ||
53 | |||
48 | static inline u64 kvmppc_mmu_hash_vpte(u64 vpage) | 54 | static inline u64 kvmppc_mmu_hash_vpte(u64 vpage) |
49 | { | 55 | { |
50 | return hash_64(vpage & 0xfffffffffULL, HPTEG_HASH_BITS_VPTE); | 56 | return hash_64(vpage & 0xfffffffffULL, HPTEG_HASH_BITS_VPTE); |
@@ -66,6 +72,11 @@ void kvmppc_mmu_hpte_cache_map(struct kvm_vcpu *vcpu, struct hpte_cache *pte) | |||
66 | index = kvmppc_mmu_hash_pte(pte->pte.eaddr); | 72 | index = kvmppc_mmu_hash_pte(pte->pte.eaddr); |
67 | hlist_add_head_rcu(&pte->list_pte, &vcpu->arch.hpte_hash_pte[index]); | 73 | hlist_add_head_rcu(&pte->list_pte, &vcpu->arch.hpte_hash_pte[index]); |
68 | 74 | ||
75 | /* Add to ePTE_long list */ | ||
76 | index = kvmppc_mmu_hash_pte_long(pte->pte.eaddr); | ||
77 | hlist_add_head_rcu(&pte->list_pte_long, | ||
78 | &vcpu->arch.hpte_hash_pte_long[index]); | ||
79 | |||
69 | /* Add to vPTE list */ | 80 | /* Add to vPTE list */ |
70 | index = kvmppc_mmu_hash_vpte(pte->pte.vpage); | 81 | index = kvmppc_mmu_hash_vpte(pte->pte.vpage); |
71 | hlist_add_head_rcu(&pte->list_vpte, &vcpu->arch.hpte_hash_vpte[index]); | 82 | hlist_add_head_rcu(&pte->list_vpte, &vcpu->arch.hpte_hash_vpte[index]); |
@@ -99,6 +110,7 @@ static void invalidate_pte(struct kvm_vcpu *vcpu, struct hpte_cache *pte) | |||
99 | spin_lock(&vcpu->arch.mmu_lock); | 110 | spin_lock(&vcpu->arch.mmu_lock); |
100 | 111 | ||
101 | hlist_del_init_rcu(&pte->list_pte); | 112 | hlist_del_init_rcu(&pte->list_pte); |
113 | hlist_del_init_rcu(&pte->list_pte_long); | ||
102 | hlist_del_init_rcu(&pte->list_vpte); | 114 | hlist_del_init_rcu(&pte->list_vpte); |
103 | hlist_del_init_rcu(&pte->list_vpte_long); | 115 | hlist_del_init_rcu(&pte->list_vpte_long); |
104 | 116 | ||
@@ -150,10 +162,28 @@ static void kvmppc_mmu_pte_flush_page(struct kvm_vcpu *vcpu, ulong guest_ea) | |||
150 | rcu_read_unlock(); | 162 | rcu_read_unlock(); |
151 | } | 163 | } |
152 | 164 | ||
153 | void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong guest_ea, ulong ea_mask) | 165 | static void kvmppc_mmu_pte_flush_long(struct kvm_vcpu *vcpu, ulong guest_ea) |
154 | { | 166 | { |
155 | u64 i; | 167 | struct hlist_head *list; |
168 | struct hlist_node *node; | ||
169 | struct hpte_cache *pte; | ||
170 | |||
171 | /* Find the list of entries in the map */ | ||
172 | list = &vcpu->arch.hpte_hash_pte_long[ | ||
173 | kvmppc_mmu_hash_pte_long(guest_ea)]; | ||
156 | 174 | ||
175 | rcu_read_lock(); | ||
176 | |||
177 | /* Check the list for matching entries and invalidate */ | ||
178 | hlist_for_each_entry_rcu(pte, node, list, list_pte_long) | ||
179 | if ((pte->pte.eaddr & 0x0ffff000UL) == guest_ea) | ||
180 | invalidate_pte(vcpu, pte); | ||
181 | |||
182 | rcu_read_unlock(); | ||
183 | } | ||
184 | |||
185 | void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong guest_ea, ulong ea_mask) | ||
186 | { | ||
157 | dprintk_mmu("KVM: Flushing %d Shadow PTEs: 0x%lx & 0x%lx\n", | 187 | dprintk_mmu("KVM: Flushing %d Shadow PTEs: 0x%lx & 0x%lx\n", |
158 | vcpu->arch.hpte_cache_count, guest_ea, ea_mask); | 188 | vcpu->arch.hpte_cache_count, guest_ea, ea_mask); |
159 | 189 | ||
@@ -164,9 +194,7 @@ void kvmppc_mmu_pte_flush(struct kvm_vcpu *vcpu, ulong guest_ea, ulong ea_mask) | |||
164 | kvmppc_mmu_pte_flush_page(vcpu, guest_ea); | 194 | kvmppc_mmu_pte_flush_page(vcpu, guest_ea); |
165 | break; | 195 | break; |
166 | case 0x0ffff000: | 196 | case 0x0ffff000: |
167 | /* 32-bit flush w/o segment, go through all possible segments */ | 197 | kvmppc_mmu_pte_flush_long(vcpu, guest_ea); |
168 | for (i = 0; i < 0x100000000ULL; i += 0x10000000ULL) | ||
169 | kvmppc_mmu_pte_flush(vcpu, guest_ea | i, ~0xfffUL); | ||
170 | break; | 198 | break; |
171 | case 0: | 199 | case 0: |
172 | /* Doing a complete flush -> start from scratch */ | 200 | /* Doing a complete flush -> start from scratch */ |
@@ -292,6 +320,8 @@ int kvmppc_mmu_hpte_init(struct kvm_vcpu *vcpu) | |||
292 | /* init hpte lookup hashes */ | 320 | /* init hpte lookup hashes */ |
293 | kvmppc_mmu_hpte_init_hash(vcpu->arch.hpte_hash_pte, | 321 | kvmppc_mmu_hpte_init_hash(vcpu->arch.hpte_hash_pte, |
294 | ARRAY_SIZE(vcpu->arch.hpte_hash_pte)); | 322 | ARRAY_SIZE(vcpu->arch.hpte_hash_pte)); |
323 | kvmppc_mmu_hpte_init_hash(vcpu->arch.hpte_hash_pte_long, | ||
324 | ARRAY_SIZE(vcpu->arch.hpte_hash_pte_long)); | ||
295 | kvmppc_mmu_hpte_init_hash(vcpu->arch.hpte_hash_vpte, | 325 | kvmppc_mmu_hpte_init_hash(vcpu->arch.hpte_hash_vpte, |
296 | ARRAY_SIZE(vcpu->arch.hpte_hash_vpte)); | 326 | ARRAY_SIZE(vcpu->arch.hpte_hash_vpte)); |
297 | kvmppc_mmu_hpte_init_hash(vcpu->arch.hpte_hash_vpte_long, | 327 | kvmppc_mmu_hpte_init_hash(vcpu->arch.hpte_hash_vpte_long, |