aboutsummaryrefslogtreecommitdiffstats
path: root/arch
diff options
context:
space:
mode:
authorTakuya Yoshikawa <yoshikawa.takuya@oss.ntt.co.jp>2012-07-02 04:58:48 -0400
committerMarcelo Tosatti <mtosatti@redhat.com>2012-07-18 15:55:04 -0400
commitf395302e09ef783b8f82d1160510a95aa8c66dbc (patch)
tree0a9b6a0520533cf2ae44e5cc1084ad04ccd68dc6 /arch
parent048212d0bc0b1769a4bbecd7ace8c8d237577d1b (diff)
KVM: MMU: Push trace_kvm_age_page() into kvm_age_rmapp()
This restricts the tracing to page aging and makes it possible to optimize kvm_handle_hva_range() further in the following patch. Signed-off-by: Takuya Yoshikawa <yoshikawa.takuya@oss.ntt.co.jp> Signed-off-by: Marcelo Tosatti <mtosatti@redhat.com>
Diffstat (limited to 'arch')
-rw-r--r--arch/x86/kvm/mmu.c23
1 files changed, 10 insertions, 13 deletions
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
index dfd7a9a31154..58adec384489 100644
--- a/arch/x86/kvm/mmu.c
+++ b/arch/x86/kvm/mmu.c
@@ -1269,8 +1269,7 @@ static int kvm_handle_hva_range(struct kvm *kvm,
1269 unsigned long data)) 1269 unsigned long data))
1270{ 1270{
1271 int j; 1271 int j;
1272 int ret; 1272 int ret = 0;
1273 int retval = 0;
1274 struct kvm_memslots *slots; 1273 struct kvm_memslots *slots;
1275 struct kvm_memory_slot *memslot; 1274 struct kvm_memory_slot *memslot;
1276 1275
@@ -1293,8 +1292,6 @@ static int kvm_handle_hva_range(struct kvm *kvm,
1293 gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot); 1292 gfn_end = hva_to_gfn_memslot(hva_end + PAGE_SIZE - 1, memslot);
1294 1293
1295 for (; gfn < gfn_end; ++gfn) { 1294 for (; gfn < gfn_end; ++gfn) {
1296 ret = 0;
1297
1298 for (j = PT_PAGE_TABLE_LEVEL; 1295 for (j = PT_PAGE_TABLE_LEVEL;
1299 j < PT_PAGE_TABLE_LEVEL + KVM_NR_PAGE_SIZES; ++j) { 1296 j < PT_PAGE_TABLE_LEVEL + KVM_NR_PAGE_SIZES; ++j) {
1300 unsigned long *rmapp; 1297 unsigned long *rmapp;
@@ -1302,14 +1299,10 @@ static int kvm_handle_hva_range(struct kvm *kvm,
1302 rmapp = __gfn_to_rmap(gfn, j, memslot); 1299 rmapp = __gfn_to_rmap(gfn, j, memslot);
1303 ret |= handler(kvm, rmapp, memslot, data); 1300 ret |= handler(kvm, rmapp, memslot, data);
1304 } 1301 }
1305 trace_kvm_age_page(memslot->userspace_addr +
1306 (gfn - memslot->base_gfn) * PAGE_SIZE,
1307 memslot, ret);
1308 retval |= ret;
1309 } 1302 }
1310 } 1303 }
1311 1304
1312 return retval; 1305 return ret;
1313} 1306}
1314 1307
1315static int kvm_handle_hva(struct kvm *kvm, unsigned long hva, 1308static int kvm_handle_hva(struct kvm *kvm, unsigned long hva,
@@ -1351,8 +1344,10 @@ static int kvm_age_rmapp(struct kvm *kvm, unsigned long *rmapp,
1351 * This has some overhead, but not as much as the cost of swapping 1344 * This has some overhead, but not as much as the cost of swapping
1352 * out actively used pages or breaking up actively used hugepages. 1345 * out actively used pages or breaking up actively used hugepages.
1353 */ 1346 */
1354 if (!shadow_accessed_mask) 1347 if (!shadow_accessed_mask) {
1355 return kvm_unmap_rmapp(kvm, rmapp, slot, data); 1348 young = kvm_unmap_rmapp(kvm, rmapp, slot, data);
1349 goto out;
1350 }
1356 1351
1357 for (sptep = rmap_get_first(*rmapp, &iter); sptep; 1352 for (sptep = rmap_get_first(*rmapp, &iter); sptep;
1358 sptep = rmap_get_next(&iter)) { 1353 sptep = rmap_get_next(&iter)) {
@@ -1364,7 +1359,9 @@ static int kvm_age_rmapp(struct kvm *kvm, unsigned long *rmapp,
1364 (unsigned long *)sptep); 1359 (unsigned long *)sptep);
1365 } 1360 }
1366 } 1361 }
1367 1362out:
1363 /* @data has hva passed to kvm_age_hva(). */
1364 trace_kvm_age_page(data, slot, young);
1368 return young; 1365 return young;
1369} 1366}
1370 1367
@@ -1413,7 +1410,7 @@ static void rmap_recycle(struct kvm_vcpu *vcpu, u64 *spte, gfn_t gfn)
1413 1410
1414int kvm_age_hva(struct kvm *kvm, unsigned long hva) 1411int kvm_age_hva(struct kvm *kvm, unsigned long hva)
1415{ 1412{
1416 return kvm_handle_hva(kvm, hva, 0, kvm_age_rmapp); 1413 return kvm_handle_hva(kvm, hva, hva, kvm_age_rmapp);
1417} 1414}
1418 1415
1419int kvm_test_age_hva(struct kvm *kvm, unsigned long hva) 1416int kvm_test_age_hva(struct kvm *kvm, unsigned long hva)