aboutsummaryrefslogtreecommitdiffstats
path: root/arch/x86/kvm
diff options
context:
space:
mode:
authorMarcelo Tosatti <mtosatti@redhat.com>2008-12-01 19:32:02 -0500
committerAvi Kivity <avi@redhat.com>2008-12-31 09:55:43 -0500
commit60c8aec6e2c9923492dabbd6b67e34692bd26c20 (patch)
treefa2d3b7f90f8dff8669a6fb6c85a41b9afb0f4c5 /arch/x86/kvm
parentfbce554e940a983d005e29849636d0ef54b3eb18 (diff)
KVM: MMU: use page array in unsync walk
Instead of invoking the handler directly collect pages into an array so the caller can work with it. Simplifies TLB flush collapsing. Signed-off-by: Marcelo Tosatti <mtosatti@redhat.com> Signed-off-by: Avi Kivity <avi@redhat.com>
Diffstat (limited to 'arch/x86/kvm')
-rw-r--r--arch/x86/kvm/mmu.c195
1 files changed, 140 insertions, 55 deletions
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
index dd20b199a7c0..7ce92f78f337 100644
--- a/arch/x86/kvm/mmu.c
+++ b/arch/x86/kvm/mmu.c
@@ -908,8 +908,9 @@ static void kvm_mmu_update_unsync_bitmap(u64 *spte)
908 struct kvm_mmu_page *sp = page_header(__pa(spte)); 908 struct kvm_mmu_page *sp = page_header(__pa(spte));
909 909
910 index = spte - sp->spt; 910 index = spte - sp->spt;
911 __set_bit(index, sp->unsync_child_bitmap); 911 if (!__test_and_set_bit(index, sp->unsync_child_bitmap))
912 sp->unsync_children = 1; 912 sp->unsync_children++;
913 WARN_ON(!sp->unsync_children);
913} 914}
914 915
915static void kvm_mmu_update_parents_unsync(struct kvm_mmu_page *sp) 916static void kvm_mmu_update_parents_unsync(struct kvm_mmu_page *sp)
@@ -936,7 +937,6 @@ static void kvm_mmu_update_parents_unsync(struct kvm_mmu_page *sp)
936 937
937static int unsync_walk_fn(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) 938static int unsync_walk_fn(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)
938{ 939{
939 sp->unsync_children = 1;
940 kvm_mmu_update_parents_unsync(sp); 940 kvm_mmu_update_parents_unsync(sp);
941 return 1; 941 return 1;
942} 942}
@@ -967,18 +967,41 @@ static void nonpaging_invlpg(struct kvm_vcpu *vcpu, gva_t gva)
967{ 967{
968} 968}
969 969
970#define KVM_PAGE_ARRAY_NR 16
971
972struct kvm_mmu_pages {
973 struct mmu_page_and_offset {
974 struct kvm_mmu_page *sp;
975 unsigned int idx;
976 } page[KVM_PAGE_ARRAY_NR];
977 unsigned int nr;
978};
979
970#define for_each_unsync_children(bitmap, idx) \ 980#define for_each_unsync_children(bitmap, idx) \
971 for (idx = find_first_bit(bitmap, 512); \ 981 for (idx = find_first_bit(bitmap, 512); \
972 idx < 512; \ 982 idx < 512; \
973 idx = find_next_bit(bitmap, 512, idx+1)) 983 idx = find_next_bit(bitmap, 512, idx+1))
974 984
975static int mmu_unsync_walk(struct kvm_mmu_page *sp, 985int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp,
976 struct kvm_unsync_walk *walker) 986 int idx)
977{ 987{
978 int i, ret; 988 int i;
979 989
980 if (!sp->unsync_children) 990 if (sp->unsync)
981 return 0; 991 for (i=0; i < pvec->nr; i++)
992 if (pvec->page[i].sp == sp)
993 return 0;
994
995 pvec->page[pvec->nr].sp = sp;
996 pvec->page[pvec->nr].idx = idx;
997 pvec->nr++;
998 return (pvec->nr == KVM_PAGE_ARRAY_NR);
999}
1000
1001static int __mmu_unsync_walk(struct kvm_mmu_page *sp,
1002 struct kvm_mmu_pages *pvec)
1003{
1004 int i, ret, nr_unsync_leaf = 0;
982 1005
983 for_each_unsync_children(sp->unsync_child_bitmap, i) { 1006 for_each_unsync_children(sp->unsync_child_bitmap, i) {
984 u64 ent = sp->spt[i]; 1007 u64 ent = sp->spt[i];
@@ -988,17 +1011,22 @@ static int mmu_unsync_walk(struct kvm_mmu_page *sp,
988 child = page_header(ent & PT64_BASE_ADDR_MASK); 1011 child = page_header(ent & PT64_BASE_ADDR_MASK);
989 1012
990 if (child->unsync_children) { 1013 if (child->unsync_children) {
991 ret = mmu_unsync_walk(child, walker); 1014 if (mmu_pages_add(pvec, child, i))
992 if (ret) 1015 return -ENOSPC;
1016
1017 ret = __mmu_unsync_walk(child, pvec);
1018 if (!ret)
1019 __clear_bit(i, sp->unsync_child_bitmap);
1020 else if (ret > 0)
1021 nr_unsync_leaf += ret;
1022 else
993 return ret; 1023 return ret;
994 __clear_bit(i, sp->unsync_child_bitmap);
995 } 1024 }
996 1025
997 if (child->unsync) { 1026 if (child->unsync) {
998 ret = walker->entry(child, walker); 1027 nr_unsync_leaf++;
999 __clear_bit(i, sp->unsync_child_bitmap); 1028 if (mmu_pages_add(pvec, child, i))
1000 if (ret) 1029 return -ENOSPC;
1001 return ret;
1002 } 1030 }
1003 } 1031 }
1004 } 1032 }
@@ -1006,7 +1034,17 @@ static int mmu_unsync_walk(struct kvm_mmu_page *sp,
1006 if (find_first_bit(sp->unsync_child_bitmap, 512) == 512) 1034 if (find_first_bit(sp->unsync_child_bitmap, 512) == 512)
1007 sp->unsync_children = 0; 1035 sp->unsync_children = 0;
1008 1036
1009 return 0; 1037 return nr_unsync_leaf;
1038}
1039
1040static int mmu_unsync_walk(struct kvm_mmu_page *sp,
1041 struct kvm_mmu_pages *pvec)
1042{
1043 if (!sp->unsync_children)
1044 return 0;
1045
1046 mmu_pages_add(pvec, sp, 0);
1047 return __mmu_unsync_walk(sp, pvec);
1010} 1048}
1011 1049
1012static struct kvm_mmu_page *kvm_mmu_lookup_page(struct kvm *kvm, gfn_t gfn) 1050static struct kvm_mmu_page *kvm_mmu_lookup_page(struct kvm *kvm, gfn_t gfn)
@@ -1056,30 +1094,81 @@ static int kvm_sync_page(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp)
1056 return 0; 1094 return 0;
1057} 1095}
1058 1096
1059struct sync_walker { 1097struct mmu_page_path {
1060 struct kvm_vcpu *vcpu; 1098 struct kvm_mmu_page *parent[PT64_ROOT_LEVEL-1];
1061 struct kvm_unsync_walk walker; 1099 unsigned int idx[PT64_ROOT_LEVEL-1];
1062}; 1100};
1063 1101
1064static int mmu_sync_fn(struct kvm_mmu_page *sp, struct kvm_unsync_walk *walk) 1102#define for_each_sp(pvec, sp, parents, i) \
1103 for (i = mmu_pages_next(&pvec, &parents, -1), \
1104 sp = pvec.page[i].sp; \
1105 i < pvec.nr && ({ sp = pvec.page[i].sp; 1;}); \
1106 i = mmu_pages_next(&pvec, &parents, i))
1107
1108int mmu_pages_next(struct kvm_mmu_pages *pvec, struct mmu_page_path *parents,
1109 int i)
1065{ 1110{
1066 struct sync_walker *sync_walk = container_of(walk, struct sync_walker, 1111 int n;
1067 walker); 1112
1068 struct kvm_vcpu *vcpu = sync_walk->vcpu; 1113 for (n = i+1; n < pvec->nr; n++) {
1114 struct kvm_mmu_page *sp = pvec->page[n].sp;
1115
1116 if (sp->role.level == PT_PAGE_TABLE_LEVEL) {
1117 parents->idx[0] = pvec->page[n].idx;
1118 return n;
1119 }
1069 1120
1070 kvm_sync_page(vcpu, sp); 1121 parents->parent[sp->role.level-2] = sp;
1071 return (need_resched() || spin_needbreak(&vcpu->kvm->mmu_lock)); 1122 parents->idx[sp->role.level-1] = pvec->page[n].idx;
1123 }
1124
1125 return n;
1072} 1126}
1073 1127
1074static void mmu_sync_children(struct kvm_vcpu *vcpu, struct kvm_mmu_page *sp) 1128void mmu_pages_clear_parents(struct mmu_page_path *parents)
1075{ 1129{
1076 struct sync_walker walker = { 1130 struct kvm_mmu_page *sp;
1077 .walker = { .entry = mmu_sync_fn, }, 1131 unsigned int level = 0;
1078 .vcpu = vcpu, 1132
1079 }; 1133 do {
1134 unsigned int idx = parents->idx[level];
1135
1136 sp = parents->parent[level];
1137 if (!sp)
1138 return;
1139
1140 --sp->unsync_children;
1141 WARN_ON((int)sp->unsync_children < 0);
1142 __clear_bit(idx, sp->unsync_child_bitmap);
1143 level++;
1144 } while (level < PT64_ROOT_LEVEL-1 && !sp->unsync_children);
1145}
1146
1147static void kvm_mmu_pages_init(struct kvm_mmu_page *parent,
1148 struct mmu_page_path *parents,
1149 struct kvm_mmu_pages *pvec)
1150{
1151 parents->parent[parent->role.level-1] = NULL;
1152 pvec->nr = 0;
1153}
1080 1154
1081 while (mmu_unsync_walk(sp, &walker.walker)) 1155static void mmu_sync_children(struct kvm_vcpu *vcpu,
1156 struct kvm_mmu_page *parent)
1157{
1158 int i;
1159 struct kvm_mmu_page *sp;
1160 struct mmu_page_path parents;
1161 struct kvm_mmu_pages pages;
1162
1163 kvm_mmu_pages_init(parent, &parents, &pages);
1164 while (mmu_unsync_walk(parent, &pages)) {
1165 for_each_sp(pages, sp, parents, i) {
1166 kvm_sync_page(vcpu, sp);
1167 mmu_pages_clear_parents(&parents);
1168 }
1082 cond_resched_lock(&vcpu->kvm->mmu_lock); 1169 cond_resched_lock(&vcpu->kvm->mmu_lock);
1170 kvm_mmu_pages_init(parent, &parents, &pages);
1171 }
1083} 1172}
1084 1173
1085static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu, 1174static struct kvm_mmu_page *kvm_mmu_get_page(struct kvm_vcpu *vcpu,
@@ -1245,33 +1334,29 @@ static void kvm_mmu_unlink_parents(struct kvm *kvm, struct kvm_mmu_page *sp)
1245 } 1334 }
1246} 1335}
1247 1336
1248struct zap_walker { 1337static int mmu_zap_unsync_children(struct kvm *kvm,
1249 struct kvm_unsync_walk walker; 1338 struct kvm_mmu_page *parent)
1250 struct kvm *kvm;
1251 int zapped;
1252};
1253
1254static int mmu_zap_fn(struct kvm_mmu_page *sp, struct kvm_unsync_walk *walk)
1255{
1256 struct zap_walker *zap_walk = container_of(walk, struct zap_walker,
1257 walker);
1258 kvm_mmu_zap_page(zap_walk->kvm, sp);
1259 zap_walk->zapped = 1;
1260 return 0;
1261}
1262
1263static int mmu_zap_unsync_children(struct kvm *kvm, struct kvm_mmu_page *sp)
1264{ 1339{
1265 struct zap_walker walker = { 1340 int i, zapped = 0;
1266 .walker = { .entry = mmu_zap_fn, }, 1341 struct mmu_page_path parents;
1267 .kvm = kvm, 1342 struct kvm_mmu_pages pages;
1268 .zapped = 0,
1269 };
1270 1343
1271 if (sp->role.level == PT_PAGE_TABLE_LEVEL) 1344 if (parent->role.level == PT_PAGE_TABLE_LEVEL)
1272 return 0; 1345 return 0;
1273 mmu_unsync_walk(sp, &walker.walker); 1346
1274 return walker.zapped; 1347 kvm_mmu_pages_init(parent, &parents, &pages);
1348 while (mmu_unsync_walk(parent, &pages)) {
1349 struct kvm_mmu_page *sp;
1350
1351 for_each_sp(pages, sp, parents, i) {
1352 kvm_mmu_zap_page(kvm, sp);
1353 mmu_pages_clear_parents(&parents);
1354 }
1355 zapped += pages.nr;
1356 kvm_mmu_pages_init(parent, &parents, &pages);
1357 }
1358
1359 return zapped;
1275} 1360}
1276 1361
1277static int kvm_mmu_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp) 1362static int kvm_mmu_zap_page(struct kvm *kvm, struct kvm_mmu_page *sp)