aboutsummaryrefslogtreecommitdiffstats
path: root/arch/x86/kvm
diff options
context:
space:
mode:
authorIzik Eidus <izike@qumranet.com>2008-02-10 11:04:15 -0500
committerAvi Kivity <avi@qumranet.com>2008-03-04 08:19:40 -0500
commit72dc67a69690288538142df73a7e3ac66fea68dc (patch)
treef40cc5ef0c66686a469977fd438e5b6786f16280 /arch/x86/kvm
parentc7ac679c160db864810920df61a6ed14275011aa (diff)
KVM: remove the usage of the mmap_sem for the protection of the memory slots.
This patch replaces the mmap_sem lock for the memory slots with a new kvm private lock, it is needed beacuse untill now there were cases where kvm accesses user memory while holding the mmap semaphore. Signed-off-by: Izik Eidus <izike@qumranet.com> Signed-off-by: Avi Kivity <avi@qumranet.com>
Diffstat (limited to 'arch/x86/kvm')
-rw-r--r--arch/x86/kvm/mmu.c24
-rw-r--r--arch/x86/kvm/paging_tmpl.h13
-rw-r--r--arch/x86/kvm/vmx.c7
-rw-r--r--arch/x86/kvm/x86.c65
4 files changed, 71 insertions, 38 deletions
diff --git a/arch/x86/kvm/mmu.c b/arch/x86/kvm/mmu.c
index 8efdcdbebb03..26037106ad19 100644
--- a/arch/x86/kvm/mmu.c
+++ b/arch/x86/kvm/mmu.c
@@ -876,11 +876,18 @@ static void page_header_update_slot(struct kvm *kvm, void *pte, gfn_t gfn)
876 876
877struct page *gva_to_page(struct kvm_vcpu *vcpu, gva_t gva) 877struct page *gva_to_page(struct kvm_vcpu *vcpu, gva_t gva)
878{ 878{
879 struct page *page;
880
879 gpa_t gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, gva); 881 gpa_t gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, gva);
880 882
881 if (gpa == UNMAPPED_GVA) 883 if (gpa == UNMAPPED_GVA)
882 return NULL; 884 return NULL;
883 return gfn_to_page(vcpu->kvm, gpa >> PAGE_SHIFT); 885
886 down_read(&current->mm->mmap_sem);
887 page = gfn_to_page(vcpu->kvm, gpa >> PAGE_SHIFT);
888 up_read(&current->mm->mmap_sem);
889
890 return page;
884} 891}
885 892
886static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *shadow_pte, 893static void mmu_set_spte(struct kvm_vcpu *vcpu, u64 *shadow_pte,
@@ -1020,15 +1027,18 @@ static int nonpaging_map(struct kvm_vcpu *vcpu, gva_t v, int write, gfn_t gfn)
1020 1027
1021 struct page *page; 1028 struct page *page;
1022 1029
1030 down_read(&vcpu->kvm->slots_lock);
1031
1023 down_read(&current->mm->mmap_sem); 1032 down_read(&current->mm->mmap_sem);
1024 page = gfn_to_page(vcpu->kvm, gfn); 1033 page = gfn_to_page(vcpu->kvm, gfn);
1034 up_read(&current->mm->mmap_sem);
1025 1035
1026 spin_lock(&vcpu->kvm->mmu_lock); 1036 spin_lock(&vcpu->kvm->mmu_lock);
1027 kvm_mmu_free_some_pages(vcpu); 1037 kvm_mmu_free_some_pages(vcpu);
1028 r = __nonpaging_map(vcpu, v, write, gfn, page); 1038 r = __nonpaging_map(vcpu, v, write, gfn, page);
1029 spin_unlock(&vcpu->kvm->mmu_lock); 1039 spin_unlock(&vcpu->kvm->mmu_lock);
1030 1040
1031 up_read(&current->mm->mmap_sem); 1041 up_read(&vcpu->kvm->slots_lock);
1032 1042
1033 return r; 1043 return r;
1034} 1044}
@@ -1362,6 +1372,7 @@ static void mmu_guess_page_from_pte_write(struct kvm_vcpu *vcpu, gpa_t gpa,
1362 gfn_t gfn; 1372 gfn_t gfn;
1363 int r; 1373 int r;
1364 u64 gpte = 0; 1374 u64 gpte = 0;
1375 struct page *page;
1365 1376
1366 if (bytes != 4 && bytes != 8) 1377 if (bytes != 4 && bytes != 8)
1367 return; 1378 return;
@@ -1389,6 +1400,11 @@ static void mmu_guess_page_from_pte_write(struct kvm_vcpu *vcpu, gpa_t gpa,
1389 if (!is_present_pte(gpte)) 1400 if (!is_present_pte(gpte))
1390 return; 1401 return;
1391 gfn = (gpte & PT64_BASE_ADDR_MASK) >> PAGE_SHIFT; 1402 gfn = (gpte & PT64_BASE_ADDR_MASK) >> PAGE_SHIFT;
1403
1404 down_read(&current->mm->mmap_sem);
1405 page = gfn_to_page(vcpu->kvm, gfn);
1406 up_read(&current->mm->mmap_sem);
1407
1392 vcpu->arch.update_pte.gfn = gfn; 1408 vcpu->arch.update_pte.gfn = gfn;
1393 vcpu->arch.update_pte.page = gfn_to_page(vcpu->kvm, gfn); 1409 vcpu->arch.update_pte.page = gfn_to_page(vcpu->kvm, gfn);
1394} 1410}
@@ -1496,9 +1512,9 @@ int kvm_mmu_unprotect_page_virt(struct kvm_vcpu *vcpu, gva_t gva)
1496 gpa_t gpa; 1512 gpa_t gpa;
1497 int r; 1513 int r;
1498 1514
1499 down_read(&current->mm->mmap_sem); 1515 down_read(&vcpu->kvm->slots_lock);
1500 gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, gva); 1516 gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, gva);
1501 up_read(&current->mm->mmap_sem); 1517 up_read(&vcpu->kvm->slots_lock);
1502 1518
1503 spin_lock(&vcpu->kvm->mmu_lock); 1519 spin_lock(&vcpu->kvm->mmu_lock);
1504 r = kvm_mmu_unprotect_page(vcpu->kvm, gpa >> PAGE_SHIFT); 1520 r = kvm_mmu_unprotect_page(vcpu->kvm, gpa >> PAGE_SHIFT);
diff --git a/arch/x86/kvm/paging_tmpl.h b/arch/x86/kvm/paging_tmpl.h
index 03ba8608fe0f..2009c6e9dc4d 100644
--- a/arch/x86/kvm/paging_tmpl.h
+++ b/arch/x86/kvm/paging_tmpl.h
@@ -91,7 +91,10 @@ static bool FNAME(cmpxchg_gpte)(struct kvm *kvm,
91 pt_element_t *table; 91 pt_element_t *table;
92 struct page *page; 92 struct page *page;
93 93
94 down_read(&current->mm->mmap_sem);
94 page = gfn_to_page(kvm, table_gfn); 95 page = gfn_to_page(kvm, table_gfn);
96 up_read(&current->mm->mmap_sem);
97
95 table = kmap_atomic(page, KM_USER0); 98 table = kmap_atomic(page, KM_USER0);
96 99
97 ret = CMPXCHG(&table[index], orig_pte, new_pte); 100 ret = CMPXCHG(&table[index], orig_pte, new_pte);
@@ -378,7 +381,7 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr,
378 if (r) 381 if (r)
379 return r; 382 return r;
380 383
381 down_read(&current->mm->mmap_sem); 384 down_read(&vcpu->kvm->slots_lock);
382 /* 385 /*
383 * Look up the shadow pte for the faulting address. 386 * Look up the shadow pte for the faulting address.
384 */ 387 */
@@ -392,11 +395,13 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr,
392 pgprintk("%s: guest page fault\n", __FUNCTION__); 395 pgprintk("%s: guest page fault\n", __FUNCTION__);
393 inject_page_fault(vcpu, addr, walker.error_code); 396 inject_page_fault(vcpu, addr, walker.error_code);
394 vcpu->arch.last_pt_write_count = 0; /* reset fork detector */ 397 vcpu->arch.last_pt_write_count = 0; /* reset fork detector */
395 up_read(&current->mm->mmap_sem); 398 up_read(&vcpu->kvm->slots_lock);
396 return 0; 399 return 0;
397 } 400 }
398 401
402 down_read(&current->mm->mmap_sem);
399 page = gfn_to_page(vcpu->kvm, walker.gfn); 403 page = gfn_to_page(vcpu->kvm, walker.gfn);
404 up_read(&current->mm->mmap_sem);
400 405
401 spin_lock(&vcpu->kvm->mmu_lock); 406 spin_lock(&vcpu->kvm->mmu_lock);
402 kvm_mmu_free_some_pages(vcpu); 407 kvm_mmu_free_some_pages(vcpu);
@@ -413,14 +418,14 @@ static int FNAME(page_fault)(struct kvm_vcpu *vcpu, gva_t addr,
413 */ 418 */
414 if (shadow_pte && is_io_pte(*shadow_pte)) { 419 if (shadow_pte && is_io_pte(*shadow_pte)) {
415 spin_unlock(&vcpu->kvm->mmu_lock); 420 spin_unlock(&vcpu->kvm->mmu_lock);
416 up_read(&current->mm->mmap_sem); 421 up_read(&vcpu->kvm->slots_lock);
417 return 1; 422 return 1;
418 } 423 }
419 424
420 ++vcpu->stat.pf_fixed; 425 ++vcpu->stat.pf_fixed;
421 kvm_mmu_audit(vcpu, "post page fault (fixed)"); 426 kvm_mmu_audit(vcpu, "post page fault (fixed)");
422 spin_unlock(&vcpu->kvm->mmu_lock); 427 spin_unlock(&vcpu->kvm->mmu_lock);
423 up_read(&current->mm->mmap_sem); 428 up_read(&vcpu->kvm->slots_lock);
424 429
425 return write_pt; 430 return write_pt;
426} 431}
diff --git a/arch/x86/kvm/vmx.c b/arch/x86/kvm/vmx.c
index ad36447e696e..86f5bf121838 100644
--- a/arch/x86/kvm/vmx.c
+++ b/arch/x86/kvm/vmx.c
@@ -1477,7 +1477,7 @@ static int alloc_apic_access_page(struct kvm *kvm)
1477 struct kvm_userspace_memory_region kvm_userspace_mem; 1477 struct kvm_userspace_memory_region kvm_userspace_mem;
1478 int r = 0; 1478 int r = 0;
1479 1479
1480 down_write(&current->mm->mmap_sem); 1480 down_write(&kvm->slots_lock);
1481 if (kvm->arch.apic_access_page) 1481 if (kvm->arch.apic_access_page)
1482 goto out; 1482 goto out;
1483 kvm_userspace_mem.slot = APIC_ACCESS_PAGE_PRIVATE_MEMSLOT; 1483 kvm_userspace_mem.slot = APIC_ACCESS_PAGE_PRIVATE_MEMSLOT;
@@ -1487,9 +1487,12 @@ static int alloc_apic_access_page(struct kvm *kvm)
1487 r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, 0); 1487 r = __kvm_set_memory_region(kvm, &kvm_userspace_mem, 0);
1488 if (r) 1488 if (r)
1489 goto out; 1489 goto out;
1490
1491 down_read(&current->mm->mmap_sem);
1490 kvm->arch.apic_access_page = gfn_to_page(kvm, 0xfee00); 1492 kvm->arch.apic_access_page = gfn_to_page(kvm, 0xfee00);
1493 up_read(&current->mm->mmap_sem);
1491out: 1494out:
1492 up_write(&current->mm->mmap_sem); 1495 up_write(&kvm->slots_lock);
1493 return r; 1496 return r;
1494} 1497}
1495 1498
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
index 338764fa5391..6b01552bd1f1 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -184,7 +184,7 @@ int load_pdptrs(struct kvm_vcpu *vcpu, unsigned long cr3)
184 int ret; 184 int ret;
185 u64 pdpte[ARRAY_SIZE(vcpu->arch.pdptrs)]; 185 u64 pdpte[ARRAY_SIZE(vcpu->arch.pdptrs)];
186 186
187 down_read(&current->mm->mmap_sem); 187 down_read(&vcpu->kvm->slots_lock);
188 ret = kvm_read_guest_page(vcpu->kvm, pdpt_gfn, pdpte, 188 ret = kvm_read_guest_page(vcpu->kvm, pdpt_gfn, pdpte,
189 offset * sizeof(u64), sizeof(pdpte)); 189 offset * sizeof(u64), sizeof(pdpte));
190 if (ret < 0) { 190 if (ret < 0) {
@@ -201,7 +201,7 @@ int load_pdptrs(struct kvm_vcpu *vcpu, unsigned long cr3)
201 201
202 memcpy(vcpu->arch.pdptrs, pdpte, sizeof(vcpu->arch.pdptrs)); 202 memcpy(vcpu->arch.pdptrs, pdpte, sizeof(vcpu->arch.pdptrs));
203out: 203out:
204 up_read(&current->mm->mmap_sem); 204 up_read(&vcpu->kvm->slots_lock);
205 205
206 return ret; 206 return ret;
207} 207}
@@ -215,13 +215,13 @@ static bool pdptrs_changed(struct kvm_vcpu *vcpu)
215 if (is_long_mode(vcpu) || !is_pae(vcpu)) 215 if (is_long_mode(vcpu) || !is_pae(vcpu))
216 return false; 216 return false;
217 217
218 down_read(&current->mm->mmap_sem); 218 down_read(&vcpu->kvm->slots_lock);
219 r = kvm_read_guest(vcpu->kvm, vcpu->arch.cr3 & ~31u, pdpte, sizeof(pdpte)); 219 r = kvm_read_guest(vcpu->kvm, vcpu->arch.cr3 & ~31u, pdpte, sizeof(pdpte));
220 if (r < 0) 220 if (r < 0)
221 goto out; 221 goto out;
222 changed = memcmp(pdpte, vcpu->arch.pdptrs, sizeof(pdpte)) != 0; 222 changed = memcmp(pdpte, vcpu->arch.pdptrs, sizeof(pdpte)) != 0;
223out: 223out:
224 up_read(&current->mm->mmap_sem); 224 up_read(&vcpu->kvm->slots_lock);
225 225
226 return changed; 226 return changed;
227} 227}
@@ -359,7 +359,7 @@ void set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3)
359 */ 359 */
360 } 360 }
361 361
362 down_read(&current->mm->mmap_sem); 362 down_read(&vcpu->kvm->slots_lock);
363 /* 363 /*
364 * Does the new cr3 value map to physical memory? (Note, we 364 * Does the new cr3 value map to physical memory? (Note, we
365 * catch an invalid cr3 even in real-mode, because it would 365 * catch an invalid cr3 even in real-mode, because it would
@@ -375,7 +375,7 @@ void set_cr3(struct kvm_vcpu *vcpu, unsigned long cr3)
375 vcpu->arch.cr3 = cr3; 375 vcpu->arch.cr3 = cr3;
376 vcpu->arch.mmu.new_cr3(vcpu); 376 vcpu->arch.mmu.new_cr3(vcpu);
377 } 377 }
378 up_read(&current->mm->mmap_sem); 378 up_read(&vcpu->kvm->slots_lock);
379} 379}
380EXPORT_SYMBOL_GPL(set_cr3); 380EXPORT_SYMBOL_GPL(set_cr3);
381 381
@@ -1232,12 +1232,12 @@ static int kvm_vm_ioctl_set_nr_mmu_pages(struct kvm *kvm,
1232 if (kvm_nr_mmu_pages < KVM_MIN_ALLOC_MMU_PAGES) 1232 if (kvm_nr_mmu_pages < KVM_MIN_ALLOC_MMU_PAGES)
1233 return -EINVAL; 1233 return -EINVAL;
1234 1234
1235 down_write(&current->mm->mmap_sem); 1235 down_write(&kvm->slots_lock);
1236 1236
1237 kvm_mmu_change_mmu_pages(kvm, kvm_nr_mmu_pages); 1237 kvm_mmu_change_mmu_pages(kvm, kvm_nr_mmu_pages);
1238 kvm->arch.n_requested_mmu_pages = kvm_nr_mmu_pages; 1238 kvm->arch.n_requested_mmu_pages = kvm_nr_mmu_pages;
1239 1239
1240 up_write(&current->mm->mmap_sem); 1240 up_write(&kvm->slots_lock);
1241 return 0; 1241 return 0;
1242} 1242}
1243 1243
@@ -1286,7 +1286,7 @@ static int kvm_vm_ioctl_set_memory_alias(struct kvm *kvm,
1286 < alias->target_phys_addr) 1286 < alias->target_phys_addr)
1287 goto out; 1287 goto out;
1288 1288
1289 down_write(&current->mm->mmap_sem); 1289 down_write(&kvm->slots_lock);
1290 1290
1291 p = &kvm->arch.aliases[alias->slot]; 1291 p = &kvm->arch.aliases[alias->slot];
1292 p->base_gfn = alias->guest_phys_addr >> PAGE_SHIFT; 1292 p->base_gfn = alias->guest_phys_addr >> PAGE_SHIFT;
@@ -1300,7 +1300,7 @@ static int kvm_vm_ioctl_set_memory_alias(struct kvm *kvm,
1300 1300
1301 kvm_mmu_zap_all(kvm); 1301 kvm_mmu_zap_all(kvm);
1302 1302
1303 up_write(&current->mm->mmap_sem); 1303 up_write(&kvm->slots_lock);
1304 1304
1305 return 0; 1305 return 0;
1306 1306
@@ -1376,7 +1376,7 @@ int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm,
1376 struct kvm_memory_slot *memslot; 1376 struct kvm_memory_slot *memslot;
1377 int is_dirty = 0; 1377 int is_dirty = 0;
1378 1378
1379 down_write(&current->mm->mmap_sem); 1379 down_write(&kvm->slots_lock);
1380 1380
1381 r = kvm_get_dirty_log(kvm, log, &is_dirty); 1381 r = kvm_get_dirty_log(kvm, log, &is_dirty);
1382 if (r) 1382 if (r)
@@ -1392,7 +1392,7 @@ int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm,
1392 } 1392 }
1393 r = 0; 1393 r = 0;
1394out: 1394out:
1395 up_write(&current->mm->mmap_sem); 1395 up_write(&kvm->slots_lock);
1396 return r; 1396 return r;
1397} 1397}
1398 1398
@@ -1570,7 +1570,7 @@ int emulator_read_std(unsigned long addr,
1570 void *data = val; 1570 void *data = val;
1571 int r = X86EMUL_CONTINUE; 1571 int r = X86EMUL_CONTINUE;
1572 1572
1573 down_read(&current->mm->mmap_sem); 1573 down_read(&vcpu->kvm->slots_lock);
1574 while (bytes) { 1574 while (bytes) {
1575 gpa_t gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, addr); 1575 gpa_t gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, addr);
1576 unsigned offset = addr & (PAGE_SIZE-1); 1576 unsigned offset = addr & (PAGE_SIZE-1);
@@ -1592,7 +1592,7 @@ int emulator_read_std(unsigned long addr,
1592 addr += tocopy; 1592 addr += tocopy;
1593 } 1593 }
1594out: 1594out:
1595 up_read(&current->mm->mmap_sem); 1595 up_read(&vcpu->kvm->slots_lock);
1596 return r; 1596 return r;
1597} 1597}
1598EXPORT_SYMBOL_GPL(emulator_read_std); 1598EXPORT_SYMBOL_GPL(emulator_read_std);
@@ -1611,9 +1611,9 @@ static int emulator_read_emulated(unsigned long addr,
1611 return X86EMUL_CONTINUE; 1611 return X86EMUL_CONTINUE;
1612 } 1612 }
1613 1613
1614 down_read(&current->mm->mmap_sem); 1614 down_read(&vcpu->kvm->slots_lock);
1615 gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, addr); 1615 gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, addr);
1616 up_read(&current->mm->mmap_sem); 1616 up_read(&vcpu->kvm->slots_lock);
1617 1617
1618 /* For APIC access vmexit */ 1618 /* For APIC access vmexit */
1619 if ((gpa & PAGE_MASK) == APIC_DEFAULT_PHYS_BASE) 1619 if ((gpa & PAGE_MASK) == APIC_DEFAULT_PHYS_BASE)
@@ -1651,14 +1651,14 @@ static int emulator_write_phys(struct kvm_vcpu *vcpu, gpa_t gpa,
1651{ 1651{
1652 int ret; 1652 int ret;
1653 1653
1654 down_read(&current->mm->mmap_sem); 1654 down_read(&vcpu->kvm->slots_lock);
1655 ret = kvm_write_guest(vcpu->kvm, gpa, val, bytes); 1655 ret = kvm_write_guest(vcpu->kvm, gpa, val, bytes);
1656 if (ret < 0) { 1656 if (ret < 0) {
1657 up_read(&current->mm->mmap_sem); 1657 up_read(&vcpu->kvm->slots_lock);
1658 return 0; 1658 return 0;
1659 } 1659 }
1660 kvm_mmu_pte_write(vcpu, gpa, val, bytes); 1660 kvm_mmu_pte_write(vcpu, gpa, val, bytes);
1661 up_read(&current->mm->mmap_sem); 1661 up_read(&vcpu->kvm->slots_lock);
1662 return 1; 1662 return 1;
1663} 1663}
1664 1664
@@ -1670,9 +1670,9 @@ static int emulator_write_emulated_onepage(unsigned long addr,
1670 struct kvm_io_device *mmio_dev; 1670 struct kvm_io_device *mmio_dev;
1671 gpa_t gpa; 1671 gpa_t gpa;
1672 1672
1673 down_read(&current->mm->mmap_sem); 1673 down_read(&vcpu->kvm->slots_lock);
1674 gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, addr); 1674 gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, addr);
1675 up_read(&current->mm->mmap_sem); 1675 up_read(&vcpu->kvm->slots_lock);
1676 1676
1677 if (gpa == UNMAPPED_GVA) { 1677 if (gpa == UNMAPPED_GVA) {
1678 kvm_inject_page_fault(vcpu, addr, 2); 1678 kvm_inject_page_fault(vcpu, addr, 2);
@@ -1749,7 +1749,7 @@ static int emulator_cmpxchg_emulated(unsigned long addr,
1749 char *kaddr; 1749 char *kaddr;
1750 u64 val; 1750 u64 val;
1751 1751
1752 down_read(&current->mm->mmap_sem); 1752 down_read(&vcpu->kvm->slots_lock);
1753 gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, addr); 1753 gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, addr);
1754 1754
1755 if (gpa == UNMAPPED_GVA || 1755 if (gpa == UNMAPPED_GVA ||
@@ -1760,13 +1760,17 @@ static int emulator_cmpxchg_emulated(unsigned long addr,
1760 goto emul_write; 1760 goto emul_write;
1761 1761
1762 val = *(u64 *)new; 1762 val = *(u64 *)new;
1763
1764 down_read(&current->mm->mmap_sem);
1763 page = gfn_to_page(vcpu->kvm, gpa >> PAGE_SHIFT); 1765 page = gfn_to_page(vcpu->kvm, gpa >> PAGE_SHIFT);
1766 up_read(&current->mm->mmap_sem);
1767
1764 kaddr = kmap_atomic(page, KM_USER0); 1768 kaddr = kmap_atomic(page, KM_USER0);
1765 set_64bit((u64 *)(kaddr + offset_in_page(gpa)), val); 1769 set_64bit((u64 *)(kaddr + offset_in_page(gpa)), val);
1766 kunmap_atomic(kaddr, KM_USER0); 1770 kunmap_atomic(kaddr, KM_USER0);
1767 kvm_release_page_dirty(page); 1771 kvm_release_page_dirty(page);
1768 emul_write: 1772 emul_write:
1769 up_read(&current->mm->mmap_sem); 1773 up_read(&vcpu->kvm->slots_lock);
1770 } 1774 }
1771#endif 1775#endif
1772 1776
@@ -2159,10 +2163,10 @@ int kvm_emulate_pio_string(struct kvm_vcpu *vcpu, struct kvm_run *run, int in,
2159 kvm_x86_ops->skip_emulated_instruction(vcpu); 2163 kvm_x86_ops->skip_emulated_instruction(vcpu);
2160 2164
2161 for (i = 0; i < nr_pages; ++i) { 2165 for (i = 0; i < nr_pages; ++i) {
2162 down_read(&current->mm->mmap_sem); 2166 down_read(&vcpu->kvm->slots_lock);
2163 page = gva_to_page(vcpu, address + i * PAGE_SIZE); 2167 page = gva_to_page(vcpu, address + i * PAGE_SIZE);
2164 vcpu->arch.pio.guest_pages[i] = page; 2168 vcpu->arch.pio.guest_pages[i] = page;
2165 up_read(&current->mm->mmap_sem); 2169 up_read(&vcpu->kvm->slots_lock);
2166 if (!page) { 2170 if (!page) {
2167 kvm_inject_gp(vcpu, 0); 2171 kvm_inject_gp(vcpu, 0);
2168 free_pio_guest_pages(vcpu); 2172 free_pio_guest_pages(vcpu);
@@ -2485,8 +2489,9 @@ static void vapic_enter(struct kvm_vcpu *vcpu)
2485 2489
2486 down_read(&current->mm->mmap_sem); 2490 down_read(&current->mm->mmap_sem);
2487 page = gfn_to_page(vcpu->kvm, apic->vapic_addr >> PAGE_SHIFT); 2491 page = gfn_to_page(vcpu->kvm, apic->vapic_addr >> PAGE_SHIFT);
2488 vcpu->arch.apic->vapic_page = page;
2489 up_read(&current->mm->mmap_sem); 2492 up_read(&current->mm->mmap_sem);
2493
2494 vcpu->arch.apic->vapic_page = page;
2490} 2495}
2491 2496
2492static void vapic_exit(struct kvm_vcpu *vcpu) 2497static void vapic_exit(struct kvm_vcpu *vcpu)
@@ -2959,9 +2964,9 @@ int kvm_arch_vcpu_ioctl_translate(struct kvm_vcpu *vcpu,
2959 gpa_t gpa; 2964 gpa_t gpa;
2960 2965
2961 vcpu_load(vcpu); 2966 vcpu_load(vcpu);
2962 down_read(&current->mm->mmap_sem); 2967 down_read(&vcpu->kvm->slots_lock);
2963 gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, vaddr); 2968 gpa = vcpu->arch.mmu.gva_to_gpa(vcpu, vaddr);
2964 up_read(&current->mm->mmap_sem); 2969 up_read(&vcpu->kvm->slots_lock);
2965 tr->physical_address = gpa; 2970 tr->physical_address = gpa;
2966 tr->valid = gpa != UNMAPPED_GVA; 2971 tr->valid = gpa != UNMAPPED_GVA;
2967 tr->writeable = 1; 2972 tr->writeable = 1;
@@ -3234,11 +3239,13 @@ int kvm_arch_set_memory_region(struct kvm *kvm,
3234 */ 3239 */
3235 if (!user_alloc) { 3240 if (!user_alloc) {
3236 if (npages && !old.rmap) { 3241 if (npages && !old.rmap) {
3242 down_write(&current->mm->mmap_sem);
3237 memslot->userspace_addr = do_mmap(NULL, 0, 3243 memslot->userspace_addr = do_mmap(NULL, 0,
3238 npages * PAGE_SIZE, 3244 npages * PAGE_SIZE,
3239 PROT_READ | PROT_WRITE, 3245 PROT_READ | PROT_WRITE,
3240 MAP_SHARED | MAP_ANONYMOUS, 3246 MAP_SHARED | MAP_ANONYMOUS,
3241 0); 3247 0);
3248 up_write(&current->mm->mmap_sem);
3242 3249
3243 if (IS_ERR((void *)memslot->userspace_addr)) 3250 if (IS_ERR((void *)memslot->userspace_addr))
3244 return PTR_ERR((void *)memslot->userspace_addr); 3251 return PTR_ERR((void *)memslot->userspace_addr);
@@ -3246,8 +3253,10 @@ int kvm_arch_set_memory_region(struct kvm *kvm,
3246 if (!old.user_alloc && old.rmap) { 3253 if (!old.user_alloc && old.rmap) {
3247 int ret; 3254 int ret;
3248 3255
3256 down_write(&current->mm->mmap_sem);
3249 ret = do_munmap(current->mm, old.userspace_addr, 3257 ret = do_munmap(current->mm, old.userspace_addr,
3250 old.npages * PAGE_SIZE); 3258 old.npages * PAGE_SIZE);
3259 up_write(&current->mm->mmap_sem);
3251 if (ret < 0) 3260 if (ret < 0)
3252 printk(KERN_WARNING 3261 printk(KERN_WARNING
3253 "kvm_vm_ioctl_set_memory_region: " 3262 "kvm_vm_ioctl_set_memory_region: "