diff options
Diffstat (limited to 'drivers/kvm/kvm_main.c')
-rw-r--r-- | drivers/kvm/kvm_main.c | 263 |
1 files changed, 149 insertions, 114 deletions
diff --git a/drivers/kvm/kvm_main.c b/drivers/kvm/kvm_main.c index 6fb36c80e3e8..a593d092d85b 100644 --- a/drivers/kvm/kvm_main.c +++ b/drivers/kvm/kvm_main.c | |||
@@ -96,6 +96,9 @@ struct segment_descriptor_64 { | |||
96 | 96 | ||
97 | #endif | 97 | #endif |
98 | 98 | ||
99 | static long kvm_vcpu_ioctl(struct file *file, unsigned int ioctl, | ||
100 | unsigned long arg); | ||
101 | |||
99 | static struct inode *kvmfs_inode(struct file_operations *fops) | 102 | static struct inode *kvmfs_inode(struct file_operations *fops) |
100 | { | 103 | { |
101 | int error = -ENOMEM; | 104 | int error = -ENOMEM; |
@@ -246,24 +249,30 @@ int kvm_write_guest(struct kvm_vcpu *vcpu, gva_t addr, unsigned long size, | |||
246 | } | 249 | } |
247 | EXPORT_SYMBOL_GPL(kvm_write_guest); | 250 | EXPORT_SYMBOL_GPL(kvm_write_guest); |
248 | 251 | ||
249 | static int vcpu_slot(struct kvm_vcpu *vcpu) | 252 | /* |
253 | * Switches to specified vcpu, until a matching vcpu_put() | ||
254 | */ | ||
255 | static void vcpu_load(struct kvm_vcpu *vcpu) | ||
250 | { | 256 | { |
251 | return vcpu - vcpu->kvm->vcpus; | 257 | mutex_lock(&vcpu->mutex); |
258 | kvm_arch_ops->vcpu_load(vcpu); | ||
252 | } | 259 | } |
253 | 260 | ||
254 | /* | 261 | /* |
255 | * Switches to specified vcpu, until a matching vcpu_put() | 262 | * Switches to specified vcpu, until a matching vcpu_put(). Will return NULL |
263 | * if the slot is not populated. | ||
256 | */ | 264 | */ |
257 | static struct kvm_vcpu *vcpu_load(struct kvm *kvm, int vcpu_slot) | 265 | static struct kvm_vcpu *vcpu_load_slot(struct kvm *kvm, int slot) |
258 | { | 266 | { |
259 | struct kvm_vcpu *vcpu = &kvm->vcpus[vcpu_slot]; | 267 | struct kvm_vcpu *vcpu = &kvm->vcpus[slot]; |
260 | 268 | ||
261 | mutex_lock(&vcpu->mutex); | 269 | mutex_lock(&vcpu->mutex); |
262 | if (unlikely(!vcpu->vmcs)) { | 270 | if (!vcpu->vmcs) { |
263 | mutex_unlock(&vcpu->mutex); | 271 | mutex_unlock(&vcpu->mutex); |
264 | return NULL; | 272 | return NULL; |
265 | } | 273 | } |
266 | return kvm_arch_ops->vcpu_load(vcpu); | 274 | kvm_arch_ops->vcpu_load(vcpu); |
275 | return vcpu; | ||
267 | } | 276 | } |
268 | 277 | ||
269 | static void vcpu_put(struct kvm_vcpu *vcpu) | 278 | static void vcpu_put(struct kvm_vcpu *vcpu) |
@@ -336,9 +345,10 @@ static void kvm_free_physmem(struct kvm *kvm) | |||
336 | 345 | ||
337 | static void kvm_free_vcpu(struct kvm_vcpu *vcpu) | 346 | static void kvm_free_vcpu(struct kvm_vcpu *vcpu) |
338 | { | 347 | { |
339 | if (!vcpu_load(vcpu->kvm, vcpu_slot(vcpu))) | 348 | if (!vcpu->vmcs) |
340 | return; | 349 | return; |
341 | 350 | ||
351 | vcpu_load(vcpu); | ||
342 | kvm_mmu_destroy(vcpu); | 352 | kvm_mmu_destroy(vcpu); |
343 | vcpu_put(vcpu); | 353 | vcpu_put(vcpu); |
344 | kvm_arch_ops->vcpu_free(vcpu); | 354 | kvm_arch_ops->vcpu_free(vcpu); |
@@ -725,7 +735,7 @@ raced: | |||
725 | for (i = 0; i < KVM_MAX_VCPUS; ++i) { | 735 | for (i = 0; i < KVM_MAX_VCPUS; ++i) { |
726 | struct kvm_vcpu *vcpu; | 736 | struct kvm_vcpu *vcpu; |
727 | 737 | ||
728 | vcpu = vcpu_load(kvm, i); | 738 | vcpu = vcpu_load_slot(kvm, i); |
729 | if (!vcpu) | 739 | if (!vcpu) |
730 | continue; | 740 | continue; |
731 | kvm_mmu_reset_context(vcpu); | 741 | kvm_mmu_reset_context(vcpu); |
@@ -791,8 +801,9 @@ static int kvm_vm_ioctl_get_dirty_log(struct kvm *kvm, | |||
791 | if (any) { | 801 | if (any) { |
792 | cleared = 0; | 802 | cleared = 0; |
793 | for (i = 0; i < KVM_MAX_VCPUS; ++i) { | 803 | for (i = 0; i < KVM_MAX_VCPUS; ++i) { |
794 | struct kvm_vcpu *vcpu = vcpu_load(kvm, i); | 804 | struct kvm_vcpu *vcpu; |
795 | 805 | ||
806 | vcpu = vcpu_load_slot(kvm, i); | ||
796 | if (!vcpu) | 807 | if (!vcpu) |
797 | continue; | 808 | continue; |
798 | if (!cleared) { | 809 | if (!cleared) { |
@@ -1461,8 +1472,7 @@ void kvm_resched(struct kvm_vcpu *vcpu) | |||
1461 | { | 1472 | { |
1462 | vcpu_put(vcpu); | 1473 | vcpu_put(vcpu); |
1463 | cond_resched(); | 1474 | cond_resched(); |
1464 | /* Cannot fail - no vcpu unplug yet. */ | 1475 | vcpu_load(vcpu); |
1465 | vcpu_load(vcpu->kvm, vcpu_slot(vcpu)); | ||
1466 | } | 1476 | } |
1467 | EXPORT_SYMBOL_GPL(kvm_resched); | 1477 | EXPORT_SYMBOL_GPL(kvm_resched); |
1468 | 1478 | ||
@@ -1484,17 +1494,11 @@ void save_msrs(struct vmx_msr_entry *e, int n) | |||
1484 | } | 1494 | } |
1485 | EXPORT_SYMBOL_GPL(save_msrs); | 1495 | EXPORT_SYMBOL_GPL(save_msrs); |
1486 | 1496 | ||
1487 | static int kvm_vm_ioctl_run(struct kvm *kvm, struct kvm_run *kvm_run) | 1497 | static int kvm_vcpu_ioctl_run(struct kvm_vcpu *vcpu, struct kvm_run *kvm_run) |
1488 | { | 1498 | { |
1489 | struct kvm_vcpu *vcpu; | ||
1490 | int r; | 1499 | int r; |
1491 | 1500 | ||
1492 | if (!valid_vcpu(kvm_run->vcpu)) | 1501 | vcpu_load(vcpu); |
1493 | return -EINVAL; | ||
1494 | |||
1495 | vcpu = vcpu_load(kvm, kvm_run->vcpu); | ||
1496 | if (!vcpu) | ||
1497 | return -ENOENT; | ||
1498 | 1502 | ||
1499 | /* re-sync apic's tpr */ | 1503 | /* re-sync apic's tpr */ |
1500 | vcpu->cr8 = kvm_run->cr8; | 1504 | vcpu->cr8 = kvm_run->cr8; |
@@ -1517,16 +1521,10 @@ static int kvm_vm_ioctl_run(struct kvm *kvm, struct kvm_run *kvm_run) | |||
1517 | return r; | 1521 | return r; |
1518 | } | 1522 | } |
1519 | 1523 | ||
1520 | static int kvm_vm_ioctl_get_regs(struct kvm *kvm, struct kvm_regs *regs) | 1524 | static int kvm_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, |
1525 | struct kvm_regs *regs) | ||
1521 | { | 1526 | { |
1522 | struct kvm_vcpu *vcpu; | 1527 | vcpu_load(vcpu); |
1523 | |||
1524 | if (!valid_vcpu(regs->vcpu)) | ||
1525 | return -EINVAL; | ||
1526 | |||
1527 | vcpu = vcpu_load(kvm, regs->vcpu); | ||
1528 | if (!vcpu) | ||
1529 | return -ENOENT; | ||
1530 | 1528 | ||
1531 | kvm_arch_ops->cache_regs(vcpu); | 1529 | kvm_arch_ops->cache_regs(vcpu); |
1532 | 1530 | ||
@@ -1563,16 +1561,10 @@ static int kvm_vm_ioctl_get_regs(struct kvm *kvm, struct kvm_regs *regs) | |||
1563 | return 0; | 1561 | return 0; |
1564 | } | 1562 | } |
1565 | 1563 | ||
1566 | static int kvm_vm_ioctl_set_regs(struct kvm *kvm, struct kvm_regs *regs) | 1564 | static int kvm_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, |
1565 | struct kvm_regs *regs) | ||
1567 | { | 1566 | { |
1568 | struct kvm_vcpu *vcpu; | 1567 | vcpu_load(vcpu); |
1569 | |||
1570 | if (!valid_vcpu(regs->vcpu)) | ||
1571 | return -EINVAL; | ||
1572 | |||
1573 | vcpu = vcpu_load(kvm, regs->vcpu); | ||
1574 | if (!vcpu) | ||
1575 | return -ENOENT; | ||
1576 | 1568 | ||
1577 | vcpu->regs[VCPU_REGS_RAX] = regs->rax; | 1569 | vcpu->regs[VCPU_REGS_RAX] = regs->rax; |
1578 | vcpu->regs[VCPU_REGS_RBX] = regs->rbx; | 1570 | vcpu->regs[VCPU_REGS_RBX] = regs->rbx; |
@@ -1609,16 +1601,12 @@ static void get_segment(struct kvm_vcpu *vcpu, | |||
1609 | return kvm_arch_ops->get_segment(vcpu, var, seg); | 1601 | return kvm_arch_ops->get_segment(vcpu, var, seg); |
1610 | } | 1602 | } |
1611 | 1603 | ||
1612 | static int kvm_vm_ioctl_get_sregs(struct kvm *kvm, struct kvm_sregs *sregs) | 1604 | static int kvm_vcpu_ioctl_get_sregs(struct kvm_vcpu *vcpu, |
1605 | struct kvm_sregs *sregs) | ||
1613 | { | 1606 | { |
1614 | struct kvm_vcpu *vcpu; | ||
1615 | struct descriptor_table dt; | 1607 | struct descriptor_table dt; |
1616 | 1608 | ||
1617 | if (!valid_vcpu(sregs->vcpu)) | 1609 | vcpu_load(vcpu); |
1618 | return -EINVAL; | ||
1619 | vcpu = vcpu_load(kvm, sregs->vcpu); | ||
1620 | if (!vcpu) | ||
1621 | return -ENOENT; | ||
1622 | 1610 | ||
1623 | get_segment(vcpu, &sregs->cs, VCPU_SREG_CS); | 1611 | get_segment(vcpu, &sregs->cs, VCPU_SREG_CS); |
1624 | get_segment(vcpu, &sregs->ds, VCPU_SREG_DS); | 1612 | get_segment(vcpu, &sregs->ds, VCPU_SREG_DS); |
@@ -1660,18 +1648,14 @@ static void set_segment(struct kvm_vcpu *vcpu, | |||
1660 | return kvm_arch_ops->set_segment(vcpu, var, seg); | 1648 | return kvm_arch_ops->set_segment(vcpu, var, seg); |
1661 | } | 1649 | } |
1662 | 1650 | ||
1663 | static int kvm_vm_ioctl_set_sregs(struct kvm *kvm, struct kvm_sregs *sregs) | 1651 | static int kvm_vcpu_ioctl_set_sregs(struct kvm_vcpu *vcpu, |
1652 | struct kvm_sregs *sregs) | ||
1664 | { | 1653 | { |
1665 | struct kvm_vcpu *vcpu; | ||
1666 | int mmu_reset_needed = 0; | 1654 | int mmu_reset_needed = 0; |
1667 | int i; | 1655 | int i; |
1668 | struct descriptor_table dt; | 1656 | struct descriptor_table dt; |
1669 | 1657 | ||
1670 | if (!valid_vcpu(sregs->vcpu)) | 1658 | vcpu_load(vcpu); |
1671 | return -EINVAL; | ||
1672 | vcpu = vcpu_load(kvm, sregs->vcpu); | ||
1673 | if (!vcpu) | ||
1674 | return -ENOENT; | ||
1675 | 1659 | ||
1676 | set_segment(vcpu, &sregs->cs, VCPU_SREG_CS); | 1660 | set_segment(vcpu, &sregs->cs, VCPU_SREG_CS); |
1677 | set_segment(vcpu, &sregs->ds, VCPU_SREG_DS); | 1661 | set_segment(vcpu, &sregs->ds, VCPU_SREG_DS); |
@@ -1777,20 +1761,14 @@ static int do_set_msr(struct kvm_vcpu *vcpu, unsigned index, u64 *data) | |||
1777 | * | 1761 | * |
1778 | * @return number of msrs set successfully. | 1762 | * @return number of msrs set successfully. |
1779 | */ | 1763 | */ |
1780 | static int __msr_io(struct kvm *kvm, struct kvm_msrs *msrs, | 1764 | static int __msr_io(struct kvm_vcpu *vcpu, struct kvm_msrs *msrs, |
1781 | struct kvm_msr_entry *entries, | 1765 | struct kvm_msr_entry *entries, |
1782 | int (*do_msr)(struct kvm_vcpu *vcpu, | 1766 | int (*do_msr)(struct kvm_vcpu *vcpu, |
1783 | unsigned index, u64 *data)) | 1767 | unsigned index, u64 *data)) |
1784 | { | 1768 | { |
1785 | struct kvm_vcpu *vcpu; | ||
1786 | int i; | 1769 | int i; |
1787 | 1770 | ||
1788 | if (!valid_vcpu(msrs->vcpu)) | 1771 | vcpu_load(vcpu); |
1789 | return -EINVAL; | ||
1790 | |||
1791 | vcpu = vcpu_load(kvm, msrs->vcpu); | ||
1792 | if (!vcpu) | ||
1793 | return -ENOENT; | ||
1794 | 1772 | ||
1795 | for (i = 0; i < msrs->nmsrs; ++i) | 1773 | for (i = 0; i < msrs->nmsrs; ++i) |
1796 | if (do_msr(vcpu, entries[i].index, &entries[i].data)) | 1774 | if (do_msr(vcpu, entries[i].index, &entries[i].data)) |
@@ -1806,7 +1784,7 @@ static int __msr_io(struct kvm *kvm, struct kvm_msrs *msrs, | |||
1806 | * | 1784 | * |
1807 | * @return number of msrs set successfully. | 1785 | * @return number of msrs set successfully. |
1808 | */ | 1786 | */ |
1809 | static int msr_io(struct kvm *kvm, struct kvm_msrs __user *user_msrs, | 1787 | static int msr_io(struct kvm_vcpu *vcpu, struct kvm_msrs __user *user_msrs, |
1810 | int (*do_msr)(struct kvm_vcpu *vcpu, | 1788 | int (*do_msr)(struct kvm_vcpu *vcpu, |
1811 | unsigned index, u64 *data), | 1789 | unsigned index, u64 *data), |
1812 | int writeback) | 1790 | int writeback) |
@@ -1834,7 +1812,7 @@ static int msr_io(struct kvm *kvm, struct kvm_msrs __user *user_msrs, | |||
1834 | if (copy_from_user(entries, user_msrs->entries, size)) | 1812 | if (copy_from_user(entries, user_msrs->entries, size)) |
1835 | goto out_free; | 1813 | goto out_free; |
1836 | 1814 | ||
1837 | r = n = __msr_io(kvm, &msrs, entries, do_msr); | 1815 | r = n = __msr_io(vcpu, &msrs, entries, do_msr); |
1838 | if (r < 0) | 1816 | if (r < 0) |
1839 | goto out_free; | 1817 | goto out_free; |
1840 | 1818 | ||
@@ -1853,38 +1831,31 @@ out: | |||
1853 | /* | 1831 | /* |
1854 | * Translate a guest virtual address to a guest physical address. | 1832 | * Translate a guest virtual address to a guest physical address. |
1855 | */ | 1833 | */ |
1856 | static int kvm_vm_ioctl_translate(struct kvm *kvm, struct kvm_translation *tr) | 1834 | static int kvm_vcpu_ioctl_translate(struct kvm_vcpu *vcpu, |
1835 | struct kvm_translation *tr) | ||
1857 | { | 1836 | { |
1858 | unsigned long vaddr = tr->linear_address; | 1837 | unsigned long vaddr = tr->linear_address; |
1859 | struct kvm_vcpu *vcpu; | ||
1860 | gpa_t gpa; | 1838 | gpa_t gpa; |
1861 | 1839 | ||
1862 | vcpu = vcpu_load(kvm, tr->vcpu); | 1840 | vcpu_load(vcpu); |
1863 | if (!vcpu) | 1841 | spin_lock(&vcpu->kvm->lock); |
1864 | return -ENOENT; | ||
1865 | spin_lock(&kvm->lock); | ||
1866 | gpa = vcpu->mmu.gva_to_gpa(vcpu, vaddr); | 1842 | gpa = vcpu->mmu.gva_to_gpa(vcpu, vaddr); |
1867 | tr->physical_address = gpa; | 1843 | tr->physical_address = gpa; |
1868 | tr->valid = gpa != UNMAPPED_GVA; | 1844 | tr->valid = gpa != UNMAPPED_GVA; |
1869 | tr->writeable = 1; | 1845 | tr->writeable = 1; |
1870 | tr->usermode = 0; | 1846 | tr->usermode = 0; |
1871 | spin_unlock(&kvm->lock); | 1847 | spin_unlock(&vcpu->kvm->lock); |
1872 | vcpu_put(vcpu); | 1848 | vcpu_put(vcpu); |
1873 | 1849 | ||
1874 | return 0; | 1850 | return 0; |
1875 | } | 1851 | } |
1876 | 1852 | ||
1877 | static int kvm_vm_ioctl_interrupt(struct kvm *kvm, struct kvm_interrupt *irq) | 1853 | static int kvm_vcpu_ioctl_interrupt(struct kvm_vcpu *vcpu, |
1854 | struct kvm_interrupt *irq) | ||
1878 | { | 1855 | { |
1879 | struct kvm_vcpu *vcpu; | ||
1880 | |||
1881 | if (!valid_vcpu(irq->vcpu)) | ||
1882 | return -EINVAL; | ||
1883 | if (irq->irq < 0 || irq->irq >= 256) | 1856 | if (irq->irq < 0 || irq->irq >= 256) |
1884 | return -EINVAL; | 1857 | return -EINVAL; |
1885 | vcpu = vcpu_load(kvm, irq->vcpu); | 1858 | vcpu_load(vcpu); |
1886 | if (!vcpu) | ||
1887 | return -ENOENT; | ||
1888 | 1859 | ||
1889 | set_bit(irq->irq, vcpu->irq_pending); | 1860 | set_bit(irq->irq, vcpu->irq_pending); |
1890 | set_bit(irq->irq / BITS_PER_LONG, &vcpu->irq_summary); | 1861 | set_bit(irq->irq / BITS_PER_LONG, &vcpu->irq_summary); |
@@ -1894,17 +1865,12 @@ static int kvm_vm_ioctl_interrupt(struct kvm *kvm, struct kvm_interrupt *irq) | |||
1894 | return 0; | 1865 | return 0; |
1895 | } | 1866 | } |
1896 | 1867 | ||
1897 | static int kvm_vm_ioctl_debug_guest(struct kvm *kvm, | 1868 | static int kvm_vcpu_ioctl_debug_guest(struct kvm_vcpu *vcpu, |
1898 | struct kvm_debug_guest *dbg) | 1869 | struct kvm_debug_guest *dbg) |
1899 | { | 1870 | { |
1900 | struct kvm_vcpu *vcpu; | ||
1901 | int r; | 1871 | int r; |
1902 | 1872 | ||
1903 | if (!valid_vcpu(dbg->vcpu)) | 1873 | vcpu_load(vcpu); |
1904 | return -EINVAL; | ||
1905 | vcpu = vcpu_load(kvm, dbg->vcpu); | ||
1906 | if (!vcpu) | ||
1907 | return -ENOENT; | ||
1908 | 1874 | ||
1909 | r = kvm_arch_ops->set_guest_debug(vcpu, dbg); | 1875 | r = kvm_arch_ops->set_guest_debug(vcpu, dbg); |
1910 | 1876 | ||
@@ -1913,6 +1879,59 @@ static int kvm_vm_ioctl_debug_guest(struct kvm *kvm, | |||
1913 | return r; | 1879 | return r; |
1914 | } | 1880 | } |
1915 | 1881 | ||
1882 | static int kvm_vcpu_release(struct inode *inode, struct file *filp) | ||
1883 | { | ||
1884 | struct kvm_vcpu *vcpu = filp->private_data; | ||
1885 | |||
1886 | fput(vcpu->kvm->filp); | ||
1887 | return 0; | ||
1888 | } | ||
1889 | |||
1890 | static struct file_operations kvm_vcpu_fops = { | ||
1891 | .release = kvm_vcpu_release, | ||
1892 | .unlocked_ioctl = kvm_vcpu_ioctl, | ||
1893 | .compat_ioctl = kvm_vcpu_ioctl, | ||
1894 | }; | ||
1895 | |||
1896 | /* | ||
1897 | * Allocates an inode for the vcpu. | ||
1898 | */ | ||
1899 | static int create_vcpu_fd(struct kvm_vcpu *vcpu) | ||
1900 | { | ||
1901 | int fd, r; | ||
1902 | struct inode *inode; | ||
1903 | struct file *file; | ||
1904 | |||
1905 | atomic_inc(&vcpu->kvm->filp->f_count); | ||
1906 | inode = kvmfs_inode(&kvm_vcpu_fops); | ||
1907 | if (IS_ERR(inode)) { | ||
1908 | r = PTR_ERR(inode); | ||
1909 | goto out1; | ||
1910 | } | ||
1911 | |||
1912 | file = kvmfs_file(inode, vcpu); | ||
1913 | if (IS_ERR(file)) { | ||
1914 | r = PTR_ERR(file); | ||
1915 | goto out2; | ||
1916 | } | ||
1917 | |||
1918 | r = get_unused_fd(); | ||
1919 | if (r < 0) | ||
1920 | goto out3; | ||
1921 | fd = r; | ||
1922 | fd_install(fd, file); | ||
1923 | |||
1924 | return fd; | ||
1925 | |||
1926 | out3: | ||
1927 | fput(file); | ||
1928 | out2: | ||
1929 | iput(inode); | ||
1930 | out1: | ||
1931 | fput(vcpu->kvm->filp); | ||
1932 | return r; | ||
1933 | } | ||
1934 | |||
1916 | /* | 1935 | /* |
1917 | * Creates some virtual cpus. Good luck creating more than one. | 1936 | * Creates some virtual cpus. Good luck creating more than one. |
1918 | */ | 1937 | */ |
@@ -1955,7 +1974,11 @@ static int kvm_vm_ioctl_create_vcpu(struct kvm *kvm, int n) | |||
1955 | if (r < 0) | 1974 | if (r < 0) |
1956 | goto out_free_vcpus; | 1975 | goto out_free_vcpus; |
1957 | 1976 | ||
1958 | return 0; | 1977 | r = create_vcpu_fd(vcpu); |
1978 | if (r < 0) | ||
1979 | goto out_free_vcpus; | ||
1980 | |||
1981 | return r; | ||
1959 | 1982 | ||
1960 | out_free_vcpus: | 1983 | out_free_vcpus: |
1961 | kvm_free_vcpu(vcpu); | 1984 | kvm_free_vcpu(vcpu); |
@@ -1964,26 +1987,21 @@ out: | |||
1964 | return r; | 1987 | return r; |
1965 | } | 1988 | } |
1966 | 1989 | ||
1967 | static long kvm_vm_ioctl(struct file *filp, | 1990 | static long kvm_vcpu_ioctl(struct file *filp, |
1968 | unsigned int ioctl, unsigned long arg) | 1991 | unsigned int ioctl, unsigned long arg) |
1969 | { | 1992 | { |
1970 | struct kvm *kvm = filp->private_data; | 1993 | struct kvm_vcpu *vcpu = filp->private_data; |
1971 | void __user *argp = (void __user *)arg; | 1994 | void __user *argp = (void __user *)arg; |
1972 | int r = -EINVAL; | 1995 | int r = -EINVAL; |
1973 | 1996 | ||
1974 | switch (ioctl) { | 1997 | switch (ioctl) { |
1975 | case KVM_CREATE_VCPU: | ||
1976 | r = kvm_vm_ioctl_create_vcpu(kvm, arg); | ||
1977 | if (r) | ||
1978 | goto out; | ||
1979 | break; | ||
1980 | case KVM_RUN: { | 1998 | case KVM_RUN: { |
1981 | struct kvm_run kvm_run; | 1999 | struct kvm_run kvm_run; |
1982 | 2000 | ||
1983 | r = -EFAULT; | 2001 | r = -EFAULT; |
1984 | if (copy_from_user(&kvm_run, argp, sizeof kvm_run)) | 2002 | if (copy_from_user(&kvm_run, argp, sizeof kvm_run)) |
1985 | goto out; | 2003 | goto out; |
1986 | r = kvm_vm_ioctl_run(kvm, &kvm_run); | 2004 | r = kvm_vcpu_ioctl_run(vcpu, &kvm_run); |
1987 | if (r < 0 && r != -EINTR) | 2005 | if (r < 0 && r != -EINTR) |
1988 | goto out; | 2006 | goto out; |
1989 | if (copy_to_user(argp, &kvm_run, sizeof kvm_run)) { | 2007 | if (copy_to_user(argp, &kvm_run, sizeof kvm_run)) { |
@@ -1995,10 +2013,8 @@ static long kvm_vm_ioctl(struct file *filp, | |||
1995 | case KVM_GET_REGS: { | 2013 | case KVM_GET_REGS: { |
1996 | struct kvm_regs kvm_regs; | 2014 | struct kvm_regs kvm_regs; |
1997 | 2015 | ||
1998 | r = -EFAULT; | 2016 | memset(&kvm_regs, 0, sizeof kvm_regs); |
1999 | if (copy_from_user(&kvm_regs, argp, sizeof kvm_regs)) | 2017 | r = kvm_vcpu_ioctl_get_regs(vcpu, &kvm_regs); |
2000 | goto out; | ||
2001 | r = kvm_vm_ioctl_get_regs(kvm, &kvm_regs); | ||
2002 | if (r) | 2018 | if (r) |
2003 | goto out; | 2019 | goto out; |
2004 | r = -EFAULT; | 2020 | r = -EFAULT; |
@@ -2013,7 +2029,7 @@ static long kvm_vm_ioctl(struct file *filp, | |||
2013 | r = -EFAULT; | 2029 | r = -EFAULT; |
2014 | if (copy_from_user(&kvm_regs, argp, sizeof kvm_regs)) | 2030 | if (copy_from_user(&kvm_regs, argp, sizeof kvm_regs)) |
2015 | goto out; | 2031 | goto out; |
2016 | r = kvm_vm_ioctl_set_regs(kvm, &kvm_regs); | 2032 | r = kvm_vcpu_ioctl_set_regs(vcpu, &kvm_regs); |
2017 | if (r) | 2033 | if (r) |
2018 | goto out; | 2034 | goto out; |
2019 | r = 0; | 2035 | r = 0; |
@@ -2022,10 +2038,8 @@ static long kvm_vm_ioctl(struct file *filp, | |||
2022 | case KVM_GET_SREGS: { | 2038 | case KVM_GET_SREGS: { |
2023 | struct kvm_sregs kvm_sregs; | 2039 | struct kvm_sregs kvm_sregs; |
2024 | 2040 | ||
2025 | r = -EFAULT; | 2041 | memset(&kvm_sregs, 0, sizeof kvm_sregs); |
2026 | if (copy_from_user(&kvm_sregs, argp, sizeof kvm_sregs)) | 2042 | r = kvm_vcpu_ioctl_get_sregs(vcpu, &kvm_sregs); |
2027 | goto out; | ||
2028 | r = kvm_vm_ioctl_get_sregs(kvm, &kvm_sregs); | ||
2029 | if (r) | 2043 | if (r) |
2030 | goto out; | 2044 | goto out; |
2031 | r = -EFAULT; | 2045 | r = -EFAULT; |
@@ -2040,7 +2054,7 @@ static long kvm_vm_ioctl(struct file *filp, | |||
2040 | r = -EFAULT; | 2054 | r = -EFAULT; |
2041 | if (copy_from_user(&kvm_sregs, argp, sizeof kvm_sregs)) | 2055 | if (copy_from_user(&kvm_sregs, argp, sizeof kvm_sregs)) |
2042 | goto out; | 2056 | goto out; |
2043 | r = kvm_vm_ioctl_set_sregs(kvm, &kvm_sregs); | 2057 | r = kvm_vcpu_ioctl_set_sregs(vcpu, &kvm_sregs); |
2044 | if (r) | 2058 | if (r) |
2045 | goto out; | 2059 | goto out; |
2046 | r = 0; | 2060 | r = 0; |
@@ -2052,7 +2066,7 @@ static long kvm_vm_ioctl(struct file *filp, | |||
2052 | r = -EFAULT; | 2066 | r = -EFAULT; |
2053 | if (copy_from_user(&tr, argp, sizeof tr)) | 2067 | if (copy_from_user(&tr, argp, sizeof tr)) |
2054 | goto out; | 2068 | goto out; |
2055 | r = kvm_vm_ioctl_translate(kvm, &tr); | 2069 | r = kvm_vcpu_ioctl_translate(vcpu, &tr); |
2056 | if (r) | 2070 | if (r) |
2057 | goto out; | 2071 | goto out; |
2058 | r = -EFAULT; | 2072 | r = -EFAULT; |
@@ -2067,7 +2081,7 @@ static long kvm_vm_ioctl(struct file *filp, | |||
2067 | r = -EFAULT; | 2081 | r = -EFAULT; |
2068 | if (copy_from_user(&irq, argp, sizeof irq)) | 2082 | if (copy_from_user(&irq, argp, sizeof irq)) |
2069 | goto out; | 2083 | goto out; |
2070 | r = kvm_vm_ioctl_interrupt(kvm, &irq); | 2084 | r = kvm_vcpu_ioctl_interrupt(vcpu, &irq); |
2071 | if (r) | 2085 | if (r) |
2072 | goto out; | 2086 | goto out; |
2073 | r = 0; | 2087 | r = 0; |
@@ -2079,12 +2093,38 @@ static long kvm_vm_ioctl(struct file *filp, | |||
2079 | r = -EFAULT; | 2093 | r = -EFAULT; |
2080 | if (copy_from_user(&dbg, argp, sizeof dbg)) | 2094 | if (copy_from_user(&dbg, argp, sizeof dbg)) |
2081 | goto out; | 2095 | goto out; |
2082 | r = kvm_vm_ioctl_debug_guest(kvm, &dbg); | 2096 | r = kvm_vcpu_ioctl_debug_guest(vcpu, &dbg); |
2083 | if (r) | 2097 | if (r) |
2084 | goto out; | 2098 | goto out; |
2085 | r = 0; | 2099 | r = 0; |
2086 | break; | 2100 | break; |
2087 | } | 2101 | } |
2102 | case KVM_GET_MSRS: | ||
2103 | r = msr_io(vcpu, argp, get_msr, 1); | ||
2104 | break; | ||
2105 | case KVM_SET_MSRS: | ||
2106 | r = msr_io(vcpu, argp, do_set_msr, 0); | ||
2107 | break; | ||
2108 | default: | ||
2109 | ; | ||
2110 | } | ||
2111 | out: | ||
2112 | return r; | ||
2113 | } | ||
2114 | |||
2115 | static long kvm_vm_ioctl(struct file *filp, | ||
2116 | unsigned int ioctl, unsigned long arg) | ||
2117 | { | ||
2118 | struct kvm *kvm = filp->private_data; | ||
2119 | void __user *argp = (void __user *)arg; | ||
2120 | int r = -EINVAL; | ||
2121 | |||
2122 | switch (ioctl) { | ||
2123 | case KVM_CREATE_VCPU: | ||
2124 | r = kvm_vm_ioctl_create_vcpu(kvm, arg); | ||
2125 | if (r < 0) | ||
2126 | goto out; | ||
2127 | break; | ||
2088 | case KVM_SET_MEMORY_REGION: { | 2128 | case KVM_SET_MEMORY_REGION: { |
2089 | struct kvm_memory_region kvm_mem; | 2129 | struct kvm_memory_region kvm_mem; |
2090 | 2130 | ||
@@ -2107,12 +2147,6 @@ static long kvm_vm_ioctl(struct file *filp, | |||
2107 | goto out; | 2147 | goto out; |
2108 | break; | 2148 | break; |
2109 | } | 2149 | } |
2110 | case KVM_GET_MSRS: | ||
2111 | r = msr_io(kvm, argp, get_msr, 1); | ||
2112 | break; | ||
2113 | case KVM_SET_MSRS: | ||
2114 | r = msr_io(kvm, argp, do_set_msr, 0); | ||
2115 | break; | ||
2116 | default: | 2150 | default: |
2117 | ; | 2151 | ; |
2118 | } | 2152 | } |
@@ -2182,6 +2216,7 @@ static int kvm_dev_ioctl_create_vm(void) | |||
2182 | r = PTR_ERR(file); | 2216 | r = PTR_ERR(file); |
2183 | goto out3; | 2217 | goto out3; |
2184 | } | 2218 | } |
2219 | kvm->filp = file; | ||
2185 | 2220 | ||
2186 | r = get_unused_fd(); | 2221 | r = get_unused_fd(); |
2187 | if (r < 0) | 2222 | if (r < 0) |