diff options
Diffstat (limited to 'arch/ia64/kvm/kvm-ia64.c')
-rw-r--r-- | arch/ia64/kvm/kvm-ia64.c | 60 |
1 files changed, 31 insertions, 29 deletions
diff --git a/arch/ia64/kvm/kvm-ia64.c b/arch/ia64/kvm/kvm-ia64.c index af1464f7a6ad..43e45f6afcda 100644 --- a/arch/ia64/kvm/kvm-ia64.c +++ b/arch/ia64/kvm/kvm-ia64.c | |||
@@ -698,27 +698,24 @@ out: | |||
698 | return r; | 698 | return r; |
699 | } | 699 | } |
700 | 700 | ||
701 | /* | ||
702 | * Allocate 16M memory for every vm to hold its specific data. | ||
703 | * Its memory map is defined in kvm_host.h. | ||
704 | */ | ||
705 | static struct kvm *kvm_alloc_kvm(void) | 701 | static struct kvm *kvm_alloc_kvm(void) |
706 | { | 702 | { |
707 | 703 | ||
708 | struct kvm *kvm; | 704 | struct kvm *kvm; |
709 | uint64_t vm_base; | 705 | uint64_t vm_base; |
710 | 706 | ||
707 | BUG_ON(sizeof(struct kvm) > KVM_VM_STRUCT_SIZE); | ||
708 | |||
711 | vm_base = __get_free_pages(GFP_KERNEL, get_order(KVM_VM_DATA_SIZE)); | 709 | vm_base = __get_free_pages(GFP_KERNEL, get_order(KVM_VM_DATA_SIZE)); |
712 | 710 | ||
713 | if (!vm_base) | 711 | if (!vm_base) |
714 | return ERR_PTR(-ENOMEM); | 712 | return ERR_PTR(-ENOMEM); |
715 | printk(KERN_DEBUG"kvm: VM data's base Address:0x%lx\n", vm_base); | ||
716 | 713 | ||
717 | /* Zero all pages before use! */ | ||
718 | memset((void *)vm_base, 0, KVM_VM_DATA_SIZE); | 714 | memset((void *)vm_base, 0, KVM_VM_DATA_SIZE); |
719 | 715 | kvm = (struct kvm *)(vm_base + | |
720 | kvm = (struct kvm *)(vm_base + KVM_VM_OFS); | 716 | offsetof(struct kvm_vm_data, kvm_vm_struct)); |
721 | kvm->arch.vm_base = vm_base; | 717 | kvm->arch.vm_base = vm_base; |
718 | printk(KERN_DEBUG"kvm: vm's data area:0x%lx\n", vm_base); | ||
722 | 719 | ||
723 | return kvm; | 720 | return kvm; |
724 | } | 721 | } |
@@ -760,21 +757,12 @@ static void kvm_build_io_pmt(struct kvm *kvm) | |||
760 | 757 | ||
761 | static void kvm_init_vm(struct kvm *kvm) | 758 | static void kvm_init_vm(struct kvm *kvm) |
762 | { | 759 | { |
763 | long vm_base; | ||
764 | |||
765 | BUG_ON(!kvm); | 760 | BUG_ON(!kvm); |
766 | 761 | ||
767 | kvm->arch.metaphysical_rr0 = GUEST_PHYSICAL_RR0; | 762 | kvm->arch.metaphysical_rr0 = GUEST_PHYSICAL_RR0; |
768 | kvm->arch.metaphysical_rr4 = GUEST_PHYSICAL_RR4; | 763 | kvm->arch.metaphysical_rr4 = GUEST_PHYSICAL_RR4; |
769 | kvm->arch.vmm_init_rr = VMM_INIT_RR; | 764 | kvm->arch.vmm_init_rr = VMM_INIT_RR; |
770 | 765 | ||
771 | vm_base = kvm->arch.vm_base; | ||
772 | if (vm_base) { | ||
773 | kvm->arch.vhpt_base = vm_base + KVM_VHPT_OFS; | ||
774 | kvm->arch.vtlb_base = vm_base + KVM_VTLB_OFS; | ||
775 | kvm->arch.vpd_base = vm_base + KVM_VPD_OFS; | ||
776 | } | ||
777 | |||
778 | /* | 766 | /* |
779 | *Fill P2M entries for MMIO/IO ranges | 767 | *Fill P2M entries for MMIO/IO ranges |
780 | */ | 768 | */ |
@@ -864,7 +852,7 @@ int kvm_arch_vcpu_ioctl_set_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) | |||
864 | goto out; | 852 | goto out; |
865 | r = copy_from_user(vcpu + 1, regs->saved_stack + | 853 | r = copy_from_user(vcpu + 1, regs->saved_stack + |
866 | sizeof(struct kvm_vcpu), | 854 | sizeof(struct kvm_vcpu), |
867 | IA64_STK_OFFSET - sizeof(struct kvm_vcpu)); | 855 | KVM_STK_OFFSET - sizeof(struct kvm_vcpu)); |
868 | if (r) | 856 | if (r) |
869 | goto out; | 857 | goto out; |
870 | vcpu->arch.exit_data = | 858 | vcpu->arch.exit_data = |
@@ -1166,10 +1154,11 @@ int kvm_arch_vcpu_init(struct kvm_vcpu *vcpu) | |||
1166 | /*Set entry address for first run.*/ | 1154 | /*Set entry address for first run.*/ |
1167 | regs->cr_iip = PALE_RESET_ENTRY; | 1155 | regs->cr_iip = PALE_RESET_ENTRY; |
1168 | 1156 | ||
1169 | /*Initilize itc offset for vcpus*/ | 1157 | /*Initialize itc offset for vcpus*/ |
1170 | itc_offset = 0UL - ia64_getreg(_IA64_REG_AR_ITC); | 1158 | itc_offset = 0UL - ia64_getreg(_IA64_REG_AR_ITC); |
1171 | for (i = 0; i < MAX_VCPU_NUM; i++) { | 1159 | for (i = 0; i < KVM_MAX_VCPUS; i++) { |
1172 | v = (struct kvm_vcpu *)((char *)vcpu + VCPU_SIZE * i); | 1160 | v = (struct kvm_vcpu *)((char *)vcpu + |
1161 | sizeof(struct kvm_vcpu_data) * i); | ||
1173 | v->arch.itc_offset = itc_offset; | 1162 | v->arch.itc_offset = itc_offset; |
1174 | v->arch.last_itc = 0; | 1163 | v->arch.last_itc = 0; |
1175 | } | 1164 | } |
@@ -1183,7 +1172,7 @@ int kvm_arch_vcpu_init(struct kvm_vcpu *vcpu) | |||
1183 | vcpu->arch.apic->vcpu = vcpu; | 1172 | vcpu->arch.apic->vcpu = vcpu; |
1184 | 1173 | ||
1185 | p_ctx->gr[1] = 0; | 1174 | p_ctx->gr[1] = 0; |
1186 | p_ctx->gr[12] = (unsigned long)((char *)vmm_vcpu + IA64_STK_OFFSET); | 1175 | p_ctx->gr[12] = (unsigned long)((char *)vmm_vcpu + KVM_STK_OFFSET); |
1187 | p_ctx->gr[13] = (unsigned long)vmm_vcpu; | 1176 | p_ctx->gr[13] = (unsigned long)vmm_vcpu; |
1188 | p_ctx->psr = 0x1008522000UL; | 1177 | p_ctx->psr = 0x1008522000UL; |
1189 | p_ctx->ar[40] = FPSR_DEFAULT; /*fpsr*/ | 1178 | p_ctx->ar[40] = FPSR_DEFAULT; /*fpsr*/ |
@@ -1218,12 +1207,12 @@ int kvm_arch_vcpu_init(struct kvm_vcpu *vcpu) | |||
1218 | vcpu->arch.hlt_timer.function = hlt_timer_fn; | 1207 | vcpu->arch.hlt_timer.function = hlt_timer_fn; |
1219 | 1208 | ||
1220 | vcpu->arch.last_run_cpu = -1; | 1209 | vcpu->arch.last_run_cpu = -1; |
1221 | vcpu->arch.vpd = (struct vpd *)VPD_ADDR(vcpu->vcpu_id); | 1210 | vcpu->arch.vpd = (struct vpd *)VPD_BASE(vcpu->vcpu_id); |
1222 | vcpu->arch.vsa_base = kvm_vsa_base; | 1211 | vcpu->arch.vsa_base = kvm_vsa_base; |
1223 | vcpu->arch.__gp = kvm_vmm_gp; | 1212 | vcpu->arch.__gp = kvm_vmm_gp; |
1224 | vcpu->arch.dirty_log_lock_pa = __pa(&kvm->arch.dirty_log_lock); | 1213 | vcpu->arch.dirty_log_lock_pa = __pa(&kvm->arch.dirty_log_lock); |
1225 | vcpu->arch.vhpt.hash = (struct thash_data *)VHPT_ADDR(vcpu->vcpu_id); | 1214 | vcpu->arch.vhpt.hash = (struct thash_data *)VHPT_BASE(vcpu->vcpu_id); |
1226 | vcpu->arch.vtlb.hash = (struct thash_data *)VTLB_ADDR(vcpu->vcpu_id); | 1215 | vcpu->arch.vtlb.hash = (struct thash_data *)VTLB_BASE(vcpu->vcpu_id); |
1227 | init_ptce_info(vcpu); | 1216 | init_ptce_info(vcpu); |
1228 | 1217 | ||
1229 | r = 0; | 1218 | r = 0; |
@@ -1273,12 +1262,22 @@ struct kvm_vcpu *kvm_arch_vcpu_create(struct kvm *kvm, | |||
1273 | int r; | 1262 | int r; |
1274 | int cpu; | 1263 | int cpu; |
1275 | 1264 | ||
1265 | BUG_ON(sizeof(struct kvm_vcpu) > VCPU_STRUCT_SIZE/2); | ||
1266 | |||
1267 | r = -EINVAL; | ||
1268 | if (id >= KVM_MAX_VCPUS) { | ||
1269 | printk(KERN_ERR"kvm: Can't configure vcpus > %ld", | ||
1270 | KVM_MAX_VCPUS); | ||
1271 | goto fail; | ||
1272 | } | ||
1273 | |||
1276 | r = -ENOMEM; | 1274 | r = -ENOMEM; |
1277 | if (!vm_base) { | 1275 | if (!vm_base) { |
1278 | printk(KERN_ERR"kvm: Create vcpu[%d] error!\n", id); | 1276 | printk(KERN_ERR"kvm: Create vcpu[%d] error!\n", id); |
1279 | goto fail; | 1277 | goto fail; |
1280 | } | 1278 | } |
1281 | vcpu = (struct kvm_vcpu *)(vm_base + KVM_VCPU_OFS + VCPU_SIZE * id); | 1279 | vcpu = (struct kvm_vcpu *)(vm_base + offsetof(struct kvm_vm_data, |
1280 | vcpu_data[id].vcpu_struct)); | ||
1282 | vcpu->kvm = kvm; | 1281 | vcpu->kvm = kvm; |
1283 | 1282 | ||
1284 | cpu = get_cpu(); | 1283 | cpu = get_cpu(); |
@@ -1396,7 +1395,7 @@ int kvm_arch_vcpu_ioctl_get_regs(struct kvm_vcpu *vcpu, struct kvm_regs *regs) | |||
1396 | sizeof(union context)); | 1395 | sizeof(union context)); |
1397 | if (r) | 1396 | if (r) |
1398 | goto out; | 1397 | goto out; |
1399 | r = copy_to_user(regs->saved_stack, (void *)vcpu, IA64_STK_OFFSET); | 1398 | r = copy_to_user(regs->saved_stack, (void *)vcpu, KVM_STK_OFFSET); |
1400 | if (r) | 1399 | if (r) |
1401 | goto out; | 1400 | goto out; |
1402 | SAVE_REGS(mp_state); | 1401 | SAVE_REGS(mp_state); |
@@ -1457,6 +1456,9 @@ int kvm_arch_set_memory_region(struct kvm *kvm, | |||
1457 | struct kvm_memory_slot *memslot = &kvm->memslots[mem->slot]; | 1456 | struct kvm_memory_slot *memslot = &kvm->memslots[mem->slot]; |
1458 | unsigned long base_gfn = memslot->base_gfn; | 1457 | unsigned long base_gfn = memslot->base_gfn; |
1459 | 1458 | ||
1459 | if (base_gfn + npages > (KVM_MAX_MEM_SIZE >> PAGE_SHIFT)) | ||
1460 | return -ENOMEM; | ||
1461 | |||
1460 | for (i = 0; i < npages; i++) { | 1462 | for (i = 0; i < npages; i++) { |
1461 | pfn = gfn_to_pfn(kvm, base_gfn + i); | 1463 | pfn = gfn_to_pfn(kvm, base_gfn + i); |
1462 | if (!kvm_is_mmio_pfn(pfn)) { | 1464 | if (!kvm_is_mmio_pfn(pfn)) { |
@@ -1631,8 +1633,8 @@ static int kvm_ia64_sync_dirty_log(struct kvm *kvm, | |||
1631 | struct kvm_memory_slot *memslot; | 1633 | struct kvm_memory_slot *memslot; |
1632 | int r, i; | 1634 | int r, i; |
1633 | long n, base; | 1635 | long n, base; |
1634 | unsigned long *dirty_bitmap = (unsigned long *)((void *)kvm - KVM_VM_OFS | 1636 | unsigned long *dirty_bitmap = (unsigned long *)(kvm->arch.vm_base + |
1635 | + KVM_MEM_DIRTY_LOG_OFS); | 1637 | offsetof(struct kvm_vm_data, kvm_mem_dirty_log)); |
1636 | 1638 | ||
1637 | r = -EINVAL; | 1639 | r = -EINVAL; |
1638 | if (log->slot >= KVM_MEMORY_SLOTS) | 1640 | if (log->slot >= KVM_MEMORY_SLOTS) |