aboutsummaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--arch/x86/kvm/x86.c37
-rw-r--r--virt/kvm/eventfd.c2
-rw-r--r--virt/kvm/kvm_main.c23
3 files changed, 50 insertions, 12 deletions
diff --git a/arch/x86/kvm/x86.c b/arch/x86/kvm/x86.c
index 58f517b59645..934b4c6b0bf9 100644
--- a/arch/x86/kvm/x86.c
+++ b/arch/x86/kvm/x86.c
@@ -2395,9 +2395,9 @@ static void do_cpuid_ent(struct kvm_cpuid_entry2 *entry, u32 function,
2395 int i; 2395 int i;
2396 2396
2397 entry->flags |= KVM_CPUID_FLAG_SIGNIFCANT_INDEX; 2397 entry->flags |= KVM_CPUID_FLAG_SIGNIFCANT_INDEX;
2398 for (i = 1; *nent < maxnent; ++i) { 2398 for (i = 1; *nent < maxnent && i < 64; ++i) {
2399 if (entry[i - 1].eax == 0 && i != 2) 2399 if (entry[i].eax == 0)
2400 break; 2400 continue;
2401 do_cpuid_1_ent(&entry[i], function, i); 2401 do_cpuid_1_ent(&entry[i], function, i);
2402 entry[i].flags |= 2402 entry[i].flags |=
2403 KVM_CPUID_FLAG_SIGNIFCANT_INDEX; 2403 KVM_CPUID_FLAG_SIGNIFCANT_INDEX;
@@ -4958,12 +4958,6 @@ struct kvm_cpuid_entry2 *kvm_find_cpuid_entry(struct kvm_vcpu *vcpu,
4958 best = e; 4958 best = e;
4959 break; 4959 break;
4960 } 4960 }
4961 /*
4962 * Both basic or both extended?
4963 */
4964 if (((e->function ^ function) & 0x80000000) == 0)
4965 if (!best || e->function > best->function)
4966 best = e;
4967 } 4961 }
4968 return best; 4962 return best;
4969} 4963}
@@ -4983,6 +4977,27 @@ not_found:
4983 return 36; 4977 return 36;
4984} 4978}
4985 4979
4980/*
4981 * If no match is found, check whether we exceed the vCPU's limit
4982 * and return the content of the highest valid _standard_ leaf instead.
4983 * This is to satisfy the CPUID specification.
4984 */
4985static struct kvm_cpuid_entry2* check_cpuid_limit(struct kvm_vcpu *vcpu,
4986 u32 function, u32 index)
4987{
4988 struct kvm_cpuid_entry2 *maxlevel;
4989
4990 maxlevel = kvm_find_cpuid_entry(vcpu, function & 0x80000000, 0);
4991 if (!maxlevel || maxlevel->eax >= function)
4992 return NULL;
4993 if (function & 0x80000000) {
4994 maxlevel = kvm_find_cpuid_entry(vcpu, 0, 0);
4995 if (!maxlevel)
4996 return NULL;
4997 }
4998 return kvm_find_cpuid_entry(vcpu, maxlevel->eax, index);
4999}
5000
4986void kvm_emulate_cpuid(struct kvm_vcpu *vcpu) 5001void kvm_emulate_cpuid(struct kvm_vcpu *vcpu)
4987{ 5002{
4988 u32 function, index; 5003 u32 function, index;
@@ -4995,6 +5010,10 @@ void kvm_emulate_cpuid(struct kvm_vcpu *vcpu)
4995 kvm_register_write(vcpu, VCPU_REGS_RCX, 0); 5010 kvm_register_write(vcpu, VCPU_REGS_RCX, 0);
4996 kvm_register_write(vcpu, VCPU_REGS_RDX, 0); 5011 kvm_register_write(vcpu, VCPU_REGS_RDX, 0);
4997 best = kvm_find_cpuid_entry(vcpu, function, index); 5012 best = kvm_find_cpuid_entry(vcpu, function, index);
5013
5014 if (!best)
5015 best = check_cpuid_limit(vcpu, function, index);
5016
4998 if (best) { 5017 if (best) {
4999 kvm_register_write(vcpu, VCPU_REGS_RAX, best->eax); 5018 kvm_register_write(vcpu, VCPU_REGS_RAX, best->eax);
5000 kvm_register_write(vcpu, VCPU_REGS_RBX, best->ebx); 5019 kvm_register_write(vcpu, VCPU_REGS_RBX, best->ebx);
diff --git a/virt/kvm/eventfd.c b/virt/kvm/eventfd.c
index 36d8092dbb3f..73358d256fa2 100644
--- a/virt/kvm/eventfd.c
+++ b/virt/kvm/eventfd.c
@@ -90,7 +90,7 @@ irqfd_shutdown(struct work_struct *work)
90 * We know no new events will be scheduled at this point, so block 90 * We know no new events will be scheduled at this point, so block
91 * until all previously outstanding events have completed 91 * until all previously outstanding events have completed
92 */ 92 */
93 flush_work(&irqfd->inject); 93 flush_work_sync(&irqfd->inject);
94 94
95 /* 95 /*
96 * It is now safe to release the object's resources 96 * It is now safe to release the object's resources
diff --git a/virt/kvm/kvm_main.c b/virt/kvm/kvm_main.c
index 556e3efe5325..6330653480e4 100644
--- a/virt/kvm/kvm_main.c
+++ b/virt/kvm/kvm_main.c
@@ -1037,6 +1037,17 @@ static pfn_t get_fault_pfn(void)
1037 return fault_pfn; 1037 return fault_pfn;
1038} 1038}
1039 1039
1040int get_user_page_nowait(struct task_struct *tsk, struct mm_struct *mm,
1041 unsigned long start, int write, struct page **page)
1042{
1043 int flags = FOLL_TOUCH | FOLL_NOWAIT | FOLL_HWPOISON | FOLL_GET;
1044
1045 if (write)
1046 flags |= FOLL_WRITE;
1047
1048 return __get_user_pages(tsk, mm, start, 1, flags, page, NULL, NULL);
1049}
1050
1040static inline int check_user_page_hwpoison(unsigned long addr) 1051static inline int check_user_page_hwpoison(unsigned long addr)
1041{ 1052{
1042 int rc, flags = FOLL_TOUCH | FOLL_HWPOISON | FOLL_WRITE; 1053 int rc, flags = FOLL_TOUCH | FOLL_HWPOISON | FOLL_WRITE;
@@ -1070,7 +1081,14 @@ static pfn_t hva_to_pfn(struct kvm *kvm, unsigned long addr, bool atomic,
1070 if (writable) 1081 if (writable)
1071 *writable = write_fault; 1082 *writable = write_fault;
1072 1083
1073 npages = get_user_pages_fast(addr, 1, write_fault, page); 1084 if (async) {
1085 down_read(&current->mm->mmap_sem);
1086 npages = get_user_page_nowait(current, current->mm,
1087 addr, write_fault, page);
1088 up_read(&current->mm->mmap_sem);
1089 } else
1090 npages = get_user_pages_fast(addr, 1, write_fault,
1091 page);
1074 1092
1075 /* map read fault as writable if possible */ 1093 /* map read fault as writable if possible */
1076 if (unlikely(!write_fault) && npages == 1) { 1094 if (unlikely(!write_fault) && npages == 1) {
@@ -1093,7 +1111,8 @@ static pfn_t hva_to_pfn(struct kvm *kvm, unsigned long addr, bool atomic,
1093 return get_fault_pfn(); 1111 return get_fault_pfn();
1094 1112
1095 down_read(&current->mm->mmap_sem); 1113 down_read(&current->mm->mmap_sem);
1096 if (check_user_page_hwpoison(addr)) { 1114 if (npages == -EHWPOISON ||
1115 (!async && check_user_page_hwpoison(addr))) {
1097 up_read(&current->mm->mmap_sem); 1116 up_read(&current->mm->mmap_sem);
1098 get_page(hwpoison_page); 1117 get_page(hwpoison_page);
1099 return page_to_pfn(hwpoison_page); 1118 return page_to_pfn(hwpoison_page);