diff options
author | Cédric Le Goater <clg@fr.ibm.com> | 2014-01-09 05:51:16 -0500 |
---|---|---|
committer | Alexander Graf <agraf@suse.de> | 2014-01-27 10:00:39 -0500 |
commit | 736017752d2f6ed0d64f5e15cf48e79779b11c85 (patch) | |
tree | f8f11f7b4daf14e76e8e3d716c85f5a56a949476 /arch | |
parent | 7a8ff56be68239bd36a2b639cb40bfbcfc58dad3 (diff) |
KVM: PPC: Book3S: MMIO emulation support for little endian guests
MMIO emulation reads the last instruction executed by the guest
and then emulates. If the guest is running in Little Endian order,
or more generally in a different endian order of the host, the
instruction needs to be byte-swapped before being emulated.
This patch adds a helper routine which tests the endian order of
the host and the guest in order to decide whether a byteswap is
needed or not. It is then used to byteswap the last instruction
of the guest in the endian order of the host before MMIO emulation
is performed.
Finally, kvmppc_handle_load() of kvmppc_handle_store() are modified
to reverse the endianness of the MMIO if required.
Signed-off-by: Cédric Le Goater <clg@fr.ibm.com>
[agraf: add booke handling]
Signed-off-by: Alexander Graf <agraf@suse.de>
Diffstat (limited to 'arch')
-rw-r--r-- | arch/powerpc/include/asm/kvm_book3s.h | 8 | ||||
-rw-r--r-- | arch/powerpc/include/asm/kvm_booke.h | 6 | ||||
-rw-r--r-- | arch/powerpc/include/asm/kvm_ppc.h | 7 | ||||
-rw-r--r-- | arch/powerpc/kvm/book3s_64_mmu_hv.c | 2 | ||||
-rw-r--r-- | arch/powerpc/kvm/emulate.c | 1 | ||||
-rw-r--r-- | arch/powerpc/kvm/powerpc.c | 28 |
6 files changed, 42 insertions, 10 deletions
diff --git a/arch/powerpc/include/asm/kvm_book3s.h b/arch/powerpc/include/asm/kvm_book3s.h index f8b23201c105..1e9c26f45d18 100644 --- a/arch/powerpc/include/asm/kvm_book3s.h +++ b/arch/powerpc/include/asm/kvm_book3s.h | |||
@@ -264,6 +264,11 @@ static inline ulong kvmppc_get_pc(struct kvm_vcpu *vcpu) | |||
264 | return vcpu->arch.pc; | 264 | return vcpu->arch.pc; |
265 | } | 265 | } |
266 | 266 | ||
267 | static inline bool kvmppc_need_byteswap(struct kvm_vcpu *vcpu) | ||
268 | { | ||
269 | return (vcpu->arch.shared->msr & MSR_LE) != (MSR_KERNEL & MSR_LE); | ||
270 | } | ||
271 | |||
267 | static inline u32 kvmppc_get_last_inst_internal(struct kvm_vcpu *vcpu, ulong pc) | 272 | static inline u32 kvmppc_get_last_inst_internal(struct kvm_vcpu *vcpu, ulong pc) |
268 | { | 273 | { |
269 | /* Load the instruction manually if it failed to do so in the | 274 | /* Load the instruction manually if it failed to do so in the |
@@ -271,7 +276,8 @@ static inline u32 kvmppc_get_last_inst_internal(struct kvm_vcpu *vcpu, ulong pc) | |||
271 | if (vcpu->arch.last_inst == KVM_INST_FETCH_FAILED) | 276 | if (vcpu->arch.last_inst == KVM_INST_FETCH_FAILED) |
272 | kvmppc_ld(vcpu, &pc, sizeof(u32), &vcpu->arch.last_inst, false); | 277 | kvmppc_ld(vcpu, &pc, sizeof(u32), &vcpu->arch.last_inst, false); |
273 | 278 | ||
274 | return vcpu->arch.last_inst; | 279 | return kvmppc_need_byteswap(vcpu) ? swab32(vcpu->arch.last_inst) : |
280 | vcpu->arch.last_inst; | ||
275 | } | 281 | } |
276 | 282 | ||
277 | static inline u32 kvmppc_get_last_inst(struct kvm_vcpu *vcpu) | 283 | static inline u32 kvmppc_get_last_inst(struct kvm_vcpu *vcpu) |
diff --git a/arch/powerpc/include/asm/kvm_booke.h b/arch/powerpc/include/asm/kvm_booke.h index dd8f61510dfd..80d46b5a7efb 100644 --- a/arch/powerpc/include/asm/kvm_booke.h +++ b/arch/powerpc/include/asm/kvm_booke.h | |||
@@ -63,6 +63,12 @@ static inline u32 kvmppc_get_xer(struct kvm_vcpu *vcpu) | |||
63 | return vcpu->arch.xer; | 63 | return vcpu->arch.xer; |
64 | } | 64 | } |
65 | 65 | ||
66 | static inline bool kvmppc_need_byteswap(struct kvm_vcpu *vcpu) | ||
67 | { | ||
68 | /* XXX Would need to check TLB entry */ | ||
69 | return false; | ||
70 | } | ||
71 | |||
66 | static inline u32 kvmppc_get_last_inst(struct kvm_vcpu *vcpu) | 72 | static inline u32 kvmppc_get_last_inst(struct kvm_vcpu *vcpu) |
67 | { | 73 | { |
68 | return vcpu->arch.last_inst; | 74 | return vcpu->arch.last_inst; |
diff --git a/arch/powerpc/include/asm/kvm_ppc.h b/arch/powerpc/include/asm/kvm_ppc.h index c8317fbf92c4..629277df4798 100644 --- a/arch/powerpc/include/asm/kvm_ppc.h +++ b/arch/powerpc/include/asm/kvm_ppc.h | |||
@@ -54,12 +54,13 @@ extern void kvmppc_handler_highmem(void); | |||
54 | extern void kvmppc_dump_vcpu(struct kvm_vcpu *vcpu); | 54 | extern void kvmppc_dump_vcpu(struct kvm_vcpu *vcpu); |
55 | extern int kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu, | 55 | extern int kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu, |
56 | unsigned int rt, unsigned int bytes, | 56 | unsigned int rt, unsigned int bytes, |
57 | int is_bigendian); | 57 | int is_default_endian); |
58 | extern int kvmppc_handle_loads(struct kvm_run *run, struct kvm_vcpu *vcpu, | 58 | extern int kvmppc_handle_loads(struct kvm_run *run, struct kvm_vcpu *vcpu, |
59 | unsigned int rt, unsigned int bytes, | 59 | unsigned int rt, unsigned int bytes, |
60 | int is_bigendian); | 60 | int is_default_endian); |
61 | extern int kvmppc_handle_store(struct kvm_run *run, struct kvm_vcpu *vcpu, | 61 | extern int kvmppc_handle_store(struct kvm_run *run, struct kvm_vcpu *vcpu, |
62 | u64 val, unsigned int bytes, int is_bigendian); | 62 | u64 val, unsigned int bytes, |
63 | int is_default_endian); | ||
63 | 64 | ||
64 | extern int kvmppc_emulate_instruction(struct kvm_run *run, | 65 | extern int kvmppc_emulate_instruction(struct kvm_run *run, |
65 | struct kvm_vcpu *vcpu); | 66 | struct kvm_vcpu *vcpu); |
diff --git a/arch/powerpc/kvm/book3s_64_mmu_hv.c b/arch/powerpc/kvm/book3s_64_mmu_hv.c index f3ff587a8b7d..efb8aa544876 100644 --- a/arch/powerpc/kvm/book3s_64_mmu_hv.c +++ b/arch/powerpc/kvm/book3s_64_mmu_hv.c | |||
@@ -558,7 +558,7 @@ static int kvmppc_hv_emulate_mmio(struct kvm_run *run, struct kvm_vcpu *vcpu, | |||
558 | * we just return and retry the instruction. | 558 | * we just return and retry the instruction. |
559 | */ | 559 | */ |
560 | 560 | ||
561 | if (instruction_is_store(vcpu->arch.last_inst) != !!is_store) | 561 | if (instruction_is_store(kvmppc_get_last_inst(vcpu)) != !!is_store) |
562 | return RESUME_GUEST; | 562 | return RESUME_GUEST; |
563 | 563 | ||
564 | /* | 564 | /* |
diff --git a/arch/powerpc/kvm/emulate.c b/arch/powerpc/kvm/emulate.c index 2f9a0873b44f..c2b887be2c29 100644 --- a/arch/powerpc/kvm/emulate.c +++ b/arch/powerpc/kvm/emulate.c | |||
@@ -219,7 +219,6 @@ static int kvmppc_emulate_mfspr(struct kvm_vcpu *vcpu, int sprn, int rt) | |||
219 | * lmw | 219 | * lmw |
220 | * stmw | 220 | * stmw |
221 | * | 221 | * |
222 | * XXX is_bigendian should depend on MMU mapping or MSR[LE] | ||
223 | */ | 222 | */ |
224 | /* XXX Should probably auto-generate instruction decoding for a particular core | 223 | /* XXX Should probably auto-generate instruction decoding for a particular core |
225 | * from opcode tables in the future. */ | 224 | * from opcode tables in the future. */ |
diff --git a/arch/powerpc/kvm/powerpc.c b/arch/powerpc/kvm/powerpc.c index 7ca9e0a80499..026dfaaa4772 100644 --- a/arch/powerpc/kvm/powerpc.c +++ b/arch/powerpc/kvm/powerpc.c | |||
@@ -673,9 +673,19 @@ static void kvmppc_complete_mmio_load(struct kvm_vcpu *vcpu, | |||
673 | } | 673 | } |
674 | 674 | ||
675 | int kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu, | 675 | int kvmppc_handle_load(struct kvm_run *run, struct kvm_vcpu *vcpu, |
676 | unsigned int rt, unsigned int bytes, int is_bigendian) | 676 | unsigned int rt, unsigned int bytes, |
677 | int is_default_endian) | ||
677 | { | 678 | { |
678 | int idx, ret; | 679 | int idx, ret; |
680 | int is_bigendian; | ||
681 | |||
682 | if (kvmppc_need_byteswap(vcpu)) { | ||
683 | /* Default endianness is "little endian". */ | ||
684 | is_bigendian = !is_default_endian; | ||
685 | } else { | ||
686 | /* Default endianness is "big endian". */ | ||
687 | is_bigendian = is_default_endian; | ||
688 | } | ||
679 | 689 | ||
680 | if (bytes > sizeof(run->mmio.data)) { | 690 | if (bytes > sizeof(run->mmio.data)) { |
681 | printk(KERN_ERR "%s: bad MMIO length: %d\n", __func__, | 691 | printk(KERN_ERR "%s: bad MMIO length: %d\n", __func__, |
@@ -711,21 +721,31 @@ EXPORT_SYMBOL_GPL(kvmppc_handle_load); | |||
711 | 721 | ||
712 | /* Same as above, but sign extends */ | 722 | /* Same as above, but sign extends */ |
713 | int kvmppc_handle_loads(struct kvm_run *run, struct kvm_vcpu *vcpu, | 723 | int kvmppc_handle_loads(struct kvm_run *run, struct kvm_vcpu *vcpu, |
714 | unsigned int rt, unsigned int bytes, int is_bigendian) | 724 | unsigned int rt, unsigned int bytes, |
725 | int is_default_endian) | ||
715 | { | 726 | { |
716 | int r; | 727 | int r; |
717 | 728 | ||
718 | vcpu->arch.mmio_sign_extend = 1; | 729 | vcpu->arch.mmio_sign_extend = 1; |
719 | r = kvmppc_handle_load(run, vcpu, rt, bytes, is_bigendian); | 730 | r = kvmppc_handle_load(run, vcpu, rt, bytes, is_default_endian); |
720 | 731 | ||
721 | return r; | 732 | return r; |
722 | } | 733 | } |
723 | 734 | ||
724 | int kvmppc_handle_store(struct kvm_run *run, struct kvm_vcpu *vcpu, | 735 | int kvmppc_handle_store(struct kvm_run *run, struct kvm_vcpu *vcpu, |
725 | u64 val, unsigned int bytes, int is_bigendian) | 736 | u64 val, unsigned int bytes, int is_default_endian) |
726 | { | 737 | { |
727 | void *data = run->mmio.data; | 738 | void *data = run->mmio.data; |
728 | int idx, ret; | 739 | int idx, ret; |
740 | int is_bigendian; | ||
741 | |||
742 | if (kvmppc_need_byteswap(vcpu)) { | ||
743 | /* Default endianness is "little endian". */ | ||
744 | is_bigendian = !is_default_endian; | ||
745 | } else { | ||
746 | /* Default endianness is "big endian". */ | ||
747 | is_bigendian = is_default_endian; | ||
748 | } | ||
729 | 749 | ||
730 | if (bytes > sizeof(run->mmio.data)) { | 750 | if (bytes > sizeof(run->mmio.data)) { |
731 | printk(KERN_ERR "%s: bad MMIO length: %d\n", __func__, | 751 | printk(KERN_ERR "%s: bad MMIO length: %d\n", __func__, |