[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v3 2/3] x86/emulate: add support of emulating SSE2 instruction {, v}movd mm, r32/m32 and {, v}movq mm, r64
>>> On 01.08.16 at 14:53, <mdontu@xxxxxxxxxxxxxxx> wrote: > On Monday 01 August 2016 10:52:12 Andrew Cooper wrote: >> On 01/08/16 03:52, Mihai Donțu wrote: >> > Found that Windows driver was using a SSE2 instruction MOVD. >> > >> > Signed-off-by: Zhi Wang <zhi.a.wang@xxxxxxxxx> >> > Signed-off-by: Mihai Donțu <mdontu@xxxxxxxxxxxxxxx> >> > --- >> > Picked from the XenServer 7 patch queue, as suggested by Andrew Cooper >> > >> > Changed since v2: >> > * handle the case where the destination is a GPR >> > --- >> > xen/arch/x86/x86_emulate/x86_emulate.c | 38 > +++++++++++++++++++++++++++++++--- >> > 1 file changed, 35 insertions(+), 3 deletions(-) >> > >> > diff --git a/xen/arch/x86/x86_emulate/x86_emulate.c > b/xen/arch/x86/x86_emulate/x86_emulate.c >> > index 44de3b6..9f89ada 100644 >> > --- a/xen/arch/x86/x86_emulate/x86_emulate.c >> > +++ b/xen/arch/x86/x86_emulate/x86_emulate.c >> > @@ -204,7 +204,7 @@ static uint8_t twobyte_table[256] = { >> > /* 0x60 - 0x6F */ >> > 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ImplicitOps|ModRM, >> > /* 0x70 - 0x7F */ >> > - 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ImplicitOps|ModRM, >> > + 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, ImplicitOps|ModRM, > ImplicitOps|ModRM, >> > /* 0x80 - 0x87 */ >> > ImplicitOps, ImplicitOps, ImplicitOps, ImplicitOps, >> > ImplicitOps, ImplicitOps, ImplicitOps, ImplicitOps, >> > @@ -4409,6 +4409,10 @@ x86_emulate( >> > case 0x6f: /* movq mm/m64,mm */ >> > /* {,v}movdq{a,u} xmm/m128,xmm */ >> > /* vmovdq{a,u} ymm/m256,ymm */ >> > + case 0x7e: /* movd mm,r/m32 */ >> > + /* movq mm,r/m64 */ >> > + /* {,v}movd xmm,r/m32 */ >> > + /* {,v}movq xmm,r/m64 */ >> > case 0x7f: /* movq mm,mm/m64 */ >> > /* {,v}movdq{a,u} xmm,xmm/m128 */ >> > /* vmovdq{a,u} ymm,ymm/m256 */ >> > @@ -4432,7 +4436,17 @@ x86_emulate( >> > host_and_vcpu_must_have(sse2); >> > buf[0] = 0x66; /* SSE */ >> > get_fpu(X86EMUL_FPU_xmm, &fic); >> > - ea.bytes = (b == 0xd6 ? 8 : 16); >> > + switch ( b ) >> > + { >> > + case 0x7e: >> > + ea.bytes = 4; >> > + break; >> > + case 0xd6: >> > + ea.bytes = 8; >> > + break; >> > + default: >> > + ea.bytes = 16; >> > + } >> > break; >> > case vex_none: >> > if ( b != 0xe7 ) >> > @@ -4452,7 +4466,17 @@ x86_emulate( >> > ((vex.pfx != vex_66) && (vex.pfx != vex_f3))); >> > host_and_vcpu_must_have(avx); >> > get_fpu(X86EMUL_FPU_ymm, &fic); >> > - ea.bytes = (b == 0xd6 ? 8 : (16 << vex.l)); >> > + switch ( b ) >> > + { >> > + case 0x7e: >> > + ea.bytes = 4; >> > + break; >> > + case 0xd6: >> > + ea.bytes = 8; >> > + break; >> > + default: >> > + ea.bytes = 16 << vex.l; >> > + } >> > } >> > if ( ea.type == OP_MEM ) >> > { >> > @@ -4468,6 +4492,14 @@ x86_emulate( >> > vex.b = 1; >> > buf[4] &= 0x38; >> > } >> > + else if ( b == 0x7e ) >> > + { >> > + /* convert the GPR destination to (%rAX) */ >> > + *((unsigned long *)&mmvalp) = (unsigned long)ea.reg; >> > + rex_prefix &= ~REX_B; >> > + vex.b = 1; >> > + buf[4] &= 0x38; >> > + } >> >> Thankyou for doing this. However, looking at it, it has some code in >> common with the "ea.type == OP_MEM" clause. >> >> Would this work? >> >> diff --git a/xen/arch/x86/x86_emulate/x86_emulate.c >> b/xen/arch/x86/x86_emulate/x86_emulate.c >> index fe594ba..90db067 100644 >> --- a/xen/arch/x86/x86_emulate/x86_emulate.c >> +++ b/xen/arch/x86/x86_emulate/x86_emulate.c >> @@ -4453,16 +4453,25 @@ x86_emulate( >> get_fpu(X86EMUL_FPU_ymm, &fic); >> ea.bytes = 16 << vex.l; >> } >> - if ( ea.type == OP_MEM ) >> + if ( ea.type == OP_MEM || ea.type == OP_REG ) >> { >> - /* XXX enable once there is ops->ea() or equivalent >> - generate_exception_if((vex.pfx == vex_66) && >> - (ops->ea(ea.mem.seg, ea.mem.off) >> - & (ea.bytes - 1)), EXC_GP, 0); */ >> - if ( b == 0x6f ) >> - rc = ops->read(ea.mem.seg, ea.mem.off+0, mmvalp, >> - ea.bytes, ctxt); >> /* convert memory operand to (%rAX) */ >> + >> + if ( ea.type == OP_MEM) >> + { >> + /* XXX enable once there is ops->ea() or equivalent >> + generate_exception_if((vex.pfx == vex_66) && >> + (ops->ea(ea.mem.seg, ea.mem.off) >> + & (ea.bytes - 1)), EXC_GP, 0); */ >> + if ( b == 0x6f ) >> + rc = ops->read(ea.mem.seg, ea.mem.off+0, mmvalp, >> + ea.bytes, ctxt); >> + } >> + else if ( ea.type == OP_REG ) >> + { >> + *((unsigned long *)&mmvalp) = (unsigned long)ea.reg; >> + } >> + >> rex_prefix &= ~REX_B; >> vex.b = 1; >> buf[4] &= 0x38; >> >> >> This is untested, but avoids duplicating this bit of state maniupulation. > > Your suggestion makes sense, but I'm starting to doubt my initial > patch. :-) I'm testing "movq xmm1, xmm1" and noticing that it takes the > GPR-handling route and I can't seem to be able to easily prevent it > with !(rex_prefix & REX_B), as rex_prefix == 0 and vex.b == 1. I need > to take a harder look at how that class of instructions is coded. You obviously need to distinguish the two kinds of register sources/ destinations: GPRs need suitable re-writing of the instruction (without having looked at the most recent version of the patch yet I btw doubt converting register to memory operands is the most efficient model), while MMs, XMMs, and YMMs can retain their register encoding. Jan _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |