|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH 4/5] x86/shadow: Drop all emulation for PV vcpus
Emulation is only performed for paging_mode_refcount() domains, which in
practice means HVM domains only.
Drop the PV emulation code. As it always set addr_side and sp_size to
BITS_PER_LONG, it can't have worked correctly for PV guests running in a
different mode to Xen.
Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
---
CC: Jan Beulich <JBeulich@xxxxxxxx>
CC: Tim Deegan <tim@xxxxxxx>
CC: George Dunlap <george.dunlap@xxxxxxxxxxxxx>
---
xen/arch/x86/mm/shadow/common.c | 111 +++++++---------------------------------
xen/arch/x86/mm/shadow/multi.c | 21 ++------
2 files changed, 22 insertions(+), 110 deletions(-)
diff --git a/xen/arch/x86/mm/shadow/common.c b/xen/arch/x86/mm/shadow/common.c
index 84a87f3..2525a57 100644
--- a/xen/arch/x86/mm/shadow/common.c
+++ b/xen/arch/x86/mm/shadow/common.c
@@ -318,75 +318,6 @@ static const struct x86_emulate_ops
hvm_shadow_emulator_ops = {
.cpuid = hvmemul_cpuid,
};
-static int
-pv_emulate_read(enum x86_segment seg,
- unsigned long offset,
- void *p_data,
- unsigned int bytes,
- struct x86_emulate_ctxt *ctxt)
-{
- unsigned int rc;
-
- if ( !is_x86_user_segment(seg) )
- return X86EMUL_UNHANDLEABLE;
-
- if ( (rc = copy_from_user(p_data, (void *)offset, bytes)) != 0 )
- {
- x86_emul_pagefault(0, offset + bytes - rc, ctxt); /* Read fault. */
- return X86EMUL_EXCEPTION;
- }
-
- return X86EMUL_OKAY;
-}
-
-static int
-pv_emulate_write(enum x86_segment seg,
- unsigned long offset,
- void *p_data,
- unsigned int bytes,
- struct x86_emulate_ctxt *ctxt)
-{
- struct sh_emulate_ctxt *sh_ctxt =
- container_of(ctxt, struct sh_emulate_ctxt, ctxt);
- struct vcpu *v = current;
- if ( !is_x86_user_segment(seg) )
- return X86EMUL_UNHANDLEABLE;
- return v->arch.paging.mode->shadow.x86_emulate_write(
- v, offset, p_data, bytes, sh_ctxt);
-}
-
-static int
-pv_emulate_cmpxchg(enum x86_segment seg,
- unsigned long offset,
- void *p_old,
- void *p_new,
- unsigned int bytes,
- struct x86_emulate_ctxt *ctxt)
-{
- struct sh_emulate_ctxt *sh_ctxt =
- container_of(ctxt, struct sh_emulate_ctxt, ctxt);
- unsigned long old, new;
- struct vcpu *v = current;
-
- if ( !is_x86_user_segment(seg) || bytes > sizeof(long) )
- return X86EMUL_UNHANDLEABLE;
-
- old = new = 0;
- memcpy(&old, p_old, bytes);
- memcpy(&new, p_new, bytes);
-
- return v->arch.paging.mode->shadow.x86_emulate_cmpxchg(
- v, offset, old, new, bytes, sh_ctxt);
-}
-
-static const struct x86_emulate_ops pv_shadow_emulator_ops = {
- .read = pv_emulate_read,
- .insn_fetch = pv_emulate_read,
- .write = pv_emulate_write,
- .cmpxchg = pv_emulate_cmpxchg,
- .cpuid = pv_emul_cpuid,
-};
-
const struct x86_emulate_ops *shadow_init_emulation(
struct sh_emulate_ctxt *sh_ctxt, struct cpu_user_regs *regs)
{
@@ -394,17 +325,13 @@ const struct x86_emulate_ops *shadow_init_emulation(
struct vcpu *v = current;
unsigned long addr;
+ ASSERT(has_hvm_container_vcpu(v));
+
memset(sh_ctxt, 0, sizeof(*sh_ctxt));
sh_ctxt->ctxt.regs = regs;
sh_ctxt->ctxt.swint_emulate = x86_swint_emulate_none;
- if ( is_pv_vcpu(v) )
- {
- sh_ctxt->ctxt.addr_size = sh_ctxt->ctxt.sp_size = BITS_PER_LONG;
- return &pv_shadow_emulator_ops;
- }
-
/* Segment cache initialisation. Primed with CS. */
creg = hvm_get_seg_reg(x86_seg_cs, sh_ctxt);
@@ -441,24 +368,24 @@ void shadow_continue_emulation(struct sh_emulate_ctxt
*sh_ctxt,
struct vcpu *v = current;
unsigned long addr, diff;
- /* We don't refetch the segment bases, because we don't emulate
- * writes to segment registers */
+ ASSERT(has_hvm_container_vcpu(v));
- if ( is_hvm_vcpu(v) )
- {
- diff = regs->eip - sh_ctxt->insn_buf_eip;
- if ( diff > sh_ctxt->insn_buf_bytes )
- {
- /* Prefetch more bytes. */
- sh_ctxt->insn_buf_bytes =
- (!hvm_translate_linear_addr(
- x86_seg_cs, regs->eip, sizeof(sh_ctxt->insn_buf),
- hvm_access_insn_fetch, sh_ctxt, &addr) &&
- !hvm_fetch_from_guest_linear(
- sh_ctxt->insn_buf, addr, sizeof(sh_ctxt->insn_buf), 0,
NULL))
- ? sizeof(sh_ctxt->insn_buf) : 0;
- sh_ctxt->insn_buf_eip = regs->eip;
- }
+ /*
+ * We don't refetch the segment bases, because we don't emulate
+ * writes to segment registers
+ */
+ diff = regs->eip - sh_ctxt->insn_buf_eip;
+ if ( diff > sh_ctxt->insn_buf_bytes )
+ {
+ /* Prefetch more bytes. */
+ sh_ctxt->insn_buf_bytes =
+ (!hvm_translate_linear_addr(
+ x86_seg_cs, regs->eip, sizeof(sh_ctxt->insn_buf),
+ hvm_access_insn_fetch, sh_ctxt, &addr) &&
+ !hvm_fetch_from_guest_linear(
+ sh_ctxt->insn_buf, addr, sizeof(sh_ctxt->insn_buf), 0, NULL))
+ ? sizeof(sh_ctxt->insn_buf) : 0;
+ sh_ctxt->insn_buf_eip = regs->eip;
}
}
diff --git a/xen/arch/x86/mm/shadow/multi.c b/xen/arch/x86/mm/shadow/multi.c
index 67c98b9..713f23d 100644
--- a/xen/arch/x86/mm/shadow/multi.c
+++ b/xen/arch/x86/mm/shadow/multi.c
@@ -3382,12 +3382,7 @@ static int sh_page_fault(struct vcpu *v,
(((emul_ctxt.ctxt.event.vector == TRAP_gp_fault) ||
(emul_ctxt.ctxt.event.vector == TRAP_stack_error)) &&
emul_ctxt.ctxt.event.error_code == 0)) )
- {
- if ( has_hvm_container_domain(d) )
- hvm_inject_event(&emul_ctxt.ctxt.event);
- else
- pv_inject_event(&emul_ctxt.ctxt.event);
- }
+ hvm_inject_event(&emul_ctxt.ctxt.event);
else
{
SHADOW_PRINTK(
@@ -3447,12 +3442,7 @@ static int sh_page_fault(struct vcpu *v,
#endif
if ( emul_ctxt.ctxt.retire.singlestep )
- {
- if ( has_hvm_container_domain(d) )
- hvm_inject_hw_exception(TRAP_debug, X86_EVENT_NO_EC);
- else
- pv_inject_hw_exception(TRAP_debug, X86_EVENT_NO_EC);
- }
+ hvm_inject_hw_exception(TRAP_debug, X86_EVENT_NO_EC);
#if GUEST_PAGING_LEVELS == 3 /* PAE guest */
/*
@@ -3493,12 +3483,7 @@ static int sh_page_fault(struct vcpu *v,
TRACE_SHADOW_PATH_FLAG(TRCE_SFLAG_EMULATION_LAST_FAILED);
if ( emul_ctxt.ctxt.retire.singlestep )
- {
- if ( has_hvm_container_domain(d) )
- hvm_inject_hw_exception(TRAP_debug, X86_EVENT_NO_EC);
- else
- pv_inject_hw_exception(TRAP_debug, X86_EVENT_NO_EC);
- }
+ hvm_inject_hw_exception(TRAP_debug, X86_EVENT_NO_EC);
break; /* Don't emulate again if we failed! */
}
--
2.1.4
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
https://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |