[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH 2/6] xen/pvh: Extend vcpu_guest_context, p2m, event, and xenbus to support PVH.



On Sat, 20 Oct 2012, Konrad Rzeszutek Wilk wrote:
> From: Mukesh Rathor <mukesh.rathor@xxxxxxxxxx>
> 
> make gdt_frames[]/gdt_ents into a union with {gdtaddr, gdtsz}, as PVH
> only needs to send down gdtaddr and gdtsz.
> 
> For interrupts, PVH uses native_irq_ops.
> vcpu hotplug is currently not available for PVH.
> 
> For events we follow what PVHVM does - to use callback vector.
> Lastly, also use HVM path to setup XenBus.
> 
> Signed-off-by: Mukesh Rathor <mukesh.rathor@xxxxxxxxxx>
> Signed-off-by: Konrad Rzeszutek Wilk <konrad.wilk@xxxxxxxxxx>
> ---
>  arch/x86/include/asm/xen/interface.h |   11 +++++-
>  arch/x86/xen/irq.c                   |    5 ++-
>  arch/x86/xen/p2m.c                   |    2 +-
>  arch/x86/xen/smp.c                   |   75 
> ++++++++++++++++++++++------------
>  drivers/xen/cpu_hotplug.c            |    4 +-
>  drivers/xen/events.c                 |    9 ++++-
>  drivers/xen/xenbus/xenbus_client.c   |    3 +-
>  7 files changed, 77 insertions(+), 32 deletions(-)
> 
> diff --git a/arch/x86/include/asm/xen/interface.h 
> b/arch/x86/include/asm/xen/interface.h
> index 6d2f75a..4c08f23 100644
> --- a/arch/x86/include/asm/xen/interface.h
> +++ b/arch/x86/include/asm/xen/interface.h
> @@ -144,7 +144,16 @@ struct vcpu_guest_context {
>      struct cpu_user_regs user_regs;         /* User-level CPU registers     
> */
>      struct trap_info trap_ctxt[256];        /* Virtual IDT                  
> */
>      unsigned long ldt_base, ldt_ents;       /* LDT (linear address, # ents) 
> */
> -    unsigned long gdt_frames[16], gdt_ents; /* GDT (machine frames, # ents) 
> */
> +    union {
> +     struct {
> +             /* PV: GDT (machine frames, # ents).*/
> +             unsigned long gdt_frames[16], gdt_ents;
> +     } pv;
> +     struct {
> +             /* PVH: GDTR addr and size */
> +             unsigned long gdtaddr, gdtsz;
> +     } pvh;
> +    } u;
>      unsigned long kernel_ss, kernel_sp;     /* Virtual TSS (only SS1/SP1)   
> */
>      /* NB. User pagetable on x86/64 is placed in ctrlreg[1]. */
>      unsigned long ctrlreg[8];               /* CR0-CR7 (control registers)  
> */
> diff --git a/arch/x86/xen/irq.c b/arch/x86/xen/irq.c
> index 01a4dc0..fcbe56a 100644
> --- a/arch/x86/xen/irq.c
> +++ b/arch/x86/xen/irq.c
> @@ -5,6 +5,7 @@
>  #include <xen/interface/xen.h>
>  #include <xen/interface/sched.h>
>  #include <xen/interface/vcpu.h>
> +#include <xen/features.h>
>  #include <xen/events.h>
>  
>  #include <asm/xen/hypercall.h>
> @@ -129,6 +130,8 @@ static const struct pv_irq_ops xen_irq_ops __initconst = {
>  
>  void __init xen_init_irq_ops(void)
>  {
> -     pv_irq_ops = xen_irq_ops;
> +     /* For PVH we use default pv_irq_ops settings */
> +     if (!xen_feature(XENFEAT_hvm_callback_vector))
> +             pv_irq_ops = xen_irq_ops;
>       x86_init.irqs.intr_init = xen_init_IRQ;
>  }
> diff --git a/arch/x86/xen/p2m.c b/arch/x86/xen/p2m.c
> index 95fb2aa..ea553c8 100644
> --- a/arch/x86/xen/p2m.c
> +++ b/arch/x86/xen/p2m.c
> @@ -798,7 +798,7 @@ bool __set_phys_to_machine(unsigned long pfn, unsigned 
> long mfn)
>  {
>       unsigned topidx, mididx, idx;
>  
> -     if (unlikely(xen_feature(XENFEAT_auto_translated_physmap))) {
> +     if (xen_feature(XENFEAT_auto_translated_physmap)) {
>               BUG_ON(pfn != mfn && mfn != INVALID_P2M_ENTRY);
>               return true;
>       }
> diff --git a/arch/x86/xen/smp.c b/arch/x86/xen/smp.c
> index 353c50f..df400349 100644
> --- a/arch/x86/xen/smp.c
> +++ b/arch/x86/xen/smp.c
> @@ -68,9 +68,11 @@ static void __cpuinit cpu_bringup(void)
>       touch_softlockup_watchdog();
>       preempt_disable();
>  
> -     xen_enable_sysenter();
> -     xen_enable_syscall();
> -
> +     /* PVH runs in ring 0 and allows us to do native syscalls. Yay! */
> +     if (!xen_feature(XENFEAT_supervisor_mode_kernel)) {
> +             xen_enable_sysenter();
> +             xen_enable_syscall();
> +     }
>       cpu = smp_processor_id();
>       smp_store_cpu_info(cpu);
>       cpu_data(cpu).x86_max_cores = 1;
> @@ -230,10 +232,11 @@ static void __init xen_smp_prepare_boot_cpu(void)
>       BUG_ON(smp_processor_id() != 0);
>       native_smp_prepare_boot_cpu();
>  
> -     /* We've switched to the "real" per-cpu gdt, so make sure the
> -        old memory can be recycled */
> -     make_lowmem_page_readwrite(xen_initial_gdt);
> -
> +     if (!xen_feature(XENFEAT_writable_page_tables)) {
> +             /* We've switched to the "real" per-cpu gdt, so make sure the
> +              * old memory can be recycled */
> +             make_lowmem_page_readwrite(xen_initial_gdt);
> +     }
>       xen_filter_cpu_maps();
>       xen_setup_vcpu_info_placement();
>  }
> @@ -300,8 +303,6 @@ cpu_initialize_context(unsigned int cpu, struct 
> task_struct *idle)
>       gdt = get_cpu_gdt_table(cpu);
>  
>       ctxt->flags = VGCF_IN_KERNEL;
> -     ctxt->user_regs.ds = __USER_DS;
> -     ctxt->user_regs.es = __USER_DS;
>       ctxt->user_regs.ss = __KERNEL_DS;
>  #ifdef CONFIG_X86_32
>       ctxt->user_regs.fs = __KERNEL_PERCPU;
> @@ -310,35 +311,57 @@ cpu_initialize_context(unsigned int cpu, struct 
> task_struct *idle)
>       ctxt->gs_base_kernel = per_cpu_offset(cpu);
>  #endif
>       ctxt->user_regs.eip = (unsigned long)cpu_bringup_and_idle;
> -     ctxt->user_regs.eflags = 0x1000; /* IOPL_RING1 */
>  
>       memset(&ctxt->fpu_ctxt, 0, sizeof(ctxt->fpu_ctxt));
>  
> -     xen_copy_trap_info(ctxt->trap_ctxt);
> +     /* check for autoxlated to get it right for 32bit kernel */

I am not sure what this comment means, considering that in another
comment below you say that we don't support 32bit PVH kernels.


> +     if (xen_feature(XENFEAT_auto_translated_physmap) &&
> +         xen_feature(XENFEAT_supervisor_mode_kernel)) {
>  
> -     ctxt->ldt_ents = 0;
> +             ctxt->user_regs.ds = __KERNEL_DS;
> +             ctxt->user_regs.es = 0;
> +             ctxt->user_regs.gs = 0;
>  
> -     BUG_ON((unsigned long)gdt & ~PAGE_MASK);
> +             ctxt->u.pvh.gdtaddr = (unsigned long)gdt;
> +             ctxt->u.pvh.gdtsz = (unsigned long)(GDT_SIZE - 1);
>  
> -     gdt_mfn = arbitrary_virt_to_mfn(gdt);
> -     make_lowmem_page_readonly(gdt);
> -     make_lowmem_page_readonly(mfn_to_virt(gdt_mfn));
> +#ifdef CONFIG_X86_64
> +             /* Note: PVH is not supported on x86_32. */
> +             ctxt->gs_base_user = (unsigned long)
> +                                     per_cpu(irq_stack_union.gs_base, cpu);
> +#endif
> +     } else {
> +             ctxt->user_regs.eflags = 0x1000; /* IOPL_RING1 */
> +             ctxt->user_regs.ds = __USER_DS;
> +             ctxt->user_regs.es = __USER_DS;
>  
> -     ctxt->gdt_frames[0] = gdt_mfn;
> -     ctxt->gdt_ents      = GDT_ENTRIES;
> +             xen_copy_trap_info(ctxt->trap_ctxt);
>  
> -     ctxt->user_regs.cs = __KERNEL_CS;
> -     ctxt->user_regs.esp = idle->thread.sp0 - sizeof(struct pt_regs);
> +             ctxt->ldt_ents = 0;
>  
> -     ctxt->kernel_ss = __KERNEL_DS;
> -     ctxt->kernel_sp = idle->thread.sp0;
> +             BUG_ON((unsigned long)gdt & ~PAGE_MASK);
> +
> +             gdt_mfn = arbitrary_virt_to_mfn(gdt);
> +             make_lowmem_page_readonly(gdt);
> +             make_lowmem_page_readonly(mfn_to_virt(gdt_mfn));
> +
> +             ctxt->u.pv.gdt_frames[0] = gdt_mfn;
> +             ctxt->u.pv.gdt_ents      = GDT_ENTRIES;
> +
> +             ctxt->kernel_ss = __KERNEL_DS;
> +             ctxt->kernel_sp = idle->thread.sp0;
>  
>  #ifdef CONFIG_X86_32
> -     ctxt->event_callback_cs     = __KERNEL_CS;
> -     ctxt->failsafe_callback_cs  = __KERNEL_CS;
> +             ctxt->event_callback_cs     = __KERNEL_CS;
> +             ctxt->failsafe_callback_cs  = __KERNEL_CS;
>  #endif
> -     ctxt->event_callback_eip    = (unsigned long)xen_hypervisor_callback;
> -     ctxt->failsafe_callback_eip = (unsigned long)xen_failsafe_callback;
> +             ctxt->event_callback_eip    =
> +                                     (unsigned long)xen_hypervisor_callback;
> +             ctxt->failsafe_callback_eip =
> +                                     (unsigned long)xen_failsafe_callback;
> +     }
> +     ctxt->user_regs.cs = __KERNEL_CS;
> +     ctxt->user_regs.esp = idle->thread.sp0 - sizeof(struct pt_regs);
>  
>       per_cpu(xen_cr3, cpu) = __pa(swapper_pg_dir);
>       ctxt->ctrlreg[3] = xen_pfn_to_cr3(virt_to_mfn(swapper_pg_dir));

The tradional path looks the same as before, however it is hard to tell
whether the PVH path is correct without the Xen side. For example, what
is gdtsz?


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.