[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH RFC v1 58/74] xen/pvshim: add migration support
From: Roger Pau Monne <roger.pau@xxxxxxxxxx> Signed-off-by: Roger Pau Monné <roger.pau@xxxxxxxxxx> --- xen/arch/x86/guest/xen.c | 29 +++++++ xen/arch/x86/mm.c | 3 +- xen/arch/x86/pv/shim.c | 163 +++++++++++++++++++++++++++++++++++++- xen/common/domain.c | 11 ++- xen/common/schedule.c | 3 +- xen/drivers/char/xen_pv_console.c | 2 +- xen/include/asm-x86/guest/xen.h | 5 ++ xen/include/asm-x86/pv/shim.h | 5 +- xen/include/xen/sched.h | 2 +- 9 files changed, 206 insertions(+), 17 deletions(-) diff --git a/xen/arch/x86/guest/xen.c b/xen/arch/x86/guest/xen.c index 653a7366ab..54c997b9e0 100644 --- a/xen/arch/x86/guest/xen.c +++ b/xen/arch/x86/guest/xen.c @@ -361,6 +361,35 @@ uint32_t hypervisor_cpuid_base(void) return xen_cpuid_base; } +void ap_resume(void *unused) +{ + map_vcpuinfo(); + init_evtchn(); +} + +void hypervisor_resume(void) +{ + /* Reset shared info page. */ + map_shared_info(); + + /* + * Reset vcpu_info. Just clean the mapped bitmap and try to map the vcpu + * area again. On failure to map (when it was previously mapped) panic + * since it's impossible to safely shut down running guest vCPUs in order + * to meet the new XEN_LEGACY_MAX_VCPUS requirement. + */ + memset(vcpu_info_mapped, 0, sizeof(vcpu_info_mapped)); + if ( map_vcpuinfo() && nr_cpu_ids > XEN_LEGACY_MAX_VCPUS ) + panic("unable to remap vCPU info and vCPUs > legacy limit"); + + /* Setup event channel upcall vector. */ + init_evtchn(); + smp_call_function(ap_resume, NULL, 1); + + if ( pv_console ) + pv_console_init(); +} + /* * Local variables: * mode: C diff --git a/xen/arch/x86/mm.c b/xen/arch/x86/mm.c index 4332d3bb39..e726c62064 100644 --- a/xen/arch/x86/mm.c +++ b/xen/arch/x86/mm.c @@ -466,8 +466,7 @@ void share_xen_page_with_guest( spin_unlock(&d->page_alloc_lock); } -int __init unshare_xen_page_with_guest(struct page_info *page, - struct domain *d) +int unshare_xen_page_with_guest(struct page_info *page, struct domain *d) { if ( page_get_owner(page) != d || !is_xen_heap_page(page) ) return -EINVAL; diff --git a/xen/arch/x86/pv/shim.c b/xen/arch/x86/pv/shim.c index 74b24bf950..56ecaea2d2 100644 --- a/xen/arch/x86/pv/shim.c +++ b/xen/arch/x86/pv/shim.c @@ -151,10 +151,167 @@ void __init pv_shim_setup_dom(struct domain *d, l4_pgentry_t *l4start, } } -void pv_shim_shutdown(uint8_t reason) +static void write_start_info(struct domain *d) { - /* XXX: handle suspend */ - xen_hypercall_shutdown(reason); + struct cpu_user_regs *regs = guest_cpu_user_regs(); + start_info_t *si = map_domain_page(_mfn(is_pv_32bit_domain(d) ? regs->edx + : regs->rdx)); + uint64_t param; + + BUG_ON(!si); + + snprintf(si->magic, sizeof(si->magic), "xen-3.0-x86_%s", + is_pv_32bit_domain(d) ? "32p" : "64"); + si->nr_pages = d->tot_pages; + si->shared_info = virt_to_maddr(d->shared_info); + si->flags = (xen_processor_pmbits << 8) & SIF_PM_MASK; + BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_PFN, &si->store_mfn)); + BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_EVTCHN, ¶m)); + si->store_evtchn = param; + BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_EVTCHN, ¶m)); + si->console.domU.evtchn = param; + if ( !pv_console ) + BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_PFN, + &si->console.domU.mfn)); + else + si->console.domU.mfn = virt_to_mfn(consoled_get_ring_addr()); + + if ( is_pv_32bit_domain(d) ) + xlat_start_info(si, XLAT_start_info_console_domU); + + unmap_domain_page(si); +} + +int pv_shim_shutdown(uint8_t reason) +{ + long rc; + + if ( reason == SHUTDOWN_suspend ) + { + struct domain *d = current->domain; + struct vcpu *v; + unsigned int i; + uint64_t old_store_pfn, old_console_pfn = 0, store_pfn, console_pfn; + uint64_t store_evtchn, console_evtchn; + + BUG_ON(current->vcpu_id != 0); + + BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_PFN, + &old_store_pfn)); + if ( !pv_console ) + BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_PFN, + &old_console_pfn)); + + /* Pause the other vcpus before starting the migration. */ + for_each_vcpu(d, v) + if ( v != current ) + vcpu_pause_by_systemcontroller(v); + + rc = xen_hypercall_shutdown(SHUTDOWN_suspend); + if ( rc ) + { + for_each_vcpu(d, v) + if ( v != current ) + vcpu_unpause_by_systemcontroller(v); + + return rc; + } + + /* Resume the shim itself first. */ + hypervisor_resume(); + + /* + * ATM there's nothing Xen can do if the console/store pfn changes, + * because Xen won't have a page_info struct for it. + */ + BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_PFN, + &store_pfn)); + BUG_ON(old_store_pfn != store_pfn); + if ( !pv_console ) + { + BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_PFN, + &console_pfn)); + BUG_ON(old_console_pfn != console_pfn); + } + + /* Update domain id. */ + d->domain_id = get_dom0_domid(); + + /* Clean the iomem range. */ + BUG_ON(iomem_deny_access(d, 0, ~0UL)); + + /* Clean grant frames. */ + xfree(grant_frames); + grant_frames = NULL; + nr_grant_list = 0; + + /* Clean event channels. */ + for ( i = 0; i < EVTCHN_2L_NR_CHANNELS; i++ ) + { + if ( !port_is_valid(d, i) ) + continue; + + if ( evtchn_handled(d, i) ) + evtchn_close(d, i, false); + else + evtchn_free(d, evtchn_from_port(d, i)); + } + + /* Reserve store/console event channel. */ + BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_STORE_EVTCHN, + &store_evtchn)); + BUG_ON(evtchn_allocate_port(d, store_evtchn)); + evtchn_reserve(d, store_evtchn); + BUG_ON(xen_hypercall_hvm_get_param(HVM_PARAM_CONSOLE_EVTCHN, + &console_evtchn)); + BUG_ON(evtchn_allocate_port(d, console_evtchn)); + evtchn_reserve(d, console_evtchn); + + /* Clean watchdogs. */ + watchdog_domain_destroy(d); + watchdog_domain_init(d); + + /* Clean the PIRQ EOI page. */ + if ( d->arch.pirq_eoi_map != NULL ) + { + unmap_domain_page_global(d->arch.pirq_eoi_map); + put_page_and_type(mfn_to_page(d->arch.pirq_eoi_map_mfn)); + d->arch.pirq_eoi_map = NULL; + d->arch.pirq_eoi_map_mfn = 0; + d->arch.auto_unmask = 0; + } + + /* + * NB: there's no need to fixup the p2m, since the mfns assigned + * to the PV guest have not changed at all. Just re-write the + * start_info fields with the appropriate value. + */ + write_start_info(d); + + for_each_vcpu(d, v) + { + /* Unmap guest vcpu_info pages. */ + unmap_vcpu_info(v); + + /* Reset the periodic timer to the default value. */ + v->periodic_period = MILLISECS(10); + /* Stop the singleshot timer. */ + stop_timer(&v->singleshot_timer); + + if ( test_bit(_VPF_down, &v->pause_flags) ) + BUG_ON(vcpu_reset(v)); + + if ( v != current ) + vcpu_unpause_by_systemcontroller(v); + else + vcpu_force_reschedule(v); + } + } + else + /* Forward to L0. */ + rc = xen_hypercall_shutdown(reason); + + return rc; } long pv_shim_event_channel_op(int cmd, XEN_GUEST_HANDLE_PARAM(void) arg) diff --git a/xen/common/domain.c b/xen/common/domain.c index d653a0b0bb..bc2ceb2d36 100644 --- a/xen/common/domain.c +++ b/xen/common/domain.c @@ -701,15 +701,12 @@ void __domain_crash_synchronous(void) } -void domain_shutdown(struct domain *d, u8 reason) +int domain_shutdown(struct domain *d, u8 reason) { struct vcpu *v; if ( pv_shim ) - { - pv_shim_shutdown(reason); - return; - } + return pv_shim_shutdown(reason); spin_lock(&d->shutdown_lock); @@ -723,7 +720,7 @@ void domain_shutdown(struct domain *d, u8 reason) if ( d->is_shutting_down ) { spin_unlock(&d->shutdown_lock); - return; + return 0; } d->is_shutting_down = 1; @@ -745,6 +742,8 @@ void domain_shutdown(struct domain *d, u8 reason) __domain_finalise_shutdown(d); spin_unlock(&d->shutdown_lock); + + return 0; } void domain_resume(struct domain *d) diff --git a/xen/common/schedule.c b/xen/common/schedule.c index 88279213e8..b7884263f2 100644 --- a/xen/common/schedule.c +++ b/xen/common/schedule.c @@ -1149,11 +1149,10 @@ ret_t do_sched_op(int cmd, XEN_GUEST_HANDLE_PARAM(void) arg) if ( copy_from_guest(&sched_shutdown, arg, 1) ) break; - ret = 0; TRACE_3D(TRC_SCHED_SHUTDOWN, current->domain->domain_id, current->vcpu_id, sched_shutdown.reason); - domain_shutdown(current->domain, (u8)sched_shutdown.reason); + ret = domain_shutdown(current->domain, (u8)sched_shutdown.reason); break; } diff --git a/xen/drivers/char/xen_pv_console.c b/xen/drivers/char/xen_pv_console.c index fb5a7893be..d2a28478ac 100644 --- a/xen/drivers/char/xen_pv_console.c +++ b/xen/drivers/char/xen_pv_console.c @@ -37,7 +37,7 @@ static DEFINE_SPINLOCK(tx_lock); bool pv_console; -void __init pv_console_init(void) +void pv_console_init(void) { struct evtchn_unmask unmask; long r; diff --git a/xen/include/asm-x86/guest/xen.h b/xen/include/asm-x86/guest/xen.h index 94f781c30f..c244569e83 100644 --- a/xen/include/asm-x86/guest/xen.h +++ b/xen/include/asm-x86/guest/xen.h @@ -40,6 +40,7 @@ void hypervisor_fixup_e820(struct e820map *e820); void hypervisor_init_memory(void); const unsigned long *hypervisor_reserved_pages(unsigned int *size); uint32_t hypervisor_cpuid_base(void); +void hypervisor_resume(void); #else @@ -91,6 +92,10 @@ static inline uint32_t hypervisor_cpuid_base(void) ASSERT_UNREACHABLE(); return 0; }; +static inline void hypervisor_resume(void) +{ + ASSERT_UNREACHABLE(); +}; #endif /* CONFIG_XEN_GUEST */ #endif /* __X86_GUEST_XEN_H__ */ diff --git a/xen/include/asm-x86/pv/shim.h b/xen/include/asm-x86/pv/shim.h index 47bc4267af..0207348a85 100644 --- a/xen/include/asm-x86/pv/shim.h +++ b/xen/include/asm-x86/pv/shim.h @@ -35,7 +35,7 @@ void pv_shim_setup_dom(struct domain *d, l4_pgentry_t *l4start, unsigned long va_start, unsigned long store_va, unsigned long console_va, unsigned long vphysmap, start_info_t *si); -void pv_shim_shutdown(uint8_t reason); +int pv_shim_shutdown(uint8_t reason); long pv_shim_event_channel_op(int cmd, XEN_GUEST_HANDLE_PARAM(void) arg); void pv_shim_inject_evtchn(unsigned int port); long pv_shim_grant_table_op(unsigned int cmd, XEN_GUEST_HANDLE_PARAM(void) uop, @@ -53,9 +53,10 @@ static inline void pv_shim_setup_dom(struct domain *d, l4_pgentry_t *l4start, { ASSERT_UNREACHABLE(); } -static inline void pv_shim_shutdown(uint8_t reason) +static inline int pv_shim_shutdown(uint8_t reason) { ASSERT_UNREACHABLE(); + return 0; } static inline long pv_shim_event_channel_op(int cmd, XEN_GUEST_HANDLE_PARAM(void) arg) diff --git a/xen/include/xen/sched.h b/xen/include/xen/sched.h index ac65d0c0df..70db377eae 100644 --- a/xen/include/xen/sched.h +++ b/xen/include/xen/sched.h @@ -605,7 +605,7 @@ static inline struct domain *rcu_lock_current_domain(void) struct domain *get_domain_by_id(domid_t dom); void domain_destroy(struct domain *d); int domain_kill(struct domain *d); -void domain_shutdown(struct domain *d, u8 reason); +int domain_shutdown(struct domain *d, u8 reason); void domain_resume(struct domain *d); void domain_pause_for_debugger(void); -- 2.11.0 _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxxxxxxxxx https://lists.xenproject.org/mailman/listinfo/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |