[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] HVM restore broken?



berger,

you run 32 linux on 64b HV, right?

i can reproduce this bug in such combination on latest change set, but have no 
console MFN & set maxmem err msg.

i doubt some memory restore issue here.

thanks,

On Sat, Jan 27, 2007 at 11:57:09AM -0500, Stefan Berger wrote:
>    I tried suspend/resume with a Linux guest with 1 processor.
> 
>    I get the following errors in xend's log file upon resume.
> 
>    [...]
>    [2007-01-27 11:44:06 3629] DEBUG (XendDomainInfo:775) Storing domain
>    details: {'console/port': '3', 'name': 'TCG-TEST', 'console/limit':
>    '1048576', 'vm': '/vm/b0aaf4ff-ede8-994d-551c-23da391edbf2', 'domid': '4',
>    'cpu/0/availability': 'online', 'memory/target': '196608', 'store/port':
>    '2'}
>    [2007-01-27 11:44:06 3629] INFO (XendCheckpoint:207) restore hvm domain 4,
>    mem=192, apic=1, pae=1
>    [2007-01-27 11:44:06 3629] DEBUG (XendCheckpoint:226) restore:shadow=0x3,
>    _static_max=0xd0, _static_min=0xc0, nr_pfns=0xc000.
>    [2007-01-27 11:44:06 3629] DEBUG (balloon:127) Balloon: 258044 KiB free;
>    need 199680; done.
>    [2007-01-27 11:44:06 3629] DEBUG (XendCheckpoint:236) [xc_restore]:
>    /usr/lib/xen/bin/xc_restore 22 4 49152 2 3 192 1 1
>    [2007-01-27 11:44:06 3629] INFO (XendCheckpoint:340) xc_hvm_restore:dom=4,
>    nr_pfns=0xc000, store_evtchn=2, *store_mfn=192, console_evtchn=3,
>    *console_mfn=-1208080352, pae=1, apic=1.
>    [2007-01-27 11:44:06 3629] INFO (XendCheckpoint:340) xc_hvm_restore start:
>    max_pfn = c000, max_mfn = 3f700, hvirt_start=f5800000, pt_levels=3
>    [2007-01-27 11:44:07 3629] INFO (XendCheckpoint:340) hvm restore:calculate
>    new store_mfn=0xbffe,v_end=0xc000000..
>    [2007-01-27 11:44:07 3629] INFO (XendCheckpoint:340) hvm restore:get
>    nr_vcpus=1.
>    [2007-01-27 11:44:07 3629] INFO (XendCheckpoint:340) Restore exit with
>    rc=0
>    [2007-01-27 11:44:07 3629] DEBUG (XendCheckpoint:311) store-mfn 49150
>    [2007-01-27 11:44:07 3629] DEBUG (XendCheckpoint:311) console-mfn
>    -1208080352
>    [2007-01-27 11:44:07 3629] DEBUG (XendDomainInfo:1558)
>    XendDomainInfo.destroy: domid=4
>    [2007-01-27 11:44:07 3629] DEBUG (XendDomainInfo:1566)
>    XendDomainInfo.destroyDomain(4)
>    [2007-01-27 11:44:07 3629] ERROR (XendDomain:1030) Restore failed
>    Traceback (most recent call last):
>     File "//usr/lib/python/xen/xend/XendDomain.py", line 1025, in
>    domain_restore_fd
>       return XendCheckpoint.restore(self, fd, paused=paused)
>     File "//usr/lib/python/xen/xend/XendCheckpoint.py", line 243, in restore
>       raise XendError('Could not read store/console MFN')
>    XendError: Could not read store/console MFN
> 
>    In the qemu dm's log file I see this here  --- notice the 'error 22':
> 
>    domid: 3
>    qemu: the number of cpus is 1
>    qemu_map_cache_init nr_buckets = c00
>    shared page at pfn:bfff
>    buffered io page at pfn:bffd
>    xs_read(): vncpasswd get error.
>    /vm/b0aaf4ff-ede8-994d-551c-23da391edbf2/vncpasswd.
>    I/O request not ready: 0, ptr: 0, port: 0, data: 0, count: 0, size: 0
>    suspend sig handler called with requested=0!
>    device model received suspend signal!
>    set maxmem returned error 22
>    cirrus_stop_acc:unset_vram_mapping.
> 
>    'xm dmesg' shows this here (trying multiple times to resume):
> 
>    (XEN) HVM S/R Loading "xen_hvm_i8259" instance 0x20
>    (XEN) HVM S/R Loading "xen_hvm_i8259" instance 0xa0
>    (XEN) HVM S/R Loading "xen_hvm_ioapic" instance 0x0
>    (XEN) HVM S/R Loading "xen_hvm_cpu" instance 0x0
>    (XEN) HVM S/R Loading "xen_hvm_lapic" instance 0x0
>    (XEN) HVM S/R Loading "xen_hvm_i8254" instance 0x40
>    (XEN) HVM S/R Loading "xen_hvm_shpage" instance 0x10
>    (XEN) HVM S/R Loading "xen_hvm_i8259" instance 0x20
>    (XEN) HVM S/R Loading "xen_hvm_i8259" instance 0xa0
>    (XEN) HVM S/R Loading "xen_hvm_ioapic" instance 0x0
>    (XEN) HVM S/R Loading "xen_hvm_cpu" instance 0x0
>    (XEN) HVM S/R Loading "xen_hvm_lapic" instance 0x0
>    (XEN) HVM S/R Loading "xen_hvm_i8254" instance 0x40
>    (XEN) HVM S/R Loading "xen_hvm_shpage" instance 0x10
> 
>       Stefan
> 
>    xen-devel-bounces@xxxxxxxxxxxxxxxxxxx wrote on 01/27/2007 08:53:07 AM:
> 
>    > Mats,
>    >
>    > at least 32b up windows on 32b HV save/restore works here.
>    >
>    > what's your configuration? 32 or 64 windows/HV? up or smp?
>    >
>    > BTW, can you have a try with linux guest?
>    >
>    >
>    > On Fri, Jan 26, 2007 at 06:41:13PM +0100, Petersson, Mats wrote:
>    > > I got latest (13601) yesterday evening. This doesn't seem to work to
>    do
>    > > Restore (at least of the Windows test-image that I've been using for
>    > > testing previously).
>    > >
>    > > The VM restores reasonably OK, but it jumps to an invalid address
>    > > shortly after restoring, giving a D1 blue-screen error
>    > > (DRIVER_IRQL_LESS_OR_EQUAL), which turns out to be "page-fault in
>    > > driver" after I looked at the memory dump in windbg. (The address it
>    > > jumps to is consistenly a0000ca5, if that's of any meaning to anyone).
>    > >
>    > > I've compared my svm.c that I had previously and the current one that
>    I
>    > > got from mercurial, and they are identical.
>    > >
>    > > I went back to my 13568 build of the hypervisor, and it works there...
>    > > There's no obvious changes in between...
>    > >
>    > > Has anyone else tried this, does anyone have an idea of what's going
>    > > wrong?
>    > >
>    > > --
>    > > Mats
>    > >
>    > >
>    > >
>    > > _______________________________________________
>    > > Xen-devel mailing list
>    > > Xen-devel@xxxxxxxxxxxxxxxxxxx
>    > > http://lists.xensource.com/xen-devel
>    > >
>    >
>    > --
>    > best rgds,
>    > edwin
>    >
>    > _______________________________________________
>    > Xen-devel mailing list
>    > Xen-devel@xxxxxxxxxxxxxxxxxxx
>    > http://lists.xensource.com/xen-devel

-- 
best rgds,
edwin

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.