[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-devel] Test results for save/restore with upstream pv_ops domU kernels



Hello,

I just tried some save/restore tests with Fedora 12 Linux 2.6.31.12 kernels.
The exact Fedora kernel versions are: 2.6.31.12-174.2.3.fc12.i686.PAE and 
2.6.31.12-174.2.3.fc12.x86_64.

Dom0 for these tests was CentOS 5.4 (Xen 3.1.2).

- F12 32bit 1vcpu PV guest: 
        save+restore OK, BUG() in guest dmesg after restore [1]

- F12 64bit 1vcpu PV guest:
        save+restore OK, BUG() in guest dmesg after restore [2]

- F12 32bit 2vcpu PV guest:
        save doesn't work, guest stays as "migrating-f12test32" in "xm list" 
forever and has to be "xm destroy"ed.

- F12 64bit 2vcpu PV guest:
        save doesn't work, guest stays as "migrating-f12test64" in "xm list" 
forever and has to be "xm destroy"ed.


What's the best way to debug failing "xm save" ? There was no errors in "xm 
log", or in "xm dmesg".
Also the guest "xm console" doesn't show anything before it dies.

Is it possible some of the save/restore related patches didn't make it to 
2.6.31.x stable kernels? 


BUG()s after "xm restore":

[1] F12 32b 1vcpu:

suspending xenstore...
BUG: sleeping function called from invalid context at kernel/mutex.c:94
in_atomic(): 0, irqs_disabled(): 1, pid: 1136, name: kstop/0
Pid: 1136, comm: kstop/0 Tainted: G        W  2.6.31.12-174.2.3.fc12.i686.PAE #1
Call Trace:
 [<c042f3fa>] __might_sleep+0xc6/0xcb
 [<c077a276>] mutex_lock+0x1d/0x3c
 [<c06404a3>] dpm_resume_noirq+0x17/0xe1
 [<c060959d>] xen_suspend+0xc4/0xcf
 [<c046f96c>] stop_cpu+0x75/0xb4
 [<c044d440>] worker_thread+0x13c/0x1bc
 [<c046f8f7>] ? stop_cpu+0x0/0xb4
 [<c0450db9>] ? autoremove_wake_function+0x0/0x34
 [<c044d304>] ? worker_thread+0x0/0x1bc
 [<c0450b0f>] kthread+0x70/0x75
 [<c0450a9f>] ? kthread+0x0/0x75
 [<c0409c07>] kernel_thread_helper+0x7/0x10
eth0: no IPv6 routers present


[2] F12 64b 1vcpu:

suspending xenstore...
trying to map vcpu_info 0 at ffffc9000000b020, mfn 1d9997, offset 32
cpu 0 using vcpu_info at ffffc9000000b020
BUG: sleeping function called from invalid context at kernel/mutex.c:94
in_atomic(): 0, irqs_disabled(): 1, pid: 1112, name: kstop/0
Pid: 1112, comm: kstop/0 Tainted: G        W  2.6.31.12-174.2.3.fc12.x86_64 #1
Call Trace:
 [<ffffffff8104026a>] __might_sleep+0xe6/0xe8
 [<ffffffff8141c844>] mutex_lock+0x22/0x4e
 [<ffffffff812b266a>] dpm_resume_noirq+0x21/0x11f
 [<ffffffff812752a9>] xen_suspend+0xca/0xd1
 [<ffffffff8108c33e>] stop_cpu+0x8c/0xd2
 [<ffffffff810635a0>] worker_thread+0x18a/0x224
 [<ffffffff81067b37>] ? autoremove_wake_function+0x0/0x39
 [<ffffffff8141d6e9>] ? _spin_unlock_irqrestore+0x19/0x1b
 [<ffffffff81063416>] ? worker_thread+0x0/0x224
 [<ffffffff810677b5>] kthread+0x91/0x99
 [<ffffffff81012daa>] child_rip+0xa/0x20
 [<ffffffff81011f97>] ? int_ret_from_sys_call+0x7/0x1b
 [<ffffffff8101271d>] ? retint_restore_args+0x5/0x6
 [<ffffffff81012da0>] ? child_rip+0x0/0x20


-- Pasi


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.