WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-devel

Re: [Xen-devel] [xen-unstable test] 1959: FAIL [and 1 more messages]

To: "Xu, Dongxiao" <dongxiao.xu@xxxxxxxxx>
Subject: Re: [Xen-devel] [xen-unstable test] 1959: FAIL [and 1 more messages]
From: Jeremy Fitzhardinge <jeremy@xxxxxxxx>
Date: Tue, 10 Aug 2010 10:38:49 -0700
Cc: Ian Campbell <Ian.Campbell@xxxxxxxxxxxxx>, "Xen-devel@xxxxxxxxxxxxxxxxxxx" <Xen-devel@xxxxxxxxxxxxxxxxxxx>, Ian Jackson <Ian.Jackson@xxxxxxxxxxxxx>
Delivery-date: Tue, 10 Aug 2010 10:39:34 -0700
Envelope-to: www-data@xxxxxxxxxxxxxxxxxxx
In-reply-to: <1280481518.24292.1882.camel@xxxxxxxxxxxxxxxxxxxxxx>
List-help: <mailto:xen-devel-request@lists.xensource.com?subject=help>
List-id: Xen developer discussion <xen-devel.lists.xensource.com>
List-post: <mailto:xen-devel@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=unsubscribe>
References: <E1OeOpE-0008C1-Jl@xxxxxxxxxxxxxxxxxxx> <E1OcaPH-0005MQ-Ay@xxxxxxxxxxxxxxxxxxx> <4C4B1B8A.6070403@xxxxxxxx> <19533.23953.645991.205290@xxxxxxxxxxxxxxxxxxxxxxxx> <4C4DC4AF.2090901@xxxxxxxx> <19535.145.995122.44883@xxxxxxxxxxxxxxxxxxxxxxxx> <4C4F063C.9030301@xxxxxxxx> <19535.2235.492925.965460@xxxxxxxxxxxxxxxxxxxxxxxx> <4C4F0B18.5080800@xxxxxxxx> <19535.3062.112242.870153@xxxxxxxxxxxxxxxxxxxxxxxx> <4C4F0FDB.2010607@xxxxxxxx> <19537.40568.143464.400709@xxxxxxxxxxxxxxxxxxxxxxxx> <4C51A084.5010903@xxxxxxxx> <19537.44367.380875.758088@xxxxxxxxxxxxxxxxxxxxxxxx> <4C51AE8C.4000206@xxxxxxxx> <D5AB6E638E5A3E4B8F4406B113A5A19A28CCF46E@xxxxxxxxxxxxxxxxxxxxxxxxxxxx> <1280481518.24292.1882.camel@xxxxxxxxxxxxxxxxxxxxxx>
Sender: xen-devel-bounces@xxxxxxxxxxxxxxxxxxx
User-agent: Mozilla/5.0 (X11; U; Linux x86_64; en-US; rv:1.9.2.7) Gecko/20100720 Fedora/3.1.1-1.fc13 Lightning/1.0b2pre Thunderbird/3.1.1
 On 07/30/2010 02:18 AM, Ian Campbell wrote:
diff --git a/drivers/net/xen-netfront.c b/drivers/net/xen-netfront.c
index 328fe40..394bbc8 100644
--- a/drivers/net/xen-netfront.c
+++ b/drivers/net/xen-netfront.c
@@ -1340,6 +1340,10 @@ static enum hrtimer_restart smart_poll_function(struct 
hrtimer *timer)
        np = netdev_priv(dev);

        spin_lock_irqsave(&np->tx_lock, flags);
+
+       if (!np->rx.sring)
+               goto end;
+
Isn't there a period until the end of xennet_disconnect_backend is
reached on resume where rx.sring will still point to the old shared
ring? If so is it safe to drop through in that case?

Would it be safer to add an explicit suspend handler which stopped the
timer?

Dongxiao, do you have a comment/updated patch for this? I'm going to revert the smartpoll stuff in the meantime, because its causing tests to fail.

Thanks,
    J

Ian.

        np->smart_poll.counter++;xennet_disconnect_backend

        if (likely(netif_carrier_ok(dev))) {
--
1.6.3



Jeremy Fitzhardinge wrote:
   On 07/29/2010 09:33 AM, Ian Jackson wrote:
Jeremy Fitzhardinge writes ("Re: [Xen-devel] [xen-unstable test]
1959: FAIL [and 1 more messages]"):
On 07/29/2010 08:30 AM, Ian Jackson wrote:
Is save/restore supposed to work in pvops ?  (Using your kernel for
both dom0 and domU.)  That would seem to be the next thing to pick
off the list ...
Yes.  IanC has been tracking down a bug where it fails after a few
thousand iterations (now fixed?), but aside from that its
apparently OK.
Well, I was able to reproduce the failure that the automatic test was
getting.  There's a problem with it not getting a copy of the console
output for some reason but I was able to get this oops from the
guest.

Just after the restore I was able to connect to the PV console and it
echoed a few of my CRs (before saving I had logged in on the
console),
and then it produced the oops.  Now it's apparently completely
wedged.

Ah, it looks like the netfront smartpoll stuff isn't coping with
save/restore.

      J

Ian.



[   63.681260] BUG: unable to handle kernel NULL pointer dereference
at 00000010 [   63.681293] IP: [<c1300381>]
smart_poll_function+0xbb/0xf2 [   63.681320] *pdpt =
000000001fee5027 *pde = 0000000000000000 [   63.681344] Oops: 0000
[#1] SMP [   63.681362] last sysfs file: /sys/kernel/uevent_seqnum
[   63.681376] Modules linked in: [last unloaded: scsi_wait_scan] [
63.681398] [   63.681410] Pid: 5, comm: events/0 Not tainted
(2.6.32.16 #1) [   63.681424] EIP: 0061:[<c1300381>] EFLAGS:
00010002 CPU: 0 [   63.681438] EIP is at
smart_poll_function+0xbb/0xf2 [   63.681451] EAX: 00000000 EBX:
dfea8320 ECX: 00000001 EDX: 00000062 [   63.681465] ESI: 00000064
EDI: 00000000 EBP: df849cfc ESP: df849cdc [   63.681479]  DS: 007b
ES: 007b FS: 00d8 GS: 0000 SS: 0069 [   63.681493] Process events/0
(pid: 5, ti=df848000 task=df839480 task.ti=df848000) [   63.681508]
Stack: [   63.681516]  dfeac828 00000002 dfeac828 dfea8368 dfea0068
0000002c dfeac828 c13002c6 [   63.681555]<0>   df849d1c c1079ab8
df849d48 c2386358 c2386328 0000002c 00000000 c2386328 [
63.681598]<0>   df849d5c c1079ce8 000028cd 88bbcad9 d3b2cae7 0000000e
d3b2cae7 0000000e [   63.681646] Call Trace: [   63.681662]
[<c13002c6>] ? smart_poll_function+0x0/0xf2 [   63.681683]
[<c1079ab8>] ? __run_hrtimer+0xa9/0xf6 [   63.681701]  [<c1079ce8>]
? hrtimer_interrupt+0xcd/0x1c8 [   63.681719]  [<c102cef0>] ?
xen_timer_interrupt+0x2b/0x224 [   63.681737]  [<c102cb3f>] ?
xen_force_evtchn_callback+0xf/0x14 [   63.681755]  [<c102d2ac>] ?
check_events+0x8/0xc [   63.681776]  [<c102d2a3>] ?
xen_restore_fl_direct_end+0x0/0x1 [   63.681795]  [<c14d1471>] ?
_spin_unlock_irqrestore+0x2f/0x31 [   63.681814]  [<c105eb9e>] ?
try_to_wake_up+0x2fa/0x304 [   63.681832]  [<c102cb3f>] ?
xen_force_evtchn_callback+0xf/0x14 [   63.681850]  [<c10a0cc4>] ?
handle_IRQ_event+0x5f/0x122 [   63.681867]  [<c10a233f>] ?
handle_level_irq+0x58/0xa9 [   63.681886]  [<c121983d>] ?
__xen_evtchn_do_upcall+0xab/0x131 [   63.681904]  [<c1219c71>] ?
xen_evtchn_do_upcall+0x20/0x30 [   63.682179]  [<c102ffe7>] ?
xen_do_upcall+0x7/0xc [   63.682179]  [<c102007b>] ?
apic_reg_write+0xa5/0x52f [   63.682179]  [<c1002227>] ?
hypercall_page+0x227/0x1005 [   63.682179]  [<c102cb3f>] ?
xen_force_evtchn_callback+0xf/0x14 [   63.682179]  [<c102d2ac>] ?
check_events+0x8/0xc [   63.682179]  [<c102d26b>] ?
xen_irq_enable_direct_end+0x0/0x1 [   63.682179]  [<c105cd0d>] ?
finish_task_switch+0x4f/0xa6 [   63.682179]  [<c14d0186>] ?
schedule+0x7dd/0x861 [   63.682179]  [<c106dd9d>] ?
__mod_timer+0x135/0x140 [   63.682179]  [<c102d2ac>] ?
check_events+0x8/0xc [   63.682179]  [<c102d2a3>] ?
xen_restore_fl_direct_end+0x0/0x1 [   63.682179]  [<c14d1471>] ?
_spin_unlock_irqrestore+0x2f/0x31 [   63.682179]  [<c107750b>] ?
prepare_to_wait+0x43/0x48 [   63.682179]  [<c10742a2>] ?
worker_thread+0x94/0x1d2 [   63.682179]  [<c10ccd2d>] ?
vmstat_update+0x0/0x2f [   63.682179]  [<c1077357>] ?
autoremove_wake_function+0x0/0x33 [   63.682179]  [<c107420e>] ?
worker_thread+0x0/0x1d2 [   63.682179]  [<c1077120>] ?
kthread+0x61/0x66 [   63.682179]  [<c10770bf>] ? kthread+0x0/0x66 [
63.682179]  [<c102ff97>] ? kernel_thread_helper+0x7/0x10 [
63.682179] Code: c6 89 d0 31 d2 f7 f6 85 d2 75 1a 85 c9 75 0c 8b 83
90 08 00 00 c6 40 10 00 eb 32 c7 83 54 45 00 00 00 00 00 00 8b 83 90
08 00 00<80>   78 10 00 74 1c 8b 4d e8 b8 00 ca 9a 3b 31 d2 f7 71 44
31 c9 [   63.682179] EIP: [<c1300381>] smart_poll_function+0xbb/0xf2
SS:ESP 0069:df849cdc [   63.682179] CR2: 0000000000000010 [
63.682179] ---[ end trace 760037e75e5675c8 ]--- [   63.682179]
Kernel panic - not syncing: Fatal exception in interrupt [
63.682179] Pid: 5, comm: events/0 Tainted: G      D    2.6.32.16 #1
[   63.682179] Call Trace: [   63.682179]  [<c14cf7b1>] ?
printk+0xf/0x11 [   63.682179]  [<c14cf6ee>] panic+0x39/0xed [
63.682179]  [<c14d2085>] oops_end+0xa1/0xb0 [   63.682179]
[<c104b426>] no_context+0x137/0x141 [   63.682179]  [<c104b56f>]
__bad_area_nosemaphore+0x13f/0x147 [   63.682179]  [<c104b584>]
bad_area_nosemaphore+0xd/0x10 [   63.682179]  [<c14d3275>]
do_page_fault+0x1c6/0x32b [   63.682179]  [<c14d30af>] ?
do_page_fault+0x0/0x32b [   63.682179]  [<c14d16c6>]
error_code+0x66/0x6c [   63.682179]  [<c14d30af>] ?
do_page_fault+0x0/0x32b [   63.682179]  [<c1300381>] ?
smart_poll_function+0xbb/0xf2 [   63.682179]  [<c13002c6>] ?
smart_poll_function+0x0/0xf2 [   63.682179]  [<c1079ab8>]
__run_hrtimer+0xa9/0xf6 [   63.682179]  [<c1079ce8>]
hrtimer_interrupt+0xcd/0x1c8 [   63.682179]  [<c102cef0>]
xen_timer_interrupt+0x2b/0x224 [   63.682179]  [<c102cb3f>] ?
xen_force_evtchn_callback+0xf/0x14 [   63.682179]  [<c102d2ac>] ?
check_events+0x8/0xc [   63.682179]  [<c102d2a3>] ?
xen_restore_fl_direct_end+0x0/0x1 [   63.682179]  [<c14d1471>] ?
_spin_unlock_irqrestore+0x2f/0x31 [   63.682179]  [<c105eb9e>] ?
try_to_wake_up+0x2fa/0x304 [   63.682179]  [<c102cb3f>] ?
xen_force_evtchn_callback+0xf/0x14 [   63.682179]  [<c10a0cc4>]
handle_IRQ_event+0x5f/0x122 [   63.682179]  [<c10a233f>]
handle_level_irq+0x58/0xa9 [   63.682179]  [<c121983d>]
__xen_evtchn_do_upcall+0xab/0x131 [   63.682179]  [<c1219c71>]
xen_evtchn_do_upcall+0x20/0x30 [   63.682179]  [<c102ffe7>]
xen_do_upcall+0x7/0xc [   63.682179]  [<c102007b>] ?
apic_reg_write+0xa5/0x52f [   63.682179]  [<c1002227>] ?
hypercall_page+0x227/0x1005 [   63.682179]  [<c102cb3f>] ?
xen_force_evtchn_callback+0xf/0x14 [   63.682179]  [<c102d2ac>]
check_events+0x8/0xc [   63.682179]  [<c102d26b>] ?
xen_irq_enable_direct_end+0x0/0x1 [   63.682179]  [<c105cd0d>] ?
finish_task_switch+0x4f/0xa6 [   63.682179]  [<c14d0186>]
schedule+0x7dd/0x861 [   63.682179]  [<c106dd9d>] ?
__mod_timer+0x135/0x140 [   63.682179]  [<c102d2ac>] ?
check_events+0x8/0xc [   63.682179]  [<c102d2a3>] ?
xen_restore_fl_direct_end+0x0/0x1 [   63.682179]  [<c14d1471>] ?
_spin_unlock_irqrestore+0x2f/0x31 [   63.682179]  [<c107750b>] ?
prepare_to_wait+0x43/0x48 [   63.682179]  [<c10742a2>]
worker_thread+0x94/0x1d2 [   63.682179]  [<c10ccd2d>] ?
vmstat_update+0x0/0x2f [   63.682179]  [<c1077357>] ?
autoremove_wake_function+0x0/0x33 [   63.682179]  [<c107420e>] ?
worker_thread+0x0/0x1d2 [   63.682179]  [<c1077120>]
kthread+0x61/0x66 [   63.682179]  [<c10770bf>] ? kthread+0x0/0x66 [
63.682179]  [<c102ff97>] kernel_thread_helper+0x7/0x10



_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel