[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

RE: [Xen-devel] Biweekly VMX status report. Xen: #21438 & Xen0: #a3e7c7...



BTW, I get following failure after loop cpu o*l for about 95 times. 

(XEN) Xen call trace:
(XEN)    [<ffff82c48014b3e9>] clear_page_sse2+0x9/0x30
(XEN)    [<ffff82c4801b9922>] vmx_cpu_up_prepare+0x43/0x88
(XEN)    [<ffff82c4801a13fa>] cpu_callback+0x4a/0x94
(XEN)    [<ffff82c480112d95>] notifier_call_chain+0x68/0x84
(XEN)    [<ffff82c480100e5b>] cpu_up+0x7b/0x12f
(XEN)    [<ffff82c480173b7d>] arch_do_sysctl+0x770/0x833
(XEN)    [<ffff82c480121672>] do_sysctl+0x992/0x9ec
(XEN)    [<ffff82c4801fa3cf>] syscall_enter+0xef/0x149
(XEN)
(XEN) Pagetable walk from ffff83022fe1d000:
(XEN)  L4[0x106] = 00000000cfc8d027 5555555555555555
(XEN)  L3[0x008] = 00000000cfef9063 5555555555555555
(XEN)  L2[0x17f] = 000000022ff2a063 5555555555555555
(XEN)  L1[0x01d] = 000000022fe1d262 5555555555555555

I really can't imagine how this can happen considering the vmx_alloc_vmcs() is 
so straight-forward. My test machine is really magic.

Another fault as following:

(XEN) Xen call trace:
(XEN)    [<ffff82c480173459>] memcpy+0x11/0x1e
(XEN)    [<ffff82c4801722bf>] cpu_smpboot_callback+0x207/0x235
(XEN)    [<ffff82c480112d95>] notifier_call_chain+0x68/0x84
(XEN)    [<ffff82c480100e5b>] cpu_up+0x7b/0x12f
(XEN)    [<ffff82c480173c1d>] arch_do_sysctl+0x770/0x833
(XEN)    [<ffff82c480121712>] do_sysctl+0x992/0x9ec
(XEN)    [<ffff82c4801fa46f>] syscall_enter+0xef/0x149
(XEN)
(XEN) Pagetable walk from ffff830228ce5000:
(XEN)  L4[0x106] = 00000000cfc8d027 5555555555555555
(XEN)  L3[0x008] = 00000000cfef9063 5555555555555555
(XEN)  L2[0x146] = 000000022fea3063 5555555555555555
(XEN)  L1[0x0e5] = 0000000228ce5262 000000000001fd49
(XEN)
(XEN) ****************************************
(XEN) Panic on CPU 1:
(XEN) FATAL PAGE FAULT
(XEN) [error_code=0002]
(XEN) Faulting linear address: ffff830228ce5000
(XEN) ****************************************
(XEN)

--jyh

>-----Original Message-----
>From: Keir Fraser [mailto:keir.fraser@xxxxxxxxxxxxx]
>Sent: Wednesday, June 02, 2010 4:01 PM
>To: Jiang, Yunhong; Xu, Jiajun; xen-devel@xxxxxxxxxxxxxxxxxxx
>Subject: Re: [Xen-devel] Biweekly VMX status report. Xen: #21438 & Xen0:
>#a3e7c7...
>
>On 02/06/2010 08:28, "Jiang, Yunhong" <yunhong.jiang@xxxxxxxxx> wrote:
>
>>> Which xen-unstable changeset are you testing? All timers should be
>>> automatically migrated off a dead CPU and onto CPU0 by changeset 21424. Is
>>> that not working okay for you?
>>
>> We are testing on 21492.
>>
>> After more investigation, the root cause is the periodic_timer is stopped
>> before take_cpu_down (in schedule()), so that it is not covred by 21424.
>> When v->periodic_period ==0, next vcpu's p_timer is not updated by the
>> schedule(), thus, later in next schedule round, it will cause trouble for
>> stop_timer().
>>
>> With following small patch, it works, but I'm not sure if this is good
>> solution.
>
>I forgot about inactive timers in c/s 21424. Hm, I will fix this in the
>timer subsystem and get back to you.
>
> -- Keir
>


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.