[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [xen-unstable bisection] complete test-amd64-amd64-xl-qemuu-ovmf-amd64



On Wed, May 25, 2016 at 11:10:21AM +0100, Wei Liu wrote:
> On Tue, May 24, 2016 at 06:50:23PM +0100, Wei Liu wrote:
> > On Sun, May 22, 2016 at 05:37:51AM +0000, osstest service owner wrote:
> > > branch xen-unstable
> > > xenbranch xen-unstable
> > > job test-amd64-amd64-xl-qemuu-ovmf-amd64
> > > testid guest-start/debianhvm.repeat
> > > 
> > > Tree: linux git://xenbits.xen.org/linux-pvops.git
> > > Tree: linuxfirmware git://xenbits.xen.org/osstest/linux-firmware.git
> > > Tree: qemu git://xenbits.xen.org/qemu-xen-traditional.git
> > > Tree: qemuu git://xenbits.xen.org/qemu-xen.git
> > > Tree: xen git://xenbits.xen.org/xen.git
> > > 
> > > *** Found and reproduced problem changeset ***
> > > 
> > >   Bug is in tree:  xen git://xenbits.xen.org/xen.git
> > >   Bug introduced:  1542efcea893df874b13b1ea78101e1ff6a55c41
> > >   Bug not present: c32381352cce9744e640bf239d2704dae4af4fc7
> > >   Last fail repro: http://logs.test-lab.xenproject.org/osstest/logs/94689/
> > > 
> > > 
> > >   commit 1542efcea893df874b13b1ea78101e1ff6a55c41
> > >   Author: Wei Liu <wei.liu2@xxxxxxxxxx>
> > >   Date:   Wed May 18 11:48:25 2016 +0100
> > >   
> > >       Config.mk: update ovmf changeset
> > >       
> > >       Signed-off-by: Wei Liu <wei.liu2@xxxxxxxxxx>
> > > 
> > 
> > I did some tests and analysis today.
> > 
> > * Manual tests
> > 
> > Seconds between starting a guest to receiving ping, test three times
> > 
> >   xl create guest.cfg ;\
> >   s=`date +%s`; date --date="@$s"; \
> >   while true;  do \
> >    ping -c 1  -q -W 1 172.16.147.190 2>&1 1>/dev/null;\
> >    if [ $? = 0 ]; then break; fi ;\
> >    done;\
> >   e=`date +%s`; date --date="@$e";\
> >   expr $e - $s
> > 
> >                           merlot0             tg06
> > old ovmf, 5000mb ram     98 99 96            33 32 31
> > old ovmf, 768mb  ram     97 100 100          31 31 32
> > new ovmf, 5000mb ram     158 158 157         25 26 25
> > new ovmf, 768mb  ram     151 156 160         26 25 25
> > 
> > Old ovmf refers to 52a99493 (currently in master)
> > New ovmf refers to b41ef325 (the fingered one)
> > 
> > Tg06 and merlot0 have the same changeset git:983aae0.
> > 
> > Note that the guest runs on tg06 has a different version of Debian, so it is
> > not really comparable to the guest on merlot0.  Also note that we can't
> > extrapolate from my manual test that osstest will or will not see timeout on
> > merlot0 because the technique to test that is not the same.
> > 
> > The conclusions are: we now know the results are consistent and the guest
> > memory size doesn't affect the time taken to start the guest.
> > 
> > * Osstest report
> > 
> > Pick the ovmf flight that tested the fingered changeset:
> > 
> > http://logs.test-lab.xenproject.org/osstest/logs/94519/test-amd64-amd64-xl-qemuu-ovmf-amd64/17.ts-repeat-test.log
> > 
> > 2016-05-18 05:40:26 Z guest debianhvm.guest.osstest 5a:36:0e:37:00:01 22 
> > link/ip/tcp: ok. (185s) 
> > 2016-05-18 05:44:13 Z guest debianhvm.guest.osstest 5a:36:0e:37:00:01 22 
> > link/ip/tcp: ok. (184s) 
> > 2016-05-18 05:47:55 Z guest debianhvm.guest.osstest 5a:36:0e:37:00:01 22 
> > link/ip/tcp: ok. (184s) 
> > ...
> > 
> > The time out for checking if a guest is up is 200 seconds so 180 seconds 
> > should
> > be fine.
> > 
> > The new ovmf failure reported by bisector is the controller timed out when
> > trying to check if the guest is up.
> > 
> >     
> > http://logs.test-lab.xenproject.org/osstest/logs/94689/test-amd64-amd64-xl-qemuu-ovmf-amd64/17.ts-repeat-test.log
> > 
> > The old ovmf passed on merlot0:
> > 
> >     
> > http://logs.test-lab.xenproject.org/osstest/logs/94680/test-amd64-amd64-xl-qemuu-ovmf-amd64/17.ts-repeat-test.log
> >     2016-05-22 02:49:57 Z guest debianhvm.guest.osstest 5a:36:0e:d8:00:01 
> > 22 link/ip/tcp: ok. (141s) 
> > 
> > The old ovmf passed on other machine:
> > 
> >     
> > http://logs.test-lab.xenproject.org/osstest/logs/94580/test-amd64-amd64-xl-qemuu-ovmf-amd64/17.ts-repeat-test.log
> > 
> >     2016-05-19 22:45:39 Z guest debianhvm.guest.osstest 5a:36:0e:74:00:3c 
> > 22 link/ip/tcp: ok. (122s) 
> > 
> > The two numbers suggest that merlot is slower than the other machine.
> > 
> > Pick one of the recent test report for OVMF:
> > 
> > http://logs.test-lab.xenproject.org/osstest/logs/94739/test-amd64-amd64-xl-qemuu-ovmf-amd64/17.ts-repeat-test.log
> > 
> > The same metric (guest creation to guest responding to network traffic) is a
> > lot shorter (on a non-merlot machine):
> > 
> > 2016-05-24 14:21:59 Z guest debianhvm.guest.osstest 5a:36:0e:13:00:02 22 
> > link/ip/tcp: ok. (49s) 
> > 2016-05-24 14:23:28 Z guest debianhvm.guest.osstest 5a:36:0e:13:00:02 22 
> > link/ip/tcp: ok. (49s) 
> > 2016-05-24 14:25:03 Z guest debianhvm.guest.osstest 5a:36:0e:13:00:02 22 
> > link/ip/tcp: ok. (48s) 
> > 2016-05-24 14:26:45 Z guest debianhvm.guest.osstest 5a:36:0e:13:00:02 22 
> > link/ip/tcp: ok. (48s) 
> > ...
> > 
> > It looks like it's getting better.
> > 
> > I don't have a conclusion on this issue because I can't eliminate all
> > variables.  I'm inclined to push another a newer ovmf changeset to see what
> > happens, because:
> > 
> > 1. merlot is slower than other machine, the time difference is about 20s.
> > 2. new ovmf on other machine already takes ~180s to come up (less than 20s 
> > to
> >    200s timeout).
> > 3. the time taken to come up seems to get shorter, though I didn't see why 
> > when
> >    I skimmed ovmf changelog.
> 
> I'm going to hold off the attempt because the latest ovmf flight was
> scheduled on merlot1 and failed.
> 
> http://logs.test-lab.xenproject.org/osstest/logs/94753/

In my experience, OVMF takes a lot longer to start on an AMD box
compared to an Intel one. I think the time is spent on
self-decompression, so very early.

-- 
Anthony PERARD

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.