[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

[Xen-devel] [xen-4.7-testing test] 99961: regressions - FAIL



flight 99961 xen-4.7-testing real [real]
http://logs.test-lab.xenproject.org/osstest/logs/99961/

Regressions :-(

Tests which did not succeed and are blocking,
including tests which could not be run:
 test-armhf-armhf-xl-arndale   7 host-ping-check-xen       fail REGR. vs. 99754

Regressions which are regarded as allowable (not blocking):
 test-amd64-i386-xl-qemuu-win7-amd64 16 guest-stop              fail like 99754
 test-amd64-i386-xl-qemut-win7-amd64 16 guest-stop              fail like 99754

Tests which did not succeed, but are not blocking:
 test-amd64-i386-rumpuserxen-i386  1 build-check(1)               blocked  n/a
 test-amd64-amd64-rumpuserxen-amd64  1 build-check(1)               blocked n/a
 test-amd64-amd64-xl-pvh-intel 11 guest-start                  fail  never pass
 build-i386-rumpuserxen        6 xen-build                    fail   never pass
 build-amd64-rumpuserxen       6 xen-build                    fail   never pass
 test-amd64-i386-libvirt      12 migrate-support-check        fail   never pass
 test-armhf-armhf-libvirt     14 guest-saverestore            fail   never pass
 test-armhf-armhf-libvirt     12 migrate-support-check        fail   never pass
 test-armhf-armhf-libvirt-xsm 12 migrate-support-check        fail   never pass
 test-armhf-armhf-libvirt-xsm 14 guest-saverestore            fail   never pass
 test-amd64-amd64-libvirt     12 migrate-support-check        fail   never pass
 test-amd64-i386-libvirt-xsm  12 migrate-support-check        fail   never pass
 test-amd64-amd64-libvirt-xsm 12 migrate-support-check        fail   never pass
 test-amd64-amd64-qemuu-nested-amd 16 debian-hvm-install/l1/l2  fail never pass
 test-amd64-amd64-libvirt-qemuu-debianhvm-amd64-xsm 10 migrate-support-check 
fail never pass
 test-amd64-i386-libvirt-qemuu-debianhvm-amd64-xsm 10 migrate-support-check 
fail never pass
 test-amd64-amd64-xl-pvh-amd  11 guest-start                  fail   never pass
 test-armhf-armhf-xl-multivcpu 13 saverestore-support-check    fail  never pass
 test-armhf-armhf-xl-multivcpu 12 migrate-support-check        fail  never pass
 test-armhf-armhf-xl          12 migrate-support-check        fail   never pass
 test-armhf-armhf-xl          13 saverestore-support-check    fail   never pass
 test-armhf-armhf-xl-xsm      13 saverestore-support-check    fail   never pass
 test-armhf-armhf-xl-xsm      12 migrate-support-check        fail   never pass
 test-armhf-armhf-xl-credit2  13 saverestore-support-check    fail   never pass
 test-armhf-armhf-xl-credit2  12 migrate-support-check        fail   never pass
 test-amd64-amd64-xl-qemut-win7-amd64 16 guest-stop             fail never pass
 test-armhf-armhf-xl-cubietruck 12 migrate-support-check        fail never pass
 test-armhf-armhf-xl-cubietruck 13 saverestore-support-check    fail never pass
 test-armhf-armhf-libvirt-qcow2 11 migrate-support-check        fail never pass
 test-armhf-armhf-libvirt-qcow2 13 guest-saverestore            fail never pass
 test-amd64-amd64-libvirt-vhd 11 migrate-support-check        fail   never pass
 test-armhf-armhf-libvirt-raw 13 guest-saverestore            fail   never pass
 test-armhf-armhf-libvirt-raw 11 migrate-support-check        fail   never pass
 test-armhf-armhf-xl-rtds     13 saverestore-support-check    fail   never pass
 test-armhf-armhf-xl-rtds     12 migrate-support-check        fail   never pass
 test-armhf-armhf-xl-vhd      11 migrate-support-check        fail   never pass
 test-armhf-armhf-xl-vhd      12 saverestore-support-check    fail   never pass

version targeted for testing:
 xen                  f2160ba6e60e990060de96f2fc9be645f51f5995
baseline version:
 xen                  899495b60a6e55fc2afa69d4616cb08af212de12

Last test of basis    99754  2016-07-28 14:03:00 Z    8 days
Testing same since    99961  2016-08-05 12:21:29 Z    0 days    1 attempts

------------------------------------------------------------
People who touched revisions under test:
  Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
  George Dunlap <george.dunlap@xxxxxxxxxx>
  Jan Beulich <jbeulich@xxxxxxxx>

jobs:
 build-amd64-xsm                                              pass    
 build-armhf-xsm                                              pass    
 build-i386-xsm                                               pass    
 build-amd64                                                  pass    
 build-armhf                                                  pass    
 build-i386                                                   pass    
 build-amd64-libvirt                                          pass    
 build-armhf-libvirt                                          pass    
 build-i386-libvirt                                           pass    
 build-amd64-prev                                             pass    
 build-i386-prev                                              pass    
 build-amd64-pvops                                            pass    
 build-armhf-pvops                                            pass    
 build-i386-pvops                                             pass    
 build-amd64-rumpuserxen                                      fail    
 build-i386-rumpuserxen                                       fail    
 test-amd64-amd64-xl                                          pass    
 test-armhf-armhf-xl                                          pass    
 test-amd64-i386-xl                                           pass    
 test-amd64-amd64-xl-qemut-debianhvm-amd64-xsm                pass    
 test-amd64-i386-xl-qemut-debianhvm-amd64-xsm                 pass    
 test-amd64-amd64-libvirt-qemuu-debianhvm-amd64-xsm           pass    
 test-amd64-i386-libvirt-qemuu-debianhvm-amd64-xsm            pass    
 test-amd64-amd64-xl-qemuu-debianhvm-amd64-xsm                pass    
 test-amd64-i386-xl-qemuu-debianhvm-amd64-xsm                 pass    
 test-amd64-amd64-xl-qemut-stubdom-debianhvm-amd64-xsm        pass    
 test-amd64-i386-xl-qemut-stubdom-debianhvm-amd64-xsm         pass    
 test-amd64-amd64-libvirt-xsm                                 pass    
 test-armhf-armhf-libvirt-xsm                                 fail    
 test-amd64-i386-libvirt-xsm                                  pass    
 test-amd64-amd64-xl-xsm                                      pass    
 test-armhf-armhf-xl-xsm                                      pass    
 test-amd64-i386-xl-xsm                                       pass    
 test-amd64-amd64-qemuu-nested-amd                            fail    
 test-amd64-amd64-xl-pvh-amd                                  fail    
 test-amd64-i386-qemut-rhel6hvm-amd                           pass    
 test-amd64-i386-qemuu-rhel6hvm-amd                           pass    
 test-amd64-amd64-xl-qemut-debianhvm-amd64                    pass    
 test-amd64-i386-xl-qemut-debianhvm-amd64                     pass    
 test-amd64-amd64-xl-qemuu-debianhvm-amd64                    pass    
 test-amd64-i386-xl-qemuu-debianhvm-amd64                     pass    
 test-amd64-i386-freebsd10-amd64                              pass    
 test-amd64-amd64-xl-qemuu-ovmf-amd64                         pass    
 test-amd64-i386-xl-qemuu-ovmf-amd64                          pass    
 test-amd64-amd64-rumpuserxen-amd64                           blocked 
 test-amd64-amd64-xl-qemut-win7-amd64                         fail    
 test-amd64-i386-xl-qemut-win7-amd64                          fail    
 test-amd64-amd64-xl-qemuu-win7-amd64                         pass    
 test-amd64-i386-xl-qemuu-win7-amd64                          fail    
 test-armhf-armhf-xl-arndale                                  fail    
 test-amd64-amd64-xl-credit2                                  pass    
 test-armhf-armhf-xl-credit2                                  pass    
 test-armhf-armhf-xl-cubietruck                               pass    
 test-amd64-i386-freebsd10-i386                               pass    
 test-amd64-i386-rumpuserxen-i386                             blocked 
 test-amd64-amd64-qemuu-nested-intel                          pass    
 test-amd64-amd64-xl-pvh-intel                                fail    
 test-amd64-i386-qemut-rhel6hvm-intel                         pass    
 test-amd64-i386-qemuu-rhel6hvm-intel                         pass    
 test-amd64-amd64-libvirt                                     pass    
 test-armhf-armhf-libvirt                                     fail    
 test-amd64-i386-libvirt                                      pass    
 test-amd64-amd64-migrupgrade                                 pass    
 test-amd64-i386-migrupgrade                                  pass    
 test-amd64-amd64-xl-multivcpu                                pass    
 test-armhf-armhf-xl-multivcpu                                pass    
 test-amd64-amd64-pair                                        pass    
 test-amd64-i386-pair                                         pass    
 test-amd64-amd64-libvirt-pair                                pass    
 test-amd64-i386-libvirt-pair                                 pass    
 test-amd64-amd64-amd64-pvgrub                                pass    
 test-amd64-amd64-i386-pvgrub                                 pass    
 test-amd64-amd64-pygrub                                      pass    
 test-armhf-armhf-libvirt-qcow2                               fail    
 test-amd64-amd64-xl-qcow2                                    pass    
 test-armhf-armhf-libvirt-raw                                 fail    
 test-amd64-i386-xl-raw                                       pass    
 test-amd64-amd64-xl-rtds                                     pass    
 test-armhf-armhf-xl-rtds                                     pass    
 test-amd64-i386-xl-qemut-winxpsp3-vcpus1                     pass    
 test-amd64-i386-xl-qemuu-winxpsp3-vcpus1                     pass    
 test-amd64-amd64-libvirt-vhd                                 pass    
 test-armhf-armhf-xl-vhd                                      pass    
 test-amd64-amd64-xl-qemut-winxpsp3                           pass    
 test-amd64-i386-xl-qemut-winxpsp3                            pass    
 test-amd64-amd64-xl-qemuu-winxpsp3                           pass    
 test-amd64-i386-xl-qemuu-winxpsp3                            pass    


------------------------------------------------------------
sg-report-flight on osstest.test-lab.xenproject.org
logs: /home/logs/logs
images: /home/logs/images

Logs, config files, etc. are available at
    http://logs.test-lab.xenproject.org/osstest/logs

Explanation of these reports, and of osstest in general, is at
    http://xenbits.xen.org/gitweb/?p=osstest.git;a=blob;f=README.email;hb=master
    http://xenbits.xen.org/gitweb/?p=osstest.git;a=blob;f=README;hb=master

Test harness code can be found at
    http://xenbits.xen.org/gitweb?p=osstest.git;a=summary


Not pushing.

------------------------------------------------------------
commit f2160ba6e60e990060de96f2fc9be645f51f5995
Author: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
Date:   Fri Aug 5 13:40:08 2016 +0200

    x86/mmcfg: Fix initalisation of variables in pci_mmcfg_nvidia_mcp55()
    
    Shifting into the sign bit of an integer is undefined behaviour.
    
    Only the first integer is actually undefined, but switch all the shifts
    for consistency.
    
    Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
    Acked-by: Jan Beulich <JBeulich@xxxxxxxx>
    master commit: ab8fc3937eeb9332b83d7e14d81e37f0b0ef1841
    master date: 2016-08-03 18:46:59 +0100

commit 471a151964da662d917ec61dffcbdb10a09fb285
Author: George Dunlap <george.dunlap@xxxxxxxxxx>
Date:   Fri Aug 5 13:39:36 2016 +0200

    xen: Remove buggy initial placement algorithm
    
    The initial placement algorithm sometimes picks cpus outside of the
    mask it's given, does a lot of unnecessary bitmasking, does its own
    separate load calculation, and completely ignores vcpu hard and soft
    affinities.  Just get rid of it and rely on the schedulers to do
    initial placement.
    
    Signed-off-by: George Dunlap <george.dunlap@xxxxxxxxxx>
    Reviewed-by: Dario Faggioli <dario.faggioli@xxxxxxxxxx>
    Acked-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
    master commit: d5438accceecc8172db2d37d98b695eb8bc43afc
    master date: 2016-07-26 10:44:06 +0100

commit c732d3c7e49ef6a69b34988167e2c20d37bfd49a
Author: George Dunlap <george.dunlap@xxxxxxxxxx>
Date:   Fri Aug 5 13:39:05 2016 +0200

    xen: Have schedulers revise initial placement
    
    The generic domain creation logic in
    xen/common/domctl.c:default_vcpu0_location() attempts to try to do
    initial placement load-balancing by placing vcpu 0 on the least-busy
    non-primary hyperthread available.  Unfortunately, the logic can end
    up picking a pcpu that's not in the online mask.  When this is passed
    to a scheduler such which assumes that the initial assignment is
    valid, it causes a null pointer dereference looking up the runqueue.
    
    Furthermore, this initial placement doesn't take into account hard or
    soft affinity, or any scheduler-specific knowledge (such as historic
    runqueue load, as in credit2).
    
    To solve this, when inserting a vcpu, always call the per-scheduler
    "pick" function to revise the initial placement.  This will
    automatically take all knowledge the scheduler has into account.
    
    csched2_cpu_pick ASSERTs that the vcpu's pcpu scheduler lock has been
    taken.  Grab and release the lock to minimize time spend with irqs
    disabled.
    
    Signed-off-by: George Dunlap <george.dunlap@xxxxxxxxxx>
    Reviewed-by: Meng Xu <mengxu@xxxxxxxxxxxxx>
    Reviwed-by: Dario Faggioli <dario.faggioli@xxxxxxxxxx>
    master commit: 9f358ddd69463fa8fb65cf67beb5f6f0d3350e32
    master date: 2016-07-26 10:42:49 +0100

commit d37c2b9db312f42ac40164c55201c4aca527beae
Author: Jan Beulich <jbeulich@xxxxxxxx>
Date:   Fri Aug 5 13:37:46 2016 +0200

    x86/EFI + Live Patch: avoid symbol address truncation
    
    ld associates __init_end, placed outside of any section by the linker
    script, with the following section, resulting in a huge (wrapped, as it
    would be negative) section relative offset. COFF symbol tables store
    section relative addresses, and hence the above leads to assembler
    truncation warnings when all symbols get included in the symbol table
    (for Live Patching code). To overcome this, move __init_end past both
    ALIGN() directives. The consuming code (init_done()) is fine with such
    an adjustment (the distinction really would only be relevant for the
    loop claring the pages, and I think it's acceptable to clear a few
    more on - for now - EFI). This effectively results in the
    (__init_begin,__init_end) and (__2M_init_start,__2M_init_end) pairs to
    become identical, with their different names only serving documentation
    purposes now.
    
    Note that moving __init_end and __2M_init_end into .init is not a good
    idea, as that would significantly grow xen.efi binary size.
    
    While inspecting symbol table and ld behavior I also noticed that
    __2M_text_start gets put at address zero in the EFI case, which hasn't
    caused problems solely because we don't actually reference that symbol.
    Correct the setting of the initial address, and comment out said symbol
    for the time being, as with the initial address correction it would in
    turn cause an assembler truncation warning similar to the one mentioned
    above.
    
    While checking init_done() for correctness with the above changes I
    noticed that code can easily be folded there, at once correcting the
    logged amount of memory which has got freed for the 2M-alignment case
    (i.e. EFI right now).
    
    Signed-off-by: Jan Beulich <jbeulich@xxxxxxxx>
    Reviewed-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
    master commit: 72e4a1b40a8b62ad3abf14869f78e7ffa5c680a0
    master date: 2016-06-29 16:38:50 +0200
(qemu changes not included)

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
https://lists.xen.org/xen-devel

 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.