Forwarding to xen-devel ..
Jeremy: Have you heard of this before? save/restore/migration
takes 4-10x longer on pvops 2.6.32 compared to xenlinux 2.6.32.
He verified the DEBUG options in kernel .configs are the same.
-- Pasi
On Wed, May 12, 2010 at 05:28:26PM -0400, Pierre POMES wrote:
> Hi,
>
> First sorry for the double posting...
>
> I just built a 2.6.32.10 kernel with Andrew Lyon patches (so it is a
> "xenlinux" kernel, not a "pvops" kernel).
>
> Live migration and restore operations are between 4 and 10 times faster
> with this kernel . Furthermore, during live migration, hangs time in
> domu are shorter (1-2 seconds versus 1 to 15 seconds for a domu with
> 256M RAM).
>
> Error messages "Error when reading batch size" / "error when buffering
> batch, finishing" are still in my logs.
>
> Regarding times, all is now similar to what I had with Xen 3.x on top of
> xenlinux kernels.
>
> Regards,
> Pierre
>
>
>
>> Hi all,
>>
>> I am using Xen 4.0.0 on top of Ubuntu Lucid (amd64), with the Jeremy
>> kernel taken from git (xen/stable-2.6.32.x branch, 2.6.32.12 when I am
>> writing this email). This kernel is also used in my domu.
>>
>> I can save a domu without any problem, but restoring it may need from
>> 2 to 5 minutes, from a 1G checkpoint file (domu has 1GB RAM). There
>> also errors in /var/log/xen/xend.log, "Error when reading batch size"
>> and "Error when reading batch size":
>>
>> [2010-05-08 04:23:16 9497] DEBUG (XendDomainInfo:1804) Storing domain
>> details: {'image/entry': '18446744071587529216', 'console/port': '2',
>> 'image/loader': 'generic', 'vm':
>> '/vm/156ea44d-6707-cbe6-2d58-7bea4792dff4',
>> 'control/platform-feature-multiprocessor-suspend': '1',
>> 'image/hv-start-low': '18446603336221196288', 'image/guest-os':
>> 'linux', 'image/virt-base': '18446744071562067968', 'memory/target':
>> '1048576', 'image/guest-version': '2.6', 'image/pae-mode': 'yes',
>> 'description': '', 'console/limit': '1048576', 'image/paddr-offset':
>> '0', 'image/hypercall-page': '18446744071578882048',
>> 'image/suspend-cancel': '1', 'cpu/0/availability': 'online',
>> 'image/features/pae-pgdir-above-4gb': '1',
>> 'image/features/writable-page-tables': '0', 'console/type':
>> 'xenconsoled', 'name': 'domusample', 'domid': '10',
>> 'image/xen-version': 'xen-3.0', 'store/port': '1'}
>> [2010-05-08 04:23:16 9497] DEBUG (XendCheckpoint:286)
>> restore:shadow=0x0, _static_max=0x40000000, _static_min=0x0,
>> [2010-05-08 04:23:16 9497] DEBUG (XendCheckpoint:305) [xc_restore]:
>> /usr/lib/xen/bin/xc_restore 22 10 1 2 0 0 0 0
>> [2010-05-08 04:23:16 9497] INFO (XendCheckpoint:423) xc_domain_restore
>> start: p2m_size = 40000
>> [2010-05-08 04:23:16 9497] INFO (XendCheckpoint:423) Reloading memory
>> pages: 0%
>> [2010-05-08 04:25:53 9497] INFO (XendCheckpoint:423) ERROR Internal
>> error: Error when reading batch size
>> [2010-05-08 04:25:53 9497] INFO (XendCheckpoint:423) ERROR Internal
>> error: error when buffering batch, finishing
>> [2010-05-08 04:25:53 9497] INFO (XendCheckpoint:423)
>> [2010-05-08 04:25:53 9497] INFO (XendCheckpoint:423) ^H^H^H^H100%
>> [2010-05-08 04:25:53 9497] INFO (XendCheckpoint:423) Memory reloaded
>> (0 pages)
>> [2010-05-08 04:25:53 9497] INFO (XendCheckpoint:423) read VCPU 0
>>
>> Live migration has the same problem, it may need several minutes to
>> complete. Please note that restore and migration do not fail, but
>> there are very long.
>>
>> My domu is on top of DRBD, and the config file is:
>>
>>
>> -------------
>> kernel = '/boot/vmlinuz-2.6.32.12-it-xen'
>> ramdisk = '/boot/initrd.img-2.6.32.12-it-xen'
>> memory = '1024'
>>
>> #
>> # Disk device(s).
>> #
>> root = '/dev/xvda2 ro'
>> disk = [
>> 'drbd:domusampleswap,xvda1,w',
>> 'drbd:domusampleslash,xvda2,w',
>> ]
>>
>>
>>
>> #
>> # Hostname
>> #
>> name = 'domusample'
>>
>> #
>> # Networking
>> #
>> vif = [ 'mac=00:16:3E:58:FC:F9' ]
>>
>> #
>> # Behaviour
>> #
>> on_poweroff = 'destroy'
>> on_reboot = 'restart'
>> on_crash = 'restart'
>>
>> extra = '2 console=hvc0'
>> ----------
>>
>> I do not have any idea here.
>>
>> Did somebody already have (and solve ?) this issue ?
>>
>> Thanks.
>> Pierre
>>
>>
>> *
>> *
>>
>> _______________________________________________
>> Xen-users mailing list
>> Xen-users@xxxxxxxxxxxxxxxxxxx
>> http://lists.xensource.com/xen-users
>
>
> _______________________________________________
> Xen-users mailing list
> Xen-users@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-users
_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users
|