WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-users

Re: [Xen-users] Re: Xen Disk I/O performance vs nativeperformance:XenI/O

To: xen-users <xen-users@xxxxxxxxxxxxxxxxxxx>
Subject: Re: [Xen-users] Re: Xen Disk I/O performance vs nativeperformance:XenI/O is definitely super super super slow
From: "Olivier B." <xen.list@xxxxxxxxx>
Date: Thu, 26 Feb 2009 10:22:15 +0100
Delivery-date: Thu, 26 Feb 2009 01:22:52 -0800
Envelope-to: www-data@xxxxxxxxxxxxxxxxxxx
In-reply-to: <7309E5BCEDC4DC4BA820EF9497269EAD0461B2A8@xxxxxxxxxxxxxxxxxxxxxxxxxxxxx>
List-help: <mailto:xen-users-request@lists.xensource.com?subject=help>
List-id: Xen user discussion <xen-users.lists.xensource.com>
List-post: <mailto:xen-users@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/mailman/listinfo/xen-users>, <mailto:xen-users-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/mailman/listinfo/xen-users>, <mailto:xen-users-request@lists.xensource.com?subject=unsubscribe>
References: <7309E5BCEDC4DC4BA820EF9497269EAD0461B244@xxxxxxxxxxxxxxxxxxxxxxxxxxxxx><7207d96f0902120602t1be864acm684fbe6b8f0f18aa@xxxxxxxxxxxxxx><7309E5BCEDC4DC4BA820EF9497269EAD0461B246@xxxxxxxxxxxxxxxxxxxxxxxxxxxxx><7207d96f0902122011x541c63eewe33fe0ef922cd0c9@xxxxxxxxxxxxxx><7309E5BCEDC4DC4BA820EF9497269EAD0461B24D@xxxxxxxxxxxxxxxxxxxxxxxxxxxxx><7309E5BCEDC4DC4BA820EF9497269EAD0461B24F@xxxxxxxxxxxxxxxxxxxxxxxxxxxxx><7207d96f0902132122n409d71ceg2c19e3ec70f52f45@xxxxxxxxxxxxxx> <7309E5BCEDC4DC4BA820EF9497269EAD0461B292@xxxxxxxxxxxxxxxxxxxxxxxxxxxxx> <7309E5BCEDC4DC4BA820EF9497269EAD0461B295@xxxxxxxxxxxxxxxxxxxxxxxxxxxxx><7309E5BCEDC4DC4BA820EF9497269EAD0461B2A3@xxxxxxxxxxxxxxxxxxxxxxxxxxxxx> <49A5A04E.60100@xxxxxxxxx> <7309E5BCEDC4DC4BA820EF9497269EAD0461B2A8@xxxxxxxxxxxxxxxxxxxxxxxxxxxxx>
Sender: xen-users-bounces@xxxxxxxxxxxxxxxxxxx
User-agent: Mozilla-Thunderbird 2.0.0.19 (X11/20090103)
To be sure I redo all tests with adding "conv=fdatasync" to dd :
   dd conv=fdatasync if=/dev/zero of=TEST bs=4k count=128000

Results don't really change.

For memory, all my domU have 512Mo of memory.

On dom0, I checked with 200Mo or 4Go free memory, this doesn't change results. I adjust dd count parameter to create 512Mo or 8Go file, and have similar results (there is a small variation of course, about 10%).

I checked also with the Debian Lenny Xen kernel for DomU (so without ext4), and have same results to.

I really don't think it's a FS issue ; but I suppose the writeback feature of ext4 avoid this write problems (I didn't try to disable it).

Olivier

DOGUET Emmanuel a écrit :
If you have differences between ext3 and ext4, it can be due to the FS cache?
In your tests what are the memory of dom0 and domU?
Very strange our write problem to RAID harware :/

For my tests, I use "standard" redhat kernel.


  Bye.

-----Message d'origine-----
De : xen-users-bounces@xxxxxxxxxxxxxxxxxxx [mailto:xen-users-
bounces@xxxxxxxxxxxxxxxxxxx] De la part de Olivier B.
Envoyé : mercredi 25 février 2009 20:47
Cc : xen-users
Objet : Re: [Xen-users] Re: Xen Disk I/O performance vs
nativeperformance:XenI/O is definitely super super super slow

I did some tests to on some servers ( time dd if=/dev/zero of=TEST bs=4k
count=512000 )

First server : raid 1 hardware with a PCI 32bits 3ware card.
dom0 (ext3) :    39MB/s
domU (ext3) :     1.4MB/s !!!
domU (ext4) :    40MB/s

Second server : raid 1 software with 2 SATA disks
dom0 (ext3) :     96MB/s
domU (ext3) :     91MB/s
domU (ext4) :     94MB/s

Note : I use vanilla kernel on DomU.

So :
- I see a big write problem from domU on hardware raid
- the writeback feature of ext4 seem to "erase" this problem

Olivier

DOGUET Emmanuel a écrit :
I have finished my tests on 3 servers. On each we loose some bandwidth
with XEN. On our 10 platform ... We always loose some bandwidth, I think
it's normal. Just the bench method who must differ?
I have made bench (write only) between hardware and software RAID under
XEN (see attachment).
Linux Software RAID is always faster than HP Raid. I must try too the
"512MB+Cache Write" option for the HP Raid.
So my problems seem to be here.


-------------------------
HP DL 380
Quad core
-------------------------
Test: dd if=/dev/zero of=TEST bs=4k count=1250000



          Hardware     Hardware    Software     Software
           RAID 5        RAID 5      RAID 5       RAID 5
          4 x 146G      8 x 146G   4 x 146G     8 x 146G
dom0
(1024MB,
 1 cpu)     32MB        22MB        88MB (*)    144MB (*)

domU
( 512MB,
 1 cpu)      8MB        5MB         34MB        31MB

domU
 (4096MB,
 2 cpu)     --          7MB         51MB        35MB



*: don't understand this difference.


This performance seems to be good for you?




         Best regards.





-----Message d'origine-----
De : DOGUET Emmanuel
Envoyé : mardi 24 février 2009 17:50
À : DOGUET Emmanuel; Fajar A. Nugraha
Cc : xen-users; Joris Dobbelsteen
Objet : RE: [Xen-users] Re: Xen Disk I/O performance vs
nativeperformance:Xen I/O is definitely super super super slow

For resuming :

on RAID 0

        dom0: 80MB      domU:   56MB                    Loose: 30M

on RAID1

        dom0: 80MB      domU:  55 MB            Loose: 32%

on RAID5:

        dom0: 30MB      domU:   9MB                     Loose: 70%



So loose seem to be "exponantial" ?




-----Message d'origine-----
De : xen-users-bounces@xxxxxxxxxxxxxxxxxxx
[mailto:xen-users-bounces@xxxxxxxxxxxxxxxxxxx] De la part de
DOGUET Emmanuel
Envoyé : mardi 24 février 2009 14:22
À : Fajar A. Nugraha
Cc : xen-users; Joris Dobbelsteen
Objet : RE: [Xen-users] Re: Xen Disk I/O performance vs
nativeperformance:Xen I/O is definitely super super super slow


I have made another test on another server (DL 380)

And same thing!

I'm always use this test :

dd if=/dev/zero of=TEST bs=4k count=1250000

(be careful with memory cache)


TEST WITH 2 RAID 5 (include system on RAID 5, 3x146G + 3x146G)
---------------------------------------------------------------

dom0: 1GO, 1CPU, 2 RAID 5

       rootvg(c0d0p1):         4596207616 bytes (4.6 GB)
copied, 158.284 seconds, 29.0 MB/s
       datavg(c0d1p1):         5120000000 bytes (5.1 GB)
copied, 155.414 seconds, 32.9 MB/s

domU: 512M, 1CPU         on System LVM/RAID5 (rootvg)

       5120000000 bytes (5.1 GB) copied, 576.923 seconds, 8.9 MB/s

domU: 512M, 1CPU         on DATA LVM/RAID5 (datavg)

       5120000000 bytes (5.1 GB) copied, 582.611 seconds, 8.8 MB/s

domU: 512M, 1 CPU on same RAID without LVM

       5120000000 bytes (5.1 GB) copied, 808.957 seconds, 6.3 MB/s


TEST WITH RAID 0 (dom0 system on RAID 1)
---------------------------------------

dom0   1GO RAM 1CPU

       on system (RAID1):
       i3955544064 bytes (4.0 GB) copied, 57.4314 seconds, 68.9 MB/s

       on direct HD (RAID 0 of cssiss), no LVM
       5120000000 bytes (5.1 GB) copied, 62.5497 seconds, 81.9 MB/s

dom0   4GO RAM 4CPU



domU:  4GO, 4 CPU

       on direct HD (RAID 0), no LVM.
       5120000000 bytes (5.1 GB) copied, 51.2684 seconds, 99.9 MB/s


domU: 4GO, 4CPU  same HD but ONE LVM on it

       5120000000 bytes (5.1 GB) copied, 51.5937 seconds, 99.2 MB/s


TEST with only ONE RAID 5 (6 x 146G)
------------------------------------

dom0 : 1024MB - 1CPUI (RHEL 5.3)

       5120000000 bytes (5.1 GB) copied, 231.113 seconds, 22.2 MB/s


512MB - 1 CPU
       5120000000 bytes (5.1 GB) copied, 1039.42 seconds, 4.9 MB/s


512MB - 1 CPU - ONLY 1 VDB [LVM] (root, no swap)

       (too slow ..stopped :P)
       4035112960 bytes (4.0 GB) copied, 702.883 seconds, 5.7 MB/s

512MB - 1 CPU - On a file (root, no swap)

       1822666752 bytes (1.8 GB) copied, 2753.91 seconds, 662 kB/s

4GB - 2 CPU
       5120000000 bytes (5.1 GB) copied, 698.681 seconds, 7.3 MB/s





-----Message d'origine-----
De : Fajar A. Nugraha [mailto:fajar@xxxxxxxxx]
Envoyé : samedi 14 février 2009 06:23
À : DOGUET Emmanuel
Cc : xen-users
Objet : Re: [Xen-users] Re: Xen Disk I/O performance vs native
performance:Xen I/O is definitely super super super slow

2009/2/13 DOGUET Emmanuel <Emmanuel.DOGUET@xxxxxxxx>:

I have mount domU partition on dom0 for testing and it's OK.
But same partiton on domU side is slow.

Strange.

Strange indeed. At least that ruled-out hardware problems :)
Could try with a "simple" domU?
- 1 vcpu
- 512 M memory
- only one vbd

this should isolate whether or not the problem is on your particular
domU (e.g. some config parameter actually make domU slower).

Your config file should have only few lines, like this

memory = "512"
vcpus=1
disk = ['phy:/dev/rootvg/bdd-root,xvda1,w' ]
vif = [ "mac=00:22:64:A1:56:BF,bridge=xenbr0" ]
vfb =['type=vnc']
bootloader="/usr/bin/pygrub"

Regards,

Fajar


_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users


_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users

_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users


_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users

<Prev in Thread] Current Thread [Next in Thread>