Simon Bettison wrote:
> Hi All,
>
> This is my first time on this (or in fact any as I recall!) list so be
> gentle :)
>
> I am seeing some peculiar load behaviour with a gentoo linux domu which
> is acting as mailserver. The load average starts out fine but then grows
> over the course of several hours topping out at around 120 (one hundred
> and twenty!)
>
> The strange thing is the machine seems responsive, imap connections
> working ok, mail being delivered, snappy shell access etc
>
> This makes me suspicious that the load average is more a reflection of
> the vm load on the cpu slice its being given (ie its only getting say
> 0.01 of the cpu time, and would be running at say 0.11 if it was not
> virtualised) maybe I'm way off though!
>
> The setup is not overly complex:
>
> # xm list
> Name ID Mem VCPUs State
> Time(s)
> Domain-0 0 1708 1 r-----
> 1682.4
> gentoo-spud 5 1024 1 -b----
> 2792.4
> solaris 2 1024 1 -b----
> 1757.4
>
> its a gentoo-linux dom0, its running on amd-vt hardware, do both domUs
> are unmodified guests...
>
> xen # cat /etc/xen/gentoo-spud
> name = "gentoo-spud";
> memory = 1024;
> disk = [ "file:/var/xen/gentoo/spud/domU,xvda,w" ];
> kernel = "/var/xen/gentoo/spud/vmlinuz";
> root = "/dev/xvda ro";
> vif = [ "mac=00:16:3E:00:00:12" ];
>
> xen # cat /etc/xen/solaris.py
> name = 'solaris'
> memory = '1024'
> disk = [ 'file:/var/xen/solaris/domU,0,w' , 'phy:/dev/sdb,1,w' ,
> 'phy:/dev/sdc,2,w' ,'phy:/dev/sdd,3,w' ]
> vif = [ 'mac=00:16:3E:00:00:11' ]
>
> The solaris vm is providing zfs on the 3 physical devices ie
>
>
> solaris # zpool list
> NAME SIZE USED AVAIL CAP HEALTH ALTROOT
> tank 832G 200G 632G 24% ONLINE -
>
> solaris # zpool status
> pool: tank
> state: ONLINE
> scrub: none requested
> config:
>
> NAME STATE READ WRITE CKSUM
> tank ONLINE 0 0 0
> raidz1 ONLINE 0 0 0
> c0d1 ONLINE 0 0 0
> c0d2 ONLINE 0 0 0
> c0d3 ONLINE 0 0 0
>
> solaris # zfs list
> NAME USED AVAIL REFER MOUNTPOINT
> tank 133G 412G 24.0K /tank
> tank/mail 1.32G 412G 1.32G /export/zfs/mail
>
> the tank/mail share is exported over nfs:
>
> solaris # zfs get sharenfs tank/mail
> NAME PROPERTY VALUE SOURCE
> tank/mail sharenfs rw,anon=0 local
>
> the linux box is running:
> postfix
> courier-imapd-ssl
> courier-authlib
> amavisd
> spamassassin
> clamav
>
> mail is routed through postfix on port 25 to amavis on port 10024 which
> takes care of spam and av then back into postfix on 10025 for local delivery
>
> all domains/users are virtual on postfix with delivery to
> /home/vmail/<user@domain>/
>
> /home/vmail is a symbolic link to /mnt/zfs/mail
> /mnt/zfs/mail is nfs mounted share from the solaris vm
>
> through a process of elimination I found that restarting
> courier-imapd-ssl would cause the loadavg to head back towards zero,
> where it would stay for short while but then eventually start to
> escalate again.
>
> I can't find any explanation....
>
> # vmstat 1
> procs -----------memory---------- ---swap-- -----io---- -system--
> ----cpu----
> r b swpd free buff cache si so bi bo in cs us sy
> id wa
> 0 0 0 495912 14488 185348 0 0 34 64 399 296 2 1
> 96 0
> 0 0 0 494296 14488 185352 0 0 0 0 269 97 5 1
> 93 0
> 0 0 0 494296 14488 185352 0 0 0 0 207 82 0 0
> 99 0
> 0 0 0 494288 14488 185400 0 0 0 0 131 64 0 0
> 99 0
> 0 0 0 494288 14488 185452 0 0 0 0 128 60 0 0
> 100 0
> 0 0 0 494304 14488 185472 0 0 0 8 130 74 0 0
> 99 0
> 0 0 0 494304 14488 185472 0 0 0 0 144 67 0 0
> 100 0
> 0 0 0 494368 14488 185512 0 0 0 0 127 62 0 0
> 99 0
> 0 0 0 494260 14488 185512 0 0 0 0 141 73 0 0
> 100 0
> 0 0 0 494252 14488 185556 0 0 0 0 143 65 0 0
> 99 0
> 0 0 0 494252 14520 185576 0 0 0 32 116 68 0 0
> 99 0
> 0 0 0 494300 14520 185596 0 0 0 0 127 69 0 0
> 100 0
> 0 0 0 494192 14520 185596 0 0 0 0 141 68 0 0
> 99 0
> 0 0 0 494184 14520 185640 0 0 0 0 153 60 0 0
> 99 0
> 0 0 0 494184 14520 185640 0 0 0 0 122 59 0 0
> 100 0
> 0 0 0 494216 14520 185680 0 0 0 0 124 72 0 0
> 99 0
> 0 0 0 494108 14520 185680 0 0 0 0 123 67 0 0
> 100 0
> 0 0 0 494092 14520 185720 0 0 0 0 117 63 0 0
> 99 0
> 0 0 0 494092 14520 185720 0 0 0 0 119 64 0 0
> 100 0
> 0 0 0 494124 14520 185760 0 0 0 0 123 66 0 0
> 99 0
> 0 0 0 494124 14520 185760 0 0 0 0 125 70 0 0
> 99 0
> 0 0 0 494016 14520 185800 0 0 0 0 118 60 0 0
> 100 0
> 0 0 0 494016 14520 185800 0 0 0 0 93 42 0 0
> 97 0
> 0 0 0 494032 14520 185820 0 0 0 0 25 18 0 0
> 99 0
> 0 0 0 494032 14520 185820 0 0 0 0 11 12 0 0
> 100 0
> 0 0 0 494032 14520 185828 0 0 0 0 89 55 0 0
> 100 0
> 0 0 0 494032 14520 185828 0 0 0 0 138 68 0 1
> 98 0
> 0 0 0 494040 14520 185872 0 0 0 0 118 66 0 0
> 99 0
> 0 0 0 493932 14520 185872 0 0 0 0 122 63 0 0
> 99 0
> 0 0 0 493916 14520 185912 0 0 0 0 129 62 0 0
> 100 0
> 0 0 0 493916 14552 185932 0 0 0 44 126 76 0 0
> 99 0
> 0 0 0 493916 14552 185956 0 0 0 0 123 65 0 0
> 100 0
> 0 0 0 493808 14552 185956 0 0 0 0 119 58 0 0
> 99 0
> 0 0 0 493800 14552 185996 0 0 0 0 132 71 0 0
> 100 0
> 0 0 0 493800 14552 185996 0 0 0 0 120 62 0 0
> 99 0
> 0 0 0 493800 14552 186040 0 0 0 36 131 73 0 0
> 99 0
> 0 0 0 493692 14552 186040 0 0 0 0 133 68 0 0
> 100 0
> 0 0 0 493676 14552 186080 0 0 0 0 130 66 0 0
> 99 0
> 0 0 0 493676 14552 186080 0 0 0 0 125 63 0 0
> 99 0
> 0 0 0 493684 14552 186124 0 0 0 0 139 79 0 0
> 100 0
> 0 0 0 493684 14552 186124 0 0 0 0 126 68 0 0
> 99 0
> 0 0 0 493584 14552 186164 0 0 0 0 125 62 0 0
> 99 0
> 0 0 0 493584 14552 186164 0 0 0 0 127 60 0 0
> 100 0
> 0 0 0 493584 14552 186204 0 0 0 0 130 80 0 0
> 99 0
> 0 0 0 493584 14552 186204 0 0 0 0 132 69 0 0
> 99 0
> 0 0 0 493460 14564 186248 0 0 0 12 119 74 0 0
> 100 0
> 0 0 0 493460 14564 186248 0 0 0 0 129 68 0 0
> 99 0
> 0 0 0 493452 14564 186288 0 0 0 0 145 75 0 0
> 100 0
>
> thought maybe it was nfs...?
>
> # iostat -n 1
> Linux 2.6.21-xen (spud) 05/28/08
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 10363.60 13.49 0.00
> 0.00 11.30 13.49
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 24.18 24.19 0.00
> 0.00 0.00 18.67
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 0.00 0.05 0.00
> 0.00 0.00 5.56
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 1.98 0.06 0.00
> 0.00 0.00 0.05
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 4137.37 0.06 0.00
> 0.00 0.00 0.07
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 12051.17 4.16 0.00
> 0.00 0.00 4.16
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 6151.21 7.25 0.00
> 0.00 0.00 7.26
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 15.48 15.49 0.00
> 0.00 0.00 15.49
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 26.00 26.00 0.00
> 0.00 0.00 26.00
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 26.98 27.00 0.00
> 0.00 0.00 21.00
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 26.79 26.79 0.00
> 0.00 0.00 29.09
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 20.99 21.00 0.00
> 0.00 0.00 18.77
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 21.20 21.24 0.00
> 0.00 0.00 27.23
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 0.00 0.05 0.00
> 0.00 0.00 0.05
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 1.98 0.05 0.00
> 0.00 0.00 0.05
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 0.00 0.05 0.00
> 0.00 0.00 0.05
>
> Filesystem: rBlk_nor/s wBlk_nor/s rBlk_dir/s
> wBlk_dir/s rBlk_svr/s wBlk_svr/s
> swanney:/export/zfs/mail 2.75 2.79 0.00
> 0.00 0.00 2.79
>
> the machine is hardly begin touched - only two mailboxes on it, very
> litle mail passing through.
>
> any pointers would be really appreciated
>
> Simon Bettison
>
I'm not sure what exactly the issue was, but i think it might have
something to do with courier-imap + thousands of e-mails (~20k total
with some maildirs having 5k in) + NFS
installed dovecot insted and this seems to be working fine.
Simon
_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users
|