| On Tue, May 6, 2008 at 11:05 PM, jim burns <jim_burn@xxxxxxxxxxxxx> wrote:
> On Tuesday May 06 2008 03:21:29 am Pasi Kärkkäinen wrote:
>
> > OK. I think measuring pv domU is worth trying too :)
>
>  Ok, let's try a few things. Repeating my original 0.8.9 numbers, with the new
>  processor:
>
>
>  pattern 4k, 50% read, 0% random
>
>  dynamo on?  |   io/s   |  MB/s  | Avg. i/o time(ms} | max i/o time(ms) | %CPU
>  domu w/gplpv|   501.7  |   1.96 |        2.90       |       0          | 
> 31.68
>  domu w/qemu |   187.5  |   0.73 |        5.87       |       0          | 
> 29.89
>  dom0 w/4Gb  |  1102.3  |   4.31 |        0.91       |      445.5       |  0
>  dom0 w/4Gb  |  1125.8  |   4.40 |        0.89       |      332.1       |  0
>  (2nd dom0 numbers from when booted w/o /gplpv)
>
>  pattern 32k, 50% read, 0% random
>
>  domu w/gplpv|   238.3  |   7.45 |        4.09       |        0         | 
> 22.48
>  domu w/qemu |   157.4  |   4.92 |        6.35       |        0         | 
> 20.51
>  dom0 w/4Gb  |    52.5  |   1.64 |       19.05       |     1590.0       |  0
>  dom0 w/4Gb  |    87.8  |   2.74 |       11.39       |     1286.4       |  0
>
>  Now, that was with all workers running on domu and dom0 simultaneously. Let's
>  try one at a time. On hvm w/gplpv, first the 4k pattern, then later the 32k
>  pattern, with dom0 using the 'idle' task:
>
>   4k pattern |  1026.6  |   4.01 |       39.37       |       0          | 
> 49.70
>  32k pattern |   311.1  |   9.72 |       45.33       |       0          | 
> 26.21
>
>  Now test dom0, with the hvm running the 'idle' task:
>
>   4k pattern |  1376.7  |   5.38 |        0.73       |      365.7       |  0
>  32k pattern |   165.9  |   5.19 |        6.02       |      226.6       |  0
>
>  As expected, all numbers are significantly faster. Compare this to 'dd'
>  creating the 4GB /iobw.tst file on dom0 at a 22MB/s rate.
>
>  Now, to test a fedora pv, since space is tight on my fedora xen server, I
>  just 'xm block-attach'-ed dom0's /iobw.tst as a new partition on the domu,
>  and in the domu, did mkfs, mount, and created a new /iobw.tst on that
>  partition. Results:
>
>   4k pattern |  1160.5  |   4.53 |        0.86       |      247.1       |  0
>  32k pattern |   284.1  |   8.88 |        3.52       |      326.4       |  0
>
>  The numbers are very similar to the hvm, including the 32k pattern being
>  faster than dom0, which you pointed out is due to caching. This compares
>  to 'dd' creating the 3.7GB iobw.tst on the mounted new partition at an 18MB/s
>  rate.
>
>
>  > Configure dom0 for 1 vcpu and domU for 1 vcpu and pin the domains to have a
>  > dedicated core. This way you're not sharing any pcpu's between the domains.
>  > I think this is the "recommended" setup from xen developers for getting
>  > maximum performance.
>  >
>  > I think the performance will be worse when you have more vcpus in use than
>  > your actual pcpu count..
>
>  Now I rebooted dom0, after editing xend-config.sxp to include '(dom0-cpus 
> 1)',
>  and then did the following pins:
>
>  [576] > xm create winxp
>  Using config file "/etc/xen/winxp".
>  Started domain winxp
>  root@Insp6400 05/06/08 10:32PM:~
>  [577] > xm vcpu-pin 0 all 0
>  root@Insp6400 05/06/08 10:32PM:~
>  [578] > xm vcpu-pin winxp all 1
>  root@Insp6400 05/06/08 10:32PM:~
>  [579] > xm vcpu-list
>
> Name                               ID  VCPU   CPU State   Time(s) CPU Affinity
>  Domain-0                             0     0     0   r--     228.7 0
>  Domain-0                             0     1     -   --p      16.0 0
>  winxp                                5     0     1   r--      36.4 1
>
>  Note I also had to set vcpus=1, because with two, I was again getting that
>  extremely sluggish response in my hvm.
>
>  Going back to simultaneous execution of all workers, to compare against the
>  numbers at the top of this post, I got:
>
>
>  pattern 4k, 50% read, 0% random
>
>  dynamo on?  |   io/s   |  MB/s  | Avg. i/o time(ms} | max i/o time(ms) | %CPU
>  domu w/gplpv|   286.4  |   1.12 |        3.49       |      564.9       | 
> 36.97
>  dom0 w/4Gb  |  1173.9  |   4.59 |        0.85       |      507.3       |  0
>
>
>  pattern 32k, 50% read, 0% random
>
>  domu w/gplpv|   217.9  |   6.81 |        4.57       |     1633.5       | 
> 22.93
>  dom0 w/4Gb  |    63.3  |   1.97 |       15.85       |     1266.5       |  0
>
>  which is somewhat slower. Recommendations of the xen developers aside, my
>  experience is that allowing xen to schedule any vcpu on any pcpu is most
>  efficient.
>
I think that your experience (allowing Xen to do the scheduling itself is most
efficient and only try to tweak the scheduling in very special cases and/or you
really know what you are doing) should be considered conventional wisdom.
Can you refresh me on the recommendations of the Xen developers that you
are referring to?
Thanks,
Todd
>
>
>  _______________________________________________
>  Xen-users mailing list
>  Xen-users@xxxxxxxxxxxxxxxxxxx
>  http://lists.xensource.com/xen-users
>
_______________________________________________
Xen-users mailing list
Xen-users@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-users
 |