[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

RE: [Xen-devel] [PATCH 0/5] [POST-4.0]: RFC: HVM NUMA guest support



Andre Przywara wrote:
> Cui, Dexuan wrote:
>> Hi Andre,  will you re-post your patches?
> Yes, I will do in the next days. I plan to add the missing automatic
> assignment patch before posting.
Glad to know this.
BTW: To support PV NUMA, on this Monday, Dulloor posted some paches that change 
libxc and the hypervisor, too.

Hi Andre, Dulloor,
I believe we should have some coordination to share the code and to avoid 
duplicate efforts.
e.g., Dullor's linux-01-sync-interface.patch is similar to Andre's old patch 
http://lists.xensource.com/archives/html/xen-devel/2008-07/msg00254.html, 
though the formar is for PV kernel and the latter  is for libxc and hypervisor. 
:-)
e.g., Dullor's xen-02-exact-node-request.patch has implemented the the 
MEMF_exact_node flag, which I intended to do. :-)
e.g., Dullor's xen-03-guest-numa-interface.patch implements a hypercall to 
export host numa info -- actually Nitin has sent out a patch to export more 
useful numa info: 
http://old.nabble.com/Host-Numa-informtion-in-dom0-td27379527.html and I 
suppose Nitin will re-send it soon.
e.g., Dullor's xen-04-node-mem-allocation.patch's xc_select_best_fit_nodes() is 
similar to  the Andre's xc_getnodeload(): 
http://lists.xensource.com/archives/html/xen-devel/2010-02/msg00284.html.
e.g., In Dulloer's xen-05-basic-cpumap-utils.patch and 
xen-07-tools-arch-setup.patch, I think some parts could be shared by pv/hvm 
numa implementations if we make some necessary changes to them.
 
>> Now I think for the first implementation, we can make things simple,
>  > e.g, we should specify how many guest nodes (the "guestnodes"
>  option > in your patch -- I think "numa_nodes", or "nodes", may be a
>  better > naming) the hvm guest will see, and we distribute guest
>  memory and > vcpus uniformly among the guest nodes.
> I agree, making things simple in the first step was also my intention.
> We have enough time to make it better later if we have more experience
> with it.
> To be honest, my first try also used "nodes" and later "numa_nodes" to
> specify the number, but I learned that it confuses users who don't see
> the difference between host and guest NUMA functionality. So I wanted
> to make sure that it is clear that this number is from the guest's
> point of view.
> 
>> And we should add one more option "uniform_nodes" -- this boolean
>  > option's default value can be True, meaning if we can't construct
>  > uniform nodes to guest(e.g., on the related host node, no enough
>> memory as expected can be allocated to the guest),  the guest
>  > creation should fail. This option is useful to users who want
>  > predictable guest performance.
> I agree that we need to avoid missing user influence, although I'd
> prefer to have the word "strict" somewhere in this name. As I wrote in
> one my earlier mails, I'd opt for a single option describing the
> policy, the "strict" meaning could be integrated in there:
> numa_policy="strict|uniform|automatic|none|single|..."
Hi Andre,
I think this looks too complex for the first simple implementation and it's 
very likely a real user will be bewildered. :-)
I think ideally we can have 2 options:
guest_nodes=n
uniform_nodes=True|False (the default is True)

Thanks,
-- Dexuan

_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.