|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH 00/19] Virtual NUMA for PV and HVM
Hi all
This patch series implements virtual NUMA support for both PV and HVM guest.
That is, admin can configure via libxl what virtual NUMA topology the guest
sees.
This is the stage 1 (basic vNUMA support) and part of stage 2 (vNUMA-ware
ballooning, hypervisor side) described in my previous email to xen-devel [0].
This series is broken into several parts:
1. xen patches: vNUMA debug output and vNUMA-aware memory hypercall support.
2. libxc/libxl support for PV vNUMA.
3. libxc/libxl support for HVM vNUMA.
4. xl vNUMA configuration documentation and parser.
I think one significant difference from Elena's work is that this patch series
makes use of multiple vmemranges should there be a memory hole, instead of
shrinking ram. This matches the behaviour of real hardware.
The vNUMA auto placement algorithm is missing at the moment and Dario is
working on it.
This series can be found at:
git://xenbits.xen.org/people/liuw/xen.git wip.vnuma-v1
With this series, the following configuration can be used to enabled virtual
NUMA support, and it works for both PV and HVM guests.
memory = 6000
vnuma_memory = [3000, 3000]
vnuma_vcpu_map = [0, 1]
vnuma_pnode_map = [0, 0]
vnuma_vdistances = [10, 30] # optional
dmesg output for HVM guest:
[ 0.000000] ACPI: SRAT 00000000fc009ff0 000C8 (v01 Xen HVM 00000000
HVML 00000000)
[ 0.000000] ACPI: SLIT 00000000fc00a0c0 00030 (v01 Xen HVM 00000000
HVML 00000000)
<...snip...>
[ 0.000000] SRAT: PXM 0 -> APIC 0x00 -> Node 0
[ 0.000000] SRAT: PXM 1 -> APIC 0x02 -> Node 1
[ 0.000000] SRAT: Node 0 PXM 0 [mem 0x00000000-0xbb7fffff]
[ 0.000000] SRAT: Node 1 PXM 1 [mem 0xbb800000-0xefffffff]
[ 0.000000] SRAT: Node 1 PXM 1 [mem 0x100000000-0x186ffffff]
[ 0.000000] NUMA: Initialized distance table, cnt=2
[ 0.000000] NUMA: Node 1 [mem 0xbb800000-0xefffffff] + [mem
0x100000000-0x1867fffff] -> [mem 0xbb800000-0x1867fffff]
[ 0.000000] Initmem setup node 0 [mem 0x00000000-0xbb7fffff]
[ 0.000000] NODE_DATA [mem 0xbb7fc000-0xbb7fffff]
[ 0.000000] Initmem setup node 1 [mem 0xbb800000-0x1867fffff]
[ 0.000000] NODE_DATA [mem 0x1867f7000-0x1867fafff]
[ 0.000000] [ffffea0000000000-ffffea00029fffff] PMD ->
[ffff8800b8600000-ffff8800baffffff] on node 0
[ 0.000000] [ffffea0002a00000-ffffea00055fffff] PMD ->
[ffff880183000000-ffff8801859fffff] on node 1
<...snip...>
[ 0.000000] Early memory node ranges
[ 0.000000] node 0: [mem 0x00001000-0x0009efff]
[ 0.000000] node 0: [mem 0x00100000-0xbb7fffff]
[ 0.000000] node 1: [mem 0xbb800000-0xefffefff]
[ 0.000000] node 1: [mem 0x100000000-0x1867fffff]
numactl output for HVM guest:
available: 2 nodes (0-1)
node 0 cpus: 0
node 0 size: 2999 MB
node 0 free: 2546 MB
node 1 cpus: 1
node 1 size: 2991 MB
node 1 free: 2144 MB
node distances:
node 0 1
0: 10 30
1: 30 10
dmesg output for PV guest:
[ 0.000000] NUMA: Initialized distance table, cnt=2
[ 0.000000] NUMA: Node 1 [mem 0xbb800000-0xce68efff] + [mem
0x100000000-0x1a8970fff] -> [mem 0xbb800000-0x1a8970fff]
[ 0.000000] NODE_DATA(0) allocated [mem 0xbb7fc000-0xbb7fffff]
[ 0.000000] NODE_DATA(1) allocated [mem 0x1a8969000-0x1a896cfff]
numactl output for PV guest:
available: 2 nodes (0-1)
node 0 cpus: 0
node 0 size: 2944 MB
node 0 free: 2917 MB
node 1 cpus: 1
node 1 size: 2934 MB
node 1 free: 2904 MB
node distances:
node 0 1
0: 10 30
1: 30 10
And for a HVM guest on a real NUMA-capable machine:
(XEN) Memory location of each domain:
(XEN) Domain 0 (total: 262144):
(XEN) Node 0: 245758
(XEN) Node 1: 16386
(XEN) Domain 2 (total: 2097226):
(XEN) Node 0: 1046335
(XEN) Node 1: 1050891
(XEN) 2 vnodes, 4 vcpus
(XEN) vnode 0 - pnode 0
(XEN) 3840 MB: 0x0 - 0xf0000000
(XEN) 256 MB: 0x100000000 - 0x110000000
(XEN) vcpus: 0 1
(XEN) vnode 1 - pnode 1
(XEN) 4096 MB: 0x110000000 - 0x210000000
(XEN) vcpus: 2 3
Wei.
[0] <20141111173606.GC21312@xxxxxxxxxxxxxxxxxxxxx>
Wei Liu (19):
xen: dump vNUMA information with debug key "u"
xen: make two memory hypercalls vNUMA-aware
libxc: allocate memory with vNUMA information for PV guest
libxl: add emacs local variables in libxl_{x86,arm}.c
libxl: introduce vNUMA types
libxl: add vmemrange to libxl__domain_build_state
libxl: introduce libxl__vnuma_config_check
libxl: x86: factor out e820_host_sanitize
libxl: functions to build vmemranges for PV guest
libxl: build, check and pass vNUMA info to Xen for PV guest
hvmloader: add new fields for vNUMA information
hvmloader: construct SRAT
hvmloader: construct SLIT
hvmloader: disallow memory relocation when vNUMA is enabled
libxc: allocate memory with vNUMA information for HVM guest
libxl: build, check and pass vNUMA info to Xen for HVM guest
libxl: refactor hvm_build_set_params
libxl: fill vNUMA information in hvm info
xl: vNUMA support
docs/man/xl.cfg.pod.5 | 32 +++++
tools/firmware/hvmloader/acpi/acpi2_0.h | 61 +++++++++
tools/firmware/hvmloader/acpi/build.c | 104 ++++++++++++++
tools/firmware/hvmloader/pci.c | 13 ++
tools/libxc/include/xc_dom.h | 5 +
tools/libxc/include/xenguest.h | 7 +
tools/libxc/xc_dom_x86.c | 72 ++++++++--
tools/libxc/xc_hvm_build_x86.c | 224 +++++++++++++++++++-----------
tools/libxc/xc_private.h | 2 +
tools/libxl/Makefile | 2 +-
tools/libxl/libxl_arch.h | 6 +
tools/libxl/libxl_arm.c | 17 +++
tools/libxl/libxl_create.c | 9 ++
tools/libxl/libxl_dom.c | 172 ++++++++++++++++++++---
tools/libxl/libxl_internal.h | 18 +++
tools/libxl/libxl_types.idl | 9 ++
tools/libxl/libxl_vnuma.c | 228 +++++++++++++++++++++++++++++++
tools/libxl/libxl_x86.c | 113 +++++++++++++--
tools/libxl/xl_cmdimpl.c | 151 ++++++++++++++++++++
xen/arch/x86/numa.c | 46 ++++++-
xen/common/memory.c | 58 +++++++-
xen/include/public/features.h | 3 +
xen/include/public/hvm/hvm_info_table.h | 19 +++
xen/include/public/memory.h | 2 +
24 files changed, 1247 insertions(+), 126 deletions(-)
create mode 100644 tools/libxl/libxl_vnuma.c
--
1.7.10.4
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |