[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: Increasing domain memory beyond initial maxmem
 
 
On 05.04.22 18:24, Marek Marczykowski-Górecki wrote:
 
On Tue, Apr 05, 2022 at 01:03:57PM +0200, Juergen Gross wrote:
 
Hi Marek,
On 31.03.22 14:36, Marek Marczykowski-Górecki wrote:
 
On Thu, Mar 31, 2022 at 02:22:03PM +0200, Juergen Gross wrote:
 
Maybe some kernel config differences, or other udev rules (memory onlining
is done via udev in my guest)?
I'm seeing:
# zgrep MEMORY_HOTPLUG /proc/config.gz
CONFIG_ARCH_ENABLE_MEMORY_HOTPLUG=y
CONFIG_MEMORY_HOTPLUG=y
# CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE is not set
CONFIG_XEN_BALLOON_MEMORY_HOTPLUG=y
CONFIG_XEN_MEMORY_HOTPLUG_LIMIT=512
 
 
I have:
# zgrep MEMORY_HOTPLUG /proc/config.gz
CONFIG_ARCH_ENABLE_MEMORY_HOTPLUG=y
CONFIG_MEMORY_HOTPLUG=y
CONFIG_MEMORY_HOTPLUG_DEFAULT_ONLINE=y
CONFIG_XEN_BALLOON_MEMORY_HOTPLUG=y
CONFIG_XEN_MEMORY_HOTPLUG_LIMIT=512
Not sure if relevant, but I also have:
CONFIG_XEN_UNPOPULATED_ALLOC=y
on top of that, I have a similar udev rule too:
SUBSYSTEM=="memory", ACTION=="add", ATTR{state}=="offline", ATTR{state}="online"
But I don't think they are conflicting.
What type of guest are you using? Mine was a PVH guest.
 
 
PVH here too.
 
 
Would you like to try the attached patch? It seemed to work for me.
 
 
Unfortunately it doesn't help, now the behavior is different:
Initially guest started with 800M:
     [root@personal ~]# free -m
                   total        used        free      shared  buff/cache   
available
     Mem:            740         223         272           2         243        
 401
     Swap:          1023           0        1023
Then increased:
     [root@dom0 ~]$ xl mem-max personal 2048
     [root@dom0 ~]$ xenstore-write /local/domain/$(xl domid 
personal)/memory/static-max $((2048*1024))
     [root@dom0 ~]$ xl mem-set personal 2000
And guest shows now only a little more memory, but not full 2000M:
     [root@personal ~]# [   37.657046] xen:balloon: Populating new zone
     [   37.658206] Fallback order for Node 0: 0
     [   37.658219] Built 1 zonelists, mobility grouping on.  Total pages: 
175889
     [   37.658233] Policy zone: Normal
     [root@personal ~]#
     [root@personal ~]# free -m
                   total        used        free      shared  buff/cache   
available
     Mem:            826         245         337           2         244        
 462
     Swap:          1023           0        1023
I've applied the patch on top of 5.16.18. If you think 5.17 would make a
difference, I can try that too.
 
Hmm, weird.
Can you please post the output of
cat /proc/buddyinfo
cat /proc/iomem
in the guest before and after the operations?
Juergen
 Attachment:
OpenPGP_0xB0DE9DD628BF132F.asc 
Description: OpenPGP public key 
Attachment:
OpenPGP_signature 
Description: OpenPGP digital signature 
 
    
     |