|
|
|
|
|
|
|
|
|
|
xen-devel
Re: [Xen-devel] OOM problems
> What do the guests use for storage? (e.g. "blktap2 for VHD files on
an iscsi mounted ext3 volume")
Simple sparse .img files on a local ext4 RAID volume, using "file:".
> It might be worth looking at /proc/kernel/slabinfo to see if there's
anything suspicious.
I didn't see anything suspicious in there, but I'm not sure what I'm
looking for.
Here is the first page of slabtop as it currently stands, if that helps.
It looks a bit easier to read.
Active / Total Objects (% used) : 274753 / 507903 (54.1%)
Active / Total Slabs (% used) : 27573 / 27582 (100.0%)
Active / Total Caches (% used) : 85 / 160 (53.1%)
Active / Total Size (% used) : 75385.52K / 107127.41K (70.4%)
Minimum / Average / Maximum Object : 0.02K / 0.21K / 4096.00K
OBJS ACTIVE USE OBJ SIZE SLABS OBJ/SLAB CACHE SIZE NAME
306397 110621 36% 0.10K 8281 37 33124K buffer_head
37324 26606 71% 0.54K 5332 7 21328K radix_tree_node
25640 25517 99% 0.19K 1282 20 5128K size-192
23472 23155 98% 0.08K 489 48 1956K sysfs_dir_cache
19964 19186 96% 0.95K 4991 4 19964K ext4_inode_cache
17860 13026 72% 0.19K 893 20 3572K dentry
14896 13057 87% 0.03K 133 112 532K size-32
8316 6171 74% 0.17K 378 22 1512K vm_area_struct
8142 5053 62% 0.06K 138 59 552K size-64
4320 3389 78% 0.12K 144 30 576K size-128
3760 2226 59% 0.19K 188 20 752K filp
3456 1875 54% 0.02K 24 144 96K anon_vma
3380 3001 88% 1.00K 845 4 3380K size-1024
3380 3365 99% 0.76K 676 5 2704K shmem_inode_cache
2736 2484 90% 0.50K 342 8 1368K size-512
2597 2507 96% 0.07K 49 53 196K Acpi-Operand
2100 1095 52% 0.25K 140 15 560K skbuff_head_cache
1920 819 42% 0.12K 64 30 256K cred_jar
1361 1356 99% 4.00K 1361 1 5444K size-4096
1230 628 51% 0.12K 41 30 164K pid
1008 907 89% 0.03K 9 112 36K Acpi-Namespace
959 496 51% 0.57K 137 7 548K inode_cache
891 554 62% 0.81K 99 9 792K signal_cache
888 115 12% 0.10K 24 37 96K ext4_prealloc_space
885 122 13% 0.06K 15 59 60K fs_cache
850 642 75% 1.45K 170 5 1360K task_struct
820 769 93% 0.19K 41 20 164K bio-0
666 550 82% 2.06K 222 3 1776K sighand_cache
576 211 36% 0.50K 72 8 288K task_xstate
529 379 71% 0.16K 23 23 92K cfq_queue
518 472 91% 2.00K 259 2 1036K size-2048
506 375 74% 0.16K 22 23 88K cfq_io_context
495 353 71% 0.33K 45 11 180K blkdev_requests
465 422 90% 0.25K 31 15 124K size-256
418 123 29% 0.69K 38 11 304K files_cache
360 207 57% 0.69K 72 5 288K sock_inode_cache
360 251 69% 0.12K 12 30 48K scsi_sense_cache
336 115 34% 0.08K 7 48 28K blkdev_ioc
285 236 82% 0.25K 19 15 76K scsi_cmd_cache
> BTW: 24 vCPUs in dom0 seems a excessive, especially if you're using
stubdoms. You may get better performance by dropping that to e.g. 2 or 3.
I will test that. Do you think it will make a difference in this case?
-John
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|
|
|
|
|