[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Xen-devel] [PATCH 1 of 4] p2m: Keep statistics on order of p2m entries



On 05/06/11 17:00, Tim Deegan wrote:
Hello,

At 15:40 +0100 on 06 May (1304696401), Christoph Egger wrote:
Can you please use defines for the indexes of p2m->stats.entries[],
please? That makes it easier to read what you are counting.

I'm not sure that's a great idea.  The array is indexed in several
places by integer already, and the defines are unlikely to be much more
illuminating unless you can think of some better naming scheme than
"#define LEVEL_0 0" etc.

What about this:

#define PAGE_ORDER_4K  0
#define PAGE_ORDER_2M  9
#define PAGE_ORDER_1G  18


On the other hand, maybe the array itself could have a more descriptive
name than "stats.entries".

Tim.

On 05/06/11 16:01, George Dunlap wrote:
Count the number of 4kiB, 2MiB, and 1GiB p2m entries.

Signed-off-by: George Dunlap<george.dunlap@xxxxxxxxxxxxx>

diff -r 4b0692880dfa -r be5d93d38f28 xen/arch/x86/mm/hap/p2m-ept.c
--- a/xen/arch/x86/mm/hap/p2m-ept.c     Thu May 05 17:40:34 2011 +0100
+++ b/xen/arch/x86/mm/hap/p2m-ept.c     Fri May 06 15:01:08 2011 +0100
@@ -39,6 +39,8 @@

   #define is_epte_present(ept_entry)      ((ept_entry)->epte&   0x7)
   #define is_epte_superpage(ept_entry)    ((ept_entry)->sp)
+#define is_epte_countable(ept_entry)    (is_epte_present(ept_entry) \
+                                         || ((ept_entry)->sa_p2mt == 
p2m_populate_on_demand))

   /* Non-ept "lock-and-check" wrapper */
   static int ept_pod_check_and_populate(struct p2m_domain *p2m, unsigned long 
gfn,
@@ -167,11 +169,14 @@
   void ept_free_entry(struct p2m_domain *p2m, ept_entry_t *ept_entry, int 
level)
   {
       /* End if the entry is a leaf entry. */
-    if ( level == 0 || !is_epte_present(ept_entry) ||
-         is_epte_superpage(ept_entry) )
+    if ( level == 0 || !is_epte_present(ept_entry) || 
is_epte_superpage(ept_entry) )
+    {
+        if ( is_epte_countable(ept_entry) )
+            p2m->stats.entries[level]--;
           return;
+    }

-    if ( level>   1 )
+    if ( level>   0 )
       {
           ept_entry_t *epte = map_domain_page(ept_entry->mfn);
           for ( int i = 0; i<   EPT_PAGETABLE_ENTRIES; i++ )
@@ -217,7 +222,10 @@
           ept_p2m_type_to_flags(epte, epte->sa_p2mt, epte->access);

           if ( (level - 1) == target )
+        {
+            p2m->stats.entries[target]++;
               continue;
+        }

           ASSERT(is_epte_superpage(epte));

@@ -400,6 +408,10 @@
               ept_p2m_type_to_flags(&new_entry, p2mt, p2ma);
           }

+        /* old_entry will be handled by ept_free_entry below */
+        if ( is_epte_countable(&new_entry) )
+            p2m->stats.entries[i]++;
+
           atomic_write_ept_entry(ept_entry, new_entry);
       }
       else
@@ -412,12 +424,16 @@

           split_ept_entry = atomic_read_ept_entry(ept_entry);

+        /* Accounting should be OK here; split_ept_entry bump the counts,
+         * free_entry will reduce them. */
           if ( !ept_split_super_page(p2m,&split_ept_entry, i, target) )
           {
               ept_free_entry(p2m,&split_ept_entry, i);
               goto out;
           }

+        /* We know this was countable or we wouldn't be here.*/
+        p2m->stats.entries[i]--;
           /* now install the newly split ept sub-tree */
           /* NB: please make sure domian is paused and no in-fly VT-d DMA. */
           atomic_write_ept_entry(ept_entry, split_ept_entry);
@@ -449,9 +465,13 @@

           ept_p2m_type_to_flags(&new_entry, p2mt, p2ma);

+        /* old_entry will be handled by ept_free_entry below */
+        if ( is_epte_countable(&new_entry) )
+            p2m->stats.entries[i]++;
+
           atomic_write_ept_entry(ept_entry, new_entry);
       }
-
+
       /* Track the highest gfn for which we have ever had a valid mapping */
       if ( mfn_valid(mfn_x(mfn))&&
            (gfn + (1UL<<   order) - 1>   p2m->max_mapped_pfn) )
diff -r 4b0692880dfa -r be5d93d38f28 xen/arch/x86/mm/p2m.c
--- a/xen/arch/x86/mm/p2m.c     Thu May 05 17:40:34 2011 +0100
+++ b/xen/arch/x86/mm/p2m.c     Fri May 06 15:01:08 2011 +0100
@@ -184,11 +184,15 @@
   {
       /* End if the entry is a leaf entry. */
       if ( page_order == 0
-         || !(l1e_get_flags(*p2m_entry)&   _PAGE_PRESENT)
+         || !(l1e_get_flags(*p2m_entry)&   _PAGE_PRESENT)
            || (l1e_get_flags(*p2m_entry)&   _PAGE_PSE) )
+    {
+        if ( l1e_get_flags(*p2m_entry) )
+            p2m->stats.entries[page_order/9]--;
           return;
-
-    if ( page_order>   9 )
+    }
+
+    if ( page_order )
       {
           l1_pgentry_t *l3_table = map_domain_page(l1e_get_pfn(*p2m_entry));
           for ( int i = 0; i<   L3_PAGETABLE_ENTRIES; i++ )
@@ -242,6 +246,7 @@
           new_entry = l1e_from_pfn(mfn_x(page_to_mfn(pg)),
                                    __PAGE_HYPERVISOR | _PAGE_USER);

+        /* Stats: Empty entry, no mods needed */
           switch ( type ) {
           case PGT_l3_page_table:
               p2m_add_iommu_flags(&new_entry, 3, 
IOMMUF_readable|IOMMUF_writable);
@@ -285,10 +290,12 @@
           {
               new_entry = l1e_from_pfn(pfn + (i * L1_PAGETABLE_ENTRIES), 
flags);
               p2m_add_iommu_flags(&new_entry, 1, 
IOMMUF_readable|IOMMUF_writable);
+            p2m->stats.entries[1]++;
               p2m->write_p2m_entry(p2m, gfn,
                   l1_entry+i, *table_mfn, new_entry, 2);
           }
           unmap_domain_page(l1_entry);
+        p2m->stats.entries[2]--;
           new_entry = l1e_from_pfn(mfn_x(page_to_mfn(pg)),
                                    __PAGE_HYPERVISOR|_PAGE_USER); //disable PSE
           p2m_add_iommu_flags(&new_entry, 2, IOMMUF_readable|IOMMUF_writable);
@@ -320,6 +327,7 @@
           {
               new_entry = l1e_from_pfn(pfn + i, flags);
               p2m_add_iommu_flags(&new_entry, 0, 0);
+            p2m->stats.entries[0]++;
               p2m->write_p2m_entry(p2m, gfn,
                   l1_entry+i, *table_mfn, new_entry, 1);
           }
@@ -328,6 +336,7 @@
           new_entry = l1e_from_pfn(mfn_x(page_to_mfn(pg)),
                                    __PAGE_HYPERVISOR|_PAGE_USER);
           p2m_add_iommu_flags(&new_entry, 1, IOMMUF_readable|IOMMUF_writable);
+        p2m->stats.entries[1]--;
           p2m->write_p2m_entry(p2m, gfn,
               p2m_entry, *table_mfn, new_entry, 2);
       }
@@ -908,6 +917,15 @@
   void
   p2m_pod_dump_data(struct p2m_domain *p2m)
   {
+    int i;
+    long entries;
+    printk("    P2M entry stats:\n");
+    for ( i=0; i<3; i++)
+        if ( (entries=p2m->stats.entries[i]) )
+            printk("     L%d: %8ld entries, %ld bytes\n",
+                   i+1,
+                   entries,
+                   entries<<(i*9+12));
       printk("    PoD entries=%d cachesize=%d\n",
              p2m->pod.entry_count, p2m->pod.count);
   }
@@ -1475,6 +1493,12 @@
               old_mfn = l1e_get_pfn(*p2m_entry);
           }

+        /* Adjust count for present/not-present entries added */
+        if ( l1e_get_flags(*p2m_entry) )
+            p2m->stats.entries[page_order/9]--;
+        if ( l1e_get_flags(entry_content) )
+            p2m->stats.entries[page_order/9]++;
+
           p2m->write_p2m_entry(p2m, gfn, p2m_entry, table_mfn, entry_content, 
3);
           /* NB: paging_write_p2m_entry() handles tlb flushes properly */

@@ -1519,6 +1543,13 @@
               p2m_add_iommu_flags(&entry_content, 0, iommu_pte_flags);
               old_mfn = l1e_get_pfn(*p2m_entry);
           }
+
+        /* Adjust count for present/not-present entries added */
+        if ( l1e_get_flags(*p2m_entry) )
+            p2m->stats.entries[page_order/9]--;
+        if ( l1e_get_flags(entry_content) )
+            p2m->stats.entries[page_order/9]++;
+
           /* level 1 entry */
           p2m->write_p2m_entry(p2m, gfn, p2m_entry, table_mfn, entry_content, 
1);
           /* NB: paging_write_p2m_entry() handles tlb flushes properly */
@@ -1556,6 +1587,12 @@
               old_mfn = l1e_get_pfn(*p2m_entry);
           }

+        /* Adjust count for present/not-present entries added */
+        if ( l1e_get_flags(*p2m_entry) )
+            p2m->stats.entries[page_order/9]--;
+        if ( l1e_get_flags(entry_content) )
+            p2m->stats.entries[page_order/9]++;
+
           p2m->write_p2m_entry(p2m, gfn, p2m_entry, table_mfn, entry_content, 
2);
           /* NB: paging_write_p2m_entry() handles tlb flushes properly */

@@ -2750,6 +2787,8 @@
                   continue;
               }

+            /* STATS: Should change only type; no stats should need adjustment 
*/
+
               l2mfn = _mfn(l3e_get_pfn(l3e[i3]));
               l2e = map_domain_page(l3e_get_pfn(l3e[i3]));
               for ( i2 = 0; i2<   L2_PAGETABLE_ENTRIES; i2++ )
diff -r 4b0692880dfa -r be5d93d38f28 xen/include/asm-x86/p2m.h
--- a/xen/include/asm-x86/p2m.h Thu May 05 17:40:34 2011 +0100
+++ b/xen/include/asm-x86/p2m.h Fri May 06 15:01:08 2011 +0100
@@ -278,6 +278,10 @@
           unsigned         reclaim_single; /* Last gpfn of a scan */
           unsigned         max_guest;    /* gpfn of max guest demand-populate 
*/
       } pod;
+
+    struct {
+        long entries[3];
+    } stats;
   };

   /* get host p2m table */


--
---to satisfy European Law for business letters:
Advanced Micro Devices GmbH
Einsteinring 24, 85689 Dornach b. Muenchen
Geschaeftsfuehrer: Alberto Bozzo, Andrew Bowd
Sitz: Dornach, Gemeinde Aschheim, Landkreis Muenchen
Registergericht Muenchen, HRB Nr. 43632


_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel


 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.