WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-ia64-devel

RE: [Xen-ia64-devel] VHPT implementation issues

To: "Arun Sharma" <arun.sharma@xxxxxxxxx>, <Xen-ia64-devel@xxxxxxxxxxxxxxxxxxx>
Subject: RE: [Xen-ia64-devel] VHPT implementation issues
From: "Magenheimer, Dan (HP Labs Fort Collins)" <dan.magenheimer@xxxxxx>
Date: Thu, 2 Jun 2005 07:30:56 -0700
Delivery-date: Thu, 02 Jun 2005 14:30:13 +0000
Envelope-to: www-data@xxxxxxxxxxxxxxxxxxx
List-help: <mailto:xen-ia64-devel-request@lists.xensource.com?subject=help>
List-id: DIscussion of the ia64 port of Xen <xen-ia64-devel.lists.xensource.com>
List-post: <mailto:xen-ia64-devel@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/cgi-bin/mailman/listinfo/xen-ia64-devel>, <mailto:xen-ia64-devel-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/cgi-bin/mailman/listinfo/xen-ia64-devel>, <mailto:xen-ia64-devel-request@lists.xensource.com?subject=unsubscribe>
Sender: xen-ia64-devel-bounces@xxxxxxxxxxxxxxxxxxx
Thread-index: AcVf22ZBW4HrVu9ORjuuNXUCQeZNdgHpB+Lg
Thread-topic: [Xen-ia64-devel] VHPT implementation issues
> We first eliminated 1 and 3 as they have scalability issues on large 
> scale SMP systems.

I forgot to ask at the time...

Could you explain the scalability issues more? 

> -----Original Message-----
> From: xen-ia64-devel-bounces@xxxxxxxxxxxxxxxxxxx 
> [mailto:xen-ia64-devel-bounces@xxxxxxxxxxxxxxxxxxx] On Behalf 
> Of Arun Sharma
> Sent: Monday, May 23, 2005 3:07 PM
> To: Xen-ia64-devel@xxxxxxxxxxxxxxxxxxx
> Subject: [Xen-ia64-devel] VHPT implementation issues
> 
> 
> This email contains minutes of a internal discussion we had 
> at Intel a 
> few weeks ago on the thread about global vs per domain VHPT.
> 
> As someone else on the list suggested, there are really 4 
> options on the 
> table (the original thread dealt primarily with global vs per domain):
> 
> 1. Global
> 2. one per logical processor (logical processor = hardware 
> thread as in 
> SMT/SOEMT)
> 3. one per domain
> 4. one per virtual processor
> 
> Generally speaking, the list is in the ascending order of number of 
> VHPTs (although there are exceptions).
> 
> We first eliminated 1 and 3 as they have scalability issues on large 
> scale SMP systems.
> 
> So it was really a 2 vs 4 and I was initially arguing for 2 
> and against 
> 4. Before we go further, I think I should explain some 
> details about how 
> we propose to implement 4.
> 
> The idea is that we set aside a fixed amount of memory for VHPT 
> purposes. In all of the above algorithms, the VHPT will be sized 
> proportional to the amount of physical memory on the system.
> 
> In the earlier arguments on the thread IIRC, it was argued 
> that since #4 
> has more VHPTs of the same size, it must be consuming more 
> memory. But 
> in our modified proposal above, 2 and 4 consume the same amount of 
> memory, which is set aside at boot time (so no issues with finding 
> contiguous memory for VHPT etc).
> 
> So now the argument comes down to how best to use this memory 
> for VHPT 
> caching purposes. Because this is a cache, it can be thrown 
> away at will 
> without compromising correctness. So if we can't find enough 
> VHPT memory 
>   when we create a domain, we can steal some from another domain.
> 
> The arguments for 4:
> 
> - less interference between domains due to similar access patterns
> - easy to reuse rids
>       - with 2, when a domain terminates and we want to reuse 
> it's RIDs, we 
> need to walk through all entries in the VHPT and clear them. 
> There is no 
> easy way to walk VHPT by RID
>       - with 4, it comes for free.
> 
> The arguments for 2:
> 
> - Potential for better VHPT hit ratio for workloads with 
> large working 
> sets. In some sense, this is like the argument between 1 vs 
> 2. But most 
> people (including Gelato for Linux) seem to choose 2 over 1 
> because SMP 
> locking overhead and cache benefits trump VHPT hit ratio.
> 
> Another way to look at this is, instead of statically 
> partitioning the 
> VHPT memory between domains (or not parititioning it at all 
> as in 1), we 
> can do it more dynamically.
> 
> Does this new proposal address some of the concerns expressed earlier?
> 
>       -Arun
> 
> 
> _______________________________________________
> Xen-ia64-devel mailing list
> Xen-ia64-devel@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-ia64-devel
> 

_______________________________________________
Xen-ia64-devel mailing list
Xen-ia64-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-ia64-devel

<Prev in Thread] Current Thread [Next in Thread>