|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [PATCH] x86/mm: Improve bitops in vcpumask_to_pcpumask()
This loop is for_each_set_bit() in disguise.
No functional change.
Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
---
CC: Jan Beulich <JBeulich@xxxxxxxx>
CC: Roger Pau Monné <roger.pau@xxxxxxxxxx>
More detailed analysis:
add/remove: 0/0 grow/shrink: 0/1 up/down: 0/-46 (-46)
Function old new delta
vcpumask_to_pcpumask 519 473 -46
Re-shifting 1 just to clear it is horrible code generation.
While the ffsl() is abtracted away, it doesn't logically move in terms of the
loop position. All that is happening is a (more efficient) clearing is moved
to the post condition, and 3/4 of a cacheline of logic disappears.
---
xen/arch/x86/mm.c | 6 ++----
1 file changed, 2 insertions(+), 4 deletions(-)
diff --git a/xen/arch/x86/mm.c b/xen/arch/x86/mm.c
index 0b787ba55312..66c15a3c864f 100644
--- a/xen/arch/x86/mm.c
+++ b/xen/arch/x86/mm.c
@@ -3411,7 +3411,7 @@ int new_guest_cr3(mfn_t mfn)
static int vcpumask_to_pcpumask(
struct domain *d, XEN_GUEST_HANDLE_PARAM(const_void) bmap, cpumask_t
*pmask)
{
- unsigned int vcpu_id, vcpu_bias, offs;
+ unsigned int vcpu_bias, offs;
unsigned long vmask;
struct vcpu *v;
bool is_native = !is_pv_32bit_domain(d);
@@ -3432,12 +3432,10 @@ static int vcpumask_to_pcpumask(
return -EFAULT;
}
- while ( vmask )
+ for_each_set_bit ( vcpu_id, vmask )
{
unsigned int cpu;
- vcpu_id = ffsl(vmask) - 1;
- vmask &= ~(1UL << vcpu_id);
vcpu_id += vcpu_bias;
if ( (vcpu_id >= d->max_vcpus) )
return 0;
--
2.39.5
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |