[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [RFC PATCH 3/3] misra: address rule 5.5 bitops
Signed-off-by: Dmytro Prokopchuk <dmytro_prokopchuk1@xxxxxxxx> --- xen/arch/arm/dm.c | 2 +- xen/arch/arm/domctl.c | 2 +- xen/arch/arm/gic-vgic.c | 26 +++++++++++----------- xen/arch/arm/gic.c | 8 +++---- xen/arch/arm/include/asm/cpufeature.h | 2 +- xen/arch/arm/include/asm/gic.h | 2 +- xen/arch/arm/include/asm/guest_atomics.h | 4 ++-- xen/arch/arm/irq.c | 16 ++++++------- xen/arch/arm/tee/ffa_partinfo.c | 2 +- xen/arch/arm/tee/ffa_private.h | 2 +- xen/arch/arm/traps.c | 2 +- xen/arch/arm/vgic-v3-its.c | 6 ++--- xen/arch/arm/vgic.c | 10 ++++----- xen/arch/arm/vgic/vgic-mmio.c | 2 +- xen/arch/arm/vgic/vgic.c | 2 +- xen/arch/arm/vpsci.c | 4 ++-- xen/arch/riscv/cpufeature.c | 2 +- xen/arch/riscv/include/asm/guest_atomics.h | 2 +- xen/arch/x86/cpu-policy.c | 14 ++++++------ xen/arch/x86/cpu/amd.c | 2 +- xen/arch/x86/cpu/common.c | 14 ++++++------ xen/arch/x86/cpu/mcheck/x86_mca.h | 2 +- xen/arch/x86/domain.c | 2 +- xen/arch/x86/e820.c | 4 ++-- xen/arch/x86/guest/xen/xen.c | 2 +- xen/arch/x86/hvm/emulate.c | 2 +- xen/arch/x86/hvm/hpet.c | 6 ++--- xen/arch/x86/hvm/irq.c | 12 +++++----- xen/arch/x86/hvm/svm/nestedsvm.c | 10 ++++----- xen/arch/x86/hvm/svm/svm.c | 2 +- xen/arch/x86/hvm/viridian/time.c | 4 ++-- xen/arch/x86/hvm/vlapic.c | 2 +- xen/arch/x86/hvm/vmx/vmcs.c | 8 +++---- xen/arch/x86/hvm/vmx/vmx.c | 4 ++-- xen/arch/x86/include/asm/guest_atomics.h | 2 +- xen/arch/x86/include/asm/hvm/vlapic.h | 2 +- xen/arch/x86/include/asm/hvm/vmx/vmx.h | 2 +- xen/arch/x86/include/asm/mpspec.h | 2 +- xen/arch/x86/irq.c | 26 +++++++++++----------- xen/arch/x86/mm.c | 8 +++---- xen/arch/x86/mm/hap/hap.c | 2 +- xen/arch/x86/mm/paging.c | 4 ++-- xen/arch/x86/mm/shadow/common.c | 2 +- xen/arch/x86/mm/shadow/hvm.c | 4 ++-- xen/arch/x86/monitor.c | 4 ++-- xen/arch/x86/msi.c | 2 +- xen/arch/x86/psr.c | 2 +- xen/arch/x86/pv/dom0_build.c | 4 ++-- xen/arch/x86/pv/emul-priv-op.c | 2 +- xen/arch/x86/pv/shim.c | 2 +- xen/arch/x86/traps.c | 2 +- xen/arch/x86/x86_64/mm.c | 2 +- xen/arch/x86/x86_64/mmconfig_64.c | 2 +- xen/arch/x86/xstate.c | 4 ++-- xen/common/domain.c | 4 ++-- xen/common/efi/runtime.c | 2 +- xen/common/event_2l.c | 2 +- xen/common/kexec.c | 16 ++++++------- xen/common/keyhandler.c | 2 +- xen/common/multicall.c | 2 +- xen/common/numa.c | 4 ++-- xen/common/page_alloc.c | 4 ++-- xen/common/pdx.c | 2 +- xen/common/sched/core.c | 12 +++++----- xen/common/sched/credit.c | 12 +++++----- xen/common/sched/credit2.c | 6 ++--- xen/common/sched/rt.c | 4 ++-- xen/common/vmap.c | 12 +++++----- xen/drivers/passthrough/iommu.c | 2 +- xen/drivers/passthrough/vtd/dmar.c | 4 ++-- xen/drivers/passthrough/vtd/iommu.c | 4 ++-- xen/drivers/passthrough/x86/hvm.c | 2 +- xen/drivers/passthrough/x86/iommu.c | 2 +- xen/drivers/vpci/vpci.c | 4 ++-- xen/include/xen/bitops.h | 8 +++---- xen/include/xen/cpumask.h | 6 ++--- xen/include/xen/nodemask.h | 2 +- xen/include/xen/sched.h | 4 ++-- 78 files changed, 194 insertions(+), 194 deletions(-) diff --git a/xen/arch/arm/dm.c b/xen/arch/arm/dm.c index fdb3d967ec..928a51aac2 100644 --- a/xen/arch/arm/dm.c +++ b/xen/arch/arm/dm.c @@ -101,7 +101,7 @@ int dm_op(const struct dmop_args *op_args) * Allow to set the logical level of a line for non-allocated * interrupts only. */ - if ( test_bit(data->irq, d->arch.vgic.allocated_irqs) ) + if ( TEST_BIT(data->irq, d->arch.vgic.allocated_irqs) ) { rc = -EINVAL; break; diff --git a/xen/arch/arm/domctl.c b/xen/arch/arm/domctl.c index ad914c915f..8cb0e41483 100644 --- a/xen/arch/arm/domctl.c +++ b/xen/arch/arm/domctl.c @@ -196,7 +196,7 @@ void arch_get_info_guest(struct vcpu *v, vcpu_guest_context_u c) ctxt->ttbr1 = v->arch.ttbr1; ctxt->ttbcr = v->arch.ttbcr; - if ( !test_bit(_VPF_down, &v->pause_flags) ) + if ( !TEST_BIT(_VPF_down, &v->pause_flags) ) ctxt->flags |= VGCF_online; } diff --git a/xen/arch/arm/gic-vgic.c b/xen/arch/arm/gic-vgic.c index ea48c5375a..2b9b9b7752 100644 --- a/xen/arch/arm/gic-vgic.c +++ b/xen/arch/arm/gic-vgic.c @@ -76,7 +76,7 @@ void gic_raise_inflight_irq(struct vcpu *v, unsigned int virtual_irq) ASSERT(spin_is_locked(&v->arch.vgic.lock)); /* Don't try to update the LR if the interrupt is disabled */ - if ( !test_bit(GIC_IRQ_GUEST_ENABLED, &n->status) ) + if ( !TEST_BIT(GIC_IRQ_GUEST_ENABLED, &n->status) ) return; if ( list_empty(&n->lr_queue) ) @@ -106,7 +106,7 @@ static unsigned int gic_find_unused_lr(struct vcpu *v, ASSERT(spin_is_locked(&v->arch.vgic.lock)); - if ( unlikely(test_bit(GIC_IRQ_GUEST_PRISTINE_LPI, &p->status)) ) + if ( unlikely(TEST_BIT(GIC_IRQ_GUEST_PRISTINE_LPI, &p->status)) ) { for_each_set_bit ( used_lr, *lr_mask ) { @@ -182,7 +182,7 @@ static void gic_update_one_lr(struct vcpu *v, int i) if ( lr_val.active ) { set_bit(GIC_IRQ_GUEST_ACTIVE, &p->status); - if ( test_bit(GIC_IRQ_GUEST_ENABLED, &p->status) && + if ( TEST_BIT(GIC_IRQ_GUEST_ENABLED, &p->status) && test_and_clear_bit(GIC_IRQ_GUEST_QUEUED, &p->status) ) { if ( p->desc == NULL ) @@ -216,9 +216,9 @@ static void gic_update_one_lr(struct vcpu *v, int i) clear_bit(GIC_IRQ_GUEST_VISIBLE, &p->status); clear_bit(GIC_IRQ_GUEST_ACTIVE, &p->status); p->lr = GIC_INVALID_LR; - if ( test_bit(GIC_IRQ_GUEST_ENABLED, &p->status) && - test_bit(GIC_IRQ_GUEST_QUEUED, &p->status) && - !test_bit(GIC_IRQ_GUEST_MIGRATING, &p->status) ) + if ( TEST_BIT(GIC_IRQ_GUEST_ENABLED, &p->status) && + TEST_BIT(GIC_IRQ_GUEST_QUEUED, &p->status) && + !TEST_BIT(GIC_IRQ_GUEST_MIGRATING, &p->status) ) gic_raise_guest_irq(v, irq, p->priority); else { list_del_init(&p->inflight); @@ -229,7 +229,7 @@ static void gic_update_one_lr(struct vcpu *v, int i) * accesses to inflight. */ smp_wmb(); - if ( test_bit(GIC_IRQ_GUEST_MIGRATING, &p->status) ) + if ( TEST_BIT(GIC_IRQ_GUEST_MIGRATING, &p->status) ) { struct vcpu *v_target = vgic_get_target_vcpu(v, irq); irq_set_affinity(p->desc, cpumask_of(v_target->processor)); @@ -290,8 +290,8 @@ static void gic_restore_pending_irqs(struct vcpu *v) { if ( p_r->priority == p->priority ) goto out; - if ( test_bit(GIC_IRQ_GUEST_VISIBLE, &p_r->status) && - !test_bit(GIC_IRQ_GUEST_ACTIVE, &p_r->status) ) + if ( TEST_BIT(GIC_IRQ_GUEST_VISIBLE, &p_r->status) && + !TEST_BIT(GIC_IRQ_GUEST_ACTIVE, &p_r->status) ) goto found; } /* We didn't find a victim this time, and we won't next @@ -370,7 +370,7 @@ int vgic_vcpu_pending_irq(struct vcpu *v) goto out; if ( GIC_PRI_TO_GUEST(p->priority) >= active_priority ) goto out; - if ( test_bit(GIC_IRQ_GUEST_ENABLED, &p->status) ) + if ( TEST_BIT(GIC_IRQ_GUEST_ENABLED, &p->status) ) { rc = 1; goto out; @@ -446,9 +446,9 @@ int vgic_connect_hw_irq(struct domain *d, struct vcpu *v, unsigned int virq, * active/pending in the guest. */ if ( !p->desc && - !test_bit(GIC_IRQ_GUEST_ENABLED, &p->status) && - !test_bit(GIC_IRQ_GUEST_VISIBLE, &p->status) && - !test_bit(GIC_IRQ_GUEST_ACTIVE, &p->status) ) + !TEST_BIT(GIC_IRQ_GUEST_ENABLED, &p->status) && + !TEST_BIT(GIC_IRQ_GUEST_VISIBLE, &p->status) && + !TEST_BIT(GIC_IRQ_GUEST_ACTIVE, &p->status) ) p->desc = desc; else ret = -EBUSY; diff --git a/xen/arch/arm/gic.c b/xen/arch/arm/gic.c index e80fe0ca24..f7e377a79c 100644 --- a/xen/arch/arm/gic.c +++ b/xen/arch/arm/gic.c @@ -93,7 +93,7 @@ void gic_set_irq_type(struct irq_desc *desc, unsigned int type) * IRQ must be disabled before configuring it (see 4.3.13 in ARM IHI * 0048B.b). We rely on the caller to do it. */ - ASSERT(test_bit(_IRQ_DISABLED, &desc->status)); + ASSERT(TEST_BIT(_IRQ_DISABLED, &desc->status)); ASSERT(spin_is_locked(&desc->lock)); ASSERT(type != IRQ_TYPE_INVALID); @@ -112,7 +112,7 @@ void gic_route_irq_to_xen(struct irq_desc *desc, unsigned int priority) { ASSERT(priority <= 0xff); /* Only 8 bits of priority */ ASSERT(desc->irq < gic_number_lines());/* Can't route interrupts that don't exist */ - ASSERT(test_bit(_IRQ_DISABLED, &desc->status)); + ASSERT(TEST_BIT(_IRQ_DISABLED, &desc->status)); ASSERT(spin_is_locked(&desc->lock)); desc->handler = gic_hw_ops->gic_host_irq_type; @@ -158,7 +158,7 @@ int gic_remove_irq_from_guest(struct domain *d, unsigned int virq, int ret; ASSERT(spin_is_locked(&desc->lock)); - ASSERT(test_bit(_IRQ_GUEST, &desc->status)); + ASSERT(TEST_BIT(_IRQ_GUEST, &desc->status)); ASSERT(!is_lpi(virq)); /* @@ -171,7 +171,7 @@ int gic_remove_irq_from_guest(struct domain *d, unsigned int virq, desc->handler->shutdown(desc); /* EOI the IRQ if it has not been done by the guest */ - if ( test_bit(_IRQ_INPROGRESS, &desc->status) ) + if ( TEST_BIT(_IRQ_INPROGRESS, &desc->status) ) gic_hw_ops->deactivate_irq(desc); clear_bit(_IRQ_INPROGRESS, &desc->status); diff --git a/xen/arch/arm/include/asm/cpufeature.h b/xen/arch/arm/include/asm/cpufeature.h index b6df188011..06c3e5bd93 100644 --- a/xen/arch/arm/include/asm/cpufeature.h +++ b/xen/arch/arm/include/asm/cpufeature.h @@ -95,7 +95,7 @@ static inline bool cpus_have_cap(unsigned int num) if ( num >= ARM_NCAPS ) return false; - return test_bit(num, cpu_hwcaps); + return TEST_BIT(num, cpu_hwcaps); } /* System capability check for constant cap */ diff --git a/xen/arch/arm/include/asm/gic.h b/xen/arch/arm/include/asm/gic.h index 541f0eeb80..366c0ebcb4 100644 --- a/xen/arch/arm/include/asm/gic.h +++ b/xen/arch/arm/include/asm/gic.h @@ -415,7 +415,7 @@ static inline unsigned int gic_get_nr_lrs(void) */ static inline void gic_set_active_state(struct irq_desc *irqd, bool state) { - ASSERT(test_bit(_IRQ_GUEST, &irqd->status)); + ASSERT(TEST_BIT(_IRQ_GUEST, &irqd->status)); gic_hw_ops->set_active_state(irqd, state); } diff --git a/xen/arch/arm/include/asm/guest_atomics.h b/xen/arch/arm/include/asm/guest_atomics.h index 8893eb9a55..681f13c336 100644 --- a/xen/arch/arm/include/asm/guest_atomics.h +++ b/xen/arch/arm/include/asm/guest_atomics.h @@ -64,8 +64,8 @@ guest_bitop(change_bit) #undef guest_bitop -/* test_bit does not use load-store atomic operations */ -#define guest_test_bit(d, nr, p) ((void)(d), test_bit(nr, p)) +/* TEST_BIT does not use load-store atomic operations */ +#define guest_test_bit(d, nr, p) ((void)(d), TEST_BIT(nr, p)) guest_testop(test_and_set_bit) guest_testop(test_and_clear_bit) diff --git a/xen/arch/arm/irq.c b/xen/arch/arm/irq.c index 03fbb90c6c..177644f936 100644 --- a/xen/arch/arm/irq.c +++ b/xen/arch/arm/irq.c @@ -160,7 +160,7 @@ void __init init_IRQ(void) static inline struct irq_guest *irq_get_guest_info(struct irq_desc *desc) { ASSERT(spin_is_locked(&desc->lock)); - ASSERT(test_bit(_IRQ_GUEST, &desc->status)); + ASSERT(TEST_BIT(_IRQ_GUEST, &desc->status)); ASSERT(desc->action != NULL); return desc->action->dev_id; @@ -246,7 +246,7 @@ void do_IRQ(struct cpu_user_regs *regs, unsigned int irq, int is_fiq) } #endif - if ( test_bit(_IRQ_GUEST, &desc->status) ) + if ( TEST_BIT(_IRQ_GUEST, &desc->status) ) { struct irq_guest *info = irq_get_guest_info(desc); @@ -264,7 +264,7 @@ void do_IRQ(struct cpu_user_regs *regs, unsigned int irq, int is_fiq) goto out_no_end; } - if ( test_bit(_IRQ_DISABLED, &desc->status) ) + if ( TEST_BIT(_IRQ_DISABLED, &desc->status) ) goto out; set_bit(_IRQ_INPROGRESS, &desc->status); @@ -331,7 +331,7 @@ void release_irq(unsigned int irq, const void *dev_id) spin_unlock_irqrestore(&desc->lock,flags); /* Wait to make sure it's not being used on another CPU */ - do { smp_mb(); } while ( test_bit(_IRQ_INPROGRESS, &desc->status) ); + do { smp_mb(); } while ( TEST_BIT(_IRQ_INPROGRESS, &desc->status) ); if ( action->free_on_release ) xfree(action); @@ -348,7 +348,7 @@ static int __setup_irq(struct irq_desc *desc, unsigned int irqflags, * - if the IRQ is marked as shared * - dev_id is not NULL when IRQF_SHARED is set */ - if ( desc->action != NULL && (!test_bit(_IRQF_SHARED, &desc->status) || !shared) ) + if ( desc->action != NULL && (!TEST_BIT(_IRQF_SHARED, &desc->status) || !shared) ) return -EINVAL; if ( shared && new->dev_id == NULL ) return -EINVAL; @@ -375,7 +375,7 @@ int setup_irq(unsigned int irq, unsigned int irqflags, struct irqaction *new) spin_lock_irqsave(&desc->lock, flags); - if ( test_bit(_IRQ_GUEST, &desc->status) ) + if ( TEST_BIT(_IRQ_GUEST, &desc->status) ) { struct domain *d = irq_get_domain(desc); @@ -501,7 +501,7 @@ int route_irq_to_guest(struct domain *d, unsigned int virq, */ if ( desc->action != NULL ) { - if ( test_bit(_IRQ_GUEST, &desc->status) ) + if ( TEST_BIT(_IRQ_GUEST, &desc->status) ) { struct domain *ad = irq_get_domain(desc); @@ -570,7 +570,7 @@ int release_guest_irq(struct domain *d, unsigned int virq) spin_lock_irqsave(&desc->lock, flags); ret = -EINVAL; - if ( !test_bit(_IRQ_GUEST, &desc->status) ) + if ( !TEST_BIT(_IRQ_GUEST, &desc->status) ) goto unlock; info = irq_get_guest_info(desc); diff --git a/xen/arch/arm/tee/ffa_partinfo.c b/xen/arch/arm/tee/ffa_partinfo.c index c0510ceb83..3479d6a7e4 100644 --- a/xen/arch/arm/tee/ffa_partinfo.c +++ b/xen/arch/arm/tee/ffa_partinfo.c @@ -445,7 +445,7 @@ bool ffa_partinfo_domain_destroy(struct domain *d) for ( n = 0; n < subscr_vm_destroyed_count; n++ ) { - if ( !test_bit(n, ctx->vm_destroy_bitmap) ) + if ( !TEST_BIT(n, ctx->vm_destroy_bitmap) ) continue; res = ffa_direct_req_send_vm(subscr_vm_destroyed[n], ffa_get_vm_id(d), diff --git a/xen/arch/arm/tee/ffa_private.h b/xen/arch/arm/tee/ffa_private.h index c4cd655389..57efa76b06 100644 --- a/xen/arch/arm/tee/ffa_private.h +++ b/xen/arch/arm/tee/ffa_private.h @@ -447,7 +447,7 @@ static inline bool ffa_fw_supports_fid(uint32_t fid) if ( FFA_ABI_BITNUM(fid) > FFA_ABI_BITMAP_SIZE) return false; - return test_bit(FFA_ABI_BITNUM(fid), ffa_fw_abi_supported); + return TEST_BIT(FFA_ABI_BITNUM(fid), ffa_fw_abi_supported); } #endif /*__FFA_PRIVATE_H__*/ diff --git a/xen/arch/arm/traps.c b/xen/arch/arm/traps.c index 445e7378dd..bab7c82393 100644 --- a/xen/arch/arm/traps.c +++ b/xen/arch/arm/traps.c @@ -998,7 +998,7 @@ static void show_guest_stack(struct vcpu *v, const struct cpu_user_regs *regs) void *mapped; unsigned long *stack, addr; - if ( test_bit(_VPF_down, &v->pause_flags) ) + if ( TEST_BIT(_VPF_down, &v->pause_flags) ) { printk("No stack trace, VCPU offline\n"); return; diff --git a/xen/arch/arm/vgic-v3-its.c b/xen/arch/arm/vgic-v3-its.c index bc738614bb..284459140b 100644 --- a/xen/arch/arm/vgic-v3-its.c +++ b/xen/arch/arm/vgic-v3-its.c @@ -370,7 +370,7 @@ static int its_handle_clear(struct virt_its *its, uint64_t *cmdptr) * no effect on that LPI anymore. Since LPIs are edge triggered and * have no active state, we don't need to care about this here. */ - if ( !test_bit(GIC_IRQ_GUEST_VISIBLE, &p->status) ) + if ( !TEST_BIT(GIC_IRQ_GUEST_VISIBLE, &p->status) ) vgic_remove_irq_from_queues(vcpu, p); spin_unlock_irqrestore(&vcpu->arch.vgic.lock, flags); @@ -432,10 +432,10 @@ static void update_lpi_vgic_status(struct vcpu *v, struct pending_irq *p) { ASSERT(spin_is_locked(&v->arch.vgic.lock)); - if ( test_bit(GIC_IRQ_GUEST_ENABLED, &p->status) ) + if ( TEST_BIT(GIC_IRQ_GUEST_ENABLED, &p->status) ) { if ( !list_empty(&p->inflight) && - !test_bit(GIC_IRQ_GUEST_VISIBLE, &p->status) ) + !TEST_BIT(GIC_IRQ_GUEST_VISIBLE, &p->status) ) gic_raise_guest_irq(v, p->irq, p->lpi_priority); } else diff --git a/xen/arch/arm/vgic.c b/xen/arch/arm/vgic.c index c563ba93af..2ed6c6e046 100644 --- a/xen/arch/arm/vgic.c +++ b/xen/arch/arm/vgic.c @@ -265,7 +265,7 @@ bool vgic_migrate_irq(struct vcpu *old, struct vcpu *new, unsigned int irq) } /* migration already in progress, no need to do anything */ - if ( test_bit(GIC_IRQ_GUEST_MIGRATING, &p->status) ) + if ( TEST_BIT(GIC_IRQ_GUEST_MIGRATING, &p->status) ) { gprintk(XENLOG_WARNING, "irq %u migration failed: requested while in progress\n", irq); spin_unlock_irqrestore(&old->arch.vgic.lock, flags); @@ -322,7 +322,7 @@ void arch_move_irqs(struct vcpu *v) v_target = vgic_get_target_vcpu(v, i); p = irq_to_pending(v_target, i); - if ( v_target == v && !test_bit(GIC_IRQ_GUEST_MIGRATING, &p->status) ) + if ( v_target == v && !TEST_BIT(GIC_IRQ_GUEST_MIGRATING, &p->status) ) irq_set_affinity(p->desc, cpu_mask); } } @@ -398,7 +398,7 @@ void vgic_enable_irqs(struct vcpu *v, uint32_t r, unsigned int n) spin_lock_irqsave(&v_target->arch.vgic.lock, flags); p = irq_to_pending(v_target, irq); set_bit(GIC_IRQ_GUEST_ENABLED, &p->status); - if ( !list_empty(&p->inflight) && !test_bit(GIC_IRQ_GUEST_VISIBLE, &p->status) ) + if ( !list_empty(&p->inflight) && !TEST_BIT(GIC_IRQ_GUEST_VISIBLE, &p->status) ) gic_raise_guest_irq(v_target, irq, p->priority); spin_unlock_irqrestore(&v_target->arch.vgic.lock, flags); if ( p->desc != NULL ) @@ -598,7 +598,7 @@ void vgic_inject_irq(struct domain *d, struct vcpu *v, unsigned int virq, } /* vcpu offline */ - if ( test_bit(_VPF_down, &v->pause_flags) ) + if ( TEST_BIT(_VPF_down, &v->pause_flags) ) { spin_unlock_irqrestore(&v->arch.vgic.lock, flags); return; @@ -616,7 +616,7 @@ void vgic_inject_irq(struct domain *d, struct vcpu *v, unsigned int virq, n->priority = priority; /* the irq is enabled */ - if ( test_bit(GIC_IRQ_GUEST_ENABLED, &n->status) ) + if ( TEST_BIT(GIC_IRQ_GUEST_ENABLED, &n->status) ) gic_raise_guest_irq(v, virq, priority); list_for_each_entry ( iter, &v->arch.vgic.inflight_irqs, inflight ) diff --git a/xen/arch/arm/vgic/vgic-mmio.c b/xen/arch/arm/vgic/vgic-mmio.c index d2bcb32cd7..a8da401ba3 100644 --- a/xen/arch/arm/vgic/vgic-mmio.c +++ b/xen/arch/arm/vgic/vgic-mmio.c @@ -463,7 +463,7 @@ void vgic_mmio_write_config(struct vcpu *vcpu, irq = vgic_get_irq(vcpu->domain, vcpu, intid + i); spin_lock_irqsave(&irq->irq_lock, flags); - if ( test_bit(i * 2 + 1, &val) ) + if ( TEST_BIT(i * 2 + 1, &val) ) irq->config = VGIC_CONFIG_EDGE; else irq->config = VGIC_CONFIG_LEVEL; diff --git a/xen/arch/arm/vgic/vgic.c b/xen/arch/arm/vgic/vgic.c index 6cabd0496d..3bd78223bc 100644 --- a/xen/arch/arm/vgic/vgic.c +++ b/xen/arch/arm/vgic/vgic.c @@ -933,7 +933,7 @@ void vgic_sync_hardware_irq(struct domain *d, * interrrupt is disabled before trying to change the config. */ if ( irq_type_set_by_domain(d) && - test_bit(_IRQ_DISABLED, &desc->status) ) + TEST_BIT(_IRQ_DISABLED, &desc->status) ) gic_set_irq_type(desc, translate_irq_type(irq->config)); if ( irq->target_vcpu ) diff --git a/xen/arch/arm/vpsci.c b/xen/arch/arm/vpsci.c index 7ba9ccd94b..74a36856ea 100644 --- a/xen/arch/arm/vpsci.c +++ b/xen/arch/arm/vpsci.c @@ -29,7 +29,7 @@ static int do_common_cpu_on(register_t target_cpu, register_t entry_point, if ( is_64bit_domain(d) && is_thumb ) return PSCI_INVALID_ADDRESS; - if ( !test_bit(_VPF_down, &v->pause_flags) ) + if ( !TEST_BIT(_VPF_down, &v->pause_flags) ) return PSCI_ALREADY_ON; if ( (ctxt = alloc_vcpu_guest_context()) == NULL ) @@ -173,7 +173,7 @@ static int32_t do_psci_0_2_affinity_info(register_t target_affinity, v = d->vcpu[vcpuid]; if ( ( ( v->arch.vmpidr & tmask ) == target_affinity ) - && ( !test_bit(_VPF_down, &v->pause_flags) ) ) + && ( !TEST_BIT(_VPF_down, &v->pause_flags) ) ) return PSCI_0_2_AFFINITY_LEVEL_ON; } diff --git a/xen/arch/riscv/cpufeature.c b/xen/arch/riscv/cpufeature.c index b846a106a3..92d8e8a85e 100644 --- a/xen/arch/riscv/cpufeature.c +++ b/xen/arch/riscv/cpufeature.c @@ -473,7 +473,7 @@ bool riscv_isa_extension_available(const unsigned long *isa_bitmap, if ( id >= RISCV_ISA_EXT_MAX ) return false; - return test_bit(id, isa_bitmap); + return TEST_BIT(id, isa_bitmap); } void __init riscv_fill_hwcap(void) diff --git a/xen/arch/riscv/include/asm/guest_atomics.h b/xen/arch/riscv/include/asm/guest_atomics.h index 22a7551804..c1b605bfc5 100644 --- a/xen/arch/riscv/include/asm/guest_atomics.h +++ b/xen/arch/riscv/include/asm/guest_atomics.h @@ -30,7 +30,7 @@ guest_testop(test_and_change_bit) #undef guest_testop -#define guest_test_bit(d, nr, p) ((void)(d), test_bit(nr, p)) +#define guest_test_bit(d, nr, p) ((void)(d), TEST_BIT(nr, p)) #endif /* ASM__RISCV__GUEST_ATOMICS_H */ diff --git a/xen/arch/x86/cpu-policy.c b/xen/arch/x86/cpu-policy.c index 42dccdce52..dc43cdca30 100644 --- a/xen/arch/x86/cpu-policy.c +++ b/xen/arch/x86/cpu-policy.c @@ -476,7 +476,7 @@ static void __init guest_common_max_feature_adjustments(uint32_t *fs) * * If RTM-capable, we can run a VM which has seen RTM_ALWAYS_ABORT. */ - if ( test_bit(X86_FEATURE_RTM, fs) ) + if ( TEST_BIT(X86_FEATURE_RTM, fs) ) __set_bit(X86_FEATURE_RTM_ALWAYS_ABORT, fs); break; @@ -613,9 +613,9 @@ static void __init guest_common_feature_adjustments(uint32_t *fs) * nop on non-HT hardware, and has this behaviour to make heterogeneous * setups easier to manage. */ - if ( test_bit(X86_FEATURE_IBRSB, fs) ) + if ( TEST_BIT(X86_FEATURE_IBRSB, fs) ) __set_bit(X86_FEATURE_STIBP, fs); - if ( test_bit(X86_FEATURE_IBRS, fs) ) + if ( TEST_BIT(X86_FEATURE_IBRS, fs) ) __set_bit(X86_FEATURE_AMD_STIBP, fs); /* @@ -715,10 +715,10 @@ static void __init calculate_pv_def_policy(void) * MSR_ARCH_CAPS, reflect the appropriate RSBA/RRSBA property to the guest * depending on the visibility of eIBRS. */ - if ( test_bit(X86_FEATURE_ARCH_CAPS, fs) && + if ( TEST_BIT(X86_FEATURE_ARCH_CAPS, fs) && (cpu_has_rsba || cpu_has_rrsba) ) { - bool eibrs = test_bit(X86_FEATURE_EIBRS, fs); + bool eibrs = TEST_BIT(X86_FEATURE_EIBRS, fs); __set_bit(eibrs ? X86_FEATURE_RRSBA : X86_FEATURE_RSBA, fs); @@ -884,10 +884,10 @@ static void __init calculate_hvm_def_policy(void) * MSR_ARCH_CAPS, reflect the appropriate RSBA/RRSBA property to the guest * depending on the visibility of eIBRS. */ - if ( test_bit(X86_FEATURE_ARCH_CAPS, fs) && + if ( TEST_BIT(X86_FEATURE_ARCH_CAPS, fs) && (cpu_has_rsba || cpu_has_rrsba) ) { - bool eibrs = test_bit(X86_FEATURE_EIBRS, fs); + bool eibrs = TEST_BIT(X86_FEATURE_EIBRS, fs); __set_bit(eibrs ? X86_FEATURE_RRSBA : X86_FEATURE_RSBA, fs); diff --git a/xen/arch/x86/cpu/amd.c b/xen/arch/x86/cpu/amd.c index eb428f284e..3c3892eff4 100644 --- a/xen/arch/x86/cpu/amd.c +++ b/xen/arch/x86/cpu/amd.c @@ -755,7 +755,7 @@ void amd_init_lfence(struct cpuinfo_x86 *c) * Some hardware has LFENCE dispatch serialising always enabled, * nothing to do on that case. */ - if (test_bit(X86_FEATURE_LFENCE_DISPATCH, c->x86_capability)) + if (TEST_BIT(X86_FEATURE_LFENCE_DISPATCH, c->x86_capability)) return; /* diff --git a/xen/arch/x86/cpu/common.c b/xen/arch/x86/cpu/common.c index e063fe790a..f96efe2b38 100644 --- a/xen/arch/x86/cpu/common.c +++ b/xen/arch/x86/cpu/common.c @@ -68,10 +68,10 @@ void __init setup_clear_cpu_cap(unsigned int cap) const uint32_t *dfs; unsigned int i; - if (__test_and_set_bit(cap, cleared_caps)) + if (TEST_AND_SET_BIT(cap, cleared_caps)) return; - if (test_bit(cap, forced_caps)) + if (TEST_BIT(cap, forced_caps)) printk("%pS clearing previously forced feature %#x\n", __builtin_return_address(0), cap); @@ -94,10 +94,10 @@ void __init setup_clear_cpu_cap(unsigned int cap) void __init setup_force_cpu_cap(unsigned int cap) { - if (__test_and_set_bit(cap, forced_caps)) + if (TEST_AND_SET_BIT(cap, forced_caps)) return; - if (test_bit(cap, cleared_caps)) { + if (TEST_BIT(cap, cleared_caps)) { printk("%pS tries to force previously cleared feature %#x\n", __builtin_return_address(0), cap); return; @@ -108,7 +108,7 @@ void __init setup_force_cpu_cap(unsigned int cap) bool __init is_forced_cpu_cap(unsigned int cap) { - return test_bit(cap, forced_caps); + return TEST_BIT(cap, forced_caps); } static void cf_check default_init(struct cpuinfo_x86 * c) @@ -411,7 +411,7 @@ void __init early_cpu_init(bool verbose) &c->x86_capability[FEATURESET_7c0], &c->x86_capability[FEATURESET_7d0]); - if (test_bit(X86_FEATURE_ARCH_CAPS, c->x86_capability)) + if (TEST_BIT(X86_FEATURE_ARCH_CAPS, c->x86_capability)) rdmsr(MSR_ARCH_CAPABILITIES, c->x86_capability[FEATURESET_m10Al], c->x86_capability[FEATURESET_m10Ah]); @@ -558,7 +558,7 @@ static void generic_identify(struct cpuinfo_x86 *c) &c->x86_capability[FEATURESET_Da1], &tmp, &tmp, &tmp); - if (test_bit(X86_FEATURE_ARCH_CAPS, c->x86_capability)) + if (TEST_BIT(X86_FEATURE_ARCH_CAPS, c->x86_capability)) rdmsr(MSR_ARCH_CAPABILITIES, c->x86_capability[FEATURESET_m10Al], c->x86_capability[FEATURESET_m10Ah]); diff --git a/xen/arch/x86/cpu/mcheck/x86_mca.h b/xen/arch/x86/cpu/mcheck/x86_mca.h index 18116737af..f0d8ff565a 100644 --- a/xen/arch/x86/cpu/mcheck/x86_mca.h +++ b/xen/arch/x86/cpu/mcheck/x86_mca.h @@ -110,7 +110,7 @@ static inline int mcabanks_test(int bit, struct mca_banks* banks) { if (!banks || !banks->bank_map || bit >= banks->num) return 0; - return test_bit(bit, banks->bank_map); + return TEST_BIT(bit, banks->bank_map); } struct mca_banks *mcabanks_alloc(unsigned int nr); diff --git a/xen/arch/x86/domain.c b/xen/arch/x86/domain.c index 56c3816187..0736229760 100644 --- a/xen/arch/x86/domain.c +++ b/xen/arch/x86/domain.c @@ -1377,7 +1377,7 @@ int arch_set_info_guest( */ if ( d != current->domain && !VM_ASSIST(d, m2p_strict) && is_pv_64bit_domain(d) && - test_bit(VMASST_TYPE_m2p_strict, &c.nat->vm_assist) && + TEST_BIT(VMASST_TYPE_m2p_strict, &c.nat->vm_assist) && atomic_read(&d->arch.pv.nr_l4_pages) ) { bool done = false; diff --git a/xen/arch/x86/e820.c b/xen/arch/x86/e820.c index ca577c0bde..036b6f10dc 100644 --- a/xen/arch/x86/e820.c +++ b/xen/arch/x86/e820.c @@ -449,7 +449,7 @@ static uint64_t __init mtrr_top_of_ram(void) printk(" MTRR cap: %"PRIx64" type: %"PRIx64"\n", mtrr_cap, mtrr_def); /* MTRRs enabled, and default memory type is not writeback? */ - if ( !test_bit(11, &mtrr_def) || ((uint8_t)mtrr_def == X86_MT_WB) ) + if ( !TEST_BIT(11, &mtrr_def) || ((uint8_t)mtrr_def == X86_MT_WB) ) return 0; /* @@ -466,7 +466,7 @@ static uint64_t __init mtrr_top_of_ram(void) printk(" MTRR[%d]: base %"PRIx64" mask %"PRIx64"\n", i, base, mask); - if ( !test_bit(11, &mask) || ((uint8_t)base != X86_MT_WB) ) + if ( !TEST_BIT(11, &mask) || ((uint8_t)base != X86_MT_WB) ) continue; base &= addr_mask; mask &= addr_mask; diff --git a/xen/arch/x86/guest/xen/xen.c b/xen/arch/x86/guest/xen/xen.c index 5c9f393c75..44481bb039 100644 --- a/xen/arch/x86/guest/xen/xen.c +++ b/xen/arch/x86/guest/xen/xen.c @@ -135,7 +135,7 @@ static int map_vcpuinfo(void) return 0; } - if ( test_bit(vcpu, vcpu_info_mapped) ) + if ( TEST_BIT(vcpu, vcpu_info_mapped) ) { this_cpu(vcpu_info) = &vcpu_info[vcpu]; return 0; diff --git a/xen/arch/x86/hvm/emulate.c b/xen/arch/x86/hvm/emulate.c index 19e809de66..584fcd1405 100644 --- a/xen/arch/x86/hvm/emulate.c +++ b/xen/arch/x86/hvm/emulate.c @@ -3040,7 +3040,7 @@ struct segment_register *hvmemul_get_seg_reg( if ( idx >= ARRAY_SIZE(hvmemul_ctxt->seg_reg) ) return ERR_PTR(-X86EMUL_UNHANDLEABLE); - if ( !__test_and_set_bit(idx, &hvmemul_ctxt->seg_reg_accessed) ) + if ( !TEST_AND_SET_BIT(idx, &hvmemul_ctxt->seg_reg_accessed) ) hvm_get_segment_register(current, idx, &hvmemul_ctxt->seg_reg[idx]); return &hvmemul_ctxt->seg_reg[idx]; } diff --git a/xen/arch/x86/hvm/hpet.c b/xen/arch/x86/hvm/hpet.c index f0e5f877f4..dc0cd33d1e 100644 --- a/xen/arch/x86/hvm/hpet.c +++ b/xen/arch/x86/hvm/hpet.c @@ -213,7 +213,7 @@ static void cf_check hpet_timer_fired(struct vcpu *v, void *data) HPETState *h = vcpu_vhpet(v); write_lock(&h->lock); - if ( __test_and_set_bit(tn, &h->hpet.isr) ) + if ( TEST_AND_SET_BIT(tn, &h->hpet.isr) ) ASSERT_UNREACHABLE(); write_unlock(&h->lock); } @@ -241,7 +241,7 @@ static void hpet_set_timer(HPETState *h, unsigned int tn, } if ( !timer_enabled(h, tn) || - (timer_level(h, tn) && test_bit(tn, &h->hpet.isr)) ) + (timer_level(h, tn) && TEST_BIT(tn, &h->hpet.isr)) ) return; if ( !timer_int_route_valid(h, tn) ) @@ -414,7 +414,7 @@ static int cf_check hpet_write( if ( i >= HPET_TIMER_NUM ) break; __clear_bit(i, &new_val); - active = __test_and_clear_bit(i, &h->hpet.isr); + active = TEST_AND_CLEAR_BIT(i, &h->hpet.isr); if ( active ) { hvm_ioapic_deassert(v->domain, timer_int_route(h, i)); diff --git a/xen/arch/x86/hvm/irq.c b/xen/arch/x86/hvm/irq.c index 5f64361113..4451ec85d4 100644 --- a/xen/arch/x86/hvm/irq.c +++ b/xen/arch/x86/hvm/irq.c @@ -95,7 +95,7 @@ static void __hvm_pci_intx_assert( ASSERT((device <= 31) && (intx <= 3)); - if ( __test_and_set_bit(device*4 + intx, &hvm_irq->pci_intx.i) ) + if ( TEST_AND_SET_BIT(device*4 + intx, &hvm_irq->pci_intx.i) ) return; gsi = hvm_pci_intx_gsi(device, intx); @@ -130,7 +130,7 @@ static void __hvm_pci_intx_deassert( ASSERT((device <= 31) && (intx <= 3)); - if ( !__test_and_clear_bit(device*4 + intx, &hvm_irq->pci_intx.i) ) + if ( !TEST_AND_CLEAR_BIT(device*4 + intx, &hvm_irq->pci_intx.i) ) return; gsi = hvm_pci_intx_gsi(device, intx); @@ -213,7 +213,7 @@ int hvm_isa_irq_assert(struct domain *d, unsigned int isa_irq, spin_lock(&d->arch.hvm.irq_lock); - if ( !__test_and_set_bit(isa_irq, &hvm_irq->isa_irq.i) && + if ( !TEST_AND_SET_BIT(isa_irq, &hvm_irq->isa_irq.i) && (hvm_irq->gsi_assert_count[gsi]++ == 0) ) assert_irq(d, gsi, isa_irq); @@ -235,7 +235,7 @@ void hvm_isa_irq_deassert( spin_lock(&d->arch.hvm.irq_lock); - if ( __test_and_clear_bit(isa_irq, &hvm_irq->isa_irq.i) && + if ( TEST_AND_CLEAR_BIT(isa_irq, &hvm_irq->isa_irq.i) && (--hvm_irq->gsi_assert_count[gsi] == 0) ) deassert_irq(d, isa_irq); @@ -720,7 +720,7 @@ static int cf_check irq_load_pci(struct domain *d, hvm_domain_context_t *h) /* Recalculate the counts from the IRQ line state */ for ( dev = 0; dev < 32; dev++ ) for ( intx = 0; intx < 4; intx++ ) - if ( test_bit(dev*4 + intx, &hvm_irq->pci_intx.i) ) + if ( TEST_BIT(dev*4 + intx, &hvm_irq->pci_intx.i) ) { /* Direct GSI assert */ gsi = hvm_pci_intx_gsi(dev, intx); @@ -745,7 +745,7 @@ static int cf_check irq_load_isa(struct domain *d, hvm_domain_context_t *h) /* Adjust the GSI assert counts for the ISA IRQ line state. * This relies on the PCI IRQ state being loaded first. */ for ( irq = 0; platform_legacy_irq(irq); irq++ ) - if ( test_bit(irq, &hvm_irq->isa_irq.i) ) + if ( TEST_BIT(irq, &hvm_irq->isa_irq.i) ) hvm_irq->gsi_assert_count[hvm_isa_irq_to_gsi(irq)]++; return 0; diff --git a/xen/arch/x86/hvm/svm/nestedsvm.c b/xen/arch/x86/hvm/svm/nestedsvm.c index dc2b6a4253..e2ca118606 100644 --- a/xen/arch/x86/hvm/svm/nestedsvm.c +++ b/xen/arch/x86/hvm/svm/nestedsvm.c @@ -376,8 +376,8 @@ static int nsvm_vmrun_permissionmap(struct vcpu *v, bool viopm) ns_viomap = hvm_map_guest_frame_ro(svm->ns_iomap_pa >> PAGE_SHIFT, 0); if ( ns_viomap ) { - ioport_80 = test_bit(0x80, ns_viomap); - ioport_ed = test_bit(0xed, ns_viomap); + ioport_80 = TEST_BIT(0x80, ns_viomap); + ioport_ed = TEST_BIT(0xed, ns_viomap); hvm_unmap_guest_frame(ns_viomap, 0); } @@ -866,10 +866,10 @@ nsvm_vmcb_guest_intercepts_msr(unsigned long *msr_bitmap, if ( write ) /* write access */ - enabled = test_bit(msr * 2 + 1, msr_bit); + enabled = TEST_BIT(msr * 2 + 1, msr_bit); else /* read access */ - enabled = test_bit(msr * 2, msr_bit); + enabled = TEST_BIT(msr * 2, msr_bit); if ( !enabled ) return NESTEDHVM_VMEXIT_HOST; @@ -906,7 +906,7 @@ nsvm_vmcb_guest_intercepts_ioio(paddr_t iopm_pa, uint64_t exitinfo1) for ( io_bitmap = hvm_map_guest_frame_ro(gfn, 0); ; ) { - enabled = io_bitmap && test_bit(port, io_bitmap); + enabled = io_bitmap && TEST_BIT(port, io_bitmap); if ( !enabled || !--size ) break; if ( unlikely(++port == 8 * PAGE_SIZE) ) diff --git a/xen/arch/x86/hvm/svm/svm.c b/xen/arch/x86/hvm/svm/svm.c index db1799bb43..14b7785961 100644 --- a/xen/arch/x86/hvm/svm/svm.c +++ b/xen/arch/x86/hvm/svm/svm.c @@ -1078,7 +1078,7 @@ static void svm_host_osvw_init(void) * choose the worst case (i.e. if erratum is present on one processor and * not on another assume that the erratum is present everywhere). */ - if ( test_bit(X86_FEATURE_OSVW, &boot_cpu_data.x86_capability) ) + if ( TEST_BIT(X86_FEATURE_OSVW, &boot_cpu_data.x86_capability) ) { uint64_t len, status; diff --git a/xen/arch/x86/hvm/viridian/time.c b/xen/arch/x86/hvm/viridian/time.c index 137577384f..3cacac5f7b 100644 --- a/xen/arch/x86/hvm/viridian/time.c +++ b/xen/arch/x86/hvm/viridian/time.c @@ -225,7 +225,7 @@ static void poll_stimer(struct vcpu *v, unsigned int stimerx) return; } - if ( !test_bit(stimerx, &vv->stimer_pending) ) + if ( !TEST_BIT(stimerx, &vv->stimer_pending) ) return; if ( !viridian_synic_deliver_timer_msg(v, vs->config.sintx, @@ -462,7 +462,7 @@ int viridian_time_rdmsr(const struct vcpu *v, uint32_t idx, uint64_t *val) * If the timer is single-shot and it has expired, make sure * the enabled flag is clear. */ - if ( !config.periodic && test_bit(stimerx, &vv->stimer_pending) ) + if ( !config.periodic && TEST_BIT(stimerx, &vv->stimer_pending) ) config.enable = 0; *val = config.as_uint64; diff --git a/xen/arch/x86/hvm/vlapic.c b/xen/arch/x86/hvm/vlapic.c index 993e972cd7..51b0d7ea94 100644 --- a/xen/arch/x86/hvm/vlapic.c +++ b/xen/arch/x86/hvm/vlapic.c @@ -684,7 +684,7 @@ int guest_rdmsr_x2apic(const struct vcpu *v, uint32_t msr, uint64_t *val) return X86EMUL_EXCEPTION; reg = array_index_nospec(reg, sizeof(readable) * 8); - if ( !test_bit(reg, readable) ) + if ( !TEST_BIT(reg, readable) ) return X86EMUL_EXCEPTION; offset = reg << 4; diff --git a/xen/arch/x86/hvm/vmx/vmcs.c b/xen/arch/x86/hvm/vmx/vmcs.c index ab8b1c87ec..b51d36dca4 100644 --- a/xen/arch/x86/hvm/vmx/vmcs.c +++ b/xen/arch/x86/hvm/vmx/vmcs.c @@ -736,7 +736,7 @@ static int _vmx_cpu_up(bool bsp) { eax = IA32_FEATURE_CONTROL_LOCK; eax |= IA32_FEATURE_CONTROL_ENABLE_VMXON_OUTSIDE_SMX; - if ( test_bit(X86_FEATURE_SMX, &boot_cpu_data.x86_capability) ) + if ( TEST_BIT(X86_FEATURE_SMX, &boot_cpu_data.x86_capability) ) eax |= IA32_FEATURE_CONTROL_ENABLE_VMXON_INSIDE_SMX; wrmsr(MSR_IA32_FEATURE_CONTROL, eax, 0); } @@ -774,7 +774,7 @@ static int _vmx_cpu_up(bool bsp) vmxon_fault: if ( bios_locked && - test_bit(X86_FEATURE_SMX, &boot_cpu_data.x86_capability) && + TEST_BIT(X86_FEATURE_SMX, &boot_cpu_data.x86_capability) && (!(eax & IA32_FEATURE_CONTROL_ENABLE_VMXON_OUTSIDE_SMX) || !(eax & IA32_FEATURE_CONTROL_ENABLE_VMXON_INSIDE_SMX)) ) { @@ -976,10 +976,10 @@ bool vmx_msr_is_intercepted(struct vmx_msr_bitmap *msr_bitmap, unsigned int msr, bool is_write) { if ( msr <= 0x1fff ) - return test_bit(msr, is_write ? msr_bitmap->write_low + return TEST_BIT(msr, is_write ? msr_bitmap->write_low : msr_bitmap->read_low); else if ( (msr >= 0xc0000000U) && (msr <= 0xc0001fffU) ) - return test_bit(msr & 0x1fff, is_write ? msr_bitmap->write_high + return TEST_BIT(msr & 0x1fff, is_write ? msr_bitmap->write_high : msr_bitmap->read_high); else /* MSRs outside the bitmap ranges are always intercepted. */ diff --git a/xen/arch/x86/hvm/vmx/vmx.c b/xen/arch/x86/hvm/vmx/vmx.c index ce538668c7..951e28f84a 100644 --- a/xen/arch/x86/hvm/vmx/vmx.c +++ b/xen/arch/x86/hvm/vmx/vmx.c @@ -132,7 +132,7 @@ static void vmx_pi_switch_from(struct vcpu *v) { struct pi_desc *pi_desc = &v->arch.hvm.vmx.pi_desc; - if ( test_bit(_VPF_blocked, &v->pause_flags) ) + if ( TEST_BIT(_VPF_blocked, &v->pause_flags) ) return; pi_set_sn(pi_desc); @@ -189,7 +189,7 @@ static void vmx_pi_unblock_vcpu(struct vcpu *v) static void vmx_pi_do_resume(struct vcpu *v) { - ASSERT(!test_bit(_VPF_blocked, &v->pause_flags)); + ASSERT(!TEST_BIT(_VPF_blocked, &v->pause_flags)); vmx_pi_unblock_vcpu(v); } diff --git a/xen/arch/x86/include/asm/guest_atomics.h b/xen/arch/x86/include/asm/guest_atomics.h index c2dec0d650..f901150204 100644 --- a/xen/arch/x86/include/asm/guest_atomics.h +++ b/xen/arch/x86/include/asm/guest_atomics.h @@ -10,7 +10,7 @@ #define guest_set_bit(d, nr, p) ((void)(d), set_bit(nr, p)) #define guest_clear_bit(d, nr, p) ((void)(d), clear_bit(nr, p)) #define guest_change_bit(d, nr, p) ((void)(d), change_bit(nr, p)) -#define guest_test_bit(d, nr, p) ((void)(d), test_bit(nr, p)) +#define guest_test_bit(d, nr, p) ((void)(d), TEST_BIT(nr, p)) #define guest_test_and_set_bit(d, nr, p) \ ((void)(d), test_and_set_bit(nr, p)) diff --git a/xen/arch/x86/include/asm/hvm/vlapic.h b/xen/arch/x86/include/asm/hvm/vlapic.h index c388551198..1eb777d694 100644 --- a/xen/arch/x86/include/asm/hvm/vlapic.h +++ b/xen/arch/x86/include/asm/hvm/vlapic.h @@ -52,7 +52,7 @@ #define VEC_POS(v) ((v) % 32) #define REG_POS(v) (((v) / 32) * 0x10) #define vlapic_test_vector(vec, bitmap) \ - test_bit(VEC_POS(vec), (const uint32_t *)((bitmap) + REG_POS(vec))) + TEST_BIT(VEC_POS(vec), (const uint32_t *)((bitmap) + REG_POS(vec))) #define vlapic_test_and_set_vector(vec, bitmap) \ test_and_set_bit(VEC_POS(vec), (uint32_t *)((bitmap) + REG_POS(vec))) #define vlapic_test_and_clear_vector(vec, bitmap) \ diff --git a/xen/arch/x86/include/asm/hvm/vmx/vmx.h b/xen/arch/x86/include/asm/hvm/vmx/vmx.h index da04752e17..0803ef8e06 100644 --- a/xen/arch/x86/include/asm/hvm/vmx/vmx.h +++ b/xen/arch/x86/include/asm/hvm/vmx/vmx.h @@ -91,7 +91,7 @@ static inline int pi_test_and_set_pir(uint8_t vector, struct pi_desc *pi_desc) static inline int pi_test_pir(uint8_t vector, const struct pi_desc *pi_desc) { - return test_bit(vector, pi_desc->pir); + return TEST_BIT(vector, pi_desc->pir); } static inline int pi_test_and_set_on(struct pi_desc *pi_desc) diff --git a/xen/arch/x86/include/asm/mpspec.h b/xen/arch/x86/include/asm/mpspec.h index 4781c253fa..d45fa1f2a6 100644 --- a/xen/arch/x86/include/asm/mpspec.h +++ b/xen/arch/x86/include/asm/mpspec.h @@ -44,7 +44,7 @@ typedef struct physid_mask physid_mask_t; #define physid_set(physid, map) set_bit(physid, (map).mask) #define physid_clear(physid, map) clear_bit(physid, (map).mask) -#define physid_isset(physid, map) test_bit(physid, (map).mask) +#define physid_isset(physid, map) TEST_BIT(physid, (map).mask) #define physid_test_and_set(physid, map) test_and_set_bit(physid, (map).mask) #define first_physid(map) find_first_bit((map).mask, \ diff --git a/xen/arch/x86/irq.c b/xen/arch/x86/irq.c index e70c7829b4..9c15846fd5 100644 --- a/xen/arch/x86/irq.c +++ b/xen/arch/x86/irq.c @@ -121,7 +121,7 @@ static void release_old_vec(struct irq_desc *desc) ASSERT_UNREACHABLE(); else if ( desc->arch.used_vectors ) { - ASSERT(test_bit(vector, desc->arch.used_vectors)); + ASSERT(TEST_BIT(vector, desc->arch.used_vectors)); clear_bit(vector, desc->arch.used_vectors); } } @@ -172,7 +172,7 @@ static int __init _bind_irq_vector(struct irq_desc *desc, int vector, cpumask_copy(desc->arch.cpu_mask, &online_mask); if ( desc->arch.used_vectors ) { - ASSERT(!test_bit(vector, desc->arch.used_vectors)); + ASSERT(!TEST_BIT(vector, desc->arch.used_vectors)); set_bit(vector, desc->arch.used_vectors); } desc->arch.used = IRQ_USED; @@ -218,7 +218,7 @@ static void _clear_irq_vector(struct irq_desc *desc) if ( desc->arch.used_vectors ) { - ASSERT(test_bit(vector, desc->arch.used_vectors)); + ASSERT(TEST_BIT(vector, desc->arch.used_vectors)); clear_bit(vector, desc->arch.used_vectors); } @@ -499,7 +499,7 @@ static vmask_t *irq_get_used_vector_mask(int irq) printk(XENLOG_INFO "IRQ%d already assigned vector %02x\n", irq, vector); - ASSERT(!test_bit(vector, ret)); + ASSERT(!TEST_BIT(vector, ret)); set_bit(vector, ret); } @@ -587,7 +587,7 @@ static int _assign_irq_vector(struct irq_desc *desc, const cpumask_t *mask) desc->arch.move_cleanup_count = 0; if ( desc->arch.used_vectors ) { - ASSERT(test_bit(old_vector, desc->arch.used_vectors)); + ASSERT(TEST_BIT(old_vector, desc->arch.used_vectors)); clear_bit(old_vector, desc->arch.used_vectors); } @@ -637,11 +637,11 @@ next: if (unlikely(current_vector == vector)) continue; - if (test_bit(vector, used_vectors)) + if (TEST_BIT(vector, used_vectors)) goto next; if (irq_used_vectors - && test_bit(vector, irq_used_vectors) ) + && TEST_BIT(vector, irq_used_vectors) ) goto next; if ( cpumask_test_cpu(0, vec_mask) && @@ -696,7 +696,7 @@ next: if ( desc->arch.used_vectors ) { - ASSERT(!test_bit(vector, desc->arch.used_vectors)); + ASSERT(!TEST_BIT(vector, desc->arch.used_vectors)); set_bit(vector, desc->arch.used_vectors); } @@ -2267,7 +2267,7 @@ int map_domain_pirq( { int vector = desc->arch.vector; - ASSERT(!test_bit(vector, desc->arch.used_vectors)); + ASSERT(!teTEST_BITector, desc->arch.used_vectors)); set_bit(vector, desc->arch.used_vectors); } } @@ -2329,12 +2329,12 @@ int map_domain_pirq( } while ( nr ) { - if ( irq >= 0 && test_bit(nr, granted) && + if ( irq >= 0 && teTEST_BITr, granted) && irq_deny_access(d, irq) ) printk(XENLOG_G_ERR "dom%d: could not revoke access to IRQ%d (pirq %d)\n", d->domain_id, irq, pirq); - if ( info && test_bit(nr, prepared) ) + if ( info && teTEST_BITr, prepared) ) cleanup_domain_irq_pirq(d, irq, info); info = pirq_info(d, pirq + --nr); irq = info->arch.irq; @@ -2358,10 +2358,10 @@ int map_domain_pirq( done: if ( ret ) { - if ( test_bit(0, prepared) ) + if ( teTEST_BIT, prepared) ) cleanup_domain_irq_pirq(d, irq, info); revoke: - if ( test_bit(0, granted) && irq_deny_access(d, irq) ) + if ( teTEST_BIT, granted) && irq_deny_access(d, irq) ) printk(XENLOG_G_ERR "dom%d: could not revoke access to IRQ%d (pirq %d)\n", d->domain_id, irq, pirq); diff --git a/xen/arch/x86/mm.c b/xen/arch/x86/mm.c index e7fd56c7ce..5b37505691 100644 --- a/xen/arch/x86/mm.c +++ b/xen/arch/x86/mm.c @@ -935,7 +935,7 @@ get_page_from_l1e( if ( !pci_mmcfg_decode(mfn, &seg, &bdf) || ((ro_map = pci_get_ro_map(seg)) != NULL && - test_bit(bdf, ro_map)) ) + TEST_BIT(bdf, ro_map)) ) printk(XENLOG_G_WARNING "d%d: Forcing read-only access to MFN %lx\n", l1e_owner->domain_id, mfn); @@ -2504,7 +2504,7 @@ static int cleanup_page_mappings(struct page_info *page) * mappings of this page. This forces the page to be coherent before it * is freed back to the heap. */ - if ( __test_and_clear_bit(_PGT_non_coherent, &page->u.inuse.type_info) ) + if ( TEST_AND_CLEAR_BIT(_PGT_non_coherent, &page->u.inuse.type_info) ) { void *addr = __map_domain_page(page); @@ -4960,7 +4960,7 @@ static int __init subpage_mmio_ro_add_page( for ( i = offset_s; i <= offset_e; i += MMIO_RO_SUBPAGE_GRAN ) { - bool oldbit = __test_and_set_bit(i / MMIO_RO_SUBPAGE_GRAN, + bool oldbit = TEST_AND_SET_BIT(i / MMIO_RO_SUBPAGE_GRAN, entry->ro_elems); ASSERT(!oldbit); } @@ -5115,7 +5115,7 @@ void subpage_mmio_write_emulate( /* Do not print message for pages without any writable parts. */ return; - if ( test_bit(offset / MMIO_RO_SUBPAGE_GRAN, entry->ro_elems) ) + if ( TEST_BIT(offset / MMIO_RO_SUBPAGE_GRAN, entry->ro_elems) ) { write_ignored: gprintk(XENLOG_WARNING, diff --git a/xen/arch/x86/mm/hap/hap.c b/xen/arch/x86/mm/hap/hap.c index ec5043a8aa..ca9a319bd6 100644 --- a/xen/arch/x86/mm/hap/hap.c +++ b/xen/arch/x86/mm/hap/hap.c @@ -717,7 +717,7 @@ static pagetable_t cf_check hap_update_cr3(struct vcpu *v, bool noflush) static bool flush_vcpu(const struct vcpu *v, const unsigned long *vcpu_bitmap) { - return !vcpu_bitmap || test_bit(v->vcpu_id, vcpu_bitmap); + return !vcpu_bitmap || TEST_BIT(v->vcpu_id, vcpu_bitmap); } /* Flush TLB of selected vCPUs. NULL for all. */ diff --git a/xen/arch/x86/mm/paging.c b/xen/arch/x86/mm/paging.c index c77f4c1dac..73e24148b4 100644 --- a/xen/arch/x86/mm/paging.c +++ b/xen/arch/x86/mm/paging.c @@ -311,7 +311,7 @@ void paging_mark_pfn_dirty(struct domain *d, pfn_t pfn) goto out; l1 = map_domain_page(mfn); - changed = !__test_and_set_bit(i1, l1); + changed = !TEST_AND_SET_BIT(i1, l1); unmap_domain_page(l1); if ( changed ) { @@ -383,7 +383,7 @@ bool paging_mfn_is_dirty(const struct domain *d, mfn_t gmfn) return false; l1 = map_domain_page(mfn); - dirty = test_bit(L1_LOGDIRTY_IDX(pfn), l1); + dirty = TEST_BIT(L1_LOGDIRTY_IDX(pfn), l1); unmap_domain_page(l1); return dirty; diff --git a/xen/arch/x86/mm/shadow/common.c b/xen/arch/x86/mm/shadow/common.c index 0176e33bc9..0691ffbca2 100644 --- a/xen/arch/x86/mm/shadow/common.c +++ b/xen/arch/x86/mm/shadow/common.c @@ -755,7 +755,7 @@ void shadow_demote(struct domain *d, mfn_t gmfn, u32 type) { struct page_info *page = mfn_to_page(gmfn); - ASSERT(test_bit(_PGC_shadowed_pt, &page->count_info)); + ASSERT(TEST_BIT(_PGC_shadowed_pt, &page->count_info)); ASSERT(page->shadow_flags & (1u << type)); page->shadow_flags &= ~(1u << type); diff --git a/xen/arch/x86/mm/shadow/hvm.c b/xen/arch/x86/mm/shadow/hvm.c index 114957a3e1..d8ba46bbfa 100644 --- a/xen/arch/x86/mm/shadow/hvm.c +++ b/xen/arch/x86/mm/shadow/hvm.c @@ -53,7 +53,7 @@ static struct segment_register *hvm_get_seg_reg( return ERR_PTR(-X86EMUL_UNHANDLEABLE); seg_reg = &sh_ctxt->seg_reg[idx]; - if ( !__test_and_set_bit(idx, &sh_ctxt->valid_seg_regs) ) + if ( !TEST_AND_SET_BIT(idx, &sh_ctxt->valid_seg_regs) ) hvm_get_segment_register(current, idx, seg_reg); return seg_reg; } @@ -696,7 +696,7 @@ static void sh_emulate_unmap_dest(struct vcpu *v, void *addr, static bool flush_vcpu(const struct vcpu *v, const unsigned long *vcpu_bitmap) { - return !vcpu_bitmap || test_bit(v->vcpu_id, vcpu_bitmap); + return !vcpu_bitmap || TEST_BIT(v->vcpu_id, vcpu_bitmap); } /* Flush TLB of selected vCPUs. NULL for all. */ diff --git a/xen/arch/x86/monitor.c b/xen/arch/x86/monitor.c index dc336c239a..26466eced1 100644 --- a/xen/arch/x86/monitor.c +++ b/xen/arch/x86/monitor.c @@ -122,7 +122,7 @@ bool monitored_msr(const struct domain *d, u32 msr) if ( !bitmap ) return false; - return test_bit(msr, bitmap); + return TEST_BIT(msr, bitmap); } bool monitored_msr_onchangeonly(const struct domain *d, u32 msr) @@ -137,7 +137,7 @@ bool monitored_msr_onchangeonly(const struct domain *d, u32 msr) if ( !bitmap ) return false; - return test_bit(msr + sizeof(struct monitor_msr_bitmap) * 8, bitmap); + return TEST_BIT(msr + sizeof(struct monitor_msr_bitmap) * 8, bitmap); } int arch_monitor_domctl_event(struct domain *d, diff --git a/xen/arch/x86/msi.c b/xen/arch/x86/msi.c index 5389bc0867..2d0ef3fb35 100644 --- a/xen/arch/x86/msi.c +++ b/xen/arch/x86/msi.c @@ -53,7 +53,7 @@ static int msix_fixmap_alloc(void) spin_lock(&msix_fixmap_lock); for ( i = 0; i < FIX_MSIX_MAX_PAGES; i++ ) - if ( !test_bit(i, &msix_fixmap_pages) ) + if ( !TEST_BIT(i, &msix_fixmap_pages) ) break; if ( i == FIX_MSIX_MAX_PAGES ) goto out; diff --git a/xen/arch/x86/psr.c b/xen/arch/x86/psr.c index 499d320e61..e229e53143 100644 --- a/xen/arch/x86/psr.c +++ b/xen/arch/x86/psr.c @@ -796,7 +796,7 @@ void psr_ctxt_switch_to(struct domain *d) struct psr_socket_info *info = socket_info + socket; unsigned int cos = 0; - if ( likely(test_bit(d->domain_id, info->dom_set)) && + if ( likely(TEST_BIT(d->domain_id, info->dom_set)) && d->arch.psr_cos_ids ) cos = d->arch.psr_cos_ids[socket]; diff --git a/xen/arch/x86/pv/dom0_build.c b/xen/arch/x86/pv/dom0_build.c index c37bea9454..73f4ebc593 100644 --- a/xen/arch/x86/pv/dom0_build.c +++ b/xen/arch/x86/pv/dom0_build.c @@ -496,7 +496,7 @@ static int __init dom0_construct(const struct boot_domain *bd) if ( parms.elf_notes[XEN_ELFNOTE_SUPPORTED_FEATURES].type != XEN_ENT_NONE ) { - if ( !pv_shim && !test_bit(XENFEAT_dom0, parms.f_supported) ) + if ( !pv_shim && !TEST_BIT(XENFEAT_dom0, parms.f_supported) ) { printk("Kernel does not support Dom0 operation\n"); return -EINVAL; @@ -1028,7 +1028,7 @@ static int __init dom0_construct(const struct boot_domain *bd) */ pv_destroy_gdt(v); - if ( test_bit(XENFEAT_supervisor_mode_kernel, parms.f_required) ) + if ( TEST_BIT(XENFEAT_supervisor_mode_kernel, parms.f_required) ) panic("Dom0 requires supervisor-mode execution\n"); rc = dom0_setup_permissions(d); diff --git a/xen/arch/x86/pv/emul-priv-op.c b/xen/arch/x86/pv/emul-priv-op.c index f3f012f8fb..c559b25a30 100644 --- a/xen/arch/x86/pv/emul-priv-op.c +++ b/xen/arch/x86/pv/emul-priv-op.c @@ -237,7 +237,7 @@ static bool pci_cfg_ok(struct domain *currd, unsigned int start, { const unsigned long *ro_map = pci_get_ro_map(0); - if ( ro_map && test_bit(machine_bdf, ro_map) ) + if ( ro_map && TEST_BIT(machine_bdf, ro_map) ) return false; } start |= CF8_ADDR_LO(currd->arch.pci_cf8); diff --git a/xen/arch/x86/pv/shim.c b/xen/arch/x86/pv/shim.c index bc2a7dd5fa..0af34bb07e 100644 --- a/xen/arch/x86/pv/shim.c +++ b/xen/arch/x86/pv/shim.c @@ -398,7 +398,7 @@ int pv_shim_shutdown(uint8_t reason) /* Stop the singleshot timer. */ stop_timer(&v->singleshot_timer); - if ( test_bit(_VPF_down, &v->pause_flags) ) + if ( TEST_BIT(_VPF_down, &v->pause_flags) ) BUG_ON(vcpu_reset(v)); if ( v != current ) diff --git a/xen/arch/x86/traps.c b/xen/arch/x86/traps.c index 34dc077cad..a6060ebe33 100644 --- a/xen/arch/x86/traps.c +++ b/xen/arch/x86/traps.c @@ -626,7 +626,7 @@ void vcpu_show_execution_state(struct vcpu *v) { unsigned long flags = 0; - if ( test_bit(_VPF_down, &v->pause_flags) ) + if ( TEST_BIT(_VPF_down, &v->pause_flags) ) { printk("*** %pv is offline ***\n", v); return; diff --git a/xen/arch/x86/x86_64/mm.c b/xen/arch/x86/x86_64/mm.c index d4e6a9c0a2..e2c462d11a 100644 --- a/xen/arch/x86/x86_64/mm.c +++ b/xen/arch/x86/x86_64/mm.c @@ -835,7 +835,7 @@ static int extend_frame_table(struct mem_hotadd_info *info) ASSERT( mfn_to_pdx(epfn) <= (DIRECTMAP_SIZE >> PAGE_SHIFT) && mfn_to_pdx(epfn) <= FRAMETABLE_NR ); - if ( test_bit(cidx, pdx_group_valid) ) + if ( TEST_BIT(cidx, pdx_group_valid) ) cidx = find_next_zero_bit(pdx_group_valid, eidx, cidx); if ( cidx >= eidx ) diff --git a/xen/arch/x86/x86_64/mmconfig_64.c b/xen/arch/x86/x86_64/mmconfig_64.c index ffdc62700d..6bcb00890a 100644 --- a/xen/arch/x86/x86_64/mmconfig_64.c +++ b/xen/arch/x86/x86_64/mmconfig_64.c @@ -202,7 +202,7 @@ bool pci_ro_mmcfg_decode(unsigned long mfn, unsigned int *seg, unsigned int *bdf return pci_mmcfg_decode(mfn, seg, bdf) && ((ro_map = pci_get_ro_map(*seg)) == NULL || - !test_bit(*bdf, ro_map)); + !TEST_BIT(*bdf, ro_map)); } int __init pci_mmcfg_arch_init(void) diff --git a/xen/arch/x86/xstate.c b/xen/arch/x86/xstate.c index e8e218caed..433be12e6e 100644 --- a/xen/arch/x86/xstate.c +++ b/xen/arch/x86/xstate.c @@ -128,7 +128,7 @@ static int setup_xstate_features(bool bsp) &ebx, &ecx, &edx); BUG_ON(eax != xstate_sizes[leaf]); BUG_ON(ebx != xstate_offsets[leaf]); - BUG_ON(!(ecx & XSTATE_ALIGN64) != !test_bit(leaf, &xstate_align)); + BUG_ON(!(ecx & XSTATE_ALIGN64) != !TEST_BIT(leaf, &xstate_align)); } } @@ -156,7 +156,7 @@ static void setup_xstate_comp(uint16_t *comp_offsets, { if ( (1UL << i) & xcomp_bv ) { - if ( test_bit(i, &xstate_align) ) + if ( TEST_BIT(i, &xstate_align) ) offset = ROUNDUP(offset, 64); comp_offsets[i] = offset; offset += xstate_sizes[i]; diff --git a/xen/common/domain.c b/xen/common/domain.c index 303c338ef2..1821864f91 100644 --- a/xen/common/domain.c +++ b/xen/common/domain.c @@ -2082,7 +2082,7 @@ long common_vcpu_op(int cmd, struct vcpu *v, XEN_GUEST_HANDLE_PARAM(void) arg) case VCPUOP_down: for_each_vcpu ( d, v ) - if ( v->vcpu_id != vcpuid && !test_bit(_VPF_down, &v->pause_flags) ) + if ( v->vcpu_id != vcpuid && !TEST_BIT(_VPF_down, &v->pause_flags) ) { rc = 1; break; @@ -2272,7 +2272,7 @@ long do_vm_assist(unsigned int cmd, unsigned int type) struct domain *currd = current->domain; const unsigned long valid = arch_vm_assist_valid_mask(currd); - if ( type >= BITS_PER_LONG || !test_bit(type, &valid) ) + if ( type >= BITS_PER_LONG || !TEST_BIT(type, &valid) ) return -EINVAL; switch ( cmd ) diff --git a/xen/common/efi/runtime.c b/xen/common/efi/runtime.c index 7e1fce291d..bbc599da7f 100644 --- a/xen/common/efi/runtime.c +++ b/xen/common/efi/runtime.c @@ -76,7 +76,7 @@ const struct efi_pci_rom *__read_mostly efi_pci_roms; bool efi_enabled(unsigned int feature) { - return test_bit(feature, &efi_flags); + return TEST_BIT(feature, &efi_flags); } #ifndef CONFIG_ARM /* TODO - disabled until implemented on ARM */ diff --git a/xen/common/event_2l.c b/xen/common/event_2l.c index d40dd51ab5..addf39ad3a 100644 --- a/xen/common/event_2l.c +++ b/xen/common/event_2l.c @@ -94,7 +94,7 @@ static void cf_check evtchn_2l_print_state( { struct vcpu *v = d->vcpu[evtchn->notify_vcpu_id]; - printk("%d", !!test_bit(evtchn->port / BITS_PER_EVTCHN_WORD(d), + printk("%d", !!TEST_BIT(evtchn->port / BITS_PER_EVTCHN_WORD(d), &vcpu_info(v, evtchn_pending_sel))); } diff --git a/xen/common/kexec.c b/xen/common/kexec.c index 84fe8c3597..4044069f10 100644 --- a/xen/common/kexec.c +++ b/xen/common/kexec.c @@ -379,8 +379,8 @@ void kexec_crash(enum crash_reason reason) keyhandler_crash_action(reason); - pos = (test_bit(KEXEC_FLAG_CRASH_POS, &kexec_flags) != 0); - if ( !test_bit(KEXEC_IMAGE_CRASH_BASE + pos, &kexec_flags) ) + pos = (TEST_BIT(KEXEC_FLAG_CRASH_POS, &kexec_flags) != 0); + if ( !TEST_BIT(KEXEC_IMAGE_CRASH_BASE + pos, &kexec_flags) ) return; kexecing = true; @@ -829,10 +829,10 @@ static int kexec_exec(XEN_GUEST_HANDLE_PARAM(void) uarg) if ( kexec_load_get_bits(exec.type, &base, &bit) ) return -EINVAL; - pos = (test_bit(bit, &kexec_flags) != 0); + pos = (TEST_BIT(bit, &kexec_flags) != 0); /* Only allow kexec/kdump into loaded images */ - if ( !test_bit(base + pos, &kexec_flags) ) + if ( !TEST_BIT(base + pos, &kexec_flags) ) return -ENOENT; switch (exec.type) @@ -857,15 +857,15 @@ static int kexec_swap_images(int type, struct kexec_image *new, *old = NULL; - if ( test_bit(KEXEC_FLAG_IN_PROGRESS, &kexec_flags) ) + if ( TEST_BIT(KEXEC_FLAG_IN_PROGRESS, &kexec_flags) ) return -EBUSY; if ( kexec_load_get_bits(type, &base, &bit) ) return -EINVAL; - ASSERT(test_bit(KEXEC_FLAG_IN_HYPERCALL, &kexec_flags)); + ASSERT(TEST_BIT(KEXEC_FLAG_IN_HYPERCALL, &kexec_flags)); - pos = (test_bit(bit, &kexec_flags) != 0); + pos = (TEST_BIT(bit, &kexec_flags) != 0); old_slot = base + pos; new_slot = base + !pos; @@ -1210,7 +1210,7 @@ static int kexec_status(XEN_GUEST_HANDLE_PARAM(void) uarg) if ( kexec_load_get_bits(status.type, &base, &bit) ) return -EINVAL; - return !!test_bit(bit, &kexec_flags); + return !!TEST_BIT(bit, &kexec_flags); } static int do_kexec_op_internal(unsigned long op, diff --git a/xen/common/keyhandler.c b/xen/common/keyhandler.c index b0a2051408..2cd3d37323 100644 --- a/xen/common/keyhandler.c +++ b/xen/common/keyhandler.c @@ -303,7 +303,7 @@ static void cf_check dump_domains(unsigned char key) d->handle[12], d->handle[13], d->handle[14], d->handle[15], d->vm_assist); for ( i = 0 ; i < NR_DOMAIN_WATCHDOG_TIMERS; i++ ) - if ( test_bit(i, &d->watchdog_inuse_map) ) + if ( TEST_BIT(i, &d->watchdog_inuse_map) ) printk(" watchdog %d expires in %d seconds\n", i, (u32)((d->watchdog_timer[i].expires - NOW()) >> 30)); diff --git a/xen/common/multicall.c b/xen/common/multicall.c index ce394c5efc..4169b513f8 100644 --- a/xen/common/multicall.c +++ b/xen/common/multicall.c @@ -42,7 +42,7 @@ ret_t do_multicall( int rc = 0; enum mc_disposition disp = mc_continue; - if ( unlikely(__test_and_set_bit(_MCSF_in_multicall, &mcs->flags)) ) + if ( unlikely(TEST_AND_SET_BIT(_MCSF_in_multicall, &mcs->flags)) ) { gdprintk(XENLOG_INFO, "Multicall reentry is disallowed.\n"); return -EINVAL; diff --git a/xen/common/numa.c b/xen/common/numa.c index ad75955a16..6439820a52 100644 --- a/xen/common/numa.c +++ b/xen/common/numa.c @@ -184,7 +184,7 @@ bool __init numa_update_node_memblks(nodeid_t node, unsigned int arch_nid, case OVERLAP: if ( memblk_nodeid[i] == node ) { - bool mismatch = !hotplug != !test_bit(i, memblk_hotplug); + bool mismatch = !hotplug != !TEST_BIT(i, memblk_hotplug); printk("%sNUMA: %s %u [%"PRIpaddr", %"PRIpaddr"] overlaps with itself [%"PRIpaddr", %"PRIpaddr"]\n", mismatch ? KERN_ERR : KERN_WARNING, numa_fw_nid_name, @@ -252,7 +252,7 @@ bool __init numa_update_node_memblks(nodeid_t node, unsigned int arch_nid, { bool prev = next; - next = test_bit(i, memblk_hotplug); + next = TEST_BIT(i, memblk_hotplug); if ( prev ) __set_bit(i, memblk_hotplug); else diff --git a/xen/common/page_alloc.c b/xen/common/page_alloc.c index ec9dec365e..36ef07cac4 100644 --- a/xen/common/page_alloc.c +++ b/xen/common/page_alloc.c @@ -1232,7 +1232,7 @@ static int reserve_offlined_page(struct page_info *head) ASSERT(total_avail_pages >= 0); page_list_add_tail(cur_head, - test_bit(_PGC_broken, &cur_head->count_info) ? + TEST_BIT(_PGC_broken, &cur_head->count_info) ? &page_broken_list : &page_offlined_list); count++; @@ -1377,7 +1377,7 @@ bool scrub_free_pages(void) for ( i = pg->u.free.first_dirty; i < (1U << order); i++) { - if ( test_bit(_PGC_need_scrub, &pg[i].count_info) ) + if ( TEST_BIT(_PGC_need_scrub, &pg[i].count_info) ) { scrub_one_page(&pg[i], true); /* diff --git a/xen/common/pdx.c b/xen/common/pdx.c index b8384e6189..7d78abe345 100644 --- a/xen/common/pdx.c +++ b/xen/common/pdx.c @@ -41,7 +41,7 @@ bool __mfn_valid(unsigned long mfn) if ( unlikely(evaluate_nospec(invalid)) ) return false; - return test_bit(pfn_to_pdx(mfn) / PDX_GROUP_COUNT, pdx_group_valid); + return TEST_BIT(pfn_to_pdx(mfn) / PDX_GROUP_COUNT, pdx_group_valid); } void set_pdx_range(unsigned long smfn, unsigned long emfn) diff --git a/xen/common/sched/core.c b/xen/common/sched/core.c index ea95dea65a..32aa7eb4d5 100644 --- a/xen/common/sched/core.c +++ b/xen/common/sched/core.c @@ -243,7 +243,7 @@ static inline void vcpu_urgent_count_update(struct vcpu *v) if ( unlikely(v->is_urgent) ) { if ( !(v->pause_flags & VPF_blocked) || - !test_bit(v->vcpu_id, v->domain->poll_mask) ) + !TEST_BIT(v->vcpu_id, v->domain->poll_mask) ) { v->is_urgent = 0; atomic_dec(&per_cpu(sched_urgent_count, v->processor)); @@ -252,7 +252,7 @@ static inline void vcpu_urgent_count_update(struct vcpu *v) else { if ( unlikely(v->pause_flags & VPF_blocked) && - unlikely(test_bit(v->vcpu_id, v->domain->poll_mask)) ) + unlikely(TEST_BIT(v->vcpu_id, v->domain->poll_mask)) ) { v->is_urgent = 1; atomic_inc(&per_cpu(sched_urgent_count, v->processor)); @@ -1109,7 +1109,7 @@ static void sched_unit_migrate_finish(struct sched_unit *unit) if ( unit->is_running ) return; for_each_sched_unit_vcpu ( unit, v ) - if ( !test_bit(_VPF_migrating, &v->pause_flags) ) + if ( !TEST_BIT(_VPF_migrating, &v->pause_flags) ) return; old_cpu = new_cpu = unit->res->master_cpu; @@ -1469,8 +1469,8 @@ static long do_poll(const struct sched_poll *sched_poll) */ rc = 0; if ( (v->poll_evtchn == 0) || - !test_bit(_VPF_blocked, &v->pause_flags) || - !test_bit(v->vcpu_id, d->poll_mask) ) + !TEST_BIT(_VPF_blocked, &v->pause_flags) || + !TEST_BIT(v->vcpu_id, d->poll_mask) ) goto out; #endif @@ -1571,7 +1571,7 @@ static long domain_watchdog(struct domain *d, uint32_t id, uint32_t timeout) } id -= 1; - if ( !test_bit(id, &d->watchdog_inuse_map) ) + if ( !TEST_BIT(id, &d->watchdog_inuse_map) ) { spin_unlock(&d->watchdog_lock); return -EINVAL; diff --git a/xen/common/sched/credit.c b/xen/common/sched/credit.c index 6dcf6b2c8b..d24606e0b1 100644 --- a/xen/common/sched/credit.c +++ b/xen/common/sched/credit.c @@ -298,7 +298,7 @@ __runq_insert(struct csched_unit *svc) /* If the unit yielded, try to put it behind one lower-priority * runnable unit if we can. The next runq_sort will bring it forward * within 30ms if the queue too long. */ - if ( test_bit(CSCHED_FLAG_UNIT_YIELD, &svc->flags) + if ( TEST_BIT(CSCHED_FLAG_UNIT_YIELD, &svc->flags) && __runq_elem(iter)->pri > CSCHED_PRI_IDLE && iter->next != runq ) iter=iter->next; @@ -384,7 +384,7 @@ static inline void __runq_tickle(const struct csched_unit *new) * Try to identify the vast majority of these situations, and deal * with them quickly. */ - if ( unlikely(test_bit(CSCHED_FLAG_UNIT_PINNED, &new->flags) && + if ( unlikely(TEST_BIT(CSCHED_FLAG_UNIT_PINNED, &new->flags) && cpumask_test_cpu(cpu, &idle_mask)) ) { ASSERT(cpumask_cycle(cpu, unit->cpu_hard_affinity) == cpu); @@ -1162,7 +1162,7 @@ csched_unit_wake(const struct scheduler *ops, struct sched_unit *unit) migrating = test_and_clear_bit(CSCHED_FLAG_UNIT_MIGRATING, &svc->flags); if ( !migrating && svc->pri == CSCHED_PRI_TS_UNDER && - !test_bit(CSCHED_FLAG_UNIT_PARKED, &svc->flags) ) + !TEST_BIT(CSCHED_FLAG_UNIT_PARKED, &svc->flags) ) { TRACE_TIME(TRC_CSCHED_BOOST_START, unit->domain->domain_id, unit->unit_id); SCHED_STAT_CRANK(unit_boost); @@ -1542,7 +1542,7 @@ static void cf_check csched_acct(void* dummy) svc->pri = CSCHED_PRI_TS_UNDER; /* Unpark any capped domains whose credits go positive */ - if ( test_bit(CSCHED_FLAG_UNIT_PARKED, &svc->flags) ) + if ( TEST_BIT(CSCHED_FLAG_UNIT_PARKED, &svc->flags) ) { /* * It's important to unset the flag AFTER the unpause() @@ -1643,7 +1643,7 @@ csched_runq_steal(int peer_cpu, int cpu, int pri, int balance_step) * Don't steal a UNIT which has yielded; it's waiting for a * reason */ - if ( test_bit(CSCHED_FLAG_UNIT_YIELD, &speer->flags) ) + if ( TEST_BIT(CSCHED_FLAG_UNIT_YIELD, &speer->flags) ) continue; /* Is this UNIT runnable on our PCPU? */ @@ -1912,7 +1912,7 @@ static void cf_check csched_schedule( * In fact, it may be the case that scurr is about to spin, and there's * no point forcing it to do so until rate limiting expires. */ - if ( !test_bit(CSCHED_FLAG_UNIT_YIELD, &scurr->flags) + if ( !TEST_BIT(CSCHED_FLAG_UNIT_YIELD, &scurr->flags) && !tasklet_work_scheduled && prv->ratelimit && unit_runnable_state(unit) diff --git a/xen/common/sched/credit2.c b/xen/common/sched/credit2.c index 0b3b61df57..3c7849abde 100644 --- a/xen/common/sched/credit2.c +++ b/xen/common/sched/credit2.c @@ -2333,7 +2333,7 @@ csched2_context_saved(const struct scheduler *ops, struct sched_unit *unit) * it seems a bit pointless; especially as we have plenty of * bits free. */ - if ( __test_and_clear_bit(__CSFLAG_delayed_runq_add, &svc->flags) + if ( TEST_AND_CLEAR_BIT(__CSFLAG_delayed_runq_add, &svc->flags) && likely(unit_runnable(unit)) ) { ASSERT(!unit_on_runq(svc)); @@ -2399,7 +2399,7 @@ csched2_res_pick(const struct scheduler *ops, const struct sched_unit *unit) * First check to see if we're here because someone else suggested a place * for us to move. */ - if ( __test_and_clear_bit(__CSFLAG_runq_migrate_request, &svc->flags) ) + if ( TEST_AND_CLEAR_BIT(__CSFLAG_runq_migrate_request, &svc->flags) ) { if ( unlikely(svc->migrate_rqd->id < 0) ) { @@ -3393,7 +3393,7 @@ runq_candidate(struct csched2_runqueue_data *rqd, goto check_runq; } - yield = __test_and_clear_bit(__CSFLAG_unit_yield, &scurr->flags); + yield = TEST_AND_CLEAR_BIT(__CSFLAG_unit_yield, &scurr->flags); /* * Return the current unit if it has executed for less than ratelimit. diff --git a/xen/common/sched/rt.c b/xen/common/sched/rt.c index 7b1f64a779..a2a7de7522 100644 --- a/xen/common/sched/rt.c +++ b/xen/common/sched/rt.c @@ -1349,7 +1349,7 @@ rt_context_saved(const struct scheduler *ops, struct sched_unit *unit) if ( is_idle_unit(unit) ) goto out; - if ( __test_and_clear_bit(__RTDS_delayed_runq_add, &svc->flags) && + if ( TEST_AND_CLEAR_BIT(__RTDS_delayed_runq_add, &svc->flags) && likely(unit_runnable(unit)) ) { runq_insert(ops, svc); @@ -1534,7 +1534,7 @@ static void cf_check repl_timer_handler(void *data) if ( compare_unit_priority(svc, next_on_runq) < 0 ) runq_tickle(ops, next_on_runq); } - else if ( __test_and_clear_bit(__RTDS_depleted, &svc->flags) && + else if ( TEST_AND_CLEAR_BIT(__RTDS_depleted, &svc->flags) && unit_on_q(svc) ) runq_tickle(ops, svc); diff --git a/xen/common/vmap.c b/xen/common/vmap.c index d6991421f3..1f87b20c18 100644 --- a/xen/common/vmap.c +++ b/xen/common/vmap.c @@ -76,7 +76,7 @@ static void *vm_alloc(unsigned int nr, unsigned int align, { mfn_t mfn; - ASSERT(vm_low[t] == vm_top[t] || !test_bit(vm_low[t], vm_bitmap(t))); + ASSERT(vm_low[t] == vm_top[t] || !TEST_BIT(vm_low[t], vm_bitmap(t))); for ( start = vm_low[t]; start < vm_top[t]; ) { bit = find_next_bit(vm_bitmap(t), vm_top[t], start + 1); @@ -151,7 +151,7 @@ static void *vm_alloc(unsigned int nr, unsigned int align, for ( bit = start; bit < start + nr; ++bit ) __set_bit(bit, vm_bitmap(t)); if ( bit < vm_top[t] ) - ASSERT(!test_bit(bit, vm_bitmap(t))); + ASSERT(!TEST_BIT(bit, vm_bitmap(t))); else ASSERT(bit == vm_top[t]); if ( start <= vm_low[t] + 2 ) @@ -175,8 +175,8 @@ static unsigned int vm_index(const void *va, enum vmap_region type) return 0; idx = PFN_DOWN(va - vm_base[type]); - return !test_bit(idx - 1, vm_bitmap(type)) && - test_bit(idx, vm_bitmap(type)) ? idx : 0; + return !TEST_BIT(idx - 1, vm_bitmap(type)) && + TEST_BIT(idx, vm_bitmap(type)) ? idx : 0; } static unsigned int vm_size(const void *va, enum vmap_region type) @@ -212,10 +212,10 @@ static void vm_free(const void *va) if ( bit < vm_low[type] ) { vm_low[type] = bit - 1; - while ( !test_bit(vm_low[type] - 1, vm_bitmap(type)) ) + while ( !TEST_BIT(vm_low[type] - 1, vm_bitmap(type)) ) --vm_low[type]; } - while ( __test_and_clear_bit(bit, vm_bitmap(type)) ) + while ( TEST_AND_CLEAR_BIT(bit, vm_bitmap(type)) ) if ( ++bit == vm_top[type] ) break; spin_unlock(&vm_lock); diff --git a/xen/drivers/passthrough/iommu.c b/xen/drivers/passthrough/iommu.c index c9425d6971..994ce2bcea 100644 --- a/xen/drivers/passthrough/iommu.c +++ b/xen/drivers/passthrough/iommu.c @@ -691,7 +691,7 @@ int iommu_get_reserved_device_memory(iommu_grdm_t *func, void *ctxt) bool iommu_has_feature(struct domain *d, enum iommu_feature feature) { - return is_iommu_enabled(d) && test_bit(feature, dom_iommu(d)->features); + return is_iommu_enabled(d) && TEST_BIT(feature, dom_iommu(d)->features); } #define MAX_EXTRA_RESERVED_RANGES 20 diff --git a/xen/drivers/passthrough/vtd/dmar.c b/xen/drivers/passthrough/vtd/dmar.c index 91c22b8330..be7bbe32bd 100644 --- a/xen/drivers/passthrough/vtd/dmar.c +++ b/xen/drivers/passthrough/vtd/dmar.c @@ -225,7 +225,7 @@ struct acpi_drhd_unit *acpi_find_matched_drhd_unit(const struct pci_dev *pdev) if ( drhd->scope.devices[i] == PCI_BDF(bus, devfn) ) return drhd; - if ( test_bit(bus, drhd->scope.buses) ) + if ( TEST_BIT(bus, drhd->scope.buses) ) return drhd; if ( drhd->include_all ) @@ -244,7 +244,7 @@ struct acpi_atsr_unit *acpi_find_matched_atsr_unit(const struct pci_dev *pdev) if ( atsr->segment != pdev->seg ) continue; - if ( test_bit(pdev->bus, atsr->scope.buses) ) + if ( TEST_BIT(pdev->bus, atsr->scope.buses) ) return atsr; if ( atsr->all_ports ) diff --git a/xen/drivers/passthrough/vtd/iommu.c b/xen/drivers/passthrough/vtd/iommu.c index c55f02c97e..34116ff369 100644 --- a/xen/drivers/passthrough/vtd/iommu.c +++ b/xen/drivers/passthrough/vtd/iommu.c @@ -240,7 +240,7 @@ domid_t did_to_domain_id(const struct vtd_iommu *iommu, unsigned int did) if ( !domid_mapping(iommu) ) return convert_domid(iommu, did); - if ( !test_bit(did, iommu->domid_bitmap) ) + if ( !TEST_BIT(did, iommu->domid_bitmap) ) return DOMID_INVALID; return iommu->domid_map[did]; @@ -763,7 +763,7 @@ static int __must_check cf_check iommu_flush_iotlb(struct domain *d, dfn_t dfn, iommu = drhd->iommu; - if ( !test_bit(iommu->index, hd->arch.vtd.iommu_bitmap) ) + if ( !TEST_BIT(iommu->index, hd->arch.vtd.iommu_bitmap) ) continue; flush_dev_iotlb = !!find_ats_dev_drhd(iommu); diff --git a/xen/drivers/passthrough/x86/hvm.c b/xen/drivers/passthrough/x86/hvm.c index 1c545ed89d..ab656d6b00 100644 --- a/xen/drivers/passthrough/x86/hvm.c +++ b/xen/drivers/passthrough/x86/hvm.c @@ -967,7 +967,7 @@ static void hvm_dpci_isairq_eoi(struct domain *d, unsigned int isairq) dpci = domain_get_irq_dpci(d); - if ( dpci && test_bit(isairq, dpci->isairq_map) ) + if ( dpci && TEST_BIT(isairq, dpci->isairq_map) ) { /* Multiple mirq may be mapped to one isa irq */ pt_pirq_iterate(d, _hvm_dpci_isairq_eoi, (void *)(long)isairq); diff --git a/xen/drivers/passthrough/x86/iommu.c b/xen/drivers/passthrough/x86/iommu.c index 0954cc4922..30ad0e74ca 100644 --- a/xen/drivers/passthrough/x86/iommu.c +++ b/xen/drivers/passthrough/x86/iommu.c @@ -578,7 +578,7 @@ void iommu_free_domid(domid_t domid, unsigned long *map) ASSERT(domid > DOMID_MASK); - if ( !__test_and_clear_bit(domid & DOMID_MASK, map) ) + if ( !TEST_AND_CLEAR_BIT(domid & DOMID_MASK, map) ) BUG(); } diff --git a/xen/drivers/vpci/vpci.c b/xen/drivers/vpci/vpci.c index 09988f04c2..6e1de0df23 100644 --- a/xen/drivers/vpci/vpci.c +++ b/xen/drivers/vpci/vpci.c @@ -142,7 +142,7 @@ int vpci_assign_device(struct pci_dev *pdev) /* No vPCI for r/o devices. */ ro_map = pci_get_ro_map(pdev->sbdf.seg); - if ( ro_map && test_bit(pdev->sbdf.bdf, ro_map) ) + if ( ro_map && TEST_BIT(pdev->sbdf.bdf, ro_map) ) return 0; pdev->vpci = xzalloc(struct vpci); @@ -628,7 +628,7 @@ void vpci_write(pci_sbdf_t sbdf, unsigned int reg, unsigned int size, write_unlock(&d->pci_lock); - if ( !ro_map || !test_bit(sbdf.bdf, ro_map) ) + if ( !ro_map || !TEST_BIT(sbdf.bdf, ro_map) ) vpci_write_hw(sbdf, reg, size, data); return; } diff --git a/xen/include/xen/bitops.h b/xen/include/xen/bitops.h index a4d31ec02a..11f0fd16fa 100644 --- a/xen/include/xen/bitops.h +++ b/xen/include/xen/bitops.h @@ -137,7 +137,7 @@ __test_and_set_bit(int nr, volatile void *addr) return arch__test_and_set_bit(nr, addr); } -#define __test_and_set_bit(nr, addr) ({ \ +#define TEST_AND_SET_BIT(nr, addr) ({ \ if ( bitop_bad_size(addr) ) __bitop_bad_size(); \ __test_and_set_bit(nr, addr); \ }) @@ -160,7 +160,7 @@ __test_and_clear_bit(int nr, volatile void *addr) return arch__test_and_clear_bit(nr, addr); } -#define __test_and_clear_bit(nr, addr) ({ \ +#define TEST_AND_CLEAR_BIT(nr, addr) ({ \ if ( bitop_bad_size(addr) ) __bitop_bad_size(); \ __test_and_clear_bit(nr, addr); \ }) @@ -183,7 +183,7 @@ __test_and_change_bit(int nr, volatile void *addr) return arch__test_and_change_bit(nr, addr); } -#define __test_and_change_bit(nr, addr) ({ \ +#define TEST_AND_CHANGE_BIT(nr, addr) ({ \ if ( bitop_bad_size(addr) ) __bitop_bad_size(); \ __test_and_change_bit(nr, addr); \ }) @@ -205,7 +205,7 @@ static always_inline bool test_bit(int nr, const volatile void *addr) return arch_test_bit(nr, addr); } -#define test_bit(nr, addr) ({ \ +#define TEST_BIT(nr, addr) ({ \ if ( bitop_bad_size(addr) ) __bitop_bad_size(); \ test_bit(nr, addr); \ }) diff --git a/xen/include/xen/cpumask.h b/xen/include/xen/cpumask.h index b713bb69a9..f4e2981f92 100644 --- a/xen/include/xen/cpumask.h +++ b/xen/include/xen/cpumask.h @@ -117,7 +117,7 @@ static inline void cpumask_clear(cpumask_t *dstp) static inline bool cpumask_test_cpu(unsigned int cpu, const cpumask_t *src) { - return test_bit(cpumask_check(cpu), src->bits); + return TEST_BIT(cpumask_check(cpu), src->bits); } static inline int cpumask_test_and_set_cpu(int cpu, volatile cpumask_t *addr) @@ -127,7 +127,7 @@ static inline int cpumask_test_and_set_cpu(int cpu, volatile cpumask_t *addr) static inline int __cpumask_test_and_set_cpu(int cpu, cpumask_t *addr) { - return __test_and_set_bit(cpumask_check(cpu), addr->bits); + return TEST_AND_SET_BIT(cpumask_check(cpu), addr->bits); } static inline int cpumask_test_and_clear_cpu(int cpu, volatile cpumask_t *addr) @@ -137,7 +137,7 @@ static inline int cpumask_test_and_clear_cpu(int cpu, volatile cpumask_t *addr) static inline int __cpumask_test_and_clear_cpu(int cpu, cpumask_t *addr) { - return __test_and_clear_bit(cpumask_check(cpu), addr->bits); + return TEST_AND_CLEAR_BIT(cpumask_check(cpu), addr->bits); } static inline void cpumask_and(cpumask_t *dstp, const cpumask_t *src1p, diff --git a/xen/include/xen/nodemask.h b/xen/include/xen/nodemask.h index 1dd6c7458e..b931382f07 100644 --- a/xen/include/xen/nodemask.h +++ b/xen/include/xen/nodemask.h @@ -95,7 +95,7 @@ static inline void __nodes_clear(nodemask_t *dstp, int nbits) static inline bool nodemask_test(unsigned int node, const nodemask_t *dst) { - return test_bit(node, dst->bits); + return TEST_BIT(node, dst->bits); } #define node_test_and_set(node, nodemask) \ diff --git a/xen/include/xen/sched.h b/xen/include/xen/sched.h index fe53d4fab7..d521e3cd86 100644 --- a/xen/include/xen/sched.h +++ b/xen/include/xen/sched.h @@ -1148,7 +1148,7 @@ static always_inline bool is_control_domain(const struct domain *d) return evaluate_nospec(d->is_privileged); } -#define VM_ASSIST(d, t) (test_bit(VMASST_TYPE_ ## t, &(d)->vm_assist)) +#define VM_ASSIST(d, t) (TEST_BIT(VMASST_TYPE_ ## t, &(d)->vm_assist)) static always_inline bool is_pv_domain(const struct domain *d) { @@ -1218,7 +1218,7 @@ static inline bool is_hwdom_pinned_vcpu(const struct vcpu *v) static inline bool is_vcpu_online(const struct vcpu *v) { - return !test_bit(_VPF_down, &v->pause_flags); + return !TEST_BIT(_VPF_down, &v->pause_flags); } static inline bool is_xenstore_domain(const struct domain *d) -- 2.43.0
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |