|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCHv1 2/2] passthrough: improve locking when iterating over interrupts bound to VMs
On Fri, Oct 23, 2015 at 12:05:22PM +0100, David Vrabel wrote:
> radix_tree_gang_lookup() only requires a RCU read lock, not the
> per-domain event_lock.
Don't you need to make some form of 'spin_lock_init' call?
>
> Introduce a new RCU read lock and take the per-interrupt lock before
> calling the callback instead.
>
> This eliminates all contention on the event_lock when injecting
> interrupts from passthrough devices.
This pt_pirq_iterate is called from:
hvm_migrate_pirqs (which only uses irq_desc lock) - so you could
also drop the event_lock there.
pt_irq_time_out (which is only called for legacy interrupts and
only at 8ms). Uses event_lock.
pci_clean_dpci_irqs (during domain shutdown, uses event_lock).
Also should this rcu lock be put in 'pirq_guest_unmask' (which oddly
enough has no lock?!)?
Reading the radix-tree.h t says: "
..
The notable exceptions to this rule are the following functions:
...
radix_tree_gang_lookup
The first 7 functions are able to be called locklessly, using RCU. The
caller must ensure calls to these functions are made within rcu_read_lock()
regions."
Great, so that is what your patch does. Thought it would imply that
everybody using radix_tree_gang_lookup on pirq_tree needs to use RCU?
>
> Signed-off-by: David Vrabel <david.vrabel@xxxxxxxxxx>
> ---
> xen/drivers/passthrough/io.c | 10 +++++++---
> xen/include/xen/sched.h | 1 +
> 2 files changed, 8 insertions(+), 3 deletions(-)
>
> diff --git a/xen/drivers/passthrough/io.c b/xen/drivers/passthrough/io.c
> index 7c86c20..9b51ef0 100644
> --- a/xen/drivers/passthrough/io.c
> +++ b/xen/drivers/passthrough/io.c
> @@ -601,7 +601,7 @@ int pt_pirq_iterate(struct domain *d,
> unsigned int pirq = 0, n, i;
> struct pirq *pirqs[8];
>
> - ASSERT(spin_is_locked(&d->event_lock));
> + rcu_read_lock(&d->pirq_rcu_lock);
>
> do {
> n = radix_tree_gang_lookup(&d->pirq_tree, (void **)pirqs, pirq,
> @@ -610,12 +610,18 @@ int pt_pirq_iterate(struct domain *d,
> {
> struct hvm_pirq_dpci *pirq_dpci = pirq_dpci(pirqs[i]);
>
> + spin_lock(&pirq_dpci->lock);
> +
> pirq = pirqs[i]->pirq;
> if ( (pirq_dpci->flags & HVM_IRQ_DPCI_MAPPED) )
> rc = cb(d, pirq_dpci, arg);
> +
> + spin_unlock(&pirq_dpci->lock);
> }
> } while ( !rc && ++pirq < d->nr_pirqs && n == ARRAY_SIZE(pirqs) );
>
> + rcu_read_unlock(&d->pirq_rcu_lock);
> +
> return rc;
> }
>
> @@ -678,9 +684,7 @@ void hvm_dpci_msi_eoi(struct domain *d, int vector)
> if ( !iommu_enabled || !d->arch.hvm_domain.irq.dpci )
> return;
>
> - spin_lock(&d->event_lock);
> pt_pirq_iterate(d, _hvm_dpci_msi_eoi, (void *)(long)vector);
> - spin_unlock(&d->event_lock);
> }
>
> static void hvm_dirq_assist(struct domain *d, struct hvm_pirq_dpci
> *pirq_dpci)
> diff --git a/xen/include/xen/sched.h b/xen/include/xen/sched.h
> index 3729b0f..ae98c1e 100644
> --- a/xen/include/xen/sched.h
> +++ b/xen/include/xen/sched.h
> @@ -355,6 +355,7 @@ struct domain
> */
> struct radix_tree_root pirq_tree;
> unsigned int nr_pirqs;
> + rcu_read_lock_t pirq_rcu_lock;
>
> enum guest_type guest_type;
>
> --
> 2.1.4
>
>
> _______________________________________________
> Xen-devel mailing list
> Xen-devel@xxxxxxxxxxxxx
> http://lists.xen.org/xen-devel
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |