[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v5 12/17] Update IRTE according to guest interrupt config changes
> -----Original Message----- > From: Konrad Rzeszutek Wilk [mailto:konrad.wilk@xxxxxxxxxx] > Sent: Thursday, August 13, 2015 12:43 AM > To: Wu, Feng > Cc: xen-devel@xxxxxxxxxxxxx; Jan Beulich > Subject: Re: [Xen-devel] [PATCH v5 12/17] Update IRTE according to guest > interrupt config changes > > On Wed, Aug 12, 2015 at 10:35:33AM +0800, Feng Wu wrote: > > When guest changes its interrupt configuration (such as, vector, etc.) > > for direct-assigned devices, we need to update the associated IRTE > > with the new guest vector, so external interrupts from the assigned > > devices can be injected to guests without VM-Exit. > > > > For lowest-priority interrupts, we use vector-hashing mechamisn to find > > the destination vCPU. This follows the hardware behavior, since modern > > Intel CPUs use vector hashing to handle the lowest-priority interrupt. > > > > For multicast/broadcast vCPU, we cannot handle it via interrupt posting, > > still use interrupt remapping. > > > > CC: Jan Beulich <jbeulich@xxxxxxxx> > > Signed-off-by: Feng Wu <feng.wu@xxxxxxxxx> > > --- > > v5: > > - Make 'struct vcpu *vcpu' const > > > > v4: > > - Make some 'int' variables 'unsigned int' in pi_find_dest_vcpu() > > - Make 'dest_id' uint32_t > > - Rename 'size' to 'bitmap_array_size' > > - find_next_bit() and find_first_bit() always return unsigned int, > > so no need to check whether the return value is less than 0. > > - Message error level XENLOG_G_WARNING -> XENLOG_G_INFO > > - Remove useless warning message > > - Create a seperate function vector_hashing_dest() to find the > > - destination of lowest-priority interrupts. > > - Change some comments > > > > v3: > > - Use bitmap to store the all the possible destination vCPUs of an > > interrupt, then trying to find the right destination from the bitmap > > - Typo and some small changes > > > > xen/drivers/passthrough/io.c | 124 > ++++++++++++++++++++++++++++++++++++++++++- > > 1 file changed, 123 insertions(+), 1 deletion(-) > > > > diff --git a/xen/drivers/passthrough/io.c b/xen/drivers/passthrough/io.c > > index bda9374..f62f86c 100644 > > --- a/xen/drivers/passthrough/io.c > > +++ b/xen/drivers/passthrough/io.c > > @@ -25,6 +25,7 @@ > > #include <asm/hvm/iommu.h> > > #include <asm/hvm/support.h> > > #include <xen/hvm/irq.h> > > +#include <asm/io_apic.h> > > > > static DEFINE_PER_CPU(struct list_head, dpci_list); > > > > @@ -198,6 +199,108 @@ void free_hvm_irq_dpci(struct hvm_irq_dpci *dpci) > > xfree(dpci); > > } > > > > +/* > > + * This routine handles lowest-priority interrupts using vector-hashing > > + * mechanism. As an example, modern Intel CPUs use this method to handle > > + * lowest-priority interrupts. > > + * > > + * Here is the details about the vector-hashing mechanism: > > + * 1. For lowest-priority interrupts, store all the possible destination > > + * vCPUs in an array. > > + * 2. Use "gvec % max number of destination vCPUs" to find the right > > + * destination vCPU in the array for the lowest-priority interrupt. > > + */ > > +static struct vcpu *vector_hashing_dest(const struct domain *d, > > + uint32_t dest_id, > > + bool_t dest_mode, > > + uint8_t gvec) > > + > > +{ > > + unsigned long *dest_vcpu_bitmap; > > + unsigned int dest_vcpu_num = 0, idx; > > + unsigned int bitmap_array_size = BITS_TO_LONGS(d->max_vcpus); > > + struct vcpu *v, *dest = NULL; > > + unsigned int i; > > + > > + dest_vcpu_bitmap = xzalloc_array(unsigned long, bitmap_array_size); > > + if ( !dest_vcpu_bitmap ) > > + { > > + dprintk(XENLOG_G_INFO, > > + "dom%d: failed to allocate memory\n", d->domain_id); > > + return NULL; > > + } > > + > > + for_each_vcpu ( d, v ) > > + { > > + if ( !vlapic_match_dest(vcpu_vlapic(v), NULL, 0, > > s/0/APIC_DEST_NOSHORT/ > > > + dest_id, dest_mode) ) > > + continue; > > + > > + __set_bit(v->vcpu_id, dest_vcpu_bitmap); > > + dest_vcpu_num++; > > Perhaps change the variable to: > > dest_vcpus > > ? > > + } > > + > > + if ( dest_vcpu_num != 0 ) > > + { > > + idx = 0; > > + > > + for ( i = gvec % dest_vcpu_num; i >= 0; i--) > > That loop is not good as it will overflow. > > Imagine gvec = 40, dest_vcpu_num = 2 > On first iteration i = 0, on the next i = -1 (aka 0xfffffff), and so on. > So we should define 'i' as int, right? Thanks, Feng > > + { > > + idx = find_next_bit(dest_vcpu_bitmap, d->max_vcpus, idx) + > 1; > > + BUG_ON(idx >= d->max_vcpus); > > + } > > + idx--; > > + > > + dest = d->vcpu[idx]; > > + } > > + > > + xfree(dest_vcpu_bitmap); > > + > > + return dest; > > +} > > + > > +/* > > + * The purpose of this routine is to find the right destination vCPU for > > + * an interrupt which will be delivered by VT-d posted-interrupt. There > > + * are several cases as below: > > + * > > + * - For lowest-priority interrupts, use vector-hashing mechanism to find > > + * the destination. > > + * - Otherwise, for single destination interrupt, it is straightforward to > > + * find the destination vCPU and return true. > > + * - For multicast/broadcast vCPU, we cannot handle it via interrupt > > posting, > > + * so return NULL. > > + */ > > +static struct vcpu *pi_find_dest_vcpu(const struct domain *d, uint32_t > dest_id, > > + bool_t dest_mode, uint8_t > delivery_mode, > > + uint8_t gvec) > > +{ > > + unsigned int dest_vcpu_num = 0; > > + struct vcpu *v, *dest = NULL; > > + > > + if ( delivery_mode == dest_LowestPrio ) > > + return vector_hashing_dest(d, dest_id, dest_mode, gvec); > > + > > + for_each_vcpu ( d, v ) > > + { > > + if ( !vlapic_match_dest(vcpu_vlapic(v), NULL, 0, > > s/0/APIC_DEST_NOSHORT/ > > + dest_id, dest_mode) ) > > + continue; > > + > > + dest_vcpu_num++; > > + dest = v; > > + } > > + > > + /* > > + * For fixed destination, we only handle single-destination > > + * interrupts. > > + */ > > + if ( dest_vcpu_num == 1 ) > > + return dest; > > Something is off with the tabs here. > > > + > > + return NULL; > > +} > > + > > int pt_irq_create_bind( > > struct domain *d, xen_domctl_bind_pt_irq_t *pt_irq_bind) > > { > > @@ -256,7 +359,7 @@ int pt_irq_create_bind( > > { > > case PT_IRQ_TYPE_MSI: > > { > > - uint8_t dest, dest_mode; > > + uint8_t dest, dest_mode, delivery_mode; > > int dest_vcpu_id; > > > > if ( !(pirq_dpci->flags & HVM_IRQ_DPCI_MAPPED) ) > > @@ -329,11 +432,30 @@ int pt_irq_create_bind( > > /* Calculate dest_vcpu_id for MSI-type pirq migration. */ > > dest = pirq_dpci->gmsi.gflags & VMSI_DEST_ID_MASK; > > dest_mode = !!(pirq_dpci->gmsi.gflags & VMSI_DM_MASK); > > + delivery_mode = (pirq_dpci->gmsi.gflags >> > GFLAGS_SHIFT_DELIV_MODE) & > > + VMSI_DELIV_MASK; > > dest_vcpu_id = hvm_girq_dest_2_vcpu_id(d, dest, dest_mode); > > pirq_dpci->gmsi.dest_vcpu_id = dest_vcpu_id; > > spin_unlock(&d->event_lock); > > if ( dest_vcpu_id >= 0 ) > > hvm_migrate_pirqs(d->vcpu[dest_vcpu_id]); > > + > > + /* Use interrupt posting if it is supported */ > > Missing full stop. > > > + if ( iommu_intpost ) > > + { > > + const struct vcpu *vcpu = pi_find_dest_vcpu(d, dest, > dest_mode, > > + delivery_mode, > pirq_dpci->gmsi.gvec); > > + > > + if ( vcpu ) > > + { > > + rc = pi_update_irte( vcpu, info, pirq_dpci->gmsi.gvec ); > > + if ( unlikely(rc) ) > > + dprintk(XENLOG_G_INFO, > > + "%pv: failed to update PI IRTE, > gvec:%02x\n", > > + vcpu, pirq_dpci->gmsi.gvec); > > + } > > + } > > + > > break; > > } > > > > -- > > 2.1.0 > > > > > > _______________________________________________ > > Xen-devel mailing list > > Xen-devel@xxxxxxxxxxxxx > > http://lists.xen.org/xen-devel _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |