[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Xen-devel] [Patch RFC 11/13] vt-d: If the Device-TLB flush is still not completed when
 
- To: Quan Xu <quan.xu@xxxxxxxxx>, <andrew.cooper3@xxxxxxxxxx>,	<eddie.dong@xxxxxxxxx>, <ian.campbell@xxxxxxxxxx>,	<ian.jackson@xxxxxxxxxxxxx>, <jbeulich@xxxxxxxx>,	<jun.nakajima@xxxxxxxxx>, 	<keir@xxxxxxx>, <kevin.tian@xxxxxxxxx>, <tim@xxxxxxx>,	<yang.z.zhang@xxxxxxxxx>, <george.dunlap@xxxxxxxxxxxxx>
 
- From: Julien Grall <julien.grall@xxxxxxxxxx>
 
- Date: Wed, 16 Sep 2015 10:56:29 +0100
 
- Cc: xen-devel@xxxxxxxxxxxxx
 
- Delivery-date: Wed, 16 Sep 2015 09:56:46 +0000
 
- List-id: Xen developer discussion <xen-devel.lists.xen.org>
 
 
 
Hi Quan,
On 16/09/2015 14:24, Quan Xu wrote:
 
to destroy virtual machine, schedule and wait on a waitqueue
until the Device-TLB flush is completed.
Signed-off-by: Quan Xu <quan.xu@xxxxxxxxx>
---
  xen/common/domain.c                 | 10 ++++++++++
  xen/drivers/passthrough/vtd/iommu.c |  9 +++++++++
  xen/include/xen/hvm/iommu.h         |  6 ++++++
  3 files changed, 25 insertions(+)
 
 
Same remarks as the previous patches.
 
diff --git a/xen/common/domain.c b/xen/common/domain.c
index 1f62e3b..8ccc1a5 100644
--- a/xen/common/domain.c
+++ b/xen/common/domain.c
@@ -867,6 +867,16 @@ void domain_destroy(struct domain *d)
      rcu_assign_pointer(*pd, d->next_in_hashbucket);
      spin_unlock(&domlist_update_lock);
+#ifdef HAS_PASSTHROUGH
+    /*
+     * If the Device-TLB flush is still not completed, schedule
+     * and wait on a waitqueue until the Device-TLB flush is
+     * completed.
+     */
+    if ( need_iommu(d) && QI_FLUSHING(d) )
+        wait_for_qi_flushing(d);
+#endif
+
      /* Schedule RCU asynchronous completion of domain destroy. */
      call_rcu(&d->rcu, complete_domain_destroy);
  }
diff --git a/xen/drivers/passthrough/vtd/iommu.c 
b/xen/drivers/passthrough/vtd/iommu.c
index 1297dea..3d98fea 100644
--- a/xen/drivers/passthrough/vtd/iommu.c
+++ b/xen/drivers/passthrough/vtd/iommu.c
@@ -1070,6 +1070,11 @@ static hw_irq_controller dma_msi_type = {
  };
  /* IOMMU Queued Invalidation(QI). */
+void wait_for_qi_flushing(struct domain *d)
+{
+    wait_event(qi_wq(d), !QI_FLUSHING(d));
+}
+
  static void qi_clear_iwc(struct iommu *iommu)
  {
      unsigned long flags;
@@ -1188,6 +1193,7 @@ scan_again:
                  }
                  spin_unlock(&qi_page_lock(d));
                  QI_FLUSHING(d) = 0;
+                wake_up_all(&qi_wq(d));
              }
              rcu_unlock_domain(d);
          }
@@ -1494,6 +1500,7 @@ static int intel_iommu_domain_init(struct domain *d)
      hd->arch.agaw = width_to_agaw(DEFAULT_DOMAIN_ADDRESS_WIDTH);
      INIT_PAGE_LIST_HEAD(&qi_hold_page_list(d));
      spin_lock_init(&qi_page_lock(d));
+    init_waitqueue_head(&qi_wq(d));
      return 0;
  }
@@ -1925,6 +1932,8 @@ static void iommu_domain_teardown(struct domain *d)
      if ( list_empty(&acpi_drhd_units) )
          return;
+    destroy_waitqueue_head(&qi_wq(d));
+
      list_for_each_entry_safe ( mrmrr, tmp, &hd->arch.mapped_rmrrs, list )
      {
          list_del(&mrmrr->list);
diff --git a/xen/include/xen/hvm/iommu.h b/xen/include/xen/hvm/iommu.h
index 5dc0033..f661c8c 100644
--- a/xen/include/xen/hvm/iommu.h
+++ b/xen/include/xen/hvm/iommu.h
@@ -20,6 +20,7 @@
  #define __XEN_HVM_IOMMU_H__
  #include <xen/iommu.h>
+#include <xen/wait.h>
  #include <xen/list.h>
  #include <asm/hvm/iommu.h>
@@ -56,12 +57,15 @@ struct hvm_iommu {
      struct page_list_head qi_hold_page_list;
      spinlock_t qi_lock;
+    struct waitqueue_head qi_wq;
+
      /* Features supported by the IOMMU */
      DECLARE_BITMAP(features, IOMMU_FEAT_count);
  };
  void do_qi_flushing(struct domain *d);
  void qi_hold_page(struct domain *d, struct page_info *pg);
+void wait_for_qi_flushing(struct domain *d);
  #define iommu_set_feature(d, f)   set_bit((f), domain_hvm_iommu(d)->features)
  #define iommu_clear_feature(d, f) clear_bit((f), 
domain_hvm_iommu(d)->features)
@@ -76,5 +80,7 @@ void qi_hold_page(struct domain *d, struct page_info *pg);
      (d->arch.hvm_domain.hvm_iommu.qi_hold_page_list)
  #define qi_page_lock(d) \
      (d->arch.hvm_domain.hvm_iommu.qi_lock)
+#define qi_wq(d) \
+    (d->arch.hvm_domain.hvm_iommu.qi_wq)
  #endif /* __XEN_HVM_IOMMU_H__ */
 
Regards,
--
Julien Grall
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
 
 
    
     |