|  |  | 
  
    |  |  | 
 
  |   |  | 
  
    |  |  | 
  
    |  |  | 
  
    |   xen-devel
[Xen-devel] Re: Question about x86/mm/gup.c's use of disabled	interrupts 
| 
Jeremy Fitzhardinge wrote:
 
Avi Kivity wrote:
 
Hm, awkward if flush_tlb_others doesn't IPI...
 
How can it avoid flushing the tlb on cpu [01]?  It's it's 
gup_fast()ing a pte, it may as well load it into the tlb.
 
xen_flush_tlb_others uses a hypercall rather than an IPI, so none of 
the logic which depends on there being an IPI will work.
 
Right, of course, that's what we were talking about.  I thought 
optimizations to avoid IPIs if an mm never visited a cpu. 
 
 Simplest fix is to make gup_get_pte() a pvop, but that does seem 
like putting a red flag in front of an inner-loop hotspot, or 
something...
The per-cpu tlb-flush exclusion flag might really be the way to go.
 
I don't see how it will work, without changing Xen to look at the flag?
local_irq_disable() is used here to lock out a remote cpu, I don't 
see why deferring the flush helps.
 
Well, no, not deferring.  Making xen_flush_tlb_others() spin waiting 
for "doing_gup" to clear on the target cpu.  Or add an explicit notion 
of a "pte update barrier" rather than implicitly relying on the tlb 
IPI (which is extremely convenient when available...).
 
Pick up a percpu flag from all cpus and spin on each?  Nasty.
You could use the irq enabled flag; it's available and what native spins 
on (but also means I'll need to add one if I implement this). 
--
I have a truly marvellous patch that fixes the bug which this
signature is too narrow to contain.
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
 | 
 |  | 
  
    |  |  |