[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v2] mem_access: sanitize code around sending vm_event request
On 03/08/16 16:18, Tamas K Lengyel wrote: > On Wed, Aug 3, 2016 at 8:41 AM, George Dunlap <george.dunlap@xxxxxxxxxx> > wrote: >> On 01/08/16 17:52, Tamas K Lengyel wrote: >>> The two functions monitor_traps and mem_access_send_req duplicate some of >>> the >>> same functionality. The mem_access_send_req however leaves a lot of the >>> standard vm_event fields to be filled by other functions. >>> >>> Remove mem_access_send_req() completely, making use of monitor_traps() to >>> put >>> requests into the monitor ring. This in turn causes some cleanup around the >>> old callsites of mem_access_send_req(), and on ARM, the introduction of the >>> __p2m_mem_access_send_req() helper to fill in common mem_access information. >>> We also update monitor_traps to now include setting the common vcpu_id field >>> so that all other call-sites can ommit this step. >>> >>> Finally, this change identifies that errors from mem_access_send_req() were >>> never checked. As errors constitute a problem with the monitor ring, >>> crashing the domain is the most appropriate action to take. >>> >>> Signed-off-by: Tamas K Lengyel <tamas.lengyel@xxxxxxxxxxxx> >>> Reviewed-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx> >> >> This appears to be v3, not v2? > > No, it's still just v2. > >> >>> diff --git a/xen/arch/x86/mm/p2m.c b/xen/arch/x86/mm/p2m.c >>> index 812dbf6..27f9d26 100644 >>> --- a/xen/arch/x86/mm/p2m.c >>> +++ b/xen/arch/x86/mm/p2m.c >>> @@ -1728,13 +1728,8 @@ bool_t p2m_mem_access_check(paddr_t gpa, unsigned >>> long gla, >>> if ( req ) >>> { >>> *req_ptr = req; >>> - req->reason = VM_EVENT_REASON_MEM_ACCESS; >>> - >>> - /* Pause the current VCPU */ >>> - if ( p2ma != p2m_access_n2rwx ) >>> - req->flags |= VM_EVENT_FLAG_VCPU_PAUSED; >>> >>> - /* Send request to mem event */ >>> + req->reason = VM_EVENT_REASON_MEM_ACCESS; >>> req->u.mem_access.gfn = gfn; >>> req->u.mem_access.offset = gpa & ((1 << PAGE_SHIFT) - 1); >>> if ( npfec.gla_valid ) >>> @@ -1750,23 +1745,10 @@ bool_t p2m_mem_access_check(paddr_t gpa, unsigned >>> long gla, >>> req->u.mem_access.flags |= npfec.read_access ? MEM_ACCESS_R : 0; >>> req->u.mem_access.flags |= npfec.write_access ? MEM_ACCESS_W : 0; >>> req->u.mem_access.flags |= npfec.insn_fetch ? MEM_ACCESS_X : 0; >>> - req->vcpu_id = v->vcpu_id; >>> - >>> - vm_event_fill_regs(req); >>> - >>> - if ( altp2m_active(v->domain) ) >>> - { >>> - req->flags |= VM_EVENT_FLAG_ALTERNATE_P2M; >>> - req->altp2m_idx = vcpu_altp2m(v).p2midx; >>> - } >>> } >>> >>> - /* Pause the current VCPU */ >>> - if ( p2ma != p2m_access_n2rwx ) >>> - vm_event_vcpu_pause(v); >>> - >>> - /* VCPU may be paused, return whether we promoted automatically */ >>> - return (p2ma == p2m_access_n2rwx); >>> + /* Return whether vCPU pause is required (aka. sync event) */ >>> + return (p2ma != p2m_access_n2rwx); >>> } >>> >>> static inline >> >> p2m-bits: >> >> Acked-by: George Dunlap <george.dunlap@xxxxxxxxxx> >> >> But I agree with Julien -- this patch has several independent changes >> which makes it quite difficult to tell what's going on. I'm sure it's >> taken the two of us a lot more time together to figure out what is and >> is not happening than it would have for you to break it down into >> several little chunks. >> >> If you're not already familiar with it, I would recommend looking into >> stackgit. My modus operandi for things like this is to get things >> working in one big patch, then pop it off the stack and apply bits of it >> at a time to make a series. >> >> It's not only more considerate of your reviewers, but it's also a >> helpful exercise for yourself. >> > > The extra work doesn't just come from splitting the code itself > (although I don't know which bits would really make sense to split > here that would worth the effort) but testing a series on various > platforms. I don't understand this statement -- why is testing a 3-patch series more difficult than testing a one-patch series? Are you testing each individual patch? -George _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |