[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH 12/14] xen-blkback: safely unmap grants in case they are still in use
On 14/01/15 16:22, Boris Ostrovsky wrote: > On 01/14/2015 11:00 AM, David Vrabel wrote: >> On 14/01/15 15:47, Boris Ostrovsky wrote: >>>> +static void xen_blkbk_unmap_and_respond_callback(int result, struct >>>> gntab_unmap_queue_data *data) >>>> +{ >>>> + struct pending_req* pending_req = (struct pending_req*) >>>> (data->data); >>>> + struct xen_blkif *blkif = pending_req->blkif; >>>> + >>>> + /* BUG_ON used to reproduce existing behaviour, >>>> + but is this the best way to deal with this? */ >>>> + BUG_ON(result); >>>> + /* free page after unmap */ >>>> + put_free_pages(blkif, data->pages, data->count); >>>> + >>>> + /* respond */ >>>> + make_response(blkif, pending_req->id, >>>> + pending_req->operation, pending_req->status); >>>> + free_req(blkif, pending_req); >>>> + /* >>>> + * Make sure the request is freed before releasing blkif, >>>> + * or there could be a race between free_req and the >>>> + * cleanup done in xen_blkif_free during shutdown. >>>> + * >>>> + * NB: The fact that we might try to wake up pending_free_wq >>>> + * before drain_complete (in case there's a drain going on) >>>> + * it's not a problem with our current implementation >>>> + * because we can assure there's no thread waiting on >>>> + * pending_free_wq if there's a drain going on, but it has >>>> + * to be taken into account if the current model is changed. >>>> + */ >>>> + if (atomic_dec_and_test(&blkif->inflight) && >>>> atomic_read(&blkif->drain)) { >>>> + complete(&blkif->drain_complete); >>>> + } >>>> + xen_blkif_put(blkif); >>>> +} >>>> + >>>> +static void xen_blkbk_unmap_populate(struct pending_req *req) >>>> +{ >>>> + >>>> + struct gntab_unmap_queue_data* work = &req->gnttab_unmap_data; >>>> + struct xen_blkif *blkif = req->blkif; >>>> + struct grant_page **pages = req->segments; >>>> + unsigned int i, invcount = 0; >>>> + >>>> + for (i = 0; i < req->nr_pages; i++) { >>>> + if (pages[i]->persistent_gnt != NULL) { >>>> + put_persistent_gnt(blkif, pages[i]->persistent_gnt); >>>> + continue; >>>> + } >>>> + if (pages[i]->handle == BLKBACK_INVALID_HANDLE) >>>> + continue; >>>> + >>>> + req->unmap_pages[invcount] = pages[i]->page; >>>> + gnttab_set_unmap_op(&req->unmap[invcount], >>>> vaddr(pages[invcount]->page), >>>> + GNTMAP_host_map, pages[i]->handle); >>>> + pages[i]->handle = BLKBACK_INVALID_HANDLE; >>>> + invcount++; >>>> + } >>>> + >>>> + work->data = req; >>>> + work->done = &xen_blkbk_unmap_and_respond_callback; >>>> + work->unmap_ops = req->unmap; >>>> + work->kunmap_ops = NULL; >>>> + work->pages = req->unmap_pages; >>>> + work->count = invcount; >>>> + >>>> +} >>>> + >>>> +static void xen_blkbk_unmap_and_respond(struct pending_req *req) >>>> +{ >>>> + xen_blkbk_unmap_populate(req); >>>> + gnttab_unmap_refs_async(&req->gnttab_unmap_data); >>>> +} >>> (One more for this patch) >>> >>> If you are using async unmap here, I thought new interface requires you >>> to wait for completion (which should be called from >>> xen_blkbk_unmap_and_respond_callback()). >> We complete the request in the done callback so we don't need to wait. >> > > > Is end_block_io_op() allowed to return before request is completed? Provided we retain a ref to the original bio, it should be fine I think. David _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx http://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |