At 14:49 +0100 on 23 Jul (1279896553), Tim Deegan wrote:
> There are a few places in Xen where we walk a domain's page lists
> without holding the page_alloc lock. They race with updates to the page
> lists, which are normally rare but can be quite common under PoD when
> the domain is close to its memory limit and the PoD reclaimer is busy.
> This patch protects those places by taking the page_alloc lock.
I should say that the other place I found is in construct_dom0(), which
I left because it (a) can't really race with allocations and (b) calls
process_pending_softirqs() within the page_list_for_each().
Tim.
> I think this is OK for the two debug-key printouts - they don't run from
> irq context and look deadlock-free. The tboot change seems safe too
> unless tboot shutdown functions are called from irq context or with the
> page_alloc lock held. The p2m one is the scariest but there are already
> code paths in PoD that take the page_alloc lock with the p2m lock held
> so it's no worse than existing code.
>
> Signed-off-by: Tim Deegan <Tim.Deegan@xxxxxxxxxx>
>
> diff -r e8dbc1262f52 xen/arch/x86/domain.c
> --- a/xen/arch/x86/domain.c Wed Jul 21 09:02:10 2010 +0100
> +++ b/xen/arch/x86/domain.c Fri Jul 23 14:33:22 2010 +0100
> @@ -139,12 +139,14 @@ void dump_pageframe_info(struct domain *
> }
> else
> {
> + spin_lock(&d->page_alloc_lock);
> page_list_for_each ( page, &d->page_list )
> {
> printk(" DomPage %p: caf=%08lx, taf=%" PRtype_info "\n",
> _p(page_to_mfn(page)),
> page->count_info, page->u.inuse.type_info);
> }
> + spin_unlock(&d->page_alloc_lock);
> }
>
> if ( is_hvm_domain(d) )
> @@ -152,12 +154,14 @@ void dump_pageframe_info(struct domain *
> p2m_pod_dump_data(d);
> }
>
> + spin_lock(&d->page_alloc_lock);
> page_list_for_each ( page, &d->xenpage_list )
> {
> printk(" XenPage %p: caf=%08lx, taf=%" PRtype_info "\n",
> _p(page_to_mfn(page)),
> page->count_info, page->u.inuse.type_info);
> }
> + spin_unlock(&d->page_alloc_lock);
> }
>
> struct domain *alloc_domain_struct(void)
> diff -r e8dbc1262f52 xen/arch/x86/mm/p2m.c
> --- a/xen/arch/x86/mm/p2m.c Wed Jul 21 09:02:10 2010 +0100
> +++ b/xen/arch/x86/mm/p2m.c Fri Jul 23 14:33:22 2010 +0100
> @@ -1833,6 +1833,7 @@ int p2m_alloc_table(struct domain *d,
> goto error;
>
> /* Copy all existing mappings from the page list and m2p */
> + spin_lock(&d->page_alloc_lock);
> page_list_for_each(page, &d->page_list)
> {
> mfn = page_to_mfn(page);
> @@ -1848,13 +1849,16 @@ int p2m_alloc_table(struct domain *d,
> #endif
> && gfn != INVALID_M2P_ENTRY
> && !set_p2m_entry(d, gfn, mfn, 0, p2m_ram_rw) )
> - goto error;
> + goto error_unlock;
> }
> + spin_unlock(&d->page_alloc_lock);
>
> P2M_PRINTK("p2m table initialised (%u pages)\n", page_count);
> p2m_unlock(p2m);
> return 0;
>
> +error_unlock:
> + spin_unlock(&d->page_alloc_lock);
> error:
> P2M_PRINTK("failed to initialize p2m table, gfn=%05lx, mfn=%"
> PRI_mfn "\n", gfn, mfn_x(mfn));
> diff -r e8dbc1262f52 xen/arch/x86/numa.c
> --- a/xen/arch/x86/numa.c Wed Jul 21 09:02:10 2010 +0100
> +++ b/xen/arch/x86/numa.c Fri Jul 23 14:33:22 2010 +0100
> @@ -385,11 +385,13 @@ static void dump_numa(unsigned char key)
> for_each_online_node(i)
> page_num_node[i] = 0;
>
> + spin_lock(&d->page_alloc_lock);
> page_list_for_each(page, &d->page_list)
> {
> i = phys_to_nid((paddr_t)page_to_mfn(page) <<
> PAGE_SHIFT);
> page_num_node[i]++;
> }
> + spin_unlock(&d->page_alloc_lock);
>
> for_each_online_node(i)
> printk(" Node %u: %u\n", i, page_num_node[i]);
> diff -r e8dbc1262f52 xen/arch/x86/tboot.c
> --- a/xen/arch/x86/tboot.c Wed Jul 21 09:02:10 2010 +0100
> +++ b/xen/arch/x86/tboot.c Fri Jul 23 14:33:22 2010 +0100
> @@ -211,12 +211,14 @@ static void tboot_gen_domain_integrity(c
> continue;
> printk("MACing Domain %u\n", d->domain_id);
>
> + spin_lock(&d->page_alloc_lock);
> page_list_for_each(page, &d->page_list)
> {
> void *pg = __map_domain_page(page);
> vmac_update(pg, PAGE_SIZE, &ctx);
> unmap_domain_page(pg);
> }
> + spin_unlock(&d->page_alloc_lock);
>
> if ( !is_idle_domain(d) )
> {
>
> _______________________________________________
> Xen-devel mailing list
> Xen-devel@xxxxxxxxxxxxxxxxxxx
> http://lists.xensource.com/xen-devel
--
Tim Deegan <Tim.Deegan@xxxxxxxxxx>
Principal Software Engineer, XenServer Engineering
Citrix Systems UK Ltd. (Company #02937203, SL9 0BG)
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel
|