|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [Xen-devel] [PATCH v2 3/3] xen: use idle vcpus to scrub pages
On 07/02/2014 02:27 PM, Bob Liu wrote:
>
> On 07/01/2014 08:59 PM, Jan Beulich wrote:
>>>>> On 01.07.14 at 14:25, <bob.liu@xxxxxxxxxx> wrote:
>>> On 07/01/2014 05:12 PM, Jan Beulich wrote:
>>>>>>> On 30.06.14 at 15:39, <lliubbo@xxxxxxxxx> wrote:
>>>>> @@ -948,6 +954,7 @@ static void free_heap_pages(
>>>>> {
>>>>> if ( !tainted )
>>>>> {
>>>>> + node_need_scrub[node] = 1;
>>>>> for ( i = 0; i < (1 << order); i++ )
>>>>> pg[i].count_info |= PGC_need_scrub;
>>>>> }
>>>>
>>>> Iirc it was more than this single place where you set
>>>> PGC_need_scrub, and hence where you'd now need to set the
>>>> other flag too.
>>>>
>>>
>>> I'm afraid this is the only place where PGC_need_scrub was set.
>>
>> Ah, indeed - I misremembered others, they are all tests for the flag.
>>
>>> I'm sorry for all of the coding style problems.
>>>
>>> By the way is there any script which can be used to check the code
>>> before submitting? Something like ./scripts/checkpatch.pl under linux.
>>
>> No, there isn't. But avoiding (or spotting) hard tabs should be easy
>> enough, and other things you ought to simply inspect your patch for
>> - after all that's no different from what reviewers do.
>>
>>>>> + }
>>>>> +
>>>>> + /* free percpu free list */
>>>>> + if ( !page_list_empty(local_free_list) )
>>>>> + {
>>>>> + spin_lock(&heap_lock);
>>>>> + page_list_for_each_safe( pg, tmp, local_free_list )
>>>>> + {
>>>>> + order = PFN_ORDER(pg);
>>>>> + page_list_del(pg, local_free_list);
>>>>> + for ( i = 0; i < (1 << order); i++ )
>>>>> + {
>>>>> + pg[i].count_info |= PGC_state_free;
>>>>> + pg[i].count_info &= ~PGC_need_scrub;
>>>>
>>>> This needs to happen earlier - the scrub flag should be cleared right
>>>> after scrubbing, and the free flag should imo be set when the page
>>>> gets freed. That's for two reasons:
>>>> 1) Hypervisor allocations don't need scrubbed pages, i.e. they can
>>>> allocate memory regardless of the scrub flag's state.
>>>
>>> AFAIR, the reason I set those flags here is to avoid a panic happen.
>>
>> That's pretty vague a statement.
>>
>>>> 2) You still detain the memory on the local lists from allocation. On a
>>>> many-node system, the 16Mb per node can certainly sum up (which
>>>> is not to say that I don't view the 16Mb on a single node as already
>>>> problematic).
>>>
>>> Right, but we can adjust SCRUB_BATCH_ORDER.
>>> Anyway I'll take a retry as you suggested.
>>
>> You should really drop the idea of removing pages temporarily.
>> All you need to do is make sure a page being allocated and getting
>> simultaneously scrubbed by another CPU won't get passed to the
>> caller until the scrubbing finished. In particular it's no problem if
>> the allocating CPU occasionally ends up scrubbing a page already
>> being scrubbed elsewhere.
>>
>
> Yes, I also like to drop percpu lists which can make things simper. But
> I'm afraid which also means I can't use any spinlock(&heap_lock) any
> more because of potential heavy lock contentions. I'm not sure whether
> things can work fine without heap_lock.
>
In my attempt to get rid of heap_lock, there was a panic happen when
iterating the heap free list. My implementation is like this:
scrub_free_pages()
{
for ( zone = 0; zone < NR_ZONES; zone++ )
{
for ( order = MAX_ORDER; order >= 0; order-- )
{
page_list_for_each_safe( pg, tmp, &heap(node, zone, order) )
{
if ( !test_bit(_PGC_need_scrub, &(pg->count_info)) )
continue;
for ( i = 0; i < (1 << order); i++ )
{
if ( test_bit(_PGC_need_scrub, &(pg->count_info)) )
{
scrub_one_page(&pg[i]);
clear_bit(_PGC_need_scrub, &pg[i].count_info);
}
}
}
}
}
}
The panic was in page_list_next().
I didn't find a good way to iterate the free list without holding
heap_lock, but if holding the lock it might be heavy lock contention
then I have to remove pages temporarily from heap free list to a percpu
list.
--
Regards,
-Bob
_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxx
http://lists.xen.org/xen-devel
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |