WARNING - OLD ARCHIVES

This is an archived copy of the Xen.org mailing list, which we have preserved to ensure that existing links to archives are not broken. The live archive, which contains the latest emails, can be found at http://lists.xen.org/
   
 
 
Xen 
 
Home Products Support Community News
 
   
 

xen-devel

[Xen-devel] Re: [PATCH] Fix cache flush bug of cpu offline

To: "Liu, Jinsong" <jinsong.liu@xxxxxxxxx>, "xen-devel@xxxxxxxxxxxxxxxxxxx" <xen-devel@xxxxxxxxxxxxxxxxxxx>
Subject: [Xen-devel] Re: [PATCH] Fix cache flush bug of cpu offline
From: Keir Fraser <keir.xen@xxxxxxxxx>
Date: Fri, 11 Mar 2011 19:10:22 +0000
Cc: "Jiang, Yunhong" <yunhong.jiang@xxxxxxxxx>, "Shan, Haitao" <haitao.shan@xxxxxxxxx>, "Wei, Gang" <gang.wei@xxxxxxxxx>, "Yu, Ke" <ke.yu@xxxxxxxxx>, "Li, Xin" <xin.li@xxxxxxxxx>
Delivery-date: Fri, 11 Mar 2011 11:11:50 -0800
Dkim-signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=gamma; h=domainkey-signature:user-agent:date:subject:from:to:cc:message-id :thread-topic:thread-index:in-reply-to:mime-version:content-type :content-transfer-encoding; bh=D51EmkWU1t8HJTqgLy2ckR49S7J+vx8vDa7ogn/Qlls=; b=ShAsOc/q5wPEmicf5RSX0d42zdk9fp3rwWQmCZGN0f+lXuqXJITmtuYkpVA3zRNb5y EOkbQOt4V7X0v+CMSLq6x4iVt3dlwZ9yy9Nr+qioIz4a/DUPQj3zW8zKigYUP6st6IKj ZYTcW8BAAWaD+xO2eUNZ7L2jrqGVeRGu4Jjzc=
Domainkey-signature: a=rsa-sha1; c=nofws; d=gmail.com; s=gamma; h=user-agent:date:subject:from:to:cc:message-id:thread-topic :thread-index:in-reply-to:mime-version:content-type :content-transfer-encoding; b=sX5OW54iMUX4DdJJvIxiVXDb/NMulv+qhS+eEuFzyK3ACtW7Dn2QhDUULWbjbJRj4J C0eUboIGrAMBx3PG7MitWnHLQjHC3U2XI/GXezXYEnJRU7XpBXb+WHbWgPIotwB+zZK0 MWhW+OmKPiFtmD6tvz8udqlsLtolKdW4DloQ0=
Envelope-to: www-data@xxxxxxxxxxxxxxxxxxx
In-reply-to: <BC00F5384FCFC9499AF06F92E8B78A9E1FCCF29CDD@xxxxxxxxxxxxxxxxxxxxxxxxxxxx>
List-help: <mailto:xen-devel-request@lists.xensource.com?subject=help>
List-id: Xen developer discussion <xen-devel.lists.xensource.com>
List-post: <mailto:xen-devel@lists.xensource.com>
List-subscribe: <http://lists.xensource.com/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=subscribe>
List-unsubscribe: <http://lists.xensource.com/mailman/listinfo/xen-devel>, <mailto:xen-devel-request@lists.xensource.com?subject=unsubscribe>
Sender: xen-devel-bounces@xxxxxxxxxxxxxxxxxxx
Thread-index: Acvf+Wq8eUV5g7VvTPu6ARIzvUA0UgAD5esZAAA/DUAAAffKGQAAcZpAAAMU9s8=
Thread-topic: [PATCH] Fix cache flush bug of cpu offline
User-agent: Microsoft-Entourage/12.28.0.101117
On 11/03/2011 17:53, "Liu, Jinsong" <jinsong.liu@xxxxxxxxx> wrote:

> Keir Fraser wrote:
>> On 11/03/2011 16:50, "Liu, Jinsong" <jinsong.liu@xxxxxxxxx> wrote:
>> 
>>> we did experiment, if did wbinvd at current position (at play_dead),
>>> sometimes it bring strange issue when repeatly cpu offline/online.
>>> so for cpu dead, the near wbinvd to last step, the safer.
>>> wbinvd would better be the last ops before cpu dead, to avoid
>>> potential cache coherency break.
>> 
>> Okay, I applied your patches. However in a follow-up patch (c/s
>> 23025) I have removed the WBINVD instructions from the default paths
>> (i.e., the HLT loops) as the CPU still does cache coherency while in
>> HLT/C1 state. 
>> 
>> Does that look okay to you?
>> 
>>  -- Keir
> 
> It's right that cpu continue snoop bus transaction to keep cache coherency
> when HLT/C1.
> 
> However, I think it better not remove wbinvd from HLT loops.
> I'm not quite sure if 'enter Cx' == 'cpu play dead'. After all, resume from Cx
> different from cpu booting: when cpu online, it start from SIPI-SIPI.
> I worried some unknown hardware behavior will bring issue if we treat 'cpu
> play dead' totally same as 'enter Cx'.
> Maybe that's the reason why kernel insist on reserveing wbinvd before HLT when
> play dead.

An APIC INIT doesn't touch the memory hierarchy, cache controls, nor CR0.CD
and CR0.NW. The care over cache invalidation I believe solely comes down to
the deep sleep that the CPU can be placed into via MWAIT, in which the
cache-control logic gets switched off.

 -- Keir

> Thanks,
> Jinsong
> 
>> 
>>> In fact, it can do wbinvd inside loop, but as cpu_offline_3.patch
>>> said, 
>>> at Xen 7400 when hyperthreading, the offlined thread may be
>>> spuriously waken up by its brother, and frequently waken inside the
>>> dead loop. 
>>> In such case, considering heavy workload of wbinvd, we add a
>>> light-weight clflush instruction inside loop.
>>> 
>>> Thanks,
>>> Jinsong
>>> 
>>> 
>>>> 
>>>>> Signed-off-by: Liu, Jinsong <jinsong.liu@xxxxxxxxx>
>>>>> 
>>>>> diff -r 2dc3c1cc1bba xen/arch/x86/acpi/cpu_idle.c
>>>>> --- a/xen/arch/x86/acpi/cpu_idle.c Mon Mar 07 05:31:46 2022 +0800
>>>>> +++ b/xen/arch/x86/acpi/cpu_idle.c Thu Mar 10 23:40:51 2022 +0800
>>>>> @@ -562,11 +562,14 @@ static void acpi_dead_idle(void)
>>>>>      if ( (cx = &power->states[power->count-1]) == NULL )
>>>>> goto default_halt;
>>>>> 
>>>>> +    /*
>>>>> +     * cache must be flashed as the last ops before cpu going into
>>>>> dead, +     * otherwise, cpu may dead with dirty data breaking
>>>>> cache coherency, +     * leading to strange errors.
>>>>> +     */
>>>>> +    wbinvd();
>>>>>      for ( ; ; )
>>>>>      {
>>>>> -        if ( !power->flags.bm_check && cx->type == ACPI_STATE_C3 )
>>>>> -            ACPI_FLUSH_CPU_CACHE();
>>>>> -
>>>>>          switch ( cx->entry_method )
>>>>>          {
>>>>>              case ACPI_CSTATE_EM_FFH:
>>>>> @@ -584,6 +587,7 @@ static void acpi_dead_idle(void)      }
>>>>> 
>>>>>  default_halt:
>>>>> +    wbinvd();
>>>>>      for ( ; ; )
>>>>>          halt();
>>>>>  }
>>>>> diff -r 2dc3c1cc1bba xen/arch/x86/domain.c
>>>>> --- a/xen/arch/x86/domain.c Mon Mar 07 05:31:46 2022 +0800
>>>>> +++ b/xen/arch/x86/domain.c Thu Mar 10 23:40:51 2022 +0800
>>>>> @@ -93,6 +93,12 @@ static void default_idle(void)
>>>>> 
>>>>>  static void default_dead_idle(void)
>>>>>  {
>>>>> +    /*
>>>>> +     * cache must be flashed as the last ops before cpu going into
>>>>> dead, +     * otherwise, cpu may dead with dirty data breaking
>>>>> cache coherency, +     * leading to strange errors.
>>>>> +     */
>>>>> +    wbinvd();
>>>>>      for ( ; ; )
>>>>>          halt();
>>>>>  }
>>>>> @@ -100,7 +106,6 @@ static void play_dead(void)
>>>>>  static void play_dead(void)
>>>>>  {
>>>>>      local_irq_disable();
>>>>> -    wbinvd();
>>>>> 
>>>>>      /*
>>>>>       * NOTE: After cpu_exit_clear, per-cpu variables are no longer
>>>>> accessible,
> 



_______________________________________________
Xen-devel mailing list
Xen-devel@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-devel