[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [PATCH] x86/boot: Fix load_system_tables() to be NMI/#MC-safe



On 27/05/2020 14:19, Jan Beulich wrote:
> On 27.05.2020 15:06, Andrew Cooper wrote:
>> @@ -720,30 +721,26 @@ void load_system_tables(void)
>>              .limit = (IDT_ENTRIES * sizeof(idt_entry_t)) - 1,
>>      };
>>  
>> -    *tss = (struct tss64){
>> -            /* Main stack for interrupts/exceptions. */
>> -            .rsp0 = stack_bottom,
>> -
>> -            /* Ring 1 and 2 stacks poisoned. */
>> -            .rsp1 = 0x8600111111111111ul,
>> -            .rsp2 = 0x8600111111111111ul,
>> -
>> -            /*
>> -             * MCE, NMI and Double Fault handlers get their own stacks.
>> -             * All others poisoned.
>> -             */
>> -            .ist = {
>> -                    [IST_MCE - 1] = stack_top + IST_MCE * PAGE_SIZE,
>> -                    [IST_DF  - 1] = stack_top + IST_DF  * PAGE_SIZE,
>> -                    [IST_NMI - 1] = stack_top + IST_NMI * PAGE_SIZE,
>> -                    [IST_DB  - 1] = stack_top + IST_DB  * PAGE_SIZE,
>> -
>> -                    [IST_MAX ... ARRAY_SIZE(tss->ist) - 1] =
>> -                            0x8600111111111111ul,
>> -            },
>> -
>> -            .bitmap = IOBMP_INVALID_OFFSET,
>> -    };
>> +    /*
>> +     * Set up the TSS.  Warning - may be live, and the NMI/#MC must remain
>> +     * valid on every instruction boundary.  (Note: these are all
>> +     * semantically ACCESS_ONCE() due to tss's volatile qualifier.)
>> +     *
>> +     * rsp0 refers to the primary stack.  #MC, #DF, NMI and #DB handlers
>> +     * each get their own stacks.  No IO Bitmap.
>> +     */
>> +    tss->rsp0 = stack_bottom;
>> +    tss->ist[IST_MCE - 1] = stack_top + IST_MCE * PAGE_SIZE;
>> +    tss->ist[IST_DF  - 1] = stack_top + IST_DF  * PAGE_SIZE;
>> +    tss->ist[IST_NMI - 1] = stack_top + IST_NMI * PAGE_SIZE;
>> +    tss->ist[IST_DB  - 1] = stack_top + IST_DB  * PAGE_SIZE;
>> +    tss->bitmap = IOBMP_INVALID_OFFSET;
>> +
>> +    /* All other stack pointers poisioned. */
>> +    for ( i = IST_MAX; i < ARRAY_SIZE(tss->ist); ++i )
>> +            tss->ist[i] = 0x8600111111111111ul;
>> +    tss->rsp1 = 0x8600111111111111ul;
>> +    tss->rsp2 = 0x8600111111111111ul;
> ACCESS_ONCE() unfortunately only has one of the two needed effects:
> It guarantees that each memory location gets accessed exactly once
> (which I assume can also be had with just the volatile addition,
> but without the moving away from using an initializer), but it does
> not guarantee single-insn accesses.

Linux's memory-barriers.txt disagrees, and specifically gives an example
with a misaligned int (vs two shorts) and the use volatile cast (by way
of {READ,WRITE}_ONCE()) to prevent load/store tearing, as the memory
location is of a size which can be fit in a single access.

I'm fairly sure we're safe here.

>  I consider this in particular
> relevant here because all of the 64-bit fields are misaligned. By
> doing it like you do, we're setting us up to have to re-do this yet
> again in a couple of years time (presumably using write_atomic()
> instead then).
>
> Nevertheless it is a clear improvement, so if you want to leave it
> like this
> Reviewed-by: Jan Beulich <jbeulich@xxxxxxxx>

Thanks,

~Andrew



 


Rackspace

Lists.xenproject.org is hosted with RackSpace, monitoring our
servers 24x7x365 and backed by RackSpace's Fanatical Support®.