[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [PATCH] x86/boot: Fix load_system_tables() to be NMI/#MC-safe
On 27/05/2020 14:19, Jan Beulich wrote: > On 27.05.2020 15:06, Andrew Cooper wrote: >> @@ -720,30 +721,26 @@ void load_system_tables(void) >> .limit = (IDT_ENTRIES * sizeof(idt_entry_t)) - 1, >> }; >> >> - *tss = (struct tss64){ >> - /* Main stack for interrupts/exceptions. */ >> - .rsp0 = stack_bottom, >> - >> - /* Ring 1 and 2 stacks poisoned. */ >> - .rsp1 = 0x8600111111111111ul, >> - .rsp2 = 0x8600111111111111ul, >> - >> - /* >> - * MCE, NMI and Double Fault handlers get their own stacks. >> - * All others poisoned. >> - */ >> - .ist = { >> - [IST_MCE - 1] = stack_top + IST_MCE * PAGE_SIZE, >> - [IST_DF - 1] = stack_top + IST_DF * PAGE_SIZE, >> - [IST_NMI - 1] = stack_top + IST_NMI * PAGE_SIZE, >> - [IST_DB - 1] = stack_top + IST_DB * PAGE_SIZE, >> - >> - [IST_MAX ... ARRAY_SIZE(tss->ist) - 1] = >> - 0x8600111111111111ul, >> - }, >> - >> - .bitmap = IOBMP_INVALID_OFFSET, >> - }; >> + /* >> + * Set up the TSS. Warning - may be live, and the NMI/#MC must remain >> + * valid on every instruction boundary. (Note: these are all >> + * semantically ACCESS_ONCE() due to tss's volatile qualifier.) >> + * >> + * rsp0 refers to the primary stack. #MC, #DF, NMI and #DB handlers >> + * each get their own stacks. No IO Bitmap. >> + */ >> + tss->rsp0 = stack_bottom; >> + tss->ist[IST_MCE - 1] = stack_top + IST_MCE * PAGE_SIZE; >> + tss->ist[IST_DF - 1] = stack_top + IST_DF * PAGE_SIZE; >> + tss->ist[IST_NMI - 1] = stack_top + IST_NMI * PAGE_SIZE; >> + tss->ist[IST_DB - 1] = stack_top + IST_DB * PAGE_SIZE; >> + tss->bitmap = IOBMP_INVALID_OFFSET; >> + >> + /* All other stack pointers poisioned. */ >> + for ( i = IST_MAX; i < ARRAY_SIZE(tss->ist); ++i ) >> + tss->ist[i] = 0x8600111111111111ul; >> + tss->rsp1 = 0x8600111111111111ul; >> + tss->rsp2 = 0x8600111111111111ul; > ACCESS_ONCE() unfortunately only has one of the two needed effects: > It guarantees that each memory location gets accessed exactly once > (which I assume can also be had with just the volatile addition, > but without the moving away from using an initializer), but it does > not guarantee single-insn accesses. Linux's memory-barriers.txt disagrees, and specifically gives an example with a misaligned int (vs two shorts) and the use volatile cast (by way of {READ,WRITE}_ONCE()) to prevent load/store tearing, as the memory location is of a size which can be fit in a single access. I'm fairly sure we're safe here. > I consider this in particular > relevant here because all of the 64-bit fields are misaligned. By > doing it like you do, we're setting us up to have to re-do this yet > again in a couple of years time (presumably using write_atomic() > instead then). > > Nevertheless it is a clear improvement, so if you want to leave it > like this > Reviewed-by: Jan Beulich <jbeulich@xxxxxxxx> Thanks, ~Andrew
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |