|
[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] Re: [PATCH] x86/pv: Sanity check bytes_per_rep in rep_outs()
On Sat, May 11, 2024 at 04:16:42PM +0100, Andrew Cooper wrote:
> I was playing with GCC-14, and gave the new static analyser (-fanalyze) a
> try. One of the more serious things it came back with was this:
>
> In file included from ./arch/x86/include/asm/mm.h:10,
> from ./include/xen/mm.h:233,
> from ./include/xen/domain_page.h:12,
> from arch/x86/pv/emul-priv-op.c:10:
> In function '__copy_from_guest_pv',
> inlined from 'rep_outs' at arch/x86/pv/emul-priv-op.c:718:20:
> ./arch/x86/include/asm/uaccess.h:174:9: warning: stack-based buffer
> overflow [CWE-121] [-Wanalyzer-out-of-bounds]
> 174 | __asm__ __volatile__(
> \
> | ^~~~~~~
> ./arch/x86/include/asm/uaccess.h:229:13: note: in expansion of macro
> 'get_unsafe_asm'
> 229 | case 8: get_unsafe_asm(x, ptr, grd, retval, "", "=r", errret);
> break; \
> | ^~~~~~~~~~~~~~
> ./arch/x86/include/asm/uaccess.h:236:5: note: in expansion of macro
> 'get_unsafe_size'
> 236 | get_unsafe_size(x, ptr, size, UA_KEEP, retval, errret)
> | ^~~~~~~~~~~~~~~
> ./arch/x86/include/asm/uaccess.h:308:13: note: in expansion of macro
> 'get_guest_size'
> 308 | get_guest_size(*(uint64_t *)to, from, 8, ret, 8);
> | ^~~~~~~~~~~~~~
> 'rep_outs': events 1-3
> |
> |arch/x86/pv/emul-priv-op.c:674:21:
> | 674 | static int cf_check rep_outs(
> | | ^~~~~~~~
> | | |
> | | (1) entry to 'rep_outs'
> |......
> | 688 | if ( !guest_io_okay(port, bytes_per_rep, curr, ctxt->regs)
> )
> | | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> | | |
> | | (3) calling 'guest_io_okay' from 'rep_outs'
> |......
> | 710 | unsigned int data = 0;
> | | ~~~~
> | | |
> | | (2) capacity: 4 bytes
> |
> +--> 'guest_io_okay': events 4-5
> |
> | 159 | static bool guest_io_okay(unsigned int port, unsigned
> int bytes,
> | | ^~~~~~~~~~~~~
> | | |
> | | (4) entry to 'guest_io_okay'
> |......
> | 165 | if ( iopl_ok(v, regs) )
> | | ~~~~~~~~~~~~~~~~
> | | |
> | | (5) calling 'iopl_ok' from 'guest_io_okay'
> |
> +--> 'iopl_ok': event 6
> |
> | 148 | static bool iopl_ok(const struct vcpu *v, const
> struct cpu_user_regs *regs)
> | | ^~~~~~~
> | | |
> | | (6) entry to 'iopl_ok'
> |
> 'iopl_ok': event 7
> |
> |./include/xen/bug.h:141:13:
> | 141 | do { if ( unlikely(!(p)) )
> assert_failed(#p); } while (0)
> | | ^
> | | |
> | | (7) following 'false' branch...
> arch/x86/pv/emul-priv-op.c:153:5: note: in expansion of macro 'ASSERT'
> | 153 | ASSERT((v->arch.pv.iopl & ~X86_EFLAGS_IOPL)
> == 0);
> | | ^~~~~~
> |
> 'iopl_ok': event 8
> |
> |./include/xen/bug.h:124:31:
> | 124 | #define assert_failed(msg) do {
> \
> | | ^
> | | |
> | | (8) ...to here
> ./include/xen/bug.h:141:32: note: in expansion of macro 'assert_failed'
> | 141 | do { if ( unlikely(!(p)) )
> assert_failed(#p); } while (0)
> | | ^~~~~~~~~~~~~
> arch/x86/pv/emul-priv-op.c:153:5: note: in expansion of macro 'ASSERT'
> | 153 | ASSERT((v->arch.pv.iopl & ~X86_EFLAGS_IOPL)
> == 0);
> | | ^~~~~~
> |
> <------+
> |
> 'guest_io_okay': event 9
> |
> | 165 | if ( iopl_ok(v, regs) )
> | | ^~~~~~~~~~~~~~~~
> | | |
> | | (9) returning to 'guest_io_okay' from 'iopl_ok'
> |
> <------+
> |
> 'rep_outs': events 10-13
> |
> | 688 | if ( !guest_io_okay(port, bytes_per_rep, curr, ctxt->regs)
> )
> | | ~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> | | | |
> | | | (10) returning to 'rep_outs' from 'guest_io_okay'
> | | (11) following 'true' branch...
> |......
> | 691 | rc = read_segment(seg, &sreg, ctxt);
> | | ~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> | | | |
> | | | (13) calling 'read_segment' from 'rep_outs'
> | | (12) ...to here
> |
> +--> 'read_segment': events 14-18
> |
> | 493 | static int cf_check read_segment(
> | | ^~~~~~~~~~~~
> | | |
> | | (14) entry to 'read_segment'
> |......
> | 510 | if ( ctxt->addr_size < 64 )
> | | ~
> | | |
> | | (15) following 'false' branch...
> |......
> | 535 | switch ( seg )
> | | ~~~~~~
> | | |
> | | (16) ...to here
> |......
> | 538 | if ( !is_x86_user_segment(seg) )
> | | ~
> | | |
> | | (17) following 'false' branch (when 'seg
> <= 5')...
> | 539 | return X86EMUL_UNHANDLEABLE;
> | 540 | reg->base = 0;
> | | ~~~
> | | |
> | | (18) ...to here
> |
> <------+
> |
> 'rep_outs': events 19-30
> |
> | 691 | rc = read_segment(seg, &sreg, ctxt);
> | | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> | | |
> | | (19) returning to 'rep_outs' from 'read_segment'
> | 692 | if ( rc != X86EMUL_OKAY )
> | | ~
> | | |
> | | (20) following 'false' branch (when 'rc == 0')...
> |......
> | 695 | if ( !sreg.p )
> | | ~~ ~
> | | | |
> | | | (22) following 'false' branch...
> | | (21) ...to here
> | 696 | return X86EMUL_UNHANDLEABLE;
> | 697 | if ( !sreg.s ||
> | | ~~ ~ ~~~~~~~~~~
> | | | | |
> | | | | (26) following 'false' branch...
> | | | (24) following 'false' branch...
> | | (23) ...to here
> | 698 | ((sreg.type & (_SEGMENT_CODE >> 8)) &&
> | | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> | | |
> | | (25) ...to here
> | 699 | !(sreg.type & (_SEGMENT_WR >> 8))) )
> | | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> |......
> | 706 | poc->bpmatch = check_guest_io_breakpoint(curr, port,
> bytes_per_rep);
> | | ~~~
> | | |
> | | (27) ...to here
> | 707 |
> | 708 | while ( *reps < goal )
> | | ~~~~~~~~~~~~
> | | |
> | | (28) following 'true' branch...
> | 709 | {
> | 710 | unsigned int data = 0;
> | | ~~~~~~~~
> | | |
> | | (29) ...to here
> |......
> | 713 | rc = pv_emul_virt_to_linear(sreg.base, offset,
> bytes_per_rep,
> | |
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> | | |
> | | (30) calling 'pv_emul_virt_to_linear' from
> 'rep_outs'
> | 714 | sreg.limit, seg, ctxt,
> &addr);
> | |
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> |
> +--> 'pv_emul_virt_to_linear': events 31-33
> |
> | 580 | static int pv_emul_virt_to_linear(unsigned long base,
> unsigned long offset,
> | | ^~~~~~~~~~~~~~~~~~~~~~
> | | |
> | | (31) entry to 'pv_emul_virt_to_linear'
> |......
> | 596 | else if ( !__addr_ok(*addr) )
> | | ~
> | | |
> | | (32) following 'false' branch...
> |......
> | 603 | return rc;
> | | ~~~~~~
> | | |
> | | (33) ...to here
> |
> <------+
> |
> 'rep_outs': events 34-37
> |
> | 713 | rc = pv_emul_virt_to_linear(sreg.base, offset,
> bytes_per_rep,
> | |
> ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> | | |
> | | (34) returning to 'rep_outs' from
> 'pv_emul_virt_to_linear'
> | 714 | sreg.limit, seg, ctxt,
> &addr);
> | |
> ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
> | 715 | if ( rc != X86EMUL_OKAY )
> | | ~
> | | |
> | | (35) following 'false' branch (when 'rc == 0')...
> |......
> | 718 | if ( (rc = __copy_from_guest_pv(&data, (void __user
> *)addr,
> | | ~~ ~
> | | | |
> | | | (37) inlined call to '__copy_from_guest_pv'
> from 'rep_outs'
> | | (36) ...to here
> |
> +--> '__copy_from_guest_pv': events 38-41
> |
> |./arch/x86/include/asm/uaccess.h:294:8:
> | 294 | if (__builtin_constant_p(n)) {
> | | ^
> | | |
> | | (38) following 'true' branch...
> | 295 | unsigned long ret;
> | | ~~~~~~~~
> | | |
> | | (39) ...to here
> | 296 |
> | 297 | switch (n) {
> | | ~~~~~~
> | | |
> | | (40) following 'case 8:' branch...
> |......
> | 307 | case 8:
> | | ~~~~
> | | |
> | | (41) ...to here
> |
> '__copy_from_guest_pv': event 42
> |
> | 174 | __asm__ __volatile__(
> \
> | | ^~~~~~~
> | | |
> | | (42) out-of-bounds write from byte 4 till byte
> 7 but 'data' ends at byte 4
> ./arch/x86/include/asm/uaccess.h:229:13: note: in expansion of macro
> 'get_unsafe_asm'
> | 229 | case 8: get_unsafe_asm(x, ptr, grd, retval, "",
> "=r", errret); break; \
> | | ^~~~~~~~~~~~~~
> ./arch/x86/include/asm/uaccess.h:236:5: note: in expansion of macro
> 'get_unsafe_size'
> | 236 | get_unsafe_size(x, ptr, size, UA_KEEP, retval,
> errret)
> | | ^~~~~~~~~~~~~~~
> ./arch/x86/include/asm/uaccess.h:308:13: note: in expansion of macro
> 'get_guest_size'
> | 308 | get_guest_size(*(uint64_t *)to, from, 8,
> ret, 8);
> | | ^~~~~~~~~~~~~~
> |
> ./arch/x86/include/asm/uaccess.h:174:9: note: write of 4 bytes to beyond
> the end of 'data'
> 174 | __asm__ __volatile__(
> \
> | ^~~~~~~
> ./arch/x86/include/asm/uaccess.h:229:13: note: in expansion of macro
> 'get_unsafe_asm'
> 229 | case 8: get_unsafe_asm(x, ptr, grd, retval, "", "=r", errret);
> break; \
> | ^~~~~~~~~~~~~~
> ./arch/x86/include/asm/uaccess.h:236:5: note: in expansion of macro
> 'get_unsafe_size'
> 236 | get_unsafe_size(x, ptr, size, UA_KEEP, retval, errret)
> | ^~~~~~~~~~~~~~~
> ./arch/x86/include/asm/uaccess.h:308:13: note: in expansion of macro
> 'get_guest_size'
> 308 | get_guest_size(*(uint64_t *)to, from, 8, ret, 8);
> | ^~~~~~~~~~~~~~
>
> ┌──────────────────────────────────────────────────────────────────────┐
> │ write of 'uint64_t' (8 bytes) │
> └──────────────────────────────────────────────────────────────────────┘
> │ │
> │ │
> v v
> ┌──────────────────────────────────┐┌──────────────────────────────────┐
> │ 'data' (type: 'unsigned int') ││ after valid range │
> └──────────────────────────────────┘└──────────────────────────────────┘
> ├────────────────┬─────────────────┤├────────────────┬─────────────────┤
> │ │
> ╭────────┴────────╮ ╭───────────┴──────────╮
> │capacity: 4 bytes│ │⚠️ overflow of 4 bytes│
> ╰─────────────────╯ ╰──────────────────────╯
>
> What is happening is that it's seen that __copy_from_guest_pv() has a case for
> 8, and it doesn't have any information about bytes_per_rep which is an input
> to the function.
>
> Architecturally, there's no 64-bit variant of OUTS in x86. Leaving an
> assertion to this effect is enough to satisfy the analyser.
>
> Signed-off-by: Andrew Cooper <andrew.cooper3@xxxxxxxxxx>
> ---
> CC: Jan Beulich <JBeulich@xxxxxxxx>
> CC: Roger Pau Monné <roger.pau@xxxxxxxxxx>
>
> A little RFC, but the analyser is right - if rep_outs() were to be called with
> anything more than 4, the stack would be clobbered, and no developer is
> finding that bug the easy way...
> ---
> xen/arch/x86/pv/emul-priv-op.c | 2 ++
> 1 file changed, 2 insertions(+)
>
> diff --git a/xen/arch/x86/pv/emul-priv-op.c b/xen/arch/x86/pv/emul-priv-op.c
> index f101510a1bab..1460ecc187be 100644
> --- a/xen/arch/x86/pv/emul-priv-op.c
> +++ b/xen/arch/x86/pv/emul-priv-op.c
> @@ -685,6 +685,8 @@ static int cf_check rep_outs(
>
> *reps = 0;
>
> + ASSERT(bytes_per_rep <= 4); /* i.e. 'data' being 4 bytes is fine. */
Don't we need this to be a BUG_ON() to satisfy the compiler also on
non-debug builds?
Or maybe:
if ( bytes_per_rep > 4 )
{
ASSERT_UNREACHABLE();
return X86EMUL_UNHANDLEABLE;
}
Would it be possible to add the check to guest_io_okay() itself?
Thanks, Roger.
|
![]() |
Lists.xenproject.org is hosted with RackSpace, monitoring our |