# HG changeset patch
# User kfraser@xxxxxxxxxxxxxxxxxxxxx
# Node ID cf98903ebb22f5a5aae73c56912bf936ab5787da
# Parent 5f7b5e5ca14b6c00d8ac23d1a2ece62fcbaebc03
[XEN] Scheduling hypercalls defer entry to the scheduler to softirq
context.
This avoids entering the scheduler with outstanding multicall state
and also happens to simplify the x86 entry protocol to the scheduler
hypercall (since we do not need to preset the return code).
Signed-off-by: Keir Fraser <keir@xxxxxxxxxxxxx>
---
xen/arch/ia64/vmx/vmx_support.c | 3 +--
xen/arch/x86/x86_32/entry.S | 22 ++++------------------
xen/arch/x86/x86_64/entry.S | 24 +++++-------------------
xen/common/domain.c | 14 ++++++++++++--
xen/common/schedule.c | 17 ++++++++---------
5 files changed, 30 insertions(+), 50 deletions(-)
diff -r 5f7b5e5ca14b -r cf98903ebb22 xen/arch/ia64/vmx/vmx_support.c
--- a/xen/arch/ia64/vmx/vmx_support.c Thu Nov 16 17:07:23 2006 +0000
+++ b/xen/arch/ia64/vmx/vmx_support.c Thu Nov 16 18:28:05 2006 +0000
@@ -95,8 +95,7 @@ void vmx_send_assist_req(struct vcpu *v)
break;
}
- /* I want to call __enter_scheduler() only */
- do_sched_op_compat(SCHEDOP_yield, 0);
+ raise_softirq(SCHEDULE_SOFTIRQ);
mb();
}
diff -r 5f7b5e5ca14b -r cf98903ebb22 xen/arch/x86/x86_32/entry.S
--- a/xen/arch/x86/x86_32/entry.S Thu Nov 16 17:07:23 2006 +0000
+++ b/xen/arch/x86/x86_32/entry.S Thu Nov 16 18:28:05 2006 +0000
@@ -597,20 +597,6 @@ ENTRY(setup_vm86_frame)
addl $16,%esp
ret
-do_arch_sched_op_compat:
- # Ensure we return success even if we return via schedule_tail()
- xorl %eax,%eax
- GET_GUEST_REGS(%ecx)
- movl %eax,UREGS_eax(%ecx)
- jmp do_sched_op_compat
-
-do_arch_sched_op:
- # Ensure we return success even if we return via schedule_tail()
- xorl %eax,%eax
- GET_GUEST_REGS(%ecx)
- movl %eax,UREGS_eax(%ecx)
- jmp do_sched_op
-
.data
ENTRY(exception_table)
@@ -642,7 +628,7 @@ ENTRY(hypercall_table)
.long do_stack_switch
.long do_set_callbacks
.long do_fpu_taskswitch /* 5 */
- .long do_arch_sched_op_compat
+ .long do_sched_op_compat
.long do_platform_op
.long do_set_debugreg
.long do_get_debugreg
@@ -665,7 +651,7 @@ ENTRY(hypercall_table)
.long do_mmuext_op
.long do_acm_op
.long do_nmi_op
- .long do_arch_sched_op
+ .long do_sched_op
.long do_callback_op /* 30 */
.long do_xenoprof_op
.long do_event_channel_op
@@ -684,7 +670,7 @@ ENTRY(hypercall_args_table)
.byte 2 /* do_stack_switch */
.byte 4 /* do_set_callbacks */
.byte 1 /* do_fpu_taskswitch */ /* 5 */
- .byte 2 /* do_arch_sched_op_compat */
+ .byte 2 /* do_sched_op_compat */
.byte 1 /* do_platform_op */
.byte 2 /* do_set_debugreg */
.byte 1 /* do_get_debugreg */
@@ -707,7 +693,7 @@ ENTRY(hypercall_args_table)
.byte 4 /* do_mmuext_op */
.byte 1 /* do_acm_op */
.byte 2 /* do_nmi_op */
- .byte 2 /* do_arch_sched_op */
+ .byte 2 /* do_sched_op */
.byte 2 /* do_callback_op */ /* 30 */
.byte 2 /* do_xenoprof_op */
.byte 2 /* do_event_channel_op */
diff -r 5f7b5e5ca14b -r cf98903ebb22 xen/arch/x86/x86_64/entry.S
--- a/xen/arch/x86/x86_64/entry.S Thu Nov 16 17:07:23 2006 +0000
+++ b/xen/arch/x86/x86_64/entry.S Thu Nov 16 18:28:05 2006 +0000
@@ -497,20 +497,6 @@ nmi_in_hypervisor_mode:
call do_nmi
jmp ret_from_intr
-do_arch_sched_op_compat:
- # Ensure we return success even if we return via schedule_tail()
- xorl %eax,%eax
- GET_GUEST_REGS(%r10)
- movq %rax,UREGS_rax(%r10)
- jmp do_sched_op_compat
-
-do_arch_sched_op:
- # Ensure we return success even if we return via schedule_tail()
- xorl %eax,%eax
- GET_GUEST_REGS(%r10)
- movq %rax,UREGS_rax(%r10)
- jmp do_sched_op
-
.data
ENTRY(exception_table)
@@ -542,7 +528,7 @@ ENTRY(hypercall_table)
.quad do_stack_switch
.quad do_set_callbacks
.quad do_fpu_taskswitch /* 5 */
- .quad do_arch_sched_op_compat
+ .quad do_sched_op_compat
.quad do_platform_op
.quad do_set_debugreg
.quad do_get_debugreg
@@ -565,7 +551,7 @@ ENTRY(hypercall_table)
.quad do_mmuext_op
.quad do_acm_op
.quad do_nmi_op
- .quad do_arch_sched_op
+ .quad do_sched_op
.quad do_callback_op /* 30 */
.quad do_xenoprof_op
.quad do_event_channel_op
@@ -584,8 +570,8 @@ ENTRY(hypercall_args_table)
.byte 2 /* do_stack_switch */
.byte 3 /* do_set_callbacks */
.byte 1 /* do_fpu_taskswitch */ /* 5 */
- .byte 2 /* do_arch_sched_op_compat */
- .byte 1 /* do_platform_op */
+ .byte 2 /* do_sched_op_compat */
+ .byte 1 /* do_platform_op */
.byte 2 /* do_set_debugreg */
.byte 1 /* do_get_debugreg */
.byte 2 /* do_update_descriptor */ /* 10 */
@@ -607,7 +593,7 @@ ENTRY(hypercall_args_table)
.byte 4 /* do_mmuext_op */
.byte 1 /* do_acm_op */
.byte 2 /* do_nmi_op */
- .byte 2 /* do_arch_sched_op */
+ .byte 2 /* do_sched_op */
.byte 2 /* do_callback_op */ /* 30 */
.byte 2 /* do_xenoprof_op */
.byte 2 /* do_event_channel_op */
diff -r 5f7b5e5ca14b -r cf98903ebb22 xen/common/domain.c
--- a/xen/common/domain.c Thu Nov 16 17:07:23 2006 +0000
+++ b/xen/common/domain.c Thu Nov 16 18:28:05 2006 +0000
@@ -258,8 +258,18 @@ void __domain_crash_synchronous(void)
{
__domain_crash(current->domain);
- /* Flush multicall state before dying. */
- this_cpu(mc_state).flags = 0;
+ /*
+ * Flush multicall state before dying if a multicall is in progress.
+ * This shouldn't be necessary, but some architectures are calling
+ * domain_crash_synchronous() when they really shouldn't (i.e., from
+ * within hypercall context).
+ */
+ if ( this_cpu(mc_state).flags != 0 )
+ {
+ dprintk(XENLOG_ERR,
+ "FIXME: synchronous domain crash during a multicall!\n");
+ this_cpu(mc_state).flags = 0;
+ }
for ( ; ; )
do_softirq();
diff -r 5f7b5e5ca14b -r cf98903ebb22 xen/common/schedule.c
--- a/xen/common/schedule.c Thu Nov 16 17:07:23 2006 +0000
+++ b/xen/common/schedule.c Thu Nov 16 18:28:05 2006 +0000
@@ -60,8 +60,6 @@ static struct scheduler *schedulers[] =
NULL
};
-static void __enter_scheduler(void);
-
static struct scheduler ops;
#define SCHED_OP(fn, ...) \
@@ -270,7 +268,7 @@ static long do_block(void)
else
{
TRACE_2D(TRC_SCHED_BLOCK, v->domain->domain_id, v->vcpu_id);
- __enter_scheduler();
+ raise_softirq(SCHEDULE_SOFTIRQ);
}
return 0;
@@ -315,9 +313,9 @@ static long do_poll(struct sched_poll *s
set_timer(&v->poll_timer, sched_poll->timeout);
TRACE_2D(TRC_SCHED_BLOCK, v->domain->domain_id, v->vcpu_id);
- __enter_scheduler();
-
- stop_timer(&v->poll_timer);
+ raise_softirq(SCHEDULE_SOFTIRQ);
+
+ return 0;
out:
clear_bit(_VCPUF_polling, &v->vcpu_flags);
@@ -329,7 +327,7 @@ static long do_yield(void)
static long do_yield(void)
{
TRACE_2D(TRC_SCHED_YIELD, current->domain->domain_id, current->vcpu_id);
- __enter_scheduler();
+ raise_softirq(SCHEDULE_SOFTIRQ);
return 0;
}
@@ -540,7 +538,7 @@ long sched_adjust(struct domain *d, stru
* - deschedule the current domain (scheduler independent).
* - pick a new domain (scheduler dependent).
*/
-static void __enter_scheduler(void)
+static void schedule(void)
{
struct vcpu *prev = current, *next = NULL;
s_time_t now = NOW();
@@ -549,6 +547,7 @@ static void __enter_scheduler(void)
s32 r_time; /* time for new dom to run */
ASSERT(!in_irq());
+ ASSERT(this_cpu(mc_state).flags == 0);
perfc_incrc(sched_run);
@@ -679,7 +678,7 @@ void __init scheduler_init(void)
{
int i;
- open_softirq(SCHEDULE_SOFTIRQ, __enter_scheduler);
+ open_softirq(SCHEDULE_SOFTIRQ, schedule);
for_each_cpu ( i )
{
_______________________________________________
Xen-changelog mailing list
Xen-changelog@xxxxxxxxxxxxxxxxxxx
http://lists.xensource.com/xen-changelog
|