[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index] [Xen-devel] [PATCH v2 08/11] xen: credit2: only marshall trace point arguments if tracing enabled
Signed-off-by: Dario Faggioli <dario.faggioli@xxxxxxxxxx> --- Cc: George Dunlap <george.dunlap@xxxxxxxxxx> Cc: Anshul Makkar <anshul.makkar@xxxxxxxxxx> Cc: David Vrabel <david.vrabel@xxxxxxxxxx> --- Changes from v1: * avoid stray code removal in balance_load(), as pointed out by George during review. --- xen/common/sched_credit2.c | 112 +++++++++++++++++++++++--------------------- 1 file changed, 59 insertions(+), 53 deletions(-) diff --git a/xen/common/sched_credit2.c b/xen/common/sched_credit2.c index a4aec73..be27ba3 100644 --- a/xen/common/sched_credit2.c +++ b/xen/common/sched_credit2.c @@ -638,6 +638,7 @@ __update_runq_load(const struct scheduler *ops, ASSERT(rqd->avgload <= STIME_MAX && rqd->b_avgload <= STIME_MAX); + if ( unlikely(tb_init_done) ) { struct { uint64_t rq_avgload, b_avgload; @@ -648,9 +649,9 @@ __update_runq_load(const struct scheduler *ops, d.rq_avgload = rqd->avgload; d.b_avgload = rqd->b_avgload; d.shift = P; - trace_var(TRC_CSCHED2_UPDATE_RUNQ_LOAD, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_UPDATE_RUNQ_LOAD, 1, + sizeof(d), + (unsigned char *)&d); } } @@ -693,6 +694,7 @@ __update_svc_load(const struct scheduler *ops, } svc->load_last_update = now; + if ( unlikely(tb_init_done) ) { struct { uint64_t v_avgload; @@ -703,9 +705,9 @@ __update_svc_load(const struct scheduler *ops, d.vcpu = svc->vcpu->vcpu_id; d.v_avgload = svc->avgload; d.shift = P; - trace_var(TRC_CSCHED2_UPDATE_VCPU_LOAD, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_UPDATE_VCPU_LOAD, 1, + sizeof(d), + (unsigned char *)&d); } } @@ -761,6 +763,7 @@ runq_insert(const struct scheduler *ops, struct csched2_vcpu *svc) pos = __runq_insert(runq, svc); + if ( unlikely(tb_init_done) ) { struct { unsigned vcpu:16, dom:16; @@ -769,9 +772,9 @@ runq_insert(const struct scheduler *ops, struct csched2_vcpu *svc) d.dom = svc->vcpu->domain->domain_id; d.vcpu = svc->vcpu->vcpu_id; d.pos = pos; - trace_var(TRC_CSCHED2_RUNQ_POS, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_RUNQ_POS, 1, + sizeof(d), + (unsigned char *)&d); } return; @@ -814,7 +817,7 @@ runq_tickle(const struct scheduler *ops, struct csched2_vcpu *new, s_time_t now) ASSERT(new->rqd == rqd); - /* TRACE */ + if ( unlikely(tb_init_done) ) { struct { unsigned vcpu:16, dom:16; @@ -824,9 +827,9 @@ runq_tickle(const struct scheduler *ops, struct csched2_vcpu *new, s_time_t now) d.vcpu = new->vcpu->vcpu_id; d.processor = new->vcpu->processor; d.credit = new->credit; - trace_var(TRC_CSCHED2_TICKLE_NEW, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_TICKLE_NEW, 1, + sizeof(d), + (unsigned char *)&d); } /* @@ -884,7 +887,8 @@ runq_tickle(const struct scheduler *ops, struct csched2_vcpu *new, s_time_t now) lowest = cur->credit; } - /* TRACE */ { + if ( unlikely(tb_init_done) ) + { struct { unsigned vcpu:16, dom:16; unsigned credit; @@ -892,9 +896,9 @@ runq_tickle(const struct scheduler *ops, struct csched2_vcpu *new, s_time_t now) d.dom = cur->vcpu->domain->domain_id; d.vcpu = cur->vcpu->vcpu_id; d.credit = cur->credit; - trace_var(TRC_CSCHED2_TICKLE_CHECK, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_TICKLE_CHECK, 1, + sizeof(d), + (unsigned char *)&d); } } @@ -912,14 +916,15 @@ runq_tickle(const struct scheduler *ops, struct csched2_vcpu *new, s_time_t now) tickle: BUG_ON(ipid == -1); - /* TRACE */ { + if ( unlikely(tb_init_done) ) + { struct { unsigned cpu:16, pad:16; } d; d.cpu = ipid; d.pad = 0; - trace_var(TRC_CSCHED2_TICKLE, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_TICKLE, 1, + sizeof(d), + (unsigned char *)&d); } __cpumask_set_cpu(ipid, &rqd->tickled); cpu_raise_softirq(ipid, SCHEDULE_SOFTIRQ); @@ -981,7 +986,8 @@ static void reset_credit(const struct scheduler *ops, int cpu, s_time_t now, svc->start_time = now; - /* TRACE */ { + if ( unlikely(tb_init_done) ) + { struct { unsigned vcpu:16, dom:16; unsigned credit_start, credit_end; @@ -992,9 +998,9 @@ static void reset_credit(const struct scheduler *ops, int cpu, s_time_t now, d.credit_start = start_credit; d.credit_end = svc->credit; d.multiplier = m; - trace_var(TRC_CSCHED2_CREDIT_RESET, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_CREDIT_RESET, 1, + sizeof(d), + (unsigned char *)&d); } } @@ -1030,7 +1036,7 @@ void burn_credits(struct csched2_runqueue_data *rqd, __func__, now, svc->start_time); } - /* TRACE */ + if ( unlikely(tb_init_done) ) { struct { unsigned vcpu:16, dom:16; @@ -1041,9 +1047,9 @@ void burn_credits(struct csched2_runqueue_data *rqd, d.vcpu = svc->vcpu->vcpu_id; d.credit = svc->credit; d.delta = delta; - trace_var(TRC_CSCHED2_CREDIT_BURN, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_CREDIT_BURN, 1, + sizeof(d), + (unsigned char *)&d); } } @@ -1079,16 +1085,16 @@ static void update_max_weight(struct csched2_runqueue_data *rqd, int new_weight, SCHED_STAT_CRANK(upd_max_weight_full); } - /* TRACE */ + if ( unlikely(tb_init_done) ) { struct { unsigned rqi:16, max_weight:16; } d; d.rqi = rqd->id; d.max_weight = rqd->max_weight; - trace_var(TRC_CSCHED2_RUNQ_MAX_WEIGHT, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_RUNQ_MAX_WEIGHT, 1, + sizeof(d), + (unsigned char *)&d); } } @@ -1168,7 +1174,7 @@ __runq_assign(struct csched2_vcpu *svc, struct csched2_runqueue_data *rqd) /* Expected new load based on adding this vcpu */ rqd->b_avgload += svc->avgload; - /* TRACE */ + if ( unlikely(tb_init_done) ) { struct { unsigned vcpu:16, dom:16; @@ -1177,9 +1183,9 @@ __runq_assign(struct csched2_vcpu *svc, struct csched2_runqueue_data *rqd) d.dom = svc->vcpu->domain->domain_id; d.vcpu = svc->vcpu->vcpu_id; d.rqi=rqd->id; - trace_var(TRC_CSCHED2_RUNQ_ASSIGN, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_RUNQ_ASSIGN, 1, + sizeof(d), + (unsigned char *)&d); } } @@ -1491,7 +1497,7 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) out_up: spin_unlock(&prv->lock); - /* TRACE */ + if ( unlikely(tb_init_done) ) { struct { uint64_t b_avgload; @@ -1503,9 +1509,9 @@ out_up: d.vcpu = vc->vcpu_id; d.rq_id = c2r(ops, new_cpu); d.new_cpu = new_cpu; - trace_var(TRC_CSCHED2_PICKED_CPU, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_PICKED_CPU, 1, + sizeof(d), + (unsigned char *)&d); } return new_cpu; @@ -1560,7 +1566,7 @@ static void migrate(const struct scheduler *ops, struct csched2_runqueue_data *trqd, s_time_t now) { - /* TRACE */ + if ( unlikely(tb_init_done) ) { struct { unsigned vcpu:16, dom:16; @@ -1570,9 +1576,9 @@ static void migrate(const struct scheduler *ops, d.vcpu = svc->vcpu->vcpu_id; d.rqi = svc->rqd->id; d.trqi = trqd->id; - trace_var(TRC_CSCHED2_MIGRATE, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_MIGRATE, 1, + sizeof(d), + (unsigned char *)&d); } if ( svc->flags & CSFLAG_scheduled ) @@ -1698,7 +1704,7 @@ retry: if ( i > cpus_max ) cpus_max = i; - /* TRACE */ + if ( unlikely(tb_init_done) ) { struct { unsigned lrq_id:16, orq_id:16; @@ -1707,9 +1713,9 @@ retry: d.lrq_id = st.lrqd->id; d.orq_id = st.orqd->id; d.load_delta = st.load_delta; - trace_var(TRC_CSCHED2_LOAD_CHECK, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_LOAD_CHECK, 1, + sizeof(d), + (unsigned char *)&d); } /* @@ -1740,7 +1746,7 @@ retry: if ( unlikely(st.orqd->id < 0) ) goto out_up; - /* TRACE */ + if ( unlikely(tb_init_done) ) { struct { uint64_t lb_avgload, ob_avgload; @@ -1750,9 +1756,9 @@ retry: d.lb_avgload = st.lrqd->b_avgload; d.orq_id = st.orqd->id; d.ob_avgload = st.orqd->b_avgload; - trace_var(TRC_CSCHED2_LOAD_BALANCE, 1, - sizeof(d), - (unsigned char *)&d); + __trace_var(TRC_CSCHED2_LOAD_BALANCE, 1, + sizeof(d), + (unsigned char *)&d); } /* Look for "swap" which gives the best load average _______________________________________________ Xen-devel mailing list Xen-devel@xxxxxxxxxxxxx https://lists.xen.org/xen-devel
|
Lists.xenproject.org is hosted with RackSpace, monitoring our |