Message ID | 146958898908.1696.15298736648669297817.stgit@Solace.fritz.box (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On 27/07/16 04:09, Dario Faggioli wrote: > In fact, when not finding a suitable runqueue where to > place a vCPU, and hence using a fallback, we either: > - don't issue any trace record (while we should, at > least, output the chosen pcpu), > - risk underruning when accessing the runqueues > array, while preparing the trace record. > > Fix both issues and, while there, also a couple of style > problems found nearby. > > Spotted by Coverity. > > Signed-off-by: Dario Faggioli <dario.faggioli@citrix.com> > Reported-by: Andrew Cooper <andrew.cooper3@citrix.com> Thanks, Reviewed-by: George Dunlap <george.dunlap@citrix.com> > --- > Cc: George Dunlap <george.dunlap@citrix.com> > Cc: Anshul Makkar <anshul.makkar@citrix.com> > --- > Changes from v2: > * use local variables to prevent unlocked access to prv, as > suggested during review. > Changes from v1: > * cite Coverity in the changelog. > --- > xen/common/sched_credit2.c | 23 ++++++++++++++--------- > 1 file changed, 14 insertions(+), 9 deletions(-) > > diff --git a/xen/common/sched_credit2.c b/xen/common/sched_credit2.c > index 7ec7f62..00d3300 100644 > --- a/xen/common/sched_credit2.c > +++ b/xen/common/sched_credit2.c > @@ -1464,7 +1464,7 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) > struct csched2_private *prv = CSCHED2_PRIV(ops); > int i, min_rqi = -1, new_cpu; > struct csched2_vcpu *svc = CSCHED2_VCPU(vc); > - s_time_t min_avgload; > + s_time_t min_avgload = MAX_LOAD; > > ASSERT(!cpumask_empty(&prv->active_queues)); > > @@ -1488,7 +1488,12 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) > { > /* We may be here because someone requested us to migrate. */ > __clear_bit(__CSFLAG_runq_migrate_request, &svc->flags); > - return get_fallback_cpu(svc); > + new_cpu = get_fallback_cpu(svc); > + /* > + * Tracing of runq and its load won't be accurate, since we could > + * not get the lock, but at least we will output the chosen pcpu. > + */ > + goto out; > } > > /* > @@ -1513,8 +1518,6 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) > /* Fall-through to normal cpu pick */ > } > > - min_avgload = MAX_LOAD; > - > /* Find the runqueue with the lowest average load. */ > for_each_cpu(i, &prv->active_queues) > { > @@ -1552,7 +1555,7 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) > if ( rqd_avgload < min_avgload ) > { > min_avgload = rqd_avgload; > - min_rqi=i; > + min_rqi = i; > } > } > > @@ -1560,6 +1563,8 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) > if ( min_rqi == -1 ) > { > new_cpu = get_fallback_cpu(svc); > + min_rqi = c2r(ops, new_cpu); > + min_avgload = prv->rqd[min_rqi].b_avgload; > goto out_up; > } > > @@ -1570,18 +1575,18 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) > > out_up: > read_unlock(&prv->lock); > - > + out: > if ( unlikely(tb_init_done) ) > { > struct { > uint64_t b_avgload; > unsigned vcpu:16, dom:16; > unsigned rq_id:16, new_cpu:16; > - } d; > - d.b_avgload = prv->rqd[min_rqi].b_avgload; > + } d; > d.dom = vc->domain->domain_id; > d.vcpu = vc->vcpu_id; > - d.rq_id = c2r(ops, new_cpu); > + d.rq_id = min_rqi; > + d.b_avgload = min_avgload; > d.new_cpu = new_cpu; > __trace_var(TRC_CSCHED2_PICKED_CPU, 1, > sizeof(d), >
diff --git a/xen/common/sched_credit2.c b/xen/common/sched_credit2.c index 7ec7f62..00d3300 100644 --- a/xen/common/sched_credit2.c +++ b/xen/common/sched_credit2.c @@ -1464,7 +1464,7 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) struct csched2_private *prv = CSCHED2_PRIV(ops); int i, min_rqi = -1, new_cpu; struct csched2_vcpu *svc = CSCHED2_VCPU(vc); - s_time_t min_avgload; + s_time_t min_avgload = MAX_LOAD; ASSERT(!cpumask_empty(&prv->active_queues)); @@ -1488,7 +1488,12 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) { /* We may be here because someone requested us to migrate. */ __clear_bit(__CSFLAG_runq_migrate_request, &svc->flags); - return get_fallback_cpu(svc); + new_cpu = get_fallback_cpu(svc); + /* + * Tracing of runq and its load won't be accurate, since we could + * not get the lock, but at least we will output the chosen pcpu. + */ + goto out; } /* @@ -1513,8 +1518,6 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) /* Fall-through to normal cpu pick */ } - min_avgload = MAX_LOAD; - /* Find the runqueue with the lowest average load. */ for_each_cpu(i, &prv->active_queues) { @@ -1552,7 +1555,7 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) if ( rqd_avgload < min_avgload ) { min_avgload = rqd_avgload; - min_rqi=i; + min_rqi = i; } } @@ -1560,6 +1563,8 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) if ( min_rqi == -1 ) { new_cpu = get_fallback_cpu(svc); + min_rqi = c2r(ops, new_cpu); + min_avgload = prv->rqd[min_rqi].b_avgload; goto out_up; } @@ -1570,18 +1575,18 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc) out_up: read_unlock(&prv->lock); - + out: if ( unlikely(tb_init_done) ) { struct { uint64_t b_avgload; unsigned vcpu:16, dom:16; unsigned rq_id:16, new_cpu:16; - } d; - d.b_avgload = prv->rqd[min_rqi].b_avgload; + } d; d.dom = vc->domain->domain_id; d.vcpu = vc->vcpu_id; - d.rq_id = c2r(ops, new_cpu); + d.rq_id = min_rqi; + d.b_avgload = min_avgload; d.new_cpu = new_cpu; __trace_var(TRC_CSCHED2_PICKED_CPU, 1, sizeof(d),
In fact, when not finding a suitable runqueue where to place a vCPU, and hence using a fallback, we either: - don't issue any trace record (while we should, at least, output the chosen pcpu), - risk underruning when accessing the runqueues array, while preparing the trace record. Fix both issues and, while there, also a couple of style problems found nearby. Spotted by Coverity. Signed-off-by: Dario Faggioli <dario.faggioli@citrix.com> Reported-by: Andrew Cooper <andrew.cooper3@citrix.com> --- Cc: George Dunlap <george.dunlap@citrix.com> Cc: Anshul Makkar <anshul.makkar@citrix.com> --- Changes from v2: * use local variables to prevent unlocked access to prv, as suggested during review. Changes from v1: * cite Coverity in the changelog. --- xen/common/sched_credit2.c | 23 ++++++++++++++--------- 1 file changed, 14 insertions(+), 9 deletions(-)