diff mbox

[v3] xen: credit2: issues in csched2_cpu_pick(), when tracing is enabled.

Message ID 146958898908.1696.15298736648669297817.stgit@Solace.fritz.box (mailing list archive)
State New, archived
Headers show

Commit Message

Dario Faggioli July 27, 2016, 3:09 a.m. UTC
In fact, when not finding a suitable runqueue where to
place a vCPU, and hence using a fallback, we either:
 - don't issue any trace record (while we should, at
   least, output the chosen pcpu),
 - risk underruning when accessing the runqueues
   array, while preparing the trace record.

Fix both issues and, while there, also a couple of style
problems found nearby.

Spotted by Coverity.

Signed-off-by: Dario Faggioli <dario.faggioli@citrix.com>
Reported-by: Andrew Cooper <andrew.cooper3@citrix.com>
---
Cc: George Dunlap <george.dunlap@citrix.com>
Cc: Anshul Makkar <anshul.makkar@citrix.com>
---
Changes from v2:
 * use local variables to prevent unlocked access to prv, as
   suggested during review.
Changes from v1:
 * cite Coverity in the changelog.
---
 xen/common/sched_credit2.c |   23 ++++++++++++++---------
 1 file changed, 14 insertions(+), 9 deletions(-)

Comments

George Dunlap Aug. 3, 2016, 11:18 a.m. UTC | #1
On 27/07/16 04:09, Dario Faggioli wrote:
> In fact, when not finding a suitable runqueue where to
> place a vCPU, and hence using a fallback, we either:
>  - don't issue any trace record (while we should, at
>    least, output the chosen pcpu),
>  - risk underruning when accessing the runqueues
>    array, while preparing the trace record.
> 
> Fix both issues and, while there, also a couple of style
> problems found nearby.
> 
> Spotted by Coverity.
> 
> Signed-off-by: Dario Faggioli <dario.faggioli@citrix.com>
> Reported-by: Andrew Cooper <andrew.cooper3@citrix.com>

Thanks,

Reviewed-by: George Dunlap <george.dunlap@citrix.com>



> ---
> Cc: George Dunlap <george.dunlap@citrix.com>
> Cc: Anshul Makkar <anshul.makkar@citrix.com>
> ---
> Changes from v2:
>  * use local variables to prevent unlocked access to prv, as
>    suggested during review.
> Changes from v1:
>  * cite Coverity in the changelog.
> ---
>  xen/common/sched_credit2.c |   23 ++++++++++++++---------
>  1 file changed, 14 insertions(+), 9 deletions(-)
> 
> diff --git a/xen/common/sched_credit2.c b/xen/common/sched_credit2.c
> index 7ec7f62..00d3300 100644
> --- a/xen/common/sched_credit2.c
> +++ b/xen/common/sched_credit2.c
> @@ -1464,7 +1464,7 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
>      struct csched2_private *prv = CSCHED2_PRIV(ops);
>      int i, min_rqi = -1, new_cpu;
>      struct csched2_vcpu *svc = CSCHED2_VCPU(vc);
> -    s_time_t min_avgload;
> +    s_time_t min_avgload = MAX_LOAD;
>  
>      ASSERT(!cpumask_empty(&prv->active_queues));
>  
> @@ -1488,7 +1488,12 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
>      {
>          /* We may be here because someone requested us to migrate. */
>          __clear_bit(__CSFLAG_runq_migrate_request, &svc->flags);
> -        return get_fallback_cpu(svc);
> +        new_cpu = get_fallback_cpu(svc);
> +        /*
> +         * Tracing of runq and its load won't be accurate, since we could
> +         * not get the lock, but at least we will output the chosen pcpu.
> +         */
> +        goto out;
>      }
>  
>      /*
> @@ -1513,8 +1518,6 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
>          /* Fall-through to normal cpu pick */
>      }
>  
> -    min_avgload = MAX_LOAD;
> -
>      /* Find the runqueue with the lowest average load. */
>      for_each_cpu(i, &prv->active_queues)
>      {
> @@ -1552,7 +1555,7 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
>          if ( rqd_avgload < min_avgload )
>          {
>              min_avgload = rqd_avgload;
> -            min_rqi=i;
> +            min_rqi = i;
>          }
>      }
>  
> @@ -1560,6 +1563,8 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
>      if ( min_rqi == -1 )
>      {
>          new_cpu = get_fallback_cpu(svc);
> +        min_rqi = c2r(ops, new_cpu);
> +        min_avgload = prv->rqd[min_rqi].b_avgload;
>          goto out_up;
>      }
>  
> @@ -1570,18 +1575,18 @@ csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
>  
>   out_up:
>      read_unlock(&prv->lock);
> -
> + out:
>      if ( unlikely(tb_init_done) )
>      {
>          struct {
>              uint64_t b_avgload;
>              unsigned vcpu:16, dom:16;
>              unsigned rq_id:16, new_cpu:16;
> -       } d;
> -        d.b_avgload = prv->rqd[min_rqi].b_avgload;
> +        } d;
>          d.dom = vc->domain->domain_id;
>          d.vcpu = vc->vcpu_id;
> -        d.rq_id = c2r(ops, new_cpu);
> +        d.rq_id = min_rqi;
> +        d.b_avgload = min_avgload;
>          d.new_cpu = new_cpu;
>          __trace_var(TRC_CSCHED2_PICKED_CPU, 1,
>                      sizeof(d),
>
diff mbox

Patch

diff --git a/xen/common/sched_credit2.c b/xen/common/sched_credit2.c
index 7ec7f62..00d3300 100644
--- a/xen/common/sched_credit2.c
+++ b/xen/common/sched_credit2.c
@@ -1464,7 +1464,7 @@  csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
     struct csched2_private *prv = CSCHED2_PRIV(ops);
     int i, min_rqi = -1, new_cpu;
     struct csched2_vcpu *svc = CSCHED2_VCPU(vc);
-    s_time_t min_avgload;
+    s_time_t min_avgload = MAX_LOAD;
 
     ASSERT(!cpumask_empty(&prv->active_queues));
 
@@ -1488,7 +1488,12 @@  csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
     {
         /* We may be here because someone requested us to migrate. */
         __clear_bit(__CSFLAG_runq_migrate_request, &svc->flags);
-        return get_fallback_cpu(svc);
+        new_cpu = get_fallback_cpu(svc);
+        /*
+         * Tracing of runq and its load won't be accurate, since we could
+         * not get the lock, but at least we will output the chosen pcpu.
+         */
+        goto out;
     }
 
     /*
@@ -1513,8 +1518,6 @@  csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
         /* Fall-through to normal cpu pick */
     }
 
-    min_avgload = MAX_LOAD;
-
     /* Find the runqueue with the lowest average load. */
     for_each_cpu(i, &prv->active_queues)
     {
@@ -1552,7 +1555,7 @@  csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
         if ( rqd_avgload < min_avgload )
         {
             min_avgload = rqd_avgload;
-            min_rqi=i;
+            min_rqi = i;
         }
     }
 
@@ -1560,6 +1563,8 @@  csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
     if ( min_rqi == -1 )
     {
         new_cpu = get_fallback_cpu(svc);
+        min_rqi = c2r(ops, new_cpu);
+        min_avgload = prv->rqd[min_rqi].b_avgload;
         goto out_up;
     }
 
@@ -1570,18 +1575,18 @@  csched2_cpu_pick(const struct scheduler *ops, struct vcpu *vc)
 
  out_up:
     read_unlock(&prv->lock);
-
+ out:
     if ( unlikely(tb_init_done) )
     {
         struct {
             uint64_t b_avgload;
             unsigned vcpu:16, dom:16;
             unsigned rq_id:16, new_cpu:16;
-       } d;
-        d.b_avgload = prv->rqd[min_rqi].b_avgload;
+        } d;
         d.dom = vc->domain->domain_id;
         d.vcpu = vc->vcpu_id;
-        d.rq_id = c2r(ops, new_cpu);
+        d.rq_id = min_rqi;
+        d.b_avgload = min_avgload;
         d.new_cpu = new_cpu;
         __trace_var(TRC_CSCHED2_PICKED_CPU, 1,
                     sizeof(d),