Message ID | 1400860385-14555-12-git-send-email-vincent.guittot@linaro.org (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Fri, May 23, 2014 at 05:53:05PM +0200, Vincent Guittot wrote: > The scheduler tries to compute how many tasks a group of CPUs can handle by > assuming that a task's load is SCHED_LOAD_SCALE and a CPU capacity is > SCHED_POWER_SCALE. > We can now have a better idea of the utilization of a group fo CPUs thanks to > group_actitvity and deduct how many capacity is still available. > > Signed-off-by: Vincent Guittot <vincent.guittot@linaro.org> > --- Right, so as Preeti already mentioned, this wrecks SMT. It also seems to loose the aggressive spread, where we want to run 1 task on each 'core' before we start 'balancing'. So I think we should be able to fix this by setting PREFER_SIBLING on the SMT domain, that way we'll get single tasks running on each SMT domain before filling them up until capacity. Now, its been a while since I looked at PREFER_SIBLING, and I've not yet looked at what your patch does to it, but it seems to me that that is the first direction we should look for an answer to this.
On Fri, May 23, 2014 at 05:53:05PM +0200, Vincent Guittot wrote: > @@ -6052,8 +6006,8 @@ static inline void update_sd_lb_stats(struct lb_env *env, struct sd_lb_stats *sd > * with a large weight task outweighs the tasks on the system). > */ > if (prefer_sibling && sds->local && > - sds->local_stat.group_has_capacity) > - sgs->group_capacity = min(sgs->group_capacity, 1U); > + sds->local_stat.group_capacity > 0) > + sgs->group_capacity = min(sgs->group_capacity, 1L); > > if (update_sd_pick_busiest(env, sds, sg, sgs)) { > sds->busiest = sg; > @@ -6228,7 +6182,7 @@ static inline void calculate_imbalance(struct lb_env *env, struct sd_lb_stats *s > * have to drop below capacity to reach cpu-load equilibrium. > */ > load_above_capacity = > - (busiest->sum_nr_running - busiest->group_capacity); > + (busiest->sum_nr_running - busiest->group_weight); > > load_above_capacity *= (SCHED_LOAD_SCALE * SCHED_POWER_SCALE); > load_above_capacity /= busiest->group_power; I think you just broke PREFER_SIBLING here.. So we want PREFER_SIBLING to work on nr_running, not utilization because we want to spread single tasks around, regardless of their utilization.
On 29 May 2014 15:55, Peter Zijlstra <peterz@infradead.org> wrote: > On Fri, May 23, 2014 at 05:53:05PM +0200, Vincent Guittot wrote: >> The scheduler tries to compute how many tasks a group of CPUs can handle by >> assuming that a task's load is SCHED_LOAD_SCALE and a CPU capacity is >> SCHED_POWER_SCALE. >> We can now have a better idea of the utilization of a group fo CPUs thanks to >> group_actitvity and deduct how many capacity is still available. >> >> Signed-off-by: Vincent Guittot <vincent.guittot@linaro.org> >> --- > > Right, so as Preeti already mentioned, this wrecks SMT. It also seems to > loose the aggressive spread, where we want to run 1 task on each 'core' > before we start 'balancing'. > > So I think we should be able to fix this by setting PREFER_SIBLING on > the SMT domain, that way we'll get single tasks running on each SMT > domain before filling them up until capacity. > > Now, its been a while since I looked at PREFER_SIBLING, and I've not yet > looked at what your patch does to it, but it seems to me that that is > the first direction we should look for an answer to this. OK, i'm going to look more deeply in PREFER_SIBLING too
On 29 May 2014 16:02, Peter Zijlstra <peterz@infradead.org> wrote: > On Fri, May 23, 2014 at 05:53:05PM +0200, Vincent Guittot wrote: >> @@ -6052,8 +6006,8 @@ static inline void update_sd_lb_stats(struct lb_env *env, struct sd_lb_stats *sd >> * with a large weight task outweighs the tasks on the system). >> */ >> if (prefer_sibling && sds->local && >> - sds->local_stat.group_has_capacity) >> - sgs->group_capacity = min(sgs->group_capacity, 1U); >> + sds->local_stat.group_capacity > 0) >> + sgs->group_capacity = min(sgs->group_capacity, 1L); >> >> if (update_sd_pick_busiest(env, sds, sg, sgs)) { >> sds->busiest = sg; >> @@ -6228,7 +6182,7 @@ static inline void calculate_imbalance(struct lb_env *env, struct sd_lb_stats *s >> * have to drop below capacity to reach cpu-load equilibrium. >> */ >> load_above_capacity = >> - (busiest->sum_nr_running - busiest->group_capacity); >> + (busiest->sum_nr_running - busiest->group_weight); >> >> load_above_capacity *= (SCHED_LOAD_SCALE * SCHED_POWER_SCALE); >> load_above_capacity /= busiest->group_power; > > I think you just broke PREFER_SIBLING here.. you mean by replacing the capacity which was reflecting the number of core for SMT by the group_weight ? > > So we want PREFER_SIBLING to work on nr_running, not utilization because > we want to spread single tasks around, regardless of their utilization. >
On Thu, May 29, 2014 at 09:56:24PM +0200, Vincent Guittot wrote: > On 29 May 2014 16:02, Peter Zijlstra <peterz@infradead.org> wrote: > > On Fri, May 23, 2014 at 05:53:05PM +0200, Vincent Guittot wrote: > >> @@ -6052,8 +6006,8 @@ static inline void update_sd_lb_stats(struct lb_env *env, struct sd_lb_stats *sd > >> * with a large weight task outweighs the tasks on the system). > >> */ > >> if (prefer_sibling && sds->local && > >> - sds->local_stat.group_has_capacity) > >> - sgs->group_capacity = min(sgs->group_capacity, 1U); > >> + sds->local_stat.group_capacity > 0) > >> + sgs->group_capacity = min(sgs->group_capacity, 1L); > >> > >> if (update_sd_pick_busiest(env, sds, sg, sgs)) { > >> sds->busiest = sg; > >> @@ -6228,7 +6182,7 @@ static inline void calculate_imbalance(struct lb_env *env, struct sd_lb_stats *s > >> * have to drop below capacity to reach cpu-load equilibrium. > >> */ > >> load_above_capacity = > >> - (busiest->sum_nr_running - busiest->group_capacity); > >> + (busiest->sum_nr_running - busiest->group_weight); > >> > >> load_above_capacity *= (SCHED_LOAD_SCALE * SCHED_POWER_SCALE); > >> load_above_capacity /= busiest->group_power; > > > > I think you just broke PREFER_SIBLING here.. > > you mean by replacing the capacity which was reflecting the number of > core for SMT by the group_weight ? Right to in the first hunk we lower group_capacity to 1 when prefer_sibling, then in the second hunk, you replace that group_capacity usage with group_weight. With the end result that prefer_sibling is now ineffective. That said, I fudged the prefer_sibling usage into the capacity logic, mostly because I could and it was already how the SMT stuff was working. But there is no reason we should continue to intertwine these two things. So I think it would be good to have a patch that implements prefer_sibling on nr_running separate from the existing capacity bits, and then convert the remaining capacity bits to utilization (or activity or whatever you did call it, see Morton's comments etc.).
On 30 May 2014 08:34, Peter Zijlstra <peterz@infradead.org> wrote: > On Thu, May 29, 2014 at 09:56:24PM +0200, Vincent Guittot wrote: >> On 29 May 2014 16:02, Peter Zijlstra <peterz@infradead.org> wrote: >> > On Fri, May 23, 2014 at 05:53:05PM +0200, Vincent Guittot wrote: >> >> @@ -6052,8 +6006,8 @@ static inline void update_sd_lb_stats(struct lb_env *env, struct sd_lb_stats *sd >> >> * with a large weight task outweighs the tasks on the system). >> >> */ >> >> if (prefer_sibling && sds->local && >> >> - sds->local_stat.group_has_capacity) >> >> - sgs->group_capacity = min(sgs->group_capacity, 1U); >> >> + sds->local_stat.group_capacity > 0) >> >> + sgs->group_capacity = min(sgs->group_capacity, 1L); >> >> >> >> if (update_sd_pick_busiest(env, sds, sg, sgs)) { >> >> sds->busiest = sg; >> >> @@ -6228,7 +6182,7 @@ static inline void calculate_imbalance(struct lb_env *env, struct sd_lb_stats *s >> >> * have to drop below capacity to reach cpu-load equilibrium. >> >> */ >> >> load_above_capacity = >> >> - (busiest->sum_nr_running - busiest->group_capacity); >> >> + (busiest->sum_nr_running - busiest->group_weight); >> >> >> >> load_above_capacity *= (SCHED_LOAD_SCALE * SCHED_POWER_SCALE); >> >> load_above_capacity /= busiest->group_power; >> > >> > I think you just broke PREFER_SIBLING here.. >> >> you mean by replacing the capacity which was reflecting the number of >> core for SMT by the group_weight ? > > Right to in the first hunk we lower group_capacity to 1 when prefer_sibling, > then in the second hunk, you replace that group_capacity usage with > group_weight. > > With the end result that prefer_sibling is now ineffective. ok > > That said, I fudged the prefer_sibling usage into the capacity logic, > mostly because I could and it was already how the SMT stuff was working. > But there is no reason we should continue to intertwine these two > things. > > So I think it would be good to have a patch that implements > prefer_sibling on nr_running separate from the existing capacity bits, > and then convert the remaining capacity bits to utilization (or activity > or whatever you did call it, see Morton's comments etc.). ok, i'm going to prepare such change > >
On 05/29/2014 07:25 PM, Peter Zijlstra wrote: > On Fri, May 23, 2014 at 05:53:05PM +0200, Vincent Guittot wrote: >> The scheduler tries to compute how many tasks a group of CPUs can handle by >> assuming that a task's load is SCHED_LOAD_SCALE and a CPU capacity is >> SCHED_POWER_SCALE. >> We can now have a better idea of the utilization of a group fo CPUs thanks to >> group_actitvity and deduct how many capacity is still available. >> >> Signed-off-by: Vincent Guittot <vincent.guittot@linaro.org> >> --- > > Right, so as Preeti already mentioned, this wrecks SMT. It also seems to > loose the aggressive spread, where we want to run 1 task on each 'core' > before we start 'balancing'. True. I just profiled the ebizzy runs and found that ebizzy threads were being packed onto a single core which is SMT-8 capable before spreading. This was a 6 core, SMT-8 machine. So for instance if I run 8 threads of ebizzy. the load balancing as record by perf sched record showed that two cores were packed upto 3 ebizzy threads and one core ran two ebizzy threads while the rest of the 3 cores were idle. I am unable to understand which part of this patch is aiding packing to a core. There is this check in this patch right? if (sgs->group_capacity < 0) return true; which should ideally prevent such packing? Because irrespective of the number of SMT threads, the capacity of a core is unchanged. And in the above scenario, we have 6 tasks on 3 cores. So shouldn't the above check have caught it? Regards Preeti U Murthy > > So I think we should be able to fix this by setting PREFER_SIBLING on > the SMT domain, that way we'll get single tasks running on each SMT > domain before filling them up until capacity. > > Now, its been a while since I looked at PREFER_SIBLING, and I've not yet > looked at what your patch does to it, but it seems to me that that is > the first direction we should look for an answer to this. >
On 2 June 2014 08:21, Preeti U Murthy <preeti@linux.vnet.ibm.com> wrote: > On 05/29/2014 07:25 PM, Peter Zijlstra wrote: >> On Fri, May 23, 2014 at 05:53:05PM +0200, Vincent Guittot wrote: >>> The scheduler tries to compute how many tasks a group of CPUs can handle by >>> assuming that a task's load is SCHED_LOAD_SCALE and a CPU capacity is >>> SCHED_POWER_SCALE. >>> We can now have a better idea of the utilization of a group fo CPUs thanks to >>> group_actitvity and deduct how many capacity is still available. >>> >>> Signed-off-by: Vincent Guittot <vincent.guittot@linaro.org> >>> --- >> >> Right, so as Preeti already mentioned, this wrecks SMT. It also seems to >> loose the aggressive spread, where we want to run 1 task on each 'core' >> before we start 'balancing'. > > True. I just profiled the ebizzy runs and found that ebizzy threads were > being packed onto a single core which is SMT-8 capable before spreading. > This was a 6 core, SMT-8 machine. So for instance if I run 8 threads of > ebizzy. the load balancing as record by perf sched record showed that > two cores were packed upto 3 ebizzy threads and one core ran two ebizzy > threads while the rest of the 3 cores were idle. > > I am unable to understand which part of this patch is aiding packing to > a core. There is this check in this patch right? > > if (sgs->group_capacity < 0) > return true; > > which should ideally prevent such packing? Because irrespective of the > number of SMT threads, the capacity of a core is unchanged. And in the > above scenario, we have 6 tasks on 3 cores. So shouldn't the above check > have caught it? yes, it should. the group_capacity should become < 0 because the CPU are fully loaded and the activity reach the max capacity value + nr_running > > Regards > Preeti U Murthy >> >> So I think we should be able to fix this by setting PREFER_SIBLING on >> the SMT domain, that way we'll get single tasks running on each SMT >> domain before filling them up until capacity. >> >> Now, its been a while since I looked at PREFER_SIBLING, and I've not yet >> looked at what your patch does to it, but it seems to me that that is >> the first direction we should look for an answer to this. >> >
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 2501e49..05b9502 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -5538,11 +5538,10 @@ struct sg_lb_stats { unsigned long group_power; unsigned long group_activity; /* Total activity of the group */ unsigned int sum_nr_running; /* Nr tasks running in the group */ - unsigned int group_capacity; + long group_capacity; unsigned int idle_cpus; unsigned int group_weight; int group_imb; /* Is there an imbalance in the group ? */ - int group_has_capacity; /* Is there extra capacity in the group? */ #ifdef CONFIG_NUMA_BALANCING unsigned int nr_numa_running; unsigned int nr_preferred_running; @@ -5785,31 +5784,6 @@ void update_group_power(struct sched_domain *sd, int cpu) } /* - * Try and fix up capacity for tiny siblings, this is needed when - * things like SD_ASYM_PACKING need f_b_g to select another sibling - * which on its own isn't powerful enough. - * - * See update_sd_pick_busiest() and check_asym_packing(). - */ -static inline int -fix_small_capacity(struct sched_domain *sd, struct sched_group *group) -{ - /* - * Only siblings can have significantly less than SCHED_POWER_SCALE - */ - if (!(sd->flags & SD_SHARE_CPUPOWER)) - return 0; - - /* - * If ~90% of the cpu_power is still there, we're good. - */ - if (group->sgp->power * 32 > group->sgp->power_orig * 29) - return 1; - - return 0; -} - -/* * Group imbalance indicates (and tries to solve) the problem where balancing * groups is inadequate due to tsk_cpus_allowed() constraints. * @@ -5843,33 +5817,6 @@ static inline int sg_imbalanced(struct sched_group *group) return group->sgp->imbalance; } -/* - * Compute the group capacity. - * - * Avoid the issue where N*frac(smt_power) >= 1 creates 'phantom' cores by - * first dividing out the smt factor and computing the actual number of cores - * and limit power unit capacity with that. - */ -static inline int sg_capacity(struct lb_env *env, struct sched_group *group) -{ - unsigned int capacity, smt, cpus; - unsigned int power, power_orig; - - power = group->sgp->power; - power_orig = group->sgp->power_orig; - cpus = group->group_weight; - - /* smt := ceil(cpus / power), assumes: 1 < smt_power < 2 */ - smt = DIV_ROUND_UP(SCHED_POWER_SCALE * cpus, power_orig); - capacity = cpus / smt; /* cores */ - - capacity = min_t(unsigned, capacity, DIV_ROUND_CLOSEST(power, SCHED_POWER_SCALE)); - if (!capacity) - capacity = fix_small_capacity(env->sd, group); - - return capacity; -} - /** * update_sg_lb_stats - Update sched_group's statistics for load balancing. * @env: The load balancing environment. @@ -5918,10 +5865,9 @@ static inline void update_sg_lb_stats(struct lb_env *env, sgs->group_weight = group->group_weight; sgs->group_imb = sg_imbalanced(group); - sgs->group_capacity = sg_capacity(env, group); + sgs->group_capacity = group->sgp->power_orig - sgs->group_activity; + - if (sgs->group_capacity > sgs->sum_nr_running) - sgs->group_has_capacity = 1; } /** @@ -5945,7 +5891,15 @@ static bool update_sd_pick_busiest(struct lb_env *env, if (sgs->avg_load <= sds->busiest_stat.avg_load) return false; - if (sgs->sum_nr_running > sgs->group_capacity) + /* The group has an obvious long run overload */ + if (sgs->group_capacity < 0) + return true; + + /* + * The group has a short run overload because more tasks than available + * CPUs are running + */ + if (sgs->sum_nr_running > sgs->group_weight) return true; /* @@ -6052,8 +6006,8 @@ static inline void update_sd_lb_stats(struct lb_env *env, struct sd_lb_stats *sd * with a large weight task outweighs the tasks on the system). */ if (prefer_sibling && sds->local && - sds->local_stat.group_has_capacity) - sgs->group_capacity = min(sgs->group_capacity, 1U); + sds->local_stat.group_capacity > 0) + sgs->group_capacity = min(sgs->group_capacity, 1L); if (update_sd_pick_busiest(env, sds, sg, sgs)) { sds->busiest = sg; @@ -6228,7 +6182,7 @@ static inline void calculate_imbalance(struct lb_env *env, struct sd_lb_stats *s * have to drop below capacity to reach cpu-load equilibrium. */ load_above_capacity = - (busiest->sum_nr_running - busiest->group_capacity); + (busiest->sum_nr_running - busiest->group_weight); load_above_capacity *= (SCHED_LOAD_SCALE * SCHED_POWER_SCALE); load_above_capacity /= busiest->group_power; @@ -6294,6 +6248,7 @@ static struct sched_group *find_busiest_group(struct lb_env *env) local = &sds.local_stat; busiest = &sds.busiest_stat; + /* ASYM feature bypasses nice load balance check */ if ((env->idle == CPU_IDLE || env->idle == CPU_NEWLY_IDLE) && check_asym_packing(env, &sds)) return sds.busiest; @@ -6313,8 +6268,8 @@ static struct sched_group *find_busiest_group(struct lb_env *env) goto force_balance; /* SD_BALANCE_NEWIDLE trumps SMP nice when underutilized */ - if (env->idle == CPU_NEWLY_IDLE && local->group_has_capacity && - !busiest->group_has_capacity) + if (env->idle == CPU_NEWLY_IDLE && (local->group_capacity > 0) + && (busiest->group_capacity < 0)) goto force_balance; /* @@ -6372,7 +6327,7 @@ static struct rq *find_busiest_queue(struct lb_env *env, int i; for_each_cpu_and(i, sched_group_cpus(group), env->cpus) { - unsigned long power, capacity, wl; + unsigned long power, wl; enum fbq_type rt; rq = cpu_rq(i); @@ -6400,18 +6355,17 @@ static struct rq *find_busiest_queue(struct lb_env *env, if (rt > env->fbq_type) continue; - power = power_of(i); - capacity = DIV_ROUND_CLOSEST(power, SCHED_POWER_SCALE); - if (!capacity) - capacity = fix_small_capacity(env->sd, group); - wl = weighted_cpuload(i); /* * When comparing with imbalance, use weighted_cpuload() * which is not scaled with the cpu power. */ - if (capacity && rq->nr_running == 1 && wl > env->imbalance) + + power = power_of(i); + + if (rq->nr_running == 1 && wl > env->imbalance && + ((power * env->sd->imbalance_pct) >= (rq->cpu_power_orig * 100))) continue; /*
The scheduler tries to compute how many tasks a group of CPUs can handle by assuming that a task's load is SCHED_LOAD_SCALE and a CPU capacity is SCHED_POWER_SCALE. We can now have a better idea of the utilization of a group fo CPUs thanks to group_actitvity and deduct how many capacity is still available. Signed-off-by: Vincent Guittot <vincent.guittot@linaro.org> --- kernel/sched/fair.c | 94 ++++++++++++++--------------------------------------- 1 file changed, 24 insertions(+), 70 deletions(-)