Re: [PATCH] sched, fair: Allow a small load imbalance between low utilisation SD_NUMA domains v4

From: Mel Gorman
Date: Fri Jan 17 2020 - 09:26:36 EST


On Fri, Jan 17, 2020 at 02:16:15PM +0100, Vincent Guittot wrote:
> > A more interesting example is the Facebook schbench which uses a
> > number of messaging threads to communicate with worker threads. In this
> > configuration, one messaging thread is used per NUMA node and the number of
> > worker threads is varied. The 50, 75, 90, 95, 99, 99.5 and 99.9 percentiles
> > for response latency is then reported.
> >
> > Lat 50.00th-qrtle-1 44.00 ( 0.00%) 37.00 ( 15.91%)
> > Lat 75.00th-qrtle-1 53.00 ( 0.00%) 41.00 ( 22.64%)
> > Lat 90.00th-qrtle-1 57.00 ( 0.00%) 42.00 ( 26.32%)
> > Lat 95.00th-qrtle-1 63.00 ( 0.00%) 43.00 ( 31.75%)
> > Lat 99.00th-qrtle-1 76.00 ( 0.00%) 51.00 ( 32.89%)
> > Lat 99.50th-qrtle-1 89.00 ( 0.00%) 52.00 ( 41.57%)
> > Lat 99.90th-qrtle-1 98.00 ( 0.00%) 55.00 ( 43.88%)
>
> Which parameter changes between above and below tests ?
>
> > Lat 50.00th-qrtle-2 42.00 ( 0.00%) 42.00 ( 0.00%)
> > Lat 75.00th-qrtle-2 48.00 ( 0.00%) 47.00 ( 2.08%)
> > Lat 90.00th-qrtle-2 53.00 ( 0.00%) 52.00 ( 1.89%)
> > Lat 95.00th-qrtle-2 55.00 ( 0.00%) 53.00 ( 3.64%)
> > Lat 99.00th-qrtle-2 62.00 ( 0.00%) 60.00 ( 3.23%)
> > Lat 99.50th-qrtle-2 63.00 ( 0.00%) 63.00 ( 0.00%)
> > Lat 99.90th-qrtle-2 68.00 ( 0.00%) 66.00 ( 2.94%
> >

The number of worker pool threads. Above is 1 worker thread, below is 2.

> > @@ -8691,16 +8687,37 @@ static inline void calculate_imbalance(struct lb_env *env, struct sd_lb_stats *s
> > env->migration_type = migrate_task;
> > lsub_positive(&nr_diff, local->sum_nr_running);
> > env->imbalance = nr_diff >> 1;
> > - return;
> > - }
> > + } else {
> >
> > - /*
> > - * If there is no overload, we just want to even the number of
> > - * idle cpus.
> > - */
> > - env->migration_type = migrate_task;
> > - env->imbalance = max_t(long, 0, (local->idle_cpus -
> > + /*
> > + * If there is no overload, we just want to even the number of
> > + * idle cpus.
> > + */
> > + env->migration_type = migrate_task;
> > + env->imbalance = max_t(long, 0, (local->idle_cpus -
> > busiest->idle_cpus) >> 1);
> > + }
> > +
> > + /* Consider allowing a small imbalance between NUMA groups */
> > + if (env->sd->flags & SD_NUMA) {
> > + unsigned int imbalance_min;
> > +
> > + /*
> > + * Compute an allowed imbalance based on a simple
> > + * pair of communicating tasks that should remain
> > + * local and ignore them.
> > + *
> > + * NOTE: Generally this would have been based on
> > + * the domain size and this was evaluated. However,
> > + * the benefit is similar across a range of workloads
> > + * and machines but scaling by the domain size adds
> > + * the risk that lower domains have to be rebalanced.
> > + */
> > + imbalance_min = 2;
> > + if (busiest->sum_nr_running <= imbalance_min)
> > + env->imbalance = 0;
>
> Out of curiosity why have you decided to use the above instead of
> env->imbalance -= min(env->imbalance, imbalance_adj);
>
> Have you seen perf regression with the min ?
>

I didn't see a regression with min() but at this point, we're only
dealing with the case of ignoring a small imbalance when the busiest
group is almost completely idle. The distinction between using min and
just ignoring the imbalance is almost irrevelant in that case.

--
Mel Gorman
SUSE Labs