top-down balance purpose discussion -- resend

From: Alex Shi
Date: Tue Jan 21 2014 - 09:04:39 EST



Current scheduler load balance is bottom-up mode, each CPU need
initiate the balance by self.

1, Like in a integrate computer system, it has smt/core/cpu/numa, 4
level scheduler domains. If there is just 2 tasks in whole system that
both running on cpu0. Current load balance need to pull task to another
smt in smt domain, then pull task to another core, then pull task to
another cpu, finally pull task to another numa. Totally it is need 4
times task moving to get system balance.

Generally, the task moving complexity is
O(nm log n), n := nr_cpus, m := nr_tasks

There is a excellent summary and explanation for this in
kernel/sched/fair.c:4605

Another weakness of current LB is that every cpu need to get the other
cpus' load info repeatedly and try to figure out busiest sched
group/queue on every sched domain level. But it just waste time, since
it may not conduct a task moving. One of reasons is that cpu can only
pull task, not pushing.

2, Consider huge cost of task moving: CS, tlb/cache refill, and the
useless remote cpu load info getting. If we can have better solution for
load balance, like reduce the balance times to.
O(m) m := nr_tasks

It will be a great win on performance. like above example, we can move
task from cpu0 direct to another numa. that only need 1 task moving,
save 3 CS and tlb/cache refill.

To get this point, a rough idea is changing the load balance behaviour
to top-down mode. And only load balance just be done by on on cpu.
Like let each of cpu report self load status on per-cpu
memory. And a baby-sitter in system to collect these cpus load info,
then decide how to move task centralize, finally send IPI to each hungry
cpu to let them pull load quota from appointed cpus.

Like in above example, the baby-sitter will fetch each cpus' load info,
then send a pull task IPI to let a cpu in another numa pull one task
from cpu0. So in the task pulling, we still just involved 2 cpus, can
reuse move_tasks() functions.

3, From power saving POV, top-down can get the whole cpu power topology
info before balance. So we can start balance with the power
consideration firstly, instead of stop balance at last minute after
getting every remote cpu info.

4, One of concern of top-down mode is that we need to remote access top
level domain cpu info in large system while each cpu keep updates its
load. That may causes cache bouncing issue. (current LB also has this
problem, but it set a long load balance interval to reduce LB times)

Paul has given lots of excellent suggestion to resolve/relief this
issue. the following is his suggestions:
===
Might be -- key thing is to measure it on a big system. If it is a
problem, here are some workarounds with varying degrees of sanity:

1. Reduce cache misses by keeping three values:

a. Current load value.
b. Last exported value. (Should be in same cache line
as a.)
c. Exported value. (Should be in separate cache line.)

The avoid writing to c unless a has deviated sufficiently from a.
If the load values stay constant for long time periods, this
can reduce the number of cache misses. On the other hand,
it introduces an extra compare and branch -- though this should
not be a problem if the load value is computed relatively
infrequently.

2. As above, but provide additional information to allow the
other CPU to extrapolate values. For example, if a CPU goes
idle, some of its values will decay exponentially. The
remote CPU could compute the decay rate, removing the need
for the subject CPU to wake up to recompute its value.
(Maybe you already do this?)

Similarly if a CPU remains CPU-bound with given runqueue
loading for some time.

3. Allow the exported values to become inaccurate, and resample
the actual values remotely if extrapolated values indicate
that action is warranted.

There are probably other approaches. I am being quite general here
because I don't have the full picture of the scheduler statistics
in my head. It is likely possible to obtain a much better approach
by considering the scheduler's specifics.

--
Thanks
Alex
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/