Re: Network slowdown due to CFS

From: Ingo Molnar
Date: Thu Sep 27 2007 - 05:46:59 EST

* Jarek Poplawski <jarkao2@xxxxx> wrote:

> > the (small) patch below fixes the iperf locking bug and removes the
> > yield() use. There are numerous immediate benefits of this patch:
> ...
> >
> > sched_yield() is almost always the symptom of broken locking or other
> > bug. In that sense CFS does the right thing by exposing such bugs =B-)
> ...Only if it were under some DEBUG option. [...]

note that i qualified my sentence both via "In that sense" and via a
smiley! So i was not suggesting that this is a general rule at all and i
was also joking :-)

> [...] Even if iperf is doing the wrong thing there is no explanation
> for such big difference in the behavior between sched_compat_yield 1
> vs. 0. It seems common interfaces should work similarly and
> predictably on various systems, and here, if I didn't miss something,
> linux looks like a different kind?

What you missed is that there is no such thing as "predictable yield
behavior" for anything but SCHED_FIFO/RR tasks (for which tasks CFS does
keep the behavior). Please read this thread on lkml for a more detailed

CFS: some bad numbers with Java/database threading [FIXED]

in short: the yield implementation was tied to the O(1) scheduler, so
the only way to have the exact same behavior would be to have the exact
same core scheduler again. If what you said was true we would not be
able to change the scheduler, ever. For something as vaguely defined of
an API as yield, there's just no way to have a different core scheduler
and still behave the same way.

So _generally_ i'd agree with you that normally we want to be bug for
bug compatible, but in this specific (iperf) case there's just no point
in preserving behavior that papers over this _clearly_ broken user-space
app/thread locking (for which now two fixes exist already, plus a third
fix is the twiddling of that sysctl).

To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at
Please read the FAQ at