Re: [PATCH -v7][RFC]: mutex: implement adaptive spinning

From: Steven Rostedt
Date: Fri Jan 09 2009 - 11:00:10 EST



On Fri, 9 Jan 2009, Peter Zijlstra wrote:

> On Fri, 2009-01-09 at 11:47 +0100, Peter Zijlstra wrote:
>
> > > So I think the bug is still there, we just hid it better by breaking out
> > > of the loop with that "if (need_resched())" always eventually triggering.
> > > And it would be ok if it really is guaranteed to _eventually_ trigger, and
> > > I guess with timeslices it eventually always will, but I suspect we could
> > > have some serious latency spikes.
> >
> > Yes, the owner getting preempted after acquiring the lock, but before
> > setting the owner can give some nasties :-(
> >
> > I initially did that preempt_disable/enable around the fast path, but I
> > agree that slowing down the fast path is unwelcome.
> >
> > Alternatively we could go back to block on !owner, with the added
> > complexity of not breaking out of the spin on lock->owner != owner
> > when !lock->owner, so that the premature owner clearing of the unlock
> > fast path will not force a schedule right before we get a chance to
> > acquire the lock.
> >
> > Let me do that..
>
> Ok a few observations..
>
> Adding that need_resched() in the outer loop utterly destroys the
> performance gain for PREEMPT=y. Voluntary preemption is mostly good, but
> somewhat unstable results.

I was going to say a while ago...
In PREEMPT=y the need_resched() is not needed at all. If you have
preemption enabled, you will get preempted in that loop. No need for the
need_resched() in the outer loop. Although I'm not sure how it would even
hit the "need_resched". If it was set, then it is most likely going to be
cleared when coming back from being preempted.

>
> Adding that blocking on !owner utterly destroys everything.

I was going to warn you about that ;-)

Without the check for !owner, you are almost guaranteed to go to sleep
every time. Here's why:

You are spinning and thus have a hot cache on that CPU.

The owner goes to unlock but will be in a cold cache. It sets lock->owner
to NULL, but is still in cold cache so it is a bit slower.

Once the spinner sees the NULL, it shoots out of the spin but sees the
lock is still not available then goes to sleep. All before the owner could
release it. This could probably happen at every contention. Thus, you lose
the benefit of spinning. You probably make things worse because you add a
spin before every sleep.

-- Steve

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/