Re: RCU idle CPU detection is broken in linux-next

From: Paul E. McKenney
Date: Fri Sep 21 2012 - 11:12:31 EST


On Fri, Sep 21, 2012 at 03:26:27PM +0200, Sasha Levin wrote:
> On 09/21/2012 02:13 PM, Paul E. McKenney wrote:
> >> This might be unrelated, but I got the following dump as well when trinity
> >> > decided it's time to reboot my guest:
> > OK, sounds like we should hold off until you reproduce, then.
>
> I'm not sure what you mean.
>
> There are basically two issues I'm seeing now, which reproduce pretty much every
> time:
>
> 1. The "using when idle" warning.
> 2. The rcu related hangs during shutdown.
>
> The first one appears early on when I start fuzzing, the other one happens when
> shutting down - so both of them are reproducible in the same session.

Ah, I misunderstood the "reboot my guest" -- I thought that you were
doing something like repeated modprobe/rmmod cycles on rcutorture while
running the guest for an extended time period. That will teach me not
to reply to email so soon after waking up. ;-)

That said, #2 is expected behavior given the RCU CPU stall warnings in
your Sept. 20 dmesg. This is because rcutorture does rcu_barrier() on
the way out, which cannot complete if grace periods are not completing.
And the later soft lockup is also likely a consequence of the stall,
because CPU hotplug does a synchronize_sched() while holding the hotplug
lock, which will then cause get_online_cpus() to hang.

Looking further down, there are hung tasks that are waiting for a
timeout, but this is also a consequence of the hang because they
are waiting for MAX_SCHEDULE_TIMEOUT -- in other words, they are
waiting to be killed at shutdown time. I could suppress this by using
schedule_timeout_interruptible() in a loop in order to reduce the noise
in this case.

The remaining traces in that email are also consequences of the stall.

So why the stall?

Using RCU from a CPU that RCU believes to be idle can cause arbitrary
bad behavior (possibly including stalls), but with very low probability.
The reason that things can go arbitrarily bad is that RCU is ignoring
the CPU, and thus not waiting for any RCU read-side critical sections.
This could of course result in abitrary corruption of memory. The reason
for the low probability is that grace periods tend to be long and RCU
read-side critical sections tend to be short.

It looks like you are running -next, which has RCU grace periods driven
by a kthread. Is it possible that this kthread is not getting a chance
to run (in fact, the "Stall ended before state dump start" is consistent
with that possibility), but in that case I would expect to see a soft
lockup from it. Furthermore, in that case, it would be expected to
start running again as soon as things started going idle during shutdown.

Or did the system somehow manage to stay busy despite being in shutdown?
Or, for that matter, are you overcommitting the physical CPUs on your
trinity test setup?

Thanx, Paul

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/