Re: [net-next] net: dsa: felix: disable always guard band bit for TAS config

From: Michael Walle
Date: Thu May 06 2021 - 09:25:19 EST


Am 2021-05-04 23:33, schrieb Vladimir Oltean:
[ trimmed the CC list, as this is most likely spam for most people ]

On Tue, May 04, 2021 at 10:23:11PM +0200, Michael Walle wrote:
Am 2021-05-04 21:17, schrieb Vladimir Oltean:
> On Tue, May 04, 2021 at 09:08:00PM +0200, Michael Walle wrote:
> > > > > > As explained in another mail in this thread, all queues are marked as
> > > > > > scheduled. So this is actually a no-op, correct? It doesn't matter if
> > > > > > it set or not set for now. Dunno why we even care for this bit then.
> > > > >
> > > > > It matters because ALWAYS_GUARD_BAND_SCH_Q reduces the available
> > > > > throughput when set.
> > > >
> > > > Ahh, I see now. All queues are "scheduled" but the guard band only
> > > > applies
> > > > for "non-scheduled" -> "scheduled" transitions. So the guard band is
> > > > never
> > > > applied, right? Is that really what we want?
> > >
> > > Xiaoliang explained that yes, this is what we want. If the end user
> > > wants a guard band they can explicitly add a "sched-entry 00" in the
> > > tc-taprio config.
> >
> > You're disabling the guard band, then. I figured, but isn't that
> > suprising for the user? Who else implements taprio? Do they do it in
> > the
> > same way? I mean this behavior is passed right to the userspace and
> > have
> > a direct impact on how it is configured. Of course a user can add it
> > manually, but I'm not sure that is what we want here. At least it
> > needs
> > to be documented somewhere. Or maybe it should be a switchable option.
> >
> > Consider the following:
> > sched-entry S 01 25000
> > sched-entry S fe 175000
> > basetime 0
> >
> > Doesn't guarantee, that queue 0 is available at the beginning of
> > the cycle, in the worst case it takes up to
> > <begin of cycle> + ~12.5us until the frame makes it through (given
> > gigabit and 1518b frames).
> >
> > Btw. there are also other implementations which don't need a guard
> > band (because they are store-and-forward and cound the remaining
> > bytes). So yes, using a guard band and scheduling is degrading the
> > performance.
>
> What is surprising for the user, and I mentioned this already in another
> thread on this patch, is that the Felix switch overruns the time gate (a
> packet taking 2 us to transmit will start transmission even if there is
> only 1 us left of its time slot, delaying the packets from the next time
> slot by 1 us). I guess that this is why the ALWAYS_GUARD_BAND_SCH_Q bit
> exists, as a way to avoid these overruns, but it is a bit of a poor tool
> for that job. Anyway, right now we disable it and live with the
> overruns.

We are talking about the same thing here. Why is that a poor tool?

It is a poor tool because it revolves around the idea of "scheduled
queues" and "non-scheduled queues".

Consider the following tc-taprio schedule:

sched-entry S 81 2000 # TC 7 and 0 open, all others closed
sched-entry S 82 2000 # TC 7 and 1 open, all others closed
sched-entry S 84 2000 # TC 7 and 2 open, all others closed
sched-entry S 88 2000 # TC 7 and 3 open, all others closed
sched-entry S 90 2000 # TC 7 and 4 open, all others closed
sched-entry S a0 2000 # TC 7 and 5 open, all others closed
sched-entry S c0 2000 # TC 7 and 6 open, all others closed

Otherwise said, traffic class 7 should be able to send any time it
wishes.

What is the use case behind that? TC7 (with the highest priority)
may always take precedence of the other TCs, thus what is the point
of having a dedicated window for the others.

Anyway, I've tried it and there are no hiccups. I've meassured
the delta between the start of successive packets and they are
always ~12370ns for a 1518b frame. TC7 is open all the time,
which makes sense. It only happens if you actually close the gate,
eg. you have a sched-entry where a TC7 bit is not set. In this case,
I can see a difference between ALWAYS_GUARD_BAND_SCH_Q set and not
set. If it is set, there is up to a ~12.5us delay added (of course
it depends on when the former frame was scheduled).

It seems that also needs to be 1->0 transition.

You've already mentioned that the switch violates the Qbv standard.
What makes you think so? IMHO before that patch, it wasn't violated.
Now it likely is (still have to confirm that). How can this
be reasonable?

If you have a look at the initial commit message, it is about
making it possible to have a smaller gate window, but that is not
possible because of the current guard band of ~12.5us. It seems
to be a shortcut for not having the MAXSDU (and thus the length
of the guard band) configurable. Yes (static) guard bands will
have a performance impact, also described in [1]. You are trading
the correctness of the TAS for performance. And it is the sole
purpose of Qbv to have a determisitc way (in terms of timing) of
sending the frames.

And telling the user, hey, we know we violate the Qbv standard,
please insert the guard bands yourself if you really need them is
not a real solution. As already mentioned, (1) it is not documented
anywhere, (2) can't be shared among other switches (unless they do
the same workaround) and (3) what am I supposed to do for TSN compliance
testing. Modifying the schedule that is about to be checked (and thus
given by the compliance suite)?

With the ALWAYS_GUARD_BAND_SCH_Q bit, there will be hiccups in packet
transmission for TC 7. For example, at the end of every time slot,
the hardware will insert a guard band for TC 7 because there is a
scheduled-queue-to-scheduled-queue transition, and it has been told to
do that. But a packet with TC 7 should be transmitted at any time,
because that's what we told the port to do!

Alternatively, we could tell the switch that TC 7 is "scheduled", and
the others are "not scheduled". Then it would implement the guard band
at the end of TCs 0-6, but it wouldn't for packets sent in TC 7. But
when you look at the overall schedule I described above, it kinds looks
like TCs 0-6 are the ones that are "scheduled" and TC 7 looks like the
one which isn't "scheduled" but can send at any time it pleases.

Odd, just odd. It's clear that someone had something in mind, it's just
not clear what. I would actually appreciate if somebody from Microchip
could chime in and say "no, you're wrong", and then explain.

If I had to make a bet, the distinction between "scheduled" and
"non-scheduled" is there to have more control for some traffic classes
you trust and where you can engineer the traffic, so you don't really
need the guard band and between arbitrary traffic where you can't really
say anything about and thus need the guard band.

> FWIW, the ENETC does not overrun the time gate, the SJA1105 does. You
> can't really tell just by looking at the driver code, just by testing.
> It's a bit of a crapshoot.

I was speaking of other switches, I see there is also a hirschmann
switch (hellcreek) supported in linux, for example.

Shouldn't the goal to make the configuration of the taprio qdisc
independent of the switch. If on one you'll have to manually define the
guard band by inserting dead-time scheduler entries and on another this
is already handled by the hardware (like it would be with
ALWAYS_GUARD_BAND_SCH_Q or if it doesn't need it at all), this goal
isn't met.

Also what do you expect if you use the following configuration:
sched-entry S 01 5000
sched-entry S fe <some large number>

Will queue 0 be able to send traffic? To me, with this patch, it seems
to me that this isn't always the case anymore. If there is a large packet
just sent at the end of the second cycle, the first might even be skipped
completely.
Will a user of the taprio (without knowledge of the underlying switch)
assume that it can send traffic up to ~600 bytes? I'd say yes.

Yeah, I think that if a switch overruns a packet's reserved time gate,
then the above tc-taprio schedule is as good as not having any. I didn't
say that overruns are not a problem, I just said that the ALWAYS_blah_blah
bit isn't as silver-bullet for a solution as you think.

See above.

-michael

[1] https://www.belden.com/hubfs/resources/knowledge/white-papers/tsn-time-sensitive-networking.pdf