Re: [PATCH v2 0/2] ic2: mux: pca9541: add delayed-release support

From: Zev Weiss
Date: Wed Mar 02 2022 - 19:43:29 EST


On Wed, Mar 02, 2022 at 06:43:31AM PST, Peter Rosin wrote:
On 2022-02-28 23:38, Zev Weiss wrote:
Hi Peter,

Thanks for the reply! (More below.)

On Mon, Feb 28, 2022 at 01:54:09PM PST, Peter Rosin wrote:
On 2022-02-01 01:18, Zev Weiss wrote:
Hello,

This series adds support for a new pca9541 device-tree property
("release-delay-us"), which delays releasing ownership of the bus
after a transaction for a configurable duration, anticipating that
another transaction may follow shortly. By avoiding a
release/reacquisition between transactions, this can provide a
substantial performance improvement for back-to-back operations -- on
a Delta AHE-50DC (ASPEED AST1250) system running OpenBMC with dozens
of LM25066 PMICs on PCA9541-arbitrated busses, a setting of 10000 (10
ms) reduces the median latency the psusensor daemon's hwmon sysfs file
reads from 2.28 ms to 0.99 ms (a 57% improvement).

Hi!

Sorry for the late reply. It seems I'm forever swamped...

There is a risk with this scheme. If you have two (or more) of these
chips on the same bus, and there are clients behind these two chips
that have the same address, accesses to one of the clients might
"leak through" the other arb to an unexpected client when its arb
is in it's release-delay state.

In other words, it is no coincidence that the segment lock is held
over the whole acquire-access-release cycle.

That's not a scenario I had considered, but I think I see what you're
saying. Just to make sure I'm understanding correctly, the problematic
situation you're describing would involve multiple (sibling, not
parent/child cascaded) arbiters at distinct addresses on the same
(master-side) bus, in effect acting as a sort of "distributed mux" in
addition to arbitrating between multiple attached masters? (So kind of
an M-to-N arrangement between M masters and N busses.)

In which case if more than one of the arbiters had their downstream
busses simultaneously connected to the same master (as could result if
one were still connected due to a delayed release while the master had
already started a subsequent transaction via another arbiter), the
resulting "combined" bus could end up with address collisions between
devices downstream of the arbiters if there are common addresses in use
between the downstream busses.

Yes, you understood me correctly. That is, if I understood you
correctly :-)


Okay, thanks for confirming.


Sure, you can always say "don't add a release-delay when ...", but I
see no such documentation.


Assuming I haven't misunderstood the above, would expanding the
description of the property in the DT binding like so be sufficient?

- release-delay-us: the number of microseconds to delay before
releasing the bus after a transaction. If unspecified the default
is zero (the bus is released immediately). Non-zero values can
reduce arbitration overhead for back-to-back transactions, at the
cost of delaying the other master's access to the bus.

If this property is employed on hardware with multiple parallel
(not cascaded) arbiters selecting between multiple downstream
busses, address conflicts can occur if a device on one of the
downstream busses uses the same address as a device on another
downstream bus. This property should thus only be used if either
(a) there is only one arbiter on the bus, (b) multiple arbiters are
strictly cascaded to a single downstream bus, or (c) all of the
devices on all downstream busses use addresses that are unique
across all of those busses.

If so I'll send a v3 with that change shortly.

I'm not sure this is a bindings thing or a quality of implementation
issue. You could imagine an implementation where the arb is
opportunistically left connected for the release-delay, but that the
arb is disconnected immediately if/when there is a change of the bus
topology, possibly only if the topology change in turn cause address
conflicts. The implication is that I'm not sure if this caveat should
be described in the bindings documentation or elsewhere. It sure would
be convenient to see it with the bindings, because if it is elsewhere
I'm sure more people will fail to take notice.


Yeah, I was thinking along similar lines -- bindings don't seem like exactly the "right" place for it, but perhaps the pragmatic place in order for people to actually see it. I guess maybe this sort of gets back into questions of what exactly DT should describe and to what extent this is a property of the hardware vs. the software running on it...

Automating an earlier release in the event of a potential conflict being detected would be nice, though given the possibility of devices on the bus that the kernel doesn't know about (e.g. things driven directly from userspace via the i2c-dev interface) it doesn't seem entirely feasible to do that soundly. A conservative approximation (e.g. releasing on the next transaction on the parent physical bus that's not from the same arbiter's virtual bus) seems like it could probably retain most of the benefit of the delayed release, though implementing it would probably require some plumbing in the i2c code that I'm guessing doesn't currently exist. Thoughts?

Another risk with the scheme is that you possibly lock out the other
master for so long time that you trigger it to force its way in, thus
possible wrecking some transaction. Or is the PCA9541 defending against
such wreckage on "hostile" takeovers? (I too lack the HW and time to
tinker with this.) If so, that might of course happen anyway, but it
might be so much more common if the bus is left connected.


Yeah, this is sort of what the "at the cost of..." bit in the dt-bindings description was getting at, if a bit obliquely. It's a risk that's not unique to this feature being in use, but yes, using it does of course increase the risk. For what my datapoint is worth, it hasn't been problematic in my use of it thus far in a system with two fairly heavily-utilized busses (and a handful of more lightly used ones) each shared by two masters, both of which are Linux hosts running this driver with a release delay of 10ms.


Thanks,
Zev