Re: [PATCH] vfio/pci: Add OpRegion 2.0 Extended VBT support.

From: Alex Williamson
Date: Thu Aug 26 2021 - 21:48:56 EST


On Fri, 27 Aug 2021 09:36:36 +0800 (CST)
Colin Xu <colin.xu@xxxxxxxxx> wrote:

> Hi Alex,
>
> In addition to the background that devices on market may still need
> OpRegion 2.0 support in vfio-pci, do you have other comments to the patch
> body?

Yes, there were further comments in my first reply below. Thanks,

Alex


> On Tue, 17 Aug 2021, Colin Xu wrote:
>
> > On Mon, 16 Aug 2021, Alex Williamson wrote:
> >
> >> On Fri, 13 Aug 2021 10:13:29 +0800
> >> Colin Xu <colin.xu@xxxxxxxxx> wrote:
> >>
> >>> Due to historical reason, some legacy shipped system doesn't follow
> >>> OpRegion 2.1 spec but still stick to OpRegion 2.0, in which the extended
> >>> VBT is not contigious after OpRegion in physical address, but any
> >>> location pointed by RVDA via absolute address. Thus it's impossible
> >>> to map a contigious range to hold both OpRegion and extended VBT as 2.1.
> >>>
> >>> Since the only difference between OpRegion 2.0 and 2.1 is where extended
> >>> VBT is stored: For 2.0, RVDA is the absolute address of extended VBT
> >>> while for 2.1, RVDA is the relative address of extended VBT to OpRegion
> >>> baes, and there is no other difference between OpRegion 2.0 and 2.1,
> >>> it's feasible to amend OpRegion support for these legacy system (before
> >>> upgrading the system firmware), by kazlloc a range to shadown OpRegion
> >>> from the beginning and stitch VBT after closely, patch the shadow
> >>> OpRegion version from 2.0 to 2.1, and patch the shadow RVDA to relative
> >>> address. So that from the vfio igd OpRegion r/w ops view, only OpRegion
> >>> 2.1 is exposed regardless the underneath host OpRegion is 2.0 or 2.1
> >>> if the extended VBT exists. vfio igd OpRegion r/w ops will return either
> >>> shadowed data (OpRegion 2.0) or directly from physical address
> >>> (OpRegion 2.1+) based on host OpRegion version and RVDA/RVDS. The shadow
> >>> mechanism makes it possible to support legacy systems on the market.
> >>
> >> Which systems does this enable? There's a suggestion above that these
> >> systems could update firmware to get OpRegion v2.1 support, why
> >> shouldn't we ask users to do that instead? When we added OpRegion v2.1
> >> support we were told that v2.0 support was essentially non-existent,
> >> why should we add code to support and old spec with few users for such
> >> a niche use case?
> > Hi Alex, there was some mis-alignment with the BIOS owner that we were told
> > the 2.0 system doesn't for retail but only for internal development. However
> > in other projects we DO see the retail market has such systems, including NUC
> > NUC6CAYB, some APL industrial PC used in RT system, and some customized APL
> > motherboard by commercial virtualization solution. We immediately contact the
> > BIOS owner to ask for a clarification and they admit it. These system won't
> > get updated BIOS for OpRegion update but still under warranty. That's why the
> > OpRegion 2.0 support is still needed.
> >
> >>
> >>> Cc: Zhenyu Wang <zhenyuw@xxxxxxxxxxxxxxx>
> >>> Cc: Hang Yuan <hang.yuan@xxxxxxxxxxxxxxx>
> >>> Cc: Swee Yee Fonn <swee.yee.fonn@xxxxxxxxx>
> >>> Cc: Fred Gao <fred.gao@xxxxxxxxx>
> >>> Signed-off-by: Colin Xu <colin.xu@xxxxxxxxx>
> >>> ---
> >>> drivers/vfio/pci/vfio_pci_igd.c | 117 ++++++++++++++++++++------------
> >>> 1 file changed, 75 insertions(+), 42 deletions(-)
> >>>
> >>> diff --git a/drivers/vfio/pci/vfio_pci_igd.c
> >>> b/drivers/vfio/pci/vfio_pci_igd.c
> >>> index 228df565e9bc..22b9436a3044 100644
> >>> --- a/drivers/vfio/pci/vfio_pci_igd.c
> >>> +++ b/drivers/vfio/pci/vfio_pci_igd.c
> >>> @@ -48,7 +48,10 @@ static size_t vfio_pci_igd_rw(struct vfio_pci_device
> >>> *vdev, char __user *buf,
> >>> static void vfio_pci_igd_release(struct vfio_pci_device *vdev,
> >>> struct vfio_pci_region *region)
> >>> {
> >>> - memunmap(region->data);
> >>> + if (is_ioremap_addr(region->data))
> >>> + memunmap(region->data);
> >>> + else
> >>> + kfree(region->data);
> >>> }
> >>>
> >>> static const struct vfio_pci_regops vfio_pci_igd_regops = {
> >>> @@ -59,10 +62,11 @@ static const struct vfio_pci_regops
> >>> vfio_pci_igd_regops = {
> >>> static int vfio_pci_igd_opregion_init(struct vfio_pci_device *vdev)
> >>> {
> >>> __le32 *dwordp = (__le32 *)(vdev->vconfig + OPREGION_PCI_ADDR);
> >>> - u32 addr, size;
> >>> - void *base;
> >>> + u32 addr, size, rvds = 0;
> >>> + void *base, *opregionvbt;
> >>> int ret;
> >>> u16 version;
> >>> + u64 rvda = 0;
> >>>
> >>> ret = pci_read_config_dword(vdev->pdev, OPREGION_PCI_ADDR, &addr);
> >>> if (ret)
> >>> @@ -89,66 +93,95 @@ static int vfio_pci_igd_opregion_init(struct
> >>> vfio_pci_device *vdev)
> >>> size *= 1024; /* In KB */
> >>>
> >>> /*
> >>> - * Support opregion v2.1+
> >>> - * When VBT data exceeds 6KB size and cannot be within mailbox #4,
> >>> then
> >>> - * the Extended VBT region next to opregion is used to hold the VBT
> >>> data.
> >>> - * RVDA (Relative Address of VBT Data from Opregion Base) and RVDS
> >>> - * (Raw VBT Data Size) from opregion structure member are used to
> >>> hold the
> >>> - * address from region base and size of VBT data. RVDA/RVDS are not
> >>> - * defined before opregion 2.0.
> >>> + * OpRegion and VBT:
> >>> + * When VBT data doesn't exceed 6KB, it's stored in Mailbox #4.
> >>> + * When VBT data exceeds 6KB size, Mailbox #4 is no longer large
> >>> enough
> >>> + * to hold the VBT data, the Extended VBT region is introduced since
> >>> + * OpRegion 2.0 to hold the VBT data. Since OpRegion 2.0, RVDA/RVDS
> >>> are
> >>> + * introduced to define the extended VBT data location and size.
> >>> + * OpRegion 2.0: RVDA defines the absolute physical address of the
> >>> + * extended VBT data, RVDS defines the VBT data size.
> >>> + * OpRegion 2.1 and above: RVDA defines the relative address of the
> >>> + * extended VBT data to OpRegion base, RVDS defines the VBT data
> >>> size.
> >>> *
> >>> - * opregion 2.1+: RVDA is unsigned, relative offset from
> >>> - * opregion base, and should point to the end of opregion.
> >>> - * otherwise, exposing to userspace to allow read access to
> >>> everything between
> >>> - * the OpRegion and VBT is not safe.
> >>> - * RVDS is defined as size in bytes.
> >>> - *
> >>> - * opregion 2.0: rvda is the physical VBT address.
> >>> - * Since rvda is HPA it cannot be directly used in guest.
> >>> - * And it should not be practically available for end user,so it is
> >>> not supported.
> >>> + * Due to the RVDA difference in OpRegion VBT (also the only diff
> >>> between
> >>> + * 2.0 and 2.1), while for OpRegion 2.1 and above it's possible to
> >>> map
> >>> + * a contigious memory to expose OpRegion and VBT r/w via the vfio
> >>> + * region, for OpRegion 2.0 shadow and amendment mechanism is used to
> >>> + * expose OpRegion and VBT r/w properly. So that from r/w ops view,
> >>> only
> >>> + * OpRegion 2.1 is exposed regardless underneath Region is 2.0 or
> >>> 2.1.
> >>> */
> >>> version = le16_to_cpu(*(__le16 *)(base + OPREGION_VERSION));
> >>> - if (version >= 0x0200) {
> >>> - u64 rvda;
> >>> - u32 rvds;
> >>>
> >>> + if (version >= 0x0200) {
> >>> rvda = le64_to_cpu(*(__le64 *)(base + OPREGION_RVDA));
> >>> rvds = le32_to_cpu(*(__le32 *)(base + OPREGION_RVDS));
> >>> +
> >>> + /* The extended VBT is valid only when RVDA/RVDS are
> >>> non-zero. */
> >>> if (rvda && rvds) {
> >>> - /* no support for opregion v2.0 with physical VBT
> >>> address */
> >>> - if (version == 0x0200) {
> >>> + size += rvds;
> >>> + }
> >>> +
> >>> + /* The extended VBT must follows OpRegion for OpRegion 2.1+
> >>> */
> >>> + if (rvda != size && version > 0x0200) {
> >>
> >> But we already added rvds to size, this is not compatible with the
> >> previous code that required rvda == size BEFORE adding rvds.
> >>
> >>> + memunmap(base);
> >>> + pci_err(vdev->pdev,
> >>> + "Extended VBT does not follow opregion on
> >>> version 0x%04x\n",
> >>> + version);
> >>> + return -EINVAL;
> >>> + }
> >>> + }
> >>> +
> >>> + if (size != OPREGION_SIZE) {
> >>> + /* Allocate memory for OpRegion and extended VBT for 2.0 */
> >>> + if (rvda && rvds && version == 0x0200) {
> >>> + void *vbt_base;
> >>> +
> >>> + vbt_base = memremap(rvda, rvds, MEMREMAP_WB);
> >>> + if (!vbt_base) {
> >>> memunmap(base);
> >>> - pci_err(vdev->pdev,
> >>> - "IGD assignment does not support
> >>> opregion v2.0 with an extended VBT region\n");
> >>> - return -EINVAL;
> >>> + return -ENOMEM;
> >>> }
> >>>
> >>> - if (rvda != size) {
> >>> + opregionvbt = kzalloc(size, GFP_KERNEL);
> >>> + if (!opregionvbt) {
> >>> memunmap(base);
> >>> - pci_err(vdev->pdev,
> >>> - "Extended VBT does not follow
> >>> opregion on version 0x%04x\n",
> >>> - version);
> >>> - return -EINVAL;
> >>> + memunmap(vbt_base);
> >>> + return -ENOMEM;
> >>> }
> >>>
> >>> - /* region size for opregion v2.0+: opregion and VBT
> >>> size. */
> >>> - size += rvds;
> >>> + /* Stitch VBT after OpRegion noncontigious */
> >>> + memcpy(opregionvbt, base, OPREGION_SIZE);
> >>> + memcpy(opregionvbt + OPREGION_SIZE, vbt_base, rvds);
> >>> +
> >>> + /* Patch OpRegion 2.0 to 2.1 */
> >>> + *(__le16 *)(opregionvbt + OPREGION_VERSION) = 0x0201;
> >>> + /* Patch RVDA to relative address after OpRegion */
> >>> + *(__le64 *)(opregionvbt + OPREGION_RVDA) =
> >>> OPREGION_SIZE;
> >>
> >> AIUI, the OpRegion is a two-way channel between the IGD device/system
> >> BIOS and the driver, numerous fields are writable by the driver. Now
> >> the driver writes to a shadow copy of the OpRegion table. What
> >> completes the write to the real OpRegion table for consumption by the
> >> device/BIOS? Likewise, what updates the fields that are written by the
> >> device/BIOS for consumption by the driver?
> >>
> >> If a shadow copy of the OpRegion detached from the physical table is
> >> sufficient here, why wouldn't we always shadow the OpRegion and prevent
> >> all userspace writes from touching the real version? Thanks,
> >>
> >> Alex
> >>
> >>> +
> >>> + memunmap(vbt_base);
> >>> + memunmap(base);
> >>> +
> >>> + /* Register shadow instead of map as vfio_region */
> >>> + base = opregionvbt;
> >>> + /* Remap OpRegion + extended VBT for 2.1+ */
> >>> + } else {
> >>> + memunmap(base);
> >>> + base = memremap(addr, size, MEMREMAP_WB);
> >>> + if (!base)
> >>> + return -ENOMEM;
> >>> }
> >>> }
> >>>
> >>> - if (size != OPREGION_SIZE) {
> >>> - memunmap(base);
> >>> - base = memremap(addr, size, MEMREMAP_WB);
> >>> - if (!base)
> >>> - return -ENOMEM;
> >>> - }
> >>> -
> >>> ret = vfio_pci_register_dev_region(vdev,
> >>> PCI_VENDOR_ID_INTEL | VFIO_REGION_TYPE_PCI_VENDOR_TYPE,
> >>> VFIO_REGION_SUBTYPE_INTEL_IGD_OPREGION,
> >>> &vfio_pci_igd_regops, size, VFIO_REGION_INFO_FLAG_READ, base);
> >>> if (ret) {
> >>> - memunmap(base);
> >>> + if (is_ioremap_addr(base))
> >>> + memunmap(base);
> >>> + else
> >>> + kfree(base);
> >>> return ret;
> >>> }
> >>>
> >>
> >>
> >
> > --
> > Best Regards,
> > Colin Xu
> >
> >
>
> --
> Best Regards,
> Colin Xu
>