Re: [PATCH v3 1/4] iommu: Add gfp parameter to iommu_ops::map

From: Tom Murphy
Date: Tue Jun 04 2019 - 16:59:57 EST


On Tue, Jun 4, 2019 at 7:11 PM Robin Murphy <robin.murphy@xxxxxxx> wrote:
>
> On 06/05/2019 19:52, Tom Murphy wrote:
> > Add a gfp_t parameter to the iommu_ops::map function.
> > Remove the needless locking in the AMD iommu driver.
> >
> > The iommu_ops::map function (or the iommu_map function which calls it)
> > was always supposed to be sleepable (according to Joerg's comment in
> > this thread: https://lore.kernel.org/patchwork/patch/977520/ ) and so
> > should probably have had a "might_sleep()" since it was written. However
> > currently the dma-iommu api can call iommu_map in an atomic context,
> > which it shouldn't do. This doesn't cause any problems because any iommu
> > driver which uses the dma-iommu api uses gfp_atomic in it's
> > iommu_ops::map function. But doing this wastes the memory allocators
> > atomic pools.
>
> Hmm, in some cases iommu_ops::unmap may need to allocate as well,
> primarily if it needs to split a hugepage mapping. Should we pass flags

Are you sure that's the case?

I don't see that in the amd or intel iommu_ops::unmap code and from
looking at the intel code it seems like we actually unmap the whole
large page:
/* Cope with horrid API which requires us to unmap more than the
size argument if it happens to be a large-page mapping. */
BUG_ON(!pfn_to_dma_pte(dmar_domain, iova >> VTD_PAGE_SHIFT, &level));

if (size < VTD_PAGE_SIZE << level_to_offset_bits(level))
size = VTD_PAGE_SIZE << level_to_offset_bits(level);



> through there as well, or are we prepared to assume that that case will
> happen rarely enough that it's fair to just assume GFP_ATOMIC? It won't
> happen for DMA ops, but it's conceivable that other users such as GPU
> drivers might make partial unmaps, and I doubt we could totally rule out
> the wackiest ones doing so from non-sleeping contexts.
>
> Robin.
>
> > We can remove the mutex lock from amd_iommu_map and amd_iommu_unmap.
> > iommu_map doesnât lock while mapping and so no two calls should touch
> > the same iova range. The AMD driver already handles the page table page
> > allocations without locks so we can safely remove the locks.
> >
> > Signed-off-by: Tom Murphy <tmurphy@xxxxxxxxxx>
> > ---
> > drivers/iommu/amd_iommu.c | 14 ++++-------
> > drivers/iommu/arm-smmu-v3.c | 2 +-
> > drivers/iommu/arm-smmu.c | 2 +-
> > drivers/iommu/dma-iommu.c | 6 ++---
> > drivers/iommu/exynos-iommu.c | 2 +-
> > drivers/iommu/intel-iommu.c | 2 +-
> > drivers/iommu/iommu.c | 43 +++++++++++++++++++++++++++++-----
> > drivers/iommu/ipmmu-vmsa.c | 2 +-
> > drivers/iommu/msm_iommu.c | 2 +-
> > drivers/iommu/mtk_iommu.c | 2 +-
> > drivers/iommu/mtk_iommu_v1.c | 2 +-
> > drivers/iommu/omap-iommu.c | 2 +-
> > drivers/iommu/qcom_iommu.c | 2 +-
> > drivers/iommu/rockchip-iommu.c | 2 +-
> > drivers/iommu/s390-iommu.c | 2 +-
> > drivers/iommu/tegra-gart.c | 2 +-
> > drivers/iommu/tegra-smmu.c | 2 +-
> > include/linux/iommu.h | 21 ++++++++++++++++-
> > 18 files changed, 78 insertions(+), 34 deletions(-)
> >
> > diff --git a/drivers/iommu/amd_iommu.c b/drivers/iommu/amd_iommu.c
> > index ebd062522cf5..ea3a5dc61bb0 100644
> > --- a/drivers/iommu/amd_iommu.c
> > +++ b/drivers/iommu/amd_iommu.c
> > @@ -3092,7 +3092,8 @@ static int amd_iommu_attach_device(struct iommu_domain *dom,
> > }
> >
> > static int amd_iommu_map(struct iommu_domain *dom, unsigned long iova,
> > - phys_addr_t paddr, size_t page_size, int iommu_prot)
> > + phys_addr_t paddr, size_t page_size, int iommu_prot,
> > + gfp_t gfp)
> > {
> > struct protection_domain *domain = to_pdomain(dom);
> > int prot = 0;
> > @@ -3106,9 +3107,7 @@ static int amd_iommu_map(struct iommu_domain *dom, unsigned long iova,
> > if (iommu_prot & IOMMU_WRITE)
> > prot |= IOMMU_PROT_IW;
> >
> > - mutex_lock(&domain->api_lock);
> > - ret = iommu_map_page(domain, iova, paddr, page_size, prot, GFP_KERNEL);
> > - mutex_unlock(&domain->api_lock);
> > + ret = iommu_map_page(domain, iova, paddr, page_size, prot, gfp);
> >
> > domain_flush_np_cache(domain, iova, page_size);
> >
> > @@ -3119,16 +3118,11 @@ static size_t amd_iommu_unmap(struct iommu_domain *dom, unsigned long iova,
> > size_t page_size)
> > {
> > struct protection_domain *domain = to_pdomain(dom);
> > - size_t unmap_size;
> >
> > if (domain->mode == PAGE_MODE_NONE)
> > return 0;
> >
> > - mutex_lock(&domain->api_lock);
> > - unmap_size = iommu_unmap_page(domain, iova, page_size);
> > - mutex_unlock(&domain->api_lock);
> > -
> > - return unmap_size;
> > + return iommu_unmap_page(domain, iova, page_size);
> > }
> >
> > static phys_addr_t amd_iommu_iova_to_phys(struct iommu_domain *dom,
> > diff --git a/drivers/iommu/arm-smmu-v3.c b/drivers/iommu/arm-smmu-v3.c
> > index d3880010c6cf..e5c48089b49f 100644
> > --- a/drivers/iommu/arm-smmu-v3.c
> > +++ b/drivers/iommu/arm-smmu-v3.c
> > @@ -1777,7 +1777,7 @@ static int arm_smmu_attach_dev(struct iommu_domain *domain, struct device *dev)
> > }
> >
> > static int arm_smmu_map(struct iommu_domain *domain, unsigned long iova,
> > - phys_addr_t paddr, size_t size, int prot)
> > + phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
> > {
> > struct io_pgtable_ops *ops = to_smmu_domain(domain)->pgtbl_ops;
> >
> > diff --git a/drivers/iommu/arm-smmu.c b/drivers/iommu/arm-smmu.c
> > index 045d93884164..2d50db55b788 100644
> > --- a/drivers/iommu/arm-smmu.c
> > +++ b/drivers/iommu/arm-smmu.c
> > @@ -1286,7 +1286,7 @@ static int arm_smmu_attach_dev(struct iommu_domain *domain, struct device *dev)
> > }
> >
> > static int arm_smmu_map(struct iommu_domain *domain, unsigned long iova,
> > - phys_addr_t paddr, size_t size, int prot)
> > + phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
> > {
> > struct io_pgtable_ops *ops = to_smmu_domain(domain)->pgtbl_ops;
> > struct arm_smmu_device *smmu = to_smmu_domain(domain)->smmu;
> > diff --git a/drivers/iommu/dma-iommu.c b/drivers/iommu/dma-iommu.c
> > index fa5713a4f6f8..7a96c2c8f56b 100644
> > --- a/drivers/iommu/dma-iommu.c
> > +++ b/drivers/iommu/dma-iommu.c
> > @@ -440,7 +440,7 @@ static dma_addr_t __iommu_dma_map(struct device *dev, phys_addr_t phys,
> > if (!iova)
> > return DMA_MAPPING_ERROR;
> >
> > - if (iommu_map(domain, iova, phys - iova_off, size, prot)) {
> > + if (iommu_map_atomic(domain, iova, phys - iova_off, size, prot)) {
> > iommu_dma_free_iova(cookie, iova, size);
> > return DMA_MAPPING_ERROR;
> > }
> > @@ -641,7 +641,7 @@ static void *iommu_dma_alloc_remap(struct device *dev, size_t size,
> > arch_dma_prep_coherent(sg_page(sg), sg->length);
> > }
> >
> > - if (iommu_map_sg(domain, iova, sgt.sgl, sgt.orig_nents, ioprot)
> > + if (iommu_map_sg_atomic(domain, iova, sgt.sgl, sgt.orig_nents, ioprot)
> > < size)
> > goto out_free_sg;
> >
> > @@ -1003,7 +1003,7 @@ static int iommu_dma_map_sg(struct device *dev, struct scatterlist *sg,
> > * We'll leave any physical concatenation to the IOMMU driver's
> > * implementation - it knows better than we do.
> > */
> > - if (iommu_map_sg(domain, iova, sg, nents, prot) < iova_len)
> > + if (iommu_map_sg_atomic(domain, iova, sg, nents, prot) < iova_len)
> > goto out_free_iova;
> >
> > return __finalise_sg(dev, sg, nents, iova);
> > diff --git a/drivers/iommu/exynos-iommu.c b/drivers/iommu/exynos-iommu.c
> > index 05c6bc099d62..46414234c179 100644
> > --- a/drivers/iommu/exynos-iommu.c
> > +++ b/drivers/iommu/exynos-iommu.c
> > @@ -1078,7 +1078,7 @@ static int lv2set_page(sysmmu_pte_t *pent, phys_addr_t paddr, size_t size,
> > */
> > static int exynos_iommu_map(struct iommu_domain *iommu_domain,
> > unsigned long l_iova, phys_addr_t paddr, size_t size,
> > - int prot)
> > + int prot, gfp_t gfp)
> > {
> > struct exynos_iommu_domain *domain = to_exynos_domain(iommu_domain);
> > sysmmu_pte_t *entry;
> > diff --git a/drivers/iommu/intel-iommu.c b/drivers/iommu/intel-iommu.c
> > index 28cb713d728c..4f0ff28f7cb9 100644
> > --- a/drivers/iommu/intel-iommu.c
> > +++ b/drivers/iommu/intel-iommu.c
> > @@ -5137,7 +5137,7 @@ static void intel_iommu_detach_device(struct iommu_domain *domain,
> >
> > static int intel_iommu_map(struct iommu_domain *domain,
> > unsigned long iova, phys_addr_t hpa,
> > - size_t size, int iommu_prot)
> > + size_t size, int iommu_prot, gfp_t gfp)
> > {
> > struct dmar_domain *dmar_domain = to_dmar_domain(domain);
> > u64 max_addr;
> > diff --git a/drivers/iommu/iommu.c b/drivers/iommu/iommu.c
> > index 109de67d5d72..1b49841c177e 100644
> > --- a/drivers/iommu/iommu.c
> > +++ b/drivers/iommu/iommu.c
> > @@ -1584,8 +1584,8 @@ static size_t iommu_pgsize(struct iommu_domain *domain,
> > return pgsize;
> > }
> >
> > -int iommu_map(struct iommu_domain *domain, unsigned long iova,
> > - phys_addr_t paddr, size_t size, int prot)
> > +int __iommu_map(struct iommu_domain *domain, unsigned long iova,
> > + phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
> > {
> > const struct iommu_ops *ops = domain->ops;
> > unsigned long orig_iova = iova;
> > @@ -1622,8 +1622,8 @@ int iommu_map(struct iommu_domain *domain, unsigned long iova,
> >
> > pr_debug("mapping: iova 0x%lx pa %pa pgsize 0x%zx\n",
> > iova, &paddr, pgsize);
> > + ret = ops->map(domain, iova, paddr, pgsize, prot, gfp);
> >
> > - ret = ops->map(domain, iova, paddr, pgsize, prot);
> > if (ret)
> > break;
> >
> > @@ -1643,8 +1643,22 @@ int iommu_map(struct iommu_domain *domain, unsigned long iova,
> >
> > return ret;
> > }
> > +
> > +int iommu_map(struct iommu_domain *domain, unsigned long iova,
> > + phys_addr_t paddr, size_t size, int prot)
> > +{
> > + might_sleep();
> > + return __iommu_map(domain, iova, paddr, size, prot, GFP_KERNEL);
> > +}
> > EXPORT_SYMBOL_GPL(iommu_map);
> >
> > +int iommu_map_atomic(struct iommu_domain *domain, unsigned long iova,
> > + phys_addr_t paddr, size_t size, int prot)
> > +{
> > + return __iommu_map(domain, iova, paddr, size, prot, GFP_ATOMIC);
> > +}
> > +EXPORT_SYMBOL_GPL(iommu_map_atomic);
> > +
> > static size_t __iommu_unmap(struct iommu_domain *domain,
> > unsigned long iova, size_t size,
> > bool sync)
> > @@ -1719,8 +1733,9 @@ size_t iommu_unmap_fast(struct iommu_domain *domain,
> > }
> > EXPORT_SYMBOL_GPL(iommu_unmap_fast);
> >
> > -size_t iommu_map_sg(struct iommu_domain *domain, unsigned long iova,
> > - struct scatterlist *sg, unsigned int nents, int prot)
> > +size_t __iommu_map_sg(struct iommu_domain *domain, unsigned long iova,
> > + struct scatterlist *sg, unsigned int nents, int prot,
> > + gfp_t gfp)
> > {
> > size_t len = 0, mapped = 0;
> > phys_addr_t start;
> > @@ -1731,7 +1746,9 @@ size_t iommu_map_sg(struct iommu_domain *domain, unsigned long iova,
> > phys_addr_t s_phys = sg_phys(sg);
> >
> > if (len && s_phys != start + len) {
> > - ret = iommu_map(domain, iova + mapped, start, len, prot);
> > + ret = __iommu_map(domain, iova + mapped, start,
> > + len, prot, gfp);
> > +
> > if (ret)
> > goto out_err;
> >
> > @@ -1759,8 +1776,22 @@ size_t iommu_map_sg(struct iommu_domain *domain, unsigned long iova,
> > return 0;
> >
> > }
> > +
> > +size_t iommu_map_sg(struct iommu_domain *domain, unsigned long iova,
> > + struct scatterlist *sg, unsigned int nents, int prot)
> > +{
> > + might_sleep();
> > + return __iommu_map_sg(domain, iova, sg, nents, prot, GFP_KERNEL);
> > +}
> > EXPORT_SYMBOL_GPL(iommu_map_sg);
> >
> > +size_t iommu_map_sg_atomic(struct iommu_domain *domain, unsigned long iova,
> > + struct scatterlist *sg, unsigned int nents, int prot)
> > +{
> > + return __iommu_map_sg(domain, iova, sg, nents, prot, GFP_ATOMIC);
> > +}
> > +EXPORT_SYMBOL_GPL(iommu_map_sg_atomic);
> > +
> > int iommu_domain_window_enable(struct iommu_domain *domain, u32 wnd_nr,
> > phys_addr_t paddr, u64 size, int prot)
> > {
> > diff --git a/drivers/iommu/ipmmu-vmsa.c b/drivers/iommu/ipmmu-vmsa.c
> > index 9a380c10655e..e005c83d49d8 100644
> > --- a/drivers/iommu/ipmmu-vmsa.c
> > +++ b/drivers/iommu/ipmmu-vmsa.c
> > @@ -707,7 +707,7 @@ static void ipmmu_detach_device(struct iommu_domain *io_domain,
> > }
> >
> > static int ipmmu_map(struct iommu_domain *io_domain, unsigned long iova,
> > - phys_addr_t paddr, size_t size, int prot)
> > + phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
> > {
> > struct ipmmu_vmsa_domain *domain = to_vmsa_domain(io_domain);
> >
> > diff --git a/drivers/iommu/msm_iommu.c b/drivers/iommu/msm_iommu.c
> > index 9fb0eb7a4d02..3f6bf3653aa2 100644
> > --- a/drivers/iommu/msm_iommu.c
> > +++ b/drivers/iommu/msm_iommu.c
> > @@ -508,7 +508,7 @@ static void msm_iommu_detach_dev(struct iommu_domain *domain,
> > }
> >
> > static int msm_iommu_map(struct iommu_domain *domain, unsigned long iova,
> > - phys_addr_t pa, size_t len, int prot)
> > + phys_addr_t pa, size_t len, int prot, gfp_t gfp)
> > {
> > struct msm_priv *priv = to_msm_priv(domain);
> > unsigned long flags;
> > diff --git a/drivers/iommu/mtk_iommu.c b/drivers/iommu/mtk_iommu.c
> > index de3e02277b70..3176b9b54d4d 100644
> > --- a/drivers/iommu/mtk_iommu.c
> > +++ b/drivers/iommu/mtk_iommu.c
> > @@ -364,7 +364,7 @@ static void mtk_iommu_detach_device(struct iommu_domain *domain,
> > }
> >
> > static int mtk_iommu_map(struct iommu_domain *domain, unsigned long iova,
> > - phys_addr_t paddr, size_t size, int prot)
> > + phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
> > {
> > struct mtk_iommu_domain *dom = to_mtk_domain(domain);
> > unsigned long flags;
> > diff --git a/drivers/iommu/mtk_iommu_v1.c b/drivers/iommu/mtk_iommu_v1.c
> > index 52b01e3a49df..e7b1907faec1 100644
> > --- a/drivers/iommu/mtk_iommu_v1.c
> > +++ b/drivers/iommu/mtk_iommu_v1.c
> > @@ -303,7 +303,7 @@ static void mtk_iommu_detach_device(struct iommu_domain *domain,
> > }
> >
> > static int mtk_iommu_map(struct iommu_domain *domain, unsigned long iova,
> > - phys_addr_t paddr, size_t size, int prot)
> > + phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
> > {
> > struct mtk_iommu_domain *dom = to_mtk_domain(domain);
> > unsigned int page_num = size >> MT2701_IOMMU_PAGE_SHIFT;
> > diff --git a/drivers/iommu/omap-iommu.c b/drivers/iommu/omap-iommu.c
> > index d2fb347aa4ff..c1d5a71285dc 100644
> > --- a/drivers/iommu/omap-iommu.c
> > +++ b/drivers/iommu/omap-iommu.c
> > @@ -1109,7 +1109,7 @@ static u32 iotlb_init_entry(struct iotlb_entry *e, u32 da, u32 pa, int pgsz)
> > }
> >
> > static int omap_iommu_map(struct iommu_domain *domain, unsigned long da,
> > - phys_addr_t pa, size_t bytes, int prot)
> > + phys_addr_t pa, size_t bytes, int prot, gfp_t gfp)
> > {
> > struct omap_iommu_domain *omap_domain = to_omap_domain(domain);
> > struct device *dev = omap_domain->dev;
> > diff --git a/drivers/iommu/qcom_iommu.c b/drivers/iommu/qcom_iommu.c
> > index 8cdd3f059513..a01e07a4e76f 100644
> > --- a/drivers/iommu/qcom_iommu.c
> > +++ b/drivers/iommu/qcom_iommu.c
> > @@ -411,7 +411,7 @@ static void qcom_iommu_detach_dev(struct iommu_domain *domain, struct device *de
> > }
> >
> > static int qcom_iommu_map(struct iommu_domain *domain, unsigned long iova,
> > - phys_addr_t paddr, size_t size, int prot)
> > + phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
> > {
> > int ret;
> > unsigned long flags;
> > diff --git a/drivers/iommu/rockchip-iommu.c b/drivers/iommu/rockchip-iommu.c
> > index 77d4bd93fe4b..aa3507f35107 100644
> > --- a/drivers/iommu/rockchip-iommu.c
> > +++ b/drivers/iommu/rockchip-iommu.c
> > @@ -760,7 +760,7 @@ static int rk_iommu_map_iova(struct rk_iommu_domain *rk_domain, u32 *pte_addr,
> > }
> >
> > static int rk_iommu_map(struct iommu_domain *domain, unsigned long _iova,
> > - phys_addr_t paddr, size_t size, int prot)
> > + phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
> > {
> > struct rk_iommu_domain *rk_domain = to_rk_domain(domain);
> > unsigned long flags;
> > diff --git a/drivers/iommu/s390-iommu.c b/drivers/iommu/s390-iommu.c
> > index 22d4db302c1c..efa6aa68521d 100644
> > --- a/drivers/iommu/s390-iommu.c
> > +++ b/drivers/iommu/s390-iommu.c
> > @@ -265,7 +265,7 @@ static int s390_iommu_update_trans(struct s390_domain *s390_domain,
> > }
> >
> > static int s390_iommu_map(struct iommu_domain *domain, unsigned long iova,
> > - phys_addr_t paddr, size_t size, int prot)
> > + phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
> > {
> > struct s390_domain *s390_domain = to_s390_domain(domain);
> > int flags = ZPCI_PTE_VALID, rc = 0;
> > diff --git a/drivers/iommu/tegra-gart.c b/drivers/iommu/tegra-gart.c
> > index 4d8057916552..f300099852b1 100644
> > --- a/drivers/iommu/tegra-gart.c
> > +++ b/drivers/iommu/tegra-gart.c
> > @@ -190,7 +190,7 @@ static inline int __gart_iommu_map(struct gart_device *gart, unsigned long iova,
> > }
> >
> > static int gart_iommu_map(struct iommu_domain *domain, unsigned long iova,
> > - phys_addr_t pa, size_t bytes, int prot)
> > + phys_addr_t pa, size_t bytes, int prot, gfp_t gfp)
> > {
> > struct gart_device *gart = gart_handle;
> > int ret;
> > diff --git a/drivers/iommu/tegra-smmu.c b/drivers/iommu/tegra-smmu.c
> > index 5182c7d6171e..e1bf867e0607 100644
> > --- a/drivers/iommu/tegra-smmu.c
> > +++ b/drivers/iommu/tegra-smmu.c
> > @@ -641,7 +641,7 @@ static void tegra_smmu_set_pte(struct tegra_smmu_as *as, unsigned long iova,
> > }
> >
> > static int tegra_smmu_map(struct iommu_domain *domain, unsigned long iova,
> > - phys_addr_t paddr, size_t size, int prot)
> > + phys_addr_t paddr, size_t size, int prot, gfp_t gfp)
> > {
> > struct tegra_smmu_as *as = to_smmu_as(domain);
> > dma_addr_t pte_dma;
> > diff --git a/include/linux/iommu.h b/include/linux/iommu.h
> > index ffbbc7e39cee..76b8e7fe3ed0 100644
> > --- a/include/linux/iommu.h
> > +++ b/include/linux/iommu.h
> > @@ -198,7 +198,7 @@ struct iommu_ops {
> > int (*attach_dev)(struct iommu_domain *domain, struct device *dev);
> > void (*detach_dev)(struct iommu_domain *domain, struct device *dev);
> > int (*map)(struct iommu_domain *domain, unsigned long iova,
> > - phys_addr_t paddr, size_t size, int prot);
> > + phys_addr_t paddr, size_t size, int prot, gfp_t gfp);
> > size_t (*unmap)(struct iommu_domain *domain, unsigned long iova,
> > size_t size);
> > void (*flush_iotlb_all)(struct iommu_domain *domain);
> > @@ -295,12 +295,17 @@ extern struct iommu_domain *iommu_get_domain_for_dev(struct device *dev);
> > extern struct iommu_domain *iommu_get_dma_domain(struct device *dev);
> > extern int iommu_map(struct iommu_domain *domain, unsigned long iova,
> > phys_addr_t paddr, size_t size, int prot);
> > +extern int iommu_map_atomic(struct iommu_domain *domain, unsigned long iova,
> > + phys_addr_t paddr, size_t size, int prot);
> > extern size_t iommu_unmap(struct iommu_domain *domain, unsigned long iova,
> > size_t size);
> > extern size_t iommu_unmap_fast(struct iommu_domain *domain,
> > unsigned long iova, size_t size);
> > extern size_t iommu_map_sg(struct iommu_domain *domain, unsigned long iova,
> > struct scatterlist *sg,unsigned int nents, int prot);
> > +extern size_t iommu_map_sg_atomic(struct iommu_domain *domain,
> > + unsigned long iova, struct scatterlist *sg,
> > + unsigned int nents, int prot);
> > extern phys_addr_t iommu_iova_to_phys(struct iommu_domain *domain, dma_addr_t iova);
> > extern void iommu_set_fault_handler(struct iommu_domain *domain,
> > iommu_fault_handler_t handler, void *token);
> > @@ -469,6 +474,13 @@ static inline int iommu_map(struct iommu_domain *domain, unsigned long iova,
> > return -ENODEV;
> > }
> >
> > +static inline int iommu_map_atomic(struct iommu_domain *domain,
> > + unsigned long iova, phys_addr_t paddr,
> > + size_t size, int prot)
> > +{
> > + return -ENODEV;
> > +}
> > +
> > static inline size_t iommu_unmap(struct iommu_domain *domain,
> > unsigned long iova, size_t size)
> > {
> > @@ -488,6 +500,13 @@ static inline size_t iommu_map_sg(struct iommu_domain *domain,
> > return 0;
> > }
> >
> > +static inline size_t iommu_map_sg_atomic(struct iommu_domain *domain,
> > + unsigned long iova, struct scatterlist *sg,
> > + unsigned int nents, int prot)
> > +{
> > + return 0;
> > +}
> > +
> > static inline void iommu_flush_tlb_all(struct iommu_domain *domain)
> > {
> > }
> >