Re: [PATCH 1/1] mm: Fix struct page layout on 32-bit systems

From: Jesper Dangaard Brouer
Date: Thu Apr 15 2021 - 14:08:48 EST


On Wed, 14 Apr 2021 21:56:39 +0000
David Laight <David.Laight@xxxxxxxxxx> wrote:

> From: Matthew Wilcox
> > Sent: 14 April 2021 22:36
> >
> > On Wed, Apr 14, 2021 at 09:13:22PM +0200, Jesper Dangaard Brouer wrote:
> > > (If others want to reproduce). First I could not reproduce on ARM32.
> > > Then I found out that enabling CONFIG_XEN on ARCH=arm was needed to
> > > cause the issue by enabling CONFIG_ARCH_DMA_ADDR_T_64BIT.
> >
> > hmmm ... you should be able to provoke it by enabling ARM_LPAE,
> > which selects PHYS_ADDR_T_64BIT, and
> >
> > config ARCH_DMA_ADDR_T_64BIT
> > def_bool 64BIT || PHYS_ADDR_T_64BIT
> >
> > > struct page {
> > > long unsigned int flags; /* 0 4 */
> > >
> > > /* XXX 4 bytes hole, try to pack */
> > >
> > > union {
> > > struct {
> > > struct list_head lru; /* 8 8 */
> > > struct address_space * mapping; /* 16 4 */
> > > long unsigned int index; /* 20 4 */
> > > long unsigned int private; /* 24 4 */
> > > }; /* 8 20 */
> > > struct {
> > > dma_addr_t dma_addr
>
> Adding __packed here will remove the 4 byte hole before the union
> and the compiler seems clever enough to know that anything following
> a 'long' must also be 'long' aligned.

Played with __packed in below patch, and I can confirm it seems to work.

> So you don't get anything horrid like byte accesses.
> On 64bit dma_addr will remain 64bit aligned.
> On arm32 dma_addr will be 32bit aligned - but forcing two 32bit access
> won't make any difference.

See below patch. Where I swap32 the dma address to satisfy
page->compound having bit zero cleared. (It is the simplest fix I could
come up with).


[PATCH] page_pool: handling 32-bit archs with 64-bit dma_addr_t

From: Jesper Dangaard Brouer <brouer@xxxxxxxxxx>

Workaround for storing 64-bit DMA-addr on 32-bit machines in struct
page. The page->dma_addr share area with page->compound_head which
use bit zero to mark compound pages. This is okay, as DMA-addr are
aligned pointers which have bit zero cleared.

In the 32-bit case, page->compound_head is 32-bit. Thus, when
dma_addr_t is 64-bit it will be located in top 32-bit. Solve by
swapping dma_addr 32-bit segments.

Signed-off-by: Jesper Dangaard Brouer <brouer@xxxxxxxxxx>
---
include/linux/mm_types.h | 2 +-
include/linux/types.h | 1 +
include/net/page_pool.h | 21 ++++++++++++++++++++-
net/core/page_pool.c | 8 +++++---
4 files changed, 27 insertions(+), 5 deletions(-)

diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h
index 6613b26a8894..27406e3b1e1b 100644
--- a/include/linux/mm_types.h
+++ b/include/linux/mm_types.h
@@ -100,7 +100,7 @@ struct page {
* @dma_addr: might require a 64-bit value even on
* 32-bit architectures.
*/
- dma_addr_t dma_addr;
+ dma_addr_t dma_addr __packed;
};
struct { /* slab, slob and slub */
union {
diff --git a/include/linux/types.h b/include/linux/types.h
index ac825ad90e44..65fd5d630016 100644
--- a/include/linux/types.h
+++ b/include/linux/types.h
@@ -141,6 +141,7 @@ typedef u64 blkcnt_t;
*/
#ifdef CONFIG_ARCH_DMA_ADDR_T_64BIT
typedef u64 dma_addr_t;
+//typedef u64 __attribute__((aligned(sizeof(void *)))) dma_addr_t;
#else
typedef u32 dma_addr_t;
#endif
diff --git a/include/net/page_pool.h b/include/net/page_pool.h
index b5b195305346..c2329088665c 100644
--- a/include/net/page_pool.h
+++ b/include/net/page_pool.h
@@ -196,9 +196,28 @@ static inline void page_pool_recycle_direct(struct page_pool *pool,
page_pool_put_full_page(pool, page, true);
}

+static inline
+dma_addr_t page_pool_dma_addr_read(dma_addr_t dma_addr)
+{
+ /* Workaround for storing 64-bit DMA-addr on 32-bit machines in struct
+ * page. The page->dma_addr share area with page->compound_head which
+ * use bit zero to mark compound pages. This is okay, as DMA-addr are
+ * aligned pointers which have bit zero cleared.
+ *
+ * In the 32-bit case, page->compound_head is 32-bit. Thus, when
+ * dma_addr_t is 64-bit it will be located in top 32-bit. Solve by
+ * swapping dma_addr 32-bit segments.
+ */
+#ifdef CONFIG_ARCH_DMA_ADDR_T_64BIT
+ if (sizeof(long unsigned int) == 4) /* 32-bit system */
+ dma_addr = (dma_addr << 32) | (dma_addr >> 32);
+#endif
+ return dma_addr;
+}
+
static inline dma_addr_t page_pool_get_dma_addr(struct page *page)
{
- return page->dma_addr;
+ return page_pool_dma_addr_read(page->dma_addr);
}

static inline bool is_page_pool_compiled_in(void)
diff --git a/net/core/page_pool.c b/net/core/page_pool.c
index ad8b0707af04..813598ea23f6 100644
--- a/net/core/page_pool.c
+++ b/net/core/page_pool.c
@@ -174,8 +174,10 @@ static void page_pool_dma_sync_for_device(struct page_pool *pool,
struct page *page,
unsigned int dma_sync_size)
{
+ dma_addr_t dma = page_pool_dma_addr_read(page->dma_addr);
+
dma_sync_size = min(dma_sync_size, pool->p.max_len);
- dma_sync_single_range_for_device(pool->p.dev, page->dma_addr,
+ dma_sync_single_range_for_device(pool->p.dev, dma,
pool->p.offset, dma_sync_size,
pool->p.dma_dir);
}
@@ -226,7 +228,7 @@ static struct page *__page_pool_alloc_pages_slow(struct page_pool *pool,
put_page(page);
return NULL;
}
- page->dma_addr = dma;
+ page->dma_addr = page_pool_dma_addr_read(dma);

if (pool->p.flags & PP_FLAG_DMA_SYNC_DEV)
page_pool_dma_sync_for_device(pool, page, pool->p.max_len);
@@ -294,7 +296,7 @@ void page_pool_release_page(struct page_pool *pool, struct page *page)
*/
goto skip_dma_unmap;

- dma = page->dma_addr;
+ dma = page_pool_dma_addr_read(page->dma_addr);

/* When page is unmapped, it cannot be returned our pool */
dma_unmap_page_attrs(pool->p.dev, dma,


--
Best regards,
Jesper Dangaard Brouer
MSc.CS, Principal Kernel Engineer at Red Hat
LinkedIn: http://www.linkedin.com/in/brouer