Re: [PATCH] mm: limit THP alignment – performance gain observed in AI inference workloads

From: siddhartha
Date: Mon Jun 30 2025 - 07:48:21 EST


On 2025-06-30 16:24, Lorenzo Stoakes wrote:
+cc Vlastimil, please keep him cc'd on discussions here as the author of this
fix in the conversation.

On Mon, Jun 30, 2025 at 10:55:52AM +0530, Dev Jain wrote:


For this workload, do you enable mTHPs on your system? My plan is to make a
similar patch for

the mTHP case and I'd be grateful if you can get me some results : )

I'd urge caution here.

The reason there was a big perf improvement is that, for certain workloads, the
original patch by Rik caused issues with VMA fragmentation. So rather than
getting adjacent VMAs that might later be khugepage'd, you'd get a bunch of VMAs
that were auto-aligned and thus fragmented from one another.

So while you got speed ups on some workloads, you got really bad perf impact on
some that were subject to this.

The observed speed up was on a very specific benchmark also. While it's a great
improvement, it's important to understand the context (see the original patch
for details [0]).

I do think it's worth considering changing thp_get_unmapped_area_vmflags() for
mTHP, as it's currently very limited (just PMD alignment) and it'd possibly be
sensible to change this to checking against allowed THP alignments, but I'd not
assume this is going to get some crazy speed up as observed here.

Note that any such change would probably require some refactoring in THP first
to make it not quite so awful.

I also think for Siddharta's usecase mTHP isn't really relevant is it, as intel
do not support mTHP currently do they?

Regards, Lorenzo

[0]: https://lore.kernel.org/all/20241024151228.101841-2-vbabka@xxxxxxx/T/#u

Hi Lorenzo, Dev, All,

Thank you for the thoughtful responses and for engaging with the performance implications of the patch.

You're absolutely right that the observed speedup came from a specific class of workloads — in this case, token-length-variable AI inference pipelines based on Hugging Face Transformers and ONNX Runtime. These workloads trigger highly dynamic, anonymous memory allocation patterns, often in bursts aligned with model shard loading and attention map resizing. In such cases, VMA fragmentation due to PMD-aligned, non-PMD-sized mappings led to near-complete loss of THP utilization.

Once the alignment restriction was lifted (via Rik’s patch), we observed substantial restoration of THP behavior, which is where the performance gains came from. That said, I completely agree that:

Not all workloads benefit from this

Some may even regress if the underlying VMAs aren't THP-coalescible for other reasons

Still, for high-throughput inference workloads on modern Intel CPUs, this behavior isn’t a corner case. The shift toward multi-model concurrent serving (e.g., LLM-as-a-Service) means this dynamic allocation pattern is becoming common, especially in edge/latency-sensitive deployments.

🧠 On mTHP: Intel Does Support It
Regarding mTHP — yes, Intel platforms (especially server-grade Xeon processors from Cascade Lake onward) do support mapped transparent huge pages, including via:

tmpfs-backed files

madvise(MADV_HUGEPAGE) on file mappings

shmem usage with shmem_enabled in the kernel

So I’d say mTHP is certainly relevant for workloads where model weights or tensors are pre-cached or memory-mapped — a pattern we’re also seeing as Hugging Face, ONNX, and PyTorch ecosystems move toward zero-copy tensor sharing.

Given that, I'd absolutely be interested in testing any mTHP-targeted patch — and I’d be happy to help validate it, especially if it avoids the VMA fragmentation pitfall you rightly pointed out.

Thanks again for the detailed feedback, and I’ll try to replicate and share further traces (from my local testbed) since I currently don’t have access to the original Intel Developer Cloud logs.

Best regards,
Siddhartha Sharma