Re: [PATCH] mm/uffd: UFFD_FEATURE_WP_ZEROPAGE

From: Peter Xu
Date: Fri Feb 17 2023 - 18:05:40 EST


On Fri, Feb 17, 2023 at 12:41:40PM +0100, David Hildenbrand wrote:
> On 16.02.23 21:08, Peter Xu wrote:
> > On Thu, Feb 16, 2023 at 07:23:17PM +0100, David Hildenbrand wrote:
> > > On 16.02.23 18:55, Peter Xu wrote:
> > > > On Thu, Feb 16, 2023 at 06:00:51PM +0100, David Hildenbrand wrote:
> > > > > > >
> > > > > > > There are various reasons why I think a UFFD_FEATURE_WP_UNPOPULATED, using
> > > > > > > PTE markers, would be more benficial:
> > > > > > >
> > > > > > > 1) It would be applicable to anon hugetlb
> > > > > >
> > > > > > Anon hugetlb should already work with non ptes with the markers?
> > > > > >
> > > > >
> > > > > ... really? I thought we'd do the whole pte marker handling only when
> > > > > dealing with hugetlb/shmem. Interesting, thanks. (we could skip population
> > > > > in QEMU in that case as well -- we always do it for now)
> > > >
> > > > Hmm, you're talking about "anon hugetlb", so it's still hugetlb, right? :)
> > >
> > > I mean especially MAP_PRIVATE|MAP_HUGETLB|MAP_ANONYMOUS, so "in theory"
> > > without any fd and thus pagecache. ... but anon hugetlb keeps confusing me
> > > with pagecache handling.
> >
> > IIUC when mmap(fd==-1) it's the same as MAP_PRIVATE|MAP_HUGETLB.
>
> Let me rephrase my original statement: I thought we'd do the whole pte
> marker handling only when dealing with MAP_SHARED hugetlb ("hugetlb shared
> memory (shmem)"). So not on MAP_PRIVATE hugetlb where we might have anon
> hugetlb pages.

Okay, then we're good - I believe private was handled there.

>
>
> > > The focus of that paper is on CoW latency yes (and deduplication
> > > instantiating shared zeropages -- but building a covert channel using CoW
> > > latency might be rather tricky I think, because they will get deduplciated
> > > independently of a sender action ...).
> > >
> > > However, in theory, one could build a covert channel between two VMs simply
> > > by using cache flushes and reading from the shared zeropage. Measuring
> > > access time can reveal if the sender read the page (L3 filled) or not (L3
> > > not filled).
> >
> > So the attacker will know when someone reads a zeropage, but I still don't
> > get how that can leads to data leak..
>
> Oh, a covert channel is not for leaking data, but for communicating via an
> unofficial channel (bypassing firewalls etc). I also don't think one can
> really leak data with the shared zeropage ... unless that data is all zero
> :)

I see.

>
> > >
> > > Having that said, I don't think that we are going to disable the shared
> > > zeropage because of that for some workloads, I assume in most cases it will
> > > simply be way too noisy to transmit any kind of data and we have more
> > > critical covert channels to sort out if we want to.
> > >
> > > Just wanted to raise it because you asked :)
> > >
> > > >
> > > > Another note for s390: when it comes we can consider moving to pte markers
> > > > conditionally when !zeropage. But we can leave that for later.
> > >
> > > Sure, we could always have another feature flag.
> >
> > I think that doesn't need to be another feature flag. If someone will port
> > uffd-wp to s390 we can implement pte markers for WP_ZEROPAGE, then we
> > either use it when zeropage not exist, or we can switch to pte markers
> > completely too without changing the interface if we want, depending on
> > whether we think replacing zeropages with pte markers will be a major issue
> > with existing apps. I don't worry too much on that part.
>
> Then maybe the feature name/description should be more generic, such that
> it's merely an implementation detail that could change?
>
> "
> UFFD_FEATURE_WP_UNPOPULATED: for anonymous memory, if PTEs/PMDs are still
> unpopulated (no page mapped), uffd-wp protection to work will not require a
> previous manual population (e.g., using MADV_POPULATE_READ). The kernel
> might or might not populate the shared zeropage for that purpose. So after a
> uffd-wp protection with UFFD_FEATURE_WP_UNPOPULATED enabled, the PTEs/PMDs
> might or might not be populated.
> "

I would really hope we don't mention "manual population" in the man page or
header comments for app developers.

That's "how to workaound some issue", I'll read it as "with flag XXX you
don't need to work around (issue YYY) by doing ZZZ".

It's much more straightforward to state what the flag will do. The
workaround can be documented somewhere but that shouldn't be in the general
description of the feature.

I don't have a strong opinion on the name, though. I guess you would like
the new feature to avoid using the term zeropage; I'm fine with changing it.

>
> For example, it has to be clear that when doing an uffd-wp protect +
> unprotect, that there could be suddenly zeropages mapped (observable via
> uffd-missing later, /proc/pagemap).
>
> I'd be fine with something like that.
>
> [...]
>
> > > > > Using PTE markers would provide a real advantage IMHO for some users (IMHO
> > > > > background snapshots), where we might want to avoid populating
> > > > > zeropages/page tables as best as we can completely if the VM memory is
> > > > > mostly untouched.
> > > > >
> > > > > Naturally, I wonder if UFFD_FEATURE_WP_ZEROPAGE is really worth it. Is there
> > > > > is another good reason to combine the populate zeropage+wp that I am missing
> > > > > (e.g., atomicity by doing both in one operation)?
> > > >
> > > > It also makes the new WP_ASYNC and pagemap interface clean: we don't want
> > > > to have user pre-fault it every time too as a common tactic.. It's hard to
> > > > use, and the user doesn't need to know the internals of why it is needed,
> > > > either.
> > >
> > > I feel like we're building a lot of infrastructure on uffd-wp instead of
> > > having an alternative softdirty mode (using a world switch?) that works as
> > > expected and doesn't require that many uffd-wp extensions. ;)
> >
> > We used to discuss this WP_ZEROPAGE before, and I thought we were all happy
> > to have that. Obviously you changed your mind. :)
>
> As I said, I'd be sold on the PTE marker idea, or designing the feature more
> generic that it is an implementation detail. :)
>
> The whole "let's place zeropages" is rather a hack and IMHO suboptimal.
>
> For example, when uffd-wp unprotecting again (e.g., after a VM snapshot),
> we'd be left with zeropages mapped. Getting rid of them (for example, to
> reclaim page tables), will require TLB flushes, mmu notifiers, because we
> had present PTEs. In comparison, the nice thing about a PTE marker is that
> it is !present and you can just rip it out.

MMU notifiers do not make a difference, afaict, because that's
unconditional and irrelevant of what's changed underneath.

For TLB flush - I'm not sure how large will there be a difference
(e.g. considering tlb batching being there), but it's a very valid point.

>
> Similarly, with zeropages (as in your current patch), getting a THP later
> allocated requires going through khugepaged. In comparison, a PMD marker
> could more easily avoid that. The huge zeropage can work around that, but
> you'd still need an MADV_DONTNEED on the hole huge zeropage first to remove
> it, in order to replace it with a "real" THP.

IMHO relying on khugepaged is fine.

Note again that the whole wr-protect idea so far is always talking on
PAGE_SIZE. Any write happened in whatever PMD/PUD marker will go into:

- A pte marker population storm, which happens _inside_ a page fault of
the workload thread (rather than the monitor thread),

- The bigger the pte marker (PUD > PMD > PTE), the higher possibility
that it'll trigger such a storm as long as any of the page is written.

It means no matter which way we choose, as long as a write happened the
whole THP idea will get lost as long as the tracking is still in progress.

I still think zeropage should be rare in serious productions, I think it's
more likely that above will happen easily in practise and it can have an
negative impact on page faults. Even if we give more chance of having THPs
in the future, we at least pay something else. It's hard to tell which is
more worthwhile.

Not to mention that'll introduce more compexity to the kernel besides the
anonymous support, which should still be relatively straightforward.

So I'm fine with switching to pte markers, but I'd leave the PMD+ markers
alone for now.

>
> >
> > I wasn't really eager on this before because the workaround of pre-read
> > works good already (I assume slightly slower but it's fine; not until
> > someone starts to worry). But if we want to extend soft-dirty that's not
> > good at all to have any new user being requested to prefault memory and
> > figuring out why it's needed.
> >
> > >
> > > Having that said, I have the feeling that you and Muhammad have a plan to
> > > make it work using uffd-wp and I won't interfere. It would be nicer to use
> > > softdirty infrastructure IMHO, though.
> >
> > Thanks. If you have any good idea on reusing soft-dirty, please shoot.
> > I'll be perfectly happy with it as long as it resolves the issue for
> > Muhammad. Trust me - I wished the soft dirty thing worked out, but
> > unfortunately it didn't.. Because at least so far uffd-wp has two major
> > issues as I can see:
> >
> > (1) Memory type limitations (e.g. general fs memories stop working)

[1]

> > (2) Tracing uffd application is, afaict, impossible
>
> I guess the nice thing is, that you can only track individual ranges, you
> don't have to enable it for the whole process. I assume softdirty tracking
> could be similarly extended, but with uffd-wp this comes naturally.
>
> >
> > So if there's better way to do with soft-dirty or anything else (and I
> > assume it'll not be limited to any of above) it's time to say..
>
> I started discussing that [1] but as nobody seemed to care about my input I
> decided to not further spend my time on that. But maybe it's a more
> fundamental issue we'd have to solve to get this clean.

The idea of merging VMAs during clear_refs sounds interesting.

>
> The problem I had with the original approach is that it required precise
> softdirty tracking even when nobody cared about softdirty tracking, and that
> it made wrong assumptions about the meaning of VM_SOFTDIRTY. We shouldn't
> have to worry about any of that if it's disabled (just like with uffd-wp).
>
>
> The semantical difference I see is that with uffd-wp, initially (pte_none())
> all PTEs are "dirty". With soft-dirty, initially (pte_none()) all PTEs are
> "clean". We work around that with VM_SOFTIDRTY, which sets all PTEs of a VMA
> effectively dirty.
>
>
> Maybe we should rethink that approach and logically track soft-clean
> instead. That is, everything is assumed to be soft-dirty until we set it
> clean on a PTE/PMD/PUD level. Setting something clean is wp'ing a PTE and
> marking it soft-clean (which is, clearing the soft-dirty bit logically).

To me both solutions should be mostly the same but just inverted bits. The
bits (for async) doesn't make sense before the "trigger" is pulled, then
it's about 0/1 with inverted meanings to me. It's just that soft-dirty
will naturally work well with none ptes but uffd-wp is not (since none pte
has all bits 0).

>
> The special case is when we don't have anything mapped yet, which is the
> same thing we are trying to handle AFAICS for uffd-wp here. A PTE
> (pte_none()) in which case we have to install a soft-dirty PTE/PMD marker to
> remember that we marked it as clean -- or map the shared zeropage to mark
> that one clean (not set the soft-dirty bit).

Yes, soft-dirty is another valid user to pte markers; that's also in my
very original proposal of it. If we can have async-uffd-wp then it's
easier because the uffd-wp marker will simply be reused naturally.

>
> Further, I would propose to have VM_SOFTDIRTY_ENABLED flags per VMA and
> interfaces to (a) enable/disable it either for some VMAs or the whole MM and
> (b) a process toggle to automatically enable softclean tracking on new VMAs.
> So we can really only care about it when someone cares about softdirty
> tracking. The old "VM_SOFTDIRTY" flag could go, because everything
> (pte_none()) starts out softdirty. So VM_SOFTDIRTY -> VM_SOFTDIRTY_ENABLED
> with changed semantics.
>
> Enabling softdirty-tracking on a VMA might have to go over PTEs, to make
> sure we really don't have any soft-clean leftovers due to imprecise
> soft-dirty tracking on VMA level while it was disabled (setting all PTEs
> soft-dirty if they not already are). Might require a thought if it's really
> required.
>
> Note that I think this resembles what we are doing with uffd-wp. Not sure if
> we'd still have to handle some unmap handling on pagecache pages. We might
> want to remember that they are soft-clean using a PTE marker.
>
>
> Requires some more thought, but I'm going to throw it out here that I think
> there might be ways to modify softdirty tracking.

As you mentioned, I think the new proposal is growing into uffd-wp-alike.
I think that's (maybe) also another reason why reusing uffd here is a good
idea, because we don't need to reinvent everything.

Need a new vma flag to identify from old? Uffd-wp already has it.

What happens if shmem swapped out? Uffd-wp handles it.

Need to be accurate and walk the pgtables? Uffd-wp does it..

One thing I didn't mention before (mostly referring to the 1st major
"defect" of using uffd-wp above I said [1] on memory types): _maybe_ we can
someday extend at least async mode of uffd-wp to all memory types, so it'll
even get everything covered. So far I don't see a strong requirement of
doing so, but I don't see a major blocker either.

While the other "uffd cannot be nested" defect is actually the same to
soft-dirty (no way to have a tracee being able to clear_refs itself or
it'll also go a mess), it's just that we can still use soft-dirty to track
an uffd application.

>
> > > >
> > > > The other thing is it provides a way to make anon and !anon behave the same
> > > > on empty ptes; it's a pity that it was not already like that.
> > >
> > > In an ideal world, we'd simply be using PTE markers unconditionally I think
> > > and avoid this zeropage feature :/
> > >
> > > Is there any particular reason to have UFFD_FEATURE_WP_ZEROPAGE and not
> > > simply always do that unconditionally? (sure, we have to indicate to user
> > > space that it now works as expected) Are we really expecting to break user
> > > space by protecting what was asked for to protect?
> >
> > I suspect so.
> >
> > From high level, the major functional changes will be:
> >
> > (1) The user will start to receive more WP message with zero page being
> > reported,
> >
> > (2) Wr-protecting a very sparse memory can be much slower
> >
> > I would expect there're cases where the app just works as usual.
> >
> > However in some other cases the user may really not care about zero pages
> > at all, and I had a feeling that's actually the majority.
> >
> > Live snapshot is actually special because IIUC the old semantics should
> > work perfectly if the guest OS won't try to sanity check freed pages being
> > all zeros.. IOW that's some corner case, and if we can control that we may
> > not even need WP_ZEROPAGE too for QEMU, iiuc. For many other apps people
> > may leverage this (ignoring mem holes) and make the app faster.
> >
> > Normally when I'm not confident of any functional change, I'd rather use a
> > flag. Luckily uffd is very friendly to that, so the user can have better
> > control of what to expect. Some future app may explicitly want to always
> > ignore zero pages when on extremely sparse mem, and without the flag it
> > can't choose.
> >
> > >
> > > >
> > > > We can always optimize this behavior in the future with either
> > > > PMD/PUD/.. pte markers as you said, but IMHO that just needs further
> > > > justification on the complexity, and also on whether that's beneficial to
> > > > the majority to become the default behavior.
> > >
> > > As I said, usually any new features require good justification. Maybe there
> > > really is a measurable performance gain (less syscalls, less pgtable walks).
> >
> > Muhammad may have a word to say here; let's see whether he has any comment.
> >
> > Besides that, as I replied above I'll collect some data in my next post
> > regardless, with an attempt to optimize with huge zeropages on top.
>
> If we can agree on making the shared zeropage an implementation detail, that
> would be great and I'd see long-term value in that.
>
> [1]
> https://lkml.kernel.org/r/d95d59d7-308d-831c-d8bd-16d06e66e8af@xxxxxxxxxx

So I assume there's no major issue to not continue with a new version, then
I'll move on.

Thanks,

--
Peter Xu