Re: [PATCH 4/5] mm, compaction: always update cached scanner positions

From: Vlastimil Babka
Date: Fri Oct 31 2014 - 11:54:02 EST


On 10/28/2014 08:08 AM, Joonsoo Kim wrote:

And, I guess that pageblock skip feature effectively disable pageblock
rescanning if there is no freepage during rescan.

If there's no freepage during rescan, then the cached free_pfn also
won't be pointed to the pageblock anymore. Regardless of pageblock skip
being set, there will not be second rescan. But there will still be the
first rescan to determine there are no freepages.

Yes, What I'd like to say is that these would work well. Just decreasing
few percent of scanning page doesn't look good to me to validate this
patch, because there is some facilities to reduce rescan overhead and

The mechanisms have a tradeoff, while this patch didn't seem to have negative consequences.

compaction is fundamentally time-consuming process. Moreover, failure of
compaction could cause serious system crash in some cases.

Relying on successful high-order allocation for not crashing is dangerous, success is never guaranteed. Such critical allocation should try harder than fail due to a single compaction attempt. With this argument you could aim to remove all the overhead reducing heuristics.

This patch would
eliminate effect of pageblock skip feature.

I don't think so (as explained above). Also if free pages were isolated
(and then returned and skipped over), the pageblock should remain
without skip bit, so after scanners meet and positions reset (which
doesn't go hand in hand with skip bit reset), the next round will skip
over the blocks without freepages and find quickly the blocks where free
pages were skipped in the previous round.

IIUC, compaction logic assume that there are many temporary failure
conditions. Retrying from others would reduce effect of this temporary
failure so implementation looks as is.

The implementation of pfn caching was written at time when we did not
keep isolated free pages between migration attempts in a single
compaction run. And the idea of async compaction is to try with minimal
effort (thus latency), and if there's a failure, try somewhere else.
Making sure we don't skip anything doesn't seem productive.

free_pfn is shared by async/sync compaction and unconditional updating
causes sync compaction to stop prematurely, too.

And, if this patch makes migrate/freepage scanner meet more frequently,
there is one problematic scenario.

OK, so you don't find a problem with how this patch changes migration scanner caching, just the free scanner, right?
So how about making release_freepages() return the highest freepage pfn it encountered (could perhaps do without comparing individual pfn's, the list should be ordered so it could be just the pfn of first or last page in the list, but need to check that) and updating cached free pfn with that? That should ensure rescanning only when needed.

compact_finished() doesn't check how many work we did. It just check
if both scanners meet. Even if we failed to allocate high order page
due to little work, compaction would be deffered for later user.
This scenario wouldn't happen frequently if updating cached pfn is
limited. But, this patch may enlarge the possibility of this problem.

I doubt it changes the possibility substantially, but nevermind.

This is another problem of current logic, and, should be fixed, but,
there is now.

If something needs the high-order allocation succeed that badly, then the proper GFP flags should result in further reclaim and compaction attempts (hopefully) and not give up after first sync compaction failure.

Thanks.


--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/