[PATCH V2] Allow compaction of unevictable pages

From: Eric B Munson
Date: Mon Mar 09 2015 - 13:15:38 EST


Currently, pages which are marked as unevictable are protected from
compaction, but not from other types of migration. The mlock
desctription does not promise that all page faults will be avoided, only
major ones so this protection is not necessary. This extra protection
can cause problems for applications that are using mlock to avoid
swapping pages out, but require order > 0 allocations to continue to
succeed in a fragmented environment. This patch removes the
ISOLATE_UNEVICTABLE mode and the check for it in __isolate_lru_page().

To illustrate this problem I wrote a quick test program that mmaps a
large number of 1MB files filled with random data. These maps are
created locked and read only. Then every other mmap is unmapped and I
attempt to allocate huge pages to the static huge page pool. Without
this patch I am unable to allocate any huge pages after fragmenting
memory. With it, I can allocate almost all the space freed by unmapping
as huge pages.

Signed-off-by: Eric B Munson <emunson@xxxxxxxxxx>

Cc: Vlastimil Babka <vbabka@xxxxxxx>
Cc: Thomas Gleixner <tglx@xxxxxxxxxxxxx>
Cc: Christoph Lameter <cl@xxxxxxxxx>
Cc: Peter Zijlstra <peterz@xxxxxxxxxxxxx>
Cc: Mel Gorman <mgorman@xxxxxxx>
Cc: David Rientjes <rientjes@xxxxxxxxxx>
Cc: linux-mm@xxxxxxxxx
Cc: linux-kernel@xxxxxxxxxxxxxxx

---
include/linux/mmzone.h | 2 --
mm/compaction.c | 3 +--
mm/vmscan.c | 4 ----
3 files changed, 1 insertion(+), 8 deletions(-)

diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h
index f279d9c..599fb01 100644
--- a/include/linux/mmzone.h
+++ b/include/linux/mmzone.h
@@ -232,8 +232,6 @@ struct lruvec {
#define ISOLATE_UNMAPPED ((__force isolate_mode_t)0x2)
/* Isolate for asynchronous migration */
#define ISOLATE_ASYNC_MIGRATE ((__force isolate_mode_t)0x4)
-/* Isolate unevictable pages */
-#define ISOLATE_UNEVICTABLE ((__force isolate_mode_t)0x8)

/* LRU Isolation modes. */
typedef unsigned __bitwise__ isolate_mode_t;
diff --git a/mm/compaction.c b/mm/compaction.c
index 8c0d945..4a8ea87 100644
--- a/mm/compaction.c
+++ b/mm/compaction.c
@@ -872,8 +872,7 @@ isolate_migratepages_range(struct compact_control *cc, unsigned long start_pfn,
if (!pageblock_pfn_to_page(pfn, block_end_pfn, cc->zone))
continue;

- pfn = isolate_migratepages_block(cc, pfn, block_end_pfn,
- ISOLATE_UNEVICTABLE);
+ pfn = isolate_migratepages_block(cc, pfn, block_end_pfn, 0);

/*
* In case of fatal failure, release everything that might
diff --git a/mm/vmscan.c b/mm/vmscan.c
index 5e8eadd..3b2a444 100644
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
@@ -1234,10 +1234,6 @@ int __isolate_lru_page(struct page *page, isolate_mode_t mode)
if (!PageLRU(page))
return ret;

- /* Compaction should not handle unevictable pages but CMA can do so */
- if (PageUnevictable(page) && !(mode & ISOLATE_UNEVICTABLE))
- return ret;
-
ret = -EBUSY;

/*
--
1.7.9.5

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/