[PATCH 22/22] mm: page allocator: Drain magazines for direct compact failures

From: Mel Gorman
Date: Wed May 08 2013 - 12:04:06 EST


THP allocations may fail due to pages pinned in magazines so drain them
in the event of a direct compact failure. Similarly drain the magazines
during memory hot-remove, memory failure and page isolation as before.

Signed-off-by: Mel Gorman <mgorman@xxxxxxx>
---
include/linux/gfp.h | 2 ++
mm/memory-failure.c | 1 +
mm/memory_hotplug.c | 2 ++
mm/page_alloc.c | 63 +++++++++++++++++++++++++++++++++++++++++++++--------
mm/page_isolation.c | 1 +
5 files changed, 60 insertions(+), 9 deletions(-)

diff --git a/include/linux/gfp.h b/include/linux/gfp.h
index 53844b4..fafa28b 100644
--- a/include/linux/gfp.h
+++ b/include/linux/gfp.h
@@ -375,6 +375,8 @@ extern void free_memcg_kmem_pages(unsigned long addr, unsigned int order);
#define free_page(addr) free_pages((addr), 0)

void page_alloc_init(void);
+void drain_zone_magazine(struct zone *zone);
+void drain_all_magazines(void);

/*
* gfp_allowed_mask is set to GFP_BOOT_MASK during early boot to restrict what
diff --git a/mm/memory-failure.c b/mm/memory-failure.c
index 3175ffd..cd201a3 100644
--- a/mm/memory-failure.c
+++ b/mm/memory-failure.c
@@ -237,6 +237,7 @@ void shake_page(struct page *p, int access)
lru_add_drain_all();
if (PageLRU(p))
return;
+ drain_zone_magazine(page_zone(p));
if (PageLRU(p) || is_free_buddy_page(p))
return;
}
diff --git a/mm/memory_hotplug.c b/mm/memory_hotplug.c
index 63f473c..b35c6ee 100644
--- a/mm/memory_hotplug.c
+++ b/mm/memory_hotplug.c
@@ -1526,6 +1526,7 @@ repeat:
if (drain) {
lru_add_drain_all();
cond_resched();
+ drain_all_magazines();
}

pfn = scan_lru_pages(start_pfn, end_pfn);
@@ -1546,6 +1547,7 @@ repeat:
/* drain all zone's lru pagevec, this is asynchronous... */
lru_add_drain_all();
yield();
+ drain_all_magazines();
/* check again */
offlined_pages = check_pages_isolated(start_pfn, end_pfn);
if (offlined_pages < 0) {
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index 374adf8..0f0bc18 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -1164,23 +1164,17 @@ struct page *__rmqueue_magazine(struct free_magazine *mag,
return page;
}

-static void magazine_drain(struct zone *zone, struct free_magazine *mag,
- int migratetype)
+static void __magazine_drain(struct zone *zone, struct free_magazine *mag,
+ int migratetype, int min_to_free, int to_free)
{
struct list_head *list;
struct page *page;
unsigned int batch_free = 0;
- unsigned int to_free = MAGAZINE_MAX_FREE_BATCH;
unsigned int nr_freed_cma = 0, nr_freed = 0;
unsigned long flags;
struct free_area_magazine *area = &mag->area;
LIST_HEAD(free_list);

- if (area->nr_free < MAGAZINE_LIMIT) {
- unlock_magazine(mag);
- return;
- }
-
/* Free batch number of pages */
while (to_free) {
/*
@@ -1216,7 +1210,7 @@ static void magazine_drain(struct zone *zone, struct free_magazine *mag,
} while (--to_free && --batch_free && !list_empty(list));

/* Watch for parallel contention */
- if (nr_freed > MAGAZINE_MIN_FREE_BATCH &&
+ if (nr_freed > min_to_free &&
magazine_contended(mag))
break;
}
@@ -1236,6 +1230,53 @@ static void magazine_drain(struct zone *zone, struct free_magazine *mag,
spin_unlock_irqrestore(&zone->lock, flags);
}

+static void magazine_drain(struct zone *zone, struct free_magazine *mag,
+ int migratetype)
+{
+ if (mag->area.nr_free < MAGAZINE_LIMIT) {
+ unlock_magazine(mag);
+ return;
+ }
+
+ __magazine_drain(zone, mag, migratetype, MAGAZINE_MIN_FREE_BATCH,
+ MAGAZINE_MAX_FREE_BATCH);
+}
+
+void drain_zone_magazine(struct zone *zone)
+{
+ int i;
+
+ for (i = 0; i < NR_MAGAZINES; i++) {
+ struct free_magazine *mag = &zone->noirq_magazine[i];
+
+ spin_lock(&zone->noirq_magazine[i].lock);
+ __magazine_drain(zone, mag, MIGRATE_UNMOVABLE,
+ mag->area.nr_free,
+ mag->area.nr_free);
+ spin_unlock(&zone->noirq_magazine[i].lock);
+ }
+}
+
+static void drain_zonelist_magazine(struct zonelist *zonelist,
+ enum zone_type high_zoneidx, nodemask_t *nodemask)
+{
+ struct zoneref *z;
+ struct zone *zone;
+
+ for_each_zone_zonelist_nodemask(zone, z, zonelist,
+ high_zoneidx, nodemask) {
+ drain_zone_magazine(zone);
+ }
+}
+
+void drain_all_magazines(void)
+{
+ struct zone *zone;
+
+ for_each_zone(zone)
+ drain_zone_magazine(zone);
+}
+
/* Prepare a page for freeing and return its migratetype */
static inline int free_base_page_prep(struct page *page)
{
@@ -2170,6 +2211,9 @@ __alloc_pages_direct_compact(gfp_t gfp_mask, unsigned int order,
if (*did_some_progress != COMPACT_SKIPPED) {
struct page *page;

+ /* Page migration frees to the magazine but we want merging */
+ drain_zonelist_magazine(zonelist, high_zoneidx, nodemask);
+
page = get_page_from_freelist(gfp_mask, nodemask,
order, zonelist, high_zoneidx,
alloc_flags & ~ALLOC_NO_WATERMARKS,
@@ -5766,6 +5810,7 @@ int alloc_contig_range(unsigned long start, unsigned long end,
*/

lru_add_drain_all();
+ drain_all_magazines();

order = 0;
outer_start = start;
diff --git a/mm/page_isolation.c b/mm/page_isolation.c
index af79199..1279d9d 100644
--- a/mm/page_isolation.c
+++ b/mm/page_isolation.c
@@ -62,6 +62,7 @@ out:
nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLATE);

__mod_zone_freepage_state(zone, -nr_pages, migratetype);
+ drain_zone_magazine(zone);
}

spin_unlock_irqrestore(&zone->lock, flags);
--
1.8.1.4

--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/