Re: [PATCH 2/2] mem-hotplug: use different mempolicy in alloc_migrate_target()

From: Vlastimil Babka
Date: Mon Jul 18 2016 - 04:01:00 EST


On 07/15/2016 04:51 AM, Xishi Qiu wrote:
When we offline a node, the new page should alloced from other
nodes instead of the current node, because re-migrate is a waste of
time.

Ugh, I'm surprised that it's not the case already. Maybe the allocation from same node is already prevented indirectly somehow?

So use prefer mempolicy for hotplug, use default mempolicy for cma.

IMHO CMA should prefer the same node as the migrated page, if anything. Current task's mempolicy shouldn't restrict it (it's likely migrating pages of somebody else) or even guide its preferences. Ideally it would keep the original page's mempolicy, but we can't afford to track that...

Signed-off-by: Xishi Qiu <qiuxishi@xxxxxxxxxx>
---
include/linux/page-isolation.h | 2 +-
mm/memory_hotplug.c | 5 ++++-
mm/page_alloc.c | 2 +-
mm/page_isolation.c | 8 +++++---
4 files changed, 11 insertions(+), 6 deletions(-)

diff --git a/include/linux/page-isolation.h b/include/linux/page-isolation.h
index 047d647..c163de3 100644
--- a/include/linux/page-isolation.h
+++ b/include/linux/page-isolation.h
@@ -65,7 +65,7 @@ undo_isolate_page_range(unsigned long start_pfn, unsigned long end_pfn,
int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
bool skip_hwpoisoned_pages);

-struct page *alloc_migrate_target(struct page *page, unsigned long private,
+struct page *alloc_migrate_target(struct page *page, unsigned long nid,
int **resultp);

#endif
diff --git a/mm/memory_hotplug.c b/mm/memory_hotplug.c
index e3cbdca..b5963bf 100644
--- a/mm/memory_hotplug.c
+++ b/mm/memory_hotplug.c
@@ -1510,12 +1510,15 @@ do_migrate_range(unsigned long start_pfn, unsigned long end_pfn)
int move_pages = NR_OFFLINE_AT_ONCE_PAGES;
int not_managed = 0;
int ret = 0;
+ int nid = NUMA_NO_NODE;
LIST_HEAD(source);

for (pfn = start_pfn; pfn < end_pfn && move_pages > 0; pfn++) {
if (!pfn_valid(pfn))
continue;
page = pfn_to_page(pfn);
+ if (nid == NUMA_NO_NODE)
+ nid = next_node_in(page_to_nid(page), node_online_map);

if (PageHuge(page)) {
struct page *head = compound_head(page);
@@ -1568,7 +1571,7 @@ do_migrate_range(unsigned long start_pfn, unsigned long end_pfn)
* alloc_migrate_target should be improooooved!!

So, is this patch improooooving it enough that we can delete the comment? If not, could the patch be improooooved? :)

* migrate_pages returns # of failed pages.
*/
- ret = migrate_pages(&source, alloc_migrate_target, NULL, 0,
+ ret = migrate_pages(&source, alloc_migrate_target, NULL, nid,
MIGRATE_SYNC, MR_MEMORY_HOTPLUG);
if (ret)
putback_movable_pages(&source);
diff --git a/mm/page_alloc.c b/mm/page_alloc.c
index 6903b69..b99f1c2 100644
--- a/mm/page_alloc.c
+++ b/mm/page_alloc.c
@@ -7322,7 +7322,7 @@ static int __alloc_contig_migrate_range(struct compact_control *cc,
cc->nr_migratepages -= nr_reclaimed;

ret = migrate_pages(&cc->migratepages, alloc_migrate_target,
- NULL, 0, cc->mode, MR_CMA);
+ NULL, NUMA_NO_NODE, cc->mode, MR_CMA);
}
if (ret < 0) {
putback_movable_pages(&cc->migratepages);
diff --git a/mm/page_isolation.c b/mm/page_isolation.c
index 4f32c9f..f471be6 100644
--- a/mm/page_isolation.c
+++ b/mm/page_isolation.c
@@ -279,18 +279,20 @@ int test_pages_isolated(unsigned long start_pfn, unsigned long end_pfn,
return pfn < end_pfn ? -EBUSY : 0;
}

-struct page *alloc_migrate_target(struct page *page, unsigned long private,
+struct page *alloc_migrate_target(struct page *page, unsigned long nid,
int **resultp)
{
/*
- * TODO: allocate a destination hugepage from a nearest neighbor node,
+ * hugeTLB: allocate a destination page from a nearest neighbor node,

for hugeTLB it's still a TODO, by removing the word the rest of comment doesn't make much sense

* accordance with memory policy of the user process if possible. For
* now as a simple work-around, we use the next node for destination.
+ * Normal page: use prefer mempolicy for destination if called by
+ * hotplug, use default mempolicy for destination if called by cma.
*/
if (PageHuge(page))
return alloc_huge_page_node(page_hstate(compound_head(page)),
next_node_in(page_to_nid(page),
node_online_map));
else
- return alloc_page(GFP_HIGHUSER_MOVABLE);
+ return alloc_pages_node(nid, GFP_HIGHUSER_MOVABLE, 0);
}