Re: [PATCH 2/3] Reclaim invalidated page ASAP

From: Minchan Kim
Date: Tue Nov 30 2010 - 09:04:51 EST


On Tue, Nov 30, 2010 at 09:16:18AM +0000, Mel Gorman wrote:
> On Tue, Nov 30, 2010 at 07:41:30AM +0900, Minchan Kim wrote:
> > On Mon, Nov 29, 2010 at 04:57:06PM +0000, Mel Gorman wrote:
> > > On Tue, Nov 30, 2010 at 12:23:20AM +0900, Minchan Kim wrote:
> > > > invalidate_mapping_pages is very big hint to reclaimer.
> > > > It means user doesn't want to use the page any more.
> > > > So in order to prevent working set page eviction, this patch
> > > > move the page into tail of inactive list by PG_reclaim.
> > > >
> > > > Please, remember that pages in inactive list are working set
> > > > as well as active list. If we don't move pages into inactive list's
> > > > tail, pages near by tail of inactive list can be evicted although
> > > > we have a big clue about useless pages. It's totally bad.
> > > >
> > > > Now PG_readahead/PG_reclaim is shared.
> > > > fe3cba17 added ClearPageReclaim into clear_page_dirty_for_io for
> > > > preventing fast reclaiming readahead marker page.
> > > >
> > > > In this series, PG_reclaim is used by invalidated page, too.
> > > > If VM find the page is invalidated and it's dirty, it sets PG_reclaim
> > > > to reclaim asap. Then, when the dirty page will be writeback,
> > > > clear_page_dirty_for_io will clear PG_reclaim unconditionally.
> > > > It disturbs this serie's goal.
> > > >
> > > > I think it's okay to clear PG_readahead when the page is dirty, not
> > > > writeback time. So this patch moves ClearPageReadahead.
> > > >
> > > > Signed-off-by: Minchan Kim <minchan.kim@xxxxxxxxx>
> > > > Acked-by: Rik van Riel <riel@xxxxxxxxxx>
> > > > Cc: Wu Fengguang <fengguang.wu@xxxxxxxxx>
> > > > Cc: KOSAKI Motohiro <kosaki.motohiro@xxxxxxxxxxxxxx>
> > > > Cc: Johannes Weiner <hannes@xxxxxxxxxxx>
> > > > Cc: Nick Piggin <npiggin@xxxxxxxxx>
> > > > Cc: Mel Gorman <mel@xxxxxxxxx>
> > > >
> > > > Changelog since v2:
> > > > - put ClearPageReclaim in set_page_dirty - suggested by Wu.
> > > >
> > > > Changelog since v1:
> > > > - make the invalidated page reclaim asap - suggested by Andrew.
> > > > ---
> > > > mm/page-writeback.c | 12 +++++++++++-
> > > > mm/swap.c | 48 +++++++++++++++++++++++++++++++++++-------------
> > > > 2 files changed, 46 insertions(+), 14 deletions(-)
> > > >
> > > > diff --git a/mm/page-writeback.c b/mm/page-writeback.c
> > > > index fc93802..88587a5 100644
> > > > --- a/mm/page-writeback.c
> > > > +++ b/mm/page-writeback.c
> > > > @@ -1250,6 +1250,17 @@ int set_page_dirty(struct page *page)
> > > > {
> > > > struct address_space *mapping = page_mapping(page);
> > > >
> > > > + /*
> > > > + * readahead/lru_deactivate_page could remain
> > > > + * PG_readahead/PG_reclaim due to race with end_page_writeback
> > > > + * About readahead, if the page is written, the flags would be
> > > > + * reset. So no problem.
> > > > + * About lru_deactivate_page, if the page is redirty, the flag
> > > > + * will be reset. So no problem. but if the page is used by readahead
> > > > + * it will confuse readahead and make it restart the size rampup
> > > > + * process. But it's a trivial problem.
> > > > + */
> > > > + ClearPageReclaim(page);
> > > > if (likely(mapping)) {
> > > > int (*spd)(struct page *) = mapping->a_ops->set_page_dirty;
> > > > #ifdef CONFIG_BLOCK
> > > > @@ -1307,7 +1318,6 @@ int clear_page_dirty_for_io(struct page *page)
> > > >
> > > > BUG_ON(!PageLocked(page));
> > > >
> > > > - ClearPageReclaim(page);
> > > > if (mapping && mapping_cap_account_dirty(mapping)) {
> > > > /*
> > > > * Yes, Virginia, this is indeed insane.
> > > > diff --git a/mm/swap.c b/mm/swap.c
> > > > index 19e0812..936b281 100644
> > > > --- a/mm/swap.c
> > > > +++ b/mm/swap.c
> > > > @@ -275,28 +275,50 @@ void add_page_to_unevictable_list(struct page *page)
> > > > * into inative list's head. Because the VM expects the page would
> > > > * be writeout by flusher. The flusher's writeout is much effective
> > > > * than reclaimer's random writeout.
> > > > + *
> > > > + * If the page isn't page_mapped and dirty/writeback, the page
> > > > + * could reclaim asap using PG_reclaim.
> > > > + *
> > > > + * 1. active, mapped page -> none
> > > > + * 2. active, dirty/writeback page -> inactive, head, PG_reclaim
> > > > + * 3. inactive, mapped page -> none
> > > > + * 4. inactive, dirty/writeback page -> inactive, head, PG_reclaim
> > > > + * 5. Others -> none
> > > > + *
> > > > + * In 4, why it moves inactive's head, the VM expects the page would
> > > > + * be writeout by flusher. The flusher's writeout is much effective than
> > > > + * reclaimer's random writeout.
> > > > */
> > > > static void __lru_deactivate(struct page *page, struct zone *zone)
> > > > {
> > > > int lru, file;
> > > > - unsigned long vm_flags;
> > > > + int active = 0;
> > > >
> > > > - if (!PageLRU(page) || !PageActive(page))
> > > > + if (!PageLRU(page))
> > > > return;
> > > > -
> > > > /* Some processes are using the page */
> > > > if (page_mapped(page))
> > > > return;
> > > > -
> > > > - file = page_is_file_cache(page);
> > > > - lru = page_lru_base_type(page);
> > > > - del_page_from_lru_list(zone, page, lru + LRU_ACTIVE);
> > > > - ClearPageActive(page);
> > > > - ClearPageReferenced(page);
> > > > - add_page_to_lru_list(zone, page, lru);
> > > > - __count_vm_event(PGDEACTIVATE);
> > > > -
> > > > - update_page_reclaim_stat(zone, page, file, 0);
> > > > + if (PageActive(page))
> > > > + active = 1;
> > > > +
> > > > + if (PageWriteback(page) || PageDirty(page)) {
> > > > + /*
> > > > + * PG_reclaim could be raced with end_page_writeback
> > > > + * It can make readahead confusing. But race window
> > > > + * is _really_ small and it's non-critical problem.
> > > > + */
> > > > + SetPageReclaim(page);
> > > > +
> > > > + file = page_is_file_cache(page);
> > > > + lru = page_lru_base_type(page);
> > > > + del_page_from_lru_list(zone, page, lru + active);
> > > > + ClearPageActive(page);
> > > > + ClearPageReferenced(page);
> > > > + add_page_to_lru_list(zone, page, lru);
> > > > + __count_vm_event(PGDEACTIVATE);
> > >
> > > You update PGDEACTIVATE whether the page was active or not.
> >
> > My fault.
> > Resend.
> >
> > Subject: [PATCH v3 2/3] Reclaim invalidated page ASAP
> >
> > invalidate_mapping_pages is very big hint to reclaimer.
> > It means user doesn't want to use the page any more.
> > So in order to prevent working set page eviction, this patch
> > move the page into tail of inactive list by PG_reclaim.
> >
> > Please, remember that pages in inactive list are working set
> > as well as active list. If we don't move pages into inactive list's
> > tail, pages near by tail of inactive list can be evicted although
> > we have a big clue about useless pages. It's totally bad.
> >
> > Now PG_readahead/PG_reclaim is shared.
> > fe3cba17 added ClearPageReclaim into clear_page_dirty_for_io for
> > preventing fast reclaiming readahead marker page.
> >
> > In this series, PG_reclaim is used by invalidated page, too.
> > If VM find the page is invalidated and it's dirty, it sets PG_reclaim
> > to reclaim asap. Then, when the dirty page will be writeback,
> > clear_page_dirty_for_io will clear PG_reclaim unconditionally.
> > It disturbs this serie's goal.
> >
> > I think it's okay to clear PG_readahead when the page is dirty, not
> > writeback time. So this patch moves ClearPageReadahead.
> >
> > Signed-off-by: Minchan Kim <minchan.kim@xxxxxxxxx>
> > Acked-by: Rik van Riel <riel@xxxxxxxxxx>
> > Cc: Wu Fengguang <fengguang.wu@xxxxxxxxx>
> > Cc: KOSAKI Motohiro <kosaki.motohiro@xxxxxxxxxxxxxx>
> > Cc: Johannes Weiner <hannes@xxxxxxxxxxx>
> > Cc: Nick Piggin <npiggin@xxxxxxxxx>
> > Cc: Mel Gorman <mel@xxxxxxxxx>
> >
> > Changelog since v2:
> > - put ClearPageReclaim in set_page_dirty - suggested by Wu.
> >
> > Changelog since v1:
> > - make the invalidated page reclaim asap - suggested by Andrew.
> > ---
> > mm/page-writeback.c | 12 +++++++++++-
> > mm/swap.c | 49 ++++++++++++++++++++++++++++++++++++-------------
> > 2 files changed, 47 insertions(+), 14 deletions(-)
> >
> > diff --git a/mm/page-writeback.c b/mm/page-writeback.c
> > index fc93802..88587a5 100644
> > --- a/mm/page-writeback.c
> > +++ b/mm/page-writeback.c
> > @@ -1250,6 +1250,17 @@ int set_page_dirty(struct page *page)
> > {
> > struct address_space *mapping = page_mapping(page);
> >
> > + /*
> > + * readahead/lru_deactivate_page could remain
> > + * PG_readahead/PG_reclaim due to race with end_page_writeback
> > + * About readahead, if the page is written, the flags would be
> > + * reset. So no problem.
> > + * About lru_deactivate_page, if the page is redirty, the flag
> > + * will be reset. So no problem. but if the page is used by readahead
> > + * it will confuse readahead and make it restart the size rampup
> > + * process. But it's a trivial problem.
> > + */
> > + ClearPageReclaim(page);
> > if (likely(mapping)) {
> > int (*spd)(struct page *) = mapping->a_ops->set_page_dirty;
> > #ifdef CONFIG_BLOCK
> > @@ -1307,7 +1318,6 @@ int clear_page_dirty_for_io(struct page *page)
> >
> > BUG_ON(!PageLocked(page));
> >
> > - ClearPageReclaim(page);
> > if (mapping && mapping_cap_account_dirty(mapping)) {
> > /*
> > * Yes, Virginia, this is indeed insane.
> > diff --git a/mm/swap.c b/mm/swap.c
> > index 19e0812..1f1f435 100644
> > --- a/mm/swap.c
> > +++ b/mm/swap.c
> > @@ -275,28 +275,51 @@ void add_page_to_unevictable_list(struct page *page)
> > * into inative list's head. Because the VM expects the page would
> > * be writeout by flusher. The flusher's writeout is much effective
> > * than reclaimer's random writeout.
> > + *
> > + * If the page isn't page_mapped and dirty/writeback, the page
> > + * could reclaim asap using PG_reclaim.
> > + *
> > + * 1. active, mapped page -> none
> > + * 2. active, dirty/writeback page -> inactive, head, PG_reclaim
> > + * 3. inactive, mapped page -> none
> > + * 4. inactive, dirty/writeback page -> inactive, head, PG_reclaim
> > + * 5. Others -> none
> > + *
> > + * In 4, why it moves inactive's head, the VM expects the page would
> > + * be writeout by flusher. The flusher's writeout is much effective than
> > + * reclaimer's random writeout.
> > */
> > static void __lru_deactivate(struct page *page, struct zone *zone)
> > {
> > int lru, file;
> > - unsigned long vm_flags;
> > + int active = 0;
> >
> > - if (!PageLRU(page) || !PageActive(page))
> > + if (!PageLRU(page))
> > return;
> > -
> > /* Some processes are using the page */
> > if (page_mapped(page))
> > return;
> > -
> > - file = page_is_file_cache(page);
> > - lru = page_lru_base_type(page);
> > - del_page_from_lru_list(zone, page, lru + LRU_ACTIVE);
> > - ClearPageActive(page);
> > - ClearPageReferenced(page);
> > - add_page_to_lru_list(zone, page, lru);
> > - __count_vm_event(PGDEACTIVATE);
> > -
> > - update_page_reclaim_stat(zone, page, file, 0);
> > + if (PageActive(page))
> > + active = 1;
> > +
>
> I should have said this the last time but if you do another revision,
> make active a "bool". There is a very slow migration of int to bool in
> cases it makes sense. It's not urgent though.

Okay, I will fix it.

>
> > + if (PageWriteback(page) || PageDirty(page)) {
> > + /*
> > + * PG_reclaim could be raced with end_page_writeback
> > + * It can make readahead confusing. But race window
> > + * is _really_ small and it's non-critical problem.
> > + */
> > + SetPageReclaim(page);
> > +
> > + file = page_is_file_cache(page);
> > + lru = page_lru_base_type(page);
> > + del_page_from_lru_list(zone, page, lru + active);
> > + ClearPageActive(page);
> > + ClearPageReferenced(page);
> > + add_page_to_lru_list(zone, page, lru);
> > + if (active)
> > + __count_vm_event(PGDEACTIVATE);
> > + update_page_reclaim_stat(zone, page, file, 0);
> > + }
> > }
> >
>
> Whether you update active's type or not;
>
> Acked-by: Mel Gorman <mel@xxxxxxxxx>

Thanks, Mel.

>
> --
> Mel Gorman
> Part-time Phd Student Linux Technology Center
> University of Limerick IBM Dublin Software Lab

--
Kind regards,
Minchan Kim
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/