Re: [PATCH v6 0/9] memcg: per cgroup dirty page accounting

From: Vivek Goyal
Date: Thu Mar 17 2011 - 10:49:42 EST


On Thu, Mar 17, 2011 at 01:43:50PM +0100, Johannes Weiner wrote:
> On Wed, Mar 16, 2011 at 09:41:48PM -0700, Greg Thelen wrote:
> > In '[PATCH v6 8/9] memcg: check memcg dirty limits in page writeback' Jan and
> > Vivek have had some discussion around how memcg and writeback mesh.
> > In my mind, the
> > discussions in 8/9 are starting to blend with this thread.
> >
> > I have been thinking about Johannes' struct memcg_mapping. I think the idea
> > may address several of the issues being discussed, especially
> > interaction between
> > IO-less balance_dirty_pages() and memcg writeback.
> >
> > Here is my thinking. Feedback is most welcome!
> >
> > The data structures:
> > - struct memcg_mapping {
> > struct address_space *mapping;
> > struct mem_cgroup *memcg;
> > int refcnt;
> > };
> > - each memcg contains a (radix, hash_table, etc.) mapping from bdi to memcg_bdi.
> > - each memcg_bdi contains a mapping from inode to memcg_mapping. This may be a
> > very large set representing many cached inodes.
> > - each memcg_mapping represents all pages within an bdi,inode,memcg. All
> > corresponding cached inode pages point to the same memcg_mapping via
> > pc->mapping. I assume that all pages of inode belong to no more than one bdi.
> > - manage a global list of memcg that are over their respective background dirty
> > limit.
> > - i_mapping continues to point to a traditional non-memcg mapping (no change
> > here).
> > - none of these memcg_* structures affect root cgroup or kernels with memcg
> > configured out.
>
> So structures roughly like this:
>
> struct mem_cgroup {
> ...
> /* key is struct backing_dev_info * */
> struct rb_root memcg_bdis;
> };
>
> struct memcg_bdi {
> /* key is struct address_space * */
> struct rb_root memcg_mappings;
> struct rb_node node;
> };
>
> struct memcg_mapping {
> struct address_space *mapping;
> struct mem_cgroup *memcg;
> struct rb_node node;
> atomic_t count;
> };
>
> struct page_cgroup {
> ...
> struct memcg_mapping *memcg_mapping;
> };
>
> > The routines under discussion:
> > - memcg charging a new inode page to a memcg: will use inode->mapping and inode
> > to walk memcg -> memcg_bdi -> memcg_mapping and lazily allocating missing
> > levels in data structure.
> >
> > - Uncharging a inode page from a memcg: will use pc->mapping->memcg to locate
> > memcg. If refcnt drops to zero, then remove memcg_mapping from the memcg_bdi.
> > Also delete memcg_bdi if last memcg_mapping is removed.
> >
> > - account_page_dirtied(): nothing new here, continue to set the per-page flags
> > and increment the memcg per-cpu dirty page counter. Same goes for routines
> > that mark pages in writeback and clean states.
>
> We may want to remember the dirty memcg_mappings so that on writeback
> we don't have to go through every single one that the memcg refers to?
>
> > - mem_cgroup_balance_dirty_pages(): if memcg dirty memory usage if above
> > background limit, then add memcg to global memcg_over_bg_limit list and use
> > memcg's set of memcg_bdi to wakeup each(?) corresponding bdi flusher. If over
> > fg limit, then use IO-less style foreground throttling with per-memcg per-bdi
> > (aka memcg_bdi) accounting structure.
>
> I wonder if we could just schedule a for_background work manually in
> the memcg case that writes back the corresponding memcg_bdi set (and
> e.g. having it continue until either the memcg is below bg thresh OR
> the global bg thresh is exceeded OR there is other work scheduled)?
> Then we would get away without the extra list, and it doesn't sound
> overly complex to implement.

Jan tought that design of maintaining a list of memocy groups (memcg_bdi)
in this case is cleaner. So he preferred that let the queuing of work be
for sync work and all this background writeout and IO less throttling
stuff can be covered through background writeout logic.

I think I tend to agree that keeping a list is a clean design as bdi
is shared medium and now flusher thread can decide how to divide the
bandwidth of shared bdi among the queued memory cgroups. So as long as
it does not turn out to be complex, I think maintaining a separate list
of cgroups waiting for writeback on this bdi is not a bad idea.

Thanks
Vivek
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/