Re: [PATCH 0/13] memory cgroup updates v4

From: KAMEZAWA Hiroyuki
Date: Mon Sep 22 2008 - 07:23:14 EST


On Mon, 22 Sep 2008 19:51:59 +0900
KAMEZAWA Hiroyuki <kamezawa.hiroyu@xxxxxxxxxxxxxx> wrote:
> Brief description.
>
> 1/13 .... special mapping fix. (NEW)
> => avoid accounting pages not on LRU...which we cannot reclaim.
>
> 2/13 .... account swap-cache under lock.
> => move accounting of swap-cache under lock for avoiding unnecessary race.
>
> 3/13 .... make root cgroup to be unlimited.
> => fix root cgroup's memory limit to unlimited.
>
> 4/13 .... atomic-flags for page_cgroup
> => make page_cgroup->flags to be atomic.
>
> 5/13 .... implement move_account function.
> => add a function for moving page_cgroup's account to other cgroup.
>
> 6/13 ... force_empty to migrate account
> => move all account to root cgroup rather than forget all.
>
> 7/13 ...make mapping NULL (clean up)
> => ensure page->mapping to be NULL before calling mem_cgroup_uncharge_cache().
>
> 8/13 ...optimize cpustat
> => optimize access to per-cpu statistics for memcg.
>
> 9/13 ...lookup page_cgroup (CHANGED)
> => preallocate all page_cgroup at boot and remove page->page_cgroup pointer.
>
> 10/13...page_cgroup lookaside buffer
> => helps looking up page_cgroup from page.
>
> 11/13...lazy lru freeing page_cgroup (NEW)
> => do removal from LRU in bached manner like pagevec.
>
> 12/13...lazy lru add page_cgroup (NEW)
> => do addition to LRU in bached manner like pagevec.
>
> 13/13...swap accountig fix. (NEW)
> => fix race in swap accounting (can be happen)
> and this intrduce new protocal as precharge/commit/cancel.
>
> Some patches are big but not complicated I think.
>
Sorry for crazy patch numbering...

1 -> 1
2 -> 2
3 -> 3
3.5 -> 4
3.6 -> 5
4 -> 6
5 -> 7
6 -> 8
9 -> 9
10 -> 10
11 -> 11
12 -> 12

I may not able to do quick responce, sorry.

Thanks,
-Kame



--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/