From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from m3.gw.fujitsu.co.jp ([10.0.50.73]) by fgwmail5.fujitsu.co.jp (Fujitsu Gateway) with ESMTP id mB1CHkAg028167 for (envelope-from kosaki.motohiro@jp.fujitsu.com); Mon, 1 Dec 2008 21:17:46 +0900 Received: from smail (m3 [127.0.0.1]) by outgoing.m3.gw.fujitsu.co.jp (Postfix) with ESMTP id 2A95B45DD7D for ; Mon, 1 Dec 2008 21:17:46 +0900 (JST) Received: from s3.gw.fujitsu.co.jp (s3.gw.fujitsu.co.jp [10.0.50.93]) by m3.gw.fujitsu.co.jp (Postfix) with ESMTP id EBF7745DD78 for ; Mon, 1 Dec 2008 21:17:45 +0900 (JST) Received: from s3.gw.fujitsu.co.jp (localhost.localdomain [127.0.0.1]) by s3.gw.fujitsu.co.jp (Postfix) with ESMTP id C400A1DB803C for ; Mon, 1 Dec 2008 21:17:45 +0900 (JST) Received: from m107.s.css.fujitsu.com (m107.s.css.fujitsu.com [10.249.87.107]) by s3.gw.fujitsu.co.jp (Postfix) with ESMTP id 6910A1DB803F for ; Mon, 1 Dec 2008 21:17:45 +0900 (JST) From: KOSAKI Motohiro Subject: [PATCH 08/11] memcg: make zone_reclaim_stat In-Reply-To: <20081201205810.1CCA.KOSAKI.MOTOHIRO@jp.fujitsu.com> References: <20081201205810.1CCA.KOSAKI.MOTOHIRO@jp.fujitsu.com> Message-Id: <20081201211646.1CE2.KOSAKI.MOTOHIRO@jp.fujitsu.com> MIME-Version: 1.0 Content-Type: text/plain; charset="US-ASCII" Content-Transfer-Encoding: 7bit Date: Mon, 1 Dec 2008 21:17:44 +0900 (JST) Sender: owner-linux-mm@kvack.org Return-Path: To: LKML , linux-mm , Andrew Morton , Balbir Singh , KAMEZAWA Hiroyuki , Rik van Riel Cc: kosaki.motohiro@jp.fujitsu.com List-ID: introduce mem_cgroup_per_zone::reclaim_stat member and its statics collecting function. Now, get_scan_ratio() can calculate correct value although memcg reclaim. Signed-off-by: KOSAKI Motohiro --- include/linux/memcontrol.h | 16 ++++++++++++++++ mm/memcontrol.c | 23 +++++++++++++++++++++++ mm/swap.c | 14 ++++++++++++++ mm/vmscan.c | 27 +++++++++++++-------------- 4 files changed, 66 insertions(+), 14 deletions(-) Index: b/include/linux/memcontrol.h =================================================================== --- a/include/linux/memcontrol.h +++ b/include/linux/memcontrol.h @@ -95,6 +95,10 @@ int mem_cgroup_inactive_anon_is_low(stru unsigned long mem_cgroup_zone_nr_pages(struct mem_cgroup *memcg, struct zone *zone, enum lru_list lru); +struct zone_reclaim_stat *mem_cgroup_get_reclaim_stat(struct mem_cgroup *memcg, + struct zone *zone); +struct zone_reclaim_stat* +mem_cgroup_get_reclaim_stat_by_page(struct page *page); #ifdef CONFIG_CGROUP_MEM_RES_CTLR_SWAP extern int do_swap_account; @@ -261,6 +265,18 @@ mem_cgroup_zone_nr_pages(struct mem_cgro } +static inline struct zone_reclaim_stat* +mem_cgroup_get_reclaim_stat(struct mem_cgroup *memcg, struct zone *zone) +{ + return NULL; +} + +static inline struct zone_reclaim_stat* +mem_cgroup_get_reclaim_stat_by_page(struct page *page) +{ + return NULL; +} + #endif /* CONFIG_CGROUP_MEM_CONT */ #endif /* _LINUX_MEMCONTROL_H */ Index: b/mm/memcontrol.c =================================================================== --- a/mm/memcontrol.c +++ b/mm/memcontrol.c @@ -103,6 +103,8 @@ struct mem_cgroup_per_zone { */ struct list_head lists[NR_LRU_LISTS]; unsigned long count[NR_LRU_LISTS]; + + struct zone_reclaim_stat reclaim_stat; }; /* Macro for accessing counter */ #define MEM_CGROUP_ZSTAT(mz, idx) ((mz)->count[(idx)]) @@ -458,6 +460,27 @@ unsigned long mem_cgroup_zone_nr_pages(s return MEM_CGROUP_ZSTAT(mz, lru); } +struct zone_reclaim_stat *mem_cgroup_get_reclaim_stat(struct mem_cgroup *memcg, + struct zone *zone) +{ + int nid = zone->zone_pgdat->node_id; + int zid = zone_idx(zone); + struct mem_cgroup_per_zone *mz = mem_cgroup_zoneinfo(memcg, nid, zid); + + return &mz->reclaim_stat; +} + +struct zone_reclaim_stat *mem_cgroup_get_reclaim_stat_by_page(struct page *page) +{ + struct page_cgroup *pc = lookup_page_cgroup(page); + struct mem_cgroup_per_zone *mz = page_cgroup_zoneinfo(pc); + + if (!mz) + return NULL; + + return &mz->reclaim_stat; +} + unsigned long mem_cgroup_isolate_pages(unsigned long nr_to_scan, struct list_head *dst, unsigned long *scanned, int order, Index: b/mm/swap.c =================================================================== --- a/mm/swap.c +++ b/mm/swap.c @@ -158,6 +158,7 @@ void activate_page(struct page *page) { struct zone *zone = page_zone(page); struct zone_reclaim_stat *reclaim_stat = &zone->reclaim_stat; + struct zone_reclaim_stat *memcg_reclaim_stat; spin_lock_irq(&zone->lru_lock); if (PageLRU(page) && !PageActive(page) && !PageUnevictable(page)) { @@ -172,6 +173,12 @@ void activate_page(struct page *page) reclaim_stat->recent_rotated[!!file]++; reclaim_stat->recent_scanned[!!file]++; + + memcg_reclaim_stat = mem_cgroup_get_reclaim_stat_by_page(page); + if (memcg_reclaim_stat) { + memcg_reclaim_stat->recent_rotated[!!file]++; + memcg_reclaim_stat->recent_scanned[!!file]++; + } } spin_unlock_irq(&zone->lru_lock); } @@ -400,6 +407,7 @@ void ____pagevec_lru_add(struct pagevec int i; struct zone *zone = NULL; struct zone_reclaim_stat *reclaim_stat = NULL; + struct zone_reclaim_stat *memcg_reclaim_stat = NULL; VM_BUG_ON(is_unevictable_lru(lru)); @@ -413,6 +421,8 @@ void ____pagevec_lru_add(struct pagevec spin_unlock_irq(&zone->lru_lock); zone = pagezone; reclaim_stat = &zone->reclaim_stat; + memcg_reclaim_stat = + mem_cgroup_get_reclaim_stat_by_page(page); spin_lock_irq(&zone->lru_lock); } VM_BUG_ON(PageActive(page)); @@ -421,9 +431,13 @@ void ____pagevec_lru_add(struct pagevec SetPageLRU(page); file = is_file_lru(lru); reclaim_stat->recent_scanned[file]++; + if (memcg_reclaim_stat) + memcg_reclaim_stat->recent_scanned[file]++; if (is_active_lru(lru)) { SetPageActive(page); reclaim_stat->recent_rotated[file]++; + if (memcg_reclaim_stat) + memcg_reclaim_stat->recent_rotated[file]++; } add_page_to_lru_list(zone, page, lru); } Index: b/mm/vmscan.c =================================================================== --- a/mm/vmscan.c +++ b/mm/vmscan.c @@ -134,6 +134,9 @@ static DECLARE_RWSEM(shrinker_rwsem); static struct zone_reclaim_stat *get_reclaim_stat(struct zone *zone, struct scan_control *sc) { + if (!scan_global_lru(sc)) + mem_cgroup_get_reclaim_stat(sc->mem_cgroup, zone); + return &zone->reclaim_stat; } @@ -1141,17 +1144,14 @@ static unsigned long shrink_inactive_lis __mod_zone_page_state(zone, NR_INACTIVE_ANON, -count[LRU_INACTIVE_ANON]); - if (scan_global_lru(sc)) { + if (scan_global_lru(sc)) zone->pages_scanned += nr_scan; - reclaim_stat->recent_scanned[0] += - count[LRU_INACTIVE_ANON]; - reclaim_stat->recent_scanned[0] += - count[LRU_ACTIVE_ANON]; - reclaim_stat->recent_scanned[1] += - count[LRU_INACTIVE_FILE]; - reclaim_stat->recent_scanned[1] += - count[LRU_ACTIVE_FILE]; - } + + reclaim_stat->recent_scanned[0] += count[LRU_INACTIVE_ANON]; + reclaim_stat->recent_scanned[0] += count[LRU_ACTIVE_ANON]; + reclaim_stat->recent_scanned[1] += count[LRU_INACTIVE_FILE]; + reclaim_stat->recent_scanned[1] += count[LRU_ACTIVE_FILE]; + spin_unlock_irq(&zone->lru_lock); nr_scanned += nr_scan; @@ -1209,7 +1209,7 @@ static unsigned long shrink_inactive_lis SetPageLRU(page); lru = page_lru(page); add_page_to_lru_list(zone, page, lru); - if (PageActive(page) && scan_global_lru(sc)) { + if (PageActive(page)) { int file = !!page_is_file_cache(page); reclaim_stat->recent_rotated[file]++; } @@ -1289,8 +1289,8 @@ static void shrink_active_list(unsigned */ if (scan_global_lru(sc)) { zone->pages_scanned += pgscanned; - reclaim_stat->recent_scanned[!!file] += pgmoved; } + reclaim_stat->recent_scanned[!!file] += pgmoved; if (file) __mod_zone_page_state(zone, NR_ACTIVE_FILE, -pgmoved); @@ -1323,8 +1323,7 @@ static void shrink_active_list(unsigned * This helps balance scan pressure between file and anonymous * pages in get_scan_ratio. */ - if (scan_global_lru(sc)) - reclaim_stat->recent_rotated[!!file] += pgmoved; + reclaim_stat->recent_rotated[!!file] += pgmoved; /* * Move the pages to the [file or anon] inactive list. -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org