From mboxrd@z Thu Jan 1 00:00:00 1970 Received: from m3.gw.fujitsu.co.jp ([10.0.50.73]) by fgwmail7.fujitsu.co.jp (Fujitsu Gateway) with ESMTP id mAA7he1K000850 for (envelope-from kosaki.motohiro@jp.fujitsu.com); Mon, 10 Nov 2008 16:43:40 +0900 Received: from smail (m3 [127.0.0.1]) by outgoing.m3.gw.fujitsu.co.jp (Postfix) with ESMTP id BDAE045DD7B for ; Mon, 10 Nov 2008 16:43:39 +0900 (JST) Received: from s3.gw.fujitsu.co.jp (s3.gw.fujitsu.co.jp [10.0.50.93]) by m3.gw.fujitsu.co.jp (Postfix) with ESMTP id 8886345DD7C for ; Mon, 10 Nov 2008 16:43:39 +0900 (JST) Received: from s3.gw.fujitsu.co.jp (localhost.localdomain [127.0.0.1]) by s3.gw.fujitsu.co.jp (Postfix) with ESMTP id 437161DB803A for ; Mon, 10 Nov 2008 16:43:39 +0900 (JST) Received: from m105.s.css.fujitsu.com (m105.s.css.fujitsu.com [10.249.87.105]) by s3.gw.fujitsu.co.jp (Postfix) with ESMTP id D3225E08003 for ; Mon, 10 Nov 2008 16:43:38 +0900 (JST) From: KOSAKI Motohiro Subject: Re: swappiness in 2.6.28-rc3? In-Reply-To: <20081108184625.6a64a518@diego-desktop> References: <200811081211.24000.gene.heskett@gmail.com> <20081108184625.6a64a518@diego-desktop> Message-Id: <20081110132742.6165.KOSAKI.MOTOHIRO@jp.fujitsu.com> MIME-Version: 1.0 Content-Type: text/plain; charset="US-ASCII" Content-Transfer-Encoding: 7bit Date: Mon, 10 Nov 2008 16:43:38 +0900 (JST) Sender: owner-linux-mm@kvack.org Return-Path: To: dcg Cc: kosaki.motohiro@jp.fujitsu.com, Gene Heskett , linux-kernel@vger.kernel.org, linux-mm , Rik van Riel List-ID: Hi CCed Rik van Riel > El Sat, 08 Nov 2008 12:11:23 -0500, Gene Heskett escribio: > > > Greetings; > > > > I have 2.6.28-rc3 with a 5 day uptime, and I have had to do a "swapoff -a; > > swapon -a" almost daily to clear the swap. > > > > This is about 18 hours since I last did that: > > Mem: 4151132k total, 2891180k used, 1259952k free, 281224k buffers > > Swap: 2048276k total, 85864k used, 1962412k free, 2078404k cached > > > > I don't recall having to do this with 2.6.27 or any of its -rc's. > > I've also noticed more swappiness (very probably due to the vm scanning > rework), but I can't say for sure if it's a bad thing... Could you please try to following patch? ----------------------------------------------------------------- From: Rik van Riel This patch still needs some testing under various workloads on different hardware - the approach should work but the threshold may need tweaking. When there is a lot of streaming IO going on, we do not want to scan or evict pages from the working set. The old VM used to skip any mapped page, but still evict indirect blocks and other data that is useful to cache. This patch adds logic to skip scanning the anon lists and the active file list if most of the file pages are on the inactive file list (where streaming IO pages live), while at the lowest scanning priority. If the system is not doing a lot of streaming IO, eg. the system is running a database workload, then more often used file pages will be on the active file list and this logic is automatically disabled. Signed-off-by: Rik van Riel Signed-off-by: KOSAKI Motohiro --- include/linux/mmzone.h | 1 + mm/vmscan.c | 18 ++++++++++++++++-- 2 files changed, 17 insertions(+), 2 deletions(-) Index: b/include/linux/mmzone.h =================================================================== --- a/include/linux/mmzone.h 2008-11-10 16:10:34.000000000 +0900 +++ b/include/linux/mmzone.h 2008-11-10 16:12:20.000000000 +0900 @@ -453,6 +453,7 @@ static inline int zone_is_oom_locked(con * queues ("queue_length >> 12") during an aging round. */ #define DEF_PRIORITY 12 +#define PRIO_CACHE_ONLY (DEF_PRIORITY+1) /* Maximum number of zones on a zonelist */ #define MAX_ZONES_PER_ZONELIST (MAX_NUMNODES * MAX_NR_ZONES) Index: b/mm/vmscan.c =================================================================== --- a/mm/vmscan.c 2008-11-10 16:10:34.000000000 +0900 +++ b/mm/vmscan.c 2008-11-10 16:11:30.000000000 +0900 @@ -1443,6 +1443,20 @@ static unsigned long shrink_zone(int pri } } + /* + * If there is a lot of sequential IO going on, most of the + * file pages will be on the inactive file list. We start + * out by reclaiming those pages, without putting pressure on + * the working set. We only do this if the bulk of the file pages + * are not in the working set (on the active file list). + */ + if (priority == PRIO_CACHE_ONLY && + (nr[LRU_INACTIVE_FILE] > nr[LRU_ACTIVE_FILE])) + for_each_evictable_lru(l) + /* Scan only the inactive_file list. */ + if (l != LRU_INACTIVE_FILE) + nr[l] = 0; + while (nr[LRU_INACTIVE_ANON] || nr[LRU_ACTIVE_FILE] || nr[LRU_INACTIVE_FILE]) { for_each_evictable_lru(l) { @@ -1573,7 +1587,7 @@ static unsigned long do_try_to_free_page } } - for (priority = DEF_PRIORITY; priority >= 0; priority--) { + for (priority = PRIO_CACHE_ONLY; priority >= 0; priority--) { sc->nr_scanned = 0; if (!priority) disable_swap_token(); @@ -1735,7 +1749,7 @@ loop_again: for (i = 0; i < pgdat->nr_zones; i++) temp_priority[i] = DEF_PRIORITY; - for (priority = DEF_PRIORITY; priority >= 0; priority--) { + for (priority = PRIO_CACHE_ONLY; priority >= 0; priority--) { int end_zone = 0; /* Inclusive. 0 = ZONE_DMA */ unsigned long lru_pages = 0; -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org