From mboxrd@z Thu Jan 1 00:00:00 1970 From: Peter Zijlstra Message-Id: <20060322223118.12658.36826.sendpatchset@twins.localnet> In-Reply-To: <20060322223107.12658.14997.sendpatchset@twins.localnet> References: <20060322223107.12658.14997.sendpatchset@twins.localnet> Subject: [PATCH 01/34] mm: kill-page-activate.patch Date: Wed, 22 Mar 2006 23:31:50 +0100 Sender: owner-linux-mm@kvack.org Return-Path: To: linux-mm@kvack.org, linux-kernel@vger.kernel.org Cc: Bob Picco , Andrew Morton , IWAMOTO Toshihiro , Peter Zijlstra , Christoph Lameter , Wu Fengguang , Nick Piggin , Linus Torvalds , Rik van Riel , Marcelo Tosatti List-ID: From: Marcelo Tosatti Get rid of activate_page() callers. Instead, page activation is achieved through mark_page_accessed() interface. Signed-off-by: Marcelo Tosatti Signed-off-by: Peter Zijlstra --- include/linux/swap.h | 1 - mm/swapfile.c | 4 ++-- 2 files changed, 2 insertions(+), 3 deletions(-) Index: linux-2.6/include/linux/swap.h =================================================================== --- linux-2.6.orig/include/linux/swap.h 2006-03-13 20:37:08.000000000 +0100 +++ linux-2.6/include/linux/swap.h 2006-03-13 20:37:22.000000000 +0100 @@ -164,7 +164,6 @@ extern unsigned int nr_free_pagecache_pa /* linux/mm/swap.c */ extern void FASTCALL(lru_cache_add(struct page *)); extern void FASTCALL(lru_cache_add_active(struct page *)); -extern void FASTCALL(activate_page(struct page *)); extern void FASTCALL(mark_page_accessed(struct page *)); extern void lru_add_drain(void); extern int lru_add_drain_all(void); Index: linux-2.6/mm/swapfile.c =================================================================== --- linux-2.6.orig/mm/swapfile.c 2006-03-13 20:37:08.000000000 +0100 +++ linux-2.6/mm/swapfile.c 2006-03-13 20:37:22.000000000 +0100 @@ -435,7 +435,7 @@ static void unuse_pte(struct vm_area_str * Move the page to the active list so it is not * immediately swapped out again after swapon. */ - activate_page(page); + mark_page_accessed(page); } static int unuse_pte_range(struct vm_area_struct *vma, pmd_t *pmd, @@ -537,7 +537,7 @@ static int unuse_mm(struct mm_struct *mm * Activate page so shrink_cache is unlikely to unmap its * ptes while lock is dropped, so swapoff can make progress. */ - activate_page(page); + mark_page_accessed(page); unlock_page(page); down_read(&mm->mmap_sem); lock_page(page); -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@kvack.org. For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: email@kvack.org