From: Johannes Weiner <hannes@cmpxchg.org>
To: Mel Gorman <mel@csn.ul.ie>
Cc: linux-kernel@vger.kernel.org, linux-fsdevel@vger.kernel.org,
linux-mm@kvack.org, Dave Chinner <david@fromorbit.com>,
Chris Mason <chris.mason@oracle.com>,
Nick Piggin <npiggin@suse.de>, Rik van Riel <riel@redhat.com>,
Christoph Hellwig <hch@infradead.org>,
Wu Fengguang <fengguang.wu@intel.com>,
KAMEZAWA Hiroyuki <kamezawa.hiroyu@jp.fujitsu.com>,
KOSAKI Motohiro <kosaki.motohiro@jp.fujitsu.com>,
Andrew Morton <akpm@linux-foundation.org>,
Andrea Arcangeli <aarcange@redhat.com>
Subject: Re: [PATCH 4/8] vmscan: Do not writeback filesystem pages in direct reclaim
Date: Wed, 21 Jul 2010 00:02:18 +0200 [thread overview]
Message-ID: <20100720220218.GE16031@cmpxchg.org> (raw)
In-Reply-To: <20100720134555.GU13117@csn.ul.ie>
On Tue, Jul 20, 2010 at 02:45:56PM +0100, Mel Gorman wrote:
> On Tue, Jul 20, 2010 at 12:14:20AM +0200, Johannes Weiner wrote:
> > > @@ -639,6 +694,9 @@ static noinline_for_stack void free_page_list(struct list_head *free_pages)
> > > pagevec_free(&freed_pvec);
> > > }
> > >
> > > +/* Direct lumpy reclaim waits up to 5 seconds for background cleaning */
> > > +#define MAX_SWAP_CLEAN_WAIT 50
> > > +
> > > /*
> > > * shrink_page_list() returns the number of reclaimed pages
> > > */
> > > @@ -646,13 +704,19 @@ static unsigned long shrink_page_list(struct list_head *page_list,
> > > struct scan_control *sc,
> > > enum pageout_io sync_writeback)
> > > {
> > > - LIST_HEAD(ret_pages);
> > > LIST_HEAD(free_pages);
> > > - int pgactivate = 0;
> > > + LIST_HEAD(putback_pages);
> > > + LIST_HEAD(dirty_pages);
> > > + int pgactivate;
> > > + int dirty_isolated = 0;
> > > + unsigned long nr_dirty;
> > > unsigned long nr_reclaimed = 0;
> > >
> > > + pgactivate = 0;
> > > cond_resched();
> > >
> > > +restart_dirty:
> > > + nr_dirty = 0;
> > > while (!list_empty(page_list)) {
> > > enum page_references references;
> > > struct address_space *mapping;
> > > @@ -741,7 +805,19 @@ static unsigned long shrink_page_list(struct list_head *page_list,
> > > }
> > > }
> > >
> > > - if (PageDirty(page)) {
> > > + if (PageDirty(page)) {
> > > + /*
> > > + * If the caller cannot writeback pages, dirty pages
> > > + * are put on a separate list for cleaning by either
> > > + * a flusher thread or kswapd
> > > + */
> > > + if (!reclaim_can_writeback(sc, page)) {
> > > + list_add(&page->lru, &dirty_pages);
> > > + unlock_page(page);
> > > + nr_dirty++;
> > > + goto keep_dirty;
> > > + }
> > > +
> > > if (references == PAGEREF_RECLAIM_CLEAN)
> > > goto keep_locked;
> > > if (!may_enter_fs)
> > > @@ -852,13 +928,39 @@ activate_locked:
> > > keep_locked:
> > > unlock_page(page);
> > > keep:
> > > - list_add(&page->lru, &ret_pages);
> > > + list_add(&page->lru, &putback_pages);
> > > +keep_dirty:
> > > VM_BUG_ON(PageLRU(page) || PageUnevictable(page));
> > > }
> > >
> > > + if (dirty_isolated < MAX_SWAP_CLEAN_WAIT && !list_empty(&dirty_pages)) {
> > > + /*
> > > + * Wakeup a flusher thread to clean at least as many dirty
> > > + * pages as encountered by direct reclaim. Wait on congestion
> > > + * to throttle processes cleaning dirty pages
> > > + */
> > > + wakeup_flusher_threads(nr_dirty);
> > > + congestion_wait(BLK_RW_ASYNC, HZ/10);
> > > +
> > > + /*
> > > + * As lumpy reclaim and memcg targets specific pages, wait on
> > > + * them to be cleaned and try reclaim again.
> > > + */
> > > + if (sync_writeback == PAGEOUT_IO_SYNC ||
> > > + sc->mem_cgroup != NULL) {
> > > + dirty_isolated++;
> > > + list_splice(&dirty_pages, page_list);
> > > + INIT_LIST_HEAD(&dirty_pages);
> > > + goto restart_dirty;
> > > + }
> > > + }
> >
> > I think it would turn out more natural to just return dirty pages on
> > page_list and have the whole looping logic in shrink_inactive_list().
> >
> > Mixing dirty pages with other 'please try again' pages is probably not
> > so bad anyway, it means we could retry all temporary unavailable pages
> > instead of twiddling thumbs over that particular bunch of pages until
> > the flushers catch up.
> >
> > What do you think?
> >
>
> It's worth considering! It won't be very tidy but it's workable. The reason
> it is not tidy is that dirty pages and pages that couldn't be paged will be
> on the same list so they whole lot will need to be recycled. We'd record in
> scan_control though that there were pages that need to be retried and loop
> based on that value. That is managable though.
Recycling all of them is what I had in mind, yeah. But...
> The reason why I did it this way was because of lumpy reclaim and memcg
> requiring specific pages. I considered lumpy reclaim to be the more common
> case. In that case, it's removing potentially a large number of pages from
> the LRU that are contiguous. If some of those are dirty and it selects more
> contiguous ranges for reclaim, I'd worry that lumpy reclaim would trash the
> system even worse than it currently does when the system is under load. Hence,
> this wait and retry loop is done instead of returning and isolating more pages.
I think here we missed each other. I don't want the loop to be _that_
much more in the outer scope that isolation is repeated as well. What
I had in mind is the attached patch. It is not tested and hacked up
rather quickly due to time constraints, sorry, but you should get the
idea. I hope I did not miss anything fundamental.
Note that since only kswapd enters pageout() anymore, everything
depending on PAGEOUT_IO_SYNC in there is moot, since there are no sync
cycles for kswapd. Just to mitigate the WTF-count on the patch :-)
Hannes
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
@@ -386,21 +386,17 @@ static pageout_t pageout(struct page *pa
ClearPageReclaim(page);
return PAGE_ACTIVATE;
}
-
- /*
- * Wait on writeback if requested to. This happens when
- * direct reclaiming a large contiguous area and the
- * first attempt to free a range of pages fails.
- */
- if (PageWriteback(page) && sync_writeback == PAGEOUT_IO_SYNC)
- wait_on_page_writeback(page);
-
if (!PageWriteback(page)) {
/* synchronous write or broken a_ops? */
ClearPageReclaim(page);
}
trace_mm_vmscan_writepage(page,
page_is_file_cache(page),
+ /*
+ * Humm. Only kswapd comes here and for
+ * kswapd there never is a PAGEOUT_IO_SYNC
+ * cycle...
+ */
sync_writeback == PAGEOUT_IO_SYNC);
inc_zone_page_state(page, NR_VMSCAN_WRITE);
return PAGE_SUCCESS;
@@ -643,12 +639,14 @@ static noinline_for_stack void free_page
* shrink_page_list() returns the number of reclaimed pages
*/
static unsigned long shrink_page_list(struct list_head *page_list,
- struct scan_control *sc,
- enum pageout_io sync_writeback)
+ struct scan_control *sc,
+ enum pageout_io sync_writeback,
+ int *dirty_seen)
{
LIST_HEAD(ret_pages);
LIST_HEAD(free_pages);
int pgactivate = 0;
+ unsigned long nr_dirty = 0;
unsigned long nr_reclaimed = 0;
cond_resched();
@@ -657,7 +655,7 @@ static unsigned long shrink_page_list(st
enum page_references references;
struct address_space *mapping;
struct page *page;
- int may_enter_fs;
+ int may_pageout;
cond_resched();
@@ -681,10 +679,15 @@ static unsigned long shrink_page_list(st
if (page_mapped(page) || PageSwapCache(page))
sc->nr_scanned++;
- may_enter_fs = (sc->gfp_mask & __GFP_FS) ||
+ /*
+ * To prevent stack overflows, only kswapd can enter
+ * the filesystem. Swap IO is always fine (for now).
+ */
+ may_pageout = current_is_kswapd() ||
(PageSwapCache(page) && (sc->gfp_mask & __GFP_IO));
if (PageWriteback(page)) {
+ int may_wait;
/*
* Synchronous reclaim is performed in two passes,
* first an asynchronous pass over the list to
@@ -693,7 +696,8 @@ static unsigned long shrink_page_list(st
* for any page for which writeback has already
* started.
*/
- if (sync_writeback == PAGEOUT_IO_SYNC && may_enter_fs)
+ may_wait = (sc->gfp_mask & __GFP_FS) || may_pageout;
+ if (sync_writeback == PAGEOUT_IO_SYNC && may_wait)
wait_on_page_writeback(page);
else
goto keep_locked;
@@ -719,7 +723,7 @@ static unsigned long shrink_page_list(st
goto keep_locked;
if (!add_to_swap(page))
goto activate_locked;
- may_enter_fs = 1;
+ may_pageout = 1;
}
mapping = page_mapping(page);
@@ -742,9 +746,11 @@ static unsigned long shrink_page_list(st
}
if (PageDirty(page)) {
+ nr_dirty++;
+
if (references == PAGEREF_RECLAIM_CLEAN)
goto keep_locked;
- if (!may_enter_fs)
+ if (!may_pageout)
goto keep_locked;
if (!sc->may_writepage)
goto keep_locked;
@@ -860,6 +866,7 @@ keep:
list_splice(&ret_pages, page_list);
count_vm_events(PGACTIVATE, pgactivate);
+ *dirty_seen = nr_dirty;
return nr_reclaimed;
}
@@ -1232,6 +1239,9 @@ static noinline_for_stack void update_is
reclaim_stat->recent_scanned[1] += *nr_file;
}
+/* Direct lumpy reclaim waits up to 5 seconds for background cleaning */
+#define MAX_SWAP_CLEAN_WAIT 50
+
/*
* shrink_inactive_list() is a helper for shrink_zone(). It returns the number
* of reclaimed pages
@@ -1247,6 +1257,7 @@ shrink_inactive_list(unsigned long nr_to
unsigned long nr_active;
unsigned long nr_anon;
unsigned long nr_file;
+ unsigned long nr_dirty;
while (unlikely(too_many_isolated(zone, file, sc))) {
congestion_wait(BLK_RW_ASYNC, HZ/10);
@@ -1295,26 +1306,34 @@ shrink_inactive_list(unsigned long nr_to
spin_unlock_irq(&zone->lru_lock);
- nr_reclaimed = shrink_page_list(&page_list, sc, PAGEOUT_IO_ASYNC);
-
+ nr_reclaimed = shrink_page_list(&page_list, sc,
+ PAGEOUT_IO_ASYNC,
+ &nr_dirty);
/*
* If we are direct reclaiming for contiguous pages and we do
* not reclaim everything in the list, try again and wait
* for IO to complete. This will stall high-order allocations
* but that should be acceptable to the caller
*/
- if (nr_reclaimed < nr_taken && !current_is_kswapd() &&
- sc->lumpy_reclaim_mode) {
- congestion_wait(BLK_RW_ASYNC, HZ/10);
+ if (!current_is_kswapd() && sc->lumpy_reclaim_mode || sc->mem_cgroup) {
+ int dirty_retry = MAX_SWAP_CLEAN_WAIT;
- /*
- * The attempt at page out may have made some
- * of the pages active, mark them inactive again.
- */
- nr_active = clear_active_flags(&page_list, NULL);
- count_vm_events(PGDEACTIVATE, nr_active);
+ while (nr_reclaimed < nr_taken && nr_dirty && dirty_retry--) {
+ wakeup_flusher_threads(nr_dirty);
+ congestion_wait(BLK_RW_ASYNC, HZ/10);
+ /*
+ * The attempt at page out may have made some
+ * of the pages active, mark them inactive again.
+ *
+ * Humm. Still needed?
+ */
+ nr_active = clear_active_flags(&page_list, NULL);
+ count_vm_events(PGDEACTIVATE, nr_active);
- nr_reclaimed += shrink_page_list(&page_list, sc, PAGEOUT_IO_SYNC);
+ nr_reclaimed += shrink_page_list(&page_list, sc,
+ PAGEOUT_IO_SYNC,
+ &nr_dirty);
+ }
}
local_irq_disable();
--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org. For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>
next prev parent reply other threads:[~2010-07-20 22:03 UTC|newest]
Thread overview: 87+ messages / expand[flat|nested] mbox.gz Atom feed top
2010-07-19 13:11 [PATCH 0/8] Reduce writeback from page reclaim context V4 Mel Gorman
2010-07-19 13:11 ` [PATCH 1/8] vmscan: tracing: Roll up of patches currently in mmotm Mel Gorman
2010-07-19 13:11 ` [PATCH 2/8] vmscan: tracing: Update trace event to track if page reclaim IO is for anon or file pages Mel Gorman
2010-07-19 13:24 ` Rik van Riel
2010-07-19 14:15 ` Christoph Hellwig
2010-07-19 14:24 ` Mel Gorman
2010-07-19 14:26 ` Christoph Hellwig
2010-07-19 13:11 ` [PATCH 3/8] vmscan: tracing: Update post-processing script to distinguish between anon and file IO from page reclaim Mel Gorman
2010-07-19 13:32 ` Rik van Riel
2010-07-19 13:11 ` [PATCH 4/8] vmscan: Do not writeback filesystem pages in direct reclaim Mel Gorman
2010-07-19 14:19 ` Christoph Hellwig
2010-07-19 14:26 ` Mel Gorman
2010-07-19 18:25 ` Rik van Riel
2010-07-19 22:14 ` Johannes Weiner
2010-07-20 13:45 ` Mel Gorman
2010-07-20 22:02 ` Johannes Weiner [this message]
2010-07-21 11:36 ` Johannes Weiner
2010-07-21 11:52 ` Mel Gorman
2010-07-21 12:01 ` KAMEZAWA Hiroyuki
2010-07-21 14:27 ` Mel Gorman
2010-07-21 23:57 ` KAMEZAWA Hiroyuki
2010-07-22 9:19 ` Mel Gorman
2010-07-22 9:22 ` KAMEZAWA Hiroyuki
2010-07-21 13:04 ` Johannes Weiner
2010-07-21 13:38 ` Mel Gorman
2010-07-21 14:28 ` Johannes Weiner
2010-07-21 14:31 ` Mel Gorman
2010-07-21 14:39 ` Johannes Weiner
2010-07-21 15:06 ` Mel Gorman
2010-07-26 8:29 ` Wu Fengguang
2010-07-26 9:12 ` Mel Gorman
2010-07-26 11:19 ` Wu Fengguang
2010-07-26 12:53 ` Mel Gorman
2010-07-26 13:03 ` Wu Fengguang
2010-07-19 13:11 ` [PATCH 5/8] fs,btrfs: Allow kswapd to writeback pages Mel Gorman
2010-07-19 18:27 ` Rik van Riel
2010-07-19 13:11 ` [PATCH 6/8] fs,xfs: " Mel Gorman
2010-07-19 14:20 ` Christoph Hellwig
2010-07-19 14:43 ` Mel Gorman
2010-07-19 13:11 ` [PATCH 7/8] writeback: sync old inodes first in background writeback Mel Gorman
2010-07-19 14:21 ` Christoph Hellwig
2010-07-19 14:40 ` Mel Gorman
2010-07-19 14:48 ` Christoph Hellwig
2010-07-22 8:52 ` Wu Fengguang
2010-07-22 9:02 ` Wu Fengguang
2010-07-22 9:21 ` Wu Fengguang
2010-07-22 10:48 ` Mel Gorman
2010-07-23 9:45 ` Wu Fengguang
2010-07-23 10:57 ` Mel Gorman
2010-07-23 11:49 ` Wu Fengguang
2010-07-23 12:20 ` Wu Fengguang
2010-07-25 10:43 ` KOSAKI Motohiro
2010-07-25 12:03 ` Minchan Kim
2010-07-26 3:27 ` Wu Fengguang
2010-07-26 4:11 ` Minchan Kim
2010-07-26 4:37 ` Wu Fengguang
2010-07-26 16:30 ` Minchan Kim
2010-07-26 22:48 ` Wu Fengguang
2010-07-26 3:08 ` Wu Fengguang
2010-07-26 3:11 ` Rik van Riel
2010-07-26 3:17 ` Wu Fengguang
2010-07-22 15:34 ` Minchan Kim
2010-07-23 11:59 ` Wu Fengguang
2010-07-22 9:42 ` Mel Gorman
2010-07-23 8:33 ` Wu Fengguang
2010-07-22 1:13 ` Wu Fengguang
2010-07-19 18:43 ` Rik van Riel
2010-07-19 13:11 ` [PATCH 8/8] vmscan: Kick flusher threads to clean pages when reclaim is encountering dirty pages Mel Gorman
2010-07-19 14:23 ` Christoph Hellwig
2010-07-19 14:37 ` Mel Gorman
2010-07-19 22:48 ` Johannes Weiner
2010-07-20 14:10 ` Mel Gorman
2010-07-20 22:05 ` Johannes Weiner
2010-07-19 18:59 ` Rik van Riel
2010-07-19 22:26 ` Johannes Weiner
2010-07-26 7:28 ` Wu Fengguang
2010-07-26 9:26 ` Mel Gorman
2010-07-26 11:27 ` Wu Fengguang
2010-07-26 12:57 ` Mel Gorman
2010-07-26 13:10 ` Wu Fengguang
2010-07-27 13:35 ` Mel Gorman
2010-07-27 14:24 ` Wu Fengguang
2010-07-27 14:34 ` Wu Fengguang
2010-07-27 14:40 ` Mel Gorman
2010-07-27 14:55 ` Wu Fengguang
2010-07-27 14:38 ` Mel Gorman
2010-07-27 15:21 ` Wu Fengguang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20100720220218.GE16031@cmpxchg.org \
--to=hannes@cmpxchg.org \
--cc=aarcange@redhat.com \
--cc=akpm@linux-foundation.org \
--cc=chris.mason@oracle.com \
--cc=david@fromorbit.com \
--cc=fengguang.wu@intel.com \
--cc=hch@infradead.org \
--cc=kamezawa.hiroyu@jp.fujitsu.com \
--cc=kosaki.motohiro@jp.fujitsu.com \
--cc=linux-fsdevel@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mel@csn.ul.ie \
--cc=npiggin@suse.de \
--cc=riel@redhat.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox