From: Barry Song <baohua@kernel.org>
To: Zhang Peng <zippermonkey@icloud.com>
Cc: Andrew Morton <akpm@linux-foundation.org>,
David Hildenbrand <david@kernel.org>,
Lorenzo Stoakes <ljs@kernel.org>,
"Liam R. Howlett" <Liam.Howlett@oracle.com>,
Vlastimil Babka <vbabka@kernel.org>,
Mike Rapoport <rppt@kernel.org>,
Suren Baghdasaryan <surenb@google.com>,
Michal Hocko <mhocko@suse.com>,
Johannes Weiner <hannes@cmpxchg.org>,
Qi Zheng <zhengqi.arch@bytedance.com>,
Shakeel Butt <shakeel.butt@linux.dev>,
Axel Rasmussen <axelrasmussen@google.com>,
Yuanchu Xie <yuanchu@google.com>, Wei Xu <weixugc@google.com>,
Michal Hocko <mhocko@kernel.org>,
linux-mm@kvack.org, linux-kernel@vger.kernel.org,
Kairui Song <kasong@tencent.com>,
Zhang Peng <bruzzhang@tencent.com>
Subject: Re: [PATCH v3 3/5] mm/vmscan: extract folio_free() and pageout_one()
Date: Sat, 11 Apr 2026 08:24:27 +0800 [thread overview]
Message-ID: <CAGsJ_4xvnBSRyi9OkRwfRDDM6QcxUfhdrf32R45+GMJ=ZpEerQ@mail.gmail.com> (raw)
In-Reply-To: <20260410-batch-tlb-flush-v3-3-ff0b9d3a351a@icloud.com>
On Fri, Apr 10, 2026 at 8:47 PM Zhang Peng <zippermonkey@icloud.com> wrote:
>
> From: Zhang Peng <bruzzhang@tencent.com>
>
> shrink_folio_list() contains two large self-contained sections:
> the pageout() dispatch state machine and the folio-freeing path
> (buffer release, lazyfree, __remove_mapping, folio_batch). Extract
> them into pageout_one() and folio_free() respectively to reduce the
> size of shrink_folio_list() and make each step independently readable.
This one looks good, but:
>
> No functional change
>
> Suggested-by: Kairui Song <kasong@tencent.com>
> Signed-off-by: Zhang Peng <bruzzhang@tencent.com>
> ---
> mm/vmscan.c | 270 ++++++++++++++++++++++++++++++++++--------------------------
> 1 file changed, 155 insertions(+), 115 deletions(-)
>
> diff --git a/mm/vmscan.c b/mm/vmscan.c
> index 0860a48d5bf3..c8ff742ed891 100644
> --- a/mm/vmscan.c
> +++ b/mm/vmscan.c
> @@ -1070,6 +1070,153 @@ static void folio_active_bounce(struct folio *folio, struct reclaim_stat *stat,
> }
> }
>
> +static bool folio_free(struct folio *folio, struct folio_batch *free_folios,
> + struct scan_control *sc, struct reclaim_stat *stat)
> +{
> + unsigned int nr_pages = folio_nr_pages(folio);
> + struct address_space *mapping = folio_mapping(folio);
> +
> + /*
> + * If the folio has buffers, try to free the buffer
> + * mappings associated with this folio. If we succeed
> + * we try to free the folio as well.
> + *
> + * We do this even if the folio is dirty.
> + * filemap_release_folio() does not perform I/O, but it
> + * is possible for a folio to have the dirty flag set,
> + * but it is actually clean (all its buffers are clean).
> + * This happens if the buffers were written out directly,
> + * with submit_bh(). ext3 will do this, as well as
> + * the blockdev mapping. filemap_release_folio() will
> + * discover that cleanness and will drop the buffers
> + * and mark the folio clean - it can be freed.
> + *
> + * Rarely, folios can have buffers and no ->mapping.
> + * These are the folios which were not successfully
> + * invalidated in truncate_cleanup_folio(). We try to
> + * drop those buffers here and if that worked, and the
> + * folio is no longer mapped into process address space
> + * (refcount == 1) it can be freed. Otherwise, leave
> + * the folio on the LRU so it is swappable.
> + */
> + if (folio_needs_release(folio)) {
> + if (!filemap_release_folio(folio, sc->gfp_mask)) {
> + folio_active_bounce(folio, stat, nr_pages);
> + return false;
> + }
> +
> + if (!mapping && folio_ref_count(folio) == 1) {
> + folio_unlock(folio);
> + if (folio_put_testzero(folio))
> + goto free_it;
> + else {
> + /*
> + * rare race with speculative reference.
> + * the speculative reference will free
> + * this folio shortly, so we may
> + * increment nr_reclaimed here (and
> + * leave it off the LRU).
> + */
> + stat->nr_reclaimed += nr_pages;
> + return true;
> + }
> + }
> + }
> +
> + if (folio_test_lazyfree(folio)) {
> + /* follow __remove_mapping for reference */
> + if (!folio_ref_freeze(folio, 1))
> + return false;
> + /*
> + * The folio has only one reference left, which is
> + * from the isolation. After the caller puts the
> + * folio back on the lru and drops the reference, the
> + * folio will be freed anyway. It doesn't matter
> + * which lru it goes on. So we don't bother checking
> + * the dirty flag here.
> + */
> + count_vm_events(PGLAZYFREED, nr_pages);
> + count_memcg_folio_events(folio, PGLAZYFREED, nr_pages);
> + } else if (!mapping || !__remove_mapping(mapping, folio, true,
> + sc->target_mem_cgroup))
> + return false;
> +
> + folio_unlock(folio);
> +free_it:
> + /*
> + * Folio may get swapped out as a whole, need to account
> + * all pages in it.
> + */
> + stat->nr_reclaimed += nr_pages;
> +
> + folio_unqueue_deferred_split(folio);
> + if (folio_batch_add(free_folios, folio) == 0) {
> + mem_cgroup_uncharge_folios(free_folios);
> + try_to_unmap_flush();
> + free_unref_folios(free_folios);
> + }
> + return true;
> +}
> +
> +static void pageout_one(struct folio *folio, struct list_head *ret_folios,
> + struct folio_batch *free_folios,
> + struct scan_control *sc, struct reclaim_stat *stat,
> + struct swap_iocb **plug, struct list_head *folio_list)
> +{
> + struct address_space *mapping = folio_mapping(folio);
> + unsigned int nr_pages = folio_nr_pages(folio);
> +
> + switch (pageout(folio, mapping, plug, folio_list)) {
> + case PAGE_ACTIVATE:
> + /*
> + * If shmem folio is split when writeback to swap,
> + * the tail pages will make their own pass through
> + * this function and be accounted then.
> + */
> + if (nr_pages > 1 && !folio_test_large(folio)) {
> + sc->nr_scanned -= (nr_pages - 1);
> + nr_pages = 1;
> + }
> + folio_active_bounce(folio, stat, nr_pages);
> + fallthrough;
> + case PAGE_KEEP:
> + goto locked_keepit;
> + case PAGE_SUCCESS:
> + if (nr_pages > 1 && !folio_test_large(folio)) {
> + sc->nr_scanned -= (nr_pages - 1);
> + nr_pages = 1;
> + }
> + stat->nr_pageout += nr_pages;
> +
> + if (folio_test_writeback(folio))
> + goto keepit;
> + if (folio_test_dirty(folio))
> + goto keepit;
> +
> + /*
> + * A synchronous write - probably a ramdisk. Go
> + * ahead and try to reclaim the folio.
> + */
> + if (!folio_trylock(folio))
> + goto keepit;
> + if (folio_test_dirty(folio) ||
> + folio_test_writeback(folio))
> + goto locked_keepit;
> + mapping = folio_mapping(folio);
> + fallthrough;
> + case PAGE_CLEAN:
> + ; /* try to free the folio below */
> + }
> + if (folio_free(folio, free_folios, sc, stat))
> + return;
> +locked_keepit:
> + folio_unlock(folio);
> +keepit:
> + list_add(&folio->lru, ret_folios);
> + VM_BUG_ON_FOLIO(folio_test_lru(folio) ||
> + folio_test_unevictable(folio), folio);
> +}
Can we at least move the “result” out of the function—
whether to “keep” it or not?
Can we have pageout() report its result to shrink_folio_list()?
If everything is hidden inside, it’s hard to tell what
happened to the folio.
This hides too many details that should be exposed to
shrink_folio_list(), making the reclamation flow harder
to understand.
Thanks
Barry
next prev parent reply other threads:[~2026-04-11 0:24 UTC|newest]
Thread overview: 11+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-04-10 12:43 [PATCH v3 0/5] mm: batch TLB flushing for dirty folios in vmscan Zhang Peng
2026-04-10 12:43 ` [PATCH v3 1/5] mm/vmscan: track reclaimed pages in reclaim_stat Zhang Peng
2026-04-10 23:58 ` Barry Song
2026-04-10 12:44 ` [PATCH v3 2/5] mm/vmscan: extract folio activation into folio_active_bounce() Zhang Peng
2026-04-11 0:02 ` Barry Song
2026-04-10 12:44 ` [PATCH v3 3/5] mm/vmscan: extract folio_free() and pageout_one() Zhang Peng
2026-04-11 0:24 ` Barry Song [this message]
2026-04-10 12:44 ` [PATCH v3 4/5] mm/vmscan: extract folio unmap logic into folio_try_unmap() Zhang Peng
2026-04-11 0:31 ` Barry Song
2026-04-10 12:44 ` [PATCH v3 5/5] mm/vmscan: flush TLB for every 31 folios evictions Zhang Peng
2026-04-11 0:34 ` Barry Song
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to='CAGsJ_4xvnBSRyi9OkRwfRDDM6QcxUfhdrf32R45+GMJ=ZpEerQ@mail.gmail.com' \
--to=baohua@kernel.org \
--cc=Liam.Howlett@oracle.com \
--cc=akpm@linux-foundation.org \
--cc=axelrasmussen@google.com \
--cc=bruzzhang@tencent.com \
--cc=david@kernel.org \
--cc=hannes@cmpxchg.org \
--cc=kasong@tencent.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=ljs@kernel.org \
--cc=mhocko@kernel.org \
--cc=mhocko@suse.com \
--cc=rppt@kernel.org \
--cc=shakeel.butt@linux.dev \
--cc=surenb@google.com \
--cc=vbabka@kernel.org \
--cc=weixugc@google.com \
--cc=yuanchu@google.com \
--cc=zhengqi.arch@bytedance.com \
--cc=zippermonkey@icloud.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox