linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: Barry Song <baohua@kernel.org>
To: Zhang Peng <zippermonkey@icloud.com>
Cc: Andrew Morton <akpm@linux-foundation.org>,
	David Hildenbrand <david@kernel.org>,
	 Lorenzo Stoakes <ljs@kernel.org>,
	"Liam R. Howlett" <Liam.Howlett@oracle.com>,
	 Vlastimil Babka <vbabka@kernel.org>,
	Mike Rapoport <rppt@kernel.org>,
	 Suren Baghdasaryan <surenb@google.com>,
	Michal Hocko <mhocko@suse.com>,
	 Johannes Weiner <hannes@cmpxchg.org>,
	Qi Zheng <zhengqi.arch@bytedance.com>,
	 Shakeel Butt <shakeel.butt@linux.dev>,
	Axel Rasmussen <axelrasmussen@google.com>,
	 Yuanchu Xie <yuanchu@google.com>, Wei Xu <weixugc@google.com>,
	 Michal Hocko <mhocko@kernel.org>,
	linux-mm@kvack.org, linux-kernel@vger.kernel.org,
	 Kairui Song <kasong@tencent.com>,
	Zhang Peng <bruzzhang@tencent.com>
Subject: Re: [PATCH v3 3/5] mm/vmscan: extract folio_free() and pageout_one()
Date: Sat, 11 Apr 2026 08:24:27 +0800	[thread overview]
Message-ID: <CAGsJ_4xvnBSRyi9OkRwfRDDM6QcxUfhdrf32R45+GMJ=ZpEerQ@mail.gmail.com> (raw)
In-Reply-To: <20260410-batch-tlb-flush-v3-3-ff0b9d3a351a@icloud.com>

On Fri, Apr 10, 2026 at 8:47 PM Zhang Peng <zippermonkey@icloud.com> wrote:
>
> From: Zhang Peng <bruzzhang@tencent.com>
>
> shrink_folio_list() contains two large self-contained sections:
> the pageout() dispatch state machine and the folio-freeing path
> (buffer release, lazyfree, __remove_mapping, folio_batch). Extract
> them into pageout_one() and folio_free() respectively to reduce the
> size of shrink_folio_list() and make each step independently readable.

This one looks good, but:

>
> No functional change
>
> Suggested-by: Kairui Song <kasong@tencent.com>
> Signed-off-by: Zhang Peng <bruzzhang@tencent.com>
> ---
>  mm/vmscan.c | 270 ++++++++++++++++++++++++++++++++++--------------------------
>  1 file changed, 155 insertions(+), 115 deletions(-)
>
> diff --git a/mm/vmscan.c b/mm/vmscan.c
> index 0860a48d5bf3..c8ff742ed891 100644
> --- a/mm/vmscan.c
> +++ b/mm/vmscan.c
> @@ -1070,6 +1070,153 @@ static void folio_active_bounce(struct folio *folio, struct reclaim_stat *stat,
>         }
>  }
>
> +static bool folio_free(struct folio *folio, struct folio_batch *free_folios,
> +               struct scan_control *sc, struct reclaim_stat *stat)
> +{
> +       unsigned int nr_pages = folio_nr_pages(folio);
> +       struct address_space *mapping = folio_mapping(folio);
> +
> +       /*
> +        * If the folio has buffers, try to free the buffer
> +        * mappings associated with this folio. If we succeed
> +        * we try to free the folio as well.
> +        *
> +        * We do this even if the folio is dirty.
> +        * filemap_release_folio() does not perform I/O, but it
> +        * is possible for a folio to have the dirty flag set,
> +        * but it is actually clean (all its buffers are clean).
> +        * This happens if the buffers were written out directly,
> +        * with submit_bh(). ext3 will do this, as well as
> +        * the blockdev mapping.  filemap_release_folio() will
> +        * discover that cleanness and will drop the buffers
> +        * and mark the folio clean - it can be freed.
> +        *
> +        * Rarely, folios can have buffers and no ->mapping.
> +        * These are the folios which were not successfully
> +        * invalidated in truncate_cleanup_folio().  We try to
> +        * drop those buffers here and if that worked, and the
> +        * folio is no longer mapped into process address space
> +        * (refcount == 1) it can be freed.  Otherwise, leave
> +        * the folio on the LRU so it is swappable.
> +        */
> +       if (folio_needs_release(folio)) {
> +               if (!filemap_release_folio(folio, sc->gfp_mask)) {
> +                       folio_active_bounce(folio, stat, nr_pages);
> +                       return false;
> +               }
> +
> +               if (!mapping && folio_ref_count(folio) == 1) {
> +                       folio_unlock(folio);
> +                       if (folio_put_testzero(folio))
> +                               goto free_it;
> +                       else {
> +                               /*
> +                                * rare race with speculative reference.
> +                                * the speculative reference will free
> +                                * this folio shortly, so we may
> +                                * increment nr_reclaimed here (and
> +                                * leave it off the LRU).
> +                                */
> +                               stat->nr_reclaimed += nr_pages;
> +                               return true;
> +                       }
> +               }
> +       }
> +
> +       if (folio_test_lazyfree(folio)) {
> +               /* follow __remove_mapping for reference */
> +               if (!folio_ref_freeze(folio, 1))
> +                       return false;
> +               /*
> +                * The folio has only one reference left, which is
> +                * from the isolation. After the caller puts the
> +                * folio back on the lru and drops the reference, the
> +                * folio will be freed anyway. It doesn't matter
> +                * which lru it goes on. So we don't bother checking
> +                * the dirty flag here.
> +                */
> +               count_vm_events(PGLAZYFREED, nr_pages);
> +               count_memcg_folio_events(folio, PGLAZYFREED, nr_pages);
> +       } else if (!mapping || !__remove_mapping(mapping, folio, true,
> +                                                       sc->target_mem_cgroup))
> +               return false;
> +
> +       folio_unlock(folio);
> +free_it:
> +       /*
> +        * Folio may get swapped out as a whole, need to account
> +        * all pages in it.
> +        */
> +       stat->nr_reclaimed += nr_pages;
> +
> +       folio_unqueue_deferred_split(folio);
> +       if (folio_batch_add(free_folios, folio) == 0) {
> +               mem_cgroup_uncharge_folios(free_folios);
> +               try_to_unmap_flush();
> +               free_unref_folios(free_folios);
> +       }
> +       return true;
> +}
> +
> +static void pageout_one(struct folio *folio, struct list_head *ret_folios,
> +                       struct folio_batch *free_folios,
> +                       struct scan_control *sc, struct reclaim_stat *stat,
> +                       struct swap_iocb **plug, struct list_head *folio_list)
> +{
> +       struct address_space *mapping = folio_mapping(folio);
> +       unsigned int nr_pages = folio_nr_pages(folio);
> +
> +       switch (pageout(folio, mapping, plug, folio_list)) {
> +       case PAGE_ACTIVATE:
> +               /*
> +                * If shmem folio is split when writeback to swap,
> +                * the tail pages will make their own pass through
> +                * this function and be accounted then.
> +                */
> +               if (nr_pages > 1 && !folio_test_large(folio)) {
> +                       sc->nr_scanned -= (nr_pages - 1);
> +                       nr_pages = 1;
> +               }
> +               folio_active_bounce(folio, stat, nr_pages);
> +               fallthrough;
> +       case PAGE_KEEP:
> +               goto locked_keepit;
> +       case PAGE_SUCCESS:
> +               if (nr_pages > 1 && !folio_test_large(folio)) {
> +                       sc->nr_scanned -= (nr_pages - 1);
> +                       nr_pages = 1;
> +               }
> +               stat->nr_pageout += nr_pages;
> +
> +               if (folio_test_writeback(folio))
> +                       goto keepit;
> +               if (folio_test_dirty(folio))
> +                       goto keepit;
> +
> +               /*
> +                * A synchronous write - probably a ramdisk.  Go
> +                * ahead and try to reclaim the folio.
> +                */
> +               if (!folio_trylock(folio))
> +                       goto keepit;
> +               if (folio_test_dirty(folio) ||
> +                       folio_test_writeback(folio))
> +                       goto locked_keepit;
> +               mapping = folio_mapping(folio);
> +               fallthrough;
> +       case PAGE_CLEAN:
> +               ; /* try to free the folio below */
> +       }
> +       if (folio_free(folio, free_folios, sc, stat))
> +               return;
> +locked_keepit:
> +       folio_unlock(folio);
> +keepit:
> +       list_add(&folio->lru, ret_folios);
> +       VM_BUG_ON_FOLIO(folio_test_lru(folio) ||
> +                       folio_test_unevictable(folio), folio);
> +}

Can we at least move the “result” out of the function—
whether to “keep” it or not?

Can we have pageout() report its result to shrink_folio_list()?
If everything is hidden inside, it’s hard to tell what
happened to the folio.

This hides too many details that should be exposed to
shrink_folio_list(), making the reclamation flow harder
to understand.

Thanks
Barry


  reply	other threads:[~2026-04-11  0:24 UTC|newest]

Thread overview: 11+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2026-04-10 12:43 [PATCH v3 0/5] mm: batch TLB flushing for dirty folios in vmscan Zhang Peng
2026-04-10 12:43 ` [PATCH v3 1/5] mm/vmscan: track reclaimed pages in reclaim_stat Zhang Peng
2026-04-10 23:58   ` Barry Song
2026-04-10 12:44 ` [PATCH v3 2/5] mm/vmscan: extract folio activation into folio_active_bounce() Zhang Peng
2026-04-11  0:02   ` Barry Song
2026-04-10 12:44 ` [PATCH v3 3/5] mm/vmscan: extract folio_free() and pageout_one() Zhang Peng
2026-04-11  0:24   ` Barry Song [this message]
2026-04-10 12:44 ` [PATCH v3 4/5] mm/vmscan: extract folio unmap logic into folio_try_unmap() Zhang Peng
2026-04-11  0:31   ` Barry Song
2026-04-10 12:44 ` [PATCH v3 5/5] mm/vmscan: flush TLB for every 31 folios evictions Zhang Peng
2026-04-11  0:34   ` Barry Song

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to='CAGsJ_4xvnBSRyi9OkRwfRDDM6QcxUfhdrf32R45+GMJ=ZpEerQ@mail.gmail.com' \
    --to=baohua@kernel.org \
    --cc=Liam.Howlett@oracle.com \
    --cc=akpm@linux-foundation.org \
    --cc=axelrasmussen@google.com \
    --cc=bruzzhang@tencent.com \
    --cc=david@kernel.org \
    --cc=hannes@cmpxchg.org \
    --cc=kasong@tencent.com \
    --cc=linux-kernel@vger.kernel.org \
    --cc=linux-mm@kvack.org \
    --cc=ljs@kernel.org \
    --cc=mhocko@kernel.org \
    --cc=mhocko@suse.com \
    --cc=rppt@kernel.org \
    --cc=shakeel.butt@linux.dev \
    --cc=surenb@google.com \
    --cc=vbabka@kernel.org \
    --cc=weixugc@google.com \
    --cc=yuanchu@google.com \
    --cc=zhengqi.arch@bytedance.com \
    --cc=zippermonkey@icloud.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox