From: Donet Tom <donettom@linux.ibm.com>
To: Bharata B Rao <bharata@amd.com>,
linux-kernel@vger.kernel.org, linux-mm@kvack.org
Cc: Jonathan.Cameron@huawei.com, dave.hansen@intel.com,
gourry@gourry.net, mgorman@techsingularity.net, mingo@redhat.com,
peterz@infradead.org, raghavendra.kt@amd.com, riel@surriel.com,
rientjes@google.com, sj@kernel.org, weixugc@google.com,
willy@infradead.org, ying.huang@linux.alibaba.com,
ziy@nvidia.com, dave@stgolabs.net, nifan.cxl@gmail.com,
xuezhengchu@huawei.com, yiannis@zptcorp.com,
akpm@linux-foundation.org, david@redhat.com, byungchul@sk.com,
kinseyho@google.com, joshua.hahnjy@gmail.com, yuanchu@google.com,
balbirs@nvidia.com, alok.rathore@samsung.com, shivankg@amd.com
Subject: Re: [RFC PATCH v6 2/5] mm: migrate: Add migrate_misplaced_folios_batch()
Date: Tue, 21 Apr 2026 20:55:02 +0530 [thread overview]
Message-ID: <24cd6a95-1304-4732-9273-43c73ea858b2@linux.ibm.com> (raw)
In-Reply-To: <20260323095104.238982-3-bharata@amd.com>
Hi Bharata
On 3/23/26 3:21 PM, Bharata B Rao wrote:
> From: Gregory Price <gourry@gourry.net>
>
> Tiered memory systems often require migrating multiple folios at once.
> Currently, migrate_misplaced_folio() handles only one folio per call,
> which is inefficient for batch operations. This patch introduces
> migrate_misplaced_folios_batch(), a batch variant that leverages
> migrate_pages() internally for improved performance.
>
> The caller must isolate folios beforehand using
> migrate_misplaced_folio_prepare(). On return, the folio list will be
> empty regardless of success or failure.
>
> This function will be used by pghot kmigrated thread.
>
> Signed-off-by: Gregory Price <gourry@gourry.net>
> [Rewrote commit description]
> Signed-off-by: Bharata B Rao <bharata@amd.com>
> ---
> include/linux/migrate.h | 6 ++++++
> mm/migrate.c | 48 +++++++++++++++++++++++++++++++++++++++++
> 2 files changed, 54 insertions(+)
>
> diff --git a/include/linux/migrate.h b/include/linux/migrate.h
> index d5af2b7f577b..5c1e2691cec2 100644
> --- a/include/linux/migrate.h
> +++ b/include/linux/migrate.h
> @@ -111,6 +111,7 @@ static inline void softleaf_entry_wait_on_locked(softleaf_t entry, spinlock_t *p
> int migrate_misplaced_folio_prepare(struct folio *folio,
> struct vm_area_struct *vma, int node);
> int migrate_misplaced_folio(struct folio *folio, int node);
> +int migrate_misplaced_folios_batch(struct list_head *folio_list, int node);
> #else
> static inline int migrate_misplaced_folio_prepare(struct folio *folio,
> struct vm_area_struct *vma, int node)
> @@ -121,6 +122,11 @@ static inline int migrate_misplaced_folio(struct folio *folio, int node)
> {
> return -EAGAIN; /* can't migrate now */
> }
> +static inline int migrate_misplaced_folios_batch(struct list_head *folio_list,
> + int node)
> +{
> + return -EAGAIN; /* can't migrate now */
> +}
> #endif /* CONFIG_NUMA_BALANCING */
>
> #ifdef CONFIG_MIGRATION
> diff --git a/mm/migrate.c b/mm/migrate.c
> index a15184950e65..94daec0f49ef 100644
> --- a/mm/migrate.c
> +++ b/mm/migrate.c
> @@ -2751,5 +2751,53 @@ int migrate_misplaced_folio(struct folio *folio, int node)
> BUG_ON(!list_empty(&migratepages));
> return nr_remaining ? -EAGAIN : 0;
> }
> +
> +/**
> + * migrate_misplaced_folios_batch() - Batch variant of migrate_misplaced_folio
> + * Attempts to migrate a folio list to the specified destination.
> + * @folio_list: Isolated list of folios to be batch-migrated.
> + * @node: The NUMA node ID to where the folios should be migrated.
> + *
> + * Caller is expected to have isolated the folios by calling
> + * migrate_misplaced_folio_prepare(), which will result in an
> + * elevated reference count on the folio. All the isolated folios
> + * in the list must belong to the same memcg so that NUMA_PAGE_MIGRATE
> + * stat can be attributed correctly to the memcg.
> + *
> + * This function will un-isolate the folios, drop the elevated reference
> + * and remove them from the list before returning. This is called
> + * only for batched promotion of hot pages from lower tier nodes.
> + *
> + * Return: 0 on success and -EAGAIN on failure or partial migration.
> + * On return, @folio_list will be empty regardless of success/failure.
> + */
> +int migrate_misplaced_folios_batch(struct list_head *folio_list, int node)
> +{
> + pg_data_t *pgdat = NODE_DATA(node);
> + struct mem_cgroup *memcg = NULL;
> + unsigned int nr_succeeded = 0;
> + int nr_remaining;
> +
> + if (!list_empty(folio_list)) {
>
We seem to proceed even when the list is empty. Should we instead return
early in that case?
> + struct folio *first = list_first_entry(folio_list, struct folio, lru);
> + memcg = get_mem_cgroup_from_folio(first);
I had a small question—are we ensuring that a single list contains
folios from the same memcg?
-Donet
> + }
> +
> + nr_remaining = migrate_pages(folio_list, alloc_misplaced_dst_folio,
> + NULL, node, MIGRATE_ASYNC,
> + MR_NUMA_MISPLACED, &nr_succeeded);
> + if (nr_remaining)
> + putback_movable_pages(folio_list);
> +
> + if (nr_succeeded) {
> + count_vm_numa_events(NUMA_PAGE_MIGRATE, nr_succeeded);
> + mod_node_page_state(pgdat, PGPROMOTE_SUCCESS, nr_succeeded);
> + count_memcg_events(memcg, NUMA_PAGE_MIGRATE, nr_succeeded);
> + }
> +
> + mem_cgroup_put(memcg);
> + WARN_ON(!list_empty(folio_list));
> + return nr_remaining ? -EAGAIN : 0;
> +}
> #endif /* CONFIG_NUMA_BALANCING */
> #endif /* CONFIG_NUMA */
next prev parent reply other threads:[~2026-04-21 15:25 UTC|newest]
Thread overview: 15+ messages / expand[flat|nested] mbox.gz Atom feed top
2026-03-23 9:50 [RFC PATCH v6 0/5] mm: Hot page tracking and promotion infrastructure Bharata B Rao
2026-03-23 9:51 ` [RFC PATCH v6 1/5] mm: migrate: Allow misplaced migration without VMA Bharata B Rao
2026-03-23 9:51 ` [RFC PATCH v6 2/5] mm: migrate: Add migrate_misplaced_folios_batch() Bharata B Rao
2026-03-26 5:50 ` Bharata B Rao
2026-04-21 15:25 ` Donet Tom [this message]
2026-04-21 16:05 ` Gregory Price
2026-03-23 9:51 ` [RFC PATCH v6 3/5] mm: Hot page tracking and promotion - pghot Bharata B Rao
2026-03-23 9:51 ` [RFC PATCH v6 4/5] mm: pghot: Precision mode for pghot Bharata B Rao
2026-03-26 10:41 ` Bharata B Rao
2026-03-23 9:51 ` [RFC PATCH v6 5/5] mm: sched: move NUMA balancing tiering promotion to pghot Bharata B Rao
2026-03-30 4:46 ` Bharata B Rao
2026-03-23 9:56 ` [RFC PATCH v6 0/5] mm: Hot page tracking and promotion infrastructure Bharata B Rao
2026-03-23 9:58 ` Bharata B Rao
2026-03-23 9:59 ` Bharata B Rao
2026-03-23 10:01 ` Bharata B Rao
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=24cd6a95-1304-4732-9273-43c73ea858b2@linux.ibm.com \
--to=donettom@linux.ibm.com \
--cc=Jonathan.Cameron@huawei.com \
--cc=akpm@linux-foundation.org \
--cc=alok.rathore@samsung.com \
--cc=balbirs@nvidia.com \
--cc=bharata@amd.com \
--cc=byungchul@sk.com \
--cc=dave.hansen@intel.com \
--cc=dave@stgolabs.net \
--cc=david@redhat.com \
--cc=gourry@gourry.net \
--cc=joshua.hahnjy@gmail.com \
--cc=kinseyho@google.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mgorman@techsingularity.net \
--cc=mingo@redhat.com \
--cc=nifan.cxl@gmail.com \
--cc=peterz@infradead.org \
--cc=raghavendra.kt@amd.com \
--cc=riel@surriel.com \
--cc=rientjes@google.com \
--cc=shivankg@amd.com \
--cc=sj@kernel.org \
--cc=weixugc@google.com \
--cc=willy@infradead.org \
--cc=xuezhengchu@huawei.com \
--cc=yiannis@zptcorp.com \
--cc=ying.huang@linux.alibaba.com \
--cc=yuanchu@google.com \
--cc=ziy@nvidia.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox