From: Alok Rathore <alok.rathore@samsung.com>
To: Bharata B Rao <bharata@amd.com>
Cc: linux-kernel@vger.kernel.org, linux-mm@kvack.org,
Jonathan.Cameron@huawei.com, dave.hansen@intel.com,
gourry@gourry.net, mgorman@techsingularity.net, mingo@redhat.com,
peterz@infradead.org, raghavendra.kt@amd.com, riel@surriel.com,
rientjes@google.com, sj@kernel.org, weixugc@google.com,
willy@infradead.org, ying.huang@linux.alibaba.com,
ziy@nvidia.com, dave@stgolabs.net, nifan.cxl@gmail.com,
xuezhengchu@huawei.com, yiannis@zptcorp.com,
akpm@linux-foundation.org, david@redhat.com, byungchul@sk.com,
kinseyho@google.com, joshua.hahnjy@gmail.com, yuanchu@google.com,
balbirs@nvidia.com, shivankg@amd.com, alokrathore20@gmail.com,
gost.dev@samsung.com, cpgs@samsung.com
Subject: Re: [RFC PATCH v4 8/9] mm: sched: Move hot page promotion from NUMAB=2 to pghot tracking
Date: Mon, 22 Dec 2025 15:56:55 +0530 [thread overview]
Message-ID: <1983025922.01766400002783.JavaMail.epsvc@epcpadp1new> (raw)
In-Reply-To: <20251206101423.5004-9-bharata@amd.com>
[-- Attachment #1: Type: text/plain, Size: 6607 bytes --]
On 06/12/25 03:44PM, Bharata B Rao wrote:
>Currently hot page promotion (NUMA_BALANCING_MEMORY_TIERING
>mode of NUMA Balancing) does hot page detection (via hint faults),
>hot page classification and eventual promotion, all by itself and
>sits within the scheduler.
>
>With the new hot page tracking and promotion mechanism being
>available, NUMA Balancing can limit itself to detection of
>hot pages (via hint faults) and off-load rest of the
>functionality to the common hot page tracking system.
>
>pghot_record_access(PGHOT_HINT_FAULT) API is used to feed the
>hot page info. In addition, the migration rate limiting and
>dynamic threshold logic are moved to kmigrated so that the same
>can be used for hot pages reported by other sources too.
>
>Signed-off-by: Bharata B Rao <bharata@amd.com>
<snip>
>--- a/mm/pghot.c
>+++ b/mm/pghot.c
>@@ -12,6 +12,9 @@
> * the hot pages. kmigrated runs for each lower tier node. It iterates
> * over the node's PFNs and migrates pages marked for migration into
> * their targeted nodes.
>+ *
>+ * Migration rate-limiting and dynamic threshold logic implementations
>+ * were moved from NUMA Balancing mode 2.
> */
> #include <linux/mm.h>
> #include <linux/migrate.h>
>@@ -25,6 +28,8 @@ static unsigned int pghot_freq_threshold = PGHOT_DEFAULT_FREQ_THRESHOLD;
> static unsigned int kmigrated_sleep_ms = KMIGRATED_DEFAULT_SLEEP_MS;
> static unsigned int kmigrated_batch_nr = KMIGRATED_DEFAULT_BATCH_NR;
>
>+/* Restrict the NUMA promotion throughput (MB/s) for each target node. */
>+static unsigned int sysctl_pghot_promote_rate_limit = 65536;
> static unsigned int sysctl_pghot_freq_window = PGHOT_DEFAULT_FREQ_WINDOW;
>
> static DEFINE_STATIC_KEY_FALSE(pghot_src_hwhints);
>@@ -43,6 +48,14 @@ static const struct ctl_table pghot_sysctls[] = {
> .proc_handler = proc_dointvec_minmax,
> .extra1 = SYSCTL_ZERO,
> },
>+ {
>+ .procname = "pghot_promote_rate_limit_MBps",
>+ .data = &sysctl_pghot_promote_rate_limit,
>+ .maxlen = sizeof(unsigned int),
>+ .mode = 0644,
>+ .proc_handler = proc_dointvec_minmax,
>+ .extra1 = SYSCTL_ZERO,
>+ },
> };
> #endif
>
>@@ -137,8 +150,13 @@ int pghot_record_access(unsigned long pfn, int nid, int src, unsigned long now)
> old_freq = (hotness >> PGHOT_FREQ_SHIFT) & PGHOT_FREQ_MASK;
> old_time = (hotness >> PGHOT_TIME_SHIFT) & PGHOT_TIME_MASK;
>
>- if (((time - old_time) > msecs_to_jiffies(sysctl_pghot_freq_window))
>- || (nid != NUMA_NO_NODE && old_nid != nid))
>+ /*
>+ * Bypass the new window logic for NUMA hint fault source
>+ * as it is too slow in reporting accesses.
>+ * TODO: Fix this.
>+ */
>+ if ((((time - old_time) > msecs_to_jiffies(sysctl_pghot_freq_window))
>+ && (src != PGHOT_HINT_FAULT)) || (nid != NUMA_NO_NODE && old_nid != nid))
> new_window = true;
>
> if (new_window)
>@@ -166,6 +184,110 @@ int pghot_record_access(unsigned long pfn, int nid, int src, unsigned long now)
> return 0;
> }
>
>+/*
>+ * For memory tiering mode, if there are enough free pages (more than
>+ * enough watermark defined here) in fast memory node, to take full
>+ * advantage of fast memory capacity, all recently accessed slow
>+ * memory pages will be migrated to fast memory node without
>+ * considering hot threshold.
>+ */
>+static bool pgdat_free_space_enough(struct pglist_data *pgdat)
>+{
>+ int z;
>+ unsigned long enough_wmark;
>+
>+ enough_wmark = max(1UL * 1024 * 1024 * 1024 >> PAGE_SHIFT,
>+ pgdat->node_present_pages >> 4);
>+ for (z = pgdat->nr_zones - 1; z >= 0; z--) {
>+ struct zone *zone = pgdat->node_zones + z;
>+
>+ if (!populated_zone(zone))
>+ continue;
>+
>+ if (zone_watermark_ok(zone, 0,
>+ promo_wmark_pages(zone) + enough_wmark,
>+ ZONE_MOVABLE, 0))
>+ return true;
>+ }
>+ return false;
>+}
>+
>+/*
>+ * For memory tiering mode, too high promotion/demotion throughput may
>+ * hurt application latency. So we provide a mechanism to rate limit
>+ * the number of pages that are tried to be promoted.
>+ */
>+static bool kmigrated_promotion_rate_limit(struct pglist_data *pgdat, unsigned long rate_limit,
>+ int nr, unsigned long now_ms)
>+{
>+ unsigned long nr_cand;
>+ unsigned int start;
>+
>+ mod_node_page_state(pgdat, PGPROMOTE_CANDIDATE, nr);
>+ nr_cand = node_page_state(pgdat, PGPROMOTE_CANDIDATE);
>+ start = pgdat->nbp_rl_start;
>+ if (now_ms - start > MSEC_PER_SEC &&
>+ cmpxchg(&pgdat->nbp_rl_start, start, now_ms) == start)
>+ pgdat->nbp_rl_nr_cand = nr_cand;
>+ if (nr_cand - pgdat->nbp_rl_nr_cand >= rate_limit)
>+ return true;
>+ return false;
>+}
>+
>+static void kmigrated_promotion_adjust_threshold(struct pglist_data *pgdat,
>+ unsigned long rate_limit, unsigned int ref_th,
>+ unsigned long now_ms)
>+{
>+ unsigned int start, th_period, unit_th, th;
>+ unsigned long nr_cand, ref_cand, diff_cand;
>+
>+ th_period = KMIGRATED_PROMOTION_THRESHOLD_WINDOW;
>+ start = pgdat->nbp_th_start;
>+ if (now_ms - start > th_period &&
>+ cmpxchg(&pgdat->nbp_th_start, start, now_ms) == start) {
>+ ref_cand = rate_limit *
>+ KMIGRATED_PROMOTION_THRESHOLD_WINDOW / MSEC_PER_SEC;
>+ nr_cand = node_page_state(pgdat, PGPROMOTE_CANDIDATE);
>+ diff_cand = nr_cand - pgdat->nbp_th_nr_cand;
>+ unit_th = ref_th * 2 / KMIGRATED_MIGRATION_ADJUST_STEPS;
>+ th = pgdat->nbp_threshold ? : ref_th;
>+ if (diff_cand > ref_cand * 11 / 10)
>+ th = max(th - unit_th, unit_th);
>+ else if (diff_cand < ref_cand * 9 / 10)
>+ th = min(th + unit_th, ref_th * 2);
>+ pgdat->nbp_th_nr_cand = nr_cand;
>+ pgdat->nbp_threshold = th;
>+ }
>+}
>+
>+static bool kmigrated_should_migrate_memory(unsigned long nr_pages, unsigned long nid,
>+ unsigned long time)
>+{
>+ struct pglist_data *pgdat;
>+ unsigned long rate_limit;
>+ unsigned int th, def_th;
>+ unsigned long now = jiffies;
now = jiffies & PGHOT_TIME_MASK;
>+ unsigned long now_ms = jiffies_to_msecs(now);
>+
>+ pgdat = NODE_DATA(nid);
>+ if (pgdat_free_space_enough(pgdat)) {
>+ /* workload changed, reset hot threshold */
>+ pgdat->nbp_threshold = 0;
>+ mod_node_page_state(pgdat, PGPROMOTE_CANDIDATE_NRL, nr_pages);
>+ return true;
>+ }
>+
>+ def_th = sysctl_pghot_freq_window;
>+ rate_limit = MB_TO_PAGES(sysctl_pghot_promote_rate_limit);
>+ kmigrated_promotion_adjust_threshold(pgdat, rate_limit, def_th, now_ms);
>+
>+ th = pgdat->nbp_threshold ? : def_th;
>+ if (jiffies_to_msecs(now - time) >= th)
Setting time in pfn hotness using PGHOT_TIME_MASK in pghot_record_access(). Therefore
here also it should be calculated using PGHOT_TIME_MASK. Then it'll be right comparision.
Regards,
Alok Rathore
[-- Attachment #2: Type: text/plain, Size: 0 bytes --]
next prev parent reply other threads:[~2025-12-22 10:40 UTC|newest]
Thread overview: 12+ messages / expand[flat|nested] mbox.gz Atom feed top
2025-12-06 10:14 [RFC PATCH v4 0/9] mm: Hot page tracking and promotion infrastructure Bharata B Rao
2025-12-06 10:14 ` [RFC PATCH v4 1/9] mm: migrate: Allow misplaced migration without VMA too Bharata B Rao
2025-12-06 10:14 ` [RFC PATCH v4 2/9] migrate: implement migrate_misplaced_folios_batch Bharata B Rao
2025-12-06 10:14 ` [RFC PATCH v4 3/9] mm: Hot page tracking and promotion Bharata B Rao
[not found] ` <CGME20251222101745epcas5p43ca3a6a59efe996cd62769e8d57bb81d@epcas5p4.samsung.com>
2025-12-22 10:17 ` Alok Rathore
2025-12-06 10:14 ` [RFC PATCH v4 4/9] x86: ibs: In-kernel IBS driver for memory access profiling Bharata B Rao
2025-12-06 10:14 ` [RFC PATCH v4 5/9] x86: ibs: Enable IBS profiling for memory accesses Bharata B Rao
2025-12-06 10:14 ` [RFC PATCH v4 6/9] mm: mglru: generalize page table walk Bharata B Rao
2025-12-06 10:14 ` [RFC PATCH v4 7/9] mm: klruscand: use mglru scanning for page promotion Bharata B Rao
2025-12-06 10:14 ` [RFC PATCH v4 8/9] mm: sched: Move hot page promotion from NUMAB=2 to pghot tracking Bharata B Rao
[not found] ` <CGME20251222102716epcas5p45d0893afb074ef3fa4be0c912cd0e237@epcas5p4.samsung.com>
2025-12-22 10:26 ` Alok Rathore [this message]
2025-12-06 10:14 ` [RFC PATCH v4 9/9] mm: pghot: Add folio_mark_accessed() as hotness source Bharata B Rao
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=1983025922.01766400002783.JavaMail.epsvc@epcpadp1new \
--to=alok.rathore@samsung.com \
--cc=Jonathan.Cameron@huawei.com \
--cc=akpm@linux-foundation.org \
--cc=alokrathore20@gmail.com \
--cc=balbirs@nvidia.com \
--cc=bharata@amd.com \
--cc=byungchul@sk.com \
--cc=cpgs@samsung.com \
--cc=dave.hansen@intel.com \
--cc=dave@stgolabs.net \
--cc=david@redhat.com \
--cc=gost.dev@samsung.com \
--cc=gourry@gourry.net \
--cc=joshua.hahnjy@gmail.com \
--cc=kinseyho@google.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mgorman@techsingularity.net \
--cc=mingo@redhat.com \
--cc=nifan.cxl@gmail.com \
--cc=peterz@infradead.org \
--cc=raghavendra.kt@amd.com \
--cc=riel@surriel.com \
--cc=rientjes@google.com \
--cc=shivankg@amd.com \
--cc=sj@kernel.org \
--cc=weixugc@google.com \
--cc=willy@infradead.org \
--cc=xuezhengchu@huawei.com \
--cc=yiannis@zptcorp.com \
--cc=ying.huang@linux.alibaba.com \
--cc=yuanchu@google.com \
--cc=ziy@nvidia.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox