* [PATCH v2] mm/mglru: reset page lru tier bits when activating
@ 2024-10-17 18:15 Wei Xu
2024-10-17 18:18 ` Wei Xu
0 siblings, 1 reply; 2+ messages in thread
From: Wei Xu @ 2024-10-17 18:15 UTC (permalink / raw)
To: Yu Zhao
Cc: Andrew Morton, Axel Rasmussen, Brian Geffon,
Jan Alexander Steffens, Suleiman Souhlal, linux-mm, linux-kernel,
Wei Xu
When a folio is activated, lru_gen_add_folio() moves the folio to the
youngest generation. But unlike folio_update_gen()/folio_inc_gen(),
lru_gen_add_folio() doesn't reset the folio lru tier bits
(LRU_REFS_MASK | LRU_REFS_FLAGS). This inconsistency can affect how
pages are aged via folio_mark_accessed() (e.g. fd accesses), though
no user visible impact related to this has been detected yet.
Note that lru_gen_add_folio() cannot clear PG_workingset if the
activation is due to workingset refault, otherwise PSI accounting
will be skipped. So fix lru_gen_add_folio() to clear the lru tier
bits other than PG_workingset when activating a folio, and also
clear all the lru tier bits when a folio is activated via
folio_activate() in lru_gen_look_around().
Fixes: 018ee47f1489 ("mm: multi-gen LRU: exploit locality in rmap")
Signed-off-by: Wei Xu <weixugc@google.com>
---
include/linux/mm_inline.h | 15 ++++++++++++++-
include/linux/mmzone.h | 2 ++
mm/vmscan.c | 8 ++++----
3 files changed, 20 insertions(+), 5 deletions(-)
diff --git a/include/linux/mm_inline.h b/include/linux/mm_inline.h
index 6f801c7b36e2..355cf46a01a6 100644
--- a/include/linux/mm_inline.h
+++ b/include/linux/mm_inline.h
@@ -155,6 +155,11 @@ static inline int folio_lru_refs(struct folio *folio)
return ((flags & LRU_REFS_MASK) >> LRU_REFS_PGOFF) + workingset;
}
+static inline void folio_clear_lru_refs(struct folio *folio)
+{
+ set_mask_bits(&folio->flags, LRU_REFS_MASK | LRU_REFS_FLAGS, 0);
+}
+
static inline int folio_lru_gen(struct folio *folio)
{
unsigned long flags = READ_ONCE(folio->flags);
@@ -222,6 +227,7 @@ static inline bool lru_gen_add_folio(struct lruvec *lruvec, struct folio *folio,
{
unsigned long seq;
unsigned long flags;
+ unsigned long mask;
int gen = folio_lru_gen(folio);
int type = folio_is_file_lru(folio);
int zone = folio_zonenum(folio);
@@ -257,7 +263,14 @@ static inline bool lru_gen_add_folio(struct lruvec *lruvec, struct folio *folio,
gen = lru_gen_from_seq(seq);
flags = (gen + 1UL) << LRU_GEN_PGOFF;
/* see the comment on MIN_NR_GENS about PG_active */
- set_mask_bits(&folio->flags, LRU_GEN_MASK | BIT(PG_active), flags);
+ mask = LRU_GEN_MASK;
+ /*
+ * Don't clear PG_workingset here because it can affect PSI accounting
+ * if the activation is due to workingset refault.
+ */
+ if (folio_test_active(folio))
+ mask |= LRU_REFS_MASK | BIT(PG_referenced) | BIT(PG_active);
+ set_mask_bits(&folio->flags, mask, flags);
lru_gen_update_size(lruvec, folio, -1, gen);
/* for folio_rotate_reclaimable() */
diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h
index 17506e4a2835..96dea31fb211 100644
--- a/include/linux/mmzone.h
+++ b/include/linux/mmzone.h
@@ -403,6 +403,8 @@ enum {
NR_LRU_GEN_CAPS
};
+#define LRU_REFS_FLAGS (BIT(PG_referenced) | BIT(PG_workingset))
+
#define MIN_LRU_BATCH BITS_PER_LONG
#define MAX_LRU_BATCH (MIN_LRU_BATCH * 64)
diff --git a/mm/vmscan.c b/mm/vmscan.c
index c12f78b042f3..2d0486189804 100644
--- a/mm/vmscan.c
+++ b/mm/vmscan.c
@@ -2602,8 +2602,6 @@ static bool should_clear_pmd_young(void)
* shorthand helpers
******************************************************************************/
-#define LRU_REFS_FLAGS (BIT(PG_referenced) | BIT(PG_workingset))
-
#define DEFINE_MAX_SEQ(lruvec) \
unsigned long max_seq = READ_ONCE((lruvec)->lrugen.max_seq)
@@ -4138,8 +4136,10 @@ void lru_gen_look_around(struct page_vma_mapped_walk *pvmw)
old_gen = folio_lru_gen(folio);
if (old_gen < 0)
folio_set_referenced(folio);
- else if (old_gen != new_gen)
+ else if (old_gen != new_gen) {
+ folio_clear_lru_refs(folio);
folio_activate(folio);
+ }
}
arch_leave_lazy_mmu_mode();
@@ -4370,7 +4370,7 @@ static bool isolate_folio(struct lruvec *lruvec, struct folio *folio, struct sca
/* see the comment on MAX_NR_TIERS */
if (!folio_test_referenced(folio))
- set_mask_bits(&folio->flags, LRU_REFS_MASK | LRU_REFS_FLAGS, 0);
+ folio_clear_lru_refs(folio);
/* for shrink_folio_list() */
folio_clear_reclaim(folio);
--
2.47.0.rc1.288.g06298d1525-goog
^ permalink raw reply [flat|nested] 2+ messages in thread
* Re: [PATCH v2] mm/mglru: reset page lru tier bits when activating
2024-10-17 18:15 [PATCH v2] mm/mglru: reset page lru tier bits when activating Wei Xu
@ 2024-10-17 18:18 ` Wei Xu
0 siblings, 0 replies; 2+ messages in thread
From: Wei Xu @ 2024-10-17 18:18 UTC (permalink / raw)
To: Yu Zhao
Cc: Andrew Morton, Axel Rasmussen, Brian Geffon,
Jan Alexander Steffens, Suleiman Souhlal, linux-mm, linux-kernel
Changelog since v1:
- Addressed Yu Zhao's comments on not clearing PG_workingset.
- Updated the commit message.
On Thu, Oct 17, 2024 at 11:15 AM Wei Xu <weixugc@google.com> wrote:
>
> When a folio is activated, lru_gen_add_folio() moves the folio to the
> youngest generation. But unlike folio_update_gen()/folio_inc_gen(),
> lru_gen_add_folio() doesn't reset the folio lru tier bits
> (LRU_REFS_MASK | LRU_REFS_FLAGS). This inconsistency can affect how
> pages are aged via folio_mark_accessed() (e.g. fd accesses), though
> no user visible impact related to this has been detected yet.
>
> Note that lru_gen_add_folio() cannot clear PG_workingset if the
> activation is due to workingset refault, otherwise PSI accounting
> will be skipped. So fix lru_gen_add_folio() to clear the lru tier
> bits other than PG_workingset when activating a folio, and also
> clear all the lru tier bits when a folio is activated via
> folio_activate() in lru_gen_look_around().
>
> Fixes: 018ee47f1489 ("mm: multi-gen LRU: exploit locality in rmap")
> Signed-off-by: Wei Xu <weixugc@google.com>
> ---
> include/linux/mm_inline.h | 15 ++++++++++++++-
> include/linux/mmzone.h | 2 ++
> mm/vmscan.c | 8 ++++----
> 3 files changed, 20 insertions(+), 5 deletions(-)
>
> diff --git a/include/linux/mm_inline.h b/include/linux/mm_inline.h
> index 6f801c7b36e2..355cf46a01a6 100644
> --- a/include/linux/mm_inline.h
> +++ b/include/linux/mm_inline.h
> @@ -155,6 +155,11 @@ static inline int folio_lru_refs(struct folio *folio)
> return ((flags & LRU_REFS_MASK) >> LRU_REFS_PGOFF) + workingset;
> }
>
> +static inline void folio_clear_lru_refs(struct folio *folio)
> +{
> + set_mask_bits(&folio->flags, LRU_REFS_MASK | LRU_REFS_FLAGS, 0);
> +}
> +
> static inline int folio_lru_gen(struct folio *folio)
> {
> unsigned long flags = READ_ONCE(folio->flags);
> @@ -222,6 +227,7 @@ static inline bool lru_gen_add_folio(struct lruvec *lruvec, struct folio *folio,
> {
> unsigned long seq;
> unsigned long flags;
> + unsigned long mask;
> int gen = folio_lru_gen(folio);
> int type = folio_is_file_lru(folio);
> int zone = folio_zonenum(folio);
> @@ -257,7 +263,14 @@ static inline bool lru_gen_add_folio(struct lruvec *lruvec, struct folio *folio,
> gen = lru_gen_from_seq(seq);
> flags = (gen + 1UL) << LRU_GEN_PGOFF;
> /* see the comment on MIN_NR_GENS about PG_active */
> - set_mask_bits(&folio->flags, LRU_GEN_MASK | BIT(PG_active), flags);
> + mask = LRU_GEN_MASK;
> + /*
> + * Don't clear PG_workingset here because it can affect PSI accounting
> + * if the activation is due to workingset refault.
> + */
> + if (folio_test_active(folio))
> + mask |= LRU_REFS_MASK | BIT(PG_referenced) | BIT(PG_active);
> + set_mask_bits(&folio->flags, mask, flags);
>
> lru_gen_update_size(lruvec, folio, -1, gen);
> /* for folio_rotate_reclaimable() */
> diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h
> index 17506e4a2835..96dea31fb211 100644
> --- a/include/linux/mmzone.h
> +++ b/include/linux/mmzone.h
> @@ -403,6 +403,8 @@ enum {
> NR_LRU_GEN_CAPS
> };
>
> +#define LRU_REFS_FLAGS (BIT(PG_referenced) | BIT(PG_workingset))
> +
> #define MIN_LRU_BATCH BITS_PER_LONG
> #define MAX_LRU_BATCH (MIN_LRU_BATCH * 64)
>
> diff --git a/mm/vmscan.c b/mm/vmscan.c
> index c12f78b042f3..2d0486189804 100644
> --- a/mm/vmscan.c
> +++ b/mm/vmscan.c
> @@ -2602,8 +2602,6 @@ static bool should_clear_pmd_young(void)
> * shorthand helpers
> ******************************************************************************/
>
> -#define LRU_REFS_FLAGS (BIT(PG_referenced) | BIT(PG_workingset))
> -
> #define DEFINE_MAX_SEQ(lruvec) \
> unsigned long max_seq = READ_ONCE((lruvec)->lrugen.max_seq)
>
> @@ -4138,8 +4136,10 @@ void lru_gen_look_around(struct page_vma_mapped_walk *pvmw)
> old_gen = folio_lru_gen(folio);
> if (old_gen < 0)
> folio_set_referenced(folio);
> - else if (old_gen != new_gen)
> + else if (old_gen != new_gen) {
> + folio_clear_lru_refs(folio);
> folio_activate(folio);
> + }
> }
>
> arch_leave_lazy_mmu_mode();
> @@ -4370,7 +4370,7 @@ static bool isolate_folio(struct lruvec *lruvec, struct folio *folio, struct sca
>
> /* see the comment on MAX_NR_TIERS */
> if (!folio_test_referenced(folio))
> - set_mask_bits(&folio->flags, LRU_REFS_MASK | LRU_REFS_FLAGS, 0);
> + folio_clear_lru_refs(folio);
>
> /* for shrink_folio_list() */
> folio_clear_reclaim(folio);
> --
> 2.47.0.rc1.288.g06298d1525-goog
>
^ permalink raw reply [flat|nested] 2+ messages in thread
end of thread, other threads:[~2024-10-17 18:18 UTC | newest]
Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2024-10-17 18:15 [PATCH v2] mm/mglru: reset page lru tier bits when activating Wei Xu
2024-10-17 18:18 ` Wei Xu
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox