From: Kefeng Wang <wangkefeng.wang@huawei.com>
To: Andrew Morton <akpm@linux-foundation.org>
Cc: <willy@infradead.org>, <linux-mm@kvack.org>,
<linux-kernel@vger.kernel.org>, <ying.huang@intel.com>,
<david@redhat.com>, Zi Yan <ziy@nvidia.com>,
Kefeng Wang <wangkefeng.wang@huawei.com>
Subject: [PATCH -next 7/7] mm: convert page_cpupid_xchg_last() to folio_cpupid_xchg_last()
Date: Tue, 10 Oct 2023 14:45:44 +0800 [thread overview]
Message-ID: <20231010064544.4162286-8-wangkefeng.wang@huawei.com> (raw)
In-Reply-To: <20231010064544.4162286-1-wangkefeng.wang@huawei.com>
Make page_cpupid_xchg_last() to take a folio, and rename it to
olio_cpupid_xchg_last() since all callers with a folio.
Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
---
include/linux/mm.h | 14 +++++++-------
kernel/sched/fair.c | 2 +-
mm/huge_memory.c | 2 +-
mm/memory.c | 2 +-
mm/migrate.c | 4 ++--
mm/mmzone.c | 6 +++---
6 files changed, 15 insertions(+), 15 deletions(-)
diff --git a/include/linux/mm.h b/include/linux/mm.h
index 3d59455626fa..e761642e1c00 100644
--- a/include/linux/mm.h
+++ b/include/linux/mm.h
@@ -1683,9 +1683,9 @@ static inline bool __cpupid_match_pid(pid_t task_pid, int cpupid)
#define cpupid_match_pid(task, cpupid) __cpupid_match_pid(task->pid, cpupid)
#ifdef LAST_CPUPID_NOT_IN_PAGE_FLAGS
-static inline int page_cpupid_xchg_last(struct page *page, int cpupid)
+static inline int folio_cpupid_xchg_last(struct folio *folio, int cpupid)
{
- return xchg(&page->_last_cpupid, cpupid & LAST_CPUPID_MASK);
+ return xchg(&folio->_last_cpupid, cpupid & LAST_CPUPID_MASK);
}
static inline int folio_cpupid_last(struct folio *folio)
@@ -1702,7 +1702,7 @@ static inline int folio_cpupid_last(struct folio *folio)
return (folio->flags >> LAST_CPUPID_PGSHIFT) & LAST_CPUPID_MASK;
}
-extern int page_cpupid_xchg_last(struct page *page, int cpupid);
+extern int folio_cpupid_xchg_last(struct folio *folio, int cpupid);
static inline void page_cpupid_reset_last(struct page *page)
{
@@ -1714,8 +1714,8 @@ static inline int xchg_folio_access_time(struct folio *folio, int time)
{
int last_time;
- last_time = page_cpupid_xchg_last(&folio->page,
- time >> PAGE_ACCESS_TIME_BUCKETS);
+ last_time = folio_cpupid_xchg_last(folio,
+ time >> PAGE_ACCESS_TIME_BUCKETS);
return last_time << PAGE_ACCESS_TIME_BUCKETS;
}
@@ -1729,9 +1729,9 @@ static inline void vma_set_access_pid_bit(struct vm_area_struct *vma)
}
}
#else /* !CONFIG_NUMA_BALANCING */
-static inline int page_cpupid_xchg_last(struct page *page, int cpupid)
+static inline int folio_cpupid_xchg_last(struct folio *folio, int cpupid)
{
- return page_to_nid(page); /* XXX */
+ return folio_nid(folio); /* XXX */
}
static inline int xchg_folio_access_time(struct folio *folio, int time)
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index 50b9f63099fb..5d4c7cedc6d1 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -1818,7 +1818,7 @@ bool should_numa_migrate_memory(struct task_struct *p, struct folio *folio,
}
this_cpupid = cpu_pid_to_cpupid(dst_cpu, current->pid);
- last_cpupid = page_cpupid_xchg_last(&folio->page, this_cpupid);
+ last_cpupid = folio_cpupid_xchg_last(folio, this_cpupid);
if (!(sysctl_numa_balancing_mode & NUMA_BALANCING_MEMORY_TIERING) &&
!node_is_toptier(src_nid) && !cpupid_valid(last_cpupid))
diff --git a/mm/huge_memory.c b/mm/huge_memory.c
index 3b37367eaeff..2163b1d0dad5 100644
--- a/mm/huge_memory.c
+++ b/mm/huge_memory.c
@@ -2515,7 +2515,7 @@ static void __split_huge_page_tail(struct folio *folio, int tail,
if (page_is_idle(head))
set_page_idle(page_tail);
- page_cpupid_xchg_last(page_tail, folio_cpupid_last(folio));
+ folio_cpupid_xchg_last(new_folio, folio_cpupid_last(folio));
/*
* always add to the tail because some iterators expect new
diff --git a/mm/memory.c b/mm/memory.c
index 1a1a6a6ccd58..9f3b359b46db 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -3034,7 +3034,7 @@ static inline void wp_page_reuse(struct vm_fault *vmf, struct folio *folio)
* information potentially belongs to a now completely
* unrelated process.
*/
- page_cpupid_xchg_last(vmf->page, (1 << LAST_CPUPID_SHIFT) - 1);
+ folio_cpupid_xchg_last(folio, (1 << LAST_CPUPID_SHIFT) - 1);
}
flush_cache_page(vma, vmf->address, pte_pfn(vmf->orig_pte));
diff --git a/mm/migrate.c b/mm/migrate.c
index c602bf6dec97..5642e9572d80 100644
--- a/mm/migrate.c
+++ b/mm/migrate.c
@@ -588,7 +588,7 @@ void folio_migrate_flags(struct folio *newfolio, struct folio *folio)
* Copy NUMA information to the new page, to prevent over-eager
* future migrations of this same page.
*/
- cpupid = page_cpupid_xchg_last(&folio->page, -1);
+ cpupid = folio_cpupid_xchg_last(folio, -1);
/*
* For memory tiering mode, when migrate between slow and fast
* memory node, reset cpupid, because that is used to record
@@ -601,7 +601,7 @@ void folio_migrate_flags(struct folio *newfolio, struct folio *folio)
if (f_toptier != t_toptier)
cpupid = -1;
}
- page_cpupid_xchg_last(&newfolio->page, cpupid);
+ folio_cpupid_xchg_last(newfolio, cpupid);
folio_migrate_ksm(newfolio, folio);
/*
diff --git a/mm/mmzone.c b/mm/mmzone.c
index 68e1511be12d..cd473f82b647 100644
--- a/mm/mmzone.c
+++ b/mm/mmzone.c
@@ -93,19 +93,19 @@ void lruvec_init(struct lruvec *lruvec)
}
#if defined(CONFIG_NUMA_BALANCING) && !defined(LAST_CPUPID_NOT_IN_PAGE_FLAGS)
-int page_cpupid_xchg_last(struct page *page, int cpupid)
+int folio_cpupid_xchg_last(struct folio *folio, int cpupid)
{
unsigned long old_flags, flags;
int last_cpupid;
- old_flags = READ_ONCE(page->flags);
+ old_flags = READ_ONCE(folio->flags);
do {
flags = old_flags;
last_cpupid = (flags >> LAST_CPUPID_PGSHIFT) & LAST_CPUPID_MASK;
flags &= ~(LAST_CPUPID_MASK << LAST_CPUPID_PGSHIFT);
flags |= (cpupid & LAST_CPUPID_MASK) << LAST_CPUPID_PGSHIFT;
- } while (unlikely(!try_cmpxchg(&page->flags, &old_flags, flags)));
+ } while (unlikely(!try_cmpxchg(&folio->flags, &old_flags, flags)));
return last_cpupid;
}
--
2.27.0
prev parent reply other threads:[~2023-10-10 6:47 UTC|newest]
Thread overview: 16+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-10-10 6:45 [PATCH -next 0/7] mm: convert page cpupid functions to folios Kefeng Wang
2023-10-10 6:45 ` [PATCH -next 1/7] mm_types: add _last_cpupid into folio Kefeng Wang
2023-10-10 8:17 ` Huang, Ying
2023-10-10 11:10 ` Kefeng Wang
2023-10-10 12:33 ` Matthew Wilcox
2023-10-11 3:02 ` Kefeng Wang
2023-10-11 5:55 ` Huang, Ying
2023-10-11 8:05 ` Kefeng Wang
2023-10-10 6:45 ` [PATCH -next 2/7] mm: mprotect: use a folio in change_pte_range() Kefeng Wang
2023-10-10 6:45 ` [PATCH -next 3/7] mm: huge_memory: use a folio in change_huge_pmd() Kefeng Wang
2023-10-10 6:45 ` [PATCH -next 4/7] mm: convert xchg_page_access_time to xchg_folio_access_time() Kefeng Wang
2023-10-10 12:27 ` Matthew Wilcox
2023-10-11 3:03 ` Kefeng Wang
2023-10-10 6:45 ` [PATCH -next 5/7] mm: convert page_cpupid_last() to folio_cpupid_last() Kefeng Wang
2023-10-10 6:45 ` [PATCH -next 6/7] mm: make wp_page_reuse() and finish_mkwrite_fault() to take a folio Kefeng Wang
2023-10-10 6:45 ` Kefeng Wang [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20231010064544.4162286-8-wangkefeng.wang@huawei.com \
--to=wangkefeng.wang@huawei.com \
--cc=akpm@linux-foundation.org \
--cc=david@redhat.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=willy@infradead.org \
--cc=ying.huang@intel.com \
--cc=ziy@nvidia.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox