* [PATCH v3] mm: memory: check userfaultfd_wp() in vmf_orig_pte_uffd_wp()
@ 2024-04-22 3:00 Kefeng Wang
2024-04-22 12:18 ` Peter Xu
0 siblings, 1 reply; 2+ messages in thread
From: Kefeng Wang @ 2024-04-22 3:00 UTC (permalink / raw)
To: Andrew Morton; +Cc: Peter Xu, linux-mm, Kefeng Wang
Adding userfaultfd_wp() check in vmf_orig_pte_uffd_wp() to avoid the
unnecessary FAULT_FLAG_ORIG_PTE_VALID check/pte_marker_entry_uffd_wp()
in most pagefault, note, the function vmf_orig_pte_uffd_wp() is not
inlined in the two kernel versions, the difference is shown below,
perf date,
perf report -i perf.data.before | grep vmf
0.17% 0.13% lat_pagefault [kernel.kallsyms] [k] vmf_orig_pte_uffd_wp.part.0.isra.0
perf report -i perf.data.after | grep vmf
lat_pagefault -W 5 -N 5 /tmp/XXX
latency before after diff
average(8 tests) 0.262675 0.2600375 -0.0026375
Although it's a small, but the uffd_wp is a new feature than previous
kernel, when the vma is not registered with UFFD_WP, let's avoid to
execute the new logical, also adding __always_inline attribute to
vmf_orig_pte_uffd_wp(), which make set_pte_range() only check VM_UFFD_WP
flags without the function call. In addition, directly call the
vmf_orig_pte_uffd_wp() in do_anonymous_page() and set_pte_range()
to save an uffd_wp variable.
Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
---
v3:
- move userfaultfd_wp() check uppper and add __always_inline to
vmf_orig_pte_uffd_wp()
- update changelog and add lat_pagefault results
mm/memory.c | 10 +++++-----
1 file changed, 5 insertions(+), 5 deletions(-)
diff --git a/mm/memory.c b/mm/memory.c
index d4e5cd9bc183..09ed76e5b8c0 100644
--- a/mm/memory.c
+++ b/mm/memory.c
@@ -112,8 +112,10 @@ static bool vmf_pte_changed(struct vm_fault *vmf);
* Return true if the original pte was a uffd-wp pte marker (so the pte was
* wr-protected).
*/
-static bool vmf_orig_pte_uffd_wp(struct vm_fault *vmf)
+static __always_inline bool vmf_orig_pte_uffd_wp(struct vm_fault *vmf)
{
+ if (!userfaultfd_wp(vmf->vma))
+ return false;
if (!(vmf->flags & FAULT_FLAG_ORIG_PTE_VALID))
return false;
@@ -4388,7 +4390,6 @@ static struct folio *alloc_anon_folio(struct vm_fault *vmf)
*/
static vm_fault_t do_anonymous_page(struct vm_fault *vmf)
{
- bool uffd_wp = vmf_orig_pte_uffd_wp(vmf);
struct vm_area_struct *vma = vmf->vma;
unsigned long addr = vmf->address;
struct folio *folio;
@@ -4488,7 +4489,7 @@ static vm_fault_t do_anonymous_page(struct vm_fault *vmf)
folio_add_new_anon_rmap(folio, vma, addr);
folio_add_lru_vma(folio, vma);
setpte:
- if (uffd_wp)
+ if (vmf_orig_pte_uffd_wp(vmf))
entry = pte_mkuffd_wp(entry);
set_ptes(vma->vm_mm, addr, vmf->pte, entry, nr_pages);
@@ -4663,7 +4664,6 @@ void set_pte_range(struct vm_fault *vmf, struct folio *folio,
struct page *page, unsigned int nr, unsigned long addr)
{
struct vm_area_struct *vma = vmf->vma;
- bool uffd_wp = vmf_orig_pte_uffd_wp(vmf);
bool write = vmf->flags & FAULT_FLAG_WRITE;
bool prefault = in_range(vmf->address, addr, nr * PAGE_SIZE);
pte_t entry;
@@ -4678,7 +4678,7 @@ void set_pte_range(struct vm_fault *vmf, struct folio *folio,
if (write)
entry = maybe_mkwrite(pte_mkdirty(entry), vma);
- if (unlikely(uffd_wp))
+ if (unlikely(vmf_orig_pte_uffd_wp(vmf)))
entry = pte_mkuffd_wp(entry);
/* copy-on-write page */
if (write && !(vma->vm_flags & VM_SHARED)) {
--
2.27.0
^ permalink raw reply [flat|nested] 2+ messages in thread
* Re: [PATCH v3] mm: memory: check userfaultfd_wp() in vmf_orig_pte_uffd_wp()
2024-04-22 3:00 [PATCH v3] mm: memory: check userfaultfd_wp() in vmf_orig_pte_uffd_wp() Kefeng Wang
@ 2024-04-22 12:18 ` Peter Xu
0 siblings, 0 replies; 2+ messages in thread
From: Peter Xu @ 2024-04-22 12:18 UTC (permalink / raw)
To: Kefeng Wang; +Cc: Andrew Morton, linux-mm
On Mon, Apr 22, 2024 at 11:00:39AM +0800, Kefeng Wang wrote:
> Adding userfaultfd_wp() check in vmf_orig_pte_uffd_wp() to avoid the
> unnecessary FAULT_FLAG_ORIG_PTE_VALID check/pte_marker_entry_uffd_wp()
> in most pagefault, note, the function vmf_orig_pte_uffd_wp() is not
> inlined in the two kernel versions, the difference is shown below,
>
> perf date,
>
> perf report -i perf.data.before | grep vmf
> 0.17% 0.13% lat_pagefault [kernel.kallsyms] [k] vmf_orig_pte_uffd_wp.part.0.isra.0
> perf report -i perf.data.after | grep vmf
>
> lat_pagefault -W 5 -N 5 /tmp/XXX
> latency before after diff
> average(8 tests) 0.262675 0.2600375 -0.0026375
>
> Although it's a small, but the uffd_wp is a new feature than previous
> kernel, when the vma is not registered with UFFD_WP, let's avoid to
> execute the new logical, also adding __always_inline attribute to
> vmf_orig_pte_uffd_wp(), which make set_pte_range() only check VM_UFFD_WP
> flags without the function call. In addition, directly call the
> vmf_orig_pte_uffd_wp() in do_anonymous_page() and set_pte_range()
> to save an uffd_wp variable.
>
> Signed-off-by: Kefeng Wang <wangkefeng.wang@huawei.com>
Reviewed-by: Peter Xu <peterx@redhat.com>
--
Peter Xu
^ permalink raw reply [flat|nested] 2+ messages in thread
end of thread, other threads:[~2024-04-22 12:18 UTC | newest]
Thread overview: 2+ messages (download: mbox.gz / follow: Atom feed)
-- links below jump to the message on this page --
2024-04-22 3:00 [PATCH v3] mm: memory: check userfaultfd_wp() in vmf_orig_pte_uffd_wp() Kefeng Wang
2024-04-22 12:18 ` Peter Xu
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox