From: Suren Baghdasaryan <surenb@google.com>
To: akpm@linux-foundation.org
Cc: michel@lespinasse.org, jglisse@google.com, mhocko@suse.com,
vbabka@suse.cz, hannes@cmpxchg.org, mgorman@techsingularity.net,
dave@stgolabs.net, willy@infradead.org, liam.howlett@oracle.com,
peterz@infradead.org, ldufour@linux.ibm.com,
laurent.dufour@fr.ibm.com, paulmck@kernel.org, riel@surriel.com,
luto@kernel.org, songliubraving@fb.com, peterx@redhat.com,
david@redhat.com, dhowells@redhat.com, hughd@google.com,
bigeasy@linutronix.de, kent.overstreet@linux.dev,
rientjes@google.com, axelrasmussen@google.com,
joelaf@google.com, minchan@google.com, surenb@google.com,
kernel-team@android.com, linux-mm@kvack.org,
linux-arm-kernel@lists.infradead.org,
linuxppc-dev@lists.ozlabs.org, x86@kernel.org,
linux-kernel@vger.kernel.org
Subject: [RFC PATCH 26/28] powerc/mm: try VMA lock-based page fault handling first
Date: Mon, 29 Aug 2022 21:25:29 +0000 [thread overview]
Message-ID: <20220829212531.3184856-27-surenb@google.com> (raw)
In-Reply-To: <20220829212531.3184856-1-surenb@google.com>
From: Laurent Dufour <ldufour@linux.ibm.com>
Attempt VMA lock-based page fault handling first, and fall back to the
existing mmap_lock-based handling if that fails.
Copied from "x86/mm: try VMA lock-based page fault handling first"
Signed-off-by: Laurent Dufour <ldufour@linux.ibm.com>
---
arch/powerpc/mm/fault.c | 41 +++++++++++++++++++++++++++++++++++++++++
1 file changed, 41 insertions(+)
diff --git a/arch/powerpc/mm/fault.c b/arch/powerpc/mm/fault.c
index 014005428687..c92bdfcd1796 100644
--- a/arch/powerpc/mm/fault.c
+++ b/arch/powerpc/mm/fault.c
@@ -450,6 +450,44 @@ static int ___do_page_fault(struct pt_regs *regs, unsigned long address,
if (is_exec)
flags |= FAULT_FLAG_INSTRUCTION;
+#ifdef CONFIG_PER_VMA_LOCK
+ if (!(flags & FAULT_FLAG_USER) || atomic_read(&mm->mm_users) == 1)
+ goto lock_mmap;
+
+ vma = find_and_lock_anon_vma(mm, address);
+ if (!vma)
+ goto lock_mmap;
+
+ if (unlikely(access_pkey_error(is_write, is_exec,
+ (error_code & DSISR_KEYFAULT), vma))) {
+ int rc = bad_access_pkey(regs, address, vma);
+
+ vma_read_unlock(vma);
+ return rc;
+ }
+
+ if (unlikely(access_error(is_write, is_exec, vma))) {
+ int rc = bad_access(regs, address);
+
+ vma_read_unlock(vma);
+ return rc;
+ }
+
+ fault = handle_mm_fault(vma, address, flags | FAULT_FLAG_VMA_LOCK, regs);
+ vma_read_unlock(vma);
+
+ if (!(fault & VM_FAULT_RETRY)) {
+ count_vm_vma_lock_event(VMA_LOCK_SUCCESS);
+ goto done;
+ }
+ count_vm_vma_lock_event(VMA_LOCK_RETRY);
+
+ if (fault_signal_pending(fault, regs))
+ return user_mode(regs) ? 0 : SIGBUS;
+
+lock_mmap:
+#endif /* CONFIG_PER_VMA_LOCK */
+
/* When running in the kernel we expect faults to occur only to
* addresses in user space. All other faults represent errors in the
* kernel and should generate an OOPS. Unfortunately, in the case of an
@@ -526,6 +564,9 @@ static int ___do_page_fault(struct pt_regs *regs, unsigned long address,
mmap_read_unlock(current->mm);
+#ifdef CONFIG_PER_VMA_LOCK
+done:
+#endif
if (unlikely(fault & VM_FAULT_ERROR))
return mm_fault_error(regs, address, fault);
--
2.37.2.672.g94769d06f0-goog
next prev parent reply other threads:[~2022-08-29 21:26 UTC|newest]
Thread overview: 29+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-08-29 21:25 [RFC PATCH 00/28] per-VMA locks proposal Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 01/28] mm: introduce CONFIG_PER_VMA_LOCK Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 02/28] mm: rcu safe VMA freeing Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 03/28] mm: introduce __find_vma to be used without mmap_lock protection Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 04/28] mm: move mmap_lock assert function definitions Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 05/28] mm: add per-VMA lock and helper functions to control it Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 06/28] mm: mark VMA as locked whenever vma->vm_flags are modified Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 07/28] kernel/fork: mark VMAs as locked before copying pages during fork Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 08/28] mm/khugepaged: mark VMA as locked while collapsing a hugepage Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 09/28] mm/mempolicy: mark VMA as locked when changing protection policy Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 10/28] mm/mmap: mark VMAs as locked in vma_adjust Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 11/28] mm/mmap: mark VMAs as locked before merging or splitting them Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 12/28] mm/mremap: mark VMA as locked while remapping it to a new address range Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 13/28] mm: conditionally mark VMA as locked in free_pgtables and unmap_page_range Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 14/28] mm: mark VMAs as locked before isolating them Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 15/28] mm/mmap: mark adjacent VMAs as locked if they can grow into unmapped area Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 16/28] kernel/fork: assert no VMA readers during its destruction Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 17/28] mm/mmap: prevent pagefault handler from racing with mmu_notifier registration Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 18/28] mm: add FAULT_FLAG_VMA_LOCK flag Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 19/28] mm: disallow do_swap_page to handle page faults under VMA lock Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 20/28] mm: introduce per-VMA lock statistics Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 21/28] mm: introduce find_and_lock_anon_vma to be used from arch-specific code Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 22/28] x86/mm: try VMA lock-based page fault handling first Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 23/28] x86/mm: define ARCH_SUPPORTS_PER_VMA_LOCK Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 24/28] arm64/mm: try VMA lock-based page fault handling first Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 25/28] arm64/mm: define ARCH_SUPPORTS_PER_VMA_LOCK Suren Baghdasaryan
2022-08-29 21:25 ` Suren Baghdasaryan [this message]
2022-08-29 21:25 ` [RFC PATCH 27/28] powerpc/mm: " Suren Baghdasaryan
2022-08-29 21:25 ` [RFC PATCH 28/28] kernel/fork: throttle call_rcu() calls in vm_area_free Suren Baghdasaryan
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20220829212531.3184856-27-surenb@google.com \
--to=surenb@google.com \
--cc=akpm@linux-foundation.org \
--cc=axelrasmussen@google.com \
--cc=bigeasy@linutronix.de \
--cc=dave@stgolabs.net \
--cc=david@redhat.com \
--cc=dhowells@redhat.com \
--cc=hannes@cmpxchg.org \
--cc=hughd@google.com \
--cc=jglisse@google.com \
--cc=joelaf@google.com \
--cc=kent.overstreet@linux.dev \
--cc=kernel-team@android.com \
--cc=laurent.dufour@fr.ibm.com \
--cc=ldufour@linux.ibm.com \
--cc=liam.howlett@oracle.com \
--cc=linux-arm-kernel@lists.infradead.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=linuxppc-dev@lists.ozlabs.org \
--cc=luto@kernel.org \
--cc=mgorman@techsingularity.net \
--cc=mhocko@suse.com \
--cc=michel@lespinasse.org \
--cc=minchan@google.com \
--cc=paulmck@kernel.org \
--cc=peterx@redhat.com \
--cc=peterz@infradead.org \
--cc=riel@surriel.com \
--cc=rientjes@google.com \
--cc=songliubraving@fb.com \
--cc=vbabka@suse.cz \
--cc=willy@infradead.org \
--cc=x86@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox