From: jglisse@redhat.com
To: linux-mm@kvack.org
Cc: "Andrew Morton" <akpm@linux-foundation.org>,
linux-kernel@vger.kernel.org,
"Jérôme Glisse" <jglisse@redhat.com>,
"Michal Hocko" <mhocko@suse.com>,
"Ralph Campbell" <rcampbell@nvidia.com>,
"John Hubbard" <jhubbard@nvidia.com>
Subject: [PATCH 7/7] mm/hmm: proper support for blockable mmu_notifier
Date: Fri, 24 Aug 2018 15:25:49 -0400 [thread overview]
Message-ID: <20180824192549.30844-8-jglisse@redhat.com> (raw)
In-Reply-To: <20180824192549.30844-1-jglisse@redhat.com>
From: JA(C)rA'me Glisse <jglisse@redhat.com>
When mmu_notifier calls invalidate_range_start callback with blockable
set to false we should not sleep. Properly propagate this to HMM users.
Signed-off-by: JA(C)rA'me Glisse <jglisse@redhat.com>
Cc: Michal Hocko <mhocko@suse.com>
Cc: Ralph Campbell <rcampbell@nvidia.com>
Cc: John Hubbard <jhubbard@nvidia.com>
Cc: Andrew Morton <akpm@linux-foundation.org>
---
include/linux/hmm.h | 12 +++++++++---
mm/hmm.c | 39 ++++++++++++++++++++++++++++-----------
2 files changed, 37 insertions(+), 14 deletions(-)
diff --git a/include/linux/hmm.h b/include/linux/hmm.h
index 064924bce75c..c783916f8732 100644
--- a/include/linux/hmm.h
+++ b/include/linux/hmm.h
@@ -287,11 +287,13 @@ enum hmm_update_event {
* @start: virtual start address of the range to update
* @end: virtual end address of the range to update
* @event: event triggering the update (what is happening)
+ * @blockable: can the callback block/sleep ?
*/
struct hmm_update {
unsigned long start;
unsigned long end;
enum hmm_update_event event;
+ bool blockable;
};
/*
@@ -314,6 +316,8 @@ struct hmm_mirror_ops {
*
* @mirror: pointer to struct hmm_mirror
* @update: update informations (see struct hmm_update)
+ * Returns: -EAGAIN if update.blockable false and callback need to
+ * block, 0 otherwise.
*
* This callback ultimately originates from mmu_notifiers when the CPU
* page table is updated. The device driver must update its page table
@@ -322,10 +326,12 @@ struct hmm_mirror_ops {
*
* The device driver must not return from this callback until the device
* page tables are completely updated (TLBs flushed, etc); this is a
- * synchronous call.
+ * synchronous call. If driver need to sleep and update->blockable is
+ * false then you need to abort (do not do anything that would sleep or
+ * block) and return -EAGAIN.
*/
- void (*sync_cpu_device_pagetables)(struct hmm_mirror *mirror,
- const struct hmm_update *update);
+ int (*sync_cpu_device_pagetables)(struct hmm_mirror *mirror,
+ const struct hmm_update *update);
};
/*
diff --git a/mm/hmm.c b/mm/hmm.c
index 6fe31e2bfa1e..1d8fcaa0606f 100644
--- a/mm/hmm.c
+++ b/mm/hmm.c
@@ -123,12 +123,18 @@ void hmm_mm_destroy(struct mm_struct *mm)
kfree(mm->hmm);
}
-static void hmm_invalidate_range(struct hmm *hmm, bool device,
- const struct hmm_update *update)
+static int hmm_invalidate_range(struct hmm *hmm, bool device,
+ const struct hmm_update *update)
{
struct hmm_mirror *mirror;
struct hmm_range *range;
+ /*
+ * It is fine to wait on lock here even if update->blockable is false
+ * as the hmm->lock is only held for short period of time (when adding
+ * or walking the ranges list). We could also convert the range list
+ * into a lru list and avoid the spinlock all together.
+ */
spin_lock(&hmm->lock);
list_for_each_entry(range, &hmm->ranges, list) {
unsigned long addr, idx, npages;
@@ -145,12 +151,26 @@ static void hmm_invalidate_range(struct hmm *hmm, bool device,
spin_unlock(&hmm->lock);
if (!device)
- return;
+ return 0;
+ /*
+ * It is fine to wait on mirrors_sem here even if update->blockable is
+ * false as this semaphore is only taken in write mode for short period
+ * when adding a new mirror to the list.
+ */
down_read(&hmm->mirrors_sem);
- list_for_each_entry(mirror, &hmm->mirrors, list)
- mirror->ops->sync_cpu_device_pagetables(mirror, update);
+ list_for_each_entry(mirror, &hmm->mirrors, list) {
+ int ret;
+
+ ret = mirror->ops->sync_cpu_device_pagetables(mirror, update);
+ if (!update->blockable && ret == -EAGAIN) {
+ up_read(&hmm->mirrors_sem);
+ return -EAGAIN;
+ }
+ }
up_read(&hmm->mirrors_sem);
+
+ return 0;
}
static void hmm_release(struct mmu_notifier *mn, struct mm_struct *mm)
@@ -188,17 +208,13 @@ static int hmm_invalidate_range_start(struct mmu_notifier *mn,
struct hmm_update update;
struct hmm *hmm = mm->hmm;
- if (!blockable)
- return -EAGAIN;
-
VM_BUG_ON(!hmm);
update.start = start;
update.end = end;
update.event = HMM_UPDATE_INVALIDATE;
- hmm_invalidate_range(hmm, true, &update);
-
- return 0;
+ update.blockable = blockable;
+ return hmm_invalidate_range(hmm, true, &update);
}
static void hmm_invalidate_range_end(struct mmu_notifier *mn,
@@ -214,6 +230,7 @@ static void hmm_invalidate_range_end(struct mmu_notifier *mn,
update.start = start;
update.end = end;
update.event = HMM_UPDATE_INVALIDATE;
+ update.blockable = true;
hmm_invalidate_range(hmm, false, &update);
}
--
2.17.1
next prev parent reply other threads:[~2018-08-24 19:25 UTC|newest]
Thread overview: 29+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-08-24 19:25 [PATCH 0/7] HMM updates, improvements and fixes jglisse
2018-08-24 19:25 ` [PATCH 1/7] mm/hmm: fix utf8 jglisse
2018-08-24 19:25 ` [PATCH 2/7] mm/rmap: map_pte() was not handling private ZONE_DEVICE page properly jglisse
2018-08-30 14:05 ` Balbir Singh
2018-08-30 14:34 ` Jerome Glisse
2018-08-30 14:41 ` [PATCH 3/7] mm/rmap: map_pte() was not handling private ZONE_DEVICE page properly v2 jglisse
2018-08-31 9:27 ` Balbir Singh
2018-08-31 16:19 ` Jerome Glisse
2018-09-02 6:58 ` Balbir Singh
2018-08-24 19:25 ` [PATCH 3/7] mm/hmm: fix race between hmm_mirror_unregister() and mmu_notifier callback jglisse
2018-08-30 14:14 ` Balbir Singh
2018-08-24 19:25 ` [PATCH 4/7] mm/hmm: properly handle migration pmd jglisse
2018-08-25 0:05 ` Zi Yan
2018-08-28 0:35 ` Jerome Glisse
2018-08-28 15:24 ` Michal Hocko
2018-08-28 15:36 ` Jerome Glisse
2018-08-28 15:42 ` Michal Hocko
2018-08-28 15:45 ` Michal Hocko
2018-08-28 15:54 ` Zi Yan
2018-08-28 16:06 ` Jerome Glisse
2018-08-28 16:10 ` Michal Hocko
2018-08-29 17:17 ` [PATCH 4/7] mm/hmm: properly handle migration pmd v2 jglisse
2018-08-24 19:25 ` [PATCH 5/7] mm/hmm: use a structure for update callback parameters jglisse
2018-08-30 23:11 ` Balbir Singh
2018-08-31 16:12 ` Jerome Glisse
2018-08-24 19:25 ` [PATCH 6/7] mm/hmm: invalidate device page table at start of invalidation jglisse
2018-08-24 19:25 ` jglisse [this message]
2018-10-12 18:15 ` [PATCH 0/7] HMM updates, improvements and fixes Jerome Glisse
2018-10-12 21:12 ` Andrew Morton
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20180824192549.30844-8-jglisse@redhat.com \
--to=jglisse@redhat.com \
--cc=akpm@linux-foundation.org \
--cc=jhubbard@nvidia.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mhocko@suse.com \
--cc=rcampbell@nvidia.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox