From: NeilBrown <neilb@suse.de>
To: Andrew Morton <akpm@linux-foundation.org>
Cc: Geert Uytterhoeven <geert+renesas@glider.be>,
Christoph Hellwig <hch@lst.de>, Miaohe Lin <linmiaohe@huawei.com>,
linux-nfs@vger.kernel.org, linux-mm@kvack.org,
linux-kernel@vger.kernel.org
Subject: [PATCH 1/2] MM: handle THP in swap_*page_fs()
Date: Fri, 29 Apr 2022 10:43:34 +1000 [thread overview]
Message-ID: <165119301488.15698.9457662928942765453.stgit@noble.brown> (raw)
In-Reply-To: <165119280115.15698.2629172320052218921.stgit@noble.brown>
Pages passed to swap_readpage()/swap_writepage() are not necessarily all
the same size - there may be transparent-huge-pages involves.
The BIO paths of swap_*page() handle this correctly, but the SWP_FS_OPS
path does not.
So we need to use thp_size() to find the size, not just assume
PAGE_SIZE, and we need to track the total length of the request, not
just assume it is "page * PAGE_SIZE".
Reported-by: Miaohe Lin <linmiaohe@huawei.com>
Signed-off-by: NeilBrown <neilb@suse.de>
---
mm/page_io.c | 23 +++++++++++++----------
1 file changed, 13 insertions(+), 10 deletions(-)
diff --git a/mm/page_io.c b/mm/page_io.c
index c132511f521c..d636a3531cad 100644
--- a/mm/page_io.c
+++ b/mm/page_io.c
@@ -239,6 +239,7 @@ struct swap_iocb {
struct kiocb iocb;
struct bio_vec bvec[SWAP_CLUSTER_MAX];
int pages;
+ int len;
};
static mempool_t *sio_pool;
@@ -261,7 +262,7 @@ static void sio_write_complete(struct kiocb *iocb, long ret)
struct page *page = sio->bvec[0].bv_page;
int p;
- if (ret != PAGE_SIZE * sio->pages) {
+ if (ret != sio->len) {
/*
* In the case of swap-over-nfs, this can be a
* temporary failure if the system has limited
@@ -301,7 +302,7 @@ static int swap_writepage_fs(struct page *page, struct writeback_control *wbc)
sio = *wbc->swap_plug;
if (sio) {
if (sio->iocb.ki_filp != swap_file ||
- sio->iocb.ki_pos + sio->pages * PAGE_SIZE != pos) {
+ sio->iocb.ki_pos + sio->len != pos) {
swap_write_unplug(sio);
sio = NULL;
}
@@ -312,10 +313,12 @@ static int swap_writepage_fs(struct page *page, struct writeback_control *wbc)
sio->iocb.ki_complete = sio_write_complete;
sio->iocb.ki_pos = pos;
sio->pages = 0;
+ sio->len = 0;
}
sio->bvec[sio->pages].bv_page = page;
- sio->bvec[sio->pages].bv_len = PAGE_SIZE;
+ sio->bvec[sio->pages].bv_len = thp_size(page);
sio->bvec[sio->pages].bv_offset = 0;
+ sio->len += thp_size(page);
sio->pages += 1;
if (sio->pages == ARRAY_SIZE(sio->bvec) || !wbc->swap_plug) {
swap_write_unplug(sio);
@@ -371,8 +374,7 @@ void swap_write_unplug(struct swap_iocb *sio)
struct address_space *mapping = sio->iocb.ki_filp->f_mapping;
int ret;
- iov_iter_bvec(&from, WRITE, sio->bvec, sio->pages,
- PAGE_SIZE * sio->pages);
+ iov_iter_bvec(&from, WRITE, sio->bvec, sio->pages, sio->len);
ret = mapping->a_ops->swap_rw(&sio->iocb, &from);
if (ret != -EIOCBQUEUED)
sio_write_complete(&sio->iocb, ret);
@@ -383,7 +385,7 @@ static void sio_read_complete(struct kiocb *iocb, long ret)
struct swap_iocb *sio = container_of(iocb, struct swap_iocb, iocb);
int p;
- if (ret == PAGE_SIZE * sio->pages) {
+ if (ret == sio->len) {
for (p = 0; p < sio->pages; p++) {
struct page *page = sio->bvec[p].bv_page;
@@ -415,7 +417,7 @@ static void swap_readpage_fs(struct page *page,
sio = *plug;
if (sio) {
if (sio->iocb.ki_filp != sis->swap_file ||
- sio->iocb.ki_pos + sio->pages * PAGE_SIZE != pos) {
+ sio->iocb.ki_pos + sio->len != pos) {
swap_read_unplug(sio);
sio = NULL;
}
@@ -426,10 +428,12 @@ static void swap_readpage_fs(struct page *page,
sio->iocb.ki_pos = pos;
sio->iocb.ki_complete = sio_read_complete;
sio->pages = 0;
+ sio->len = 0;
}
sio->bvec[sio->pages].bv_page = page;
- sio->bvec[sio->pages].bv_len = PAGE_SIZE;
+ sio->bvec[sio->pages].bv_len = thp_size(page);
sio->bvec[sio->pages].bv_offset = 0;
+ sio->len += thp_size(page);
sio->pages += 1;
if (sio->pages == ARRAY_SIZE(sio->bvec) || !plug) {
swap_read_unplug(sio);
@@ -521,8 +525,7 @@ void __swap_read_unplug(struct swap_iocb *sio)
struct address_space *mapping = sio->iocb.ki_filp->f_mapping;
int ret;
- iov_iter_bvec(&from, READ, sio->bvec, sio->pages,
- PAGE_SIZE * sio->pages);
+ iov_iter_bvec(&from, READ, sio->bvec, sio->pages, sio->len);
ret = mapping->a_ops->swap_rw(&sio->iocb, &from);
if (ret != -EIOCBQUEUED)
sio_read_complete(&sio->iocb, ret);
next prev parent reply other threads:[~2022-04-29 0:44 UTC|newest]
Thread overview: 13+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-04-29 0:43 [PATCH 0/2] Finalising swap-over-NFS patches NeilBrown
2022-04-29 0:43 ` [PATCH 2/2] NFS: rename nfs_direct_IO and use as ->swap_rw NeilBrown
2022-04-29 1:23 ` Andrew Morton
2022-04-29 2:05 ` NeilBrown
2022-04-29 0:43 ` NeilBrown [this message]
2022-04-29 1:21 ` [PATCH 1/2] MM: handle THP in swap_*page_fs() Andrew Morton
2022-04-29 1:57 ` NeilBrown
2022-04-29 8:13 ` Miaohe Lin
2022-04-29 19:04 ` Yang Shi
2022-05-02 4:23 ` NeilBrown
2022-05-02 17:48 ` Yang Shi
2022-05-04 23:41 ` NeilBrown
2022-05-06 2:56 ` ying.huang
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=165119301488.15698.9457662928942765453.stgit@noble.brown \
--to=neilb@suse.de \
--cc=akpm@linux-foundation.org \
--cc=geert+renesas@glider.be \
--cc=hch@lst.de \
--cc=linmiaohe@huawei.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=linux-nfs@vger.kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox