From: Barry Song <21cnbao@gmail.com>
To: Shakeel Butt <shakeel.butt@linux.dev>
Cc: hanchuanhua@oppo.com, akpm@linux-foundation.org,
linux-mm@kvack.org, baolin.wang@linux.alibaba.com,
chrisl@kernel.org, david@redhat.com, hannes@cmpxchg.org,
hughd@google.com, kaleshsingh@google.com, kasong@tencent.com,
linux-kernel@vger.kernel.org, mhocko@suse.com,
minchan@kernel.org, nphamcs@gmail.com, ryan.roberts@arm.com,
senozhatsky@chromium.org, shy828301@gmail.com,
surenb@google.com, v-songbaohua@oppo.com, willy@infradead.org,
xiang@kernel.org, ying.huang@intel.com, yosryahmed@google.com,
hch@infradead.org, ryncsn@gmail.com,
Tangquan Zheng <zhengtangquan@oppo.com>
Subject: Re: [PATCH v7 2/2] mm: support large folios swap-in for sync io devices
Date: Tue, 27 Aug 2024 07:46:19 +1200 [thread overview]
Message-ID: <CAGsJ_4wGK6pu+KNhYjpWgydp6DyjH5tE=9+mje3UyrXdFJOuNw@mail.gmail.com> (raw)
In-Reply-To: <i6jki2zocqzsjcjgraf6lyl7m3cjzv5lnsuluq5xnvznw7bsge@4easx2ucpxml>
On Sat, Aug 24, 2024 at 5:56 AM Shakeel Butt <shakeel.butt@linux.dev> wrote:
>
> Hi Barry,
>
> On Thu, Aug 22, 2024 at 05:13:06AM GMT, Barry Song wrote:
> > On Thu, Aug 22, 2024 at 1:31 AM Shakeel Butt <shakeel.butt@linux.dev> wrote:
> > >
> > > On Wed, Aug 21, 2024 at 03:45:40PM GMT, hanchuanhua@oppo.com wrote:
> > > > From: Chuanhua Han <hanchuanhua@oppo.com>
> > > >
> > > >
> > > > 3. With both mTHP swap-out and swap-in supported, we offer the option to enable
> > > > zsmalloc compression/decompression with larger granularity[2]. The upcoming
> > > > optimization in zsmalloc will significantly increase swap speed and improve
> > > > compression efficiency. Tested by running 100 iterations of swapping 100MiB
> > > > of anon memory, the swap speed improved dramatically:
> > > > time consumption of swapin(ms) time consumption of swapout(ms)
> > > > lz4 4k 45274 90540
> > > > lz4 64k 22942 55667
> > > > zstdn 4k 85035 186585
> > > > zstdn 64k 46558 118533
> > >
> > > Are the above number with the patch series at [2] or without? Also can
> > > you explain your experiment setup or how can someone reproduce these?
> >
> > Hi Shakeel,
> >
> > The data was recorded after applying both this patch (swap-in mTHP) and
> > patch [2] (compressing/decompressing mTHP instead of page). However,
> > without the swap-in series, patch [2] becomes useless because:
> >
> > If we have a large object, such as 16 pages in zsmalloc:
> > do_swap_page will happen 16 times:
> > 1. decompress the whole large object and copy one page;
> > 2. decompress the whole large object and copy one page;
> > 3. decompress the whole large object and copy one page;
> > ....
> > 16. decompress the whole large object and copy one page;
> >
> > So, patchset [2] will actually degrade performance rather than
> > enhance it if we don't have this swap-in series. This swap-in
> > series is a prerequisite for the zsmalloc/zram series.
>
> Thanks for the explanation.
>
> >
> > We reproduced the data through the following simple steps:
> > 1. Collected anonymous pages from a running phone and saved them to a file.
> > 2. Used a small program to open and read the file into a mapped anonymous
> > memory.
> > 3. Do the belows in the small program:
> > swapout_start_time
> > madv_pageout()
> > swapout_end_time
> >
> > swapin_start_time
> > read_data()
> > swapin_end_time
> >
> > We calculate the throughput of swapout and swapin using the difference between
> > end_time and start_time. Additionally, we record the memory usage of zram after
> > the swapout is complete.
> >
>
> Please correct me if I am wrong but you are saying in your experiment,
> 100 MiB took 90540 ms to compress/swapout and 45274 ms to
> decompress/swapin if backed by 4k pages but took 55667 ms and 22942 ms
> if backed by 64k pages. Basically the table shows total time to compress
> or decomress 100 MiB of memory, right?
Hi Shakeel,
Tangquan(CC'd) collected the data and double-checked the case to confirm
the answer to your question.
We have three cases:
1. no mTHP swap-in, no zsmalloc/zram multi-pages compression/decompression
2. have mTHP swap-in, no zsmalloc/zram multi-pages compression/decompression
3. have mTHP swap-in, have zsmalloc/zram multi-pages compression/decompression
The data was 1 vs 3.
To provide more precise data that covers each change, Tangquan tested
1 vs. 2 and
2 vs. 3 yesterday using LZ4 (the hardware might differ from the
previous test, but the
data shows the same trend) per my request.
1. no mTHP swapin, no zsmalloc/zram patch
swapin_ms. 30336
swapout_ms. 65651
2. have mTHP swapin, no zsmalloc/zram patch
swapin_ms. 27161
swapout_ms. 61135
3. have mTHP swapin, have zsmalloc/zram patch
swapin_ms. 13683
swapout_ms. 43305
The test pseudocode is as follows:
addr=mmap(100M)
read_anon_data_from_file_to addr();
for(i=0;i<100;i++) {
swapout_start_time;
madv_pageout();
swapout_end_time;
swapin_start_time;
read_addr_to_swapin();
swapin_end_time;
}
So, while we saw some improvement from 1 to 2, the significant gains
come from using large blocks for compression and decompression.
This mTHP swap-in series ensures that mTHPs aren't lost after the first swap-in,
so the following 99 iterations continue to involve THP swap-out and
mTHP swap-in.
The improvement from 1 to 2 is due to this mTHP swap-in series, while the
improvement from 2 to 3 comes from the zsmalloc/zram patchset [2] you
mentioned.
[2] https://lore.kernel.org/all/20240327214816.31191-1-21cnbao@gmail.com/
> > >
> >
Thanks
Barry
next prev parent reply other threads:[~2024-08-26 19:46 UTC|newest]
Thread overview: 24+ messages / expand[flat|nested] mbox.gz Atom feed top
2024-08-21 7:45 [PATCH v7 0/2] mm: Ignite large folios swap-in support hanchuanhua
2024-08-21 7:45 ` [PATCH v7 1/2] mm: add nr argument in mem_cgroup_swapin_uncharge_swap() helper to support large folios hanchuanhua
2024-08-21 7:45 ` [PATCH v7 2/2] mm: support large folios swap-in for sync io devices hanchuanhua
2024-08-21 17:31 ` Shakeel Butt
2024-08-21 21:13 ` Barry Song
2024-08-23 17:56 ` Shakeel Butt
2024-08-26 19:46 ` Barry Song [this message]
2024-08-29 1:01 ` Kanchana P Sridhar
2024-08-29 2:24 ` Barry Song
2024-08-29 2:38 ` Sridhar, Kanchana P
2024-09-03 18:24 ` Kairui Song
2024-09-03 18:38 ` Yosry Ahmed
2024-09-03 20:07 ` Andrew Morton
2024-09-03 21:36 ` Barry Song
2024-09-03 22:05 ` Yosry Ahmed
2024-09-04 21:30 ` Usama Arif
2024-09-04 23:10 ` Barry Song
2024-09-04 23:23 ` Usama Arif
2024-09-04 23:27 ` Barry Song
2024-09-04 23:35 ` Yosry Ahmed
2024-09-22 23:57 ` Barry Song
2024-09-23 10:22 ` Usama Arif
2024-09-23 12:10 ` Johannes Weiner
2024-09-23 16:53 ` Yosry Ahmed
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to='CAGsJ_4wGK6pu+KNhYjpWgydp6DyjH5tE=9+mje3UyrXdFJOuNw@mail.gmail.com' \
--to=21cnbao@gmail.com \
--cc=akpm@linux-foundation.org \
--cc=baolin.wang@linux.alibaba.com \
--cc=chrisl@kernel.org \
--cc=david@redhat.com \
--cc=hanchuanhua@oppo.com \
--cc=hannes@cmpxchg.org \
--cc=hch@infradead.org \
--cc=hughd@google.com \
--cc=kaleshsingh@google.com \
--cc=kasong@tencent.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mhocko@suse.com \
--cc=minchan@kernel.org \
--cc=nphamcs@gmail.com \
--cc=ryan.roberts@arm.com \
--cc=ryncsn@gmail.com \
--cc=senozhatsky@chromium.org \
--cc=shakeel.butt@linux.dev \
--cc=shy828301@gmail.com \
--cc=surenb@google.com \
--cc=v-songbaohua@oppo.com \
--cc=willy@infradead.org \
--cc=xiang@kernel.org \
--cc=ying.huang@intel.com \
--cc=yosryahmed@google.com \
--cc=zhengtangquan@oppo.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox