From: Alex Shi <alex.shi@linux.alibaba.com>
To: Andrew Morton <akpm@linux-foundation.org>
Cc: mgorman@techsingularity.net, tj@kernel.org, hughd@google.com,
khlebnikov@yandex-team.ru, daniel.m.jordan@oracle.com,
yang.shi@linux.alibaba.com, willy@infradead.org,
hannes@cmpxchg.org, lkp@intel.com, linux-mm@kvack.org,
linux-kernel@vger.kernel.org, cgroups@vger.kernel.org,
shakeelb@google.com, iamjoonsoo.kim@lge.com,
richard.weiyang@gmail.com
Subject: Re: [PATCH v13 00/18] per memcg lru lock
Date: Sun, 21 Jun 2020 23:44:47 +0800 [thread overview]
Message-ID: <5561f72b-8f9a-f84e-94a4-600c66084f29@linux.alibaba.com> (raw)
In-Reply-To: <20200620160807.0e0997c3e0e3ca1b18e68a53@linux-foundation.org>
在 2020/6/21 上午7:08, Andrew Morton 写道:
> On Fri, 19 Jun 2020 16:33:38 +0800 Alex Shi <alex.shi@linux.alibaba.com> wrote:
>
>> This is a new version which bases on linux-next, merged much suggestion
>> from Hugh Dickins, from compaction fix to less TestClearPageLRU and
>> comments reverse etc. Thank a lot, Hugh!
>>
>> Johannes Weiner has suggested:
>> "So here is a crazy idea that may be worth exploring:
>>
>> Right now, pgdat->lru_lock protects both PageLRU *and* the lruvec's
>> linked list.
>>
>> Can we make PageLRU atomic and use it to stabilize the lru_lock
>> instead, and then use the lru_lock only serialize list operations?
>
> I don't understand this sentence. How can a per-page flag stabilize a
> per-pgdat spinlock? Perhaps some additional description will help.
Hi Andrew,
Well, above comments miss a context, which lru_lock means new lru_lock on each
of memcg not the current per node lru_lock. Sorry!
Currently the lru bit changed under lru_lock, so isolate a page from lru just
need take lru_lock. New patch will change it with a atomic action alone from
lru_lock, so isolate a page need both actions: TestClearPageLRU and take the
lru_lock. like followings in isolate_lru_page():
The main reason for this comes from isolate_migratepages_block() in compaction.c
we have to take lru bit before lru lock, that serialized the page isolation in
memcg page charge/migration which will change page's lruvec and new lru_lock
in it. The current isolation just take lru lock directly which fails on guard
page's lruvec change(memcg change).
changes in isolate_lru_page():- if (PageLRU(page)) {
+ if (TestClearPageLRU(page)) {
pg_data_t *pgdat = page_pgdat(page);
struct lruvec *lruvec;
+ int lru = page_lru(page);
- spin_lock_irq(&pgdat->lru_lock);
+ get_page(page);
lruvec = mem_cgroup_page_lruvec(page, pgdat);
- if (PageLRU(page)) {
- int lru = page_lru(page);
- get_page(page);
- ClearPageLRU(page);
- del_page_from_lru_list(page, lruvec, lru);
- ret = 0;
- }
+ spin_lock_irq(&pgdat->lru_lock);
+ del_page_from_lru_list(page, lruvec, lru);
spin_unlock_irq(&pgdat->lru_lock);
+ ret = 0;
}
>
>>
>> Following Daniel Jordan's suggestion, I have run 208 'dd' with on 104
>> containers on a 2s * 26cores * HT box with a modefied case:
>> https://git.kernel.org/pub/scm/linux/kernel/git/wfg/vm-scalability.git/tree/case-lru-file-readtwice
>>
>> With this patchset, the readtwice performance increased about 80%
>> in concurrent containers.
>>
>> Thanks Hugh Dickins and Konstantin Khlebnikov, they both brought this
>> idea 8 years ago, and others who give comments as well: Daniel Jordan,
>> Mel Gorman, Shakeel Butt, Matthew Wilcox etc.
>>
>> Thanks for Testing support from Intel 0day and Rong Chen, Fengguang Wu,
>> and Yun Wang. Hugh Dickins also shared his kbuild-swap case. Thanks!
>>
>> ...
>>
>> 24 files changed, 500 insertions(+), 357 deletions(-)
>
> It's a large patchset and afaict the whole point is performance gain.
> 80% in one specialized test sounds nice, but is there a plan for more
> extensive quantification?
Once I got 5% aim7 performance gain on 16 cores machine, and about 20+%
readtwice performance gain. the performance gain is increased a lot following
larger cores.
Is there some suggestion for this?
>
> There isn't much sign of completed review activity here, so I'll go
> into hiding for a while.
>
Yes, it's relatively big. also much of change from comments part. :)
Anyway, thanks for look into!
Thanks
Alex
prev parent reply other threads:[~2020-06-21 15:46 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-06-19 8:33 Alex Shi
2020-06-19 8:33 ` [PATCH v13 01/18] mm/vmscan: remove unnecessary lruvec adding Alex Shi
2020-06-19 8:33 ` [PATCH v13 02/18] mm/page_idle: no unlikely double check for idle page counting Alex Shi
2020-06-19 8:33 ` [PATCH v13 03/18] mm/compaction: correct the comments of compact_defer_shift Alex Shi
2020-06-19 8:33 ` [PATCH v13 04/18] mm/compaction: rename compact_deferred as compact_should_defer Alex Shi
2020-06-19 8:33 ` [PATCH v13 05/18] mm/thp: move lru_add_page_tail func to huge_memory.c Alex Shi
2020-06-19 8:33 ` [PATCH v13 06/18] mm/thp: clean up lru_add_page_tail Alex Shi
2020-06-19 8:33 ` [PATCH v13 07/18] mm/thp: narrow lru locking Alex Shi
2020-06-19 8:33 ` [PATCH v13 08/18] mm/memcg: add debug checking in lock_page_memcg Alex Shi
2020-06-19 8:33 ` [PATCH v13 09/18] mm/swap: fold vm event PGROTATED into pagevec_move_tail_fn Alex Shi
2020-06-19 8:33 ` [PATCH v13 10/18] mm/lru: introduce TestClearPageLRU Alex Shi
2020-06-19 8:33 ` [PATCH v13 11/18] mm/compaction: do page isolation first in compaction Alex Shi
2020-06-19 8:33 ` [PATCH v13 12/18] mm/mlock: reorder isolation sequence during munlock Alex Shi
2020-06-19 8:33 ` [PATCH v13 13/18] mm/swap: serialize memcg changes during pagevec_lru_move_fn Alex Shi
2020-06-19 8:33 ` [PATCH v13 14/18] mm/lru: replace pgdat lru_lock with lruvec lock Alex Shi
2020-06-19 8:33 ` [PATCH v13 15/18] mm/lru: introduce the relock_page_lruvec function Alex Shi
2020-06-19 8:33 ` [PATCH v13 16/18] mm/vmscan: use relock for move_pages_to_lru Alex Shi
2020-06-19 8:33 ` [PATCH v13 17/18] mm/pgdat: remove pgdat lru_lock Alex Shi
2020-06-19 8:33 ` [PATCH v13 18/18] mm/lru: revise the comments of lru_lock Alex Shi
2020-06-20 23:08 ` [PATCH v13 00/18] per memcg lru lock Andrew Morton
2020-06-21 15:44 ` Alex Shi [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=5561f72b-8f9a-f84e-94a4-600c66084f29@linux.alibaba.com \
--to=alex.shi@linux.alibaba.com \
--cc=akpm@linux-foundation.org \
--cc=cgroups@vger.kernel.org \
--cc=daniel.m.jordan@oracle.com \
--cc=hannes@cmpxchg.org \
--cc=hughd@google.com \
--cc=iamjoonsoo.kim@lge.com \
--cc=khlebnikov@yandex-team.ru \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=lkp@intel.com \
--cc=mgorman@techsingularity.net \
--cc=richard.weiyang@gmail.com \
--cc=shakeelb@google.com \
--cc=tj@kernel.org \
--cc=willy@infradead.org \
--cc=yang.shi@linux.alibaba.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox