From: Johannes Weiner <hannes@cmpxchg.org>
To: Chris Down <chris@chrisdown.name>
Cc: Shakeel Butt <shakeelb@google.com>,
Andrew Morton <akpm@linux-foundation.org>,
Tejun Heo <tj@kernel.org>, Michal Hocko <mhocko@kernel.org>,
Linux MM <linux-mm@kvack.org>, Cgroups <cgroups@vger.kernel.org>,
LKML <linux-kernel@vger.kernel.org>,
Kernel Team <kernel-team@fb.com>
Subject: Re: [PATCH] mm, memcg: reclaim more aggressively before high allocator throttling
Date: Thu, 28 May 2020 16:29:44 -0400 [thread overview]
Message-ID: <20200528202944.GA76514@cmpxchg.org> (raw)
In-Reply-To: <20200528194831.GA2017@chrisdown.name>
On Thu, May 28, 2020 at 08:48:31PM +0100, Chris Down wrote:
> Shakeel Butt writes:
> > What was the initial reason to have different behavior in the first place?
>
> This differing behaviour is simply a mistake, it was never intended to be
> this deviate from what happens elsewhere. To that extent this patch is as
> much a bug fix as it is an improvement.
Yes, it was an oversight.
> > > static void high_work_func(struct work_struct *work)
> > > @@ -2378,16 +2384,20 @@ void mem_cgroup_handle_over_high(void)
> > > {
> > > unsigned long penalty_jiffies;
> > > unsigned long pflags;
> > > + unsigned long nr_reclaimed;
> > > unsigned int nr_pages = current->memcg_nr_pages_over_high;
> >
> > Is there any benefit to keep current->memcg_nr_pages_over_high after
> > this change? Why not just use SWAP_CLUSTER_MAX?
It's there for the same reason why try_to_free_pages() takes a reclaim
argument in the first place: we want to make the thread allocating the
most also do the most reclaim work. Consider a thread allocating THPs
in a loop with another thread allocating regular pages.
Remember that all callers loop. They could theoretically all just ask
for SWAP_CLUSTER_MAX pages over and over again.
The idea is to have fairness in most cases, and avoid allocation
failure, premature OOM, and containment failure in the edge cases that
are caused by the inherent raciness of page reclaim.
> I don't feel strongly either way, but current->memcg_nr_pages_over_high can
> be very large for large allocations.
>
> That said, maybe we should just reclaim `max(SWAP_CLUSTER_MAX, current -
> high)` for each loop? I agree that with this design it looks like perhaps we
> don't need it any more.
>
> Johannes, what do you think?
How about this:
Reclaim memcg_nr_pages_over_high in the first iteration, then switch
to SWAP_CLUSTER_MAX in the retries.
This acknowledges that while the page allocator and memory.max reclaim
every time an allocation is made, memory.high is currently batched and
can have larger targets. We want the allocating thread to reclaim at
least the batch size, but beyond that only what's necessary to prevent
premature OOM or failing containment.
Add a comment stating as much.
Once we reclaim memory.high synchronously instead of batched, this
exceptional handling is no longer needed and can be deleted again.
Does that sound reasonable?
next prev parent reply other threads:[~2020-05-28 20:30 UTC|newest]
Thread overview: 40+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-05-20 14:37 Chris Down
2020-05-20 16:07 ` Michal Hocko
2020-05-20 16:51 ` Johannes Weiner
2020-05-20 17:04 ` Michal Hocko
2020-05-20 17:51 ` Johannes Weiner
2020-05-21 7:32 ` Michal Hocko
2020-05-21 13:51 ` Johannes Weiner
2020-05-21 14:22 ` Johannes Weiner
2020-05-21 14:35 ` Michal Hocko
2020-05-21 15:02 ` Chris Down
2020-05-21 16:38 ` Johannes Weiner
2020-05-21 17:37 ` Michal Hocko
2020-05-21 18:45 ` Johannes Weiner
2020-05-28 16:31 ` Michal Hocko
2020-05-28 16:48 ` Chris Down
2020-05-29 7:31 ` Michal Hocko
2020-05-29 10:08 ` Chris Down
2020-05-29 10:14 ` Michal Hocko
2020-05-28 20:11 ` Johannes Weiner
2020-05-20 20:26 ` Chris Down
2020-05-21 7:19 ` Michal Hocko
2020-05-21 11:27 ` Chris Down
2020-05-21 12:04 ` Michal Hocko
2020-05-21 12:23 ` Chris Down
2020-05-21 12:24 ` Chris Down
2020-05-21 12:37 ` Michal Hocko
2020-05-21 12:57 ` Chris Down
2020-05-21 13:05 ` Chris Down
2020-05-21 13:28 ` Michal Hocko
2020-05-21 13:21 ` Michal Hocko
2020-05-21 13:41 ` Chris Down
2020-05-21 13:58 ` Michal Hocko
2020-05-21 14:22 ` Chris Down
2020-05-21 12:28 ` Michal Hocko
2020-05-28 18:02 ` Shakeel Butt
2020-05-28 19:48 ` Chris Down
2020-05-28 20:29 ` Johannes Weiner [this message]
2020-05-28 21:02 ` Shakeel Butt
2020-05-28 21:14 ` Chris Down
2020-05-29 7:25 ` Michal Hocko
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20200528202944.GA76514@cmpxchg.org \
--to=hannes@cmpxchg.org \
--cc=akpm@linux-foundation.org \
--cc=cgroups@vger.kernel.org \
--cc=chris@chrisdown.name \
--cc=kernel-team@fb.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mhocko@kernel.org \
--cc=shakeelb@google.com \
--cc=tj@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox