From: Vlastimil Babka <vbabka@suse.cz>
To: David Rientjes <rientjes@google.com>,
Tetsuo Handa <penguin-kernel@I-love.SAKURA.ne.jp>
Cc: Andrew Morton <akpm@linux-foundation.org>,
linux-mm@kvack.org, linux-kernel@vger.kernel.org,
Mel Gorman <mgorman@techsingularity.net>
Subject: Re: [patch] mm, oom: stop reclaiming if GFP_ATOMIC will start failing soon
Date: Wed, 29 Apr 2020 09:51:39 +0200 [thread overview]
Message-ID: <31f1f84d-c5fe-824b-3c28-1a9ad69fcae5@suse.cz> (raw)
In-Reply-To: <alpine.DEB.2.22.394.2004281436280.131129@chino.kir.corp.google.com>
On 4/28/20 11:48 PM, David Rientjes wrote:
> On Tue, 28 Apr 2020, Vlastimil Babka wrote:
>
> Yes, order-0 reclaim capture is interesting since the issue being reported
> here is userspace going out to lunch because it loops for an unbounded
> amount of time trying to get above a watermark where it's allowed to
> allocate and other consumers are depleting that resource.
>
> We actually prefer to oom kill earlier rather than being put in a
> perpetual state of aggressive reclaim that affects all allocators and the
> unbounded nature of those allocations leads to very poor results for
> everybody.
Sure. My vague impression is that your (and similar cloud companies) kind of
workloads are designed to maximize machine utilization, and overshooting and
killing something as a result is no big deal. Then you perhaps have more
probability of hitting this state, and on the other hand, even an occasional
premature oom kill is not a big deal?
My concers are workloads not designed in such a way, where premature oom kill
due to temporary higher reclaim activity together with burst of incoming network
packets will result in e.g. killing an important database. There, the tradeoff
looks different.
> I'm happy to scope this solely to an order-0 reclaim capture. I'm not
> sure if I'm clear on whether this has been worked on before and patches
> existed in the past?
Andrew mentioned some. I don't recall any, so it might have been before my time.
> Somewhat related to what I described in the changelog: we lost the "page
> allocation stalls" artifacts in the kernel log for 4.15. The commit
> description references an asynchronous mechanism for getting this
> information; I don't know where this mechanism currently lives.
>
next prev parent reply other threads:[~2020-04-29 7:51 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-04-24 20:48 David Rientjes
2020-04-25 0:32 ` Tetsuo Handa
2020-04-26 0:27 ` Andrew Morton
2020-04-26 3:04 ` Tetsuo Handa
2020-04-27 3:12 ` David Rientjes
2020-04-27 5:03 ` Tetsuo Handa
2020-04-27 20:30 ` Andrew Morton
2020-04-27 23:03 ` David Rientjes
2020-04-27 23:35 ` Andrew Morton
2020-04-28 7:43 ` Michal Hocko
2020-04-29 8:31 ` peter enderborg
2020-04-29 9:00 ` Michal Hocko
2020-04-28 9:38 ` Vlastimil Babka
2020-04-28 21:48 ` David Rientjes
2020-04-28 23:37 ` Tetsuo Handa
2020-04-29 7:51 ` Vlastimil Babka [this message]
2020-04-29 9:04 ` Michal Hocko
2020-04-29 10:45 ` Tetsuo Handa
2020-04-29 11:43 ` Michal Hocko
2020-04-27 8:20 ` peter enderborg
2020-04-27 15:01 ` Michal Hocko
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=31f1f84d-c5fe-824b-3c28-1a9ad69fcae5@suse.cz \
--to=vbabka@suse.cz \
--cc=akpm@linux-foundation.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mgorman@techsingularity.net \
--cc=penguin-kernel@I-love.SAKURA.ne.jp \
--cc=rientjes@google.com \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox