From: peter enderborg <peter.enderborg@sony.com>
To: Michal Hocko <mhocko@kernel.org>,
Andrew Morton <akpm@linux-foundation.org>
Cc: David Rientjes <rientjes@google.com>,
Vlastimil Babka <vbabka@suse.cz>, <linux-mm@kvack.org>,
<linux-kernel@vger.kernel.org>
Subject: Re: [patch] mm, oom: stop reclaiming if GFP_ATOMIC will start failing soon
Date: Wed, 29 Apr 2020 10:31:41 +0200 [thread overview]
Message-ID: <94f9b716-b251-79d8-2c8c-70d63a255496@sony.com> (raw)
In-Reply-To: <20200428074301.GK28637@dhcp22.suse.cz>
On 4/28/20 9:43 AM, Michal Hocko wrote:
> On Mon 27-04-20 16:35:58, Andrew Morton wrote:
> [...]
>> No consumer of GFP_ATOMIC memory should consume an unbounded amount of
>> it.
>> Subsystems such as networking will consume a certain amount and
>> will then start recycling it. The total amount in-flight will vary
>> over the longer term as workloads change. A dynamically tuning
>> threshold system will need to adapt rapidly enough to sudden load
>> shifts, which might require unreasonable amounts of headroom.
> I do agree. __GFP_HIGH/__GFP_ATOMIC are bound by the size of the
> reserves under memory pressure. Then allocatios start failing very
> quickly and users have to cope with that, usually by deferring to a
> sleepable context. Tuning reserves dynamically for heavy reserves
> consumers would be possible but I am worried that this is far from
> trivial.
>
> We definitely need to understand what is going on here. Why doesn't
> kswapd + N*direct reclaimers do not provide enough memory to satisfy
> both N threads + reserves consumers? How many times those direct
> reclaimers have to retry?
Was this not supposed to be avoided with PSI, user-space should
a fair change to take actions before it goes bad in user-space?
> We used to have the allocation stall warning as David mentioned in the
> patch description and I have seen it triggering without heavy reserves
> consumers (aka reported free pages corresponded to the min watermark).
> The underlying problem was usually kswapd being stuck on some FS locks,
> direct reclaimers stuck in shrinkers or way too overloaded system with
> dozens if not hundreds of processes stuck in the page allocator each
> racing with the reclaim and betting on luck. The last problem was the
> most annoying because it is really hard to tune for.
next prev parent reply other threads:[~2020-04-29 8:31 UTC|newest]
Thread overview: 21+ messages / expand[flat|nested] mbox.gz Atom feed top
2020-04-24 20:48 David Rientjes
2020-04-25 0:32 ` Tetsuo Handa
2020-04-26 0:27 ` Andrew Morton
2020-04-26 3:04 ` Tetsuo Handa
2020-04-27 3:12 ` David Rientjes
2020-04-27 5:03 ` Tetsuo Handa
2020-04-27 20:30 ` Andrew Morton
2020-04-27 23:03 ` David Rientjes
2020-04-27 23:35 ` Andrew Morton
2020-04-28 7:43 ` Michal Hocko
2020-04-29 8:31 ` peter enderborg [this message]
2020-04-29 9:00 ` Michal Hocko
2020-04-28 9:38 ` Vlastimil Babka
2020-04-28 21:48 ` David Rientjes
2020-04-28 23:37 ` Tetsuo Handa
2020-04-29 7:51 ` Vlastimil Babka
2020-04-29 9:04 ` Michal Hocko
2020-04-29 10:45 ` Tetsuo Handa
2020-04-29 11:43 ` Michal Hocko
2020-04-27 8:20 ` peter enderborg
2020-04-27 15:01 ` Michal Hocko
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=94f9b716-b251-79d8-2c8c-70d63a255496@sony.com \
--to=peter.enderborg@sony.com \
--cc=akpm@linux-foundation.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mhocko@kernel.org \
--cc=rientjes@google.com \
--cc=vbabka@suse.cz \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox