linux-mm.kvack.org archive mirror
 help / color / mirror / Atom feed
From: Tetsuo Handa <penguin-kernel@I-love.SAKURA.ne.jp>
To: mhocko@kernel.org
Cc: linux-mm@kvack.org, rientjes@google.com, oleg@redhat.com,
	vdavydov@parallels.com, akpm@linux-foundation.org
Subject: Re: [PATCH 6/6] mm, oom: fortify task_will_free_mem
Date: Wed, 1 Jun 2016 21:04:18 +0900	[thread overview]
Message-ID: <201606012104.FIC12458.FOFHtOFMOSLVJQ@I-love.SAKURA.ne.jp> (raw)
In-Reply-To: <20160601072549.GD26601@dhcp22.suse.cz>

Michal Hocko wrote:
> On Wed 01-06-16 00:29:45, Tetsuo Handa wrote:
> > I'm fine with task_will_free_mem(current) == true case. My question is that
> > "doesn't this patch break task_will_free_mem(current) == false case when there is
> > already TIF_MEMDIE thread" ?
> 
> OK, I see your point now. This is certainly possible, albeit unlikely. I
> think calling this a regression would be a bit an overstatement. We are
> basically replacing one unreliable heuristic by another one which is
> more likely to lead to a deterministic behavior.
> 
> If you are worried about locking up the oom killer I have another 2
> patches on top of this series which should deal with that (one of them
> was already posted [1] and another one was drafted in [2]. Both of them
> on top of this series should remove the concern of the lockup. I just
> wait to post them until this thread settles down.
> 
> [1] http://lkml.kernel.org/r/1464276476-25136-1-git-send-email-mhocko@kernel.org
> [2] http://lkml.kernel.org/r/20160527133502.GN27686@dhcp22.suse.cz

I want [1] though I don't know we should try twice. But I still can't
understand why [2] works. Setting MMF_OOM_REAPED on victim's mm_struct
allows oom_scan_process_thread() to call oom_badness() only after TIF_MEMDIE
was removed from that victim. Since oom_reap_task() is not called due to
can_oom_reap == false, nobody will be able to remove TIF_MEMDIE from that
victim if that victim got stuck at unkillable wait.
down_write_killable(&mm->mmap_sem) might not reduce possibility of lockup
when oom_reap_task() is not called.

Quoting from http://lkml.kernel.org/r/20160530115551.GU22928@dhcp22.suse.cz :
> But your bottom half would just decide to back off the same way I do
> here. And as for the bonus your bottom half would have to do the rather
> costly process iteration to find that out.

Doing the rather costly process iteration for up to one second (though most
of duration is schedule_timeout_idle(HZ/10)) gives that victim some reasonable
grace period for termination before the OOM killer selects next OOM victim.

If we set MMF_OOM_REAPED as of oom_kill_process() while also setting
TIF_MEMDIE, the OOM killer can lock up like described above.
If we set MMF_OOM_REAPED as of oom_kill_process() while not setting
TIF_MEMDIE, the OOM killer will immediately select next OOM victim
which is almost

 enum oom_scan_t oom_scan_process_thread(struct oom_control *oc,
 			struct task_struct *task, unsigned long totalpages)
 {
 	if (oom_unkillable_task(task, NULL, oc->nodemask))
 		return OOM_SCAN_CONTINUE;
 
-	/*
-	 * This task already has access to memory reserves and is being killed.
-	 * Don't allow any other task to have access to the reserves.
-	 */
-	if (!is_sysrq_oom(oc) && atomic_read(&task->signal->oom_victims))
-		return OOM_SCAN_ABORT;
 
 	/*
 	 * If task is allocating a lot of memory and has been marked to be
 	 * killed first if it triggers an oom, then select it.
 	 */
 	if (oom_task_origin(task))
 		return OOM_SCAN_SELECT;
 
 	return OOM_SCAN_OK;
 }

situation when we hit can_oom_reap == false. Since not many thread groups
will hit can_oom_reap == false condition, above situation won't kill all
thread groups. But I think that waiting for one second before backing off
is acceptable from the point of view of least killing. This resembles

Quoting from http://lkml.kernel.org/r/20160601073441.GE26601@dhcp22.suse.cz :
> > > > Third is that oom_kill_process() chooses a thread group which already
> > > > has a TIF_MEMDIE thread when the candidate select_bad_process() chose
> > > > has children because oom_badness() does not take TIF_MEMDIE into account.
> > > > This patch checks child->signal->oom_victims before calling oom_badness()
> > > > if oom_kill_process() was called by SysRq-f case. This resembles making
> > > > sure that oom_badness() is skipped by returning OOM_SCAN_CONTINUE.
> > > 
> > > This makes sense to me as well but why should be limit this to sysrq case?
> > > Does it make any sense to select a child which already got killed for
> > > normal OOM killer? Anyway I think it would be better to split this into
> > > its own patch as well.
> > 
> > The reason is described in next paragraph.
> > Do we prefer immediately killing all children of the allocating task?
> 
> I do not think we want to select them _all_. We haven't been doing that
> and I do not see a reason we should start now. But it surely doesn't
> make any sense to select a task which has already TIF_MEMDIE set.

"although selecting a TIF_MEMDIE thread group forever does not make any
sense, we haven't tried selecting next thread group as soon as some thread
group got TIF_MEMDIE".

Setting MMF_OOM_REAPED and clearing TIF_MEMDIE after some period is the key.

My bottom half does not require user visible changes. If some programs use
clone(CLONE_VM without CLONE_SIGHAND) and mix OOM_SCORE_ADJ_MIN /
OOM_SCORE_ADJ_MAX, I think they have reason they want to do so (e.g.
shrink memory usage when OOM_SCORE_ADJ_MAX thread group was OOM-killed).

--
To unsubscribe, send a message with 'unsubscribe linux-mm' in
the body to majordomo@kvack.org.  For more info on Linux MM,
see: http://www.linux-mm.org/ .
Don't email: <a href=mailto:"dont@kvack.org"> email@kvack.org </a>

  reply	other threads:[~2016-06-01 12:04 UTC|newest]

Thread overview: 45+ messages / expand[flat|nested]  mbox.gz  Atom feed  top
2016-05-30 13:05 [PATCH 0/6 -v2] Handle oom bypass more gracefully Michal Hocko
2016-05-30 13:05 ` [PATCH 1/6] proc, oom: drop bogus task_lock and mm check Michal Hocko
2016-05-30 13:49   ` Vladimir Davydov
2016-05-30 17:43   ` Oleg Nesterov
2016-05-31  7:32     ` Michal Hocko
2016-05-31 22:53       ` Oleg Nesterov
2016-06-01  6:53         ` Michal Hocko
2016-06-01 10:41           ` Tetsuo Handa
2016-06-01 10:48             ` Michal Hocko
2016-05-30 13:05 ` [PATCH 2/6] proc, oom_adj: extract oom_score_adj setting into a helper Michal Hocko
2016-05-30 13:05 ` [PATCH 3/6] mm, oom_adj: make sure processes sharing mm have same view of oom_score_adj Michal Hocko
2016-05-31  7:41   ` Michal Hocko
2016-05-30 13:05 ` [PATCH 4/6] mm, oom: skip vforked tasks from being selected Michal Hocko
2016-05-30 19:28   ` Oleg Nesterov
2016-05-31  7:42     ` Michal Hocko
2016-05-31 21:43       ` Oleg Nesterov
2016-06-01  7:09         ` Michal Hocko
2016-06-01 14:12   ` Tetsuo Handa
2016-06-01 14:25     ` Michal Hocko
2016-06-02 10:45       ` Tetsuo Handa
2016-06-02 11:20         ` Michal Hocko
2016-06-02 11:31           ` Tetsuo Handa
2016-06-02 12:55             ` Michal Hocko
2016-05-30 13:05 ` [PATCH 5/6] mm, oom: kill all tasks sharing the mm Michal Hocko
2016-05-30 18:18   ` Oleg Nesterov
2016-05-31  7:43     ` Michal Hocko
2016-05-31 21:48       ` Oleg Nesterov
2016-05-30 13:05 ` [PATCH 6/6] mm, oom: fortify task_will_free_mem Michal Hocko
2016-05-30 17:35   ` Oleg Nesterov
2016-05-31  7:46     ` Michal Hocko
2016-05-31 22:29       ` Oleg Nesterov
2016-06-01  7:03         ` Michal Hocko
2016-05-31 15:03   ` Tetsuo Handa
2016-05-31 15:10     ` Michal Hocko
2016-05-31 15:29       ` Tetsuo Handa
2016-06-01  7:25         ` Michal Hocko
2016-06-01 12:04           ` Tetsuo Handa [this message]
2016-06-01 12:43             ` Michal Hocko
2016-06-02 14:03 ` [PATCH 7/6] mm, oom: task_will_free_mem should skip oom_reaped tasks Michal Hocko
2016-06-02 15:24   ` Tetsuo Handa
2016-06-02 15:50     ` Michal Hocko
  -- strict thread matches above, loose matches on Subject: below --
2016-05-26 12:40 [PATCH 0/5] Handle oom bypass more gracefully Michal Hocko
2016-05-26 12:40 ` [PATCH 6/6] mm, oom: fortify task_will_free_mem Michal Hocko
     [not found]   ` <201605262311.FFF64092.FFQVtOLOOMJSFH@I-love.SAKURA.ne.jp>
     [not found]     ` <20160526142317.GC23675@dhcp22.suse.cz>
2016-05-26 14:41       ` Tetsuo Handa
2016-05-26 14:56         ` Michal Hocko
2016-05-27 11:07   ` Michal Hocko

Reply instructions:

You may reply publicly to this message via plain-text email
using any one of the following methods:

* Save the following mbox file, import it into your mail client,
  and reply-to-all from there: mbox

  Avoid top-posting and favor interleaved quoting:
  https://en.wikipedia.org/wiki/Posting_style#Interleaved_style

* Reply using the --to, --cc, and --in-reply-to
  switches of git-send-email(1):

  git send-email \
    --in-reply-to=201606012104.FIC12458.FOFHtOFMOSLVJQ@I-love.SAKURA.ne.jp \
    --to=penguin-kernel@i-love.sakura.ne.jp \
    --cc=akpm@linux-foundation.org \
    --cc=linux-mm@kvack.org \
    --cc=mhocko@kernel.org \
    --cc=oleg@redhat.com \
    --cc=rientjes@google.com \
    --cc=vdavydov@parallels.com \
    /path/to/YOUR_REPLY

  https://kernel.org/pub/software/scm/git/docs/git-send-email.html

* If your mail client supports setting the In-Reply-To header
  via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox