From: Yang Shi <yang.shi@linux.alibaba.com>
To: Michal Hocko <mhocko@kernel.org>,
Andrew Morton <akpm@linux-foundation.org>
Cc: willy@infradead.org, ldufour@linux.vnet.ibm.com,
peterz@infradead.org, mingo@redhat.com, acme@kernel.org,
alexander.shishkin@linux.intel.com, jolsa@redhat.com,
namhyung@kernel.org, tglx@linutronix.de, hpa@zytor.com,
linux-mm@kvack.org, x86@kernel.org, linux-kernel@vger.kernel.org
Subject: Re: [RFC v3 PATCH 4/5] mm: mmap: zap pages with read mmap_sem for large mapping
Date: Tue, 3 Jul 2018 11:22:17 -0700 [thread overview]
Message-ID: <658e4c7b-d426-11ab-ef9a-018579cbf756@linux.alibaba.com> (raw)
In-Reply-To: <20180703060921.GA16767@dhcp22.suse.cz>
On 7/2/18 11:09 PM, Michal Hocko wrote:
> On Mon 02-07-18 13:48:45, Andrew Morton wrote:
>> On Mon, 2 Jul 2018 16:05:02 +0200 Michal Hocko <mhocko@kernel.org> wrote:
>>
>>> On Fri 29-06-18 20:15:47, Andrew Morton wrote:
>>> [...]
>>>> Would one of your earlier designs have addressed all usecases? I
>>>> expect the dumb unmap-a-little-bit-at-a-time approach would have?
>>> It has been already pointed out that this will not work.
>> I said "one of". There were others.
> Well, I was aware only about two potential solutions. Either do the
> heavy lifting under the shared lock and do the rest with the exlusive
> one and this, drop the lock per parts. Maybe I have missed others?
>
>>> You simply
>>> cannot drop the mmap_sem during unmap because another thread could
>>> change the address space under your feet. So you need some form of
>>> VM_DEAD and handle concurrent and conflicting address space operations.
>> Unclear that this is a problem. If a thread does an unmap of a range
>> of virtual address space, there's no guarantee that upon return some
>> other thread has not already mapped new stuff into that address range.
>> So what's changed?
> Well, consider the following scenario:
> Thread A = calling mmap(NULL, sizeA)
> Thread B = calling munmap(addr, sizeB)
>
> They do not use any external synchronization and rely on the atomic
> munmap. Thread B only munmaps range that it knows belongs to it (e.g.
> called mmap in the past). It should be clear that ThreadA should not
> get an address from the addr, sizeB range, right? In the most simple case
> it will not happen. But let's say that the addr, sizeB range has
> unmapped holes for what ever reasons. Now anytime munmap drops the
> exclusive lock after handling one VMA, Thread A might find its sizeA
> range and use it. ThreadB then might remove this new range as soon as it
> gets its exclusive lock again.
I'm a little bit confused here. If ThreadB already has unmapped that
range, then ThreadA uses it. It sounds not like a problem since ThreadB
should just go ahead to handle the next range when it gets its exclusive
lock again, right? I don't think of why ThreadB would re-visit that
range to remove it.
But, if ThreadA uses MAP_FIXED, it might remap some ranges, then ThreadB
remove them. It might trigger SIGSEGV or SIGBUS, but this is not even
guaranteed on vanilla kernel too if the application doesn't do any
synchronization. It all depends on timing.
>
> Is such a code safe? No it is not and I would call it fragile at best
> but people tend to do weird things and atomic munmap behavior is
> something they can easily depend on.
>
> Another example would be an atomic address range probing by
> MAP_FIXED_NOREPLACE. It would simply break for similar reasons.
Yes, I agree, it could simply break MAP_FIXED_NOREPLACE.
Yang
>
> I remember my attempt to make MAP_LOCKED consistent with mlock (if the
> population fails then return -ENOMEM) and that required to drop the
> shared mmap_sem and take it in exclusive mode (because we do not
> have upgrade_read) and Linus was strongly against [1][2] for very
> similar reasons. If you drop the lock you simply do not know what
> happened under your feet.
>
> [1] http://lkml.kernel.org/r/CA+55aFydkG-BgZzry5DrTzueVh9VvEcVJdLV8iOyUphQk=0vpw@mail.gmail.com
> [2] http://lkml.kernel.org/r/CA+55aFyajquhGhw59qNWKGK4dBV0TPmDD7-1XqPo7DZWvO_hPg@mail.gmail.com
next prev parent reply other threads:[~2018-07-03 18:22 UTC|newest]
Thread overview: 48+ messages / expand[flat|nested] mbox.gz Atom feed top
2018-06-29 22:39 [RFC v3 PATCH 0/5] mm: zap pages with read mmap_sem in munmap " Yang Shi
2018-06-29 22:39 ` [RFC v3 PATCH 1/5] uprobes: make vma_has_uprobes non-static Yang Shi
2018-06-29 22:39 ` [RFC v3 PATCH 2/5] mm: introduce VM_DEAD flag Yang Shi
2018-07-02 13:40 ` Michal Hocko
2018-06-29 22:39 ` [RFC v3 PATCH 3/5] mm: refactor do_munmap() to extract the common part Yang Shi
2018-07-02 13:42 ` Michal Hocko
2018-07-02 16:59 ` Yang Shi
2018-07-02 17:58 ` Michal Hocko
2018-07-02 18:02 ` Yang Shi
2018-06-29 22:39 ` [RFC v3 PATCH 4/5] mm: mmap: zap pages with read mmap_sem for large mapping Yang Shi
2018-06-30 1:28 ` Andrew Morton
2018-06-30 2:10 ` Yang Shi
2018-06-30 1:35 ` Andrew Morton
2018-06-30 2:28 ` Yang Shi
2018-06-30 3:15 ` Andrew Morton
2018-06-30 4:26 ` Yang Shi
2018-07-03 0:01 ` Yang Shi
2018-07-02 14:05 ` Michal Hocko
2018-07-02 20:48 ` Andrew Morton
2018-07-03 6:09 ` Michal Hocko
2018-07-03 16:53 ` Yang Shi
2018-07-03 18:22 ` Yang Shi [this message]
2018-07-04 8:13 ` Michal Hocko
2018-07-02 12:33 ` Kirill A. Shutemov
2018-07-02 12:49 ` Michal Hocko
2018-07-03 8:12 ` Kirill A. Shutemov
2018-07-03 8:27 ` Michal Hocko
2018-07-03 9:19 ` Kirill A. Shutemov
2018-07-03 11:34 ` Michal Hocko
2018-07-03 12:14 ` Kirill A. Shutemov
2018-07-03 17:00 ` Yang Shi
2018-07-02 17:19 ` Yang Shi
2018-07-03 8:07 ` Kirill A. Shutemov
2018-07-02 13:53 ` Michal Hocko
2018-07-02 17:07 ` Yang Shi
2018-06-29 22:39 ` [RFC v3 PATCH 5/5] x86: check VM_DEAD flag in page fault Yang Shi
2018-07-02 8:45 ` Laurent Dufour
2018-07-02 12:15 ` Michal Hocko
2018-07-02 12:26 ` Laurent Dufour
2018-07-02 12:45 ` Michal Hocko
2018-07-02 13:33 ` Laurent Dufour
2018-07-02 13:37 ` Michal Hocko
2018-07-02 17:24 ` Yang Shi
2018-07-02 17:57 ` Michal Hocko
2018-07-02 18:10 ` Yang Shi
2018-07-03 6:17 ` Michal Hocko
2018-07-03 16:50 ` Yang Shi
2018-07-02 13:39 ` [RFC v3 PATCH 0/5] mm: zap pages with read mmap_sem in munmap for large mapping Michal Hocko
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=658e4c7b-d426-11ab-ef9a-018579cbf756@linux.alibaba.com \
--to=yang.shi@linux.alibaba.com \
--cc=acme@kernel.org \
--cc=akpm@linux-foundation.org \
--cc=alexander.shishkin@linux.intel.com \
--cc=hpa@zytor.com \
--cc=jolsa@redhat.com \
--cc=ldufour@linux.vnet.ibm.com \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=mhocko@kernel.org \
--cc=mingo@redhat.com \
--cc=namhyung@kernel.org \
--cc=peterz@infradead.org \
--cc=tglx@linutronix.de \
--cc=willy@infradead.org \
--cc=x86@kernel.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox