From: Alistair Popple <apopple@nvidia.com>
To: Matthew Wilcox <willy@infradead.org>
Cc: akpm@linux-foundation.org, linux-mm@kvack.org,
linux-kernel@vger.kernel.org
Subject: Re: [PATCH] mm/filemap.c: Always read one page in do_sync_mmap_readahead()
Date: Mon, 20 Jun 2022 19:06:38 +1000 [thread overview]
Message-ID: <87y1xr7kzs.fsf@nvdebian.thelocal> (raw)
In-Reply-To: <87sfof8x1w.fsf@nvdebian.thelocal>
Alistair Popple <apopple@nvidia.com> writes:
> Matthew Wilcox <willy@infradead.org> writes:
>
>> On Tue, Jun 07, 2022 at 06:37:14PM +1000, Alistair Popple wrote:
>>> ---
>>> include/linux/pagemap.h | 7 +++---
>>> mm/filemap.c | 47 +++++++++++++----------------------------
>>> 2 files changed, 18 insertions(+), 36 deletions(-)
>>
>> Love the diffstat ;-)
>>
>>> @@ -3011,14 +3001,8 @@ static struct file *do_sync_mmap_readahead(struct vm_fault *vmf)
>>> }
>>> #endif
>>>
>>> - /* If we don't want any read-ahead, don't bother */
>>> - if (vmf->vma->vm_flags & VM_RAND_READ)
>>> - return fpin;
>>> - if (!ra->ra_pages)
>>> - return fpin;
>>> -
>>> + fpin = maybe_unlock_mmap_for_io(vmf, fpin);
>>> if (vmf->vma->vm_flags & VM_SEQ_READ) {
>>> - fpin = maybe_unlock_mmap_for_io(vmf, fpin);
>>> page_cache_sync_ra(&ractl, ra->ra_pages);
>>> return fpin;
>>> }
>>
>> Good. Could even pull the maybe_unlock_mmap_for_io() all the way to the
>> top of the file and remove it from the VM_HUGEPAGE case?
>
> Good idea. Also while I'm here is there a reason we don't update
> ra->start or mmap_miss for the VM_HUGEPAGE case?
>
>>> @@ -3029,19 +3013,20 @@ static struct file *do_sync_mmap_readahead(struct vm_fault *vmf)
>>> WRITE_ONCE(ra->mmap_miss, ++mmap_miss);
>>>
>>> /*
>>> - * Do we miss much more than hit in this file? If so,
>>> - * stop bothering with read-ahead. It will only hurt.
>>> + * mmap read-around. If we don't want any read-ahead or if we miss more
>>> + * than we hit don't bother with read-ahead and just read a single page.
>>> */
>>> - if (mmap_miss > MMAP_LOTSAMISS)
>>> - return fpin;
>>> + if ((vmf->vma->vm_flags & VM_RAND_READ) ||
>>> + !ra->ra_pages || mmap_miss > MMAP_LOTSAMISS) {
>>> + ra->start = vmf->pgoff;
>>> + ra->size = 1;
>>> + ra->async_size = 0;
>>> + } else {
>>
>> I'd put the:
>> /* mmap read-around */
>> here
>>
>>> + ra->start = max_t(long, 0, vmf->pgoff - ra->ra_pages / 2);
>>> + ra->size = ra->ra_pages;
>>> + ra->async_size = ra->ra_pages / 4;
>>> + }
>>>
>>> - /*
>>> - * mmap read-around
>>> - */
>>> - fpin = maybe_unlock_mmap_for_io(vmf, fpin);
>>> - ra->start = max_t(long, 0, vmf->pgoff - ra->ra_pages / 2);
>>> - ra->size = ra->ra_pages;
>>> - ra->async_size = ra->ra_pages / 4;
>>> ractl._index = ra->start;
>>> page_cache_ra_order(&ractl, ra, 0);
>>> return fpin;
>>> @@ -3145,9 +3130,7 @@ vm_fault_t filemap_fault(struct vm_fault *vmf)
>>> filemap_invalidate_lock_shared(mapping);
>>> mapping_locked = true;
>>> }
>>> - folio = __filemap_get_folio(mapping, index,
>>> - FGP_CREAT|FGP_FOR_MMAP,
>>> - vmf->gfp_mask);
>>> + folio = filemap_get_folio(mapping, index);
>>> if (!folio) {
>>> if (fpin)
>>> goto out_retry;
>>
>> I think we also should remove the filemap_invalidate_lock_shared()
>> here, no?
>
> Right, afaik filemap_invalidate_lock_shared() is needed when
> instantiating pages in the page cache during fault, which this patch
> does via page_cache_ra_order() in do_sync_mmap_readahead() so I think
> you're right about removing it for filemap_get_folio().
>
> However do_sync_mmap_readahead() is the way normal (ie. !VM_RAND_READ)
> pages would get instantiated today. So shouldn't
> filemap_invalidate_lock_shared() be called before
> do_sync_mmap_readahead() anyway? Or am I missing something?
Never mind. I missed that this is normally done further down the call
stack (in page_cache_ra_unbounded()). This makes it somewhat annoying
to do this clean-up though, because to deal with this case:
if (unlikely(!folio_test_uptodate(folio))) {
/*
* The page was in cache and uptodate and now it is not.
* Strange but possible since we didn't hold the page lock all
* the time. Let's drop everything get the invalidate lock and
* try again.
*/
if (!mapping_locked) {
In this change we need to be able to call do_sync_mmap_readahead()
whilst holding invalidate_lock to ensure we can successfully get an
uptodate folio without it being removed by eg. hole punching when the
folio lock is dropped.
I am experimenting with pulling all the filemap_invalidate_lock_shared()
calls further up the stack, but that creates it's own problems.
>> We also need to handle the !folio case differently. Before, if it was
>> gone, that was definitely an OOM. Now if it's gone it might have been
>> truncated, or removed due to memory pressure, or it might be an OOM
>> situation where readahead didn't manage to create the folio.
>
> Good point, thanks for catching that.
prev parent reply other threads:[~2022-06-20 9:16 UTC|newest]
Thread overview: 5+ messages / expand[flat|nested] mbox.gz Atom feed top
2022-06-07 8:37 Alistair Popple
2022-06-07 12:01 ` William Kucharski
2022-06-07 14:01 ` Matthew Wilcox
2022-06-08 6:35 ` Alistair Popple
2022-06-20 9:06 ` Alistair Popple [this message]
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=87y1xr7kzs.fsf@nvdebian.thelocal \
--to=apopple@nvidia.com \
--cc=akpm@linux-foundation.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=willy@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox