From: Claudio Imbrenda <imbrenda@linux.ibm.com>
To: Matthew Wilcox <willy@infradead.org>
Cc: Christian Borntraeger <borntraeger@linux.ibm.com>,
Andrew Morton <akpm@linux-foundation.org>,
linux-arch@vger.kernel.org, linux-mm@kvack.org,
linux-kernel@vger.kernel.org,
Gerald Schaefer <gerald.schaefer@linux.ibm.com>,
linux-s390 <linux-s390@vger.kernel.org>
Subject: Re: [PATCH v5 00/38] New page table range API
Date: Tue, 11 Jul 2023 17:24:40 +0200 [thread overview]
Message-ID: <20230711172440.77504856@p-imbrenda> (raw)
In-Reply-To: <ZK1My5hQYC2Kb6G1@casper.infradead.org>
On Tue, 11 Jul 2023 13:36:27 +0100
Matthew Wilcox <willy@infradead.org> wrote:
> On Tue, Jul 11, 2023 at 11:07:06AM +0200, Christian Borntraeger wrote:
> > Am 10.07.23 um 22:43 schrieb Matthew Wilcox (Oracle):
> > > This patchset changes the API used by the MM to set up page table entries.
> > > The four APIs are:
> > > set_ptes(mm, addr, ptep, pte, nr)
> > > update_mmu_cache_range(vma, addr, ptep, nr)
> > > flush_dcache_folio(folio)
> > > flush_icache_pages(vma, page, nr)
> > >
> > > flush_dcache_folio() isn't technically new, but no architecture
> > > implemented it, so I've done that for them. The old APIs remain around
> > > but are mostly implemented by calling the new interfaces.
> > >
> > > The new APIs are based around setting up N page table entries at once.
> > > The N entries belong to the same PMD, the same folio and the same VMA,
> > > so ptep++ is a legitimate operation, and locking is taken care of for
> > > you. Some architectures can do a better job of it than just a loop,
> > > but I have hesitated to make too deep a change to architectures I don't
> > > understand well.
> > >
> > > One thing I have changed in every architecture is that PG_arch_1 is now a
> > > per-folio bit instead of a per-page bit. This was something that would
> > > have to happen eventually, and it makes sense to do it now rather than
> > > iterate over every page involved in a cache flush and figure out if it
> > > needs to happen.
> >
> > I think we do use PG_arch_1 on s390 for our secure page handling and
> > making this perf folio instead of physical page really seems wrong
> > and it probably breaks this code.
>
> Per-page flags are going away in the next few years, so you're going to
For each 4k physical page frame, we need to keep track whether it is
secure or not.
A bit in struct page seems the most logical choice. If that's not
possible anymore, how would you propose we should do?
> need a new design. s390 seems to do a lot of unusual things. I wish
s390 is an unusual architecture. we are working on un-weirding our
code, but it takes time
> you'd talk to the rest of us more.
next prev parent reply other threads:[~2023-07-11 15:25 UTC|newest]
Thread overview: 66+ messages / expand[flat|nested] mbox.gz Atom feed top
2023-07-10 20:43 Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 01/38] minmax: Add in_range() macro Matthew Wilcox (Oracle)
2023-07-10 23:13 ` Andrew Morton
2023-07-11 2:14 ` Matthew Wilcox
2023-07-11 15:49 ` Andrew Morton
2023-07-24 15:21 ` David Laight
2023-07-11 5:28 ` Christoph Hellwig
2023-07-21 10:14 ` Ryan Roberts
2023-07-10 20:43 ` [PATCH v5 02/38] mm: Convert page_table_check_pte_set() to page_table_check_ptes_set() Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 03/38] mm: Add generic flush_icache_pages() and documentation Matthew Wilcox (Oracle)
2023-07-10 22:23 ` Randy Dunlap
2023-07-10 20:43 ` [PATCH v5 04/38] mm: Add folio_flush_mapping() Matthew Wilcox (Oracle)
2023-07-10 23:17 ` Andrew Morton
2023-07-11 2:33 ` Matthew Wilcox
2023-07-11 16:01 ` Andrew Morton
2023-07-10 20:43 ` [PATCH v5 05/38] mm: Remove ARCH_IMPLEMENTS_FLUSH_DCACHE_FOLIO Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 06/38] mm: Add default definition of set_ptes() Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 07/38] alpha: Implement the new page table range API Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 08/38] arc: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 09/38] arm: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 10/38] arm64: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 11/38] csky: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 12/38] hexagon: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 13/38] ia64: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 14/38] loongarch: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 15/38] m68k: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 16/38] microblaze: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 17/38] mips: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 18/38] nios2: " Matthew Wilcox (Oracle)
2023-07-10 23:08 ` Dinh Nguyen
2023-07-10 20:43 ` [PATCH v5 19/38] openrisc: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 20/38] parisc: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 21/38] powerpc: " Matthew Wilcox (Oracle)
2023-07-11 4:41 ` Christophe Leroy
2023-07-10 20:43 ` [PATCH v5 22/38] riscv: " Matthew Wilcox (Oracle)
2023-07-12 13:58 ` Palmer Dabbelt
2023-07-10 20:43 ` [PATCH v5 23/38] s390: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 24/38] sh: " Matthew Wilcox (Oracle)
2023-07-11 4:00 ` John Paul Adrian Glaubitz
2023-07-11 5:19 ` Yin, Fengwei
2023-07-10 20:43 ` [PATCH v5 25/38] sparc32: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 26/38] sparc64: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 27/38] um: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 28/38] x86: " Matthew Wilcox (Oracle)
2023-07-11 11:51 ` Peter Zijlstra
2023-07-10 20:43 ` [PATCH v5 29/38] xtensa: " Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 30/38] mm: Remove page_mapping_file() Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 31/38] mm: Rationalise flush_icache_pages() and flush_icache_page() Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 32/38] mm: Tidy up set_ptes definition Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 33/38] mm: Use flush_icache_pages() in do_set_pmd() Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 34/38] filemap: Add filemap_map_folio_range() Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 35/38] rmap: add folio_add_file_rmap_range() Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 36/38] mm: Convert do_set_pte() to set_pte_range() Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 37/38] filemap: Batch PTE mappings Matthew Wilcox (Oracle)
2023-07-10 20:43 ` [PATCH v5 38/38] mm: Call update_mmu_cache_range() in more page fault handling paths Matthew Wilcox (Oracle)
2023-07-11 9:07 ` [PATCH v5 00/38] New page table range API Christian Borntraeger
2023-07-11 12:36 ` Matthew Wilcox
2023-07-11 15:24 ` Claudio Imbrenda [this message]
2023-07-11 16:52 ` Andrew Morton
2023-07-11 22:03 ` Matthew Wilcox
2023-07-12 5:29 ` Matthew Wilcox
2023-07-12 8:35 ` Claudio Imbrenda
2023-07-13 10:42 ` Christian Borntraeger
2023-07-13 13:42 ` Matthew Wilcox
2023-07-13 20:27 ` Christian Borntraeger
2023-07-13 21:22 ` Matthew Wilcox
Reply instructions:
You may reply publicly to this message via plain-text email
using any one of the following methods:
* Save the following mbox file, import it into your mail client,
and reply-to-all from there: mbox
Avoid top-posting and favor interleaved quoting:
https://en.wikipedia.org/wiki/Posting_style#Interleaved_style
* Reply using the --to, --cc, and --in-reply-to
switches of git-send-email(1):
git send-email \
--in-reply-to=20230711172440.77504856@p-imbrenda \
--to=imbrenda@linux.ibm.com \
--cc=akpm@linux-foundation.org \
--cc=borntraeger@linux.ibm.com \
--cc=gerald.schaefer@linux.ibm.com \
--cc=linux-arch@vger.kernel.org \
--cc=linux-kernel@vger.kernel.org \
--cc=linux-mm@kvack.org \
--cc=linux-s390@vger.kernel.org \
--cc=willy@infradead.org \
/path/to/YOUR_REPLY
https://kernel.org/pub/software/scm/git/docs/git-send-email.html
* If your mail client supports setting the In-Reply-To header
via mailto: links, try the mailto: link
Be sure your reply has a Subject: header at the top and a blank line
before the message body.
This is a public inbox, see mirroring instructions
for how to clone and mirror all data and code used for this inbox