From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 67549C61DA4 for ; Fri, 3 Feb 2023 13:32:14 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 039EA6B0073; Fri, 3 Feb 2023 08:32:14 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id F2BCD6B0074; Fri, 3 Feb 2023 08:32:13 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id DF47A6B0075; Fri, 3 Feb 2023 08:32:13 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0012.hostedemail.com [216.40.44.12]) by kanga.kvack.org (Postfix) with ESMTP id D00436B0073 for ; Fri, 3 Feb 2023 08:32:13 -0500 (EST) Received: from smtpin28.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay06.hostedemail.com (Postfix) with ESMTP id A30E8AB870 for ; Fri, 3 Feb 2023 13:32:13 +0000 (UTC) X-FDA: 80426069346.28.9C57252 Received: from mail-pf1-f171.google.com (mail-pf1-f171.google.com [209.85.210.171]) by imf20.hostedemail.com (Postfix) with ESMTP id D076F1C0011 for ; Fri, 3 Feb 2023 13:32:11 +0000 (UTC) Authentication-Results: imf20.hostedemail.com; dkim=pass header.d=gmail.com header.s=20210112 header.b=eTk6xbQ1; spf=pass (imf20.hostedemail.com: domain of shiyn.lin@gmail.com designates 209.85.210.171 as permitted sender) smtp.mailfrom=shiyn.lin@gmail.com; dmarc=pass (policy=none) header.from=gmail.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1675431131; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=D5Js/+CqwdLw7tA73yaGDUZkjTA5vVg5orxMLI68dSg=; b=DI/remwK97Yv7vs/uM7XelA+5lACnU2Uyg/uo2uwrP3A5+sTdgXXI4YaVJLiU5UntTYm+i PsYc6nDFKtrbeNZZ/9PBy4+PQ74Kx4aO2t6sfB/BN2oxKSOEMxX5cUzSlBbq8Aved4QvkW x1O5n0rWAVfTwT97GaU6Rk0XHVYMtnY= ARC-Authentication-Results: i=1; imf20.hostedemail.com; dkim=pass header.d=gmail.com header.s=20210112 header.b=eTk6xbQ1; spf=pass (imf20.hostedemail.com: domain of shiyn.lin@gmail.com designates 209.85.210.171 as permitted sender) smtp.mailfrom=shiyn.lin@gmail.com; dmarc=pass (policy=none) header.from=gmail.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1675431131; a=rsa-sha256; cv=none; b=lFKhc6CgVWsP1kyt9hLT+MLs6erhnmeKO+K00/poKsEB2TOhjy+hjr/mceLIPkg0e/k642 GtFWhNvXUbVx4YS0MR3B+U/6WrPzh4ih04zPd05pX8XgvskVH+t5IKpNdNMdu3Cc67qnLT ft54Cujhtk+gPeu0ikb7M2J0U3en8Yg= Received: by mail-pf1-f171.google.com with SMTP id 203so3550531pfx.6 for ; Fri, 03 Feb 2023 05:32:11 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20210112; h=in-reply-to:content-disposition:mime-version:references:message-id :subject:cc:to:from:date:from:to:cc:subject:date:message-id:reply-to; bh=D5Js/+CqwdLw7tA73yaGDUZkjTA5vVg5orxMLI68dSg=; b=eTk6xbQ1gZf4cGajr0Eq7CiWwmA+crHUzSH6bvyTjoQZF5If01Vc1iuMJ5yYLMtFAz ZakAjNR82DiB0uK4VYakJV0rXC93poKzddhg6VQQbPXavYsraRpWCByF0hK74WAGOMeR az8GIBDt67wy3q+aeVwNYkGmvZ24rQdEeTYRGDjwFwZltPld9M5LkiQBTUC7YLyE7cr5 EMLP3Vbpd4XTcZm17I/3NcY6dRglSYLxQlA5pK3fEjNBsBW8nHl465FT0jWrxwABANBw YmAbADVmktdL3SAokjU2Rn6s8RkMo90k4AyBNR+sYjGpnctKPaxD6HYMH+Mq39LMXJY/ wa8g== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; h=in-reply-to:content-disposition:mime-version:references:message-id :subject:cc:to:from:date:x-gm-message-state:from:to:cc:subject:date :message-id:reply-to; bh=D5Js/+CqwdLw7tA73yaGDUZkjTA5vVg5orxMLI68dSg=; b=SFaN0lX7RlfpgU24q514kcLjNMqm9T+bMtrDZH7mg7lX1Dz07XtCa+I+Hp/cvsgISm 7Sr373KyUsXR8byRO13RWpg3L9IqmIF8f9Q8HT6Xx4Zn0nPcZa00l6/cjTt8bH6+E38y nLmemoxXZoJIgmQfGF2I9edDBiHUTHRQfnhXF6XeDu3rmBI7bs6UyuU9hQfA5OqQXft3 tpUhVg21cHC3TVLOpfYyc9QVM8yXNXjgu7qcZHxRW8xoTUherILnB8k+nZwonqTag+n2 kAGxeBJgG75smmwCWHwZS8fMYSXgaKMRnE4z5BOE5MeNWnXuy0WMp9Hl23bmckp/6fzp v9vw== X-Gm-Message-State: AO0yUKVtlqrhbZIWh1wk3kFrudHKGKqKRwese7qhw6xLLFjy3T/YvL4J fxmt0K6cVhTY1Jei+uccnLI= X-Google-Smtp-Source: AK7set/rOo7muaMJKTXiL62n7xT48BEBwtErmon3K2lTKRiSdLZy7LKsaRTIu30L1OvIRWPpQggp/Q== X-Received: by 2002:a05:6a00:3498:b0:593:1405:943 with SMTP id cp24-20020a056a00349800b0059314050943mr10292492pfb.15.1675431130464; Fri, 03 Feb 2023 05:32:10 -0800 (PST) Received: from strix-laptop ([123.110.9.95]) by smtp.gmail.com with ESMTPSA id o185-20020a62cdc2000000b005815017d348sm1734631pfg.179.2023.02.03.05.32.08 (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256); Fri, 03 Feb 2023 05:32:09 -0800 (PST) Date: Fri, 3 Feb 2023 21:32:04 +0800 From: Chih-En Lin To: Yin Fengwei Cc: willy@infradead.org, david@redhat.com, linux-mm@kvack.org, dave.hansen@intel.com, tim.c.chen@intel.com, ying.huang@intel.com Subject: Re: [RFC PATCH v3 3/4] mm: add do_set_pte_range() Message-ID: References: <20230203131636.1648662-1-fengwei.yin@intel.com> <20230203131636.1648662-4-fengwei.yin@intel.com> MIME-Version: 1.0 Content-Type: text/plain; charset=us-ascii Content-Disposition: inline In-Reply-To: <20230203131636.1648662-4-fengwei.yin@intel.com> X-Rspam-User: X-Rspamd-Server: rspam03 X-Stat-Signature: oe8qqeadb1gpojejhknj587igfn1tzpk X-Rspamd-Queue-Id: D076F1C0011 X-HE-Tag: 1675431131-586136 X-HE-Meta: U2FsdGVkX19l/cyIxRonRDRN06BxW34xjP6wqpk4j0MlVuXEaYM259qkrPSa0Gzzrkft08yQNJtIPCBoXJc90YZaCYkpY19Vks1Me4j2ooV2oC4BytKNV2xnNdOElr3/ZFtIeccOAK33Ytd2eIBYSX1+YvwLFnab7pF6u3JIrUQ/xsExb0MYa+0ZZdrgtPpx1MLPMHXWPOKKYf0ofkOB5gi16eNI3e6qXEKm9uM3jJvIsis8OKC80BbnyXT6jqzyTJ9+3qHaaI6rxxDkZVDhg+EZIvJaX06W/jKROEu8NfTjwX/9Jd6Oc0cJtwSZrvNPZMJMKtQv6qgMXgX08jHrukl16+i5Omk6oYm4Fqc3Yk9s5CPExZivqulM5oPLe3PgINhTmgAP4oKBDiUT2bRqRNKQfuDhqbGX6VIDotbfPgJVa3WDNRMqYQe2bAfgLyfszr7D+ZVEUFy8lC91HHp0dZ0LjGzC89ZXWUYVJn23oc8AvnA1/Bo27MoZSbNvo8wA06jlM4MAp4pw9GvapJOJb4/zVe4A1DoJAUQAW1vGXuMs0WZ37N1a1Eua1O1y8kIIvfeJ0lk2vGFeQlGzOpMV5fIg8h7JN8d0So8VbZiDD2ETTHX7VGwdWfEG7ixUoBI6+4i5M9GR1VbU5fbCR6XuIUj/DAXjvNJeG1lnD0pq/xMNUkEHXxjq74D8X3AAinOHbsYSWGDbCJzYRZk6OBPuAo7dS4zFvEyOJLzLaqKWkA6801TeRebW53b0muviNXTqR9Txs5Gh4N67KXPIJ19ZEBnhGhAqPPbeq4/TmKBSCWatKF40T2+LKjYrqcYIY2N/tt2Jg0Rtbn/+Mxelcf1Di32ZwzMZTPz4nI26d5tJrUdXW6oIWVK/glrA6F0/iR4ROrWP+G8b+jolLRMKLA2EsWi39SsHik+C6FsknPQfpIJEA/RNoNU2iJNHZ2X74Z2qX1HV/9kvqiV0SGKnQNO M6naMOjK 5xG0VEEq4tidrX0+x9weaMNmuwaVSSmS8g6ze5aOb/6Ywq5FO3ciwGxj9wrtZvbPsuyUvsa/KLL8WeOe2RV0LsP4FHGislbLofFW+nJFkITo9tPEOOOZ9aB/Ta8AYBqOLEFtfZb2iRRXgYvLG3PXrR4MMsF6+DEsei99Zjgpu2LFJxQgP/Scx0eDhMCpDZvHG4Jz9aaHeEVCM2h6rd7PoqZakbfIWAlkltq2rUcZro7Y2UrvLHUs58qseX6z/3r2AFdDsj0RKhZbpIyppA34L3pGVvjSr/29ubhCxHlw7RboMlTWiBEjUDgOSaf6zdFhJvneoCZMFxul0XsijGoRZa5wnpBGGHLG849okPTPAyB76ruK37UEVgYHMrAJXxo47RT0kE/FK35bZdd/P/VnK8Ya/UAv2dj2ONIs85XzRkVQLzPq64RpUYlnoWTY5rZ4mBD0B/qgOJ2sm/JDqx2XbElYsuTurNhy1n+0FNd7agwqZNGZgKDOCFq0qu+vG8SLWOdbavEDUcBpg4E9UVPnuxewWlZ4eneeOLaWsQtYAnK3tITA= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On Fri, Feb 03, 2023 at 09:16:35PM +0800, Yin Fengwei wrote: > do_set_pte_range() allows to setup page table entries for a > specific range. It calls folio_add_file_rmap_range() to take > advantage of batched rmap update for large folio. > > Signed-off-by: Yin Fengwei > --- > include/linux/mm.h | 3 +++ > mm/filemap.c | 1 - > mm/memory.c | 59 ++++++++++++++++++++++++++++++---------------- > 3 files changed, 42 insertions(+), 21 deletions(-) > > diff --git a/include/linux/mm.h b/include/linux/mm.h > index d6f8f41514cc..93192f04b276 100644 > --- a/include/linux/mm.h > +++ b/include/linux/mm.h > @@ -1162,6 +1162,9 @@ static inline pte_t maybe_mkwrite(pte_t pte, struct vm_area_struct *vma) > > vm_fault_t do_set_pmd(struct vm_fault *vmf, struct page *page); > void do_set_pte(struct vm_fault *vmf, struct page *page, unsigned long addr); > +void do_set_pte_range(struct vm_fault *vmf, struct folio *folio, > + unsigned long addr, pte_t *pte, > + unsigned long start, unsigned int nr); > > vm_fault_t finish_fault(struct vm_fault *vmf); > vm_fault_t finish_mkwrite_fault(struct vm_fault *vmf); > diff --git a/mm/filemap.c b/mm/filemap.c > index f444684db9f2..74046a3a0ff5 100644 > --- a/mm/filemap.c > +++ b/mm/filemap.c > @@ -3386,7 +3386,6 @@ static vm_fault_t filemap_map_folio_range(struct vm_fault *vmf, > > ref_count++; > do_set_pte(vmf, page, addr); > - update_mmu_cache(vma, addr, vmf->pte); > } while (vmf->pte++, page++, addr += PAGE_SIZE, ++count < nr_pages); > > /* Restore the vmf->pte */ > diff --git a/mm/memory.c b/mm/memory.c > index 7a04a1130ec1..3754b2ef166a 100644 > --- a/mm/memory.c > +++ b/mm/memory.c > @@ -4257,36 +4257,58 @@ vm_fault_t do_set_pmd(struct vm_fault *vmf, struct page *page) > } > #endif > > -void do_set_pte(struct vm_fault *vmf, struct page *page, unsigned long addr) > +void do_set_pte_range(struct vm_fault *vmf, struct folio *folio, > + unsigned long addr, pte_t *pte, > + unsigned long start, unsigned int nr) > { > struct vm_area_struct *vma = vmf->vma; > bool uffd_wp = pte_marker_uffd_wp(vmf->orig_pte); > bool write = vmf->flags & FAULT_FLAG_WRITE; > + bool cow = write && !(vma->vm_flags & VM_SHARED); Why don't use is_cow_mapping()? Is there anything messed up with VM_MAYWRITE? > bool prefault = vmf->address != addr; > + struct page *page = folio_page(folio, start); > pte_t entry; > > - flush_icache_page(vma, page); > - entry = mk_pte(page, vma->vm_page_prot); > + if (!cow) { > + folio_add_file_rmap_range(folio, start, nr, vma, false); > + add_mm_counter(vma->vm_mm, mm_counter_file(page), nr); > + } > > - if (prefault && arch_wants_old_prefaulted_pte()) > - entry = pte_mkold(entry); > - else > - entry = pte_sw_mkyoung(entry); > + do { > + flush_icache_page(vma, page); > + entry = mk_pte(page, vma->vm_page_prot); > > - if (write) > - entry = maybe_mkwrite(pte_mkdirty(entry), vma); > - if (unlikely(uffd_wp)) > - entry = pte_mkuffd_wp(entry); > - /* copy-on-write page */ > - if (write && !(vma->vm_flags & VM_SHARED)) { > + if (prefault && arch_wants_old_prefaulted_pte()) > + entry = pte_mkold(entry); > + else > + entry = pte_sw_mkyoung(entry); > + > + if (write) > + entry = maybe_mkwrite(pte_mkdirty(entry), vma); > + if (unlikely(uffd_wp)) > + entry = pte_mkuffd_wp(entry); > + set_pte_at(vma->vm_mm, addr, pte, entry); > + > + /* no need to invalidate: a not-present page won't be cached */ > + update_mmu_cache(vma, addr, pte); > + } while (pte++, page++, addr += PAGE_SIZE, --nr > 0); > +} > + > +void do_set_pte(struct vm_fault *vmf, struct page *page, unsigned long addr) > +{ > + struct folio *folio = page_folio(page); > + struct vm_area_struct *vma = vmf->vma; > + bool cow = (vmf->flags & FAULT_FLAG_WRITE) && > + !(vma->vm_flags & VM_SHARED); Here too. > + > + if (cow) { > inc_mm_counter(vma->vm_mm, MM_ANONPAGES); > page_add_new_anon_rmap(page, vma, addr); > lru_cache_add_inactive_or_unevictable(page, vma); > - } else { > - inc_mm_counter(vma->vm_mm, mm_counter_file(page)); > - page_add_file_rmap(page, vma, false); > } > - set_pte_at(vma->vm_mm, addr, vmf->pte, entry); > + > + do_set_pte_range(vmf, folio, addr, vmf->pte, > + folio_page_idx(folio, page), 1); > } > > static bool vmf_pte_changed(struct vm_fault *vmf) > @@ -4361,9 +4383,6 @@ vm_fault_t finish_fault(struct vm_fault *vmf) > if (likely(!vmf_pte_changed(vmf))) { > do_set_pte(vmf, page, vmf->address); > > - /* no need to invalidate: a not-present page won't be cached */ > - update_mmu_cache(vma, vmf->address, vmf->pte); > - > ret = 0; > } else { > update_mmu_tlb(vma, vmf->address, vmf->pte); > -- > 2.30.2 > > Thanks, Chih-En Lin