From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 3E204C36010 for ; Fri, 4 Apr 2025 04:54:30 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id D61D26B0011; Fri, 4 Apr 2025 00:54:27 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id CEA226B0012; Fri, 4 Apr 2025 00:54:27 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id B8CF86B0022; Fri, 4 Apr 2025 00:54:27 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0016.hostedemail.com [216.40.44.16]) by kanga.kvack.org (Postfix) with ESMTP id 93E7E6B0011 for ; Fri, 4 Apr 2025 00:54:27 -0400 (EDT) Received: from smtpin27.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay06.hostedemail.com (Postfix) with ESMTP id B6794BB6D6 for ; Fri, 4 Apr 2025 04:54:28 +0000 (UTC) X-FDA: 83295145416.27.2BE3B12 Received: from foss.arm.com (foss.arm.com [217.140.110.172]) by imf05.hostedemail.com (Postfix) with ESMTP id DB27F10000D for ; Fri, 4 Apr 2025 04:54:26 +0000 (UTC) Authentication-Results: imf05.hostedemail.com; dkim=none; dmarc=pass (policy=none) header.from=arm.com; spf=pass (imf05.hostedemail.com: domain of anshuman.khandual@arm.com designates 217.140.110.172 as permitted sender) smtp.mailfrom=anshuman.khandual@arm.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1743742467; a=rsa-sha256; cv=none; b=adh8ZLUOegyPTaLBLaUeqHS+0XV806Y4lUe/sfrErIgMe6X26dnxDrhGryLBC7d2XcrfkW YU+2aXJ8VMTSqm2lmoZp7sTYkX9T1rj9aakeCYuOKV1h8BQyN5mYxaKGZ0PcBfFqHmAJsZ PxXTlnojotJJvxN/OzzoLby+RipeUKI= ARC-Authentication-Results: i=1; imf05.hostedemail.com; dkim=none; dmarc=pass (policy=none) header.from=arm.com; spf=pass (imf05.hostedemail.com: domain of anshuman.khandual@arm.com designates 217.140.110.172 as permitted sender) smtp.mailfrom=anshuman.khandual@arm.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1743742467; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=tfG/L2q3tEaCb0PV+CtwYxhiHfdHGC1ZMO8q5WmuOiw=; b=ZHqwpAtEC4kSU4phymDVneFBhLhKQhzFNTopi6Kym4h2T9TKS6P5tLhGCZtFVDDtt2gdN1 +z55ktMPy28bu4JN412gEOyS2boa4HPstQ208Ct2i/XvU4ZZMs7FR76ytKtM4IxbxTs7zd STNEGjeFEHNIRWFKAlVUv4exPDdUNBA= Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 3AE2A1516; Thu, 3 Apr 2025 21:54:28 -0700 (PDT) Received: from [10.162.40.17] (a077893.blr.arm.com [10.162.40.17]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id E5A633F59E; Thu, 3 Apr 2025 21:54:20 -0700 (PDT) Message-ID: <39d4d1d4-3d22-4d31-bfc6-1d583e9d3f9b@arm.com> Date: Fri, 4 Apr 2025 10:24:17 +0530 MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH v3 10/11] mm/vmalloc: Enter lazy mmu mode while manipulating vmalloc ptes To: Ryan Roberts , Catalin Marinas , Will Deacon , Pasha Tatashin , Andrew Morton , Uladzislau Rezki , Christoph Hellwig , David Hildenbrand , "Matthew Wilcox (Oracle)" , Mark Rutland , Alexandre Ghiti , Kevin Brodsky Cc: linux-arm-kernel@lists.infradead.org, linux-mm@kvack.org, linux-kernel@vger.kernel.org References: <20250304150444.3788920-1-ryan.roberts@arm.com> <20250304150444.3788920-11-ryan.roberts@arm.com> Content-Language: en-US From: Anshuman Khandual In-Reply-To: <20250304150444.3788920-11-ryan.roberts@arm.com> Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 7bit X-Rspam-User: X-Rspamd-Server: rspam03 X-Rspamd-Queue-Id: DB27F10000D X-Stat-Signature: 4h85sqkqpqrto1perrmbue4gar76ysod X-HE-Tag: 1743742466-280658 X-HE-Meta: U2FsdGVkX1+WgDeRSQ777JnoQ8cr3QEeDq3NrH6ZMh4dRK9KgG2cLYzN42AUnCJg2bwizMg6MdBn21TRl19V56g8E3uyCEXLtMC/D4hBPyQ+dPKdlbh9i85d5E6iMKHJiPv7+iGoF5id3qleubj+T5yjTtJt2JN9GTkFJDnla98T9vlW0Q+gSqHyVl3zG87NERo1bSgFpIpyeTgKUQh+wvpYPOoayp7EjT0WB2QBat17HX4ltENRTsM4tl8nhwGWAbgJ0fx+juQKOmWtfoMmLQLd8twBpXHOuIEclI2Kcq4NYCT7oUxb5WrCKsE/E9pRXnLdB7UFZWU34831z5Oz8C6VZJLjNbJjs5RMgexXEqd5rjsvMPAOqXyH0Kp6S1wrpsQRAk9on0rfhi/69i00cW6yqDyTgiQ2Dw0iNx/34SJfiLoD2qVHsPomMCwD4VTYc9sDp9+DYEpk8IIBZadv3I0ddYUk6uWfR5gIb4InUxcbof/xPUleEfI9OxoMba+8ikV/fKuw0oVaUfJS7L1dUuh15wCgR6kZiWLh3wZQUj/mpq6fHyoVqxgZ+j8A65azkkN+rKax15wNaXv9cpXaiR+OC+W4wwlbhp5MNR0mRu55wf3mA/IFlBAofpuYvNyXnkxJZ6whkkyC/Pi9dqCnua9zAtjBxMG1S35TI0cTfAieVJ5/0U5Gg4LnYcmnyzx0GD/m1yjfdxpuxD1Ab/g22l93P5IYaVlvB2f+Z1jhN1P57JEKJ7Tu9i9KJBWhN1IHhnpvzj2PXb1hdpTv2j2p7kkvdMcb3Au+TAYwbb+JUyMK7+fSqojr5PBJsmve4ThPOy7Q5i+eBnbTEnT2GJ+EnHEap/F/u8DFyfeRBqfWdaGe/Ir3oqf3XyMOGg1UzcOm2Y7iD1GkJ5IQA/H8WAsHgnbofmbiTAIzknFkHdfIcfbbMn5qcfuIAT3B2YiTmrL++ez1xpJlhK0eKaXhzRd TMMrOB60 E0ARc/5Hdw6iGnAEuYhzFkpRYfSN0sAHyF1PvWSll2k2dkDA/4tXXgHMqaQHDPSTtkVs7EGJEkfY+X+SNgOOGVARoV9ItO5YyHMOxd/bSUInDsJkx+EuJHNcBtlwIMcW6L+qdw0E70Iu4r40gq271s7ODpL+p4WSkGWjzD1jden9sp20KtsPI7uEOrhiFEJ41TZe2HxNZ7ObiB+Q= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: On 3/4/25 20:34, Ryan Roberts wrote: > Wrap vmalloc's pte table manipulation loops with > arch_enter_lazy_mmu_mode() / arch_leave_lazy_mmu_mode(). This provides > the arch code with the opportunity to optimize the pte manipulations. > > Note that vmap_pfn() already uses lazy mmu mode since it delegates to > apply_to_page_range() which enters lazy mmu mode for both user and > kernel mappings. > > These hooks will shortly be used by arm64 to improve vmalloc > performance. > > Signed-off-by: Ryan Roberts > --- > mm/vmalloc.c | 14 ++++++++++++++ > 1 file changed, 14 insertions(+) > > diff --git a/mm/vmalloc.c b/mm/vmalloc.c > index 6111ce900ec4..b63ca0b7dd40 100644 > --- a/mm/vmalloc.c > +++ b/mm/vmalloc.c > @@ -104,6 +104,9 @@ static int vmap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, > pte = pte_alloc_kernel_track(pmd, addr, mask); > if (!pte) > return -ENOMEM; > + > + arch_enter_lazy_mmu_mode(); > + > do { > if (unlikely(!pte_none(ptep_get(pte)))) { > if (pfn_valid(pfn)) { > @@ -127,6 +130,8 @@ static int vmap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, > set_pte_at(&init_mm, addr, pte, pfn_pte(pfn, prot)); > pfn++; > } while (pte += PFN_DOWN(size), addr += size, addr != end); > + > + arch_leave_lazy_mmu_mode(); > *mask |= PGTBL_PTE_MODIFIED; > return 0; > } > @@ -354,6 +359,8 @@ static void vunmap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, > unsigned long size = PAGE_SIZE; > > pte = pte_offset_kernel(pmd, addr); > + arch_enter_lazy_mmu_mode(); > + > do { > #ifdef CONFIG_HUGETLB_PAGE > size = arch_vmap_pte_range_unmap_size(addr, pte); > @@ -370,6 +377,8 @@ static void vunmap_pte_range(pmd_t *pmd, unsigned long addr, unsigned long end, > ptent = ptep_get_and_clear(&init_mm, addr, pte); > WARN_ON(!pte_none(ptent) && !pte_present(ptent)); > } while (pte += (size >> PAGE_SHIFT), addr += size, addr != end); > + > + arch_leave_lazy_mmu_mode(); > *mask |= PGTBL_PTE_MODIFIED; > } > > @@ -515,6 +524,9 @@ static int vmap_pages_pte_range(pmd_t *pmd, unsigned long addr, > pte = pte_alloc_kernel_track(pmd, addr, mask); > if (!pte) > return -ENOMEM; > + > + arch_enter_lazy_mmu_mode(); > + > do { > struct page *page = pages[*nr]; > > @@ -528,6 +540,8 @@ static int vmap_pages_pte_range(pmd_t *pmd, unsigned long addr, > set_pte_at(&init_mm, addr, pte, mk_pte(page, prot)); > (*nr)++; > } while (pte++, addr += PAGE_SIZE, addr != end); > + > + arch_leave_lazy_mmu_mode(); > *mask |= PGTBL_PTE_MODIFIED; > return 0; > } Reviewed-by: Anshuman Khandual