From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-8.3 required=3.0 tests=HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_PATCH,MAILING_LIST_MULTI,SIGNED_OFF_BY,SPF_HELO_NONE,SPF_PASS, USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 5AB76C3F2D1 for ; Mon, 2 Mar 2020 03:23:03 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id DD59C24699 for ; Mon, 2 Mar 2020 03:23:01 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org DD59C24699 Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=arm.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id 6D5BB6B0005; Sun, 1 Mar 2020 22:23:01 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id 65FB76B0006; Sun, 1 Mar 2020 22:23:01 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 4FF536B0007; Sun, 1 Mar 2020 22:23:01 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0241.hostedemail.com [216.40.44.241]) by kanga.kvack.org (Postfix) with ESMTP id 2DDF66B0005 for ; Sun, 1 Mar 2020 22:23:01 -0500 (EST) Received: from smtpin08.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay04.hostedemail.com (Postfix) with ESMTP id BA24B23E4A for ; Mon, 2 Mar 2020 03:23:00 +0000 (UTC) X-FDA: 76548975720.08.fear00_375042528a603 X-HE-Tag: fear00_375042528a603 X-Filterd-Recvd-Size: 21435 Received: from foss.arm.com (foss.arm.com [217.140.110.172]) by imf48.hostedemail.com (Postfix) with ESMTP for ; Mon, 2 Mar 2020 03:22:59 +0000 (UTC) Received: from usa-sjc-imap-foss1.foss.arm.com (unknown [10.121.207.14]) by usa-sjc-mx-foss1.foss.arm.com (Postfix) with ESMTP id 259AEFEC; Sun, 1 Mar 2020 19:22:59 -0800 (PST) Received: from [10.163.1.119] (unknown [10.163.1.119]) by usa-sjc-imap-foss1.foss.arm.com (Postfix) with ESMTPSA id BBB5D3F6CF; Sun, 1 Mar 2020 19:22:50 -0800 (PST) From: Anshuman Khandual Subject: Re: [PATCH] mm/debug: Add tests validating arch page table helpers for core features To: Christophe Leroy , linux-mm@kvack.org Cc: Heiko Carstens , Paul Mackerras , "H. Peter Anvin" , linux-riscv@lists.infradead.org, Will Deacon , linux-arch@vger.kernel.org, linux-s390@vger.kernel.org, x86@kernel.org, Mike Rapoport , Christian Borntraeger , Ingo Molnar , Catalin Marinas , linux-snps-arc@lists.infradead.org, Vasily Gorbik , Borislav Petkov , Paul Walmsley , "Kirill A . Shutemov" , Thomas Gleixner , linux-arm-kernel@lists.infradead.org, Vineet Gupta , linux-kernel@vger.kernel.org, Palmer Dabbelt , Andrew Morton , linuxppc-dev@lists.ozlabs.org References: <1582799637-11786-1-git-send-email-anshuman.khandual@arm.com> <2be41c29-500c-50af-f915-1493846ae9e5@c-s.fr> Message-ID: <4343eda9-7df2-a13c-0125-cf784c81ce14@arm.com> Date: Mon, 2 Mar 2020 08:52:52 +0530 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:52.0) Gecko/20100101 Thunderbird/52.9.1 MIME-Version: 1.0 In-Reply-To: <2be41c29-500c-50af-f915-1493846ae9e5@c-s.fr> Content-Type: text/plain; charset=utf-8 Content-Language: en-US Content-Transfer-Encoding: quoted-printable X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On 02/27/2020 04:59 PM, Christophe Leroy wrote: >=20 >=20 > Le 27/02/2020 =C3=A0 11:33, Anshuman Khandual a =C3=A9crit=C2=A0: >> This adds new tests validating arch page table helpers for these follo= wing >> core memory features. These tests create and test specific mapping typ= es at >> various page table levels. >> >> * SPECIAL mapping >> * PROTNONE mapping >> * DEVMAP mapping >> * SOFTDIRTY mapping >> * SWAP mapping >> * MIGRATION mapping >> * HUGETLB mapping >> * THP mapping >> >> Cc: Andrew Morton >> Cc: Mike Rapoport >> Cc: Vineet Gupta >> Cc: Catalin Marinas >> Cc: Will Deacon >> Cc: Benjamin Herrenschmidt >> Cc: Paul Mackerras >> Cc: Michael Ellerman >> Cc: Heiko Carstens >> Cc: Vasily Gorbik >> Cc: Christian Borntraeger >> Cc: Thomas Gleixner >> Cc: Ingo Molnar >> Cc: Borislav Petkov >> Cc: "H. Peter Anvin" >> Cc: Kirill A. Shutemov >> Cc: Paul Walmsley >> Cc: Palmer Dabbelt >> Cc: linux-snps-arc@lists.infradead.org >> Cc: linux-arm-kernel@lists.infradead.org >> Cc: linuxppc-dev@lists.ozlabs.org >> Cc: linux-s390@vger.kernel.org >> Cc: linux-riscv@lists.infradead.org >> Cc: x86@kernel.org >> Cc: linux-arch@vger.kernel.org >> Cc: linux-kernel@vger.kernel.org >> Suggested-by: Catalin Marinas >> Signed-off-by: Anshuman Khandual >> --- >> Tested on arm64 and x86 platforms without any test failures. But this = has >> only been built tested on several other platforms. Individual tests ne= ed >> to be verified on all current enabling platforms for the test i.e s390= , >> ppc32, arc etc. >> >> This patch must be applied on v5.6-rc3 after these patches >> >> 1. https://patchwork.kernel.org/patch/11385057/ >> 2. https://patchwork.kernel.org/patch/11407715/ >> >> OR >> >> This patch must be applied on linux-next (next-20200227) after this pa= tch >> >> 2. https://patchwork.kernel.org/patch/11407715/ >> >> =C2=A0 mm/debug_vm_pgtable.c | 310 +++++++++++++++++++++++++++++++++++= ++++++- >> =C2=A0 1 file changed, 309 insertions(+), 1 deletion(-) >> >> diff --git a/mm/debug_vm_pgtable.c b/mm/debug_vm_pgtable.c >> index 96dd7d574cef..3fb90d5b604e 100644 >> --- a/mm/debug_vm_pgtable.c >> +++ b/mm/debug_vm_pgtable.c >> @@ -41,6 +41,44 @@ >> =C2=A0=C2=A0 * wrprotect(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2= =A0 =3D A write protected and not a write entry >> =C2=A0=C2=A0 * pxx_bad(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0= =3D A mapped and non-table entry >> =C2=A0=C2=A0 * pxx_same(entry1, entry2)=C2=A0=C2=A0=C2=A0 =3D Both ent= ries hold the exact same value >> + * >> + * Specific feature operations >> + * >> + * pte_mkspecial(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D= Creates a special entry at PTE level >> + * pte_special(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D T= ests a special entry at PTE level >> + * >> + * pte_protnone(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D = Tests a no access entry at PTE level >> + * pmd_protnone(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D = Tests a no access entry at PMD level >> + * >> + * pte_mkdevmap(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D = Creates a device entry at PTE level >> + * pmd_mkdevmap(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D = Creates a device entry at PMD level >> + * pud_mkdevmap(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D = Creates a device entry at PUD level >> + * pte_devmap(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D Te= sts a device entry at PTE level >> + * pmd_devmap(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D Te= sts a device entry at PMD level >> + * pud_devmap(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D Te= sts a device entry at PUD level >> + * >> + * pte_mksoft_dirty(entry)=C2=A0=C2=A0=C2=A0 =3D Creates a soft dirty= entry at PTE level >> + * pmd_mksoft_dirty(entry)=C2=A0=C2=A0=C2=A0 =3D Creates a soft dirty= entry at PMD level >> + * pte_swp_mksoft_dirty(entry)=C2=A0=C2=A0=C2=A0 =3D Creates a soft d= irty swap entry at PTE level >> + * pmd_swp_mksoft_dirty(entry)=C2=A0=C2=A0=C2=A0 =3D Creates a soft d= irty swap entry at PMD level >> + * pte_soft_dirty(entry)=C2=A0=C2=A0=C2=A0 =3D Tests a soft dirty ent= ry at PTE level >> + * pmd_soft_dirty(entry)=C2=A0=C2=A0=C2=A0 =3D Tests a soft dirty ent= ry at PMD level >> + * pte_swp_soft_dirty(entry)=C2=A0=C2=A0=C2=A0 =3D Tests a soft dirty= swap entry at PTE level >> + * pmd_swp_soft_dirty(entry)=C2=A0=C2=A0=C2=A0 =3D Tests a soft dirty= swap entry at PMD level >> + * pte_clear_soft_dirty(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D= Clears a soft dirty entry at PTE level >> + * pmd_clear_soft_dirty(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D= Clears a soft dirty entry at PMD level >> + * pte_swp_clear_soft_dirty(entry) =3D Clears a soft dirty swap entry= at PTE level >> + * pmd_swp_clear_soft_dirty(entry) =3D Clears a soft dirty swap entry= at PMD level >> + * >> + * pte_mkhuge(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D Cr= eates a HugeTLB entry at given level >> + * pte_huge(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D Test= s a HugeTLB entry at given level >> + * >> + * pmd_trans_huge(entry)=C2=A0=C2=A0=C2=A0 =3D Tests a trans huge pag= e at PMD level >> + * pud_trans_huge(entry)=C2=A0=C2=A0=C2=A0 =3D Tests a trans huge pag= e at PUD level >> + * pmd_present(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D T= ests an entry points to memory at PMD level >> + * pud_present(entry)=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 =3D T= ests an entry points to memory at PUD level >> + * pmd_mknotpresent(entry)=C2=A0=C2=A0=C2=A0 =3D Invalidates an PMD e= ntry for MMU >> + * pud_mknotpresent(entry)=C2=A0=C2=A0=C2=A0 =3D Invalidates an PUD e= ntry for MMU >> =C2=A0=C2=A0 */ >> =C2=A0 #define VMFLAGS=C2=A0=C2=A0=C2=A0 (VM_READ|VM_WRITE|VM_EXEC) >> =C2=A0 @@ -287,6 +325,233 @@ static void __init pmd_populate_tests(str= uct mm_struct *mm, pmd_t *pmdp, >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 WARN_ON(pmd_bad(pmd)); >> =C2=A0 } >> =C2=A0 +#ifdef CONFIG_ARCH_HAS_PTE_SPECIAL >=20 > Can we avoid ifdefs unless necessary ? >=20 > In mm/memory.c I see things like the following, it means pte_special() = always exist and a #ifdef is not necessary. True, #ifdef here can be dropped here, done. >=20 > =C2=A0=C2=A0=C2=A0=C2=A0if (IS_ENABLED(CONFIG_ARCH_HAS_PTE_SPECIAL)) { > =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 if (likely(!pte_special(pte)= )) > =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 goto= check_pfn; > =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 if (vma->vm_ops && vma->vm_o= ps->find_special_page) > =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 retu= rn vma->vm_ops->find_special_page(vma, addr); > =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 if (vma->vm_flags & (VM_PFNM= AP | VM_MIXEDMAP)) > =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 retu= rn NULL; > =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 if (is_zero_pfn(pfn)) > =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 retu= rn NULL; > =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 if (pte_devmap(pte)) > =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 retu= rn NULL; >=20 > =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 print_bad_pte(vma, addr, pte= , NULL); > =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 return NULL; > =C2=A0=C2=A0=C2=A0=C2=A0} >=20 >> +static void __init pte_special_tests(unsigned long pfn, pgprot_t prot= ) >> +{ >> +=C2=A0=C2=A0=C2=A0 pte_t pte =3D pfn_pte(pfn, prot); >> + >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pte_special(pte_mkspecial(pte))); >> +} >> +#else >> +static void __init pte_special_tests(unsigned long pfn, pgprot_t prot= ) { } >> +#endif >> + >> +#ifdef CONFIG_NUMA_BALANCING >=20 > Same here, this ifdef shouldn't be necessary because in /include/asm-ge= neric/pgtable.h we have the following, so a if (IS_ENABLED()) should be e= nough. >=20 > #ifndef CONFIG_NUMA_BALANCING > /* > =C2=A0* Technically a PTE can be PROTNONE even when not doing NUMA bala= ncing but > =C2=A0* the only case the kernel cares is for NUMA balancing and is onl= y ever set > =C2=A0* when the VMA is accessible. For PROT_NONE VMAs, the PTEs are no= t marked > =C2=A0* _PAGE_PROTNONE so by by default, implement the helper as "alway= s no". It > =C2=A0* is the responsibility of the caller to distinguish between PROT= _NONE > =C2=A0* protections and NUMA hinting fault protections. > =C2=A0*/ > static inline int pte_protnone(pte_t pte) > { > =C2=A0=C2=A0=C2=A0=C2=A0return 0; > } >=20 > static inline int pmd_protnone(pmd_t pmd) > { > =C2=A0=C2=A0=C2=A0=C2=A0return 0; > } > #endif /* CONFIG_NUMA_BALANCING */ True, #ifdef here can be dropped, done. There is something I had missed before, pfn_pmd() requires #ifdef CONFIG_TRANSPARENT_HUGEPAGE instead. We need a pmd_t here with given prot. We cannot go via pfn_pte() followed by pte_pmd(), as the later is platform specific and not available in general= . >=20 >> +static void __init pte_protnone_tests(unsigned long pfn, pgprot_t pro= t) >> +{ >> +=C2=A0=C2=A0=C2=A0 pte_t pte =3D pfn_pte(pfn, prot); >> + >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pte_protnone(pte)); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pte_present(pte)); >> +} >> + >> +static void __init pmd_protnone_tests(unsigned long pfn, pgprot_t pro= t) >> +{ >> +=C2=A0=C2=A0=C2=A0 pmd_t pmd =3D pfn_pmd(pfn, prot); >> + >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pmd_protnone(pmd)); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pmd_present(pmd)); >> +} >> +#else >> +static void __init pte_protnone_tests(unsigned long pfn, pgprot_t pro= t) { } >> +static void __init pmd_protnone_tests(unsigned long pfn, pgprot_t pro= t) { } >> +#endif >> + >> +#ifdef CONFIG_ARCH_HAS_PTE_DEVMAP >=20 > Same here, in include/linux/mm.h we have: >=20 > #ifndef CONFIG_ARCH_HAS_PTE_DEVMAP > static inline int pte_devmap(pte_t pte) > { > =C2=A0=C2=A0=C2=A0=C2=A0return 0; > } > #endif >=20 >=20 >> +static void __init pte_devmap_tests(unsigned long pfn, pgprot_t prot) >> +{ >> +=C2=A0=C2=A0=C2=A0 pte_t pte =3D pfn_pte(pfn, prot); >> + >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pte_devmap(pte_mkdevmap(pte))); >> +} >> + >> +#ifdef CONFIG_TRANSPARENT_HUGEPAGE >=20 > Same. In inlude/asm-generic/pgtables.h you have: >=20 > #if !defined(CONFIG_ARCH_HAS_PTE_DEVMAP) || !defined(CONFIG_TRANSPARENT= _HUGEPAGE) > static inline int pmd_devmap(pmd_t pmd) > { > =C2=A0=C2=A0=C2=A0=C2=A0return 0; > } > static inline int pud_devmap(pud_t pud) > { > =C2=A0=C2=A0=C2=A0=C2=A0return 0; > } > static inline int pgd_devmap(pgd_t pgd) > { > =C2=A0=C2=A0=C2=A0=C2=A0return 0; > } > #endif >=20 >> +static void __init pmd_devmap_tests(unsigned long pfn, pgprot_t prot) >> +{ >> +=C2=A0=C2=A0=C2=A0 pmd_t pmd =3D pfn_pmd(pfn, prot); >> + >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pmd_devmap(pmd_mkdevmap(pmd))); >> +} >> + >> +#ifdef CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD >=20 > Same, see above Even though pxx_devmap() fallback definitions are present, pxx_mkdevmap() ones are still missing. We will have to add them first as a pre-requisite patch (which might not be popular without any non-debug use case) in orde= r to drop these #ifdefs here. >=20 >> +static void __init pud_devmap_tests(unsigned long pfn, pgprot_t prot) >> +{ >> +=C2=A0=C2=A0=C2=A0 pud_t pud =3D pfn_pud(pfn, prot); >> + >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pud_devmap(pud_mkdevmap(pud))); >> +} >> +#else >> +static void __init pud_devmap_tests(unsigned long pfn, pgprot_t prot)= { } >> +#endif >> +#else >> +static void __init pmd_devmap_tests(unsigned long pfn, pgprot_t prot)= { } >> +static void __init pud_devmap_tests(unsigned long pfn, pgprot_t prot)= { } >> +#endif >> +#else >> +static void __init pte_devmap_tests(unsigned long pfn, pgprot_t prot)= { } >> +static void __init pmd_devmap_tests(unsigned long pfn, pgprot_t prot)= { } >> +static void __init pud_devmap_tests(unsigned long pfn, pgprot_t prot)= { } >> +#endif >> + >> +#ifdef CONFIG_MEM_SOFT_DIRTY >=20 > Same, they always exist, see include/asm-generic/pgtable.h Yeah, this can be dropped. Though will have to again add TRANSPARENT_HUGE= PAGE to protect pfn_pmd() as explained before. >=20 >> +static void __init pte_soft_dirty_tests(unsigned long pfn, pgprot_t p= rot) >> +{ >> +=C2=A0=C2=A0=C2=A0 pte_t pte =3D pfn_pte(pfn, prot); >> + >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pte_soft_dirty(pte_mksoft_dirty(pte))); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(pte_soft_dirty(pte_clear_soft_dirty(pte)))= ; >> +} >> + >> +static void __init pte_swap_soft_dirty_tests(unsigned long pfn, pgpro= t_t prot) >> +{ >> +=C2=A0=C2=A0=C2=A0 pte_t pte =3D pfn_pte(pfn, prot); >> + >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pte_swp_soft_dirty(pte_swp_mksoft_dirty(p= te))); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(pte_swp_soft_dirty(pte_swp_clear_soft_dirt= y(pte))); >> +} >> + >> +#ifdef CONFIG_ARCH_ENABLE_THP_MIGRATION >=20 > Same True, #ifdef here can be dropped, done. >=20 >> +static void __init pmd_soft_dirty_tests(unsigned long pfn, pgprot_t p= rot) >> +{ >> +=C2=A0=C2=A0=C2=A0 pmd_t pmd =3D pfn_pmd(pfn, prot); >> + >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pmd_soft_dirty(pmd_mksoft_dirty(pmd))); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(pmd_soft_dirty(pmd_clear_soft_dirty(pmd)))= ; >> +} >> + >> +static void __init pmd_swap_soft_dirty_tests(unsigned long pfn, pgpro= t_t prot) >> +{ >> +=C2=A0=C2=A0=C2=A0 pmd_t pmd =3D pfn_pmd(pfn, prot); >> + >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pmd_swp_soft_dirty(pmd_swp_mksoft_dirty(p= md))); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(pmd_swp_soft_dirty(pmd_swp_clear_soft_dirt= y(pmd))); >> +} >> +#else >> +static void __init pmd_soft_dirty_tests(unsigned long pfn, pgprot_t p= rot) { } >> +static void __init pmd_swap_soft_dirty_tests(unsigned long pfn, pgpro= t_t prot) >> +{ >> +} >> +#endif >> +#else >> +static void __init pte_soft_dirty_tests(unsigned long pfn, pgprot_t p= rot) { } >> +static void __init pmd_soft_dirty_tests(unsigned long pfn, pgprot_t p= rot) { } >> +static void __init pte_swap_soft_dirty_tests(unsigned long pfn, pgpro= t_t prot) >> +{ >> +} >> +static void __init pmd_swap_soft_dirty_tests(unsigned long pfn, pgpro= t_t prot) >> +{ >> +} >> +#endif >> + >> +static void __init pte_swap_tests(unsigned long pfn, pgprot_t prot) >> +{ >> +=C2=A0=C2=A0=C2=A0 swp_entry_t swp; >> +=C2=A0=C2=A0=C2=A0 pte_t pte; >> + >> +=C2=A0=C2=A0=C2=A0 pte =3D pfn_pte(pfn, prot); >> +=C2=A0=C2=A0=C2=A0 swp =3D __pte_to_swp_entry(pte); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pte_same(pte, __swp_entry_to_pte(swp))); >> +} >> + >> +#ifdef CONFIG_ARCH_ENABLE_THP_MIGRATION >> +static void __init pmd_swap_tests(unsigned long pfn, pgprot_t prot) >> +{ >> +=C2=A0=C2=A0=C2=A0 swp_entry_t swp; >> +=C2=A0=C2=A0=C2=A0 pmd_t pmd; >> + >> +=C2=A0=C2=A0=C2=A0 pmd =3D pfn_pmd(pfn, prot); >> +=C2=A0=C2=A0=C2=A0 swp =3D __pmd_to_swp_entry(pmd); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pmd_same(pmd, __swp_entry_to_pmd(swp))); >> +} >> +#else >> +static void __init pmd_swap_tests(unsigned long pfn, pgprot_t prot) {= } >> +#endif >> + >> +#ifdef CONFIG_MIGRATION >=20 > Same. See include/linux/swapops.h True, #ifdef here can be dropped, done. Though will have to again add back TRANSPARENT_HUGEPAGE to protect pfn_pmd() as explained before. >=20 >> +static void __init swap_migration_tests(struct page *page) >> +{ >> +=C2=A0=C2=A0=C2=A0 swp_entry_t swp; >> + >> +=C2=A0=C2=A0=C2=A0 /* >> +=C2=A0=C2=A0=C2=A0=C2=A0 * make_migration_entry() expects given page = to be >> +=C2=A0=C2=A0=C2=A0=C2=A0 * locked, otherwise it stumbles upon a BUG_O= N(). >> +=C2=A0=C2=A0=C2=A0=C2=A0 */ >> +=C2=A0=C2=A0=C2=A0 __SetPageLocked(page); >> +=C2=A0=C2=A0=C2=A0 swp =3D make_migration_entry(page, 1); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!is_migration_entry(swp)); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!is_write_migration_entry(swp)); >> + >> +=C2=A0=C2=A0=C2=A0 make_migration_entry_read(&swp); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!is_migration_entry(swp)); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(is_write_migration_entry(swp)); >> + >> +=C2=A0=C2=A0=C2=A0 swp =3D make_migration_entry(page, 0); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!is_migration_entry(swp)); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(is_write_migration_entry(swp)); >> +=C2=A0=C2=A0=C2=A0 __ClearPageLocked(page); >> +} >> +#else >> +static void __init swap_migration_tests(struct page *page) { } >> +#endif >> + >> +#ifdef CONFIG_HUGETLB_PAGE >> +static void __init hugetlb_tests(unsigned long pfn, pgprot_t prot) >> +{ >> +#ifdef CONFIG_ARCH_WANT_GENERAL_HUGETLB >> +=C2=A0=C2=A0=C2=A0 pte_t pte =3D pfn_pte(pfn, prot); >> + >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pte_huge(pte_mkhuge(pte))); >=20 > We also need tests on hugepd stuff Sure, but lets discuss this on the other thread. >=20 >> +#endif >> +} >> +#else >> +static void __init hugetlb_tests(unsigned long pfn, pgprot_t prot) { = } >> +#endif >> + >> +#ifdef CONFIG_TRANSPARENT_HUGEPAGE >=20 > Same, see include/asm-generic/pgtable.h This is required to protect pxx_mknotpresent() which does not have a fall back and pfn_pmd()/pfn_pud() helpers have similar situation as well. >=20 >> +static void __init pmd_thp_tests(unsigned long pfn, pgprot_t prot) >> +{ >> +=C2=A0=C2=A0=C2=A0 pmd_t pmd; >> + >> +=C2=A0=C2=A0=C2=A0 /* >> +=C2=A0=C2=A0=C2=A0=C2=A0 * pmd_trans_huge() and pmd_present() must re= turn negative >> +=C2=A0=C2=A0=C2=A0=C2=A0 * after MMU invalidation with pmd_mknotprese= nt(). >> +=C2=A0=C2=A0=C2=A0=C2=A0 */ >> +=C2=A0=C2=A0=C2=A0 pmd =3D pfn_pmd(pfn, prot); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pmd_trans_huge(pmd_mkhuge(pmd))); >> + >> +=C2=A0=C2=A0=C2=A0 /* >> +=C2=A0=C2=A0=C2=A0=C2=A0 * Though platform specific test exclusions a= re not ideal, >> +=C2=A0=C2=A0=C2=A0=C2=A0 * in this case S390 does not define pmd_mkno= tpresent() >> +=C2=A0=C2=A0=C2=A0=C2=A0 * which should be tested on other platforms = enabling THP. >> +=C2=A0=C2=A0=C2=A0=C2=A0 */ >> +#ifndef CONFIG_S390 >> +=C2=A0=C2=A0=C2=A0 WARN_ON(pmd_trans_huge(pmd_mknotpresent(pmd))); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(pmd_present(pmd_mknotpresent(pmd))); >> +#endif >=20 > Can we add a stub on S390 instead ? Actually we dont have to. pmd_mknotpresent() is required for platforms that do not have __HAVE_ARCH_PMDP_INVALIDATE. Hence can wrap this code with !__HAVE_ARCH_PMDP_INVALIDATE to prevent build failures on such platforms like s390. >=20 >> +} >> + >> +#ifdef CONFIG_HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD >=20 > Same ? The problem is, neither pud_mknotpresent() nor pfn_pud() have a generic fallback definition. So will have to keep this #ifdef. >=20 >> +static void __init pud_thp_tests(unsigned long pfn, pgprot_t prot) >> +{ >> +=C2=A0=C2=A0=C2=A0 pud_t pud; >> + >> +=C2=A0=C2=A0=C2=A0 /* >> +=C2=A0=C2=A0=C2=A0=C2=A0 * pud_trans_huge() and pud_present() must re= turn negative >> +=C2=A0=C2=A0=C2=A0=C2=A0 * after MMU invalidation with pud_mknotprese= nt(). >> +=C2=A0=C2=A0=C2=A0=C2=A0 */ >> +=C2=A0=C2=A0=C2=A0 pud =3D pfn_pud(pfn, prot); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(!pud_trans_huge(pud_mkhuge(pud))); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(pud_trans_huge(pud_mknotpresent(pud))); >> +=C2=A0=C2=A0=C2=A0 WARN_ON(pud_present(pud_mknotpresent(pud))); >> +} >> +#else >> +static void __init pud_thp_tests(unsigned long pfn, pgprot_t prot) { = } >> +#endif >> +#else >> +static void __init pmd_thp_tests(unsigned long pfn, pgprot_t prot) { = } >> +static void __init pud_thp_tests(unsigned long pfn, pgprot_t prot) { = } >> +#endif >> + >> =C2=A0 static unsigned long __init get_random_vaddr(void) >> =C2=A0 { >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 unsigned long random_vaddr, random_page= s, total_user_pages; >> @@ -302,13 +567,14 @@ static unsigned long __init get_random_vaddr(voi= d) >> =C2=A0 void __init debug_vm_pgtable(void) >> =C2=A0 { >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 struct mm_struct *mm; >> +=C2=A0=C2=A0=C2=A0 struct page *page; >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 pgd_t *pgdp; >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 p4d_t *p4dp, *saved_p4dp; >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 pud_t *pudp, *saved_pudp; >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 pmd_t *pmdp, *saved_pmdp, pmd; >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 pte_t *ptep; >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 pgtable_t saved_ptep; >> -=C2=A0=C2=A0=C2=A0 pgprot_t prot; >> +=C2=A0=C2=A0=C2=A0 pgprot_t prot, protnone; >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 phys_addr_t paddr; >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 unsigned long vaddr, pte_aligned, pmd_a= ligned; >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 unsigned long pud_aligned, p4d_aligned,= pgd_aligned; >> @@ -322,6 +588,25 @@ void __init debug_vm_pgtable(void) >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 return; >> =C2=A0=C2=A0=C2=A0=C2=A0=C2=A0 } >> =C2=A0 +=C2=A0=C2=A0=C2=A0 /* >> +=C2=A0=C2=A0=C2=A0=C2=A0 * swap_migration_tests() requires a dedicate= d page as it needs to >> +=C2=A0=C2=A0=C2=A0=C2=A0 * be locked before creating a migration entr= y from it. Locking the >> +=C2=A0=C2=A0=C2=A0=C2=A0 * page that actually maps kernel text ('star= t_kernel') can be real >> +=C2=A0=C2=A0=C2=A0=C2=A0 * problematic. Lets allocate a dedicated pag= e explicitly for this >> +=C2=A0=C2=A0=C2=A0=C2=A0 * purpose that will be freed later. >> +=C2=A0=C2=A0=C2=A0=C2=A0 */ >> +=C2=A0=C2=A0=C2=A0 page =3D alloc_page(GFP_KERNEL); >=20 > Can we do the page allocation and freeing in swap_migration_tests() ins= tead ? Although all the resources used in the helpers have been allocated in the= main function itself before being passed down and later freed. But may be just= an exception could be made for swap_migration_tests() function as the alloca= ted page is being exclusively used here. Later on if we need this page for so= me other future tests, then will have to move it back to debug_vm_pgtable().