From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-15.3 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,NICE_REPLY_A,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED, USER_AGENT_SANE_1 autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 9B0BCC433E0 for ; Wed, 27 Jan 2021 02:10:49 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 16B5F2065D for ; Wed, 27 Jan 2021 02:10:49 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 16B5F2065D Authentication-Results: mail.kernel.org; dmarc=none (p=none dis=none) header.from=huawei.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id AB5E06B0007; Tue, 26 Jan 2021 21:10:48 -0500 (EST) Received: by kanga.kvack.org (Postfix, from userid 40) id A66736B0008; Tue, 26 Jan 2021 21:10:48 -0500 (EST) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 9A3CD6B000A; Tue, 26 Jan 2021 21:10:48 -0500 (EST) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0028.hostedemail.com [216.40.44.28]) by kanga.kvack.org (Postfix) with ESMTP id 83CCA6B0007 for ; Tue, 26 Jan 2021 21:10:48 -0500 (EST) Received: from smtpin27.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay01.hostedemail.com (Postfix) with ESMTP id 4217F180AD807 for ; Wed, 27 Jan 2021 02:10:48 +0000 (UTC) X-FDA: 77749926576.27.fall88_5d0acb127593 Received: from filter.hostedemail.com (10.5.16.251.rfc1918.com [10.5.16.251]) by smtpin27.hostedemail.com (Postfix) with ESMTP id 1BAC23D668 for ; Wed, 27 Jan 2021 02:10:48 +0000 (UTC) X-HE-Tag: fall88_5d0acb127593 X-Filterd-Recvd-Size: 5044 Received: from szxga04-in.huawei.com (szxga04-in.huawei.com [45.249.212.190]) by imf21.hostedemail.com (Postfix) with ESMTP for ; Wed, 27 Jan 2021 02:10:47 +0000 (UTC) Received: from DGGEMS408-HUB.china.huawei.com (unknown [172.30.72.58]) by szxga04-in.huawei.com (SkyGuard) with ESMTP id 4DQRqt0YPLz1607h; Wed, 27 Jan 2021 10:09:30 +0800 (CST) Received: from [10.174.179.117] (10.174.179.117) by DGGEMS408-HUB.china.huawei.com (10.3.19.208) with Microsoft SMTP Server id 14.3.498.0; Wed, 27 Jan 2021 10:10:34 +0800 Subject: Re: [PATCH v11 03/13] mm/vmalloc: rename vmap_*_range vmap_pages_*_range To: Nicholas Piggin CC: , , , Jonathan Cameron , Christoph Hellwig , Christophe Leroy , Rick Edgecombe , Ding Tianhong , Christoph Hellwig , , Andrew Morton References: <20210126044510.2491820-1-npiggin@gmail.com> <20210126044510.2491820-4-npiggin@gmail.com> From: Miaohe Lin Message-ID: <01fe6df9-8c92-72c9-94cb-797e11160ad7@huawei.com> Date: Wed, 27 Jan 2021 10:10:33 +0800 User-Agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:78.0) Gecko/20100101 Thunderbird/78.6.0 MIME-Version: 1.0 In-Reply-To: <20210126044510.2491820-4-npiggin@gmail.com> Content-Type: text/plain; charset="utf-8" Content-Language: en-US Content-Transfer-Encoding: 7bit X-Originating-IP: [10.174.179.117] X-CFilter-Loop: Reflected X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: Hi: On 2021/1/26 12:45, Nicholas Piggin wrote: > The vmalloc mapper operates on a struct page * array rather than a > linear physical address, re-name it to make this distinction clear. > > Reviewed-by: Christoph Hellwig > Signed-off-by: Nicholas Piggin > --- > mm/vmalloc.c | 16 ++++++++-------- > 1 file changed, 8 insertions(+), 8 deletions(-) > > diff --git a/mm/vmalloc.c b/mm/vmalloc.c > index 62372f9e0167..7f2f36116980 100644 > --- a/mm/vmalloc.c > +++ b/mm/vmalloc.c > @@ -189,7 +189,7 @@ void unmap_kernel_range_noflush(unsigned long start, unsigned long size) > arch_sync_kernel_mappings(start, end); > } > > -static int vmap_pte_range(pmd_t *pmd, unsigned long addr, > +static int vmap_pages_pte_range(pmd_t *pmd, unsigned long addr, > unsigned long end, pgprot_t prot, struct page **pages, int *nr, > pgtbl_mod_mask *mask) > { > @@ -217,7 +217,7 @@ static int vmap_pte_range(pmd_t *pmd, unsigned long addr, > return 0; > } > > -static int vmap_pmd_range(pud_t *pud, unsigned long addr, > +static int vmap_pages_pmd_range(pud_t *pud, unsigned long addr, > unsigned long end, pgprot_t prot, struct page **pages, int *nr, > pgtbl_mod_mask *mask) > { > @@ -229,13 +229,13 @@ static int vmap_pmd_range(pud_t *pud, unsigned long addr, > return -ENOMEM; > do { > next = pmd_addr_end(addr, end); > - if (vmap_pte_range(pmd, addr, next, prot, pages, nr, mask)) > + if (vmap_pages_pte_range(pmd, addr, next, prot, pages, nr, mask)) > return -ENOMEM; > } while (pmd++, addr = next, addr != end); > return 0; > } > > -static int vmap_pud_range(p4d_t *p4d, unsigned long addr, > +static int vmap_pages_pud_range(p4d_t *p4d, unsigned long addr, > unsigned long end, pgprot_t prot, struct page **pages, int *nr, > pgtbl_mod_mask *mask) > { > @@ -247,13 +247,13 @@ static int vmap_pud_range(p4d_t *p4d, unsigned long addr, > return -ENOMEM; > do { > next = pud_addr_end(addr, end); > - if (vmap_pmd_range(pud, addr, next, prot, pages, nr, mask)) > + if (vmap_pages_pmd_range(pud, addr, next, prot, pages, nr, mask)) > return -ENOMEM; > } while (pud++, addr = next, addr != end); > return 0; > } > > -static int vmap_p4d_range(pgd_t *pgd, unsigned long addr, > +static int vmap_pages_p4d_range(pgd_t *pgd, unsigned long addr, > unsigned long end, pgprot_t prot, struct page **pages, int *nr, > pgtbl_mod_mask *mask) > { > @@ -265,7 +265,7 @@ static int vmap_p4d_range(pgd_t *pgd, unsigned long addr, > return -ENOMEM; > do { > next = p4d_addr_end(addr, end); > - if (vmap_pud_range(p4d, addr, next, prot, pages, nr, mask)) > + if (vmap_pages_pud_range(p4d, addr, next, prot, pages, nr, mask)) > return -ENOMEM; > } while (p4d++, addr = next, addr != end); > return 0; > @@ -306,7 +306,7 @@ int map_kernel_range_noflush(unsigned long addr, unsigned long size, > next = pgd_addr_end(addr, end); > if (pgd_bad(*pgd)) > mask |= PGTBL_PGD_MODIFIED; > - err = vmap_p4d_range(pgd, addr, next, prot, pages, &nr, &mask); > + err = vmap_pages_p4d_range(pgd, addr, next, prot, pages, &nr, &mask); > if (err) > return err; > } while (pgd++, addr = next, addr != end); > Reviewed-by: Miaohe Lin