From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 699CBC369CB for ; Wed, 23 Apr 2025 08:18:51 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 608236B000D; Wed, 23 Apr 2025 04:18:46 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 3A7086B00AA; Wed, 23 Apr 2025 04:18:46 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id EC85A6B00A9; Wed, 23 Apr 2025 04:18:45 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0010.hostedemail.com [216.40.44.10]) by kanga.kvack.org (Postfix) with ESMTP id C41956B00A9 for ; Wed, 23 Apr 2025 04:18:45 -0400 (EDT) Received: from smtpin27.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay10.hostedemail.com (Postfix) with ESMTP id 9B0BEC1817 for ; Wed, 23 Apr 2025 08:18:46 +0000 (UTC) X-FDA: 83364607452.27.8CA8870 Received: from casper.infradead.org (casper.infradead.org [90.155.50.34]) by imf21.hostedemail.com (Postfix) with ESMTP id C2FD41C000A for ; Wed, 23 Apr 2025 08:18:42 +0000 (UTC) Authentication-Results: imf21.hostedemail.com; dkim=pass header.d=infradead.org header.s=casper.20170209 header.b="WEUdl/rV"; dmarc=none; spf=none (imf21.hostedemail.com: domain of BATV+cfb108382ec1185fc08d+7913+infradead.org+dwmw2@casper.srs.infradead.org has no SPF policy when checking 90.155.50.34) smtp.mailfrom=BATV+cfb108382ec1185fc08d+7913+infradead.org+dwmw2@casper.srs.infradead.org ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1745396324; h=from:from:sender:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=NJbaNH4W3yPbLyC+ZTcgmLXA2i0jKLuqPtxQoIRJ/KA=; b=PJjq6m4yBEDRsUGaiUHBIX3mcwkNL1IUxg0+QavxhNCA5Sk1CwE9kzv6MG8uIJZX0dPtou gOrBNvl+15LzdfD1j7mYJI5fCMNT2VZY3sRDgesfy4biDJAsRz/6xRmLcnGgAc/izrh2+4 JpgQeUyN8xKqXMGXhzxMyPjr8ywHGj0= ARC-Authentication-Results: i=1; imf21.hostedemail.com; dkim=pass header.d=infradead.org header.s=casper.20170209 header.b="WEUdl/rV"; dmarc=none; spf=none (imf21.hostedemail.com: domain of BATV+cfb108382ec1185fc08d+7913+infradead.org+dwmw2@casper.srs.infradead.org has no SPF policy when checking 90.155.50.34) smtp.mailfrom=BATV+cfb108382ec1185fc08d+7913+infradead.org+dwmw2@casper.srs.infradead.org ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1745396324; a=rsa-sha256; cv=none; b=cPCgXya1m1rOOvBwcu1xUBDwA6VAvm008ColwFM+XA9Ix37mBfhntppBbnEE+78R6oxAGZ Dy7RpxfGuqmBO5mzdJk1sDhwlwLjgGVN4konqyacPaYsBp/9qMM+ZOU8M++d+H2f5TBqRB k+h0V7RqTDB15BU1Rs0aPSxfnEMt4OM= DKIM-Signature: v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed; d=infradead.org; s=casper.20170209; h=Sender:Content-Transfer-Encoding: Content-Type:MIME-Version:References:In-Reply-To:Message-ID:Date:Subject:Cc: To:From:Reply-To:Content-ID:Content-Description; bh=NJbaNH4W3yPbLyC+ZTcgmLXA2i0jKLuqPtxQoIRJ/KA=; b=WEUdl/rVsKvvnnQFVRM5PtO990 ivSqRMqjmtKIjCWM5FwEH96/6j1Hh1WDWxgrhpqO88J2lqObGFBA+tDf3aQow6apiBiWlFMHp0DWu Dn6TgfjNuA3mltFv9DbPw/HUszTjJAcY2ab2IiYxXpM9edV/ELdlKyFGJwOvuXjVr6+2f/cdzWrJv 6Nvv6WTVHgoJeOPPO9eiSDnWXLjTKPhaUJ4Rud9T5q3BjgR1qABGGc/wh0p60zDs/eN1+Oz1vFktq 6EGwmw0W563E/MMI9W75F+JVUmWbakPH9s9eVFIRdmdc+DhXUFfoFk2ysaklJ+ZLt03Vd2zVW/hND mmKiLwUQ==; Received: from [2001:8b0:10b:1::ebe] (helo=i7.infradead.org) by casper.infradead.org with esmtpsa (Exim 4.98.2 #2 (Red Hat Linux)) id 1u7VJQ-000000081P4-0FRV; Wed, 23 Apr 2025 08:18:32 +0000 Received: from dwoodhou by i7.infradead.org with local (Exim 4.98.1 #2 (Red Hat Linux)) id 1u7VJO-00000002YNn-0kqz; Wed, 23 Apr 2025 09:18:30 +0100 From: David Woodhouse To: Mike Rapoport Cc: Andrew Morton , "Sauerwein, David" , Anshuman Khandual , Ard Biesheuvel , Catalin Marinas , David Hildenbrand , Marc Zyngier , Mark Rutland , Mike Rapoport , Will Deacon , kvmarm@lists.cs.columbia.edu, linux-arm-kernel@lists.infradead.org, linux-kernel@vger.kernel.org, linux-mm@kvack.org, Ruihan Li Subject: [PATCH v3 1/7] mm: Introduce for_each_valid_pfn() and use it from reserve_bootmem_region() Date: Wed, 23 Apr 2025 08:52:43 +0100 Message-ID: <20250423081828.608422-2-dwmw2@infradead.org> X-Mailer: git-send-email 2.49.0 In-Reply-To: <20250423081828.608422-1-dwmw2@infradead.org> References: <20250423081828.608422-1-dwmw2@infradead.org> MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit X-SRS-Rewrite: SMTP reverse-path rewritten from by casper.infradead.org. See http://www.infradead.org/rpr.html X-Stat-Signature: 1uircsyd7c79z9e8eej4qnqobdmg9hib X-Rspamd-Queue-Id: C2FD41C000A X-Rspam-User: X-Rspamd-Server: rspam05 X-HE-Tag: 1745396322-303797 X-HE-Meta: U2FsdGVkX18CZaTQ2Kl2lHwys7+wgEO7iBf4lv9uSCMgXldBfW/dr6fQ2KINVnO0gd3mjg/sG9Tp0xCkQ1tFOtF1nKdelSvmzpL6/sE9qaKLhLL03JVWPndP/AqITkWTuUy7eKWEAsNyrx4Mm/I9PAu8dnaHyI646d6UcrI+RvTP7IwWmqHRGYadwjYTCHfwBIQDFsxFG+3guJD72ZbkULiJpF0mbRGMTFG5ePlfin236PEB9xmAUR6+vOA2/chJKKghqwitXmOAJ51EhPODYGIflz7ic6RBxzHL/jM+bxZ/i3nNSYR8Qu+X15NjmJ2s3LD0hzZZTvIB290CcxRxmRYDJmx8t2pr59P0opfEdR1BXuvmNuLIjCZ85FJK1v8mLdhVZRyk6COnEqGj2JQdSc6hh6lZv285F9A4f3EuUI9jpkf63YpsUm0dDoUs5e57bd6ttv62gavaLavVzGZLSeoZuFs3mAEjpCDOjdx4QGO2fLN/09fpLIjBr/dS7gzOmn3BcBamYqx0Xg/C0qjC0F53A5tF9iFXv/GwiJ0y6ZgFHx1JZgohrFwQkkWhU4g5evkE+ilzph3rAPz0TH8acCqB39tIsZYBCFIl/lkZyHUFPbqzEM98fE1yCPqPXImNm1Eaudi5klYvsLxnn5vXhObRFk770Doivgz91d916FHj6METuxAZp0ZHQ6KjGHYSkK00wmhPEHobpDdzTOpX4ZI0ja2hEwdY9ZiyJBaGG/ZgyEBRBlYvaGFFGKgDy2LnaZyCf4UVplWxWbLrlKEVDFDAY8j7vJNVN0+aVGUkMg4RPKkCN8oOCAgow1bkaUO2Wur+Wph2zDGxAjJ9w0RY2rX7JB6zy1xYXj7Z80VHlLySBm+VFKJaGSSFOdUhVHdJ9Jd+hqr68We+IO9pLhuALm2+tAY7SHlRxZm4cR2DYecPrj46oRD95vVRsxhDzm5dK/ZQW8Zt4Z7uCQVprpK Zl9BHKGy 63mC9Yo754m5ER/DoW0Qy8Rtzj1ybz9JF7VPh6abvcykNjRENR70+XS5UDcmntTNghhz6nOil3SHA6hIrX1Ne9qFpdfdzwRxRAzylJaRGuffZrRNyqLX07gFRVamQwALDVNwI/8PKk5S1BTV3mxGjIAXIhtFA3NeDb4jiDLNArSynYh/8H5dYgIoHO4nMxXNdbsq3zhDC5V/maQrcrvF/ehN538b69HHRir4uJLBgY/kso/7mLhMsI5qSZIGrlsgCzqAffJltDpF2l28eylefUKWLpjxNCzPu7pheGZX5AA3rhObkn4Z46vSH3Aq2vHMocyuFQGuXJvNBMpb4PGLufWDqsEhQuNNmT9bS6tIyXlTCpmOMOsDaZksrZa53Io/DKM6us9HIPtPL4Ks3/xgMSWMlj7+K/PNb5GHxWpOL/HuGc+sfKU15fbIOT3aKH1PKdF9YeUcUVZvS6Ce/ifdiQFNFA50wj/7Wbw538MEggczYXDyFZwiIXuXah5AZM7H4El4nRmiSop+sEZOyCo0+9bwj6VdwWUPhBK82j/UMlwtHJ421GsZXTH9FhIWUMcB6L4hoy0yy+GTDOdin+lk8Wwt0aA== X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: From: David Woodhouse Especially since commit 9092d4f7a1f8 ("memblock: update initialization of reserved pages"), the reserve_bootmem_region() function can spend a significant amount of time iterating over every 4KiB PFN in a range, calling pfn_valid() on each one, and ultimately doing absolutely nothing. On a platform used for virtualization, with large NOMAP regions that eventually get used for guest RAM, this leads to a significant increase in steal time experienced during kexec for a live update. Introduce for_each_valid_pfn() and use it from reserve_bootmem_region(). This implementation is precisely the same naïve loop that the function used to have, but subsequent commits will provide optimised versions for FLATMEM and SPARSEMEM, and this version will remain for those architectures which provide their own pfn_valid() implementation, until/unless they also provide a matching for_each_valid_pfn(). Signed-off-by: David Woodhouse Reviewed-by: Mike Rapoport (Microsoft) --- include/linux/mmzone.h | 10 ++++++++++ mm/mm_init.c | 23 ++++++++++------------- 2 files changed, 20 insertions(+), 13 deletions(-) diff --git a/include/linux/mmzone.h b/include/linux/mmzone.h index 6ccec1bf2896..230a29c2ed1a 100644 --- a/include/linux/mmzone.h +++ b/include/linux/mmzone.h @@ -2177,6 +2177,16 @@ void sparse_init(void); #define subsection_map_init(_pfn, _nr_pages) do {} while (0) #endif /* CONFIG_SPARSEMEM */ +/* + * Fallback case for when the architecture provides its own pfn_valid() but + * not a corresponding for_each_valid_pfn(). + */ +#ifndef for_each_valid_pfn +#define for_each_valid_pfn(_pfn, _start_pfn, _end_pfn) \ + for ((_pfn) = (_start_pfn); (_pfn) < (_end_pfn); (_pfn)++) \ + if (pfn_valid(_pfn)) +#endif + #endif /* !__GENERATING_BOUNDS.H */ #endif /* !__ASSEMBLY__ */ #endif /* _LINUX_MMZONE_H */ diff --git a/mm/mm_init.c b/mm/mm_init.c index 9659689b8ace..41884f2155c4 100644 --- a/mm/mm_init.c +++ b/mm/mm_init.c @@ -777,22 +777,19 @@ static inline void init_deferred_page(unsigned long pfn, int nid) void __meminit reserve_bootmem_region(phys_addr_t start, phys_addr_t end, int nid) { - unsigned long start_pfn = PFN_DOWN(start); - unsigned long end_pfn = PFN_UP(end); + unsigned long pfn; - for (; start_pfn < end_pfn; start_pfn++) { - if (pfn_valid(start_pfn)) { - struct page *page = pfn_to_page(start_pfn); + for_each_valid_pfn (pfn, PFN_DOWN(start), PFN_UP(end)) { + struct page *page = pfn_to_page(pfn); - init_deferred_page(start_pfn, nid); + init_deferred_page(pfn, nid); - /* - * no need for atomic set_bit because the struct - * page is not visible yet so nobody should - * access it yet. - */ - __SetPageReserved(page); - } + /* + * no need for atomic set_bit because the struct + * page is not visible yet so nobody should + * access it yet. + */ + __SetPageReserved(page); } } -- 2.49.0