From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id 88433C369CB for ; Wed, 23 Apr 2025 06:44:13 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 0FA9D6B0005; Wed, 23 Apr 2025 02:44:11 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 0ACDB6B0007; Wed, 23 Apr 2025 02:44:11 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id EB1AA6B0008; Wed, 23 Apr 2025 02:44:10 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0012.hostedemail.com [216.40.44.12]) by kanga.kvack.org (Postfix) with ESMTP id CEFA26B0005 for ; Wed, 23 Apr 2025 02:44:10 -0400 (EDT) Received: from smtpin29.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay06.hostedemail.com (Postfix) with ESMTP id 03695BF855 for ; Wed, 23 Apr 2025 06:44:11 +0000 (UTC) X-FDA: 83364369144.29.626A776 Received: from out30-132.freemail.mail.aliyun.com (out30-132.freemail.mail.aliyun.com [115.124.30.132]) by imf26.hostedemail.com (Postfix) with ESMTP id 9BA9614000D for ; Wed, 23 Apr 2025 06:44:08 +0000 (UTC) Authentication-Results: imf26.hostedemail.com; dkim=pass header.d=linux.alibaba.com header.s=default header.b=cexsE5Vw; dmarc=pass (policy=none) header.from=linux.alibaba.com; spf=pass (imf26.hostedemail.com: domain of baolin.wang@linux.alibaba.com designates 115.124.30.132 as permitted sender) smtp.mailfrom=baolin.wang@linux.alibaba.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1745390650; a=rsa-sha256; cv=none; b=tPzZw+yE+yRDow1ZWKvUSqUoC4IvkzG+cHb1ORVnQkjIfJ+wAE+gXw5RUIrCyAORhRNo0W A7PuC1xeCuhbArvM7QD3D/2cl7eOrhvA2VmLECq9VxXoGr+J/KTG/djkTsHVLNXJoI5frJ 6mm5Q2TdjfGY3myLQKxfyTYjBdoEmGw= ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1745390650; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=JGm9lYL8xvhbTj7FN+hg66vA+WW9yw+RDmjxEEjhhuQ=; b=BQaRpt/RzsWzXmE96URhq/93WPAJVpZ81DZno4BocIyA3iU1X+IhDUe8MPof3yAVGRE14f ex8vPUiTeJb/2PHiKPwU/idAyg92AGm0CqSrRoXeiVDMSaH49ZgnMtS2DmQuFO0cepP2GU OqKxBsiwZ8VgtDQc3Rp5q2HCkbL0Xik= ARC-Authentication-Results: i=1; imf26.hostedemail.com; dkim=pass header.d=linux.alibaba.com header.s=default header.b=cexsE5Vw; dmarc=pass (policy=none) header.from=linux.alibaba.com; spf=pass (imf26.hostedemail.com: domain of baolin.wang@linux.alibaba.com designates 115.124.30.132 as permitted sender) smtp.mailfrom=baolin.wang@linux.alibaba.com DKIM-Signature:v=1; a=rsa-sha256; c=relaxed/relaxed; d=linux.alibaba.com; s=default; t=1745390645; h=Message-ID:Date:MIME-Version:Subject:To:From:Content-Type; bh=JGm9lYL8xvhbTj7FN+hg66vA+WW9yw+RDmjxEEjhhuQ=; b=cexsE5Vwg7LOTDzggIMslW+DCCPJp2BbBVZaKViyIGqiGD0m2CE01xm8cqmKzSwcAk/9v29jNt/T2sGIOstHBnyTpAMbIVchL0JeAzJHauxJSQcdxCVyBuXOxkFo9S51UUYI+BbvcEjwQljLiLZZ3yvvDCuiZp6vyitE5q1noQ0= Received: from 30.74.144.121(mailfrom:baolin.wang@linux.alibaba.com fp:SMTPD_---0WXtRz4h_1745390641 cluster:ay36) by smtp.aliyun-inc.com; Wed, 23 Apr 2025 14:44:02 +0800 Message-ID: Date: Wed, 23 Apr 2025 14:44:01 +0800 MIME-Version: 1.0 User-Agent: Mozilla Thunderbird Subject: Re: [PATCH v4 01/12] introduce khugepaged_collapse_single_pmd to unify khugepaged and madvise_collapse To: Nico Pache , linux-mm@kvack.org, linux-doc@vger.kernel.org, linux-kernel@vger.kernel.org, linux-trace-kernel@vger.kernel.org Cc: akpm@linux-foundation.org, corbet@lwn.net, rostedt@goodmis.org, mhiramat@kernel.org, mathieu.desnoyers@efficios.com, david@redhat.com, baohua@kernel.org, ryan.roberts@arm.com, willy@infradead.org, peterx@redhat.com, ziy@nvidia.com, wangkefeng.wang@huawei.com, usamaarif642@gmail.com, sunnanyong@huawei.com, vishal.moola@gmail.com, thomas.hellstrom@linux.intel.com, yang@os.amperecomputing.com, kirill.shutemov@linux.intel.com, aarcange@redhat.com, raquini@redhat.com, dev.jain@arm.com, anshuman.khandual@arm.com, catalin.marinas@arm.com, tiwai@suse.de, will@kernel.org, dave.hansen@linux.intel.com, jack@suse.cz, cl@gentwo.org, jglisse@google.com, surenb@google.com, zokeefe@google.com, hannes@cmpxchg.org, rientjes@google.com, mhocko@suse.com, rdunlap@infradead.org References: <20250417000238.74567-1-npache@redhat.com> <20250417000238.74567-2-npache@redhat.com> From: Baolin Wang In-Reply-To: <20250417000238.74567-2-npache@redhat.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 7bit X-Rspamd-Server: rspam06 X-Rspamd-Queue-Id: 9BA9614000D X-Stat-Signature: u74fq64kr3kb8wrg8ed5sfq913us9bzg X-Rspam-User: X-HE-Tag: 1745390648-359895 X-HE-Meta: U2FsdGVkX1/h06eHEQp452bPgiHtMDgAT5PDsEsPNix7cW6O/FVVxG/JMys0fjpkjF5SvtclC+7fHykESRfV3fDRTUBkpDI63AB93oZbEerAE3KOMPiQWjXVPvr7OKJvUFY3TkZ4D2SQFMy0F8EhE+lv/aqAU7RLfhQnxQrXCnyZ9AAuOCiVmnK4XwuLvCblwi/27vtRofc2K6/B59KQPJP/sZXUf5QHOZvsAdTdKy9tZn9mBZDEhGhFnsL0kt/s4UY8pb11JlM5SCDedEGQe8ofijwCWXsJbHbUbLxCneLSQeiUYI0SUbKzj8DMHy4HlFRN34zVXZ7wSqK1u1xMH7fKZ6tGZStacQBL2Piv6cinIb9iIOUCW5n0DkOYABwXss73cCEhnaE5VQ+eP1vCAD+XuoPn69fgvyHnN8A5uJ3qAHwejzDWMfeYh5N88RWrDNhEIz+TGIddZpMOCWfxRt++xyYUmanNuHXvgPCF0Pf8Jdj/kw/ZFRxUamC+n+zxTJUl3d4shpC8bk4OG3IUgiQcDnVNszdXqdTj74NkYh/qAEqC+fmuNbPfvvPjN+ND9l7GtnRnoTyFGg08L3Stag8YAClzaqFJ3lUDDUdfj0Hwx2by/sQMJ8fbGa0QXaFfe6wVkAUIWr5TyaWKwErq8vkdrwHtkRhfyMwKl5K+QJFplGhKDxlFvpOZ9gQfSBKhFMAsHG+3yxirh5wF11xluo/61s/0EVQ4fFs6EsVNiTeBRCeBQ6vEpBRlEvouqdAtWiXFgEbqYel8S7qKHVlgyvx3PHWnizjDR6R0gc0C8NdhzfjavArhod7KrlzmT43oUkGyEN+BfdflkRl2Tsw/ITyoOWQtP+6XQrtvKB090IzzGlyuhE+8Xv+89bZOqOezT2BMVmUl+KXJ+ljFhMwxZQJ/codVA57TWUzSHXcadG8UiOPiOdjlrMItGOEMxazMmAxxutFe9MIq/pO3ghF ymH1ZOsu cbcZ8koPonjffGXxwYw1yNTTi1XoOWPslzulCPeddXdPxhhkcNUYwPjKfLZMLFltAxGQBbEkMyCRjToBYpkl5ttQJt5O52H7SR913a5RT7ZLxW24MPj8Y/qj+fFfx3IU+K4583t79yf/aaDLyYXU4YVcg5PiSOHdE8B8JnYohdennPCARcozm2Ve8Df93lGGjW1nVGL1ULUjSaBkxC2O8cm3PgqXWfSGdmQ1uwYtxqxQy08iC3Ek/2j57NJgIsZ/LXuZulp41G0K6j68vSy/taWGHcy+Pxz5aJlEzRWBf6te6vYuCWPuTbmYENUmtoH9O2bSk5KZ0lyjkV2FgC5Uiilw8leDCqjwbGxndOFIvCP54lW2z/UXhsGXU7+GPFU1kDF/RsYOaZbmAUfE= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: On 2025/4/17 08:02, Nico Pache wrote: > The khugepaged daemon and madvise_collapse have two different > implementations that do almost the same thing. > > Create khugepaged_collapse_single_pmd to increase code > reuse and create an entry point for future khugepaged changes. > > Refactor madvise_collapse and khugepaged_scan_mm_slot to use > the new khugepaged_collapse_single_pmd function. > > Signed-off-by: Nico Pache Can you add a prefix 'khugepaged:' for the subject line? > --- > mm/khugepaged.c | 92 ++++++++++++++++++++++++------------------------- > 1 file changed, 46 insertions(+), 46 deletions(-) > > diff --git a/mm/khugepaged.c b/mm/khugepaged.c > index b8838ba8207a..cecadc4239e7 100644 > --- a/mm/khugepaged.c > +++ b/mm/khugepaged.c > @@ -2363,6 +2363,48 @@ static int hpage_collapse_scan_file(struct mm_struct *mm, unsigned long addr, > } > #endif > > +/* > + * Try to collapse a single PMD starting at a PMD aligned addr, and return > + * the results. > + */ > +static int khugepaged_collapse_single_pmd(unsigned long addr, > + struct vm_area_struct *vma, bool *mmap_locked, > + struct collapse_control *cc) > +{ > + int result = SCAN_FAIL; > + struct mm_struct *mm = vma->vm_mm; > + unsigned long tva_flags = cc->is_khugepaged ? TVA_ENFORCE_SYSFS : 0; > + > + if (thp_vma_allowable_order(vma, vma->vm_flags, > + tva_flags, PMD_ORDER)) { We've already checked the thp_vma_allowable_order() before calling this function, why check again? > + if (IS_ENABLED(CONFIG_SHMEM) && !vma_is_anonymous(vma)) { > + struct file *file = get_file(vma->vm_file); > + pgoff_t pgoff = linear_page_index(vma, addr); > + > + mmap_read_unlock(mm); > + *mmap_locked = false; > + result = hpage_collapse_scan_file(mm, addr, file, pgoff, > + cc); > + fput(file); > + if (result == SCAN_PTE_MAPPED_HUGEPAGE) { > + mmap_read_lock(mm); > + if (hpage_collapse_test_exit_or_disable(mm)) > + goto end; > + result = collapse_pte_mapped_thp(mm, addr, > + !cc->is_khugepaged); why drop the following check? if (*result == SCAN_PMD_MAPPED) *result = SCAN_SUCCEED; > + mmap_read_unlock(mm); > + } > + } else { > + result = hpage_collapse_scan_pmd(mm, vma, addr, > + mmap_locked, cc); > + } > + if (cc->is_khugepaged && result == SCAN_SUCCEED) > + ++khugepaged_pages_collapsed; > + } > +end: > + return result; > +} > + > static unsigned int khugepaged_scan_mm_slot(unsigned int pages, int *result, > struct collapse_control *cc) > __releases(&khugepaged_mm_lock) > @@ -2437,33 +2479,9 @@ static unsigned int khugepaged_scan_mm_slot(unsigned int pages, int *result, > VM_BUG_ON(khugepaged_scan.address < hstart || > khugepaged_scan.address + HPAGE_PMD_SIZE > > hend); > - if (IS_ENABLED(CONFIG_SHMEM) && !vma_is_anonymous(vma)) { > - struct file *file = get_file(vma->vm_file); > - pgoff_t pgoff = linear_page_index(vma, > - khugepaged_scan.address); > > - mmap_read_unlock(mm); > - mmap_locked = false; > - *result = hpage_collapse_scan_file(mm, > - khugepaged_scan.address, file, pgoff, cc); > - fput(file); > - if (*result == SCAN_PTE_MAPPED_HUGEPAGE) { > - mmap_read_lock(mm); > - if (hpage_collapse_test_exit_or_disable(mm)) > - goto breakouterloop; > - *result = collapse_pte_mapped_thp(mm, > - khugepaged_scan.address, false); > - if (*result == SCAN_PMD_MAPPED) > - *result = SCAN_SUCCEED; > - mmap_read_unlock(mm); > - } > - } else { > - *result = hpage_collapse_scan_pmd(mm, vma, > - khugepaged_scan.address, &mmap_locked, cc); > - } > - > - if (*result == SCAN_SUCCEED) > - ++khugepaged_pages_collapsed; > + *result = khugepaged_collapse_single_pmd(khugepaged_scan.address, > + vma, &mmap_locked, cc); If the khugepaged_collapse_single_pmd() returns a failure caused by hpage_collapse_test_exit_or_disable(), we should break out of the loop according to the original logic. But you've changed the action in this patch, is this intentional? > > /* move to next address */ > khugepaged_scan.address += HPAGE_PMD_SIZE; > @@ -2783,36 +2801,18 @@ int madvise_collapse(struct vm_area_struct *vma, struct vm_area_struct **prev, > mmap_assert_locked(mm); > memset(cc->node_load, 0, sizeof(cc->node_load)); > nodes_clear(cc->alloc_nmask); > - if (IS_ENABLED(CONFIG_SHMEM) && !vma_is_anonymous(vma)) { > - struct file *file = get_file(vma->vm_file); > - pgoff_t pgoff = linear_page_index(vma, addr); > > - mmap_read_unlock(mm); > - mmap_locked = false; > - result = hpage_collapse_scan_file(mm, addr, file, pgoff, > - cc); > - fput(file); > - } else { > - result = hpage_collapse_scan_pmd(mm, vma, addr, > - &mmap_locked, cc); > - } > + result = khugepaged_collapse_single_pmd(addr, vma, &mmap_locked, cc); > + > if (!mmap_locked) > *prev = NULL; /* Tell caller we dropped mmap_lock */ > > -handle_result: > switch (result) { > case SCAN_SUCCEED: > case SCAN_PMD_MAPPED: > ++thps; > break; > case SCAN_PTE_MAPPED_HUGEPAGE: > - BUG_ON(mmap_locked); > - BUG_ON(*prev); > - mmap_read_lock(mm); > - result = collapse_pte_mapped_thp(mm, addr, true); > - mmap_read_unlock(mm); > - goto handle_result; > - /* Whitelisted set of results where continuing OK */ > case SCAN_PMD_NULL: > case SCAN_PTE_NON_PRESENT: > case SCAN_PTE_UFFD_WP: