From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-12.7 required=3.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,FREEMAIL_FORGED_FROMDOMAIN,FREEMAIL_FROM, HEADER_FROM_DIFFERENT_DOMAINS,INCLUDES_CR_TRAILER,INCLUDES_PATCH, MAILING_LIST_MULTI,SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED autolearn=ham autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id D4123C4707F for ; Tue, 25 May 2021 22:45:46 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 5F3946102A for ; Tue, 25 May 2021 22:45:46 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 5F3946102A Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=gmail.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id 505206B0036; Tue, 25 May 2021 18:45:45 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 4B3B06B006C; Tue, 25 May 2021 18:45:45 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id F3BAC6B006E; Tue, 25 May 2021 18:45:44 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0250.hostedemail.com [216.40.44.250]) by kanga.kvack.org (Postfix) with ESMTP id 92F996B0036 for ; Tue, 25 May 2021 18:45:44 -0400 (EDT) Received: from smtpin08.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay05.hostedemail.com (Postfix) with ESMTP id 27B11181AEF3E for ; Tue, 25 May 2021 22:45:44 +0000 (UTC) X-FDA: 78181237008.08.61600BB Received: from mail-ej1-f51.google.com (mail-ej1-f51.google.com [209.85.218.51]) by imf06.hostedemail.com (Postfix) with ESMTP id 89DBFC0007C7 for ; Tue, 25 May 2021 22:45:37 +0000 (UTC) Received: by mail-ej1-f51.google.com with SMTP id z12so48336079ejw.0 for ; Tue, 25 May 2021 15:45:43 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=gmail.com; s=20161025; h=mime-version:references:in-reply-to:from:date:message-id:subject:to :cc; bh=ZEcQbvMhSoHLjMNo49ldT5vNI3t55RutmZvoGxMuPkg=; b=QGRKMAe+uYrh5/cVYSKtIZIu6Vpw2kbCG+B2z0JnIuuv5ol5NCCEEsxp+m0PGp5yhR YvaB1t+ZXx+v6S2fAp1yyTSWC4uxE5P5d01dgVzdGgEI6hBaAYI/N9UoNtx6kyJJWGf6 BB4JpkXBvrGg04HTrVRzlDrMh/kuJtO72rfwGRqzT21Vp7LUdKzSm7oVvnhHeHsizYqV oW9aHUV4C23+a/mAZ21hB97oUHpEtjj5sYv1rzf50Pnn9/wA7Pb33b/LRBVtbGkRMSHc xBj17wLFHvCgjg8rGK6TL5UafXc1k8LLZ6iPUwkzAX3gglRW3gCBsrldSibSnXLUGL6i 4/cw== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:mime-version:references:in-reply-to:from:date :message-id:subject:to:cc; bh=ZEcQbvMhSoHLjMNo49ldT5vNI3t55RutmZvoGxMuPkg=; b=VVNwKvi00mcM9UeqpJh5u0QxSqGN3X8sDSZJiR7NWdD8+muJgdKdZ/NSZm1ewgImGU swdhXEn/MRIyV66xiTSr8a04kgedgaevja1xKtg3m+fWKCo9fpQTpmPEvKBoXugkMipP dZpGg7TMczVmC3vNfiLpDkRgpbfe4K8Kwtzyji44iYqA1cSpZ3k6bHsIRB6CW/7GH9fl DqomMjv039tk+Bp1pV+OQ2T8eKbdCK4T83b3TgHKnAsr/jTdmCuR24OVvQZXs3o88F6p xHmeGPlkwlzJplFOCHMkgEvZZ5ceAa06LjL5ib7TJyiQoCbPAJy/M0k4MewFHg9M6xFy 5/gg== X-Gm-Message-State: AOAM53351LtCH9k/b6fJ05y1lWUvf2YwaM8rwpFQXLBCDq05o7ATJt6z 8iSiwF+v7e93uKpXxzMvCaHp0hk42aMwRr8OHls= X-Google-Smtp-Source: ABdhPJzfCBahew97KwA/E0nGsS+tI1s/DjBv3C8PfI/XV4OfaQ16fpUc3t30vmolJa2Fks9NRWlK5D9EpMVgBs5drB8= X-Received: by 2002:a17:906:4751:: with SMTP id j17mr18079906ejs.25.1621982742594; Tue, 25 May 2021 15:45:42 -0700 (PDT) MIME-Version: 1.0 References: <20210525162145.3510-1-shy828301@gmail.com> <20210525162145.3510-2-shy828301@gmail.com> In-Reply-To: From: Yang Shi Date: Tue, 25 May 2021 15:45:31 -0700 Message-ID: Subject: Re: [v3 PATCH 2/2] mm: thp: check page_mapped instead of page_mapcount for split To: Hugh Dickins Cc: Zi Yan , "Kirill A. Shutemov" , =?UTF-8?B?SE9SSUdVQ0hJIE5BT1lBKOWggOWPoyDnm7TkuZ8p?= , Wang Yugui , Andrew Morton , Linux MM , Linux Kernel Mailing List Content-Type: text/plain; charset="UTF-8" Authentication-Results: imf06.hostedemail.com; dkim=pass header.d=gmail.com header.s=20161025 header.b=QGRKMAe+; dmarc=pass (policy=none) header.from=gmail.com; spf=pass (imf06.hostedemail.com: domain of shy828301@gmail.com designates 209.85.218.51 as permitted sender) smtp.mailfrom=shy828301@gmail.com X-Stat-Signature: ihoswosuzhfs8m5q58jdqxhxf3j4ts77 X-Rspamd-Queue-Id: 89DBFC0007C7 X-Rspamd-Server: rspam02 X-HE-Tag: 1621982737-732537 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On Tue, May 25, 2021 at 3:06 PM Hugh Dickins wrote: > > On Tue, 25 May 2021, Yang Shi wrote: > > > When debugging the bug reported by Wang Yugui [1], try_to_unmap() may > > return false positive for PTE-mapped THP since page_mapcount() is used > > to check if the THP is unmapped, but it just checks compound mapount and > > head page's mapcount. If the THP is PTE-mapped and head page is not > > mapped, it may return false positive. > > But those false positives did not matter because there was a separate > DEBUG_VM check later. > > It's good to have the link to Wang Yugui's report, but that paragraph > is not really about this patch, as it has evolved now: this patch > consolidates the two DEBUG_VM checks into one VM_WARN_ON_ONCE_PAGE. > > > > > The try_to_unmap() has been changed to void function, so check > > page_mapped() after it. And changed BUG_ON to WARN_ON since it is not a > > fatal issue. > > The change from DEBUG_VM BUG to VM_WARN_ON_ONCE is the most important > part of this, and the reason it's good for stable: and the patch title > ought to highlight that, not the page_mapcount business. Will update the subject and the commit log accordingly. > > > > > [1] https://lore.kernel.org/linux-mm/20210412180659.B9E3.409509F4@e16-tech.com/ > > > > Reviewed-by: Zi Yan > > Signed-off-by: Yang Shi > > This will be required Cc: stable@vger.kernel.org > (but we don't want to Cc them on this mail). > > As I said on the other, I think this should be 1/2 not 2/2. Sure. > > > --- > > v3: Incorporated the comments from Hugh. Keep Zi Yan's reviewed-by tag > > since there is no fundamental change against v2. > > v2: Removed dead code and updated the comment of try_to_unmap() per Zi > > Yan. > > mm/huge_memory.c | 17 +++++------------ > > 1 file changed, 5 insertions(+), 12 deletions(-) > > > > diff --git a/mm/huge_memory.c b/mm/huge_memory.c > > index 80fe642d742d..72d81d8e01b1 100644 > > --- a/mm/huge_memory.c > > +++ b/mm/huge_memory.c > > @@ -2343,6 +2343,8 @@ static void unmap_page(struct page *page) > > ttu_flags |= TTU_SPLIT_FREEZE; > > > > try_to_unmap(page, ttu_flags); > > + > > + VM_WARN_ON_ONCE_PAGE(page_mapped(page), page); > > There is one useful piece of information that dump_page() will not show: > total_mapcount(page). Is there a way of crafting that into the output? > > Not with the macros available, I think. Maybe we should be optimistic > and assume I already have the fixes, so not worth trying to refine the > message (but I'm not entirely convinced of that!). > > The trouble with > if (VM_WARN_ON_ONCE_PAGE(page_mapped(page), page)) > pr_warn("total_mapcount:%d\n", total_mapcount(page)); > is that it's printed regardless of the ONCEness. Another "trouble" > is that it's printed so long after the page_mapped(page) check that > it may be 0 by now - but one can see that as itself informative. We should be able to make dump_page() print total mapcount, right? The dump_page() should be just called in some error paths so taking some extra overhead to dump more information seems harmless, or am I missing something? Of course, this can be done in a separate patch. > > I guess leave it as you have it, I don't see an easy better > (without going back to something like the old contortions). > > > } > > > > static void remap_page(struct page *page, unsigned int nr) > > @@ -2653,7 +2655,7 @@ int split_huge_page_to_list(struct page *page, struct list_head *list) > > struct deferred_split *ds_queue = get_deferred_split_queue(head); > > struct anon_vma *anon_vma = NULL; > > struct address_space *mapping = NULL; > > - int count, mapcount, extra_pins, ret; > > + int mapcount, extra_pins, ret; > > Remove mapcount too. > > > pgoff_t end; > > > > VM_BUG_ON_PAGE(is_huge_zero_page(head), head); > > @@ -2712,7 +2714,6 @@ int split_huge_page_to_list(struct page *page, struct list_head *list) > > } > > > > unmap_page(head); > > - VM_BUG_ON_PAGE(compound_mapcount(head), head); > > > > /* block interrupt reentry in xa_lock and spinlock */ > > local_irq_disable(); > > @@ -2730,7 +2731,6 @@ int split_huge_page_to_list(struct page *page, struct list_head *list) > > > > /* Prevent deferred_split_scan() touching ->_refcount */ > > spin_lock(&ds_queue->split_queue_lock); > > - count = page_count(head); > > mapcount = total_mapcount(head); > > if (!mapcount && page_ref_freeze(head, 1 + extra_pins)) { > > mapcount was useful for printing in the hand-crafted message deleted, > but serves no purpose now: just do the page_ref_freeze() without it. Aha, yes, good catch. If mapcount is not zero, the refcount freeze won't succeed. > > > if (!list_empty(page_deferred_list(head))) { > > @@ -2752,16 +2752,9 @@ int split_huge_page_to_list(struct page *page, struct list_head *list) > > __split_huge_page(page, list, end); > > ret = 0; > > } else { > > - if (IS_ENABLED(CONFIG_DEBUG_VM) && mapcount) { > > - pr_alert("total_mapcount: %u, page_count(): %u\n", > > - mapcount, count); > > - if (PageTail(page)) > > - dump_page(head, NULL); > > - dump_page(page, "total_mapcount(head) > 0"); > > - BUG(); > > - } > > spin_unlock(&ds_queue->split_queue_lock); > > -fail: if (mapping) > > +fail: > > + if (mapping) > > xa_unlock(&mapping->i_pages); > > local_irq_enable(); > > remap_page(head, thp_nr_pages(head)); > > -- > > 2.26.2