From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id E7CFEC7618D for ; Tue, 4 Apr 2023 07:07:13 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id 302726B0074; Tue, 4 Apr 2023 03:07:13 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 2B3616B0075; Tue, 4 Apr 2023 03:07:13 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 154866B0078; Tue, 4 Apr 2023 03:07:13 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0013.hostedemail.com [216.40.44.13]) by kanga.kvack.org (Postfix) with ESMTP id 04EE56B0074 for ; Tue, 4 Apr 2023 03:07:13 -0400 (EDT) Received: from smtpin03.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay01.hostedemail.com (Postfix) with ESMTP id C5CF51C6364 for ; Tue, 4 Apr 2023 07:07:12 +0000 (UTC) X-FDA: 80642827104.03.83B45CB Received: from mail-pg1-f172.google.com (mail-pg1-f172.google.com [209.85.215.172]) by imf08.hostedemail.com (Postfix) with ESMTP id E6117160017 for ; Tue, 4 Apr 2023 07:07:08 +0000 (UTC) Authentication-Results: imf08.hostedemail.com; dkim=pass header.d=bytedance.com header.s=google header.b=VQwBgBsk; spf=pass (imf08.hostedemail.com: domain of zhengqi.arch@bytedance.com designates 209.85.215.172 as permitted sender) smtp.mailfrom=zhengqi.arch@bytedance.com; dmarc=pass (policy=quarantine) header.from=bytedance.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1680592029; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=twOPbfkYHULFT9ytXIMVRlWTWlhqUczKwRzl0UgT8FE=; b=8XCSA0N4QjVxJOwmEPnZ2Brk3UN9xgG7UEuW8izmCjgRcdQuOCGvakLuHuRP/7Shg4kKor G1J/7x2Ic1FjIIgrdKW8qwogjkb+/wsBJkvT7PglBvtTVkecey3xWwL/rlaDeqijh/9/wF h/J1vzkS9315fX85OW58a9eBAmqGT3g= ARC-Authentication-Results: i=1; imf08.hostedemail.com; dkim=pass header.d=bytedance.com header.s=google header.b=VQwBgBsk; spf=pass (imf08.hostedemail.com: domain of zhengqi.arch@bytedance.com designates 209.85.215.172 as permitted sender) smtp.mailfrom=zhengqi.arch@bytedance.com; dmarc=pass (policy=quarantine) header.from=bytedance.com ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1680592029; a=rsa-sha256; cv=none; b=eyFMJsLkT/ZZMO/dzAL3rFn35Vy6pmSiOgE6t/qCadkcPHxVbXGSzBVBAwJJpXnFOJtwY6 VIpvLdNRzAr6X3PkI+O5N7+NOfud6+abAz2zZp8q/8bQrIeDLxuq8kzTzfJZ6BST13uIny AZ6F9FhQly3FLIhoHWK4HeEW49b7C0U= Received: by mail-pg1-f172.google.com with SMTP id d22so19027818pgw.2 for ; Tue, 04 Apr 2023 00:07:08 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=bytedance.com; s=google; t=1680592027; h=content-transfer-encoding:in-reply-to:references:cc:to:from :content-language:subject:user-agent:mime-version:date:message-id :from:to:cc:subject:date:message-id:reply-to; bh=twOPbfkYHULFT9ytXIMVRlWTWlhqUczKwRzl0UgT8FE=; b=VQwBgBskPIZsqNPq6QbEJQZIdmsWzqK8sOopOPe7FaLP+h0sxBo4nNSx6dHTzPuZHv fHz8ssbV+rRl7OmLUAXeXmFW2GKW7cuFsPu9bmrHmTKyzmaRn/GeTkD85RwTchhCMRKc ibmrfM20x4KzUQquHWHqG1sWs06dLlpldt0lb8rmo0S1O7l5OseeeO0dh4STd/PYbVZJ xczXHl3HVTycHtz/HdjJ+uQRiTdgTs/gtCgAxyw/4rPg7DhMm+zH+8ww/bGPeXQrf1HV uENkNM5ZoIaF1NTqM9ExALHIiqFENRxZ5U9iHEsaeL9ElyBrP92bjOoMkGoD4HBxG5G2 Ohvg== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20210112; t=1680592027; h=content-transfer-encoding:in-reply-to:references:cc:to:from :content-language:subject:user-agent:mime-version:date:message-id :x-gm-message-state:from:to:cc:subject:date:message-id:reply-to; bh=twOPbfkYHULFT9ytXIMVRlWTWlhqUczKwRzl0UgT8FE=; b=iMl1tZ4AluMBvBctq6lEKKEIbledS8T6Vul+r0Tg4nGSFlIf4/b4z6ns4RCa9avDA0 uXc6DJq4iN9eG+c+/qZC2c5DL51pagoX5YfNekYqi+Gh5X20bWptSRELt7dtYjGYrh1o vrVWDC0wOsVRr8BCYpxmg2ilvcJR/vMbCpos+7T9yxDyiX26f+aVh4+h1luM8kM7N+9c PeUysB/phD0TY3wpdEqExRmB/S5GrGY8eVzu3QzT6C3tHxjLKPAKjdqNnu9KHT0/rd5e gmTCy3xiwyjufSJrkCPmNhbMlUstsRtR60C5jM/fnAn/LuOEBdkv/Wk0XbZnFb0S9t5Y TKWg== X-Gm-Message-State: AAQBX9fw7unf5jLLoP805JOEgRbO8WELoXJcmwIsm7A6GN3WhkT1py8B L+UQ2xqrJncFEuU8sPCeZttPDQ== X-Google-Smtp-Source: AKy350Z0Dfa0bt8qwppuCspk5hwN33zGD9tFwNW1XiTonZDE6XxLhPnZ2ERAiMnwF3LWvwNCVoU1gg== X-Received: by 2002:a05:6a00:2294:b0:62d:dd14:bed1 with SMTP id f20-20020a056a00229400b0062ddd14bed1mr1579121pfe.1.1680592027328; Tue, 04 Apr 2023 00:07:07 -0700 (PDT) Received: from [10.200.9.56] ([139.177.225.248]) by smtp.gmail.com with ESMTPSA id c22-20020aa78816000000b00627e55f383dsm8055551pfo.3.2023.04.04.00.07.04 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Tue, 04 Apr 2023 00:07:06 -0700 (PDT) Message-ID: Date: Tue, 4 Apr 2023 15:07:02 +0800 MIME-Version: 1.0 User-Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.15; rv:102.0) Gecko/20100101 Thunderbird/102.9.1 Subject: Re: [PATCH 1/2] mm: swap: use folio_batch_reinit() in folio_batch_move_lru() Content-Language: en-US From: Qi Zheng To: Andrew Morton , Mel Gorman Cc: willy@infradead.org, lstoakes@gmail.com, linux-mm@kvack.org, linux-kernel@vger.kernel.org References: <20230331095858.51810-1-zhengqi.arch@bytedance.com> <20230331150430.546de954b0a7918f042c562e@linux-foundation.org> <3bd45269-557d-db5a-2224-3d960ba3159d@bytedance.com> In-Reply-To: <3bd45269-557d-db5a-2224-3d960ba3159d@bytedance.com> Content-Type: text/plain; charset=UTF-8; format=flowed Content-Transfer-Encoding: 8bit X-Stat-Signature: syqb1dokzsdfjaud7aswukm5nskawiae X-Rspam-User: X-Rspamd-Queue-Id: E6117160017 X-Rspamd-Server: rspam06 X-HE-Tag: 1680592028-603224 X-HE-Meta: U2FsdGVkX19HdBbRbd9tt63BoCKmh/kbCrSeV/KQ2lvJgK0YaOPNK/7x1JvyiZQfE4lXXIi8qtf4riPUp9EiQNA92VRfb8/m7HVb+yF4q2dtIgLrLo7pzq66uKwThIIIwD/Xm4UkycM7399KptTduhYFAu1ghkotoPNz3h+Jcd5rA1EtgP+ssoNEsn8elje0bVF1kItQkIGjA9uJtRQvBc0M0fSLgf6CdLz4VhlO+AjV3ZO8bYtKoMVASJDWkWKQKqaH6bOvNUR3bSykqELZGRu9rvx6YiQCfhNS4whDoRIOxTq8FMWYVShjJ4UPZSh3BxCMqM2iznBJd3HDp5uGStP1wkr55cfWpCHfhdbnZ9jUXmPyq3dR7+hvSscEip4v04reJ238u0w3MJgwadXkxVmZQmgmph01L0hBDIkeYmPP27IyH9xbPpi+jjocGPoII3g7MUuDCcLoyCYUCBo0M7ORgw2qKzmTT0kazUnNJAOseCwAE4DRGI4ab3ntxf7hYFg1ALiohNj1nG8JAukK3LemdXv5ZA6Y0fLnfHE5ExK2jcCrNiFtdgA4Jve16RDOSr7JlvTkLcHYqguH0Wv0BmxXXnyuV9g+tUj3vfeBBrsaPVxC4k99hHHiAGsLI72bNksWVMPrD7mfBjV8NXFWCaNd1if1cK+AJL8AHc69ODbcWbfkBcvXbhmir5PiLquQ442X0gX4QLr/6fNTUhYy+ym57speOmlesN7WDrErOeiLwZ/0rIyDyOgYoDgD/lEV/HlOQy63no+JdDyoEDMKjuAYQPjH8Wr+ASImmQHZaTCXkKvBYvDkjGM+Tc0gMcyT1d5tG52NgFwSnTRa90bkRQXfesaapB0yLBTwOgseMqFTIJltXp0hb69gEInD5hPfGzWDDB0tMU7PdJZTJ1BLJPBZ7W89g5h3CiEVVsinfGSm7CejSt82sRgiw+cC/t0+qehJdXVgwENKtVh2lK4 pY0p7TOz Oyh3NKueI75WZgDd4mN/uU82jRqJfPew3/VUAUk2tswUjpyE/CiD9EQEPMRVKa6BOUn+mQnlDqQndH1A8Ez/3GlbgnERM3ouB838K2zeaVSV+pGWeM6PBGWgjsLVf2oPhcxTEJp1o7PQlwrpT5kDmyI69cIfFDhfe7dtUukizb/d40FmktKD5KnP30XsJ91e5GoS4YaX7Ui/aGSm3E/sKMh/BcNuNW+KZ7BSMfNUkoOFiWEPNd63bbNOaT06ebGzbVpwqkJenT5QPSdO4AIJznWdNzmPVziIbAjo29U5h3OA7qJdSXf1N7AvoSLhuui9CRp4ySsG9qc4TBhuGtHClAT1kthbr5Q0C3dysnRLjUkYvh2Ojrp6E/AhCqojyKOkHtbRMPoCcUvZO76uhW8HV+eiTLZJ4w1kNeQJE5AvGqvmMPo1cmJmBnI0XT3MZ4rQ0yZmYi6G6dd/pPcrpVecws0UFzqNc5xI20/NBteD9fPqALqQ= X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On 2023/4/2 21:36, Qi Zheng wrote: > Hi Andrew, > > On 2023/4/1 06:04, Andrew Morton wrote: >> On Fri, 31 Mar 2023 17:58:57 +0800 Qi Zheng >> wrote: >> >>> In folio_batch_move_lru(), the folio_batch is not freshly >>> initialised, so it should call folio_batch_reinit() as >>> pagevec_lru_move_fn() did before. >>> >>> ... >>> >>> --- a/mm/swap.c >>> +++ b/mm/swap.c >>> @@ -222,7 +222,7 @@ static void folio_batch_move_lru(struct >>> folio_batch *fbatch, move_fn_t move_fn) >>>       if (lruvec) >>>           unlock_page_lruvec_irqrestore(lruvec, flags); >>>       folios_put(fbatch->folios, folio_batch_count(fbatch)); >>> -    folio_batch_init(fbatch); >>> +    folio_batch_reinit(fbatch); >>>   } >>>   static void folio_batch_add_and_move(struct folio_batch *fbatch, >> >> Well...  why?  This could leave the kernel falsely thinking that the >> folio's pages have been drained from the per-cpu LRU addition >> magazines. >> >> Maybe that's desirable, maybe not, but I think this change needs much >> much more explanation describing why it is beneficial. >> >> >> folio_batch_reinit() seems to be a custom thing for the mlock code - >> perhaps it just shouldn't exist, and its operation should instead be >> open-coded in mlock_folio_batch(). > > The folio_batch_reinit() corresponds to pagevec_reinit(), > the pagevec_reinit() was originally used in pagevec_lru_move_fn() > and mlock_pagevec(), not a custom thing for the mlock code. > > > The commit c2bc16817aa0 ("mm/swap: add folio_batch_move_lru()") > introduces folio_batch_move_lru() to replace pagevec_lru_move_fn(), > but calls folio_batch_init() (corresponding to pagevec_init()) instead > of folio_batch_reinit() (corresponding to pagevec_reinit()). This > change was not explained in the commit message and seems like an > oversight. > >> >> >> The dynamics and rules around ->percpu_pvec_drained are a bit >> mysterious.  A code comment which explains all of this would be >> useful. > > The commit d9ed0d08b6c6 ("mm: only drain per-cpu pagevecs once per > pagevec usage") originally introduced the ->drained (which was later > renamed to ->percpu_pvec_drained by commit 7f0b5fb953e7), which is > intended to drain per-cpu pagevecs only once per pagevec usage. > > Maybe it would be better to add the following code comment: > > diff --git a/mm/swap.c b/mm/swap.c > index 423199ee8478..107c4a13e476 100644 > --- a/mm/swap.c > +++ b/mm/swap.c > @@ -1055,6 +1055,7 @@ EXPORT_SYMBOL(release_pages); >   */ >  void __pagevec_release(struct pagevec *pvec) >  { > +       /* Only drain per-cpu pagevecs once per pagevec usage */ >         if (!pvec->percpu_pvec_drained) { >                 lru_add_drain(); >                 pvec->percpu_pvec_drained = true; > > Please let me know if I missed something. Maybe the commit message can be modified as follows: ``` The ->percpu_pvec_drained was originally introduced by commit d9ed0d08b6c6 ("mm: only drain per-cpu pagevecs once per pagevec usage") to drain per-cpu pagevecs only once per pagevec usage. But after commit c2bc16817aa0 ("mm/swap: add folio_batch_move_lru()"), the ->percpu_pvec_drained will be reset to false by calling folio_batch_init() in folio_batch_move_lru(), which may cause per-cpu pagevecs to be drained multiple times per pagevec usage. This is not what we expected, let's use folio_batch_reinit() in folio_batch_move_lru() to fix it. ``` Also +CC Mel Gorman to confirm this. :) Thanks, Qi > > Thanks, > Qi > >> > > -- Thanks, Qi