From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-5.3 required=3.0 tests=BAYES_00, HEADER_FROM_DIFFERENT_DOMAINS,MAILING_LIST_MULTI,NICE_REPLY_A,SPF_HELO_NONE, SPF_PASS,USER_AGENT_SANE_1 autolearn=no autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id 8E36DC43462 for ; Sat, 15 May 2021 02:07:15 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 0F45C61440 for ; Sat, 15 May 2021 02:07:14 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.3.2 mail.kernel.org 0F45C61440 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=huawei.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=owner-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix) id 3A49E8E0001; Fri, 14 May 2021 22:07:14 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id 37AE78D0002; Fri, 14 May 2021 22:07:14 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 1CCDA8E0001; Fri, 14 May 2021 22:07:14 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0203.hostedemail.com [216.40.44.203]) by kanga.kvack.org (Postfix) with ESMTP id D7C1A8D0002 for ; Fri, 14 May 2021 22:07:13 -0400 (EDT) Received: from smtpin33.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay01.hostedemail.com (Postfix) with ESMTP id 6A88E18034F5C for ; Sat, 15 May 2021 02:07:13 +0000 (UTC) X-FDA: 78141827946.33.E4ADB24 Received: from szxga06-in.huawei.com (szxga06-in.huawei.com [45.249.212.32]) by imf17.hostedemail.com (Postfix) with ESMTP id 8DB2340002D1 for ; Sat, 15 May 2021 02:07:11 +0000 (UTC) Received: from dggems703-chm.china.huawei.com (unknown [172.30.72.58]) by szxga06-in.huawei.com (SkyGuard) with ESMTP id 4Fhpcl200qzldZs; Sat, 15 May 2021 10:04:55 +0800 (CST) Received: from dggpemm500005.china.huawei.com (7.185.36.74) by dggems703-chm.china.huawei.com (10.3.19.180) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256) id 15.1.2176.2; Sat, 15 May 2021 10:07:08 +0800 Received: from [127.0.0.1] (10.69.30.204) by dggpemm500005.china.huawei.com (7.185.36.74) with Microsoft SMTP Server (version=TLS1_2, cipher=TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256) id 15.1.2176.2; Sat, 15 May 2021 10:07:07 +0800 Subject: Re: [PATCH net-next v5 3/5] page_pool: Allow drivers to hint on SKB recycling To: Ilias Apalodimas CC: Matteo Croce , , , Ayush Sawal , "Vinay Kumar Yadav" , Rohit Maheshwari , "David S. Miller" , Jakub Kicinski , Thomas Petazzoni , Marcin Wojtas , Russell King , Mirko Lindner , Stephen Hemminger , "Tariq Toukan" , Jesper Dangaard Brouer , "Alexei Starovoitov" , Daniel Borkmann , "John Fastabend" , Boris Pismenny , Arnd Bergmann , Andrew Morton , "Peter Zijlstra (Intel)" , Vlastimil Babka , Yu Zhao , Will Deacon , Fenghua Yu , Roman Gushchin , Hugh Dickins , Peter Xu , Jason Gunthorpe , Jonathan Lemon , Alexander Lobakin , Cong Wang , wenxu , Kevin Hao , Jakub Sitnicki , Marco Elver , Willem de Bruijn , Miaohe Lin , Guillaume Nault , , , , Matthew Wilcox , Eric Dumazet , David Ahern , Lorenzo Bianconi , Saeed Mahameed , Andrew Lunn , Paolo Abeni , Sven Auhagen References: <20210513165846.23722-1-mcroce@linux.microsoft.com> <20210513165846.23722-4-mcroce@linux.microsoft.com> <798d6dad-7950-91b2-46a5-3535f44df4e2@huawei.com> <212498cf-376b-2dac-e1cd-12c7cc7910c6@huawei.com> From: Yunsheng Lin Message-ID: Date: Sat, 15 May 2021 10:07:06 +0800 User-Agent: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:52.0) Gecko/20100101 Thunderbird/52.2.0 MIME-Version: 1.0 In-Reply-To: Content-Type: text/plain; charset="utf-8" Content-Language: en-US X-Originating-IP: [10.69.30.204] X-ClientProxiedBy: dggeme716-chm.china.huawei.com (10.1.199.112) To dggpemm500005.china.huawei.com (7.185.36.74) X-CFilter-Loop: Reflected X-Rspamd-Queue-Id: 8DB2340002D1 Authentication-Results: imf17.hostedemail.com; dkim=none; spf=pass (imf17.hostedemail.com: domain of linyunsheng@huawei.com designates 45.249.212.32 as permitted sender) smtp.mailfrom=linyunsheng@huawei.com; dmarc=pass (policy=none) header.from=huawei.com X-Rspamd-Server: rspam03 X-Stat-Signature: bbqdiw4yfh5mhrrq5jrnuidussgg6dd5 X-HE-Tag: 1621044431-372173 Content-Transfer-Encoding: quoted-printable X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On 2021/5/14 17:17, Ilias Apalodimas wrote: > On Fri, May 14, 2021 at 04:31:50PM +0800, Yunsheng Lin wrote: >> On 2021/5/14 15:36, Ilias Apalodimas wrote: >>> [...] >>>>> + return false; >>>>> + >>>>> + pp =3D (struct page_pool *)page->pp; >>>>> + >>>>> + /* Driver set this to memory recycling info. Reset it on recycle. >>>>> + * This will *not* work for NIC using a split-page memory model. >>>>> + * The page will be returned to the pool here regardless of the >>>>> + * 'flipped' fragment being in use or not. >>>>> + */ >>>>> + page->pp =3D NULL; >>>> >>>> Why not only clear the page->pp when the page can not be recycled >>>> by the page pool? so that we do not need to set and clear it every >>>> time the page is recycled=E3=80=82 >>>> >>> >>> If the page cannot be recycled, page->pp will not probably be set to = begin >>> with. Since we don't embed the feature in page_pool and we require th= e >>> driver to explicitly enable it, as part of the 'skb flow', I'd rather= keep=20 >>> it as is. When we set/clear the page->pp, the page is probably alrea= dy in=20 >>> cache, so I doubt this will have any measurable impact. >> >> The point is that we already have the skb->pp_recycle to let driver to >> explicitly enable recycling, as part of the 'skb flow, if the page poo= l keep >> the page->pp while it owns the page, then the driver may only need to = call >> one skb_mark_for_recycle() for a skb, instead of call skb_mark_for_rec= ycle() >> for each page frag of a skb. >> >=20 > The driver is meant to call skb_mark_for_recycle for the skb and > page_pool_store_mem_info() for the fragments (in order to store page->p= p). > Nothing bad will happen if you call skb_mark_for_recycle on a frag thou= gh, > but in any case you need to store the page_pool pointer of each frag to > struct page. Right. Nothing bad will happen when we keep the page_pool pointer in page->pp while page pool owns the page too, even if the skb->pp_recycle is not set, right? >=20 >> Maybe we can add a parameter in "struct page_pool_params" to let drive= r >> to decide if the page pool ptr is stored in page->pp while the page po= ol >> owns the page? >=20 > Then you'd have to check the page pool config before saving the meta-da= ta, I am not sure what the "saving the meta-data" meant? > and you would have to make the skb path aware of that as well (I assume= you > mean replace pp_recycle with this?). I meant we could set the in page->pp when the page is allocated from alloc_pages() in __page_pool_alloc_pages_slow() unconditionally or according to a newly add filed in pool->p, and only clear it in page_pool_release_page(), between which the page is owned by page pool, right? > If not and you just want to add an extra flag on page_pool_params and b= e able=20 > to enable recycling depending on that flag, we just add a patch afterwa= rds. > I am not sure we need an extra if for each packet though. In that case, the skb_mark_for_recycle() could only set the skb->pp_recyc= le, but not the pool->p. >=20 >> >> Another thing accured to me is that if the driver use page from the >> page pool to form a skb, and it does not call skb_mark_for_recycle(), >> then there will be resource leaking, right? if yes, it seems the >> skb_mark_for_recycle() call does not seems to add any value? >> >=20 > Not really, the driver has 2 choices: > - call page_pool_release_page() once it receives the payload. That will > clean up dma mappings (if page pool is responsible for them) and free= the > buffer The is only needed before SKB recycling is supported or the driver does n= ot want the SKB recycling support explicitly, right? > - call skb_mark_for_recycle(). Which will end up recycling the buffer. If the driver need to add extra flag to enable recycling based on skb instead of page pool, then adding skb_mark_for_recycle() makes sense to me too, otherwise it seems adding a field in pool->p to recycling based on skb makes more sense? >=20 > If you call none of those, you'd leak a page, but that's a driver bug. > patches [4/5, 5/5] do that for two marvell drivers. > I really want to make drivers opt-in in the feature instead of always > enabling it. >=20 > Thanks > /Ilias >> >>> >>>>> + page_pool_put_full_page(pp, virt_to_head_page(data), false); >>>>> + >>>>> C(end); >>> >>> [...] >> >> >=20 > . >=20