From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-18.0 required=3.0 tests=BAYES_00,DKIMWL_WL_HIGH, DKIM_SIGNED,DKIM_VALID,DKIM_VALID_AU,HEADER_FROM_DIFFERENT_DOMAINS, INCLUDES_CR_TRAILER,INCLUDES_PATCH,MAILING_LIST_MULTI,NICE_REPLY_A, SPF_HELO_NONE,SPF_PASS,URIBL_BLOCKED,USER_AGENT_SANE_1 autolearn=unavailable autolearn_force=no version=3.4.0 Received: from mail.kernel.org (mail.kernel.org [198.145.29.99]) by smtp.lore.kernel.org (Postfix) with ESMTP id BDAB0C4338F for ; Tue, 10 Aug 2021 14:58:19 +0000 (UTC) Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by mail.kernel.org (Postfix) with ESMTP id 4FDDD60F38 for ; Tue, 10 Aug 2021 14:58:19 +0000 (UTC) DMARC-Filter: OpenDMARC Filter v1.4.1 mail.kernel.org 4FDDD60F38 Authentication-Results: mail.kernel.org; dmarc=fail (p=none dis=none) header.from=redhat.com Authentication-Results: mail.kernel.org; spf=pass smtp.mailfrom=kvack.org Received: by kanga.kvack.org (Postfix) id AF7468D000E; Tue, 10 Aug 2021 10:58:18 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id AA7AA8D0001; Tue, 10 Aug 2021 10:58:18 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id 96F3C8D000E; Tue, 10 Aug 2021 10:58:18 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from forelay.hostedemail.com (smtprelay0185.hostedemail.com [216.40.44.185]) by kanga.kvack.org (Postfix) with ESMTP id 7A69B8D0001 for ; Tue, 10 Aug 2021 10:58:18 -0400 (EDT) Received: from smtpin36.hostedemail.com (10.5.19.251.rfc1918.com [10.5.19.251]) by forelay01.hostedemail.com (Postfix) with ESMTP id 274301826CFAD for ; Tue, 10 Aug 2021 14:58:18 +0000 (UTC) X-FDA: 78459476676.36.7CDFCF5 Received: from us-smtp-delivery-124.mimecast.com (us-smtp-delivery-124.mimecast.com [170.10.133.124]) by imf28.hostedemail.com (Postfix) with ESMTP id B9D8B9002F2F for ; Tue, 10 Aug 2021 14:58:17 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1628607497; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=NUy18vQWFqEpKaLHVXakl3mPB08V6UfeF611fK9Gam0=; b=I48BV+0BeVozDnr1x+XG/XWOlMIhXKGN4c+9gDe357BtF5F3lQTZYVLgRrKQdA2n7Nr0a9 MCjaZDatrVtkIqKP05eKB5Byq0f1Nlf648tnuUsjh5v4sJkAdUVbMGr9BWJ9opGzm7CjZe 2kGSLWzsQ8nQmraGKu7a/95LcmgV+rY= Received: from mail-wr1-f72.google.com (mail-wr1-f72.google.com [209.85.221.72]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-566-Unxo5fV-PZCHiqYdLsJQ6w-1; Tue, 10 Aug 2021 10:58:09 -0400 X-MC-Unique: Unxo5fV-PZCHiqYdLsJQ6w-1 Received: by mail-wr1-f72.google.com with SMTP id e17-20020adfa4510000b02901550e0550f3so513990wra.11 for ; Tue, 10 Aug 2021 07:58:09 -0700 (PDT) X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20161025; h=x-gm-message-state:from:cc:subject:to:references:message-id:date :user-agent:mime-version:in-reply-to:content-language :content-transfer-encoding; bh=NUy18vQWFqEpKaLHVXakl3mPB08V6UfeF611fK9Gam0=; b=aT/OHs1u3RWJnc+35vf8rS502H1q39zOkZ47zLgBGnCmeaonSTOZz2yZju0wvhS+ak yoxjTEVesBBDjsu8WvO0wuQ8ODNjhPCmpldpcz63rbhDn/9oILlkBTMVgD9EDXZOznCS XcGDkWi5sVQ14lvFnkmWctehLzQYfP7H28tnInLGsUDWm0xY5F1ouIBfI1rjjHY5L3cV XzMKXAqWSD+x0DPGryHpO43EBBmQbXzZLs2+cm43bAIl649LhbxCfXJdFvMjy1kJwsLc k/LGT6XtdAc9/IIOyo1hR/LkusKTyZ18giFuswqySpWjuz7q00IppsKxV7biJE2nO184 ZUGw== X-Gm-Message-State: AOAM530xscbHGl3kJNGSVQeeXei/i4tX7WDtqhQvA3KR+x3Fej0JTJlG TBuWkQYJC2jFbCkS/ZUbwCZWScKydzPNk8O1f6DSV+RZ8hvA9uNf5BOfqlX8V+NHM0vVu5EOusd L/g9fGN7FRwnPWBpcOy9sLehDNta2+DmfNwoZ1t5sOcnzgv9RY8He6SSmagP3iQ== X-Received: by 2002:a5d:6146:: with SMTP id y6mr30881292wrt.278.1628607488079; Tue, 10 Aug 2021 07:58:08 -0700 (PDT) X-Google-Smtp-Source: ABdhPJxevabhjt95IKMjXAm2nrxG7HgOUYrBPOJkvsPqoKg1v208UM0XWcPZ+fLqKciWDPkHANY1qw== X-Received: by 2002:a5d:6146:: with SMTP id y6mr30881232wrt.278.1628607487862; Tue, 10 Aug 2021 07:58:07 -0700 (PDT) Received: from [192.168.42.238] (3-14-107-185.static.kviknet.dk. [185.107.14.3]) by smtp.gmail.com with ESMTPSA id g5sm2883361wmh.31.2021.08.10.07.58.05 (version=TLS1_3 cipher=TLS_AES_128_GCM_SHA256 bits=128/128); Tue, 10 Aug 2021 07:58:07 -0700 (PDT) From: Jesper Dangaard Brouer X-Google-Original-From: Jesper Dangaard Brouer Cc: brouer@redhat.com, alexander.duyck@gmail.com, linux@armlinux.org.uk, mw@semihalf.com, linuxarm@openeuler.org, yisen.zhuang@huawei.com, salil.mehta@huawei.com, thomas.petazzoni@bootlin.com, hawk@kernel.org, ilias.apalodimas@linaro.org, ast@kernel.org, daniel@iogearbox.net, john.fastabend@gmail.com, akpm@linux-foundation.org, peterz@infradead.org, will@kernel.org, willy@infradead.org, vbabka@suse.cz, fenghua.yu@intel.com, guro@fb.com, peterx@redhat.com, feng.tang@intel.com, jgg@ziepe.ca, mcroce@microsoft.com, hughd@google.com, jonathan.lemon@gmail.com, alobakin@pm.me, willemb@google.com, wenxu@ucloud.cn, cong.wang@bytedance.com, haokexin@gmail.com, nogikh@google.com, elver@google.com, yhs@fb.com, kpsingh@kernel.org, andrii@kernel.org, kafai@fb.com, songliubraving@fb.com, netdev@vger.kernel.org, linux-kernel@vger.kernel.org, bpf@vger.kernel.org, chenhao288@hisilicon.com, Linux-MM Subject: Re: [PATCH net-next v2 2/4] page_pool: add interface to manipulate frag count in page pool To: Yunsheng Lin , davem@davemloft.net, kuba@kernel.org References: <1628217982-53533-1-git-send-email-linyunsheng@huawei.com> <1628217982-53533-3-git-send-email-linyunsheng@huawei.com> Message-ID: Date: Tue, 10 Aug 2021 16:58:05 +0200 User-Agent: Mozilla/5.0 (X11; Linux x86_64; rv:78.0) Gecko/20100101 Thunderbird/78.11.0 MIME-Version: 1.0 In-Reply-To: <1628217982-53533-3-git-send-email-linyunsheng@huawei.com> X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Content-Type: text/plain; charset=utf-8; format=flowed Content-Language: en-US Content-Transfer-Encoding: 7bit X-Rspamd-Server: rspam06 X-Rspamd-Queue-Id: B9D8B9002F2F Authentication-Results: imf28.hostedemail.com; dkim=pass header.d=redhat.com header.s=mimecast20190719 header.b=I48BV+0B; dmarc=pass (policy=none) header.from=redhat.com; spf=none (imf28.hostedemail.com: domain of jbrouer@redhat.com has no SPF policy when checking 170.10.133.124) smtp.mailfrom=jbrouer@redhat.com X-Stat-Signature: 3nhiaxfmo99p9hbhy7sbie7bnbtcaqzu X-HE-Tag: 1628607497-327992 X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: On 06/08/2021 04.46, Yunsheng Lin wrote: > For 32 bit systems with 64 bit dma, dma_addr[1] is used to > store the upper 32 bit dma addr, those system should be rare > those days. > > For normal system, the dma_addr[1] in 'struct page' is not > used, so we can reuse dma_addr[1] for storing frag count, > which means how many frags this page might be splited to. > > In order to simplify the page frag support in the page pool, > the PAGE_POOL_DMA_USE_PP_FRAG_COUNT macro is added to indicate > the 32 bit systems with 64 bit dma, and the page frag support > in page pool is disabled for such system. > > The newly added page_pool_set_frag_count() is called to reserve > the maximum frag count before any page frag is passed to the > user. The page_pool_atomic_sub_frag_count_return() is called > when user is done with the page frag. > > Signed-off-by: Yunsheng Lin > --- > include/linux/mm_types.h | 18 +++++++++++++----- > include/net/page_pool.h | 46 +++++++++++++++++++++++++++++++++++++++------- > net/core/page_pool.c | 4 ++++ > 3 files changed, 56 insertions(+), 12 deletions(-) > > diff --git a/include/linux/mm_types.h b/include/linux/mm_types.h > index 52bbd2b..7f8ee09 100644 > --- a/include/linux/mm_types.h > +++ b/include/linux/mm_types.h > @@ -103,11 +103,19 @@ struct page { > unsigned long pp_magic; > struct page_pool *pp; > unsigned long _pp_mapping_pad; > - /** > - * @dma_addr: might require a 64-bit value on > - * 32-bit architectures. > - */ > - unsigned long dma_addr[2]; > + unsigned long dma_addr; > + union { > + /** > + * dma_addr_upper: might require a 64-bit > + * value on 32-bit architectures. > + */ > + unsigned long dma_addr_upper; > + /** > + * For frag page support, not supported in > + * 32-bit architectures with 64-bit DMA. > + */ > + atomic_long_t pp_frag_count; > + }; > }; > struct { /* slab, slob and slub */ > union { > diff --git a/include/net/page_pool.h b/include/net/page_pool.h > index 8d7744d..42e6997 100644 > --- a/include/net/page_pool.h > +++ b/include/net/page_pool.h > @@ -45,7 +45,10 @@ > * Please note DMA-sync-for-CPU is still > * device driver responsibility > */ > -#define PP_FLAG_ALL (PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV) > +#define PP_FLAG_PAGE_FRAG BIT(2) /* for page frag feature */ > +#define PP_FLAG_ALL (PP_FLAG_DMA_MAP |\ > + PP_FLAG_DMA_SYNC_DEV |\ > + PP_FLAG_PAGE_FRAG) > > /* > * Fast allocation side cache array/stack > @@ -198,19 +201,48 @@ static inline void page_pool_recycle_direct(struct page_pool *pool, > page_pool_put_full_page(pool, page, true); > } > > +#define PAGE_POOL_DMA_USE_PP_FRAG_COUNT \ > + (sizeof(dma_addr_t) > sizeof(unsigned long)) > + > static inline dma_addr_t page_pool_get_dma_addr(struct page *page) > { > - dma_addr_t ret = page->dma_addr[0]; > - if (sizeof(dma_addr_t) > sizeof(unsigned long)) > - ret |= (dma_addr_t)page->dma_addr[1] << 16 << 16; > + dma_addr_t ret = page->dma_addr; > + > + if (PAGE_POOL_DMA_USE_PP_FRAG_COUNT) > + ret |= (dma_addr_t)page->dma_addr_upper << 16 << 16; I find the macro name confusing. I think it would be easier to read the code, if it was called: PAGE_POOL_DMA_CANNOT_USE_PP_FRAG_COUNT > + > return ret; > } > > static inline void page_pool_set_dma_addr(struct page *page, dma_addr_t addr) > { > - page->dma_addr[0] = addr; > - if (sizeof(dma_addr_t) > sizeof(unsigned long)) > - page->dma_addr[1] = upper_32_bits(addr); > + page->dma_addr = addr; > + if (PAGE_POOL_DMA_USE_PP_FRAG_COUNT) > + page->dma_addr_upper = upper_32_bits(addr); > +} > + > +static inline void page_pool_set_frag_count(struct page *page, long nr) > +{ > + atomic_long_set(&page->pp_frag_count, nr); > +} > + > +static inline long page_pool_atomic_sub_frag_count_return(struct page *page, > + long nr) > +{ > + long ret; > + > + /* As suggested by Alexander, atomic_long_read() may cover up the > + * reference count errors, so avoid calling atomic_long_read() in > + * the cases of freeing or draining the page_frags, where we would > + * not expect it to match or that are slowpath anyway. > + */ > + if (__builtin_constant_p(nr) && > + atomic_long_read(&page->pp_frag_count) == nr) > + return 0; > + > + ret = atomic_long_sub_return(nr, &page->pp_frag_count); > + WARN_ON(ret < 0); > + return ret; > } > > static inline bool is_page_pool_compiled_in(void) > diff --git a/net/core/page_pool.c b/net/core/page_pool.c > index 78838c6..68fab94 100644 > --- a/net/core/page_pool.c > +++ b/net/core/page_pool.c > @@ -67,6 +67,10 @@ static int page_pool_init(struct page_pool *pool, > */ > } > > + if (PAGE_POOL_DMA_USE_PP_FRAG_COUNT && > + pool->p.flags & PP_FLAG_PAGE_FRAG) > + return -EINVAL; I read this as: if the page_pool use pp_frag_count and have flag set, then it is invalid/no-allowed, which seems wrong. I find this code more intuitive to read: + if (PAGE_POOL_DMA_CANNOT_USE_PP_FRAG_COUNT && + pool->p.flags & PP_FLAG_PAGE_FRAG) + return -EINVAL; --Jesper