From mboxrd@z Thu Jan 1 00:00:00 1970 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from kanga.kvack.org (kanga.kvack.org [205.233.56.17]) by smtp.lore.kernel.org (Postfix) with ESMTP id E2DB3C36001 for ; Fri, 21 Mar 2025 23:13:34 +0000 (UTC) Received: by kanga.kvack.org (Postfix) id CADFE280002; Fri, 21 Mar 2025 19:13:33 -0400 (EDT) Received: by kanga.kvack.org (Postfix, from userid 40) id C5E92280001; Fri, 21 Mar 2025 19:13:33 -0400 (EDT) X-Delivered-To: int-list-linux-mm@kvack.org Received: by kanga.kvack.org (Postfix, from userid 63042) id AFE69280002; Fri, 21 Mar 2025 19:13:33 -0400 (EDT) X-Delivered-To: linux-mm@kvack.org Received: from relay.hostedemail.com (smtprelay0016.hostedemail.com [216.40.44.16]) by kanga.kvack.org (Postfix) with ESMTP id 916D0280001 for ; Fri, 21 Mar 2025 19:13:33 -0400 (EDT) Received: from smtpin11.hostedemail.com (a10.router.float.18 [10.200.18.1]) by unirelay09.hostedemail.com (Postfix) with ESMTP id 597EB8081A for ; Fri, 21 Mar 2025 23:13:33 +0000 (UTC) X-FDA: 83247111906.11.A459E97 Received: from mail-pl1-f181.google.com (mail-pl1-f181.google.com [209.85.214.181]) by imf26.hostedemail.com (Postfix) with ESMTP id 725E4140009 for ; Fri, 21 Mar 2025 23:13:31 +0000 (UTC) Authentication-Results: imf26.hostedemail.com; dkim=pass header.d=google.com header.s=20230601 header.b=P8HxYGdc; spf=pass (imf26.hostedemail.com: domain of almasrymina@google.com designates 209.85.214.181 as permitted sender) smtp.mailfrom=almasrymina@google.com; dmarc=pass (policy=reject) header.from=google.com ARC-Message-Signature: i=1; a=rsa-sha256; c=relaxed/relaxed; d=hostedemail.com; s=arc-20220608; t=1742598811; h=from:from:sender:reply-to:subject:subject:date:date: message-id:message-id:to:to:cc:cc:mime-version:mime-version: content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references:dkim-signature; bh=FbvnD5haxTLytS47gBGUy0pWnlYckQ3OAvL7U1cwfUA=; b=ydYg4ofUxKFTKbBh/ob399So4LTU7QTPKgw1540gMGfGDXHlEzI7k74CY1O0mMKHVxxWfX XAvJYJ+gXdzh80T8bbqsY86nZ4F5caOITXWvdKuoDg+x1w8m7i45gXRkGs0JTg7Zg+HDCV mQLFu1mYvb6qzUF1b+J5vGaRO5ykKKs= ARC-Seal: i=1; s=arc-20220608; d=hostedemail.com; t=1742598811; a=rsa-sha256; cv=none; b=Bq/HILN1l8bE9zWn3eCXmekMn61olmMxxi1eIGzC7XDvtw8eVzdOcVVX4/nEINnufXjsgh uO48uZmxPL8O5UPxB1Lexf+bf9HHzqcjveXSnqFJMa8NfzoDtYnz+SeDxJnUM/Dk8xziPD rNQRwsO0URaD9e2zAH4A9zMqNKjevuE= ARC-Authentication-Results: i=1; imf26.hostedemail.com; dkim=pass header.d=google.com header.s=20230601 header.b=P8HxYGdc; spf=pass (imf26.hostedemail.com: domain of almasrymina@google.com designates 209.85.214.181 as permitted sender) smtp.mailfrom=almasrymina@google.com; dmarc=pass (policy=reject) header.from=google.com Received: by mail-pl1-f181.google.com with SMTP id d9443c01a7336-2242ac37caeso34335ad.1 for ; Fri, 21 Mar 2025 16:13:31 -0700 (PDT) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=google.com; s=20230601; t=1742598810; x=1743203610; darn=kvack.org; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:from:to:cc:subject:date :message-id:reply-to; bh=FbvnD5haxTLytS47gBGUy0pWnlYckQ3OAvL7U1cwfUA=; b=P8HxYGdcuPq11Ll5LTAKtG4tGrwuozF7JzHzhNhAONVe3Vv/Jyr+RT5VmL9s00NfYE 3VQ3rOO60UnqyffnhPrTjXI5j6Y773UUF08OIZvcsIZzI3myBMjObABWY9dEr3UtacUo 9vvZdWrU7N/VvI44zWj45Hy1fNC0ouwTRgVnGy7fBIM4+DYN2rOs/+h8tbxSNW6bQ6dE EvBOrMeSdIWvBuG25xfluXLcIFZSUpCRnKGWXmAl5gLACi5RWyzNybiu01iyS+Q1WpMZ kHANLkuNTKq3Qb3uKXNhn/wzty9ZkWLwUnFFhVz6PAagQEWdJameAkT5mNXKwX7/i+Cz xfQQ== X-Google-DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=1e100.net; s=20230601; t=1742598810; x=1743203610; h=content-transfer-encoding:cc:to:subject:message-id:date:from :in-reply-to:references:mime-version:x-gm-message-state:from:to:cc :subject:date:message-id:reply-to; bh=FbvnD5haxTLytS47gBGUy0pWnlYckQ3OAvL7U1cwfUA=; b=PGhmjMgipXECnYWDWIOiNn1z8KfadY3Qiy9gHtFnfTcpNC2dLANaIxIWEby8kkegvL vSm55X8yuR8gTelzmJPWDRXDuk1POfgMlYqn1enfu1G9gZc2RCFRCieNqaMfkgjI8n0H hQaAD0ScnxLIe8ROaEQpqD4wrIy3E4fsJHquAtB5J1JvBCBQYwLxviCWFD3DJz1IeGm0 dOJfQMFpLb86+6xQhjmxzxU7Xki9kOQPCLH4I3rtT/oQCI8XUt5GBVlzrz9A7cfMSAdu d6Q54FwGaAnZhDTofDmvzy6V5vLC0+0ZljzWvJUfaFcgPKbj8FwGa6l5DemU7Q0iv9nm rubg== X-Forwarded-Encrypted: i=1; AJvYcCW2guapYO3+QYDSuRca49l5PLaCnzkvib2R79ogltaW+Tj3u42kzihPjPX+4lZnu06OidS3s+d1lg==@kvack.org X-Gm-Message-State: AOJu0YyeMXn8aXTm5w3R9k016j5/UH2BVu3Re8n62drYaATssdMd6Spi UsuNppqiWknnSt0USgNKXpvfjVoIdx+1LIeeBGx2HrGdmBdqSRZKWgbn48Y+Fyh1M/iJybVCAJW cHToZwKfSgI6GlNu8CXPGooRTs3nVJAQ0/2ZU X-Gm-Gg: ASbGncveLVdON1lwnPMPs9F4JYhN1fdgY7AnXhUVeE9fP9274APRApstmYkIx6kQUid ieUzRWxJ53uR7XRHEt9GPRY+bh9D2QszIwdu4J+caXrnEHjZPQ1X1iWf9mjZiKyMWgGAamrWxLn klL4teh8u+NbHEa8vBIg3obqHVqAvtXmnYGr1psUSZlKWrLhXT2JIsFIE= X-Google-Smtp-Source: AGHT+IGJfNj+BTaeUb7CmreHUKO1/d5/Yp0KjN22BdNR7eOnswfJOKqQxGDzzhLFJBXAju2i6SyYA+NISPX0wGE5064= X-Received: by 2002:a17:903:234f:b0:21f:465d:c588 with SMTP id d9443c01a7336-227982bcf17mr1141275ad.14.1742598809765; Fri, 21 Mar 2025 16:13:29 -0700 (PDT) MIME-Version: 1.0 References: <20250314-page-pool-track-dma-v1-0-c212e57a74c2@redhat.com> <20250314-page-pool-track-dma-v1-2-c212e57a74c2@redhat.com> In-Reply-To: <20250314-page-pool-track-dma-v1-2-c212e57a74c2@redhat.com> From: Mina Almasry Date: Fri, 21 Mar 2025 16:13:17 -0700 X-Gm-Features: AQ5f1Jr3pNjB7nwYsq0njhmiKzooW96U1O7ADhvC9S4zs33ypykXN0QQc-_3wro Message-ID: Subject: Re: [PATCH net-next 2/3] page_pool: Turn dma_sync and dma_sync_cpu fields into a bitmap To: =?UTF-8?B?VG9rZSBIw7hpbGFuZC1Kw7hyZ2Vuc2Vu?= Cc: "David S. Miller" , Jakub Kicinski , Jesper Dangaard Brouer , Saeed Mahameed , Leon Romanovsky , Tariq Toukan , Andrew Lunn , Eric Dumazet , Paolo Abeni , Ilias Apalodimas , Simon Horman , Andrew Morton , Yonglong Liu , Yunsheng Lin , Pavel Begunkov , Matthew Wilcox , netdev@vger.kernel.org, bpf@vger.kernel.org, linux-rdma@vger.kernel.org, linux-mm@kvack.org Content-Type: text/plain; charset="UTF-8" Content-Transfer-Encoding: quoted-printable X-Rspam-User: X-Rspamd-Server: rspam10 X-Rspamd-Queue-Id: 725E4140009 X-Stat-Signature: 3d9izznmrc6q95o57wz3hwujroszauhe X-HE-Tag: 1742598811-309461 X-HE-Meta: U2FsdGVkX18sE8QkC9ZDv9dvLVnnSfwPhluvf/VCDxmBWuKt0enrme+K3PxNvFoqFG7rhK5vvtimQcTtWhmNYSd8NOZyyIfbLjPANQcnzx1+5t8Gh+CxL+lkeWBq7hlwPqJJKLCIbaT3u1Uzzo4sp1DONroYzTrR7ZK6mIwZN8c51EF3bELGa2qlXle3ABWrhIu1MILb3kk2MfN3mFHGz2A3G6qrx/qapAE89ujt6LZt/XwMsLq72tE/2KlkDzKJ2i3mhkRw9dMUc7xNILQemJoIxHn3ophU+nvlcUJVl3+tQL9lHfwOswNfU+nB+aCzuZTFrGfB76og+JQDyAm1PRLwfnQzpP199smohs19rMparpgN79YHNefQCs5gd2QW4Rwrffe43wwQ+cIgPfej4vlW/pu1A+keYXEbwfWrtyHYIXq7RGTQcwP6N3268I84VyTeejPNBQiz2c2AeVSIgqkjUl50i/2QxMl+IDQCllDjUSaeJr/R8bzo7+FBerClODmtavmXl0kPffIjZ/nj1jvC/E+zHcC8aBKTxkAh/kfrGyjLXRfATNGQ0C5FiZMsZSlZFlIJlPJsJgX48N0+RI8MDAhe/lEZHJA9dnGtptxLpQDJNWBExxeOs5K3hENVgSPGP5UQXHZU71hoiVtM8E/q/4eXBLgYV4eAVHBYLuQG+e6U+pWCnAJ6ShObSkR52GjXnG7Un/wb6woGU11u5noBezQ7Q8MPXNpshxt99RDJhuDFasZesSo1yS/A7RKR3MC/gnSTrx/8k1YPTm/PMVZo+PrZ1VssthLkZ27D2bciOR+ECKc6qBRIlPovXsneyxSwV69CMg8Vq5WCEB6pU1uKNe8Qwppwad3Sf00mqSCJyYpxFlXWOWnb2JdRmsLgy14Pw5HNE5Dfn78qnYd8BvwLChfrg4kf/mDNDsWglniUfhx3QtU9R4tz1wV/50V2n38O4mwx6/HgKXqxukG NZIFXg3+ rPXykjcHndjhg5IIMvPZygB54CB9B58veLASjcT0teiKv1KiA2xKchA366HyQVi/GXqKZo6SH78YQCdyzEMUB7Crkz0zD4QxX7mfW3EYOvNp/0cDck0FJU1UKyJcFyIaSzGe+wd6fAIetKxXGx9f22v2MDl0UhgZL7uhu1vnpJPvfDNApB5QHHWM3JZn4q1t7N8AcsiZ2Xoqcj+4nVzuAyteDxwz+s892Wf3oCQ1Vka2w3Ibqs7N8zi7tOa8RmPqNgbUqQIehxux9hso1EJXrkL1B9EnWnYld9HWFTgiYBifSC8iuVb+q2HnaKljj3wkFz+09Kz0Lt23yUjjh0kMBEjYQW95GmYRq6UO+AiP2PoMMW83rX3TCsXuFCw== X-Bogosity: Ham, tests=bogofilter, spamicity=0.000000, version=1.2.4 Sender: owner-linux-mm@kvack.org Precedence: bulk X-Loop: owner-majordomo@kvack.org List-ID: List-Subscribe: List-Unsubscribe: On Fri, Mar 14, 2025 at 3:12=E2=80=AFAM Toke H=C3=B8iland-J=C3=B8rgensen wrote: > > Change the single-bit booleans for dma_sync into an unsigned long with > BIT() definitions so that a subsequent patch can write them both with a > singe WRITE_ONCE() on teardown. Also move the check for the sync_cpu > side into __page_pool_dma_sync_for_cpu() so it can be disabled for > non-netmem providers as well. > > Signed-off-by: Toke H=C3=B8iland-J=C3=B8rgensen Reviewed-by: Mina Almasry > --- > include/net/page_pool/helpers.h | 6 +++--- > include/net/page_pool/types.h | 8 ++++++-- > net/core/devmem.c | 3 +-- > net/core/page_pool.c | 9 +++++---- > 4 files changed, 15 insertions(+), 11 deletions(-) > > diff --git a/include/net/page_pool/helpers.h b/include/net/page_pool/help= ers.h > index 582a3d00cbe2315edeb92850b6a42ab21e509e45..7ed32bde4b8944deb7fb22e29= 1e95b8487be681a 100644 > --- a/include/net/page_pool/helpers.h > +++ b/include/net/page_pool/helpers.h > @@ -443,6 +443,9 @@ static inline void __page_pool_dma_sync_for_cpu(const= struct page_pool *pool, > const dma_addr_t dma_addr= , > u32 offset, u32 dma_sync_= size) > { > + if (!(READ_ONCE(pool->dma_sync) & PP_DMA_SYNC_CPU)) > + return; > + > dma_sync_single_range_for_cpu(pool->p.dev, dma_addr, > offset + pool->p.offset, dma_sync_s= ize, > page_pool_get_dma_dir(pool)); > @@ -473,9 +476,6 @@ page_pool_dma_sync_netmem_for_cpu(const struct page_p= ool *pool, > const netmem_ref netmem, u32 offset, > u32 dma_sync_size) > { > - if (!pool->dma_sync_for_cpu) > - return; > - > __page_pool_dma_sync_for_cpu(pool, > page_pool_get_dma_addr_netmem(netmem= ), > offset, dma_sync_size); I think moving the check to __page_pool_dma_sync_for_cpu is fine, but I would have preferred to keep it as-is actually. I think if we're syncing netmem we should check dma_sync_for_cpu, because the netmem may not be dma-syncable. But for pages, they will likely always be dma-syncable. Some driver may have opted to do a perf optimizations by calling __page_pool_dma_sync_for_cpu on a dma-addr that it knows came from a page to save some cycles of netmem checking. > diff --git a/include/net/page_pool/types.h b/include/net/page_pool/types.= h > index df0d3c1608929605224feb26173135ff37951ef8..fbe34024b20061e8bcd1d4474= f6ebfc70992f1eb 100644 > --- a/include/net/page_pool/types.h > +++ b/include/net/page_pool/types.h > @@ -33,6 +33,10 @@ > #define PP_FLAG_ALL (PP_FLAG_DMA_MAP | PP_FLAG_DMA_SYNC_DEV |= \ > PP_FLAG_SYSTEM_POOL | PP_FLAG_ALLOW_UNRE= ADABLE_NETMEM) > > +/* bit values used in pp->dma_sync */ > +#define PP_DMA_SYNC_DEV BIT(0) > +#define PP_DMA_SYNC_CPU BIT(1) > + > /* > * Fast allocation side cache array/stack > * > @@ -175,12 +179,12 @@ struct page_pool { > > bool has_init_callback:1; /* slow::init_callback is set */ > bool dma_map:1; /* Perform DMA mapping */ > - bool dma_sync:1; /* Perform DMA sync for device */ > - bool dma_sync_for_cpu:1; /* Perform DMA sync for cpu */ > #ifdef CONFIG_PAGE_POOL_STATS > bool system:1; /* This is a global percpu pool *= / > #endif > > + unsigned long dma_sync; > + > __cacheline_group_begin_aligned(frag, PAGE_POOL_FRAG_GROUP_ALIGN)= ; > long frag_users; > netmem_ref frag_page; > diff --git a/net/core/devmem.c b/net/core/devmem.c > index 7c6e0b5b6acb55f376ec725dfb71d1f70a4320c3..16e43752566feb510b3e47fbe= c2d8da0f26a6adc 100644 > --- a/net/core/devmem.c > +++ b/net/core/devmem.c > @@ -337,8 +337,7 @@ int mp_dmabuf_devmem_init(struct page_pool *pool) > /* dma-buf dma addresses do not need and should not be used with > * dma_sync_for_cpu/device. Force disable dma_sync. > */ > - pool->dma_sync =3D false; > - pool->dma_sync_for_cpu =3D false; > + pool->dma_sync =3D 0; > > if (pool->p.order !=3D 0) > return -E2BIG; > diff --git a/net/core/page_pool.c b/net/core/page_pool.c > index acef1fcd8ddcfd1853a6f2055c1f1820ab248e8d..d51ca4389dd62d8bc266a9a2b= 792838257173535 100644 > --- a/net/core/page_pool.c > +++ b/net/core/page_pool.c > @@ -203,7 +203,7 @@ static int page_pool_init(struct page_pool *pool, > memcpy(&pool->slow, ¶ms->slow, sizeof(pool->slow)); > > pool->cpuid =3D cpuid; > - pool->dma_sync_for_cpu =3D true; > + pool->dma_sync =3D PP_DMA_SYNC_CPU; > More pedantically this should have been pool->dma_sync |=3D PP_DMA_SYNC_CPU, but it doesn't matter since this variable is 0 initialized I think. -- Thanks, Mina